hexsha
stringlengths
40
40
size
int64
6
14.9M
ext
stringclasses
1 value
lang
stringclasses
1 value
max_stars_repo_path
stringlengths
6
260
max_stars_repo_name
stringlengths
6
119
max_stars_repo_head_hexsha
stringlengths
40
41
max_stars_repo_licenses
sequence
max_stars_count
int64
1
191k
max_stars_repo_stars_event_min_datetime
stringlengths
24
24
max_stars_repo_stars_event_max_datetime
stringlengths
24
24
max_issues_repo_path
stringlengths
6
260
max_issues_repo_name
stringlengths
6
119
max_issues_repo_head_hexsha
stringlengths
40
41
max_issues_repo_licenses
sequence
max_issues_count
int64
1
67k
max_issues_repo_issues_event_min_datetime
stringlengths
24
24
max_issues_repo_issues_event_max_datetime
stringlengths
24
24
max_forks_repo_path
stringlengths
6
260
max_forks_repo_name
stringlengths
6
119
max_forks_repo_head_hexsha
stringlengths
40
41
max_forks_repo_licenses
sequence
max_forks_count
int64
1
105k
max_forks_repo_forks_event_min_datetime
stringlengths
24
24
max_forks_repo_forks_event_max_datetime
stringlengths
24
24
avg_line_length
float64
2
1.04M
max_line_length
int64
2
11.2M
alphanum_fraction
float64
0
1
cells
sequence
cell_types
sequence
cell_type_groups
sequence
ece7b2aa4c7d1b43c991a2f12cc583b2acb823f0
68,021
ipynb
Jupyter Notebook
jupyter/pandas-tut/10-Reading and Writing Data.ipynb
Sesughter01/ds-march-univel
fbed9e39abf08f7ffdf01d491ce3d898fafaed12
[ "MIT" ]
null
null
null
jupyter/pandas-tut/10-Reading and Writing Data.ipynb
Sesughter01/ds-march-univel
fbed9e39abf08f7ffdf01d491ce3d898fafaed12
[ "MIT" ]
null
null
null
jupyter/pandas-tut/10-Reading and Writing Data.ipynb
Sesughter01/ds-march-univel
fbed9e39abf08f7ffdf01d491ce3d898fafaed12
[ "MIT" ]
2
2022-03-03T15:03:14.000Z
2022-03-15T11:42:24.000Z
28.798052
160
0.337543
[ [ [ "# Tutorial 10: Data Reading and Writting", "_____no_output_____" ] ], [ [ "import pandas as pd \nimport numpy as np ", "_____no_output_____" ], [ "df=pd.read_table(\"DataSets/data.txt\")\ndf", "_____no_output_____" ], [ "df2=pd.read_table(\"DataSets/data2.txt\")\ndf2", "_____no_output_____" ], [ "df2=pd.read_table(\"Data/data2.txt\", sep=\",\")\ndf2", "_____no_output_____" ], [ "df=pd.read_table(\"Data/data2.txt\", sep=\",\")\ndf", "_____no_output_____" ], [ "df=pd.read_table(\"DataSets/data2.txt\",\n sep=\",\",\n names=['name', \"number\", \"sex\"])\ndf", "_____no_output_____" ], [ "df=pd.read_table(\"Data/data2.txt\",\n sep=\",\",\n header=None, \n names=[\"name\",\"score\",\"sex\"])\ndf", "_____no_output_____" ], [ "df=pd.read_table(\"Data/data2.txt\",\n sep=\",\",header=None, \n names=[\"name\",\"score\",\"sex\"],\n index_col=\"name\")\ndf", "_____no_output_____" ], [ "df2=pd.read_table(\"Data/data3.txt\",\n sep=\",\")\ndf2", "_____no_output_____" ], [ "df2=pd.read_table(\"Data/data3.txt\",\n sep=\",\",\n index_col=[\"lesson\",\"name\"])\ndf2", "_____no_output_____" ], [ "df3=pd.read_table(\"Data/data4.txt\",sep=\",\")\ndf3", "_____no_output_____" ], [ "df3=pd.read_table(\"Data/data4.txt\",\n sep=\",\", \n skiprows=[0,2])\ndf3", "_____no_output_____" ], [ "df3=pd.read_table(\"Data/data4.txt\",\n sep=\",\", \n skiprows=[0,2],\n usecols=[0,1])\ndf3", "_____no_output_____" ], [ "df3=pd.read_table(\"Data/data4.txt\",\n sep=\",\", \n skiprows=[0,2],\n usecols=[0,1],\n nrows=3)\ndf3", "_____no_output_____" ] ], [ [ "## Writing Data", "_____no_output_____" ] ], [ [ "df=pd.read_csv(\"DataSets/supermarket_sales - Sheet1.csv\")", "_____no_output_____" ], [ "mask = df[\"Gender\"] == \"Male\"\nmale_sales = df[mask]\n\nmask = df[\"Gender\"] == \"Female\"\nfemale_sales = df[mask]", "_____no_output_____" ], [ "##Write male and female sales individualy to csv\nmale_sales.to_csv(\"DataSets/male_sales.csv\")\nfemale_sales.to_csv(\"DataSets/female_sales.csv\")", "_____no_output_____" ], [ "df=pd.read_csv(\"Data/data.txt\",sep=\"\\t\")\ndf", "_____no_output_____" ], [ "df.to_csv(\"Data/new_data.csv\")", "_____no_output_____" ], [ "# Read excel", "_____no_output_____" ], [ "female_df = pd.read_excel(\"DataSets/m_f_sales.xlsx\", sheet_name = \"female_sales\")\nmale_df = pd.read_excel(\"DataSets/m_f_sales.xlsx\", sheet_name = \"male_sales\")", "_____no_output_____" ], [ "branch_a_mask = female_df[\"Branch\"] == \"A\"\nbranch_b_mask = female_df[\"Branch\"] == \"B\"\n\nbranch_a = female_df[branch_a_mask]\nbranch_b = female_df[branch_b_mask]", "_____no_output_____" ], [ "with pd.ExcelWriter('branches.xlsx', mode='w') as writer: \n \n branch_a.to_excel(writer, sheet_name='Branch A')\n branch_b.to_excel(writer, sheet_name='Branch B')", "_____no_output_____" ], [ "male_df", "_____no_output_____" ], [ "#Automate process getting all different unique Product lines in different sheets", "_____no_output_____" ], [ "female_df = pd.read_excel(\"DataSets/m_f_sales.xlsx\", sheet_name = \"female_sales\")\nmale_df = pd.read_excel(\"DataSets/m_f_sales.xlsx\", sheet_name = \"male_sales\")", "_____no_output_____" ], [ "# Get Unique Product lines", "_____no_output_____" ], [ "female_df[\"Product line\"].unique()", "_____no_output_____" ], [ "male_df[\"Product line\"].unique()", "_____no_output_____" ], [ "unique_products_lines = female_df[\"Product line\"].unique()\n\nfor prod_lin in unique_products_lines:\n \n print(prod_lin)", "Health and beauty\nElectronic accessories\nHome and lifestyle\nFood and beverages\nFashion accessories\nSports and travel\n" ], [ "unique_products_lines = female_df[\"Product line\"].unique()\n\nwith pd.ExcelWriter('prod_lines.xlsx', mode='w') as writer: \n \n for prod_lin_from_forloop in unique_products_lines:\n \n prod_line_mask = female_df[\"Product line\"] == prod_lin_from_forloop\n\n prod_line_df = female_df[prod_line_mask]\n prod_line_df.to_excel(writer, sheet_name=prod_lin_from_forloop)", "_____no_output_____" ], [ "female_df.describe()", "_____no_output_____" ], [ "male_df.describe()", "_____no_output_____" ], [ "female_df.describe(include=[\"O\"])", "_____no_output_____" ], [ "male_df.describe(include=[\"O\"])", "_____no_output_____" ] ], [ [ "Don't forget to follow us on our Tirendaz Academy YouTube channel http://youtube.com/c/tirendazakademi and Medium page (http://tirendazacademy.medium.com)", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown" ]
[ [ "markdown" ], [ "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code" ], [ "markdown" ] ]
ece7b60108c6cb16545dd1ff2299fb6a86e8593b
101,611
ipynb
Jupyter Notebook
samples/shapes/train_shapes.ipynb
mksuns/maskrcnn
8daae5c1f0d1897cfc3be91c39f4d129a5f86eeb
[ "MIT" ]
null
null
null
samples/shapes/train_shapes.ipynb
mksuns/maskrcnn
8daae5c1f0d1897cfc3be91c39f4d129a5f86eeb
[ "MIT" ]
null
null
null
samples/shapes/train_shapes.ipynb
mksuns/maskrcnn
8daae5c1f0d1897cfc3be91c39f4d129a5f86eeb
[ "MIT" ]
null
null
null
97.985535
23,858
0.786322
[ [ [ "# Mask R-CNN - Train on Shapes Dataset\n\n\nThis notebook shows how to train Mask R-CNN on your own dataset. To keep things simple we use a synthetic dataset of shapes (squares, triangles, and circles) which enables fast training. You'd still need a GPU, though, because the network backbone is a Resnet101, which would be too slow to train on a CPU. On a GPU, you can start to get okay-ish results in a few minutes, and good results in less than an hour.\n\nThe code of the *Shapes* dataset is included below. It generates images on the fly, so it doesn't require downloading any data. And it can generate images of any size, so we pick a small image size to train faster. ", "_____no_output_____" ] ], [ [ "import os\nimport sys\nimport random\nimport math\nimport re\nimport time\nimport numpy as np\nimport cv2\nimport matplotlib\nimport matplotlib.pyplot as plt\n\n# Root directory of the project\nROOT_DIR = os.path.abspath(\"../../\")\n\n# Import Mask RCNN\nsys.path.append(ROOT_DIR) # To find local version of the library\nfrom mrcnn.config import Config\nfrom mrcnn import utils\nimport mrcnn.model as modellib\nfrom mrcnn import visualize\nfrom mrcnn.model import log\n\n%matplotlib inline \n\n# Directory to save logs and trained model\nMODEL_DIR = os.path.join(ROOT_DIR, \"logs\")\n\n# Local path to trained weights file\nCOCO_MODEL_PATH = os.path.join(ROOT_DIR, \"mask_rcnn_coco.h5\")\n# Download COCO trained weights from Releases if needed\nif not os.path.exists(COCO_MODEL_PATH):\n utils.download_trained_weights(COCO_MODEL_PATH)", "Using TensorFlow backend.\n" ] ], [ [ "## Configurations", "_____no_output_____" ] ], [ [ "class ShapesConfig(Config):\n \"\"\"Configuration for training on the toy shapes dataset.\n Derives from the base Config class and overrides values specific\n to the toy shapes dataset.\n \"\"\"\n # Give the configuration a recognizable name\n NAME = \"shapes\"\n\n # Train on 1 GPU and 8 images per GPU. We can put multiple images on each\n # GPU because the images are small. Batch size is 8 (GPUs * images/GPU).\n GPU_COUNT = 1\n IMAGES_PER_GPU = 8\n\n # Number of classes (including background)\n NUM_CLASSES = 1 + 3 # background + 3 shapes\n\n # Use small images for faster training. Set the limits of the small side\n # the large side, and that determines the image shape.\n IMAGE_MIN_DIM = 128\n IMAGE_MAX_DIM = 128\n\n # Use smaller anchors because our image and objects are small\n RPN_ANCHOR_SCALES = (8, 16, 32, 64, 128) # anchor side in pixels\n\n # Reduce training ROIs per image because the images are small and have\n # few objects. Aim to allow ROI sampling to pick 33% positive ROIs.\n TRAIN_ROIS_PER_IMAGE = 32\n\n # Use a small epoch since the data is simple\n STEPS_PER_EPOCH = 100\n\n # use small validation steps since the epoch is small\n VALIDATION_STEPS = 5\n \nconfig = ShapesConfig()\nconfig.display()", "\nConfigurations:\nBACKBONE_SHAPES [[32 32]\n [16 16]\n [ 8 8]\n [ 4 4]\n [ 2 2]]\nBACKBONE_STRIDES [4, 8, 16, 32, 64]\nBATCH_SIZE 8\nBBOX_STD_DEV [ 0.1 0.1 0.2 0.2]\nDETECTION_MAX_INSTANCES 100\nDETECTION_MIN_CONFIDENCE 0.5\nDETECTION_NMS_THRESHOLD 0.3\nGPU_COUNT 1\nIMAGES_PER_GPU 8\nIMAGE_MAX_DIM 128\nIMAGE_MIN_DIM 128\nIMAGE_PADDING True\nIMAGE_SHAPE [128 128 3]\nLEARNING_MOMENTUM 0.9\nLEARNING_RATE 0.002\nMASK_POOL_SIZE 14\nMASK_SHAPE [28, 28]\nMAX_GT_INSTANCES 100\nMEAN_PIXEL [ 123.7 116.8 103.9]\nMINI_MASK_SHAPE (56, 56)\nNAME SHAPES\nNUM_CLASSES 4\nPOOL_SIZE 7\nPOST_NMS_ROIS_INFERENCE 1000\nPOST_NMS_ROIS_TRAINING 2000\nROI_POSITIVE_RATIO 0.33\nRPN_ANCHOR_RATIOS [0.5, 1, 2]\nRPN_ANCHOR_SCALES (8, 16, 32, 64, 128)\nRPN_ANCHOR_STRIDE 2\nRPN_BBOX_STD_DEV [ 0.1 0.1 0.2 0.2]\nRPN_TRAIN_ANCHORS_PER_IMAGE 256\nSTEPS_PER_EPOCH 100\nTRAIN_ROIS_PER_IMAGE 32\nUSE_MINI_MASK True\nUSE_RPN_ROIS True\nVALIDATION_STEPS 50\nWEIGHT_DECAY 0.0001\n\n\n" ] ], [ [ "## Notebook Preferences", "_____no_output_____" ] ], [ [ "def get_ax(rows=1, cols=1, size=8):\n \"\"\"Return a Matplotlib Axes array to be used in\n all visualizations in the notebook. Provide a\n central point to control graph sizes.\n \n Change the default size attribute to control the size\n of rendered images\n \"\"\"\n _, ax = plt.subplots(rows, cols, figsize=(size*cols, size*rows))\n return ax", "_____no_output_____" ] ], [ [ "## Dataset\n\nCreate a synthetic dataset\n\nExtend the Dataset class and add a method to load the shapes dataset, `load_shapes()`, and override the following methods:\n\n* load_image()\n* load_mask()\n* image_reference()", "_____no_output_____" ] ], [ [ "class ShapesDataset(utils.Dataset):\n \"\"\"Generates the shapes synthetic dataset. The dataset consists of simple\n shapes (triangles, squares, circles) placed randomly on a blank surface.\n The images are generated on the fly. No file access required.\n \"\"\"\n\n def load_shapes(self, count, height, width):\n \"\"\"Generate the requested number of synthetic images.\n count: number of images to generate.\n height, width: the size of the generated images.\n \"\"\"\n # Add classes\n self.add_class(\"shapes\", 1, \"square\")\n self.add_class(\"shapes\", 2, \"circle\")\n self.add_class(\"shapes\", 3, \"triangle\")\n\n # Add images\n # Generate random specifications of images (i.e. color and\n # list of shapes sizes and locations). This is more compact than\n # actual images. Images are generated on the fly in load_image().\n for i in range(count):\n bg_color, shapes = self.random_image(height, width)\n self.add_image(\"shapes\", image_id=i, path=None,\n width=width, height=height,\n bg_color=bg_color, shapes=shapes)\n\n def load_image(self, image_id):\n \"\"\"Generate an image from the specs of the given image ID.\n Typically this function loads the image from a file, but\n in this case it generates the image on the fly from the\n specs in image_info.\n \"\"\"\n info = self.image_info[image_id]\n bg_color = np.array(info['bg_color']).reshape([1, 1, 3])\n image = np.ones([info['height'], info['width'], 3], dtype=np.uint8)\n image = image * bg_color.astype(np.uint8)\n for shape, color, dims in info['shapes']:\n image = self.draw_shape(image, shape, dims, color)\n return image\n\n def image_reference(self, image_id):\n \"\"\"Return the shapes data of the image.\"\"\"\n info = self.image_info[image_id]\n if info[\"source\"] == \"shapes\":\n return info[\"shapes\"]\n else:\n super(self.__class__).image_reference(self, image_id)\n\n def load_mask(self, image_id):\n \"\"\"Generate instance masks for shapes of the given image ID.\n \"\"\"\n info = self.image_info[image_id]\n shapes = info['shapes']\n count = len(shapes)\n mask = np.zeros([info['height'], info['width'], count], dtype=np.uint8)\n for i, (shape, _, dims) in enumerate(info['shapes']):\n mask[:, :, i:i+1] = self.draw_shape(mask[:, :, i:i+1].copy(),\n shape, dims, 1)\n # Handle occlusions\n occlusion = np.logical_not(mask[:, :, -1]).astype(np.uint8)\n for i in range(count-2, -1, -1):\n mask[:, :, i] = mask[:, :, i] * occlusion\n occlusion = np.logical_and(occlusion, np.logical_not(mask[:, :, i]))\n # Map class names to class IDs.\n class_ids = np.array([self.class_names.index(s[0]) for s in shapes])\n return mask.astype(np.bool), class_ids.astype(np.int32)\n\n def draw_shape(self, image, shape, dims, color):\n \"\"\"Draws a shape from the given specs.\"\"\"\n # Get the center x, y and the size s\n x, y, s = dims\n if shape == 'square':\n cv2.rectangle(image, (x-s, y-s), (x+s, y+s), color, -1)\n elif shape == \"circle\":\n cv2.circle(image, (x, y), s, color, -1)\n elif shape == \"triangle\":\n points = np.array([[(x, y-s),\n (x-s/math.sin(math.radians(60)), y+s),\n (x+s/math.sin(math.radians(60)), y+s),\n ]], dtype=np.int32)\n cv2.fillPoly(image, points, color)\n return image\n\n def random_shape(self, height, width):\n \"\"\"Generates specifications of a random shape that lies within\n the given height and width boundaries.\n Returns a tuple of three valus:\n * The shape name (square, circle, ...)\n * Shape color: a tuple of 3 values, RGB.\n * Shape dimensions: A tuple of values that define the shape size\n and location. Differs per shape type.\n \"\"\"\n # Shape\n shape = random.choice([\"square\", \"circle\", \"triangle\"])\n # Color\n color = tuple([random.randint(0, 255) for _ in range(3)])\n # Center x, y\n buffer = 20\n y = random.randint(buffer, height - buffer - 1)\n x = random.randint(buffer, width - buffer - 1)\n # Size\n s = random.randint(buffer, height//4)\n return shape, color, (x, y, s)\n\n def random_image(self, height, width):\n \"\"\"Creates random specifications of an image with multiple shapes.\n Returns the background color of the image and a list of shape\n specifications that can be used to draw the image.\n \"\"\"\n # Pick random background color\n bg_color = np.array([random.randint(0, 255) for _ in range(3)])\n # Generate a few random shapes and record their\n # bounding boxes\n shapes = []\n boxes = []\n N = random.randint(1, 4)\n for _ in range(N):\n shape, color, dims = self.random_shape(height, width)\n shapes.append((shape, color, dims))\n x, y, s = dims\n boxes.append([y-s, x-s, y+s, x+s])\n # Apply non-max suppression wit 0.3 threshold to avoid\n # shapes covering each other\n keep_ixs = utils.non_max_suppression(np.array(boxes), np.arange(N), 0.3)\n shapes = [s for i, s in enumerate(shapes) if i in keep_ixs]\n return bg_color, shapes", "_____no_output_____" ], [ "# Training dataset\ndataset_train = ShapesDataset()\ndataset_train.load_shapes(500, config.IMAGE_SHAPE[0], config.IMAGE_SHAPE[1])\ndataset_train.prepare()\n\n# Validation dataset\ndataset_val = ShapesDataset()\ndataset_val.load_shapes(50, config.IMAGE_SHAPE[0], config.IMAGE_SHAPE[1])\ndataset_val.prepare()", "_____no_output_____" ], [ "# Load and display random samples\nimage_ids = np.random.choice(dataset_train.image_ids, 4)\nfor image_id in image_ids:\n image = dataset_train.load_image(image_id)\n mask, class_ids = dataset_train.load_mask(image_id)\n visualize.display_top_masks(image, mask, class_ids, dataset_train.class_names)", "_____no_output_____" ] ], [ [ "## Create Model", "_____no_output_____" ] ], [ [ "# Create model in training mode\nmodel = modellib.MaskRCNN(mode=\"training\", config=config,\n model_dir=MODEL_DIR)", "_____no_output_____" ], [ "# Which weights to start with?\ninit_with = \"coco\" # imagenet, coco, or last\n\nif init_with == \"imagenet\":\n model.load_weights(model.get_imagenet_weights(), by_name=True)\nelif init_with == \"coco\":\n # Load weights trained on MS COCO, but skip layers that\n # are different due to the different number of classes\n # See README for instructions to download the COCO weights\n model.load_weights(COCO_MODEL_PATH, by_name=True,\n exclude=[\"mrcnn_class_logits\", \"mrcnn_bbox_fc\", \n \"mrcnn_bbox\", \"mrcnn_mask\"])\nelif init_with == \"last\":\n # Load the last model you trained and continue training\n model.load_weights(model.find_last(), by_name=True)", "_____no_output_____" ] ], [ [ "## Training\n\nTrain in two stages:\n1. Only the heads. Here we're freezing all the backbone layers and training only the randomly initialized layers (i.e. the ones that we didn't use pre-trained weights from MS COCO). To train only the head layers, pass `layers='heads'` to the `train()` function.\n\n2. Fine-tune all layers. For this simple example it's not necessary, but we're including it to show the process. Simply pass `layers=\"all` to train all layers.", "_____no_output_____" ] ], [ [ "# Train the head branches\n# Passing layers=\"heads\" freezes all layers except the head\n# layers. You can also pass a regular expression to select\n# which layers to train by name pattern.\nmodel.train(dataset_train, dataset_val, \n learning_rate=config.LEARNING_RATE, \n epochs=1, \n layers='heads')", "Checkpoint Path: /deepmatter/mask_rcnn/logs/shapes2017102802/mask_rcnn_{epoch:04d}.h5\nStarting at epoch 0. LR=0.002\n\nSelecting layers to train\nfpn_c5p5 (Conv2D)\nfpn_c4p4 (Conv2D)\nfpn_c3p3 (Conv2D)\nfpn_c2p2 (Conv2D)\nfpn_p5 (Conv2D)\nfpn_p2 (Conv2D)\nfpn_p3 (Conv2D)\nfpn_p4 (Conv2D)\nIn model: rpn_model\n rpn_conv_shared (Conv2D)\n rpn_class_raw (Conv2D)\n rpn_bbox_pred (Conv2D)\nmrcnn_mask_conv1 (TimeDistributed)\nmrcnn_mask_bn1 (TimeDistributed)\nmrcnn_mask_conv2 (TimeDistributed)\nmrcnn_mask_bn2 (TimeDistributed)\nmrcnn_class_conv1 (TimeDistributed)\nmrcnn_class_bn1 (TimeDistributed)\nmrcnn_mask_conv3 (TimeDistributed)\nmrcnn_mask_bn3 (TimeDistributed)\nmrcnn_class_conv2 (TimeDistributed)\nmrcnn_class_bn2 (TimeDistributed)\nmrcnn_mask_conv4 (TimeDistributed)\nmrcnn_mask_bn4 (TimeDistributed)\nmrcnn_bbox_fc (TimeDistributed)\nmrcnn_mask_deconv (TimeDistributed)\nmrcnn_class_logits (TimeDistributed)\nmrcnn_mask (TimeDistributed)\n" ], [ "# Fine tune all layers\n# Passing layers=\"all\" trains all layers. You can also \n# pass a regular expression to select which layers to\n# train by name pattern.\nmodel.train(dataset_train, dataset_val, \n learning_rate=config.LEARNING_RATE / 10,\n epochs=2, \n layers=\"all\")", "Checkpoint Path: /deepmatter/mask_rcnn/logs/shapes2017102802/mask_rcnn_{epoch:04d}.h5\nStarting at epoch 0. LR=0.0002\n\nSelecting layers to train\nconv1 (Conv2D)\nbn_conv1 (BatchNorm)\nres2a_branch2a (Conv2D)\nbn2a_branch2a (BatchNorm)\nres2a_branch2b (Conv2D)\nbn2a_branch2b (BatchNorm)\nres2a_branch2c (Conv2D)\nres2a_branch1 (Conv2D)\nbn2a_branch2c (BatchNorm)\nbn2a_branch1 (BatchNorm)\nres2b_branch2a (Conv2D)\nbn2b_branch2a (BatchNorm)\nres2b_branch2b (Conv2D)\nbn2b_branch2b (BatchNorm)\nres2b_branch2c (Conv2D)\nbn2b_branch2c (BatchNorm)\nres2c_branch2a (Conv2D)\nbn2c_branch2a (BatchNorm)\nres2c_branch2b (Conv2D)\nbn2c_branch2b (BatchNorm)\nres2c_branch2c (Conv2D)\nbn2c_branch2c (BatchNorm)\nres3a_branch2a (Conv2D)\nbn3a_branch2a (BatchNorm)\nres3a_branch2b (Conv2D)\nbn3a_branch2b (BatchNorm)\nres3a_branch2c (Conv2D)\nres3a_branch1 (Conv2D)\nbn3a_branch2c (BatchNorm)\nbn3a_branch1 (BatchNorm)\nres3b_branch2a (Conv2D)\nbn3b_branch2a (BatchNorm)\nres3b_branch2b (Conv2D)\nbn3b_branch2b (BatchNorm)\nres3b_branch2c (Conv2D)\nbn3b_branch2c (BatchNorm)\nres3c_branch2a (Conv2D)\nbn3c_branch2a (BatchNorm)\nres3c_branch2b (Conv2D)\nbn3c_branch2b (BatchNorm)\nres3c_branch2c (Conv2D)\nbn3c_branch2c (BatchNorm)\nres3d_branch2a (Conv2D)\nbn3d_branch2a (BatchNorm)\nres3d_branch2b (Conv2D)\nbn3d_branch2b (BatchNorm)\nres3d_branch2c (Conv2D)\nbn3d_branch2c (BatchNorm)\nres4a_branch2a (Conv2D)\nbn4a_branch2a (BatchNorm)\nres4a_branch2b (Conv2D)\nbn4a_branch2b (BatchNorm)\nres4a_branch2c (Conv2D)\nres4a_branch1 (Conv2D)\nbn4a_branch2c (BatchNorm)\nbn4a_branch1 (BatchNorm)\nres4b_branch2a (Conv2D)\nbn4b_branch2a (BatchNorm)\nres4b_branch2b (Conv2D)\nbn4b_branch2b (BatchNorm)\nres4b_branch2c (Conv2D)\nbn4b_branch2c (BatchNorm)\nres4c_branch2a (Conv2D)\nbn4c_branch2a (BatchNorm)\nres4c_branch2b (Conv2D)\nbn4c_branch2b (BatchNorm)\nres4c_branch2c (Conv2D)\nbn4c_branch2c (BatchNorm)\nres4d_branch2a (Conv2D)\nbn4d_branch2a (BatchNorm)\nres4d_branch2b (Conv2D)\nbn4d_branch2b (BatchNorm)\nres4d_branch2c (Conv2D)\nbn4d_branch2c (BatchNorm)\nres4e_branch2a (Conv2D)\nbn4e_branch2a (BatchNorm)\nres4e_branch2b (Conv2D)\nbn4e_branch2b (BatchNorm)\nres4e_branch2c (Conv2D)\nbn4e_branch2c (BatchNorm)\nres4f_branch2a (Conv2D)\nbn4f_branch2a (BatchNorm)\nres4f_branch2b (Conv2D)\nbn4f_branch2b (BatchNorm)\nres4f_branch2c (Conv2D)\nbn4f_branch2c (BatchNorm)\nres4g_branch2a (Conv2D)\nbn4g_branch2a (BatchNorm)\nres4g_branch2b (Conv2D)\nbn4g_branch2b (BatchNorm)\nres4g_branch2c (Conv2D)\nbn4g_branch2c (BatchNorm)\nres4h_branch2a (Conv2D)\nbn4h_branch2a (BatchNorm)\nres4h_branch2b (Conv2D)\nbn4h_branch2b (BatchNorm)\nres4h_branch2c (Conv2D)\nbn4h_branch2c (BatchNorm)\nres4i_branch2a (Conv2D)\nbn4i_branch2a (BatchNorm)\nres4i_branch2b (Conv2D)\nbn4i_branch2b (BatchNorm)\nres4i_branch2c (Conv2D)\nbn4i_branch2c (BatchNorm)\nres4j_branch2a (Conv2D)\nbn4j_branch2a (BatchNorm)\nres4j_branch2b (Conv2D)\nbn4j_branch2b (BatchNorm)\nres4j_branch2c (Conv2D)\nbn4j_branch2c (BatchNorm)\nres4k_branch2a (Conv2D)\nbn4k_branch2a (BatchNorm)\nres4k_branch2b (Conv2D)\nbn4k_branch2b (BatchNorm)\nres4k_branch2c (Conv2D)\nbn4k_branch2c (BatchNorm)\nres4l_branch2a (Conv2D)\nbn4l_branch2a (BatchNorm)\nres4l_branch2b (Conv2D)\nbn4l_branch2b (BatchNorm)\nres4l_branch2c (Conv2D)\nbn4l_branch2c (BatchNorm)\nres4m_branch2a (Conv2D)\nbn4m_branch2a (BatchNorm)\nres4m_branch2b (Conv2D)\nbn4m_branch2b (BatchNorm)\nres4m_branch2c (Conv2D)\nbn4m_branch2c (BatchNorm)\nres4n_branch2a (Conv2D)\nbn4n_branch2a (BatchNorm)\nres4n_branch2b (Conv2D)\nbn4n_branch2b (BatchNorm)\nres4n_branch2c (Conv2D)\nbn4n_branch2c (BatchNorm)\nres4o_branch2a (Conv2D)\nbn4o_branch2a (BatchNorm)\nres4o_branch2b (Conv2D)\nbn4o_branch2b (BatchNorm)\nres4o_branch2c (Conv2D)\nbn4o_branch2c (BatchNorm)\nres4p_branch2a (Conv2D)\nbn4p_branch2a (BatchNorm)\nres4p_branch2b (Conv2D)\nbn4p_branch2b (BatchNorm)\nres4p_branch2c (Conv2D)\nbn4p_branch2c (BatchNorm)\nres4q_branch2a (Conv2D)\nbn4q_branch2a (BatchNorm)\nres4q_branch2b (Conv2D)\nbn4q_branch2b (BatchNorm)\nres4q_branch2c (Conv2D)\nbn4q_branch2c (BatchNorm)\nres4r_branch2a (Conv2D)\nbn4r_branch2a (BatchNorm)\nres4r_branch2b (Conv2D)\nbn4r_branch2b (BatchNorm)\nres4r_branch2c (Conv2D)\nbn4r_branch2c (BatchNorm)\nres4s_branch2a (Conv2D)\nbn4s_branch2a (BatchNorm)\nres4s_branch2b (Conv2D)\nbn4s_branch2b (BatchNorm)\nres4s_branch2c (Conv2D)\nbn4s_branch2c (BatchNorm)\nres4t_branch2a (Conv2D)\nbn4t_branch2a (BatchNorm)\nres4t_branch2b (Conv2D)\nbn4t_branch2b (BatchNorm)\nres4t_branch2c (Conv2D)\nbn4t_branch2c (BatchNorm)\nres4u_branch2a (Conv2D)\nbn4u_branch2a (BatchNorm)\nres4u_branch2b (Conv2D)\nbn4u_branch2b (BatchNorm)\nres4u_branch2c (Conv2D)\nbn4u_branch2c (BatchNorm)\nres4v_branch2a (Conv2D)\nbn4v_branch2a (BatchNorm)\nres4v_branch2b (Conv2D)\nbn4v_branch2b (BatchNorm)\nres4v_branch2c (Conv2D)\nbn4v_branch2c (BatchNorm)\nres4w_branch2a (Conv2D)\nbn4w_branch2a (BatchNorm)\nres4w_branch2b (Conv2D)\nbn4w_branch2b (BatchNorm)\nres4w_branch2c (Conv2D)\nbn4w_branch2c (BatchNorm)\nres5a_branch2a (Conv2D)\nbn5a_branch2a (BatchNorm)\nres5a_branch2b (Conv2D)\nbn5a_branch2b (BatchNorm)\nres5a_branch2c (Conv2D)\nres5a_branch1 (Conv2D)\nbn5a_branch2c (BatchNorm)\nbn5a_branch1 (BatchNorm)\nres5b_branch2a (Conv2D)\nbn5b_branch2a (BatchNorm)\nres5b_branch2b (Conv2D)\nbn5b_branch2b (BatchNorm)\nres5b_branch2c (Conv2D)\nbn5b_branch2c (BatchNorm)\nres5c_branch2a (Conv2D)\nbn5c_branch2a (BatchNorm)\nres5c_branch2b (Conv2D)\nbn5c_branch2b (BatchNorm)\nres5c_branch2c (Conv2D)\nbn5c_branch2c (BatchNorm)\nfpn_c5p5 (Conv2D)\nfpn_c4p4 (Conv2D)\nfpn_c3p3 (Conv2D)\nfpn_c2p2 (Conv2D)\nfpn_p5 (Conv2D)\nfpn_p2 (Conv2D)\nfpn_p3 (Conv2D)\nfpn_p4 (Conv2D)\nIn model: rpn_model\n rpn_conv_shared (Conv2D)\n rpn_class_raw (Conv2D)\n rpn_bbox_pred (Conv2D)\nmrcnn_mask_conv1 (TimeDistributed)\nmrcnn_mask_bn1 (TimeDistributed)\nmrcnn_mask_conv2 (TimeDistributed)\nmrcnn_mask_bn2 (TimeDistributed)\nmrcnn_class_conv1 (TimeDistributed)\nmrcnn_class_bn1 (TimeDistributed)\nmrcnn_mask_conv3 (TimeDistributed)\nmrcnn_mask_bn3 (TimeDistributed)\nmrcnn_class_conv2 (TimeDistributed)\nmrcnn_class_bn2 (TimeDistributed)\nmrcnn_mask_conv4 (TimeDistributed)\nmrcnn_mask_bn4 (TimeDistributed)\nmrcnn_bbox_fc (TimeDistributed)\nmrcnn_mask_deconv (TimeDistributed)\nmrcnn_class_logits (TimeDistributed)\nmrcnn_mask (TimeDistributed)\n" ], [ "# Save weights\n# Typically not needed because callbacks save after every epoch\n# Uncomment to save manually\n# model_path = os.path.join(MODEL_DIR, \"mask_rcnn_shapes.h5\")\n# model.keras_model.save_weights(model_path)", "_____no_output_____" ] ], [ [ "## Detection", "_____no_output_____" ] ], [ [ "class InferenceConfig(ShapesConfig):\n GPU_COUNT = 1\n IMAGES_PER_GPU = 1\n\ninference_config = InferenceConfig()\n\n# Recreate the model in inference mode\nmodel = modellib.MaskRCNN(mode=\"inference\", \n config=inference_config,\n model_dir=MODEL_DIR)\n\n# Get path to saved weights\n# Either set a specific path or find last trained weights\n# model_path = os.path.join(ROOT_DIR, \".h5 file name here\")\nmodel_path = model.find_last()\n\n# Load trained weights\nprint(\"Loading weights from \", model_path)\nmodel.load_weights(model_path, by_name=True)", "_____no_output_____" ], [ "# Test on a random image\nimage_id = random.choice(dataset_val.image_ids)\noriginal_image, image_meta, gt_class_id, gt_bbox, gt_mask =\\\n modellib.load_image_gt(dataset_val, inference_config, \n image_id, use_mini_mask=False)\n\nlog(\"original_image\", original_image)\nlog(\"image_meta\", image_meta)\nlog(\"gt_class_id\", gt_class_id)\nlog(\"gt_bbox\", gt_bbox)\nlog(\"gt_mask\", gt_mask)\n\nvisualize.display_instances(original_image, gt_bbox, gt_mask, gt_class_id, \n dataset_train.class_names, figsize=(8, 8))", "original_image shape: (128, 128, 3) min: 108.00000 max: 236.00000\nimage_meta shape: (12,) min: 0.00000 max: 128.00000\ngt_bbox shape: (2, 5) min: 2.00000 max: 102.00000\ngt_mask shape: (128, 128, 2) min: 0.00000 max: 1.00000\n" ], [ "results = model.detect([original_image], verbose=1)\n\nr = results[0]\nvisualize.display_instances(original_image, r['rois'], r['masks'], r['class_ids'], \n dataset_val.class_names, r['scores'], ax=get_ax())", "Processing 1 images\nimage shape: (128, 128, 3) min: 108.00000 max: 236.00000\nmolded_images shape: (1, 128, 128, 3) min: -15.70000 max: 132.10000\nimage_metas shape: (1, 12) min: 0.00000 max: 128.00000\n" ] ], [ [ "## Evaluation", "_____no_output_____" ] ], [ [ "# Compute VOC-Style mAP @ IoU=0.5\n# Running on 10 images. Increase for better accuracy.\nimage_ids = np.random.choice(dataset_val.image_ids, 10)\nAPs = []\nfor image_id in image_ids:\n # Load image and ground truth data\n image, image_meta, gt_class_id, gt_bbox, gt_mask =\\\n modellib.load_image_gt(dataset_val, inference_config,\n image_id, use_mini_mask=False)\n molded_images = np.expand_dims(modellib.mold_image(image, inference_config), 0)\n # Run object detection\n results = model.detect([image], verbose=0)\n r = results[0]\n # Compute AP\n AP, precisions, recalls, overlaps =\\\n utils.compute_ap(gt_bbox, gt_class_id, gt_mask,\n r[\"rois\"], r[\"class_ids\"], r[\"scores\"], r['masks'])\n APs.append(AP)\n \nprint(\"mAP: \", np.mean(APs))", "mAP: 0.95\n" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code" ]
[ [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown" ], [ "code" ] ]
ece7bed7f9e8252c0d50d73214c55b56f531eeff
62,866
ipynb
Jupyter Notebook
exp/exp046_roberta_large_1dcnn.ipynb
TakoiHirokazu/Feedback-Prize-Evaluating-Student-Writing
8aea7e67717b146f27766bba1b532b5cd1c0774e
[ "MIT" ]
null
null
null
exp/exp046_roberta_large_1dcnn.ipynb
TakoiHirokazu/Feedback-Prize-Evaluating-Student-Writing
8aea7e67717b146f27766bba1b532b5cd1c0774e
[ "MIT" ]
null
null
null
exp/exp046_roberta_large_1dcnn.ipynb
TakoiHirokazu/Feedback-Prize-Evaluating-Student-Writing
8aea7e67717b146f27766bba1b532b5cd1c0774e
[ "MIT" ]
null
null
null
32.606846
255
0.492222
[ [ [ "# ========================================\n# library\n# ========================================\nimport pandas as pd\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom sklearn.model_selection import StratifiedKFold, KFold,GroupKFold\nfrom sklearn.metrics import mean_squared_error\n%matplotlib inline\nimport torch\nimport torch.nn as nn\nfrom torch.utils.data import Dataset, DataLoader, Subset\nimport transformers\nfrom transformers import LongformerTokenizer, LongformerModel,AutoTokenizer,RobertaModel\nfrom transformers import AdamW, get_linear_schedule_with_warmup\nfrom torch.cuda.amp import autocast, GradScaler\nimport logging\nfrom ast import literal_eval\nimport sys\nfrom contextlib import contextmanager\nimport time\nimport random\nfrom tqdm import tqdm\nimport os", "2022-02-14 01:21:48.579260: I tensorflow/stream_executor/platform/default/dso_loader.cc:49] Successfully opened dynamic library libcudart.so.11.0\n" ], [ "# ==================\n# Constant\n# ==================\nex = \"046\"\nTRAIN_PATH = \"../data/train.csv\"\nDATA_DIR = \"../data/roberta-large/\"\nif not os.path.exists(f\"../output/exp/ex{ex}\"):\n os.makedirs(f\"../output/exp/ex{ex}\")\n os.makedirs(f\"../output/exp/ex{ex}/ex{ex}_model\")\n \nOUTPUT_DIR = f\"../output/exp/ex{ex}\"\nMODEL_PATH_BASE = f\"../output/exp/ex{ex}/ex{ex}_model/ex{ex}\"\nLOGGER_PATH = f\"../output/exp/ex{ex}/ex{ex}.txt\"\ndevice = torch.device(\"cuda\" if torch.cuda.is_available() else \"cpu\")", "_____no_output_____" ], [ "# ===============\n# Configs\n# ===============\nSEED = 0\nN_SPLITS = 5\nSHUFFLE = True\nnum_workers = 4\nBATCH_SIZE = 8\n\nn_epochs = 6\nmax_len = 512\nweight_decay = 0.1\nbeta = (0.9, 0.98)\nlr = 2e-5\nnum_warmup_steps_rate = 0.1\n\nMODEL_PATH = 'roberta-large'\ntokenizer = AutoTokenizer.from_pretrained(MODEL_PATH)", "_____no_output_____" ], [ "# ===============\n# Functions\n# ===============\ndef set_seed(seed: int = 42):\n random.seed(seed)\n np.random.seed(seed)\n os.environ[\"PYTHONHASHSEED\"] = str(seed)\n torch.manual_seed(seed)\n torch.cuda.manual_seed(seed)\n torch.backends.cudnn.deterministic = True\n torch.backends.cudnn.benchmark = False\n\ndef setup_logger(out_file=None, stderr=True, stderr_level=logging.INFO, file_level=logging.DEBUG):\n LOGGER.handlers = []\n LOGGER.setLevel(min(stderr_level, file_level))\n\n if stderr:\n handler = logging.StreamHandler(sys.stderr)\n handler.setFormatter(FORMATTER)\n handler.setLevel(stderr_level)\n LOGGER.addHandler(handler)\n\n if out_file is not None:\n handler = logging.FileHandler(out_file)\n handler.setFormatter(FORMATTER)\n handler.setLevel(file_level)\n LOGGER.addHandler(handler)\n\n LOGGER.info(\"logger set up\")\n return LOGGER\n\n\n@contextmanager\ndef timer(name):\n t0 = time.time()\n yield \n LOGGER.info(f'[{name}] done in {time.time() - t0:.0f} s')\n \n \nLOGGER = logging.getLogger()\nFORMATTER = logging.Formatter(\"%(asctime)s - %(levelname)s - %(message)s\")\nsetup_logger(out_file=LOGGER_PATH)", "2022-02-14 01:21:53,687 - INFO - logger set up\n" ], [ "class TrainDataset(Dataset):\n def __init__(self, token,attentiona_mask,label=None):\n self.len = len(token)\n self.token = token\n self.attention_mask = attentiona_mask\n self.label = label\n #self.get_wids = get_wids # for validation\n\n def __getitem__(self, index):\n # GET TEXT AND WORD LABELS \n if self.label is not None:\n return {\n 'token': torch.tensor(self.token[index], dtype=torch.long),\n 'mask': torch.tensor(self.attention_mask[index], dtype=torch.long),\n \"y\":torch.tensor(self.label[index], dtype=torch.float32)\n }\n else:\n return {\n 'token': torch.tensor(self.token[index], dtype=torch.long),\n 'mask': torch.tensor(self.attention_mask[index], dtype=torch.long),\n }\n\n def __len__(self):\n return self.len\n\nclass custom_model(nn.Module):\n def __init__(self):\n super(custom_model, self).__init__()\n self.backbone = RobertaModel.from_pretrained(\n MODEL_PATH, \n )\n \n #self.dropout = nn.Dropout(p=0.2)\n self.ln = nn.LayerNorm(1024)\n \n self.conv1= nn.Conv1d(1024, 512, kernel_size=3, padding=1)\n self.conv2= nn.Conv1d(1024, 512, kernel_size=9, padding=4)\n self.conv3= nn.Conv1d(1024, 512, kernel_size=15, padding=7)\n self.conv4= nn.Conv1d(1024, 512, kernel_size=31, padding=15)\n self.ln1 = nn.Sequential(nn.LayerNorm(512),\n nn.ReLU(),\n nn.Dropout(0.2))\n self.ln2 = nn.Sequential( nn.LayerNorm(512),\n nn.ReLU(),\n nn.Dropout(0.2))\n self.ln3 = nn.Sequential( nn.LayerNorm(512),\n nn.ReLU(),\n nn.Dropout(0.2))\n self.ln4 = nn.Sequential( nn.LayerNorm(512),\n nn.ReLU(),\n nn.Dropout(0.2))\n \n self.linear1 = nn.Sequential(\n nn.Linear(2048,1024),\n nn.LayerNorm(1024),\n nn.ReLU(),\n nn.Dropout(0.2),\n nn.Linear(1024,2),\n )\n self.linear2 = nn.Sequential(\n nn.Linear(2048,1024),\n nn.LayerNorm(1024),\n nn.ReLU(),\n nn.Dropout(0.2),\n nn.Linear(1024,2),\n )\n self.linear3 = nn.Sequential(\n nn.Linear(2048,1024),\n nn.LayerNorm(1024),\n nn.ReLU(),\n nn.Dropout(0.2),\n nn.Linear(1024,2),\n )\n self.linear4 = nn.Sequential(\n nn.Linear(2048,1024),\n nn.LayerNorm(1024),\n nn.ReLU(),\n nn.Dropout(0.2),\n nn.Linear(1024,2),\n )\n self.linear5 = nn.Sequential(\n nn.Linear(2048,1024),\n nn.LayerNorm(1024),\n nn.ReLU(),\n nn.Dropout(0.2),\n nn.Linear(1024,2),\n )\n self.linear6 = nn.Sequential(\n nn.Linear(2048,1024),\n nn.LayerNorm(1024),\n nn.ReLU(),\n nn.Dropout(0.2),\n nn.Linear(1024,2),\n )\n self.linear7 = nn.Sequential(\n nn.Linear(2048,1024),\n nn.LayerNorm(1024),\n nn.ReLU(),\n nn.Dropout(0.2),\n nn.Linear(1024,2),\n )\n self.linear8 = nn.Sequential(\n nn.Linear(2048,1024),\n nn.LayerNorm(1024),\n nn.ReLU(),\n nn.Dropout(0.2),\n nn.Linear(1024,1),\n )\n def forward(self, ids, mask):\n # pooler\n emb = self.backbone(ids, attention_mask=mask)[\"last_hidden_state\"]\n output = self.ln(emb)\n output = output.permute((0, 2, 1)).contiguous()\n output1 = self.conv1(output)\n output1 = self.ln1(output1.permute((0, 2, 1)).contiguous())\n output2 = self.conv2(output)\n output2 = self.ln2(output2.permute((0, 2, 1)).contiguous())\n output3 = self.conv3(output)\n output3 = self.ln3(output3.permute((0, 2, 1)).contiguous())\n output4 = self.conv4(output)\n output4 = self.ln4(output4.permute((0, 2, 1)).contiguous())\n output_concat = torch.cat([output1,output2,output3,output4],axis=-1)\n output2_1 = self.linear1(output_concat)\n output2_2 = self.linear2(output_concat)\n output2_3 = self.linear3(output_concat)\n output2_4 = self.linear4(output_concat)\n output2_5 = self.linear5(output_concat)\n output2_6 = self.linear6(output_concat)\n output2_7= self.linear7(output_concat)\n output2_8 = self.linear8(output_concat)\n out = torch.cat(\n [output2_1,output2_2,output2_3,output2_4,\n output2_5,output2_6,output2_7,output2_8], axis=2)\n return out", "_____no_output_____" ], [ "target_map_rev = {0:'Lead', 1:'Position', 2:'Evidence', 3:'Claim', 4:'Concluding Statement',\n 5:'Counterclaim', 6:'Rebuttal', 7:'blank'}\n\ndef get_preds_collate(dataset, verbose,text_ids, preds, preds_len):\n all_predictions = []\n\n for id_num in tqdm(range(len(preds))):\n \n # GET ID\n #if (id_num%100==0)&(verbose): \n # print(id_num,', ',end='')\n n = text_ids[id_num]\n max_len = int(preds_len[id_num])\n # GET TOKEN POSITIONS IN CHARS\n name = f'../data/{dataset}/{n}.txt'\n txt = open(name, 'r').read()\n tokens = tokenizer.encode_plus(txt, max_length=max_len, padding='max_length',\n truncation=True, return_offsets_mapping=True)\n off = tokens['offset_mapping']\n \n # GET WORD POSITIONS IN CHARS\n w = []\n blank = True\n for i in range(len(txt)):\n if (txt[i]!=' ')&(txt[i]!='\\n')&(txt[i]!='\\xa0')&(txt[i]!='\\x85')&(blank==True):\n w.append(i)\n blank=False\n elif (txt[i]==' ')|(txt[i]=='\\n')|(txt[i]=='\\xa0')|(txt[i]=='\\x85'):\n blank=True\n w.append(1e6)\n \n # MAPPING FROM TOKENS TO WORDS\n word_map = -1 * np.ones(max_len,dtype='int32')\n w_i = 0\n for i in range(len(off)):\n if off[i][1]==0: continue\n while off[i][0]>=w[w_i+1]: w_i += 1\n word_map[i] = int(w_i)\n \n # CONVERT TOKEN PREDICTIONS INTO WORD LABELS\n ### KEY: ###\n # 0: LEAD_B, 1: LEAD_I\n # 2: POSITION_B, 3: POSITION_I\n # 4: EVIDENCE_B, 5: EVIDENCE_I\n # 6: CLAIM_B, 7: CLAIM_I\n # 8: CONCLUSION_B, 9: CONCLUSION_I\n # 10: COUNTERCLAIM_B, 11: COUNTERCLAIM_I\n # 12: REBUTTAL_B, 13: REBUTTAL_I\n # 14: NOTHING i.e. O\n ### NOTE THESE VALUES ARE DIVIDED BY 2 IN NEXT CODE LINE\n pred = preds[id_num,]/2.0\n \n i = 0\n while i<max_len:\n prediction = []\n start = pred[i]\n if start in [0,1,2,3,4,5,6,7]:\n prediction.append(word_map[i])\n i += 1\n if i>=max_len: break\n while pred[i]==start+0.5:\n if not word_map[i] in prediction:\n prediction.append(word_map[i])\n i += 1\n if i>=max_len: break\n else:\n i += 1\n prediction = [x for x in prediction if x!=-1]\n if len(prediction)>4:\n all_predictions.append( (n, target_map_rev[int(start)], \n ' '.join([str(x) for x in prediction]) ) )\n \n # MAKE DATAFRAME\n df = pd.DataFrame(all_predictions)\n df.columns = ['id','class','predictionstring']\n \n return df\n\n\ndef calc_overlap(row):\n \"\"\"\n Calculates the overlap between prediction and\n ground truth and overlap percentages used for determining\n true positives.\n \"\"\"\n set_pred = set(row.predictionstring_pred.split(' '))\n set_gt = set(row.predictionstring_gt.split(' '))\n # Length of each and intersection\n len_gt = len(set_gt)\n len_pred = len(set_pred)\n inter = len(set_gt.intersection(set_pred))\n overlap_1 = inter / len_gt\n overlap_2 = inter/ len_pred\n return [overlap_1, overlap_2]\n\n\ndef score_feedback_comp(pred_df, gt_df):\n \"\"\"\n A function that scores for the kaggle\n Student Writing Competition\n \n Uses the steps in the evaluation page here:\n https://www.kaggle.com/c/feedback-prize-2021/overview/evaluation\n \"\"\"\n gt_df = gt_df[['id','discourse_type','predictionstring']] \\\n .reset_index(drop=True).copy()\n pred_df = pred_df[['id','class','predictionstring']] \\\n .reset_index(drop=True).copy()\n pred_df['pred_id'] = pred_df.index\n gt_df['gt_id'] = gt_df.index\n # Step 1. all ground truths and predictions for a given class are compared.\n joined = pred_df.merge(gt_df,\n left_on=['id','class'],\n right_on=['id','discourse_type'],\n how='outer',\n suffixes=('_pred','_gt')\n )\n joined['predictionstring_gt'] = joined['predictionstring_gt'].fillna(' ')\n joined['predictionstring_pred'] = joined['predictionstring_pred'].fillna(' ')\n\n joined['overlaps'] = joined.apply(calc_overlap, axis=1)\n\n # 2. If the overlap between the ground truth and prediction is >= 0.5, \n # and the overlap between the prediction and the ground truth >= 0.5,\n # the prediction is a match and considered a true positive.\n # If multiple matches exist, the match with the highest pair of overlaps is taken.\n joined['overlap1'] = joined['overlaps'].apply(lambda x: eval(str(x))[0])\n joined['overlap2'] = joined['overlaps'].apply(lambda x: eval(str(x))[1])\n\n\n joined['potential_TP'] = (joined['overlap1'] >= 0.5) & (joined['overlap2'] >= 0.5)\n joined['max_overlap'] = joined[['overlap1','overlap2']].max(axis=1)\n tp_pred_ids = joined.query('potential_TP') \\\n .sort_values('max_overlap', ascending=False) \\\n .groupby(['id','predictionstring_gt']).first()['pred_id'].values\n\n # 3. Any unmatched ground truths are false negatives\n # and any unmatched predictions are false positives.\n fp_pred_ids = [p for p in joined['pred_id'].unique() if p not in tp_pred_ids]\n\n matched_gt_ids = joined.query('potential_TP')['gt_id'].unique()\n unmatched_gt_ids = [c for c in joined['gt_id'].unique() if c not in matched_gt_ids]\n\n # Get numbers of each type\n TP = len(tp_pred_ids)\n FP = len(fp_pred_ids)\n FN = len(unmatched_gt_ids)\n #calc microf1\n my_f1_score = TP / (TP + 0.5*(FP+FN))\n return my_f1_score\n\ndef collate(d,train=True):\n mask_len = int(d[\"mask\"].sum(axis=1).max())\n if train:\n return {\"token\" : d['token'][:,:mask_len],\n \"mask\" : d['mask'][:,:mask_len],\n \"y\" : d['y'][:,:mask_len],\n \"max_len\" : mask_len}\n else:\n return {\"token\" : d['token'][:,:mask_len],\n \"mask\" : d['mask'][:,:mask_len],\n \"max_len\" : mask_len}", "_____no_output_____" ], [ "# ================================\n# Main\n# ================================\ntrain = pd.read_csv(TRAIN_PATH)\nIDS = train.id.unique()\nid_array = np.array(IDS)", "_____no_output_____" ], [ "# ================================\n# data load\n# ================================\ntargets = np.load(DATA_DIR + f\"targets_{max_len}.npy\")\ntrain_tokens = np.load(DATA_DIR + f\"tokens_{max_len}.npy\")\ntrain_attention = np.load(DATA_DIR + f\"attention_{max_len}.npy\")", "_____no_output_____" ], [ "# ================================\n# train\n# ================================\nwith timer(\"roberta_large\"):\n set_seed(SEED)\n oof = pd.DataFrame()\n oof_pred = np.ndarray((0,max_len,15))\n kf = KFold(n_splits=N_SPLITS, shuffle=SHUFFLE, random_state=SEED)\n for fold, (train_idx, valid_idx) in enumerate(kf.split(id_array)):\n print(f\"fold{fold}:start\")\n x_train_token, x_train_attention, y_train = train_tokens[train_idx], train_attention[train_idx], targets[train_idx]\n x_val_token, x_val_attention, y_val = train_tokens[valid_idx], train_attention[valid_idx], targets[valid_idx]\n train_val = train[train.id.isin(id_array[valid_idx])].reset_index(drop=True)\n \n # dataset\n train_ = TrainDataset( x_train_token, x_train_attention, y_train)\n val_ = TrainDataset( x_val_token, x_val_attention, y_val)\n \n # loader\n train_loader = DataLoader(dataset=train_, batch_size=BATCH_SIZE, shuffle = True ,pin_memory=True)\n val_loader = DataLoader(dataset=val_, batch_size=BATCH_SIZE, shuffle = False , pin_memory=True)\n \n # model\n model = custom_model()\n model = model.to(device)\n \n # optimizer, scheduler\n param_optimizer = list(model.named_parameters())\n no_decay = ['bias', 'LayerNorm.bias', 'LayerNorm.weight']\n optimizer_grouped_parameters = [\n {'params': [p for n, p in param_optimizer if not any(nd in n for nd in no_decay)], 'weight_decay': weight_decay},\n {'params': [p for n, p in param_optimizer if any(nd in n for nd in no_decay)], 'weight_decay': 0.0}\n ]\n optimizer = AdamW(optimizer_grouped_parameters,\n lr=lr,\n betas=beta,\n weight_decay=weight_decay,\n )\n num_train_optimization_steps = int(len(train_loader) * n_epochs)\n num_warmup_steps = int(num_train_optimization_steps * num_warmup_steps_rate)\n scheduler = get_linear_schedule_with_warmup(optimizer,\n num_warmup_steps=num_warmup_steps,\n num_training_steps=num_train_optimization_steps)\n \n criterion = nn.BCEWithLogitsLoss()\n best_val = 0\n \n for epoch in range(n_epochs):\n print(f\"============start epoch:{epoch}==============\")\n model.train() \n val_losses_batch = []\n scaler = GradScaler()\n for i, d in tqdm(enumerate(train_loader),total=len(train_loader)):\n d = collate(d)\n ids = d['token'].to(device)\n mask = d['mask'].to(device)\n labels = d['y'].to(device)\n #labels = labels.unsqueeze(-1)\n optimizer.zero_grad()\n with autocast():\n output = model(ids,mask)\n loss = criterion(output[mask == 1], labels[mask == 1])\n scaler.scale(loss).backward()\n #torch.nn.utils.clip_grad_norm_(model.parameters(), clip_grad_norm)\n scaler.step(optimizer)\n scaler.update()\n scheduler.step()\n \n y_pred2 = []\n val_preds = np.ndarray((0,max_len,15))\n val_len = np.ndarray(0)\n model.eval() # switch model to the evaluation mode\n with torch.no_grad(): \n # Predicting on validation set\n \n for d in tqdm(val_loader,total=len(val_loader)):\n # =========================\n # data loader\n # =========================\n d = collate(d)\n ids = d['token'].to(device)\n mask = d['mask'].to(device)\n with autocast():\n outputs = model(ids, mask)\n outputs = np.concatenate([outputs.sigmoid().detach().cpu().numpy(),np.zeros([len(outputs),max_len - d[\"max_len\"],15])],axis=1)\n val_preds = np.concatenate([val_preds, outputs], axis=0)\n val_len = np.concatenate([val_len,np.array([d[\"max_len\"] for i in range(len(ids))])],axis=0)\n val_preds_max = np.argmax(val_preds,axis=-1)\n oof_ = get_preds_collate( dataset='train', verbose=True, text_ids=id_array[valid_idx],\n preds = val_preds_max,preds_len=val_len) \n # COMPUTE F1 SCORE\n f1s = []\n CLASSES = oof_['class'].unique()\n print()\n for c in CLASSES:\n pred_df = oof_.loc[oof_['class']==c].copy()\n gt_df = train_val.loc[train_val['discourse_type']==c].copy()\n f1 = score_feedback_comp(pred_df, gt_df)\n print(c,f1)\n f1s.append(f1)\n score = np.mean(f1s)\n LOGGER.info(f'{fold},{epoch}:{i},val_score:{score}')\n if best_val < score:\n print(\"save model weight\")\n best_val = score\n best_val_preds = val_preds\n oof_best = oof_.copy()\n torch.save(model.state_dict(), MODEL_PATH_BASE + f\"_{fold}.pth\") # Saving current best model\n oof_best[\"fold\"] = fold\n oof_best.to_csv(OUTPUT_DIR + f\"/ex{ex}_oof_{fold}.csv\",index=False)\n np.save(OUTPUT_DIR + f\"/ex{ex}_oof_npy_{fold}.npy\",best_val_preds)", "fold0:start\n" ], [ "oof = pd.DataFrame()\nfor i in range(5):\n oof__ = pd.read_csv(OUTPUT_DIR + f\"/ex{ex}_oof_{i}.csv\")\n oof = pd.concat([oof,oof__]).reset_index(drop=True)\n# COMPUTE F1 SCORE\nf1s = []\nCLASSES = oof['class'].unique()\nfor c in CLASSES:\n pred_df = oof.loc[oof['class']==c].copy()\n gt_df = train.loc[train['discourse_type']==c].copy()\n f1 = score_feedback_comp(pred_df, gt_df)\n print(c,f1)\n f1s.append(f1)\nscore = np.mean(f1s)\nLOGGER.info(f'CV:{score}')", "Lead 0.823718808693319\nClaim 0.6090094006411998\nPosition 0.7003339742940998\nEvidence 0.6761049075724875\nConcluding Statement 0.6019067703527965\nCounterclaim 0.47756956274843837\n" ] ] ]
[ "code" ]
[ [ "code", "code", "code", "code", "code", "code", "code", "code", "code", "code" ] ]
ece7c29aaad0f468e345ecef28e823833e864576
12,362
ipynb
Jupyter Notebook
AI 이노베이션 스퀘어 음성지능 과정/20200927/lab-11_6_PackedSequence.ipynb
donddog/AI_Innovation_Square_Codes
a04d50db011d25e00d8486146c24124c50242aa7
[ "MIT" ]
1
2021-02-11T16:45:21.000Z
2021-02-11T16:45:21.000Z
AI 이노베이션 스퀘어 음성지능 과정/20200927/lab-11_6_PackedSequence.ipynb
donddog/AI_Innovation_Square_Codes
a04d50db011d25e00d8486146c24124c50242aa7
[ "MIT" ]
null
null
null
AI 이노베이션 스퀘어 음성지능 과정/20200927/lab-11_6_PackedSequence.ipynb
donddog/AI_Innovation_Square_Codes
a04d50db011d25e00d8486146c24124c50242aa7
[ "MIT" ]
null
null
null
25.228571
168
0.517877
[ [ [ "### PackedSequence 와 PaddedSequence\n\n[링크: PackedSequence에 대한 PyTorch 공식 문서](https://pytorch.org/docs/stable/nn.html#packedsequence)\n\n이 튜토리얼에서는 RNN / LSTM 계열의 모델에서 sequence batch를 잘 활용할 수 있는 `PackedSequence` 와 `PaddedSequence`를 만드는 법을 배워보겠습니다.\n\nPyTorch 라이브러리 안에는 다음 4가지 함수들이 주어집니다.\n\n`pad_sequence`, `pack_sequence`, `pack_padded_sequence`, `pad_packed_sequence`", "_____no_output_____" ] ], [ [ "import torch\nimport numpy as np\nfrom torch.nn.utils.rnn import pad_sequence, pack_sequence, pack_padded_sequence, pad_packed_sequence", "_____no_output_____" ] ], [ [ "### 예제 데이터", "_____no_output_____" ] ], [ [ "data = ['hello world',\n 'midnight',\n 'calculation',\n 'path',\n 'short circuit']\n\nchar_set = ['<pad>'] + list(set(char for seq in data for char in seq)) # Get all characters and include pad token\nchar2idx = {char: idx for idx, char in enumerate(char_set)} # Constuct character to index dictionary\n\nprint('char_set:', char_set)\nprint('char_set length:', len(char_set))", "char_set: ['<pad>', 't', 'l', 'r', 'm', 'd', 'u', 'w', ' ', 'p', 'e', 'g', 's', 'o', 'i', 'a', 'h', 'n', 'c']\nchar_set length: 19\n" ], [ "X = [torch.LongTensor([char2idx[char] for char in seq]) for seq in data]\n\nfor sequence in X:\n print(sequence)", "tensor([16, 10, 2, 2, 13, 8, 7, 13, 3, 2, 5])\ntensor([ 4, 14, 5, 17, 14, 11, 16, 1])\ntensor([18, 15, 2, 18, 6, 2, 15, 1, 14, 13, 17])\ntensor([ 9, 15, 1, 16])\ntensor([12, 16, 13, 3, 1, 8, 18, 14, 3, 18, 6, 14, 1])\n" ], [ "lengths = [len(seq) for seq in X]\nprint('lengths:', lengths)", "lengths: [11, 8, 11, 4, 13]\n" ] ], [ [ "### Sequence 데이터의 경우 어떻게 batch로 묶을까요?\n\n위와같이 Text 나 audio 처럼 sequence 형식인 데이터의 경우 길이가 각각 다 다르기 때문에 \n\n하나의 batch로 만들어주기 위해서 일반적으로 제일 긴 sequence 길이에 맞춰 뒷부분에 padding을 추가해줍니다.\n\n이 방식이 일반적으로 많이 쓰이는 Padding 방식입니다.\n\n하지만 PyTorch에서는 `PackedSequence`라는 것을 쓰면 padding 없이도 정확히 필요한 부분까지만 병렬 계산을 할 수 있습니다.", "_____no_output_____" ], [ "### `pad_sequence` 함수를 이용하여 PaddedSequence (그냥 Tensor) 만들기\n\n사실, PaddedSequence는 sequence중에서 가장 긴 sequence와 길이를 맞추어주기 위해 padding을 추가한 일반적인 **Tensor**를 말합니다.\n\n(따로 PaddedSequence라는 class는 존재하지 않습니다.)\n\n이때, `pad_sequence`라는 PyTorch 기본 라이브러리 함수를 이용하면 쉽게 padding을 추가할 수 있습니다.\n\n여기서 주의하실 점은 input이 **Tensor들의 list** 로 주어져야합니다. (그냥 **Tensor** 가 아닌 **Tensor들의 list** 입니다.)\n\nlist 안에 있는 각각의 Tensor들의 shape가 `(?, a, b, ...)` 라고 할때, (여기서 ?는 각각 다른 sequence length 입니다.)\n\n`pad_sequence` 함수를 쓰면 `(T, batch_size, a, b, ...)` shape를 가지는 Tensor가 리턴됩니다. \n(여기서 `T`는 batch안에서 가장 큰 sequence length 입니다.)\n\n만약, `pad_sequence`에 명시적으로 `batch_first=True`라는 파라미터를 지정해주면, \n\n`(batch_size, T, a, b, ...)` shape를 가지는 Tensor가 리턴됩니다. \n\n기본적으로 padding 값은 0으로 되어있지만, `padding_value=42`와 같이 파라미터를 지정해주면, padding하는 값도 정할 수 있습니다.", "_____no_output_____" ] ], [ [ "padded_sequence = pad_sequence(X, batch_first=True)\n\nprint(padded_sequence)\nprint(padded_sequence.shape)", "tensor([[16, 10, 2, 2, 13, 8, 7, 13, 3, 2, 5, 0, 0],\n [ 4, 14, 5, 17, 14, 11, 16, 1, 0, 0, 0, 0, 0],\n [18, 15, 2, 18, 6, 2, 15, 1, 14, 13, 17, 0, 0],\n [ 9, 15, 1, 16, 0, 0, 0, 0, 0, 0, 0, 0, 0],\n [12, 16, 13, 3, 1, 8, 18, 14, 3, 18, 6, 14, 1]])\ntorch.Size([5, 13])\n" ] ], [ [ "### `pack_sequence` 함수를 이용하여 PackedSequence 만들기\n\nPackedSequence는 위와같이 padding token을 추가하여 sequence의 최대 길이에 맞는 Tensor를 만드는게 아닌,\n\npadding을 추가하지 않고 정확히 주어진 sequence 길이까지만 모델이 연산을 하게끔 만드는 PyTorch의 자료구조입니다.\n\n이 PackedSequence를 만들기 위해서는 한가지 조건이 필요합니다.\n- **주어지는 input (list of Tensor)는 길이에 따른 내림차순으로 정렬이 되어있어야 합니다.**\n\n따라서 먼저 input을 길이에 따른 내림차순으로 정렬해봅시다.", "_____no_output_____" ] ], [ [ "sorted_idx = sorted(range(len(lengths)), key=lengths.__getitem__, reverse=True)\nsorted_X = [X[idx] for idx in sorted_idx]\n\nfor sequence in sorted_X:\n print(sequence)", "tensor([12, 16, 13, 3, 1, 8, 18, 14, 3, 18, 6, 14, 1])\ntensor([16, 10, 2, 2, 13, 8, 7, 13, 3, 2, 5])\ntensor([18, 15, 2, 18, 6, 2, 15, 1, 14, 13, 17])\ntensor([ 4, 14, 5, 17, 14, 11, 16, 1])\ntensor([ 9, 15, 1, 16])\n" ] ], [ [ "`pack_sequence`를 이용하여 PackedSequence를 만들어보겠습니다.", "_____no_output_____" ] ], [ [ "packed_sequence = pack_sequence(sorted_X)\n\nprint(packed_sequence)", "PackedSequence(data=tensor([12, 16, 18, 4, 9, 16, 10, 15, 14, 15, 13, 2, 2, 5, 1, 3, 2, 18,\n 17, 16, 1, 13, 6, 14, 8, 8, 2, 11, 18, 7, 15, 16, 14, 13, 1, 1,\n 3, 3, 14, 18, 2, 13, 6, 5, 17, 14, 1]), batch_sizes=tensor([5, 5, 5, 5, 4, 4, 4, 4, 3, 3, 3, 1, 1]), sorted_indices=None, unsorted_indices=None)\n" ] ], [ [ "### Embedding 적용해보기\n자 이제, `PackedSequence`와 padding이 된 Tensor인 `PaddedSequence`를 만들어보았으니, RNN에 input으로 넣어서 테스트해보려고 합니다.\n\n그 전에, 위에 예제들에서는 input이 character의 index들을 가지고 있는 데이터였지만, 보통은 주로 이를 embedding한 값을 RNN의 input으로 넣어줍니다.\n\n이 튜토리얼에서는 one-hot character embedding을 해보도록 하겠습니다.", "_____no_output_____" ] ], [ [ "eye = torch.eye(len(char_set))\nembedded_tensor = eye[padded_sequence]\nprint(embedded_tensor.shape)", "torch.Size([5, 13, 19])\n" ], [ "embedded_packed_seq = pack_sequence([eye[X[idx]] for idx in sorted_idx])\n\nprint(embedded_packed_seq.data.shape)", "torch.Size([47, 19])\n" ] ], [ [ "### RNN 모델 만들기", "_____no_output_____" ] ], [ [ "rnn = torch.nn.RNN(input_size=len(char_set), hidden_size=30, batch_first=True)", "_____no_output_____" ] ], [ [ "`PaddedSequence`를 이용하여 RNN에 넣어봅시다.", "_____no_output_____" ] ], [ [ "rnn_output, hidden = rnn(embedded_tensor)\n\nprint(rnn_output.shape)\nprint(hidden.shape)", "torch.Size([5, 13, 30])\ntorch.Size([1, 5, 30])\n" ] ], [ [ "`PackedSequence`를 이용하여 RNN에 넣어봅시다.", "_____no_output_____" ] ], [ [ "rnn_output, hidden = rnn(embedded_packed_seq)\n\nprint(rnn_output.data.shape)\nprint(hidden.data.shape)", "torch.Size([47, 30])\ntorch.Size([1, 5, 30])\n" ] ], [ [ "### `pad_packed_sequence`\n\n위 함수는 `PackedSequence`를 `PaddedSequence`(Tensor)로 바꾸어주는 함수입니다.\n\n`PackedSequence`는 각 sequence에 대한 길이 정보도 가지고있기 때문에, 이 함수는 Tensor와 함께 길이에 대한 리스트를 튜플로 리턴해줍니다.\n\n리턴값: (Tensor, list_of_lengths)", "_____no_output_____" ] ], [ [ "unpacked_sequence, seq_lengths = pad_packed_sequence(embedded_packed_seq, batch_first=True)\n\nprint(unpacked_sequence.shape)\nprint(seq_lengths)", "torch.Size([5, 13, 19])\ntensor([13, 11, 11, 8, 4])\n" ] ], [ [ "## # `pack_padded_sequence`\n반대로, Padding이 된 Tensor인 `PaddedSequence`를 `PackedSequence`로 바꾸어주는 함수도 있습니다.\n\n`pack_padded_sequence` 함수는 실제 sequence길이에 대한 정보를 모르기때문에, 파라미터로 꼭 제공해주어야합니다.\n\n여기서 주의하여야 할 점은, input인 `PaddedSequence`가 아까 언급드린 **길이에 따른 내림차순으로 정렬되어야 한다는** 조건이 성립되어야 `PackedSequence`로 올바르게 변환될 수 있습니다.\n\n아까 저희가 만든 `padded_sequence` 변수는 이 조건을 만족하지 않기 때문에 다시 새로 만들어보겠습니다.", "_____no_output_____" ] ], [ [ "embedded_padded_sequence = eye[pad_sequence(sorted_X, batch_first=True)]\n\nprint(embedded_padded_sequence.shape)", "torch.Size([5, 13, 19])\n" ] ], [ [ "이제 이 padding이 된 Tensor를 `PackedSequence`로 변환해보겠습니다.", "_____no_output_____" ] ], [ [ "sorted_lengths = sorted(lengths, reverse=True)\nnew_packed_sequence = pack_padded_sequence(embedded_padded_sequence, sorted_lengths, batch_first=True)\nprint(new_packed_sequence.data.shape)\nprint(new_packed_sequence.batch_sizes)", "torch.Size([47, 19])\ntensor([5, 5, 5, 5, 4, 4, 4, 4, 3, 3, 3, 1, 1])\n" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code" ]
[ [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ] ]
ece7c7f212eb77669a8de9dca706a78c89b50114
16,845
ipynb
Jupyter Notebook
Google Analytics/Google_Analytics_Get_stats_per_country.ipynb
techthiyanes/awesome-notebooks
10ab4da1b94dfa101e908356a649609b0b17561a
[ "BSD-3-Clause" ]
null
null
null
Google Analytics/Google_Analytics_Get_stats_per_country.ipynb
techthiyanes/awesome-notebooks
10ab4da1b94dfa101e908356a649609b0b17561a
[ "BSD-3-Clause" ]
null
null
null
Google Analytics/Google_Analytics_Get_stats_per_country.ipynb
techthiyanes/awesome-notebooks
10ab4da1b94dfa101e908356a649609b0b17561a
[ "BSD-3-Clause" ]
null
null
null
29.500876
311
0.632473
[ [ [ "<img width=\"10%\" alt=\"Naas\" src=\"https://landen.imgix.net/jtci2pxwjczr/assets/5ice39g4.png?w=160\"/>", "_____no_output_____" ], [ "# Google Analytics - Get stats per country\n<a href=\"https://app.naas.ai/user-redirect/naas/downloader?url=https://raw.githubusercontent.com/jupyter-naas/awesome-notebooks/master/Google%20Analytics/Google_Analytics_Get_stats_per_country.ipynb\" target=\"_parent\"><img src=\"https://naasai-public.s3.eu-west-3.amazonaws.com/open_in_naas.svg\"/></a>", "_____no_output_____" ], [ "**Tags:** #googleanalytics #statspercountry #marketing #analytics #automation #image #csv #html #plotly", "_____no_output_____" ], [ "**Author:** [Charles Demontigny](https://www.linkedin.com/in/charles-demontigny/)", "_____no_output_____" ], [ "Pre-requisite: Create your own <a href=\"\">Google API JSON credential</a>", "_____no_output_____" ], [ "## Input", "_____no_output_____" ], [ "### Import library", "_____no_output_____" ] ], [ [ "try:\n import pycountry\nexcept:\n !pip install pycountry\n import pycountry\nimport plotly.graph_objects as go\nimport plotly.express as px\nimport naas\nfrom naas_drivers import googleanalytics", "_____no_output_____" ] ], [ [ "### Get your credential from Google Cloud Platform", "_____no_output_____" ] ], [ [ "json_path = 'naas-googleanalytics.json'", "_____no_output_____" ] ], [ [ "### Get view id from google analytics", "_____no_output_____" ] ], [ [ "view_id = \"228952707\"", "_____no_output_____" ] ], [ [ "### Schedule your notebook", "_____no_output_____" ] ], [ [ "naas.scheduler.add(cron=\"0 8 * * *\")\nnaas.dependency.add(json_path)\n\n#-> Uncomment the line below (by removing the hashtag) to remove your scheduler\n# naas.scheduler.delete()", "_____no_output_____" ] ], [ [ "## Model", "_____no_output_____" ], [ "### Visitor's country of origin", "_____no_output_____" ] ], [ [ "df_country = googleanalytics.connect(json_path=json_path).views.get_data(\n view_id,\n metrics=\"ga:sessions\",\n pivots_dimensions=\"ga:country\",\n dimensions=\"ga:month\",\n start_date=None,\n end_date=None,\n format_type=\"pivot\"\n)\ndf_country", "_____no_output_____" ], [ "sessions_per_country = googleanalytics.connect(json_path=json_path).views.get_country(view_id) # default: metrics=\"ga:sessions\"", "_____no_output_____" ], [ "sessions_per_country", "_____no_output_____" ], [ "users_per_country = googleanalytics.views.get_country(view_id, metrics=\"ga:users\") ", "_____no_output_____" ] ], [ [ "## Output", "_____no_output_____" ], [ "### Display result", "_____no_output_____" ] ], [ [ "sessions_per_country.head()", "_____no_output_____" ], [ "users_per_country.head()", "_____no_output_____" ], [ "sessions_per_country = sessions_per_country.reset_index().rename(columns={\"index\": \"Country\"})\nmapping = {country.name: country.alpha_3 for country in pycountry.countries}\nsessions_per_country['iso_alpha'] = sessions_per_country['Country'].apply(lambda x: mapping.get(x))", "_____no_output_____" ], [ "sessions_per_country", "_____no_output_____" ], [ "fig = px.choropleth(sessions_per_country, locations=\"iso_alpha\",\n color=\"Sessions\", \n hover_name=\"Country\",\n color_continuous_scale=\"Greens\")\nfig.show()", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code" ]
[ [ "markdown", "markdown", "markdown", "markdown", "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code", "code", "code", "code" ], [ "markdown", "markdown" ], [ "code", "code", "code", "code", "code" ] ]
ece7cc92cea46552674eeee3ae5eab246198c496
6,199
ipynb
Jupyter Notebook
Choropleth_map_of_India_for_Covid/realtime-visualization-of-covid-19-data-of-india.ipynb
anishsingh20/COVID-19-INDIA-TRACKER
2185caf16c08a03d92936f4fca88a2e4b870e4ed
[ "Apache-2.0" ]
6
2020-03-27T06:09:12.000Z
2022-01-24T11:42:12.000Z
Choropleth_map_of_India_for_Covid/realtime-visualization-of-covid-19-data-of-india.ipynb
anishsingh20/COVID-19-INDIA-TRACKER
2185caf16c08a03d92936f4fca88a2e4b870e4ed
[ "Apache-2.0" ]
null
null
null
Choropleth_map_of_India_for_Covid/realtime-visualization-of-covid-19-data-of-india.ipynb
anishsingh20/COVID-19-INDIA-TRACKER
2185caf16c08a03d92936f4fca88a2e4b870e4ed
[ "Apache-2.0" ]
5
2020-05-04T16:36:04.000Z
2020-10-02T08:38:44.000Z
6,199
6,199
0.724149
[ [ [ "## The global spread of corona virus has motivated to data analysts to come up with ways to visualise country-wise data using various techniques. However, I went a step ahead and made an ****attempt to visualise my country's (India) COVID-19 data****. This notebook shows choropleth map visualisations of this data. Instead of using a static dataset, I've used data directly from a webservice, where the data is updated on a daily basis. Therefore, this notebook shows new visualisations everytime it's run.", "_____no_output_____" ], [ "## A List of libraries and modules used:\n1. Geopandas\n2. Pandas\n3. requests\n4. Matplotlib", "_____no_output_____" ], [ "# Choropleth map using [Geopandas](http://https://geopandas.org/#:~:text=GeoPandas%20is%20an%20open%20source,operations%20are%20performed%20by%20shapely.) library\nFor these maps, I have used two data sources. The first is a dataset of COVID-19 cases in every state of India. The aim is to produce a map on real time data. Therefore, instead of using a static datasource, like a .csv or an excel file, I have used the data derived from a web service. This code directly fetches data from the web using a request module. The data from this service is updated on a daily basis. So, whenever the code is run, the map is updated automatically. The data can be accessed from: \n\n[https://covid-india-cases.herokuapp.com/states/](http://)\n\nThe requests module fetces data from the source provided.", "_____no_output_____" ] ], [ [ "pip install geopandas", "_____no_output_____" ], [ "import geopandas as gpd\nimport pandas as pd\nimport requests\nimport matplotlib.pyplot as plt\n%matplotlib inline", "_____no_output_____" ] ], [ [ "## Fetching COVID 19 Data for Indian States ", "_____no_output_____" ] ], [ [ "raw= requests.get(\"https://covid-india-cases.herokuapp.com/states/\")\nraw_json = raw.json()\ndf = pd.DataFrame(raw_json)\ndf.head()", "_____no_output_____" ] ], [ [ "Keeping only the desired columns in dataframe", "_____no_output_____" ] ], [ [ "df=df[[\"state\",\"active\",\"deaths\",\"cured\"]]\ndf.head()", "_____no_output_____" ] ], [ [ "## Getting data for the map\nThe next dataset that is required is a shapefile of all the states. A shape file consists of the geometry of each state on a map. For this I have used the Geopandas library", "_____no_output_____" ] ], [ [ "fp = \"../input/india-shapefile/INDIA.shp\"\nmap_df = gpd.read_file(fp)\nmap_df.head()", "_____no_output_____" ], [ "map_df.plot()", "_____no_output_____" ] ], [ [ "## Replacing the names of some countries so that they match with the ones in the data provided by the webservice", "_____no_output_____" ] ], [ [ "map_df[\"ST_NAME\"].replace({\"DADRA AND NAGAR HAVELI\": \"Dadra and Nagar Haveli and Daman and Diu\", \"Andaman & Nicobar Island\": \"Andaman and Nicobar Islands\", \"Pondicherry\": \"Puducherry\",\"Orissa\":\"Odisha\",\"Nct Of Delhi\":\"Delhi\",\"LAKSHADWEEP\":\"Lakshadweep\",\"CHANDIGARH\":\"Chandigarh\",\"DAMAN AND DIU\":\"Dadra and Nagar Haveli and Daman and Diu\",\"ANDAMAN AND NICOBAR ISLANDS\":\"Andaman and Nicobar Islands\",\"Jammu And Kashmir\":\"Jammu and Kashmir\"}, inplace=True)", "_____no_output_____" ] ], [ [ "## Merging both the datasets to obtain a final dataset with the state's name, it’s geometry and the COVID-19 data", "_____no_output_____" ] ], [ [ "merged = map_df.set_index('ST_NAME').join(df.set_index('state'))\nmerged.head()", "_____no_output_____" ] ], [ [ "# Plotting the map\nSince it's a choropleth map, we first have to specify a variable which decides the colour of a particular state.\nFor the first map, I want the criteria to be the number of active cases. Therefore, the column with the name 'active' is passed to the plot function to make the corresponding map", "_____no_output_____" ] ], [ [ "variable = 'active'\n# create figure and axes for Matplotlib\nfig, ax = plt.subplots(1,figsize=(18, 13))\nax.axis('off')\nmerged.plot(column=variable, cmap='bone_r', linewidth=0.5, ax=ax, edgecolor='0.6', legend='True')", "_____no_output_____" ] ], [ [ "# Map of recovered cases", "_____no_output_____" ] ], [ [ "variable2 = 'cured'\n# create figure and axes for Matplotlib\nfig, ax = plt.subplots(1,figsize=(18, 13))\nax.axis('off')\nmerged.plot(column=variable2, cmap='Greens', linewidth=0.5, ax=ax, edgecolor='0.6', legend='True')", "_____no_output_____" ] ], [ [ "# Map of death cases", "_____no_output_____" ] ], [ [ "variable3 = 'deaths'\n# create figure and axes for Matplotlib\nfig, ax = plt.subplots(1,figsize=(18, 13))\nax.axis('off')\nmerged.plot(column=variable3, cmap='Reds', linewidth=0.5, ax=ax, edgecolor='0.6', legend='True')", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code" ]
[ [ "markdown", "markdown", "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ] ]
ece7f5ea176c65daa1029e5cfe84246ecc3bf8e7
614,408
ipynb
Jupyter Notebook
day1/5_modelmnist.ipynb
bkamins/ComplexNetworks2019
d652364e1e66aa89e96d3a726652fbf3835c73f3
[ "MIT" ]
3
2019-08-20T13:07:16.000Z
2021-03-12T01:34:47.000Z
day1/5_modelmnist.ipynb
bkamins/ComplexNetworks2019
d652364e1e66aa89e96d3a726652fbf3835c73f3
[ "MIT" ]
null
null
null
day1/5_modelmnist.ipynb
bkamins/ComplexNetworks2019
d652364e1e66aa89e96d3a726652fbf3835c73f3
[ "MIT" ]
2
2019-08-12T23:29:15.000Z
2020-01-30T22:46:41.000Z
98.336748
40,403
0.724576
[ [ [ "empty" ] ] ]
[ "empty" ]
[ [ "empty" ] ]
ece7fb5ac8f6be4eeb479b17b8b1d5e5556bc968
5,350
ipynb
Jupyter Notebook
examples/hello_world/hello_world.ipynb
antmicro/tensorflow-arduino-examples
47662b61cea883766396c26c6e01a238c67af655
[ "Apache-2.0" ]
16
2020-12-22T17:55:06.000Z
2022-03-22T19:22:19.000Z
examples/hello_world/hello_world.ipynb
antmicro/tensorflow-arduino-examples
47662b61cea883766396c26c6e01a238c67af655
[ "Apache-2.0" ]
5
2020-12-28T08:31:41.000Z
2021-01-22T12:39:13.000Z
examples/hello_world/hello_world.ipynb
antmicro/tensorflow-arduino-examples
47662b61cea883766396c26c6e01a238c67af655
[ "Apache-2.0" ]
6
2020-12-22T17:54:32.000Z
2022-03-22T19:22:22.000Z
30.924855
323
0.616822
[ [ [ "![Renode](https://dl.antmicro.com/projects/renode/renode.png)\n<table align=\"left\">\n <td>\n <a target=\"_blank\" href=\"https://colab.research.google.com/github/antmicro/tensorflow-arduino-examples/blob/master/examples/hello_world/hello_world.ipynb\"><img src=\"https://raw.githubusercontent.com/antmicro/tensorflow-arduino-examples/master/examples/.static/view-in-colab.png\" />Run in Google Colab</a>\n </td>\n <td>\n <a target=\"_blank\" href=\"https://github.com/antmicro/tensorflow-arduino-examples/blob/master/examples/hello_world/hello_world.ipynb\"><img src=\"https://raw.githubusercontent.com/antmicro/tensorflow-arduino-examples/master/examples/.static/view-ipynb.png\" />View ipynb on GitHub</a>\n </td>\n <td>\n <a target=\"_blank\" href=\"https://github.com/antmicro/tensorflow-arduino-examples/blob/master/examples/hello_world/hello_world.py\"><img src=\"https://raw.githubusercontent.com/antmicro/tensorflow-arduino-examples/master/examples/.static/view-source.png\" />View Python source on GitHub</a>\n </td>\n</table>", "_____no_output_____" ], [ "## Install requirements", "_____no_output_____" ] ], [ [ "!pip install -q git+https://github.com/antmicro/pyrenode.git git+https://github.com/antmicro/renode-colab-tools.git\n!mkdir -p renode && cd renode && wget https://dl.antmicro.com/projects/renode/builds/renode-latest.linux-portable.tar.gz && tar -xzf renode-latest.linux-portable.tar.gz --strip 1\n!pip install -q -r renode/tests/requirements.txt\n\nimport os\nfrom renode_colab_tools import metrics\nos.environ['PATH'] = os.getcwd()+\"/renode:\"+os.environ['PATH']\nos.environ['TENSORFLOW_PATH'] = os.getcwd()+\"/tensorflow-arduino-examples/tensorflow\"", "_____no_output_____" ], [ "!mkdir -p binaries/hello_world && cd binaries/hello_world && wget https://github.com/antmicro/tensorflow-arduino-examples-binaries/raw/master/hello_world/hello_world.ino.elf # fetch prebuilt binaries", "_____no_output_____" ] ], [ [ "## Run the hello_world example in Renode", "_____no_output_____" ] ], [ [ "import time\nfrom pyrenode import *\nshutdown_renode()\nconnect_renode() # this sets up a log file, and clears the simulation (just in case)\ntell_renode('using sysbus')\ntell_renode('mach create')\ntell_renode('machine LoadPlatformDescription @platforms/boards/arduino_nano_33_ble.repl')\ntell_renode('sysbus LoadELF @binaries/hello_world/hello_world.ino.elf')\n\ntell_renode('uart0 CreateFileBackend @uart.dump true')\ntell_renode('logLevel 3')\ntell_renode('machine EnableProfiler @metrics.dump')\ntell_renode('s')\nwhile not os.path.exists('renode/uart.dump'):\n time.sleep(1) #waits for creating uart.dump\n!timeout 60 tail -c+2 -f renode/uart.dump | sed '/^1$/ q'\ntell_renode('q')\nexpect_cli('Renode is quitting')\ntime.sleep(1) #wait not to kill Renode forcefully\nshutdown_renode()", "_____no_output_____" ] ], [ [ "## Renode metrics analysis", "_____no_output_____" ] ], [ [ "from renode.tools.metrics_analyzer.metrics_parser import MetricsParser\nmetrics.init_notebook_mode(connected=False)\nparser = MetricsParser('renode/metrics.dump')", "_____no_output_____" ], [ "metrics.configure_plotly_browser_state()\nmetrics.show_executed_instructions(parser)", "_____no_output_____" ], [ "metrics.configure_plotly_browser_state()\nmetrics.show_memory_access(parser)", "_____no_output_____" ], [ "metrics.configure_plotly_browser_state()\nmetrics.show_exceptions(parser)", "_____no_output_____" ], [ "metrics.configure_plotly_browser_state()\nmetrics.show_peripheral_access(parser)", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code" ]
[ [ "markdown", "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code" ] ]
ece801e14b6d1617e9519e9c83f65300fea76f41
159,026
ipynb
Jupyter Notebook
1st component/Step 1 - Process EDA, accelerometer and skin temperature data.ipynb
ddritsa/PhD-Thesis-repository
512850326fcdd56e29ea519a054a9668d92c2046
[ "MIT" ]
null
null
null
1st component/Step 1 - Process EDA, accelerometer and skin temperature data.ipynb
ddritsa/PhD-Thesis-repository
512850326fcdd56e29ea519a054a9668d92c2046
[ "MIT" ]
null
null
null
1st component/Step 1 - Process EDA, accelerometer and skin temperature data.ipynb
ddritsa/PhD-Thesis-repository
512850326fcdd56e29ea519a054a9668d92c2046
[ "MIT" ]
null
null
null
402.597468
37,528
0.93298
[ [ [ "import sys\n# Put here the directory containing the folder where the definitions are saved\nsys.path.insert(0, r'C:\\Users\\demdr\\UTS\\Important files for PhD thesis\\Definitions')\n", "_____no_output_____" ], [ "import matplotlib.pyplot as plt\nimport pandas as pd \nimport numpy as np\n\nfrom definitions_for_opening_E4_files_and_making_a_df import create_final_dict\nfrom definitions_for_EDA_feature_extraction_and_activity_recognition import resample_EDA_df, create_EDA_features, detect_activity\nfrom definitions_for_activity_recognition import detect_change_of_intensity", "Using TensorFlow backend.\nc:\\users\\demdr\\python37\\lib\\site-packages\\tensorflow\\python\\framework\\dtypes.py:526: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) / '(1,)type'.\n _np_qint8 = np.dtype([(\"qint8\", np.int8, 1)])\nc:\\users\\demdr\\python37\\lib\\site-packages\\tensorflow\\python\\framework\\dtypes.py:527: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) / '(1,)type'.\n _np_quint8 = np.dtype([(\"quint8\", np.uint8, 1)])\nc:\\users\\demdr\\python37\\lib\\site-packages\\tensorflow\\python\\framework\\dtypes.py:528: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) / '(1,)type'.\n _np_qint16 = np.dtype([(\"qint16\", np.int16, 1)])\nc:\\users\\demdr\\python37\\lib\\site-packages\\tensorflow\\python\\framework\\dtypes.py:529: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) / '(1,)type'.\n _np_quint16 = np.dtype([(\"quint16\", np.uint16, 1)])\nc:\\users\\demdr\\python37\\lib\\site-packages\\tensorflow\\python\\framework\\dtypes.py:530: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) / '(1,)type'.\n _np_qint32 = np.dtype([(\"qint32\", np.int32, 1)])\nc:\\users\\demdr\\python37\\lib\\site-packages\\tensorflow\\python\\framework\\dtypes.py:535: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) / '(1,)type'.\n np_resource = np.dtype([(\"resource\", np.ubyte, 1)])\n" ], [ "# Put here the directory of the folder where the EDA data is saved \n# This should be the folder with the EDA data, which is downloaded from the Empatica E4 manager\n# For downloading the Empatica E4 data from the E4 manager, see here the directions: https://support.empatica.com/hc/en-us/articles/206389995-Download-your-data-from-the-E4-wristband\n# This code uses the files 'EDA.csv', 'ACC.csv', 'HR.csv' and 'TEMP.csv' from this folder\n\n# Ideally we should have a separate folder for each user, to store their data\n# Otherwise we have to tweak a bit the initial code in the 'definitions_for_opening_E4_files_and_making_a_df' file\n\n# ATTENTION: The path that we specify should correspond to a folder that has another folder (or more), containing the following csv files \n# of the session that we want to analyse: \n# 'EDA.csv', 'ACC.csv', 'HR.csv' and 'TEMP.csv' \n# DO NOT put directly the path of the final folder that contains the csv files \n# Instead, put the path of its parent folder\n# For instance, imagine the following directories: \n# ...\\Project data\\Raw data\\Participants: The name of the folder containing all the raw data from all participants\n# ...\\Project data\\Raw data\\Participants\\A : The path of the folder containing the data of participant A within that folder\n# ...\\Project data\\Raw data\\Participants\\A\\EDA data : The folder where we have all the unzipped EDA data for this participant, from many recording sessions\n# ...\\Project data\\Raw data\\Participants\\A\\EDA data\\Session 1 : This is the parent folder that we want. 'Session 1' is just a name for this session. Inside, it has another folder with unzipped EDA data for one session. \n# ...\\Project data\\Raw data\\Participants\\A\\EDA data\\Session 1\\1460023847_A00A98 : The name of the unzipped folder with the EDA data for this session\n# Inside this folder, there are csv files for EDA, ACC, HR, TEMP, like this:\n# ...\\Project data\\Raw data\\Participants\\A\\EDA data\\Session 1\\1460023847_A00A98\\EDA.csv ...and so on\n# From all these directories, we put the directory '...\\Project data\\Raw data\\Participants\\A\\EDA data\\Session 1' \n# to analyse the data of session 1 for user A\n\n#this_directory = 'XXXXXXXXX' # replace this with the appropriate directory name\n#For the example described above, we would put the following directory: \nthis_directory = r'C:\\Users\\demdr\\Desktop\\Testing the thesis functions\\Project data\\Raw data\\Participants\\A\\EDA data\\Session 1'\n#participant_ID = 'YYYYYYYY' # Put here the participant ID; this could be a randomly assigned string or integer or anything else \n# For the example described above, we would do the following:\nparticipant_ID = 'A'\n\nall_data_together = create_final_dict(this_directory)", "2019-07-16 13:09:55+10:00\n0.04247095584869385\n" ] ], [ [ "### EDA and activity feature extraction", "_____no_output_____" ] ], [ [ "for_final_EDA_df = pd.DataFrame()\n\nfor ses in all_data_together['E4_Session'].unique():\n this_df = all_data_together[all_data_together['E4_Session']==ses]\n\n # The following are the definitions for feature extraction:\n this_df_resampled = resample_EDA_df(this_df)\n this_df_resampled = create_EDA_features(this_df_resampled)\n #activity recognition\n this_df_resampled = detect_activity(this_df_resampled)\n this_df_resampled\n # Plot some of the extracted data if we want\n plt.figure(figsize=(20,3))\n plt.plot(this_df_resampled['EDA'])\n plt.plot(this_df_resampled['EDA_tonic'])\n plt.title('Extracted tonic EDA')\n plt.figure(figsize=(20,3))\n plt.plot(this_df_resampled['EDA artifact'])\n plt.title('EDA artefacts')\n plt.figure(figsize=(20,3))\n plt.plot(this_df_resampled['EDR_amplitude'])\n plt.title('EDA amplitude')\n plt.figure(figsize=(20,3))\n plt.plot(this_df_resampled['activity'])\n plt.title('Classified activity from accelerometer data')\n\n \n new_df = this_df_resampled.copy()\n new_df['Detected Activity']=new_df['activity'].copy()\n new_df = detect_change_of_intensity(new_df)\n\n\n this_df_resampled['Change of activity state'] = new_df['Change of activity state_After prediction']\n this_df_resampled['Steady state'] = new_df['Steady state']\n this_df_resampled['Change of movement intensity'] = new_df['Change of movement intensity_After prediction']\n this_df_resampled['Spontaneous movement'] = new_df['Spontaneous movement_After prediction']\n \n for_final_EDA_df = for_final_EDA_df.append(this_df_resampled)\n ", "art\npercentage of artifacts: 22 9995 0.0022011005502751376\n[ 0 80 120 280 320 520 560 1040 1080 1200 1240 1760 2920 3440\n 3560 3960 4000 4520 4560 5080 5440 6160 6200 6240 7400 7560 8200 8240\n 8280 8320 8360 8480 8560 8720 8760 8920 9040 9080 9160 9400 9440 9480\n 9520 9560 9720 9760 9800 9880 9961]\n" ] ], [ [ "#### visualise the extracted changes in activity", "_____no_output_____" ] ], [ [ "\nfor ses in for_final_EDA_df['E4_Session'].unique():\n this_df = for_final_EDA_df[for_final_EDA_df['E4_Session']==ses]\n plt.figure(figsize=(20,3))\n plt.plot(this_df['Change of activity state'])\n plt.title('Change of activity state')\n plt.figure(figsize=(20,3))\n plt.plot(this_df['Change of movement intensity'])\n plt.title('Change of movement intensity')\n\n", "_____no_output_____" ] ], [ [ "#### Construct a df with only selected columns", "_____no_output_____" ] ], [ [ "for_final_EDA_df['HR_from e4']=for_final_EDA_df['HR'].copy()\nfor_final_EDA_df['Participant_ID']=participant_ID\n\n\nfinal_EDA_df = pd.concat([for_final_EDA_df['EDA'],\nfor_final_EDA_df['ACC_1'],\nfor_final_EDA_df['ACC_2'],\nfor_final_EDA_df['ACC_3'],\nfor_final_EDA_df['HR_from e4'],\nfor_final_EDA_df['EDA_session'],\nfor_final_EDA_df['EDA artifact'],\nfor_final_EDA_df['EDA artifact'],\nfor_final_EDA_df['Stress'],\nfor_final_EDA_df['TEMP_smoothed'],\nfor_final_EDA_df['EDA_tonic'],\nfor_final_EDA_df['EDR_amplitude'],\nfor_final_EDA_df['EDA_duration'],\nfor_final_EDA_df['EDA_duration'],\nfor_final_EDA_df['percentage_of_EDA_artifacts'],\nfor_final_EDA_df['activity'],\nfor_final_EDA_df['activity_2s'],\nfor_final_EDA_df['Change of activity state'],\nfor_final_EDA_df['Steady state'],\nfor_final_EDA_df['Change of movement intensity'],\nfor_final_EDA_df['Spontaneous movement']],axis=1)\n\n\n\n\n\nfinal_EDA_df = final_EDA_df.fillna(method='bfill')\nfinal_EDA_df = final_EDA_df.fillna(method='ffill')\n", "_____no_output_____" ] ], [ [ "### Save here the extracted data", "_____no_output_____" ] ], [ [ "# Put here the directory for saving the constructed file \n#path_to_save = r'ZZZZZZZZ.csv' \n# This file contains the processed EDA and activity data,with timestamps\n\n# For the example described above, we would use the following file for saving the csv: \npath_to_save = r'C:\\Users\\demdr\\Desktop\\Testing the thesis functions\\Project data\\Analysed data\\Participants\\ID\\E4 data\\Session 1.csv'\npath_to_save = path_to_save.replace('ID',participant_ID)\nprint(path_to_save)\n\n\n\nfinal_EDA_df.to_csv(path_to_save)", "C:\\Users\\demdr\\Desktop\\Testing the thesis functions\\Project data\\Analysed data\\Participants\\A\\E4 data\\Session 1.csv\n" ] ] ]
[ "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code" ]
[ [ "code", "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ] ]
ece80540a8863916d4a3b359206d9c185469e702
385,513
ipynb
Jupyter Notebook
example/Demo.ipynb
k-utsubo/bert_score
1e5d0abc298829d693d574032b781c63409a71b3
[ "MIT" ]
null
null
null
example/Demo.ipynb
k-utsubo/bert_score
1e5d0abc298829d693d574032b781c63409a71b3
[ "MIT" ]
null
null
null
example/Demo.ipynb
k-utsubo/bert_score
1e5d0abc298829d693d574032b781c63409a71b3
[ "MIT" ]
null
null
null
561.154294
129,218
0.944848
[ [ [ "## BERTScore Tutorial", "_____no_output_____" ], [ "### Installation\nif you have not installed `bert_score`, it is very easy\nsimply uncomment the line below to install through pip", "_____no_output_____" ] ], [ [ "!pip install bert_score", "Collecting bert_score\n Using cached bert_score-0.3.10-py3-none-any.whl (59 kB)\nRequirement already satisfied: pandas>=1.0.1 in /Volumes/DATA/src/bert_score/venv/lib/python3.9/site-packages (from bert_score) (1.3.4)\nRequirement already satisfied: matplotlib in /Volumes/DATA/src/bert_score/venv/lib/python3.9/site-packages (from bert_score) (3.4.3)\nRequirement already satisfied: torch>=1.0.0 in /Volumes/DATA/src/bert_score/venv/lib/python3.9/site-packages (from bert_score) (1.10.0)\nRequirement already satisfied: packaging>=20.9 in /Volumes/DATA/src/bert_score/venv/lib/python3.9/site-packages (from bert_score) (21.2)\nRequirement already satisfied: requests in /Volumes/DATA/src/bert_score/venv/lib/python3.9/site-packages (from bert_score) (2.26.0)\nRequirement already satisfied: transformers>=3.0.0numpy in /Volumes/DATA/src/bert_score/venv/lib/python3.9/site-packages (from bert_score) (4.12.3)\nRequirement already satisfied: tqdm>=4.31.1 in /Volumes/DATA/src/bert_score/venv/lib/python3.9/site-packages (from bert_score) (4.62.3)\nRequirement already satisfied: pyparsing<3,>=2.0.2 in /Volumes/DATA/src/bert_score/venv/lib/python3.9/site-packages (from packaging>=20.9->bert_score) (2.4.7)\nRequirement already satisfied: python-dateutil>=2.7.3 in /Volumes/DATA/src/bert_score/venv/lib/python3.9/site-packages (from pandas>=1.0.1->bert_score) (2.8.2)\nRequirement already satisfied: pytz>=2017.3 in /Volumes/DATA/src/bert_score/venv/lib/python3.9/site-packages (from pandas>=1.0.1->bert_score) (2021.3)\nRequirement already satisfied: numpy>=1.20.0 in /Volumes/DATA/src/bert_score/venv/lib/python3.9/site-packages (from pandas>=1.0.1->bert_score) (1.21.4)\nRequirement already satisfied: six>=1.5 in /Volumes/DATA/src/bert_score/venv/lib/python3.9/site-packages (from python-dateutil>=2.7.3->pandas>=1.0.1->bert_score) (1.16.0)\nRequirement already satisfied: typing-extensions in /Volumes/DATA/src/bert_score/venv/lib/python3.9/site-packages (from torch>=1.0.0->bert_score) (3.10.0.2)\nRequirement already satisfied: regex!=2019.12.17 in /Volumes/DATA/src/bert_score/venv/lib/python3.9/site-packages (from transformers>=3.0.0numpy->bert_score) (2021.11.2)\nRequirement already satisfied: pyyaml>=5.1 in /Volumes/DATA/src/bert_score/venv/lib/python3.9/site-packages (from transformers>=3.0.0numpy->bert_score) (6.0)\nRequirement already satisfied: filelock in /Volumes/DATA/src/bert_score/venv/lib/python3.9/site-packages (from transformers>=3.0.0numpy->bert_score) (3.3.2)\nRequirement already satisfied: sacremoses in /Volumes/DATA/src/bert_score/venv/lib/python3.9/site-packages (from transformers>=3.0.0numpy->bert_score) (0.0.46)\nRequirement already satisfied: huggingface-hub<1.0,>=0.1.0 in /Volumes/DATA/src/bert_score/venv/lib/python3.9/site-packages (from transformers>=3.0.0numpy->bert_score) (0.1.2)\nRequirement already satisfied: tokenizers<0.11,>=0.10.1 in /Volumes/DATA/src/bert_score/venv/lib/python3.9/site-packages (from transformers>=3.0.0numpy->bert_score) (0.10.3)\nRequirement already satisfied: cycler>=0.10 in /Volumes/DATA/src/bert_score/venv/lib/python3.9/site-packages (from matplotlib->bert_score) (0.11.0)\nRequirement already satisfied: kiwisolver>=1.0.1 in /Volumes/DATA/src/bert_score/venv/lib/python3.9/site-packages (from matplotlib->bert_score) (1.3.2)\nRequirement already satisfied: pillow>=6.2.0 in /Volumes/DATA/src/bert_score/venv/lib/python3.9/site-packages (from matplotlib->bert_score) (8.4.0)\nRequirement already satisfied: urllib3<1.27,>=1.21.1 in /Volumes/DATA/src/bert_score/venv/lib/python3.9/site-packages (from requests->bert_score) (1.26.7)\nRequirement already satisfied: idna<4,>=2.5 in /Volumes/DATA/src/bert_score/venv/lib/python3.9/site-packages (from requests->bert_score) (3.3)\nRequirement already satisfied: certifi>=2017.4.17 in /Volumes/DATA/src/bert_score/venv/lib/python3.9/site-packages (from requests->bert_score) (2021.10.8)\nRequirement already satisfied: charset-normalizer~=2.0.0 in /Volumes/DATA/src/bert_score/venv/lib/python3.9/site-packages (from requests->bert_score) (2.0.7)\nRequirement already satisfied: joblib in /Volumes/DATA/src/bert_score/venv/lib/python3.9/site-packages (from sacremoses->transformers>=3.0.0numpy->bert_score) (1.1.0)\nRequirement already satisfied: click in /Volumes/DATA/src/bert_score/venv/lib/python3.9/site-packages (from sacremoses->transformers>=3.0.0numpy->bert_score) (8.0.3)\nInstalling collected packages: bert-score\nSuccessfully installed bert-score-0.3.10\n\u001b[33mWARNING: You are using pip version 21.2.3; however, version 21.3.1 is available.\nYou should consider upgrading via the '/Volumes/DATA/src/bert_score/venv/bin/python -m pip install --upgrade pip' command.\u001b[0m\n" ], [ "# check your installation\nimport bert_score\nbert_score.__version__", "_____no_output_____" ] ], [ [ "### preparation", "_____no_output_____" ] ], [ [ "# hide the loading messages\nimport logging\nimport transformers\ntransformers.tokenization_utils.logger.setLevel(logging.ERROR)\ntransformers.configuration_utils.logger.setLevel(logging.ERROR)\ntransformers.modeling_utils.logger.setLevel(logging.ERROR)", "_____no_output_____" ], [ "%matplotlib inline\nimport matplotlib.pyplot as plt\nfrom matplotlib import rcParams\n\nrcParams[\"xtick.major.size\"] = 0\nrcParams[\"xtick.minor.size\"] = 0\nrcParams[\"ytick.major.size\"] = 0\nrcParams[\"ytick.minor.size\"] = 0\n\nrcParams[\"axes.labelsize\"] = \"large\"\nrcParams[\"axes.axisbelow\"] = True\nrcParams[\"axes.grid\"] = True", "_____no_output_____" ] ], [ [ "## Function API", "_____no_output_____" ], [ "We will first demonstrate how to use the `score` function in `bert_score`, which is what you need to evaluate a set of machine generated outputs.", "_____no_output_____" ] ], [ [ "from bert_score import score", "_____no_output_____" ] ], [ [ "Inputs to `score` are a list of candidate sentences and a list of reference sentences. ", "_____no_output_____" ] ], [ [ "with open(\"hyps.txt\") as f:\n cands = [line.strip() for line in f]\n\nwith open(\"refs.txt\") as f:\n refs = [line.strip() for line in f]", "_____no_output_____" ], [ "cands", "_____no_output_____" ] ], [ [ "Let's have a look.", "_____no_output_____" ] ], [ [ "cands[0]", "_____no_output_____" ] ], [ [ "We are now ready to call the score function. Besides candidates and references, we need to speicify the bert model we are using. Since we are dealing with English sentences, we will use the *bert-base-uncased* model.", "_____no_output_____" ] ], [ [ "P, R, F1 = score(cands, refs, lang='en', verbose=True)", "Downloading: 100%|██████████| 482/482 [00:00<00:00, 202kB/s]\nDownloading: 100%|██████████| 878k/878k [00:00<00:00, 904kB/s] \nDownloading: 100%|██████████| 446k/446k [00:00<00:00, 684kB/s]\nDownloading: 100%|██████████| 1.29M/1.29M [00:01<00:00, 1.30MB/s]\nDownloading: 100%|██████████| 1.33G/1.33G [00:17<00:00, 81.4MB/s]\n" ] ], [ [ "The outputs of the `score` function are Tensors of precision, recall, and F1 respectively. Each Tensor has the same number of items with the candidate and reference lists. Each item in the list is a scalar, representing the score for the corresponding candidates and references.", "_____no_output_____" ] ], [ [ "F1", "_____no_output_____" ] ], [ [ "We can take the average of all candidate reference pairs to be the system level score.", "_____no_output_____" ] ], [ [ "print(f\"System level F1 score: {F1.mean():.3f}\")", "System level F1 score: 0.959\n" ] ], [ [ "It might also be very interestig to see the distribution of BERTScore.", "_____no_output_____" ] ], [ [ "import matplotlib.pyplot as plt", "_____no_output_____" ], [ "plt.hist(F1, bins=20)\nplt.xlabel(\"score\")\nplt.ylabel(\"counts\")\nplt.show()", "_____no_output_____" ] ], [ [ "Some contextual embedding models, like RoBERTa, often produce BERTScores in a very narrow range (as shown above, the range is roughly between 0.92 and 1). Although this artifact does not affect the ranking ability of BERTScore, it affects the readability. Therefore, we propose to apply \"baseline rescaling\" to adjust the output scores. More details on this feature can be found in [this post](https://github.com/Tiiiger/bert_score/blob/master/journal/rescale_baseline.md).", "_____no_output_____" ] ], [ [ "P, R, F1 = score(cands, refs, lang='en', rescale_with_baseline=True)", "_____no_output_____" ] ], [ [ "We can now see that the scores are much more spread out, which makes it easy to compare different examples.", "_____no_output_____" ] ], [ [ "plt.hist(F1, bins=20)\nplt.xlabel(\"score\")\nplt.ylabel(\"counts\")\nplt.show()", "_____no_output_____" ] ], [ [ "The `score` function also handles multiple references gracefully. Consider a candidate sentences with 3 references.", "_____no_output_____" ] ], [ [ "single_cands = ['I like lemons.']\nmulti_refs = [['I am proud of you.', 'I love lemons.', 'Go go go.']]", "_____no_output_____" ], [ "P_mul, R_mul, F_mul = score(single_cands, multi_refs, lang=\"en\", rescale_with_baseline=True)", "_____no_output_____" ] ], [ [ "The `score` function will return the best score among all the references automatically.", "_____no_output_____" ] ], [ [ "F_mul", "_____no_output_____" ] ], [ [ "To understand a text generation system better, we can visualize the matchings in BERTScore.", "_____no_output_____" ] ], [ [ "from bert_score import plot_example", "_____no_output_____" ], [ "plot_example(cands[0], refs[0], lang=\"en\")", "_____no_output_____" ] ], [ [ "Similarly, we can apply rescaling to adjust the similarity distribution to be more distinguishable.", "_____no_output_____" ] ], [ [ "plot_example(cands[0], refs[0], lang=\"en\", rescale_with_baseline=True)", "_____no_output_____" ] ], [ [ "## Object-oriented API", "_____no_output_____" ], [ "In practice, most of the time of calling the `score` function is spent on building the model. In situations when we want to call the `score` function repeatedly, it is better to cache the model in a `scorer` object. Hence, in `bert_score` we also provide an object-oriented API. \n\nThe `BERTScorer` class provides the two methods we have introduced above, `score` and `plot_example`.", "_____no_output_____" ] ], [ [ "from bert_score import BERTScorer", "_____no_output_____" ], [ "scorer = BERTScorer(lang=\"en\", rescale_with_baseline=True)", "_____no_output_____" ], [ "P, R, F1 = scorer.score(cands, refs)", "_____no_output_____" ], [ "F1", "_____no_output_____" ], [ "scorer.plot_example(cands[0], refs[0])", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code" ]
[ [ "markdown", "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code", "code", "code", "code", "code" ] ]
ece808d16745e560bb426e323ce843130bd938a8
150,819
ipynb
Jupyter Notebook
examples/SWAN_example.ipynb
ryancoe/MHKiT-Python
e2564d6f22a4a19aadd29568838366e58741f88c
[ "BSD-3-Clause" ]
null
null
null
examples/SWAN_example.ipynb
ryancoe/MHKiT-Python
e2564d6f22a4a19aadd29568838366e58741f88c
[ "BSD-3-Clause" ]
3
2020-06-23T17:16:38.000Z
2021-01-20T23:43:28.000Z
examples/SWAN_example.ipynb
ryancoe/MHKiT-Python
e2564d6f22a4a19aadd29568838366e58741f88c
[ "BSD-3-Clause" ]
null
null
null
104.445291
27,192
0.773351
[ [ [ "# MHKiT SWAN Example\n\nThis example notebook demonstrates the input and plotting of output data from the software [Simulating WAves Nearshore (SWAN)](http://swanmodel.sourceforge.net/) using MHKiT. In this example the [SNL-SWAN](https://github.com/SNL-WaterPower/SNL-SWAN) tutorial was run for a wave energy converter. The output was written in ASCII and binary (*.mat) files. This MHKiT example notebook demonstrates how to import these different files into MHKiT and plot the output data. First, we will import the MHKiT SWAN package and the other Python packages needed for this example. Secondly, we will create an operating system independent path to the folder housing the SWAN data used in this example `swan_data_folder` using the `join` funtion. ", "_____no_output_____" ] ], [ [ "from mhkit.wave.io import swan\nimport matplotlib.pyplot as plt\nfrom os.path import join\nimport pandas as pd\n\nswan_data_folder = join('data','wave','swan')", "_____no_output_____" ] ], [ [ "## Supported SWAN Output Files\n\nMHKiT currenlty supports block and table SWAN output files in ASCII or binary (*.mat) files. Detailed descriptions of these file types may be found in the [SWAN User Manual](http://swanmodel.sourceforge.net/download/zip/swanuse.pdf). In the following cells, SWAN table and block data will be imported, discussed, and plotted. Three SWAN output files will be imported:\n 1. An ASCII table file ('SWANOUT.DAT'), \n 2. An ASCII block file ('SWANOUTBlock.DAT') \n 3. A binary block file ('SWANOUT.mat') ", "_____no_output_____" ] ], [ [ "swan_table_file = join(swan_data_folder, 'SWANOUT.DAT')\nswan_block_file = join(swan_data_folder, 'SWANOUTBlock.DAT')\nswan_block_mat_file = join(swan_data_folder, 'SWANOUT.mat')", "_____no_output_____" ] ], [ [ "## Load SWAN Files with MHKiT\n\nTo load a supported SWAN file simply call the `swan.read_table` or `swan.read_block` as appropriate for the swan output. The MHKiT function will read in the SWAN output and return the data as a DataFrame for table data or a dictionary of DataFrames for block data with multiple quantities of interest written to the file. The MHKiT SWAN read function will also return any metadata that the file may contain which will vary based on the file type and options specified in the SWAN run. MHKiT requires that for block data written in ASCII format that the file was written with headers. The `swan.read_block` function accepts both binary and ASCII format by assuming that any non-'.mat' extension is ASCII format.", "_____no_output_____" ], [ "## SWAN Table Data and Metadata\n\nThe SWAN output table is parsed from the MHKiT funtion `swan.read_table` into a DataFrame that is displayed below. The DataFrame columns contain a series of x-points ('Xp'), y-points ('Yp'), and keyword values at a given (x,y) point. The keywords are specified in the SWAN user manual and here can be seen as: 'Hsig' (significant wave height), 'Dir' (average wave direction), 'RTpeak' (Relative peak period), 'TDir' (direction of the energy transport). ", "_____no_output_____" ] ], [ [ "swan_table, metadata_table = swan.read_table(swan_table_file)\nswan_table", "_____no_output_____" ] ], [ [ "In the cell below, metadata is written to screen and can be seen to be a dictionary of keywords which contains the SWAN run name, the type of table written, the version of SWAN run, the column headers, and the associated units.", "_____no_output_____" ] ], [ [ "metadata_table", "_____no_output_____" ] ], [ [ "## SWAN Block (ASCII) Data and Metadata\n\nMHKiT will read in block data as a Dictionary of DataFrames for each quantity of interest in the file. The Dictionary `swan_block` (shown below) is read using `swan.read_block` on the ASCII block data, and has the same four keys from the table data shown previously. In the cell below the DataFrame for the 'Significant wave height' is shown by accessing the Dictionary using the specified key. This DataFrame has indices and columns referring to a point on the grid and a value of significant wave height at each point. In the last code block, the metadata Dictionary is written to screen.", "_____no_output_____" ] ], [ [ "swan_block, metadata_block = swan.read_block(swan_block_file)\nswan_block.keys()", "_____no_output_____" ], [ "swan_block['Significant wave height']", "_____no_output_____" ], [ "metadata_block", "_____no_output_____" ] ], [ [ "## SWAN Block (.mat) Data and Metadata\n\nThe Block \"SWANOUT.mat\" file is a binary output from SWAN containing the same data as was shown for the ASCII block file above. The Dictionary `swan_block_mat` (shown below) is read using `swan.read_block` on the binary block data, and has the same four keys from the Table data shown previously. Looking at the first code block below it can be seen that the returned Dictionary keys are the SWAN variable names ('Hsig', 'Dir', 'RTpeak', 'TDir'). Looking at the DataFrame for the significant wave height ('Hsig') we can see that the indices and columns are the same as the previous block ASCII DataFrame but the values now contain six decimal places. One consideration for working with binary data is shown in the last cell block of the section where there is no metadata letting the user know the units of the data. For binary data, the user would need to check the run's SWAN input file.", "_____no_output_____" ] ], [ [ "swan_block_mat, metadata_block_mat = swan.read_block(swan_block_mat_file)\nswan_block_mat.keys()", "_____no_output_____" ], [ "swan_block_mat['Hsig']", "_____no_output_____" ], [ "metadata_block_mat", "_____no_output_____" ] ], [ [ "## Block to Table\n\nMHKiT provides functionality to convert SWAN block Dictionaries to table DataFrame format. This provides the user with the ability to easily process and manipulate data across multiple data types. The function converts each key to a column in a single DataFrame.", "_____no_output_____" ] ], [ [ "swan_block_as_table = swan.dictionary_of_block_to_table(swan_block)\nswan_block_mat_as_table = swan.dictionary_of_block_to_table(swan_block_mat)\nswan_block_as_table", "_____no_output_____" ] ], [ [ "## Example Plots from SWAN Data\n\nThis last section shows a couple of plots for the significant wave height using each of the imported results.", "_____no_output_____" ] ], [ [ "plt.figure()\nplt.tricontourf(swan_table.Xp, swan_table.Yp, \n swan_table.Hsig, levels=256)\ncbar = plt.colorbar()\ncbar.set_label('Significant wave height [m]')", "_____no_output_____" ], [ "plt.figure()\nplt.tricontourf(swan_block_mat_as_table.x, swan_block_mat_as_table.y, \n swan_block_mat_as_table.Hsig,\n levels=256, cmap='viridis')\ncbar = plt.colorbar()\ncbar.set_label('Significant wave height [m]')", "_____no_output_____" ], [ "plt.figure()\nplt.tricontourf(swan_block_as_table.x, swan_block_as_table.y, \n swan_block_as_table['Significant wave height'], \n levels=256, cmap='viridis')\ncbar = plt.colorbar()\ncbar.set_label('Significant wave height [m]')", "_____no_output_____" ] ], [ [ "## Plot Block Data\n\nLastly significant wave height is plotted from the block data using `imshow` with a reversed y-axis to show the same plot achieved above.", "_____no_output_____" ] ], [ [ "plt.figure()\nplt.imshow(swan_block_mat['Hsig'])\nplt.gca().invert_yaxis()\ncbar = plt.colorbar()\ncbar.set_label('Significant wave height [m]')", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code" ]
[ [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown" ], [ "code" ] ]
ece81604e39c41668426702675bbc2fcfb881b4b
45,857
ipynb
Jupyter Notebook
docs/tutorials/03-kglab_pytorch_geometric.ipynb
issam9/rubrix
16cafb10ca60bb75dd716a33eb8e5149bdbedb7b
[ "Apache-2.0" ]
5
2021-11-11T23:37:53.000Z
2021-12-19T03:09:43.000Z
docs/tutorials/03-kglab_pytorch_geometric.ipynb
issam9/rubrix
16cafb10ca60bb75dd716a33eb8e5149bdbedb7b
[ "Apache-2.0" ]
null
null
null
docs/tutorials/03-kglab_pytorch_geometric.ipynb
issam9/rubrix
16cafb10ca60bb75dd716a33eb8e5149bdbedb7b
[ "Apache-2.0" ]
1
2022-02-11T08:28:02.000Z
2022-02-11T08:28:02.000Z
28.57134
419
0.561223
[ [ [ "# 🧪 Node classification with `kglab` and PyTorch Geometric\n\nWe introduce the application of neural networks on knowledge graphs using `kglab` and `pytorch_geometric`. \n\nGraph Neural networks (GNNs) have gained popularity in a number of practical applications, including knowledge graphs, social networks and recommender systems. In the context of knowledge graphs, GNNs are being used for tasks such as link prediction, node classification or knowledge graph embeddings. Many use cases for these tasks are related to `Automatic Knowledge Base Construction` (AKBC) and completion.\n\nIn this tutorial, we will learn to:\n\n- use `kglab` to represent a knowledge graph as a Pytorch Tensor, a suitable structure for working with neural nets\n\n- use the widely known `pytorch_geometric` (PyG) GNN library together with `kglab`.\n\n- train a GNN with `pytorch_geometric` and `PyTorch Lightning` for semi-supervised node classification of the recipes knowledge graph.\n\n- build and iterate on training data using `rubrix` with a Human-in-the-loop (HITL) approach.\n\n## Our use case in a nutshell\n\nOur goal in this notebook will be to build a semi-supervised node classifier of recipes and ingredients from scratch using kglab, PyG and Rubrix. \n\nOur classifier will be able to classify the nodes in our 15K nodes knowledge graph according to a set of pre-defined flavour related categories: `sweet`, `salty`, `piquant`, `sour`, etc. To account for mixed flavours (e.g., sweet chili sauce), our model will be multi-class (we have several target labels), multi-label (a node can be labelled as with 0 or several categories).", "_____no_output_____" ], [ "## Install `kglab` and `Pytorch Geometric`", "_____no_output_____" ] ], [ [ "%pip install torch-scatter -f https://pytorch-geometric.com/whl/torch-1.8.0+cpu.html -qqq\n%pip install torch-sparse -f https://pytorch-geometric.com/whl/torch-1.8.0+cpu.html -qqq\n%pip install torch-cluster -f https://pytorch-geometric.com/whl/torch-1.8.0+cpu.html -qqq\n%pip install torch-spline-conv -f https://pytorch-geometric.com/whl/torch-1.8.0+cpu.html -qqq\n%pip install torch-geometric -qqq\n%pip install torch==1.8.0 -qqq\n\n%pip install kglab -qqq\n\n%pip install pytorch_lightning -qqq", "_____no_output_____" ] ], [ [ "## 1. Loading and exploring the recipes knowledge graph\n\nWe'll be working with the \"recipes\" knowledge graph, which is used throughout the `kglab` tutorial (see the [Syllabus](https://derwen.ai/docs/kgl/tutorial/)).\n\nThis version of the recipes kg contains around ~15K recipes linked to their respective ingredients, as well as some other properties such as cooking time, labels and descriptions. \n\nLet's load the knowledge graph into a `kg` object by reading from an RDF file (in Turtle):", "_____no_output_____" ] ], [ [ "import kglab\n\nNAMESPACES = {\n \"wtm\": \"http://purl.org/heals/food/\",\n \"ind\": \"http://purl.org/heals/ingredient/\",\n \"recipe\": \"https://www.food.com/recipe/\",\n }\n\nkg = kglab.KnowledgeGraph(namespaces = NAMESPACES)\n\n_ = kg.load_rdf(\"data/recipe_lg.ttl\")", "_____no_output_____" ] ], [ [ "Let's take a look at our graph structure using the `Measure` class:", "_____no_output_____" ] ], [ [ "measure = kglab.Measure()\nmeasure.measure_graph(kg)\n\nf\"Nodes: {measure.get_node_count()} ; Edges: {measure.get_edge_count()}\"", "_____no_output_____" ], [ "measure.p_gen.get_tally() # tallies the counts of predicates", "_____no_output_____" ], [ "measure.s_gen.get_tally() # tallies the counts of predicates", "_____no_output_____" ], [ "measure.o_gen.get_tally() # tallies the counts of predicates", "_____no_output_____" ], [ "measure.l_gen.get_tally() # tallies the counts of literals", "_____no_output_____" ] ], [ [ "From the above exploration, we can extract some conclusions to guide the next steps:\n\n- We have a limited number of relationships, being `hasIngredient` the most frequent.\n\n- We have rather unique literals for labels and descriptions, but a certain amount of repetition for `hasCookTime`.\n\n- As we would have expected, most frequently referenced objects are ingredients such as `Salt`, `ChikenEgg` and so on. \n\n\nNow, let's move into preparing our knowledge graph for PyTorch.", "_____no_output_____" ], [ "## 2. Representing our knowledge graph as a `PyTorch` Tensor\n\nLet's now represent our `kg` as a `PyTorch` tensor using the `kglab.SubgraphTensor` class.", "_____no_output_____" ] ], [ [ "sg = kglab.SubgraphTensor(kg)", "_____no_output_____" ], [ "def to_edge_list(g, sg, excludes):\n def exclude(rel):\n return sg.n3fy(rel) in excludes\n \n relations = sorted(set(g.predicates()))\n subjects = set(g.subjects())\n objects = set(g.objects())\n nodes = list(subjects.union(objects))\n \n relations_dict = {rel: i for i, rel in enumerate(list(relations)) if not exclude(rel)}\n \n # this offset enables consecutive indices in our final vector\n offset = len(relations_dict.keys())\n \n nodes_dict = {node: i+offset for i, node in enumerate(nodes)}\n\n \n edge_list = []\n \n for s, p, o in g.triples((None, None, None)):\n if p in relations_dict.keys(): # this means is not excluded\n src, dst, rel = nodes_dict[s], nodes_dict[o], relations_dict[p]\n edge_list.append([src, dst, 2 * rel])\n edge_list.append([dst, src, 2 * rel + 1])\n \n # turn into str keys and concat\n node_vector = [sg.n3fy(node) for node in relations_dict.keys()] + [sg.n3fy(node) for node in nodes_dict.keys()]\n return edge_list, node_vector", "_____no_output_____" ], [ "edge_list, node_vector = to_edge_list(kg.rdf_graph(), sg, excludes=['skos:description', 'skos:prefLabel'])", "_____no_output_____" ], [ "len(edge_list) , edge_list[0:5]", "_____no_output_____" ] ], [ [ "Let's create `kglab.Subgraph` to be used for encoding/decoding numerical ids and uris, which will be useful for preparing our training data, as well as making sense of the predictions of our neural net.", "_____no_output_____" ] ], [ [ "sg = kglab.Subgraph(kg=kg, preload=node_vector)", "_____no_output_____" ], [ "import torch\nfrom torch_geometric.data import Data\n\ntensor = torch.tensor(edge_list, dtype=torch.long).t().contiguous() \nedge_index, edge_type = tensor[:2], tensor[2]\ndata = Data(edge_index=edge_index)\ndata.edge_type = edge_type", "_____no_output_____" ], [ "(data.edge_index.shape, data.edge_type.shape, data.edge_type.max())", "_____no_output_____" ] ], [ [ "## 3. Building a training set with Rubrix\n\nNow that we have a tensor representation of our kg which we can feed into our neural network, let's now focus on the training data.\n\nAs we will be doing semi-supervised classification, we need to build a training set (i.e., some recipes and ingredients with ground-truth labels). \n\n\nFor this, we can use [Rubrix](https://github.com/recognai/rubrix), an open-source tool for exploring, labeling and iterating on data for AI. Rubrix allows data scientists and subject matter experts to rapidly iterate on training and evaluation data by enabling iterative, asynchronous and potentially distributed workflows.\n\nIn Rubrix, a very simple workflow during model development looks like this:\n\n1. Log unlabelled data records with `rb.log()` into a Rubrix dataset. At this step you could use weak supervision methods (e.g., Snorkel) to pre-populate and then only refine the suggested labels, or use a pretrained model to guide your annotation process. In our case, we will just log recipe and ingredient \"records\" along with some metadata (RDF types, labels, etc.).\n\n2. Rapidly explore and label records in your dataset using the webapp which follows a search-driven approach, which is especially useful with large, potentially noisy datasets and for quickly leveraging domain knowledge (e.g., recipes containing WhiteSugar are likely sweet). For the tutorial, we have spent around 30min for labelling around 600 records.\n\n3. Retrieve your annotations any time using `rb.load()`, which return a convenient `pd.Dataframe` making it quite handy to process and use for model development. In our case, we will load a dataset, filter annotated entities, do a train_test_split with scikit_learn, and then use this for training our GNN.\n\n4. After training a model, you can go back to step 1, this time using your model and its predictions, to spot improvements, quickly label other portions of the data, and so on. In our case, as we've started with a very limited training set (~600 examples), we will use our node classifier and `rb.log()` it's predictions over the rest of our data (unlabelled recipes and ingredients).", "_____no_output_____" ] ], [ [ "LABELS = ['Bitter', 'Meaty', 'Piquant', 'Salty', 'Sour', 'Sweet']", "_____no_output_____" ] ], [ [ "### Setup Rubrix\n\nIf you have not installed and launched Rubrix, check the [installation guide](https://github.com/recognai/rubrix#get-started). ", "_____no_output_____" ] ], [ [ "import rubrix as rb", "_____no_output_____" ] ], [ [ "### Preparing our raw dataset of recipes and ingredients", "_____no_output_____" ] ], [ [ "import pandas as pd\nsparql = \"\"\"\n SELECT distinct *\n WHERE {\n ?uri a wtm:Recipe .\n ?uri a ?type .\n ?uri skos:definition ?definition .\n ?uri wtm:hasIngredient ?ingredient\n } \n\"\"\"\ndf = kg.query_as_df(sparql=sparql)\n\n# We group the ingredients into one column containing lists:\nrecipes_df = df.groupby(['uri', 'definition', 'type'])['ingredient'].apply(list).reset_index(name='ingredients') ; recipes_df\n\nsparql_ingredients = \"\"\"\n SELECT distinct *\n WHERE {\n ?uri a wtm:Ingredient .\n ?uri a ?type .\n OPTIONAL { ?uri skos:prefLabel ?definition } \n }\n\"\"\"\n\ndf = kg.query_as_df(sparql=sparql_ingredients)\ndf['ingredients'] = None\n\ning_recipes_df = pd.concat([recipes_df, df]).reset_index(drop=True)\n\ning_recipes_df.fillna('', inplace=True) ; ing_recipes_df", "_____no_output_____" ] ], [ [ "### Logging into Rubrix", "_____no_output_____" ] ], [ [ "import rubrix as rb\n\nrecords = []\nfor i, r in ing_recipes_df.iterrows():\n item = rb.TextClassificationRecord(\n inputs={\n \"id\":r.uri, \n \"definition\": r.definition,\n \"ingredients\": str(r.ingredients), \n \"type\": r.type\n }, # log node fields\n prediction=[(label, 0.0) for label in LABELS], # log \"dummy\" predictions for aiding annotation\n metadata={'ingredients': [ing.replace('ind:','') for ing in r.ingredients], \"type\": r.type}, # metadata filters for quick exploration and annotation\n prediction_agent=\"kglab_tutorial\", # who's performing/logging the prediction\n multi_label=True\n )\n records.append(item)", "_____no_output_____" ], [ "len(records)", "_____no_output_____" ], [ "rb.log(records=records, name=\"kg_classification_tutorial\")", "_____no_output_____" ] ], [ [ "### Annotation session with Rubrix (optional)\n\nIn this step you can go to your rubrix dataset and annotate some examples of each class.\n\nIf you have no time to do this, just skip this part as we have prepared a dataset for you with around ~600 examples.", "_____no_output_____" ], [ "### Loading our labelled records and create a train_test split (optional)\n\nIf you have no time to do this, just skip this part as we have prepared a dataset for you.", "_____no_output_____" ] ], [ [ "rb.snapshots(name=\"kg_classification_tutorial\")", "_____no_output_____" ] ], [ [ "Once you have annotated your dataset, you will find an snapshot id on the previous list. This id should be place in the next command. In our case, it was 1620136587.907149.", "_____no_output_____" ] ], [ [ "df = rb.load(name=\"kg_classification_tutorial\", snapshot='1620136587.907149') ; df.head()", "_____no_output_____" ], [ "from sklearn.model_selection import train_test_split\n\ntrain_df, test_df = train_test_split(df)\ntrain_df.to_csv('data/train_recipes_new.csv')\ntest_df.to_csv('data/test_recipes_new.csv')", "_____no_output_____" ] ], [ [ "### Creating PyTorch train and test sets\n\nHere we take our train and test datasets and transform them into `torch.Tensor` objects with the help of our kglab `Subgraph` for turning `uris` into `torch.long` indices.", "_____no_output_____" ] ], [ [ "import pandas as pd\n\ntrain_df = pd.read_csv('data/train_recipes.csv') # use your own labelled datasets if you've created a snapshot\ntest_df = pd.read_csv('data/test_recipes.csv')\n\n# we make sure lists are parsed correctly\ntrain_df.labels = train_df.labels.apply(eval)\ntest_df.labels = test_df.labels.apply(eval)", "_____no_output_____" ], [ "train_df", "_____no_output_____" ] ], [ [ "Let's create label lookups for label to int and viceversa", "_____no_output_____" ] ], [ [ "label2id = {label:i for i,label in enumerate(LABELS)} ; \nid2label = {i:l for l,i in label2id.items()} ; (id2label, label2id)", "_____no_output_____" ] ], [ [ "The following function turns our DataFrame into numerical arrays for node indices and labels", "_____no_output_____" ] ], [ [ "import numpy as np\n\ndef create_indices_labels(df):\n # turn our dense labels into a one-hot list\n def one_hot(label_ids):\n a = np.zeros(len(LABELS))\n a.put(label_ids, np.ones(len(label_ids)))\n return a\n \n indices, labels = [], []\n for uri, label in zip(df.uri.tolist(), df.labels.tolist()):\n indices.append(sg.transform(uri))\n labels.append(one_hot([label2id[label] for label in label]))\n return indices, labels", "_____no_output_____" ] ], [ [ "Finally, let's turn our dataset into PyTorch tensors", "_____no_output_____" ] ], [ [ "train_indices, train_labels = create_indices_labels(train_df)\ntest_indices, test_labels = create_indices_labels(test_df)\n\ntrain_idx = torch.tensor(train_indices, dtype=torch.long)\ntrain_y = torch.tensor(train_labels, dtype=torch.float)\n\ntest_idx = torch.tensor(test_indices, dtype=torch.long)\ntest_y = torch.tensor(test_labels, dtype=torch.float) ; train_idx[:10], train_y", "_____no_output_____" ] ], [ [ "Let's see if we can recover the correct URIs for our numerical ids using our `kglab.Subgraph`", "_____no_output_____" ] ], [ [ "(train_df.loc[0], sg.inverse_transform(15380))", "_____no_output_____" ] ], [ [ "## 4. Creating a Subgraph of recipe and ingredient nodes\nHere we create a node list to be used as a seed for building our `PyG` subgraph (using k-hops as we will see in the next section). Our goal will be to start only with `recipes` and `ingredients`, as all nodes passed through the GNN will be classified and those are our main target. ", "_____no_output_____" ] ], [ [ "node_idx = torch.LongTensor([\n sg.transform(i) for i in ing_recipes_df.uri.values\n])", "_____no_output_____" ], [ "node_idx.max(), node_idx.shape", "_____no_output_____" ], [ "ing_recipes_df.iloc[1]", "_____no_output_____" ], [ "sg.inverse_transform(node_idx[1])", "_____no_output_____" ], [ "node_idx[0:10]", "_____no_output_____" ] ], [ [ "## 5. Semi-supervised node classification with PyTorch Geometric\n\nFor the node classification task **we are given the ground-truth labels** (our recipes and ingredients training set) **for a small subset of nodes**, and **we want to predict the labels for all the remaining nodes** (our recipes and ingredients test set and unlabelled nodes).\n\n\n### Graph Convolutional Networks\n\nTo get a great intro to GCNs we recommend you to check Kipf's [blog post](https://tkipf.github.io/graph-convolutional-networks/) on the topic. \n\nIn a nutshell, GCNs are multi-layer neural works which apply \"convolutions\" to nodes in graphs by sharing and applying the same filter parameters over all locations in the graph. \n\nAdditionally, modern GCNs such as those implemented in `PyG` use **message passing** mechanisms, where vertices exchange information with their neighbors, and send messages to each other.\n\n![GCN_Kipf](img/gcn_web.png \"GCN\")\n\n***Multi-layer Graph Convolutional Network (GCN) with first-order filters***. Source: https://tkipf.github.io/graph-convolutional-networks\n\n### Relational Graph Convolutional Networks\n\nRelational Graph Convolutional Networks (R-GCNs) were introduced by [Schlichtkrull et al. 2017](https://arxiv.org/abs/1703.06103), as an extension of GCNs to deal with **multi-relational knowledge graphs**. \n\nYou can see below the computation model for nodes:", "_____no_output_____" ], [ "![RGCN](img/rgcn.png \"RGCN\")\n\n***Computation of the update of a single graph node(red) in the R-GCN model.***. Source: https://arxiv.org/abs/1703.06103", "_____no_output_____" ], [ "### Creating a `PyG` subgraph\n\nHere we build a subgraph with `k` hops from target to source starting with all `recipe` and `ingredient` nodes:", "_____no_output_____" ] ], [ [ "from torch_geometric.utils import k_hop_subgraph\n# here we take all connected nodes with k hops\nk = 1\nnode_idx, edge_index, mapping, edge_mask = k_hop_subgraph(\n node_idx, \n k, \n data.edge_index, \n relabel_nodes=False\n)", "_____no_output_____" ] ], [ [ "We have increased the size of our node set:", "_____no_output_____" ] ], [ [ "node_idx.shape", "_____no_output_____" ], [ "data.edge_index.shape", "_____no_output_____" ] ], [ [ "Here we compute some measures needed for defining the size of our layers", "_____no_output_____" ] ], [ [ "data.edge_index = edge_index\n\ndata.num_nodes = data.edge_index.max().item() + 1\n\ndata.num_relations = data.edge_type.max().item() + 1\n\ndata.edge_type = data.edge_type[edge_mask]\n\ndata.num_classes = len(LABELS)\n\ndata.num_nodes, data.num_relations, data.num_classes", "_____no_output_____" ] ], [ [ "### Defining a basic Relational Graph Convolutional Network", "_____no_output_____" ] ], [ [ "from torch_geometric.nn import FastRGCNConv, RGCNConv\nimport torch.nn.functional as F", "_____no_output_____" ], [ "RGCNConv?", "_____no_output_____" ], [ "class RGCN(torch.nn.Module):\n def __init__(self, num_nodes, num_relations, num_classes, out_channels=16, num_bases=30, dropout=0.0, layer_type=FastRGCNConv, ):\n \n super(RGCN, self).__init__()\n \n self.conv1 = layer_type(\n num_nodes, \n out_channels, \n num_relations, \n num_bases=num_bases\n )\n self.conv2 = layer_type(\n out_channels, \n num_classes, \n num_relations, \n num_bases=num_bases\n )\n self.dropout = torch.nn.Dropout(dropout)\n\n def forward(self, edge_index, edge_type):\n x = F.relu(self.conv1(None, edge_index, edge_type))\n x = self.dropout(x)\n x = self.conv2(x, edge_index, edge_type)\n return torch.sigmoid(x)", "_____no_output_____" ] ], [ [ "### Create and visualizing our model", "_____no_output_____" ] ], [ [ "model = RGCN(\n num_nodes=data.num_nodes,\n num_relations=data.num_relations,\n num_classes=data.num_classes,\n #out_channels=64,\n dropout=0.2,\n layer_type=RGCNConv\n) ; model", "_____no_output_____" ], [ "# code adapted from https://colab.research.google.com/drive/14OvFnAXggxB8vM4e8vSURUp1TaKnovzX\n%matplotlib inline\nimport matplotlib.pyplot as plt\nfrom sklearn.manifold import TSNE\nfrom pytorch_lightning.metrics.utils import to_categorical\n\ndef visualize(h, color, labels):\n z = TSNE(n_components=2).fit_transform(h.detach().cpu().numpy())\n\n plt.figure(figsize=(10,10))\n plt.xticks([])\n plt.yticks([])\n \n scatter = plt.scatter(z[:, 0], z[:, 1], s=70, c=color, cmap=\"Set2\")\n legend = plt.legend(scatter.legend_elements()[0],labels, loc=\"upper right\", title=\"Labels\",) #*scatter.legend_elements()\n plt.show()", "_____no_output_____" ], [ "pred = model(edge_index, edge_type)", "_____no_output_____" ], [ "visualize(pred[train_idx], color=to_categorical(train_y), labels=LABELS)", "_____no_output_____" ], [ "visualize(pred[test_idx], color=to_categorical(test_y), labels=LABELS)", "_____no_output_____" ] ], [ [ "### Training our RGCN", "_____no_output_____" ] ], [ [ "device = torch.device('cpu') # ('cuda')\ndata = data.to(device)\nmodel = model.to(device)\noptimizer = torch.optim.AdamW(model.parameters())\nloss_module = torch.nn.BCELoss()\n\ndef train():\n model.train()\n optimizer.zero_grad()\n out = model(data.edge_index, data.edge_type)\n loss = loss_module(out[train_idx], train_y)\n loss.backward()\n optimizer.step()\n return loss.item()\n\ndef accuracy(predictions, y):\n predictions = np.round(predictions)\n return predictions.eq(y).to(torch.float).mean()\n\[email protected]_grad()\ndef test():\n model.eval()\n pred = model(data.edge_index, data.edge_type)\n train_acc = accuracy(pred[train_idx], train_y)\n test_acc = accuracy(pred[test_idx], test_y)\n return train_acc.item(), test_acc.item()", "_____no_output_____" ], [ "for epoch in range(1, 50):\n loss = train()\n train_acc, test_acc = test()\n print(f'Epoch: {epoch:02d}, Loss: {loss:.4f}, Train: {train_acc:.4f} '\n f'Test: {test_acc:.4f}')", "_____no_output_____" ] ], [ [ "### Model visualization", "_____no_output_____" ] ], [ [ "pred = model(edge_index, edge_type)", "_____no_output_____" ], [ "visualize(pred[train_idx], color=to_categorical(train_y), labels=LABELS)", "_____no_output_____" ], [ "visualize(pred[test_idx], color=to_categorical(test_y), labels=LABELS)", "_____no_output_____" ] ], [ [ "## 6. Using our model and analyzing its predictions with Rubrix\nLet's see the shape of our model predictions", "_____no_output_____" ] ], [ [ "pred = model(edge_index, edge_type) ; pred", "_____no_output_____" ], [ "def find(tensor, values):\n return torch.nonzero(tensor[..., None] == values)", "_____no_output_____" ] ], [ [ "### Analizing predictions over the test set", "_____no_output_____" ] ], [ [ "test_idx = find(node_idx,test_idx)[:,0] ; len(test_idx)", "_____no_output_____" ], [ "index = torch.zeros(node_idx.shape[0], dtype=bool)\nindex[test_idx] = True\nidx = node_idx[index]", "_____no_output_____" ], [ "uris = [sg.inverse_transform(i) for i in idx]\npredicted_labels = [l for l in pred[idx]]", "_____no_output_____" ], [ "predictions = list(zip(uris,predicted_labels)) ; predictions[0:2]", "_____no_output_____" ], [ "import rubrix as rb\n\nrecords = []\nfor uri,predicted_labels in predictions:\n ids = ing_recipes_df.index[ing_recipes_df.uri == uri]\n if len(ids) > 0:\n r = ing_recipes_df.iloc[ids]\n # get the gold labels from our test set\n gold_labels = test_df.iloc[test_df.index[test_df.uri == uri]].labels.values[0]\n \n item = rb.TextClassificationRecord(\n inputs={\"id\":r.uri.values[0], \"definition\": r.definition.values[0], \"ingredients\": str(r.ingredients.values[0]), \"type\": r.type.values[0]}, \n prediction=[(id2label[i], score) for i,score in enumerate(predicted_labels)],\n annotation=gold_labels,\n metadata={'ingredients': r.ingredients.values[0], \"type\": r.type.values[0]}, \n prediction_agent=\"node_classifier_v1\", \n multi_label=True\n )\n records.append(item)", "_____no_output_____" ], [ "rb.log(records, name=\"kg_classification_test_analysis\")", "_____no_output_____" ] ], [ [ "### Analizing predictions over unseen nodes (and potentially relabeling them)", "_____no_output_____" ], [ "Let's find the ids for the nodes in our training and test sets", "_____no_output_____" ] ], [ [ "train_test_idx = find(node_idx,torch.cat((test_idx, train_idx)))[:,0] ; len(train_test_idx)", "_____no_output_____" ] ], [ [ "Let's get the ids, uris and labels of the nodes which were not in our train/test datasets", "_____no_output_____" ] ], [ [ "index = torch.ones(node_idx.shape[0], dtype=bool)\nindex[train_test_idx] = False\nidx = node_idx[index]", "_____no_output_____" ] ], [ [ "We use our `SubgraphTensor` for getting back our URIs and build `uri,predicted_labels` pairs:", "_____no_output_____" ] ], [ [ "uris = [sg.inverse_transform(i) for i in idx]\npredicted_labels = [l for l in pred[idx]]", "_____no_output_____" ], [ "predictions = list(zip(uris,predicted_labels)) ; predictions[0:2]", "_____no_output_____" ], [ "import rubrix as rb\n\nrecords = []\nfor uri,predicted_labels in predictions:\n ids = ing_recipes_df.index[ing_recipes_df.uri == uri]\n if len(ids) > 0:\n r = ing_recipes_df.iloc[ids]\n item = rb.TextClassificationRecord(\n inputs={\"id\":r.uri.values[0], \"definition\": r.definition.values[0], \"ingredients\": str(r.ingredients.values[0]), \"type\": r.type.values[0]}, \n prediction=[(id2label[i], score) for i,score in enumerate(predicted_labels)], \n metadata={'ingredients': r.ingredients.values[0], \"type\": r.type.values[0]}, \n prediction_agent=\"node_classifier_v1\", \n multi_label=True\n )\n records.append(item)", "_____no_output_____" ], [ "rb.log(records, name=\"kg_node_classification_unseen_nodes_v3\")", "_____no_output_____" ] ], [ [ "## Exercise 1: Training experiments with PyTorch Lightning", "_____no_output_____" ] ], [ [ "#!pip install wandb -qqq # optional", "_____no_output_____" ], [ "!wandb login #optional", "_____no_output_____" ], [ "from torch_geometric.data import Data, DataLoader\n\ndata.train_idx = train_idx\ndata.train_y = train_y\ndata.test_idx = test_idx\ndata.test_y = test_y\n\ndataloader = DataLoader([data], batch_size=1); dataloader", "_____no_output_____" ], [ "import torch\nimport pytorch_lightning as pl\nfrom pytorch_lightning.callbacks import EarlyStopping, ModelCheckpoint\nfrom pytorch_lightning.loggers import WandbLogger\n\nclass RGCNNodeClassification(pl.LightningModule):\n \n def __init__(self, **model_kwargs):\n super().__init__()\n \n self.model = RGCN(**model_kwargs)\n self.loss_module = torch.nn.BCELoss()\n \n def forward(self, edge_index, edge_type):\n return self.model(edge_index, edge_type)\n \n def configure_optimizers(self):\n optimizer = torch.optim.Adam(self.parameters(), lr=0.01, weight_decay=0.001)\n return optimizer\n \n def training_step(self, batch, batch_idx):\n idx, y = data.train_idx, data.train_y\n edge_index, edge_type = data.edge_index, data.edge_type\n x = self.forward(edge_index, edge_type)\n loss = self.loss_module(x[idx], y)\n x = x.detach()\n self.log('train_acc', accuracy(x[idx], y), prog_bar=True)\n self.log('train_loss', loss)\n return loss \n \n def validation_step(self, batch, batch_idx):\n idx, y = data.test_idx, data.test_y\n edge_index, edge_type = data.edge_index, data.edge_type\n x = self.forward(edge_index, edge_type)\n loss = self.loss_module(x[idx], y)\n x = x.detach()\n self.log('val_acc', accuracy(x[idx], y), prog_bar=True)\n self.log('val_loss', loss)", "_____no_output_____" ], [ "pl.seed_everything()", "_____no_output_____" ], [ "model_pl = RGCNNodeClassification(\n num_nodes=data.num_nodes,\n num_relations=data.num_relations,\n num_classes=data.num_classes,\n #out_channels=64,\n dropout=0.2,\n #layer_type=RGCNConv\n)", "_____no_output_____" ], [ "early_stopping = EarlyStopping(monitor='val_acc', patience=10, mode='max')", "_____no_output_____" ], [ "trainer = pl.Trainer(\n default_root_dir='pl_runs',\n checkpoint_callback=ModelCheckpoint(save_weights_only=True, mode=\"max\", monitor=\"val_acc\"),\n max_epochs=200,\n #logger= WandbLogger(), # optional\n callbacks=[early_stopping]\n)", "_____no_output_____" ], [ "trainer.fit(model_pl, dataloader, dataloader)", "_____no_output_____" ] ], [ [ "## Exercise 2: Bootstrapping annotation with a zeroshot-classifier", "_____no_output_____" ] ], [ [ "!pip install transformers -qqq", "_____no_output_____" ], [ "from transformers import pipeline\n \npretrained_model = \"valhalla/distilbart-mnli-12-1\" # \"typeform/squeezebert-mnli\"\n\npl = pipeline('zero-shot-classification', model=pretrained_model)", "_____no_output_____" ], [ "pl(\"chocolate cake\", LABELS, hypothesis_template='The flavour is {}.',multi_label=True)", "_____no_output_____" ], [ "import rubrix as rb\n\nrecords = []\nfor i, r in ing_recipes_df[50:150].iterrows():\n preds = pl(r.definition, LABELS, hypothesis_template='The flavour is {}.', multi_label=True)\n item = rb.TextClassificationRecord(\n inputs={\n \"id\":r.uri, \n \"definition\": r.definition,\n \"ingredients\": str(r.ingredients), \n \"type\": r.type\n }, \n prediction=list(zip(preds['labels'], preds['scores'])), # TODO: here we log he predictions of our zeroshot pipeline as a list of tuples (label, score)\n metadata={'ingredients': r.ingredients, \"type\": r.type}, \n prediction_agent=\"valhalla/distilbart-mnli-12-1\", \n multi_label=True\n )\n records.append(item)", "_____no_output_____" ], [ "rb.log(records, name='kg_zeroshot')", "_____no_output_____" ] ], [ [ "## Next steps\n\n### 📚 [Rubrix documentation](https://docs.rubrix.ml) for more guides and tutorials.\n\n### 🙋‍♀️ Join the Rubrix community! A good place to start is the [discussion forum](https://github.com/recognai/rubrix/discussions).\n\n### ⭐ Rubrix [Github repo](https://github.com/recognai/rubrix) to stay updated.", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown" ]
[ [ "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code" ], [ "markdown", "markdown" ], [ "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code" ], [ "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code", "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code", "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code" ], [ "markdown" ] ]
ece8235fbcdd2217a23644426a819bd5071b68ed
283,442
ipynb
Jupyter Notebook
EHR_Only/Lasso/Comp Cardio.ipynb
shreyaskar123/EHR-Discontinuity
8d2becfd784b9cbe697f8308d60023701971ef5d
[ "MIT" ]
null
null
null
EHR_Only/Lasso/Comp Cardio.ipynb
shreyaskar123/EHR-Discontinuity
8d2becfd784b9cbe697f8308d60023701971ef5d
[ "MIT" ]
null
null
null
EHR_Only/Lasso/Comp Cardio.ipynb
shreyaskar123/EHR-Discontinuity
8d2becfd784b9cbe697f8308d60023701971ef5d
[ "MIT" ]
null
null
null
15.671901
1,735
0.29334
[ [ [ "import pandas as pd\nmedicare = pd.read_csv(\"/netapp2/home/se197/RPDR/Josh Lin/3_EHR_V2/CMS/Data/final_medicare.csv\")", "_____no_output_____" ], [ "def val_appended(card_outcomes): #MI, stroke, death\n for outcomes in card_outcomes:\n if (outcomes == 1):\n return 1\n return 0\nout_comp_cardiovascular_R1 = [] \nfor index in range(medicare.shape[0]):\n rel_outcomes = [medicare.at[index,'Out_IscheStroke_R1'], medicare.at[index,'Out_HemoStroke_R1'], medicare.at[index, 'Out_UncertainStroke_R1'], medicare.at[index, 'Out_MI_R1'], medicare.at[index,'ehr_death']]\n out_comp_cardiovascular_R1.append(val_appended(rel_outcomes))\n\nmedicare.insert(loc = 1, column = \"Out_comp_cardiovascular_R1\", value = out_comp_cardiovascular_R1)", "_____no_output_____" ], [ "from datetime import date\nimport math\nehr_claims_death = []\nfor rows in medicare[['D_Index', 'D_Death_C']].itertuples():\n d0 = date(int(rows[1][0:4]), int(rows[1][5:7]), int(rows[1][8:10]))\n try: \n math.isnan(float(rows[2]))\n ehr_claims_death.append(0)\n continue\n except ValueError:\n d1 = date(int(rows[2][0:4]), int(rows[2][5:7]), int(rows[2][8:10]))\n delta = d1 - d0\n print(delta.days)\n if (delta.days <= 365):\n ehr_claims_death.append(1)\n continue\n else:\n ehr_claims_death.append(0)\nmedicare['claims_death'] = ehr_claims_death \n ", "1111\n147\n1835\n98\n1693\n1530\n11\n1531\n116\n43\n1966\n1931\n1772\n1744\n378\n460\n509\n1827\n595\n125\n27\n218\n369\n813\n598\n1608\n807\n800\n161\n40\n770\n1061\n340\n2070\n77\n177\n1072\n1045\n833\n142\n35\n616\n1092\n1220\n661\n658\n271\n1270\n588\n461\n116\n1069\n1276\n398\n718\n2279\n325\n1154\n144\n211\n841\n1145\n331\n28\n4\n51\n528\n72\n208\n314\n1008\n1704\n1582\n38\n1794\n1189\n1963\n1900\n1682\n825\n1248\n269\n271\n49\n960\n253\n18\n1397\n1736\n812\n233\n1465\n356\n168\n544\n55\n1904\n2091\n384\n2092\n189\n1357\n1721\n348\n194\n1301\n1413\n749\n459\n1591\n170\n180\n588\n2899\n746\n1743\n1411\n592\n390\n807\n737\n3398\n195\n2010\n339\n891\n468\n2149\n658\n2071\n51\n2626\n3004\n2987\n757\n212\n1576\n725\n1320\n3555\n1429\n1759\n1860\n1061\n1449\n751\n318\n1119\n3054\n2408\n2733\n64\n151\n532\n1205\n1533\n480\n127\n79\n2179\n1081\n792\n2699\n975\n3045\n1265\n2992\n1316\n1768\n1616\n1929\n2190\n2473\n174\n2848\n1118\n778\n148\n1112\n129\n2969\n2618\n576\n2494\n385\n351\n810\n1039\n1940\n587\n1565\n2061\n3348\n497\n1314\n1374\n1629\n2723\n2065\n308\n935\n1254\n252\n417\n1479\n460\n654\n2853\n2664\n11\n1557\n7\n560\n1128\n3274\n275\n1902\n483\n2229\n1939\n1859\n2909\n2641\n2632\n2091\n475\n1646\n56\n689\n663\n1244\n2932\n576\n1458\n137\n724\n2274\n2621\n5\n1283\n1530\n878\n1719\n421\n1604\n362\n6\n69\n547\n2080\n538\n1499\n71\n876\n2606\n906\n436\n922\n40\n1402\n1756\n2978\n2969\n61\n445\n146\n541\n3565\n372\n3476\n372\n654\n1023\n186\n533\n1296\n834\n1390\n1579\n1\n1786\n2032\n373\n3232\n232\n343\n190\n2731\n1\n235\n36\n267\n2087\n1556\n427\n183\n59\n2142\n1399\n1777\n671\n1400\n43\n2487\n2465\n444\n2463\n1409\n2037\n1398\n1466\n3138\n63\n243\n1054\n717\n3190\n1217\n334\n142\n1574\n227\n24\n1417\n2699\n665\n2760\n3595\n451\n278\n406\n977\n2739\n304\n86\n503\n1493\n2505\n1453\n338\n94\n1262\n1855\n584\n215\n753\n2578\n394\n2537\n266\n1017\n2495\n352\n3406\n76\n739\n175\n1355\n1796\n2106\n1293\n1173\n162\n2207\n600\n43\n1758\n568\n660\n2310\n268\n613\n2447\n2882\n893\n576\n912\n447\n2444\n1027\n1694\n908\n1534\n267\n618\n2108\n317\n1966\n2190\n1328\n3194\n987\n2956\n491\n18\n3262\n1271\n1694\n2524\n628\n506\n229\n974\n374\n351\n1560\n2072\n1148\n3430\n3209\n728\n329\n1510\n774\n76\n1759\n1181\n1097\n42\n56\n85\n1269\n706\n1005\n529\n2156\n1412\n835\n1140\n323\n2447\n544\n1037\n848\n1688\n68\n1032\n2666\n1321\n105\n1919\n175\n794\n1431\n1279\n2568\n2346\n966\n267\n1232\n1047\n2972\n37\n1538\n279\n2215\n3546\n416\n627\n3184\n3558\n2196\n689\n1202\n1031\n221\n1577\n2227\n1328\n1392\n1799\n334\n284\n2365\n2441\n2777\n1015\n82\n1826\n615\n1510\n2618\n2073\n1297\n32\n752\n598\n176\n1641\n1274\n1558\n2961\n1693\n3065\n67\n3326\n361\n407\n503\n2120\n1913\n58\n3512\n1595\n2623\n620\n1547\n2588\n897\n2750\n733\n1492\n969\n3579\n1270\n335\n5\n1772\n556\n910\n1021\n2395\n978\n2909\n988\n789\n802\n2717\n3296\n2744\n872\n227\n1908\n2631\n1533\n97\n1556\n3300\n894\n909\n2106\n2505\n1006\n2058\n1001\n1658\n138\n275\n1512\n592\n2386\n96\n521\n1466\n449\n2262\n317\n1271\n1835\n3127\n603\n689\n1307\n1341\n2168\n141\n1574\n102\n274\n221\n577\n2121\n150\n141\n47\n1707\n109\n438\n3237\n1811\n38\n2086\n184\n3\n57\n115\n1407\n2954\n412\n2955\n1766\n2098\n82\n129\n2545\n2091\n2187\n2417\n252\n3185\n375\n2936\n2109\n3169\n1473\n1248\n896\n1729\n123\n2590\n2160\n1054\n2721\n1360\n867\n1422\n1054\n166\n1619\n3335\n54\n2255\n1415\n1391\n1547\n1400\n2015\n1890\n1773\n3547\n1088\n1932\n1485\n722\n1310\n631\n27\n362\n1242\n1327\n3225\n94\n2526\n1650\n253\n3004\n5\n2403\n2907\n3062\n928\n2230\n3215\n2360\n1809\n705\n1007\n121\n368\n807\n507\n246\n1147\n883\n806\n689\n1147\n495\n496\n2297\n1250\n2119\n1657\n1927\n1148\n985\n160\n777\n1630\n2053\n2531\n273\n696\n2236\n99\n507\n366\n3644\n2092\n2081\n3463\n2259\n1626\n2728\n788\n745\n4\n961\n2330\n2629\n299\n2437\n1296\n493\n2362\n2739\n1037\n1373\n435\n310\n2990\n1282\n890\n1436\n2938\n2979\n792\n953\n3247\n1214\n3348\n248\n2321\n1056\n28\n1025\n80\n43\n2038\n280\n770\n1207\n650\n83\n1423\n2761\n1116\n1985\n1187\n2828\n754\n1164\n95\n1175\n1871\n1012\n1440\n272\n761\n1741\n2339\n180\n128\n1175\n1518\n52\n1160\n770\n1188\n227\n424\n201\n1144\n3125\n317\n1982\n2510\n539\n308\n727\n389\n750\n572\n3641\n399\n404\n1067\n10\n1194\n795\n457\n536\n1829\n2903\n2621\n944\n1142\n2330\n2653\n88\n1499\n3458\n3518\n2035\n1731\n1258\n2781\n1639\n41\n3379\n6\n928\n2271\n484\n453\n2124\n1570\n1101\n56\n2117\n1781\n1755\n3477\n387\n3162\n1163\n861\n117\n2630\n75\n2406\n1663\n665\n2940\n3282\n794\n2669\n1575\n550\n1119\n1340\n1071\n2365\n2762\n2189\n1074\n1006\n1661\n1113\n857\n79\n2288\n2917\n2954\n511\n2632\n2497\n1054\n2470\n1196\n519\n72\n937\n435\n507\n1286\n39\n613\n669\n402\n776\n1203\n607\n1360\n315\n453\n1999\n1604\n1136\n2324\n268\n1261\n963\n554\n388\n548\n1144\n766\n240\n862\n775\n71\n1617\n79\n400\n189\n652\n1695\n1109\n850\n1035\n1254\n792\n3455\n2251\n1769\n79\n808\n2048\n1438\n1037\n2130\n1653\n489\n1468\n1474\n1670\n5\n2995\n79\n32\n970\n1495\n1701\n2055\n1340\n486\n3076\n3381\n625\n1897\n639\n1304\n1694\n1470\n226\n896\n2719\n2692\n660\n3\n328\n1736\n2585\n2988\n943\n14\n1744\n1284\n67\n1428\n265\n748\n551\n49\n1979\n444\n3577\n1286\n1273\n240\n2960\n1321\n1491\n1432\n2533\n1427\n1626\n3445\n847\n119\n2358\n844\n2094\n2928\n588\n2713\n1274\n2998\n2699\n2317\n2659\n1025\n2527\n1728\n194\n368\n104\n767\n999\n3516\n2022\n1861\n54\n519\n1502\n892\n2499\n2701\n289\n954\n15\n1773\n61\n1455\n2788\n1447\n430\n2110\n1460\n1461\n289\n2979\n672\n2177\n2612\n805\n2285\n328\n562\n1625\n880\n2806\n2213\n1771\n2466\n2353\n86\n1399\n1810\n6\n571\n26\n800\n815\n428\n1745\n29\n1781\n2129\n2698\n2181\n744\n126\n1973\n887\n511\n1767\n3234\n530\n1744\n431\n2762\n1658\n711\n1731\n1900\n366\n954\n732\n2163\n1443\n84\n3454\n228\n1315\n1978\n1724\n458\n370\n3314\n2336\n1596\n4\n2185\n2816\n3217\n85\n3109\n830\n1632\n2242\n1904\n2524\n1560\n2581\n2347\n556\n10\n1219\n3408\n281\n1672\n2501\n2820\n1550\n1544\n2304\n417\n1834\n894\n491\n637\n2612\n372\n921\n903\n1584\n1462\n2699\n1446\n592\n2535\n788\n1645\n1695\n85\n2549\n510\n1065\n1323\n2082\n1077\n2047\n617\n859\n1416\n169\n3034\n3005\n353\n181\n156\n1748\n1248\n982\n753\n53\n1553\n54\n2024\n114\n1\n729\n131\n1880\n1821\n900\n58\n944\n2478\n1836\n123\n189\n102\n690\n1712\n2598\n100\n1111\n208\n3201\n1917\n30\n479\n2293\n1607\n3050\n447\n2393\n1360\n2978\n1212\n2749\n875\n2470\n1518\n1326\n199\n532\n2361\n2204\n2673\n457\n264\n1556\n773\n1408\n3006\n2208\n1385\n1026\n455\n510\n361\n1845\n842\n663\n411\n648\n489\n2\n1798\n3304\n660\n3258\n2859\n1115\n746\n641\n315\n1975\n1290\n2164\n870\n1848\n540\n1955\n1083\n580\n597\n1729\n2858\n428\n215\n84\n601\n539\n1586\n1246\n1060\n582\n1547\n401\n1654\n1032\n32\n13\n415\n1796\n429\n323\n706\n1887\n146\n2377\n17\n848\n142\n3506\n2051\n797\n63\n2796\n1169\n1845\n1121\n1801\n3175\n1024\n1164\n1722\n604\n3204\n2759\n744\n1397\n2533\n2851\n1968\n1053\n494\n2816\n434\n1561\n726\n679\n2738\n1099\n275\n388\n226\n581\n47\n502\n172\n3626\n1043\n1956\n1188\n2385\n356\n1807\n21\n5\n48\n173\n372\n849\n57\n2499\n2915\n2605\n48\n604\n1244\n1532\n695\n903\n1950\n155\n245\n230\n1097\n2666\n2166\n2158\n2334\n1125\n1156\n60\n999\n1150\n1192\n336\n332\n5\n602\n1879\n106\n75\n1100\n249\n1464\n1130\n755\n2425\n1451\n360\n3570\n717\n2005\n2399\n774\n279\n2209\n803\n3066\n2345\n1619\n3513\n403\n329\n962\n2831\n3536\n56\n1402\n436\n2577\n749\n47\n597\n504\n373\n2003\n6\n2382\n2897\n200\n52\n212\n588\n16\n1912\n523\n418\n117\n264\n535\n2418\n2500\n1257\n187\n93\n1995\n534\n1194\n970\n2420\n729\n751\n3295\n2398\n1911\n668\n4\n469\n2539\n2323\n3584\n282\n2658\n476\n423\n377\n35\n265\n1979\n1224\n368\n558\n1714\n1993\n62\n2541\n1\n2213\n3068\n920\n14\n33\n1324\n1988\n2863\n1420\n6\n492\n16\n1219\n420\n815\n494\n189\n3107\n35\n549\n238\n681\n1059\n1554\n1457\n731\n2242\n1982\n13\n115\n226\n2090\n2046\n317\n188\n682\n1182\n67\n1730\n122\n2197\n2944\n207\n2007\n467\n170\n778\n2284\n3607\n13\n1\n321\n302\n1013\n14\n1692\n3410\n467\n862\n1179\n348\n2\n238\n55\n1873\n3522\n44\n2484\n1024\n185\n240\n3529\n411\n409\n1991\n82\n2593\n1682\n24\n520\n1448\n334\n707\n1287\n2615\n61\n2658\n321\n2350\n2952\n1413\n921\n611\n984\n1287\n3204\n2530\n3561\n667\n1\n247\n3075\n2535\n1514\n1381\n3267\n368\n1122\n3314\n771\n2596\n1125\n155\n1932\n1190\n534\n1619\n2580\n1167\n1170\n3288\n1877\n1670\n381\n121\n3537\n3594\n1482\n757\n1691\n193\n95\n2718\n976\n151\n776\n1964\n2975\n1096\n508\n343\n804\n2524\n1328\n793\n3264\n2657\n2614\n2168\n1599\n4\n3379\n3430\n2905\n523\n172\n938\n49\n2401\n2250\n35\n52\n2514\n298\n3\n3253\n1476\n567\n154\n1052\n2000\n13\n155\n2477\n1928\n2314\n1839\n28\n44\n1622\n1603\n1664\n2602\n155\n851\n2709\n199\n2737\n1689\n602\n1097\n1155\n2135\n4\n4\n1443\n2413\n571\n2902\n187\n3606\n3627\n227\n2672\n1031\n858\n2636\n2715\n1776\n620\n955\n1931\n1195\n480\n2489\n1094\n2006\n14\n18\n44\n1157\n152\n1663\n725\n552\n336\n73\n894\n287\n2357\n128\n6\n526\n3385\n1902\n3396\n1706\n226\n163\n731\n1309\n915\n6\n686\n1683\n1422\n688\n535\n1497\n225\n1749\n3056\n2205\n150\n2488\n688\n2606\n2550\n1880\n1990\n1371\n830\n1854\n146\n2467\n329\n3522\n932\n617\n1943\n2938\n383\n2819\n1102\n466\n1103\n325\n408\n186\n478\n1735\n417\n284\n1008\n6\n3409\n2710\n1\n14\n117\n5\n940\n776\n1\n375\n658\n2594\n136\n3093\n508\n1558\n2256\n1328\n73\n2080\n547\n1426\n3224\n2225\n709\n22\n1024\n245\n2896\n2355\n566\n1436\n1999\n76\n63\n7\n1375\n569\n180\n376\n3245\n1970\n2389\n18\n3600\n661\n1848\n2786\n3308\n3152\n2232\n2735\n57\n109\n2402\n779\n572\n2365\n374\n2447\n2542\n147\n1659\n2701\n333\n2429\n812\n376\n3594\n2847\n10\n1864\n1105\n2791\n1791\n503\n375\n3520\n2703\n1136\n1558\n2399\n2637\n44\n1152\n77\n3506\n812\n3\n908\n1258\n334\n3287\n3538\n2218\n1323\n1120\n326\n1576\n254\n2326\n1464\n23\n1282\n2297\n686\n839\n3147\n153\n1187\n127\n2481\n1041\n508\n752\n1357\n2262\n1613\n2615\n981\n3206\n1709\n2381\n1071\n2536\n2871\n404\n282\n568\n2310\n609\n263\n200\n188\n37\n1039\n893\n1136\n1111\n1641\n22\n2988\n158\n863\n2935\n421\n1469\n656\n375\n2725\n32\n1121\n1419\n1213\n124\n2403\n1497\n29\n398\n2945\n341\n235\n895\n1479\n606\n438\n932\n2328\n390\n1382\n1598\n1078\n15\n1517\n1338\n24\n600\n2733\n1481\n1887\n2083\n2426\n306\n2171\n818\n70\n1221\n1674\n1172\n1675\n192\n916\n364\n2686\n1525\n3399\n1689\n110\n1405\n1623\n1565\n1193\n638\n279\n3128\n83\n709\n978\n563\n14\n1464\n458\n604\n32\n600\n2398\n1225\n160\n3438\n1653\n1965\n92\n144\n1125\n1700\n510\n1746\n1063\n2347\n110\n567\n2228\n2940\n1276\n1813\n499\n142\n61\n746\n609\n827\n3008\n1962\n375\n342\n3154\n1116\n1169\n14\n48\n599\n649\n1195\n1504\n12\n589\n301\n75\n1906\n3193\n75\n950\n2337\n14\n2469\n3415\n2503\n173\n418\n1800\n1476\n240\n770\n702\n296\n39\n969\n2367\n1989\n1703\n146\n2640\n522\n1717\n69\n3307\n1031\n1545\n1136\n169\n1767\n1338\n272\n868\n546\n2523\n310\n1855\n1822\n2840\n165\n1034\n2465\n229\n2205\n2181\n1838\n3169\n3528\n23\n280\n2046\n2295\n107\n281\n1560\n81\n1097\n126\n1153\n3303\n1072\n176\n192\n2307\n43\n1761\n3064\n321\n1062\n51\n2855\n2228\n410\n482\n1245\n2306\n397\n1364\n393\n3090\n519\n2214\n790\n969\n2570\n1757\n1207\n360\n846\n215\n144\n718\n87\n1281\n3314\n664\n1329\n890\n2\n372\n1483\n1297\n3322\n1366\n2588\n228\n313\n10\n355\n1401\n388\n774\n3\n160\n3024\n1798\n116\n713\n1961\n2314\n740\n1764\n1613\n1782\n1903\n1211\n29\n1764\n1987\n1049\n2215\n869\n377\n1571\n2689\n779\n17\n535\n44\n333\n1613\n756\n1669\n42\n2759\n2342\n3147\n2671\n471\n1448\n1448\n3092\n346\n1094\n191\n268\n1954\n3602\n943\n553\n1404\n655\n827\n735\n73\n562\n2094\n134\n158\n147\n910\n441\n12\n310\n952\n2864\n4\n1034\n144\n62\n303\n893\n1180\n28\n185\n484\n7\n613\n94\n3\n2748\n9\n756\n871\n1\n2584\n3595\n808\n95\n1045\n349\n271\n2970\n1071\n1191\n33\n171\n3605\n3322\n3628\n2364\n1415\n688\n3557\n15\n1343\n2273\n2968\n1629\n1019\n1368\n278\n2264\n1882\n227\n2210\n717\n2612\n2694\n2998\n805\n965\n233\n1360\n1305\n1366\n917\n1333\n2563\n1998\n3290\n1306\n565\n1836\n3161\n3349\n2087\n2678\n648\n752\n185\n186\n1302\n21\n836\n477\n404\n972\n304\n329\n136\n2646\n3000\n918\n227\n1000\n1770\n1296\n2044\n3233\n917\n897\n2661\n350\n1851\n525\n377\n290\n732\n1212\n2176\n1469\n1971\n1313\n1445\n564\n343\n2226\n1817\n663\n1556\n556\n1442\n1576\n1269\n186\n16\n1367\n203\n523\n456\n677\n409\n382\n369\n681\n2639\n3578\n282\n1531\n419\n803\n1172\n2886\n2129\n419\n2608\n2660\n981\n1101\n1109\n1024\n90\n7\n2049\n3106\n1985\n422\n2554\n323\n1400\n549\n2797\n301\n256\n7\n677\n1916\n526\n714\n1477\n554\n376\n909\n1048\n1890\n30\n3585\n1363\n3298\n742\n1092\n596\n243\n1142\n2003\n2426\n192\n791\n1366\n773\n1153\n2030\n1842\n1441\n65\n104\n2460\n1765\n744\n1\n1508\n1811\n3000\n3097\n1563\n900\n2729\n32\n280\n841\n39\n3378\n669\n1919\n282\n713\n640\n252\n1282\n1778\n1340\n1159\n667\n43\n1945\n3478\n868\n7\n553\n227\n2737\n41\n953\n2064\n174\n1403\n663\n1838\n259\n229\n1491\n830\n11\n203\n191\n621\n18\n2507\n842\n162\n189\n431\n2636\n652\n2097\n869\n1326\n1461\n2099\n2747\n1211\n741\n1909\n941\n278\n369\n1111\n646\n247\n1113\n768\n653\n1590\n365\n359\n1258\n462\n1772\n2924\n2993\n2\n2105\n677\n485\n2981\n1\n10\n1328\n2439\n2316\n1460\n714\n9\n252\n28\n3234\n2244\n1964\n1\n98\n204\n1873\n3226\n1497\n1910\n142\n645\n2721\n3517\n2657\n185\n3203\n1120\n11\n201\n337\n42\n1234\n1731\n1591\n888\n327\n130\n3042\n542\n95\n78\n482\n115\n5\n1788\n2580\n2287\n108\n1157\n3052\n28\n2077\n33\n740\n1910\n26\n204\n2409\n2240\n237\n964\n45\n3443\n2\n417\n2050\n852\n306\n3296\n3238\n1694\n1492\n125\n2203\n566\n939\n1329\n178\n2241\n614\n3133\n794\n2704\n551\n1633\n2197\n182\n2987\n78\n482\n1477\n83\n2515\n214\n1064\n1755\n2138\n662\n2327\n15\n1925\n874\n2859\n2\n441\n1040\n1272\n701\n1875\n337\n3349\n81\n2823\n3094\n3083\n557\n2245\n943\n11\n45\n705\n5\n1270\n1172\n13\n514\n888\n1840\n1826\n1133\n2436\n215\n3043\n1225\n399\n1739\n1466\n867\n2292\n790\n2859\n13\n5\n2073\n2736\n3593\n2169\n278\n583\n1749\n773\n417\n188\n1408\n373\n1510\n23\n2392\n1846\n17\n3351\n357\n1341\n318\n159\n2517\n366\n519\n214\n1024\n3344\n716\n" ], [ "count = 0\nt_val = 0\nfor values in ehr_claims_death: \n t_val = t_val + 1\n if (values == 1): \n count = count + 1\nprint(count, t_val)", "5003 52277\n" ], [ "ehr_claims_death = []\nfor rows in medicare[['ehr_death', 'claims_death']].itertuples():\n ehr_claims_death.append(rows[1] or rows[2])\ncount = 0\nmedicare['ehr_claims_death'] = ehr_claims_death \n\n ", "_____no_output_____" ], [ "def val_appended(card_outcomes): #MI, stroke, death\n for outcomes in card_outcomes:\n if (outcomes == 1):\n return 1\n return 0\nout_comp_cardiovascular_RC1 = [] \nfor index in range(medicare.shape[0]):\n rel_outcomes = [medicare.at[index,'Out_IscheStroke_RC1'], medicare.at[index,'Out_HemoStroke_RC1'], medicare.at[index, 'Out_UncertainStroke_RC1'], medicare.at[index, 'Out_MI_RC1'], medicare.at[index,'ehr_claims_death']]\n out_comp_cardiovascular_RC1.append(val_appended(rel_outcomes))\n\nmedicare.insert(loc = 1, column = \"Out_comp_RC1\", value = out_comp_cardiovascular_RC1)", "_____no_output_____" ], [ "medicare['Out_comp_RC1']", "_____no_output_____" ], [ "count = 0\nfor values in medicare['Out_comp_RC1']:\n if (values == 1):\n count = count + 1\nprint(count)", "6672\n" ], [ "medicare.to_csv(\"/netapp2/home/Data/final2_medicare.csv\")", "_____no_output_____" ] ] ]
[ "code" ]
[ [ "code", "code", "code", "code", "code", "code", "code", "code", "code" ] ]
ece827e170096a5b2d4d3e2f73e20deab76e333d
28,970
ipynb
Jupyter Notebook
idb_tf/explore_index.ipynb
iDigBio/idb-spark
abaedf871beae749c42cc737675db1b9b14d7d57
[ "MIT" ]
1
2019-03-11T08:17:46.000Z
2019-03-11T08:17:46.000Z
idb_tf/explore_index.ipynb
iDigBio/idb-spark
abaedf871beae749c42cc737675db1b9b14d7d57
[ "MIT" ]
null
null
null
idb_tf/explore_index.ipynb
iDigBio/idb-spark
abaedf871beae749c42cc737675db1b9b14d7d57
[ "MIT" ]
null
null
null
75.639687
17,224
0.78174
[ [ [ "import pyspark.sql.functions as sql\nimport pyspark.sql.types as types\nidb_df_version = \"20161119\"\nsize=\"\"", "_____no_output_____" ], [ "idb_tf_df = sqlContext.read.parquet(\"/guoda/data/idigbio-{0}-tf{1}.parquet\".format(idb_df_version, size))\nidb_tf_df.count()", "_____no_output_____" ], [ "idb_df = sqlContext.read.parquet(\"/guoda/data/idigbio-{0}{1}.parquet\".format(idb_df_version, size))\nidb_df.count()", "_____no_output_____" ], [ "idb_df_ids = (idb_df\n .select(idb_df[\"uuid\"].alias(\"idb_uuid\"),\n idb_df[\"catalognumber\"].alias(\"idb_catalognumber\"))\n )", "_____no_output_____" ], [ "idb_df_notes = (idb_df\n .select(idb_df[\"uuid\"].alias(\"note_uuid\"),\n sql.concat_ws(\" \", idb_df[\"data.dwc:occurrenceRemarks\"],\n idb_df[\"data.dwc:eventRemarks\"],\n idb_df[\"data.dwc:fieldNotes\"]\n )\n .alias(\"note\")\n )\n )", "_____no_output_____" ], [ "joined = (idb_df_ids \n .join(idb_tf_df, on=idb_df_ids[\"idb_catalognumber\"]==idb_tf_df[\"token\"])\n .join(idb_df_notes, on=sql.column(\"uuid\")==idb_df_notes[\"note_uuid\"])\n .withColumn(\"catalognumber_len\", sql.length(sql.column(\"idb_catalognumber\")))\n )\n\njoined.count()", "_____no_output_____" ], [ "joined.head()", "_____no_output_____" ], [ "length_pd = (joined\n .groupBy(joined[\"catalognumber_len\"])\n .count()\n ).toPandas()\nlength_pd.head(2)", "_____no_output_____" ], [ "length_pd.shape", "_____no_output_____" ], [ "import numpy\nlength_pd[\"log_count\"] = numpy.log10(length_pd[\"count\"])", "_____no_output_____" ], [ "print(length_pd)", " catalognumber_len count log_catalognumber_len log_count\n0 1 1314912287 0.000000 9.118897\n1 6 138556200 0.778151 8.141626\n2 16 2 1.204120 0.301030\n3 3 212986935 0.477121 8.328353\n4 5 123868957 0.698970 8.092962\n5 9 45511 0.954243 4.658116\n6 4 531911093 0.602060 8.725839\n7 8 35533 0.903090 4.550632\n8 7 73042 0.845098 4.863573\n9 10 4 1.000000 0.602060\n10 11 7017 1.041393 3.846151\n11 14 5 1.146128 0.698970\n12 2 613337855 0.301030 8.787700\n" ], [ "import matplotlib\nimport matplotlib.pyplot as plt\nmatplotlib.style.use('ggplot')\n\nplt.figure();\n(length_pd\n .sort_values(by=\"catalognumber_len\")\n .plot(x=\"catalognumber_len\", y=\"log_count\", kind=\"bar\")\n)\n\nplt.show()", "_____no_output_____" ], [ "joined.filter(joined[\"catalognumber_len\"]==11).head(10)", "_____no_output_____" ], [ "(joined\n .write\n .parquet(\"/outputs/idb-tokens-joined-{}.parquet\".format(idb_df_version))\n)", "_____no_output_____" ] ] ]
[ "code" ]
[ [ "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code" ] ]
ece828553549f21c29ef1b27e15634fc938c7dba
10,670
ipynb
Jupyter Notebook
2-Natural Language Processing with Probabilistic Models/week-2/Parts-of-Speech Tagging.ipynb
rahulverma7788/coursera-natural-language-processing-specialization-2-3650
687df12f6da6e8fc1c72a4c72e2d8fbd5b4fbe4b
[ "Apache-2.0" ]
null
null
null
2-Natural Language Processing with Probabilistic Models/week-2/Parts-of-Speech Tagging.ipynb
rahulverma7788/coursera-natural-language-processing-specialization-2-3650
687df12f6da6e8fc1c72a4c72e2d8fbd5b4fbe4b
[ "Apache-2.0" ]
null
null
null
2-Natural Language Processing with Probabilistic Models/week-2/Parts-of-Speech Tagging.ipynb
rahulverma7788/coursera-natural-language-processing-specialization-2-3650
687df12f6da6e8fc1c72a4c72e2d8fbd5b4fbe4b
[ "Apache-2.0" ]
null
null
null
39.962547
379
0.709372
[ [ [ "empty" ] ] ]
[ "empty" ]
[ [ "empty" ] ]
ece841dc8fdd903514973d5c9126e83fe24c0bb1
7,818
ipynb
Jupyter Notebook
PredictWaterSplit/Regression package.ipynb
xqzh/Perovskite
5d66c67953f1c6571f1f3c3b89d6571619587e31
[ "MIT" ]
1
2021-08-23T09:09:45.000Z
2021-08-23T09:09:45.000Z
PredictWaterSplit/Regression package.ipynb
xqzh/Perovskite
5d66c67953f1c6571f1f3c3b89d6571619587e31
[ "MIT" ]
null
null
null
PredictWaterSplit/Regression package.ipynb
xqzh/Perovskite
5d66c67953f1c6571f1f3c3b89d6571619587e31
[ "MIT" ]
6
2018-02-13T23:36:05.000Z
2021-03-24T04:15:18.000Z
23.477477
161
0.574572
[ [ [ "import pandas as pd", "_____no_output_____" ], [ "from KRRcode import open_database", "_____no_output_____" ], [ "from KRRcode import predictbyKRR", "_____no_output_____" ], [ "from KRRcode import outlier", "_____no_output_____" ], [ "from KRRcode import featureselect", "_____no_output_____" ] ], [ [ "## This notebook is to show how we make prediction with kernel ridge regression and analyze the outliers", "_____no_output_____" ] ], [ [ "perovskite,values,data_total = open_database.read_database()", "_____no_output_____" ], [ "traindata,testdata = predictbyKRR.build_data(data_total)", "_____no_output_____" ], [ "feature_list = ['anion_X', 'anion_IE', 'A_X', 'A_IE', 'A_aff', 'B_IE', 'B_aff', 'volume', 'mass', 'A_R', 'B_R', 'aff_A+B', 'aff_A-B', 'A_R_max', 'B_R_max']", "_____no_output_____" ] ], [ [ "### Predict the band energy and heat of formation", "_____no_output_____" ] ], [ [ "CB_dir_predict_train,CB_dir_predict_test = predictbyKRR.predict_Band(feature_list,traindata,testdata,'CB_dir')", "_____no_output_____" ], [ "predictbyKRR.plot_predict(CB_dir_predict_test,testdata.CB_dir,\n CB_dir_predict_train,traindata.CB_dir,\n 'prediction of direct conduction band')", "_____no_output_____" ], [ "CB_ind_predict_train,CB_ind_predict_test = predictbyKRR.predict_Band(feature_list,traindata,testdata,'CB_ind')", "_____no_output_____" ], [ "type(CB_ind_predict_train)", "_____no_output_____" ], [ "predictbyKRR.plot_predict(CB_ind_predict_test,testdata.CB_ind,\n CB_ind_predict_train,traindata.CB_ind,\n 'prediction of indirect conduction band')", "_____no_output_____" ], [ "VB_dir_predict_train,VB_dir_predict_test = predictbyKRR.predict_Band(feature_list,traindata,testdata,'VB_dir')", "_____no_output_____" ], [ "predictbyKRR.plot_predict(VB_dir_predict_test,testdata.VB_dir,\n VB_dir_predict_train,traindata.VB_dir,\n 'prediction of direct valence band')", "_____no_output_____" ], [ "VB_ind_predict_train,VB_ind_predict_test = predictbyKRR.predict_Band(feature_list,traindata,testdata,'VB_ind')", "_____no_output_____" ], [ "predictbyKRR.plot_predict(VB_ind_predict_test,testdata.VB_ind,\n VB_ind_predict_train,traindata.VB_ind,\n 'prediction of indirect valence band')", "_____no_output_____" ], [ "hf_predict_train,hf_predict_test = predictbyKRR.predict_Band(feature_list,traindata,testdata,'heat_of_formation')", "_____no_output_____" ], [ "predictbyKRR.plot_predict(hf_predict_test,testdata.heat_of_formation,\n hf_predict_train,traindata.heat_of_formation,\n 'prediction of heat of formation')", "_____no_output_____" ] ], [ [ "### find the recurring outliers in testdata and plot them", "_____no_output_____" ] ], [ [ "pred_compare = testdata.copy()\npred_compare['CB_ind_pred'] = CB_ind_predict_test\npred_compare['VB_ind_pred'] = VB_ind_predict_test\npred_compare['CB_dir_pred'] = CB_dir_predict_test\npred_compare['VB_dir_pred'] = VB_dir_predict_test\npred_compare['hf_pred'] = hf_predict_test\n\npred_compare['ind_gap'] = pred_compare['CB_ind_pred']-pred_compare['VB_ind_pred']\npred_compare['dir_gap'] = pred_compare['CB_dir_pred']-pred_compare['VB_dir_pred']\n", "_____no_output_____" ], [ "recuroutlier_test = outlier.collectoutlier(pred_compare)", "_____no_output_____" ], [ "recuroutlier_train.groupby('anion').count()", "_____no_output_____" ], [ "recuroutlier_test.groupby('anion').count()", "_____no_output_____" ], [ "outlier.plot_outlier(CB_dir_predict_test,testdata.CB_dir,\n recuroutlier_test.CB_dir_pred,recuroutlier_test.CB_dir,\n 'Prediction of Direct Conduction Band')", "_____no_output_____" ], [ "outlier.plot_outlier(VB_dir_predict_test,testdata.VB_dir,\n recuroutlier_test.VB_dir_pred,recuroutlier_test.VB_dir,\n 'Prediction of Direct Valence Band')", "_____no_output_____" ], [ "outlier.plot_outlier(VB_ind_predict_test,testdata.VB_ind,\n recuroutlier_test.VB_ind_pred,recuroutlier_test.VB_ind,\n 'Prediction of Indirect Valence Band')", "_____no_output_____" ], [ "outlier.plot_outlier(CB_ind_predict_test,testdata.CB_ind,\n recuroutlier_test.CB_ind_pred,recuroutlier_test.CB_ind,\n 'Prediction of Indirect Valence Band')", "_____no_output_____" ] ] ]
[ "code", "markdown", "code", "markdown", "code", "markdown", "code" ]
[ [ "code", "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code", "code", "code", "code" ] ]
ece8515ee6d79d67e50d441ba62c1a3462520dae
13,025
ipynb
Jupyter Notebook
ProblemSets/PS1-julia-intro/gps_01.ipynb
gpetrini/OU_Econometrics_III
81b5f35999d6b10f65f84b3344593d1c28acf1e2
[ "MIT" ]
null
null
null
ProblemSets/PS1-julia-intro/gps_01.ipynb
gpetrini/OU_Econometrics_III
81b5f35999d6b10f65f84b3344593d1c28acf1e2
[ "MIT" ]
null
null
null
ProblemSets/PS1-julia-intro/gps_01.ipynb
gpetrini/OU_Econometrics_III
81b5f35999d6b10f65f84b3344593d1c28acf1e2
[ "MIT" ]
null
null
null
28.689427
373
0.509175
[ [ [ "empty" ] ] ]
[ "empty" ]
[ [ "empty" ] ]
ece85fa8b90d4b1f5fcdc0d01162488118a4782c
421,538
ipynb
Jupyter Notebook
nbs/001_utils.ipynb
duchaba/tsai
694c26c22fa102b236fe4763111095c40ec8d7b3
[ "Apache-2.0" ]
null
null
null
nbs/001_utils.ipynb
duchaba/tsai
694c26c22fa102b236fe4763111095c40ec8d7b3
[ "Apache-2.0" ]
null
null
null
nbs/001_utils.ipynb
duchaba/tsai
694c26c22fa102b236fe4763111095c40ec8d7b3
[ "Apache-2.0" ]
null
null
null
96.638698
76,552
0.817639
[ [ [ "# default_exp utils", "_____no_output_____" ] ], [ [ "# Utilities\n\n> Helper functions used throughout the library not related to timeseries data.", "_____no_output_____" ] ], [ [ "#export\nimport string\nfrom scipy.stats import ttest_ind, ks_2samp, pearsonr, spearmanr, normaltest, linregress\nfrom tsai.imports import *", "_____no_output_____" ], [ "# ensure these folders exist for testing purposes\nfns = ['data', 'export', 'models']\nfor fn in fns: \n path = Path('.')/fn\n if not os.path.exists(path): os.makedirs(path)", "_____no_output_____" ], [ "#export\ndef totensor(o):\n if isinstance(o, torch.Tensor): return o\n elif isinstance(o, np.ndarray): return torch.from_numpy(o)\n elif isinstance(o, pd.DataFrame): return torch.from_numpy(o.values)\n else: \n try: return torch.tensor(o)\n except: warn(f\"Can't convert {type(o)} to torch.Tensor\", Warning)\n\n\ndef toarray(o):\n if isinstance(o, np.ndarray): return o\n elif isinstance(o, torch.Tensor): return o.cpu().numpy()\n elif isinstance(o, pd.DataFrame): return o.values\n else:\n try: return np.asarray(o)\n except: warn(f\"Can't convert {type(o)} to np.array\", Warning)\n \n \ndef toL(o):\n if isinstance(o, L): return o\n elif isinstance(o, (np.ndarray, torch.Tensor)): return L(o.tolist())\n else:\n try: return L(o)\n except: warn(f'passed object needs to be of type L, list, np.ndarray or torch.Tensor but is {type(o)}', Warning)\n\n\ndef to3dtensor(o):\n o = totensor(o)\n if o.ndim == 3: return o\n elif o.ndim == 1: return o[None, None]\n elif o.ndim == 2: return o[:, None]\n assert False, f'Please, review input dimensions {o.ndim}'\n\n\ndef to2dtensor(o):\n o = totensor(o)\n if o.ndim == 2: return o\n elif o.ndim == 1: return o[None]\n elif o.ndim == 3: return o[0]\n assert False, f'Please, review input dimensions {o.ndim}'\n\n\ndef to1dtensor(o):\n o = totensor(o)\n if o.ndim == 1: return o\n elif o.ndim == 3: return o[0,0]\n if o.ndim == 2: return o[0]\n assert False, f'Please, review input dimensions {o.ndim}'\n\n\ndef to3darray(o):\n o = toarray(o)\n if o.ndim == 3: return o\n elif o.ndim == 1: return o[None, None]\n elif o.ndim == 2: return o[:, None]\n assert False, f'Please, review input dimensions {o.ndim}'\n\n\ndef to2darray(o):\n o = toarray(o)\n if o.ndim == 2: return o\n elif o.ndim == 1: return o[None]\n elif o.ndim == 3: return o[0]\n assert False, f'Please, review input dimensions {o.ndim}'\n\n\ndef to1darray(o):\n o = toarray(o)\n if o.ndim == 1: return o\n elif o.ndim == 3: o = o[0,0]\n elif o.ndim == 2: o = o[0]\n assert False, f'Please, review input dimensions {o.ndim}'\n \n \ndef to3d(o):\n if o.ndim == 3: return o\n if isinstance(o, (np.ndarray, pd.DataFrame)): return to3darray(o)\n if isinstance(o, torch.Tensor): return to3dtensor(o)\n \n \ndef to2d(o):\n if o.ndim == 2: return o\n if isinstance(o, np.ndarray): return to2darray(o)\n if isinstance(o, torch.Tensor): return to2dtensor(o)\n \n \ndef to1d(o):\n if o.ndim == 1: return o\n if isinstance(o, np.ndarray): return to1darray(o)\n if isinstance(o, torch.Tensor): return to1dtensor(o)\n \n \ndef to2dPlus(o):\n if o.ndim >= 2: return o\n if isinstance(o, np.ndarray): return to2darray(o)\n elif isinstance(o, torch.Tensor): return to2dtensor(o)\n \n \ndef to3dPlus(o):\n if o.ndim >= 3: return o\n if isinstance(o, np.ndarray): return to3darray(o)\n elif isinstance(o, torch.Tensor): return to3dtensor(o)\n \n \ndef to2dPlusTensor(o):\n return to2dPlus(totensor(o))\n\n\ndef to2dPlusArray(o):\n return to2dPlus(toarray(o))\n\n\ndef to3dPlusTensor(o):\n return to3dPlus(totensor(o))\n\n\ndef to3dPlusArray(o):\n return to3dPlus(toarray(o))\n\n\ndef todtype(dtype):\n def _to_type(o, dtype=dtype):\n if o.dtype == dtype: return o\n elif isinstance(o, torch.Tensor): o = o.to(dtype=dtype)\n elif isinstance(o, np.ndarray): o = o.astype(dtype)\n return o\n return _to_type", "_____no_output_____" ], [ "a = np.random.rand(100).astype(np.float32)\nb = torch.from_numpy(a).float()\ntest_eq(totensor(a), b)\ntest_eq(a, toarray(b))\ntest_eq(to3dtensor(a).ndim, 3)\ntest_eq(to2dtensor(a).ndim, 2)\ntest_eq(to1dtensor(a).ndim, 1)\ntest_eq(to3darray(b).ndim, 3)\ntest_eq(to2darray(b).ndim, 2)\ntest_eq(to1darray(b).ndim, 1)", "_____no_output_____" ], [ "data = np.random.rand(10, 20)\ndf = pd.DataFrame(data)\ndf['target'] = np.random.randint(0, 3, len(df))\nX = df[df.columns[:-1]]\ny = df['target']\ntest_eq(to3darray(X).shape, (10, 1, 20))\ntest_eq(toarray(y).shape, (10,))", "_____no_output_____" ], [ "#export\ndef bytes2size(\n size_bytes : int, # Number of bytes \n decimals=2 # Number of decimals in the output\n )->str:\n if size_bytes == 0: return \"0B\"\n size_name = (\"B\", \"KB\", \"MB\", \"GB\", \"TB\", \"PB\", \"EB\", \"ZB\", \"YB\")\n i = int(math.floor(math.log(size_bytes, 1024)))\n p = math.pow(1024, i)\n s = round(size_bytes / p, decimals)\n return f'{s} {size_name[i]}'\n\ndef bytes2GB(byts):\n return round(byts / math.pow(1024, 3), 2)\n\ndef get_size(\n o, # Any object \n return_str : bool = True, # True returns size in human-readable format (KB, MB, GB, ...). False in bytes.\n decimals : int = 2, # Number of decimals in the output\n ):\n s = sys.getsizeof(o)\n if return_str: return bytes2size(s, decimals=decimals)\n else: return s\n\ndef get_dir_size(\n dir_path : str, # path to directory \n return_str : bool = True, # True returns size in human-readable format (KB, MB, GB, ...). False in bytes.\n decimals : int = 2, # Number of decimals in the output\n verbose : bool = False, # Controls verbosity\n ):\n assert os.path.isdir(dir_path)\n total_size = 0\n for dirpath, dirnames, filenames in os.walk(dir_path):\n for f in filenames:\n fp = os.path.join(dirpath, f)\n # skip if it is symbolic link\n if not os.path.islink(fp):\n fp_size = os.path.getsize(fp)\n total_size += fp_size\n pv(f'file: {fp[-50:]:50} size: {fp_size}', verbose)\n if return_str: \n return bytes2size(total_size, decimals=decimals)\n return total_size\n\ndef get_file_size(\n file_path : str, # path to file \n return_str : bool = True, # True returns size in human-readable format (KB, MB, GB, ...). False in bytes.\n decimals : int = 2, # Number of decimals in the output\n ):\n assert os.path.isfile(file_path)\n fsize = os.path.getsize(file_path)\n if return_str: \n return bytes2size(fsize, decimals=decimals)\n return fsize", "_____no_output_____" ], [ "a = np.random.rand(10, 5, 3)\ntest_eq(get_size(a, True), '1.3 KB')", "_____no_output_____" ], [ "#export\ndef is_file(path):\n return os.path.isfile(path)\n\ndef is_dir(path):\n return os.path.isdir(path)", "_____no_output_____" ], [ "test_eq(is_file(\"001_utils.ipynb\"), True)\ntest_eq(is_file(\"utils.ipynb\"), False)", "_____no_output_____" ], [ "#export\ndef delete_all_in_dir(tgt_dir, exception=None):\n import shutil\n if exception is not None and len(L(exception)) > 1: exception = tuple(exception)\n for file in os.listdir(tgt_dir):\n if exception is not None and file.endswith(exception): continue\n file_path = os.path.join(tgt_dir, file)\n if os.path.isfile(file_path) or os.path.islink(file_path): os.unlink(file_path)\n elif os.path.isdir(file_path): shutil.rmtree(file_path)", "_____no_output_____" ], [ "#export\ndef reverse_dict(dictionary): \n return {v: k for k, v in dictionary.items()}", "_____no_output_____" ], [ "#export\ndef is_tuple(o): return isinstance(o, tuple)", "_____no_output_____" ], [ "#export\ndef itemify(*o, tup_id=None): \n o = [o_ for o_ in L(*o) if o_ is not None]\n items = L(o).zip()\n if tup_id is not None: return L([item[tup_id] for item in items])\n else: return items", "_____no_output_____" ], [ "a = [1, 2, 3]\nb = [4, 5, 6]\nprint(itemify(a, b))\ntest_eq(len(itemify(a, b)), len(a))\na = [1, 2, 3]\nb = None\nprint(itemify(a, b))\ntest_eq(len(itemify(a, b)), len(a))\na = [1, 2, 3]\nb = [4, 5, 6]\nc = None\nprint(itemify(a, b, c))\ntest_eq(len(itemify(a, b, c)), len(a))", "[(1, 4), (2, 5), (3, 6)]\n[(1,), (2,), (3,)]\n[(1, 4), (2, 5), (3, 6)]\n" ], [ "#export\ndef isnone(o):\n return o is None\n\ndef exists(o): return o is not None\n\ndef ifelse(a, b, c):\n \"`b` if `a` is True else `c`\"\n return b if a else c", "_____no_output_____" ], [ "a = np.array(3)\ntest_eq(isnone(a), False)\ntest_eq(exists(a), True)\nb = None\ntest_eq(isnone(b), True)\ntest_eq(exists(b), False)", "_____no_output_____" ], [ "#export\ndef is_not_close(a, b, eps=1e-5):\n \"Is `a` within `eps` of `b`\"\n if hasattr(a, '__array__') or hasattr(b, '__array__'):\n return (abs(a - b) > eps).all()\n if isinstance(a, (Iterable, Generator)) or isinstance(b, (Iterable, Generator)):\n return is_not_close(np.array(a), np.array(b), eps=eps)\n return abs(a - b) > eps\n\n\ndef test_not_close(a, b, eps=1e-5):\n \"`test` that `a` is within `eps` of `b`\"\n test(a, b, partial(is_not_close, eps=eps), 'not_close')\n\n\ndef test_type(a, b):\n return test_eq(type(a), type(b))\n\n\ndef test_ok(f, *args, **kwargs):\n try: \n f(*args, **kwargs)\n e = 0\n except: \n e = 1\n pass\n test_eq(e, 0)\n \ndef test_not_ok(f, *args, **kwargs):\n try: \n f(*args, **kwargs)\n e = 0\n except: \n e = 1\n pass\n test_eq(e, 1)\n \ndef test_error(error, f, *args, **kwargs):\n try: f(*args, **kwargs)\n except Exception as e: \n test_eq(str(e), error)\n \ndef test_eq_nan(a,b):\n \"`test` that `a==b` excluding nan values (valid for torch.Tensor and np.ndarray)\"\n mask_a = torch.isnan(a) if isinstance(a, torch.Tensor) else np.isnan(a)\n mask_b = torch.isnan(b) if isinstance(b, torch.Tensor) else np.isnan(b)\n test(a[~mask_a],b[~mask_b],equals, '==')", "_____no_output_____" ], [ "#export\ndef assert_fn(*args, **kwargs): assert False, 'assertion test'\ntest_error('assertion test', assert_fn, 35, a=3)", "_____no_output_____" ], [ "#export\ndef test_gt(a,b):\n \"`test` that `a>b`\"\n test(a,b,gt,'>')\n\ndef test_ge(a,b):\n \"`test` that `a>=b`\"\n test(a,b,ge,'>')\n \ndef test_lt(a,b):\n \"`test` that `a>b`\"\n test(a,b,lt,'<')\n\ndef test_le(a,b):\n \"`test` that `a>b`\"\n test(a,b,le,'<=')", "_____no_output_____" ], [ "test_ok(test_gt, 5, 4)\ntest_not_ok(test_gt, 4, 4)\ntest_ok(test_ge, 4, 4)\ntest_not_ok(test_ge, 3, 4)\n\ntest_ok(test_lt, 3, 4)\ntest_not_ok(test_lt, 4, 4)\ntest_ok(test_le, 4, 4)\ntest_not_ok(test_le, 5, 4)", "_____no_output_____" ], [ "t = torch.rand(100)\nt[t<.5] = np.nan\ntest_ne(t, t)\ntest_eq_nan(t, t)", "_____no_output_____" ], [ "#export\ndef stack(o, axis=0, retain=True):\n if hasattr(o, '__array__'): return o\n if isinstance(o[0], torch.Tensor):\n return retain_type(torch.stack(tuple(o), dim=axis), o[0]) if retain else torch.stack(tuple(o), dim=axis)\n else:\n return retain_type(np.stack(o, axis), o[0]) if retain else np.stack(o, axis)\n \n \ndef stack_pad(o, padding_value=np.nan):\n 'Converts a an iterable into a numpy array using padding if necessary'\n if not is_listy(o) or not is_array(o):\n if not hasattr(o, \"ndim\"): o = np.asarray([o])\n else: o = np.asarray(o)\n o_ndim = 1\n if o.ndim > 1:\n o_ndim = o.ndim\n o_shape = o.shape\n o = o.flatten()\n o = [oi if (is_array(oi) and oi.ndim > 0) or is_listy(oi) else [oi] for oi in o]\n row_length = len(max(o, key=len))\n result = np.full((len(o), row_length), padding_value)\n for i,row in enumerate(o):\n result[i, :len(row)] = row\n if o_ndim > 1:\n if row_length == 1:\n result = result.reshape(*o_shape)\n else:\n result = result.reshape(*o_shape, row_length)\n return result", "_____no_output_____" ], [ "o = [[0,1,2], [4,5,6,7]]\ntest_eq(stack_pad(o).shape, (1, 2, 4))\ntest_eq(type(stack_pad(o)), np.ndarray)\ntest_eq(np.isnan(stack_pad(o)).sum(), 1)", "/Users/nacho/opt/anaconda3/envs/py37torch110/lib/python3.7/site-packages/ipykernel_launcher.py:13: VisibleDeprecationWarning: Creating an ndarray from ragged nested sequences (which is a list-or-tuple of lists-or-tuples-or ndarrays with different lengths or shapes) is deprecated. If you meant to do this, you must specify 'dtype=object' when creating the ndarray.\n del sys.path[0]\n" ], [ "o = 3\nprint(stack_pad(o))\ntest_eq(stack_pad(o), np.array([[3.]]))\no = [4,5]\nprint(stack_pad(o))\ntest_eq(stack_pad(o), np.array([[4., 5.]]))\no = [[0,1,2], [4,5,6,7]]\nprint(stack_pad(o))\no = np.array([0, [1,2]], dtype=object)\nprint(stack_pad(o))\no = np.array([[[0], [10, 20], [100, 200, 300]], [[0, 1, 2, 3], [10, 20], [100]]], dtype=object)\nprint(stack_pad(o))\no = np.array([0, [10, 20]], dtype=object)\nprint(stack_pad(o))", "[[3.]]\n[[4. 5.]]\n[[[ 0. 1. 2. nan]\n [ 4. 5. 6. 7.]]]\n[[ 0. nan]\n [ 1. 2.]]\n[[[ 0. nan nan nan]\n [ 10. 20. nan nan]\n [100. 200. 300. nan]]\n\n [[ 0. 1. 2. 3.]\n [ 10. 20. nan nan]\n [100. nan nan nan]]]\n[[ 0. nan]\n [10. 20.]]\n" ], [ "a = np.random.rand(2, 3, 4)\nt = torch.from_numpy(a)\ntest_eq_type(stack(itemify(a, tup_id=0)), a)\ntest_eq_type(stack(itemify(t, tup_id=0)), t)", "_____no_output_____" ], [ "#export\ndef pad_sequences(\n o, # Iterable object\n maxlen:int=None, # Optional max length of the output. If None, max length of the longest individual sequence.\n dtype:(str, type)=np.float64, # Type of the output sequences. To pad sequences with variable length strings, you can use object.\n padding:str='pre', # 'pre' or 'post' pad either before or after each sequence.\n truncating:str='pre', # 'pre' or 'post' remove values from sequences larger than maxlen, either at the beginning or at the end of the sequences.\n padding_value:float=np.nan, # Value used for padding.\n):\n \"Transforms an iterable with sequences into a 3d numpy array using padding or truncating sequences if necessary\"\n \n assert padding in ['pre', 'post']\n assert truncating in ['pre', 'post']\n assert is_iter(o)\n\n if not is_array(o):\n o = [to2darray(oi) for oi in o]\n seq_len = maxlen or max(o, key=len).shape[-1]\n result = np.full((len(o), o[0].shape[-2], seq_len), padding_value, dtype=dtype)\n for i,values in enumerate(o):\n if truncating == 'pre':\n values = values[..., -seq_len:]\n else:\n values = values[..., :seq_len]\n if padding == 'pre':\n result[i, :, -values.shape[-1]:] = values\n else:\n result[i, :, :values.shape[-1]] = values \n return result", "_____no_output_____" ] ], [ [ "This function transforms a list (of length n_samples) of sequences into a 3d numpy array of shape:\n\n```bash\n [n_samples x n_vars x seq_len]\n```\n\nseq_len is either the maxlen argument if provided, or the length of the longest sequence in the list.\n\nSequences that are shorter than seq_len are padded with value until they are seq_len long.\n\nSequences longer than seq_len are truncated so that they fit the desired length.\n\nThe position where padding or truncation happens is determined by the arguments padding and truncating, respectively. Pre-padding or removing values from the beginning of the sequence is the default.\n\nInput sequences to pad_sequences may be have 1, 2 or 3 dimensions:", "_____no_output_____" ] ], [ [ "# 1 dim\na1 = np.arange(6)\na2 = np.arange(3) * 10\na3 = np.arange(2) * 100\no = [a1, a2, a3]\npadded_o = pad_sequences(o, maxlen=4, dtype=np.float64, padding='post', truncating='pre', padding_value=np.nan)\ntest_eq(padded_o.shape, (3, 1, 4))\npadded_o", "_____no_output_____" ], [ "# 2 dim\na1 = np.arange(12).reshape(2, 6)\na2 = np.arange(6).reshape(2, 3) * 10\na3 = np.arange(4).reshape(2, 2) * 100\no = [a1, a2, a3]\npadded_o = pad_sequences(o, maxlen=4, dtype=np.float64, padding='post', truncating='pre', padding_value=np.nan)\ntest_eq(padded_o.shape, (3, 2, 4))\npadded_o", "_____no_output_____" ], [ "# 3 dim\na1 = np.arange(10).reshape(1, 2, 5)\na2 = np.arange(6).reshape(1, 2, 3) * 10\na3 = np.arange(4).reshape(1, 2, 2) * 100\no = [a1, a2, a3]\npadded_o = pad_sequences(o, maxlen=None, dtype=np.float64, padding='pre', truncating='pre', padding_value=np.nan)\ntest_eq(padded_o.shape, (3, 2, 5))\npadded_o", "_____no_output_____" ], [ "# 3 dim\na1 = np.arange(10).reshape(1, 2, 5)\na2 = np.arange(6).reshape(1, 2, 3) * 10\na3 = np.arange(4).reshape(1, 2, 2) * 100\no = [a1, a2, a3]\npadded_o = pad_sequences(o, maxlen=4, dtype=np.float64, padding='pre', truncating='pre', padding_value=np.nan)\ntest_eq(padded_o.shape, (3, 2, 4))\npadded_o", "_____no_output_____" ], [ "# 3 dim\na1 = np.arange(10).reshape(1, 2, 5)\na2 = np.arange(6).reshape(1, 2, 3) * 10\na3 = np.arange(4).reshape(1, 2, 2) * 100\no = [a1, a2, a3]\npadded_o = pad_sequences(o, maxlen=4, dtype=np.float64, padding='post', truncating='pre', padding_value=np.nan)\ntest_eq(padded_o.shape, (3, 2, 4))\npadded_o", "_____no_output_____" ], [ "# 3 dim\na1 = np.arange(10).reshape(1, 2, 5)\na2 = np.arange(6).reshape(1, 2, 3) * 10\na3 = np.arange(4).reshape(1, 2, 2) * 100\no = [a1, a2, a3]\npadded_o = pad_sequences(o, maxlen=4, dtype=np.float64, padding='post', truncating='post', padding_value=np.nan)\ntest_eq(padded_o.shape, (3, 2, 4))\npadded_o", "_____no_output_____" ], [ "# iterable is a list of lists\na1 = np.arange(12).reshape(1, 2, 6).tolist()\na2 = (np.arange(6).reshape(1, 2, 3) * 10).tolist()\na3 = (np.arange(4).reshape(1, 2, 2) * 100).tolist()\no = [a1, a2, a3]\npadded_o = pad_sequences(o, maxlen=None, dtype=np.float64, padding='post', truncating='pre', padding_value=np.nan)\ntest_eq(padded_o.shape, (3, 2, 6))\npadded_o", "_____no_output_____" ], [ "#export\ndef match_seq_len(*arrays):\n max_len = stack([x.shape[-1] for x in arrays]).max()\n return [np.pad(x, pad_width=((0,0), (0,0), (max_len - x.shape[-1], 0)), mode='constant', constant_values=0) for x in arrays]", "_____no_output_____" ], [ "a = np.random.rand(10, 5, 8)\nb = np.random.rand(3, 5, 10)\nc, d = match_seq_len(a, b)\ntest_eq(c.shape[-1], d.shape[-1])", "_____no_output_____" ], [ "#export\ndef random_shuffle(o, random_state=None):\n import sklearn\n res = sklearn.utils.shuffle(o, random_state=random_state)\n if isinstance(o, L): return L(list(res))\n return res", "_____no_output_____" ], [ "a = np.arange(10)\ntest_eq_type(random_shuffle(a, 1), np.array([2, 9, 6, 4, 0, 3, 1, 7, 8, 5]))\nt = torch.arange(10)\ntest_eq_type(random_shuffle(t, 1), tensor([2, 9, 6, 4, 0, 3, 1, 7, 8, 5]))\nl = list(a)\ntest_eq(random_shuffle(l, 1), [2, 9, 6, 4, 0, 3, 1, 7, 8, 5])\nl2 = L(l)\ntest_eq_type(random_shuffle(l2, 1), L([2, 9, 6, 4, 0, 3, 1, 7, 8, 5]))", "_____no_output_____" ], [ "#export\ndef cat2int(o):\n from fastai.data.transforms import Categorize\n from fastai.data.core import TfmdLists\n cat = Categorize()\n cat.setup(o)\n return stack(TfmdLists(o, cat)[:])", "_____no_output_____" ], [ "a = np.array(['b', 'a', 'a', 'b', 'a', 'b', 'a'])\ntest_eq_type(cat2int(a), TensorCategory([1, 0, 0, 1, 0, 1, 0]))", "_____no_output_____" ], [ "TensorBase([1,2,3])", "_____no_output_____" ], [ "#export\ndef cycle_dl(dl, show_progress_bar=False):\n if show_progress_bar: \n for _ in progress_bar(dl): _\n else: \n for _ in dl: _\n \ndef cycle_dl_to_device(dl, show_progress_bar=False):\n if show_progress_bar: \n for bs in progress_bar(dl): [b.to(default_device()) for b in bs]\n else:\n for bs in dl: [b.to(default_device()) for b in bs]\n \ndef cycle_dl_estimate(dl, iters=10):\n iters = min(iters, len(dl))\n iterator = iter(dl)\n timer.start(False)\n for _ in range(iters): next(iterator)\n t = timer.stop()\n return (t/iters * len(dl)).total_seconds()", "_____no_output_____" ], [ "#export\ndef cache_data(o, slice_len=10_000, verbose=False):\n start = 0\n n_loops = (len(o) - 1) // slice_len + 1\n pv(f'{n_loops} loops', verbose)\n timer.start(False)\n for i in range(n_loops):\n o[slice(start,start + slice_len)] \n if verbose and (i+1) % 10 == 0: print(f'{i+1:4} elapsed time: {timer.elapsed()}')\n start += slice_len\n pv(f'{i+1:4} total time : {timer.stop()}\\n', verbose)\n \nmemmap2cache = cache_data\ncache_memmap = cache_data", "_____no_output_____" ], [ "#export\ndef get_func_defaults(f): \n import inspect\n fa = inspect.getfullargspec(f)\n if fa.defaults is None: return dict(zip(fa.args, [''] * (len(fa.args))))\n else: return dict(zip(fa.args, [''] * (len(fa.args) - len(fa.defaults)) + list(fa.defaults)))", "_____no_output_____" ], [ "#export\ndef get_idx_from_df_col_vals(df, col, val_list):\n return [df[df[col] == val].index[0] for val in val_list]", "_____no_output_____" ], [ "#export\ndef get_sublist_idxs(aList, bList):\n \"Get idxs that when applied to aList will return bList. aList must contain all values in bList\"\n sorted_aList = aList[np.argsort(aList)]\n return np.argsort(aList)[np.searchsorted(sorted_aList, bList)]", "_____no_output_____" ], [ "x = np.array([3, 5, 7, 1, 9, 8, 6, 2])\ny = np.array([6, 1, 5, 7])\nidx = get_sublist_idxs(x, y)\ntest_eq(x[idx], y)\nx = np.array([3, 5, 7, 1, 9, 8, 6, 6, 2])\ny = np.array([6, 1, 5, 7, 5])\nidx = get_sublist_idxs(x, y)\ntest_eq(x[idx], y)", "_____no_output_____" ], [ "#export\ndef flatten_list(l):\n return [item for sublist in l for item in sublist]", "_____no_output_____" ], [ "#export\ndef display_pd_df(df, max_rows:Union[bool, int]=False, max_columns:Union[bool, int]=False):\n if max_rows:\n old_max_rows = pd.get_option('display.max_rows')\n if max_rows is not True and isinstance(max_rows, Integral): pd.set_option('display.max_rows', max_rows)\n else: pd.set_option('display.max_rows', df.shape[0])\n if max_columns:\n old_max_columns = pd.get_option('display.max_columns')\n if max_columns is not True and isinstance(max_columns, Integral): pd.set_option('display.max_columns', max_columns)\n else: pd.set_option('display.max_columns', df.shape[1])\n display(df)\n if max_rows: pd.set_option('display.max_rows', old_max_rows)\n if max_columns: pd.set_option('display.max_columns', old_max_columns)", "_____no_output_____" ], [ "old_max_rows, old_max_columns = pd.get_option('display.max_rows'), pd.get_option('display.max_columns')\ndf = pd.DataFrame(np.random.rand(70, 25))\ndisplay_pd_df(df, max_rows=2, max_columns=3)\ntest_eq(old_max_rows, pd.get_option('display.max_rows'))\ntest_eq(old_max_columns, pd.get_option('display.max_columns'))", "_____no_output_____" ], [ "#export\ndef ttest(data1, data2, equal_var=False):\n \"Calculates t-statistic and p-value based on 2 sample distributions\"\n t_stat, p_value = ttest_ind(data1, data2, equal_var=equal_var)\n return t_stat, np.sign(t_stat) * p_value\n\ndef kstest(data1, data2, alternative='two-sided', mode='auto', by_axis=None):\n \"\"\"Performs the two-sample Kolmogorov-Smirnov test for goodness of fit.\n \n Parameters\n data1, data2: Two arrays of sample observations assumed to be drawn from a continuous distributions. Sample sizes can be different.\n alternative: {‘two-sided’, ‘less’, ‘greater’}, optional. Defines the null and alternative hypotheses. Default is ‘two-sided’. \n mode: {‘auto’, ‘exact’, ‘asymp’}, optional. Defines the method used for calculating the p-value. \n by_axis (optional, int): for arrays with more than 1 dimension, the test will be run for each variable in that axis if by_axis is not None.\n \"\"\"\n if by_axis is None:\n stat, p_value = ks_2samp(data1.flatten(), data2.flatten(), alternative=alternative, mode=mode)\n return stat, np.sign(stat) * p_value\n else:\n assert data1.shape[by_axis] == data2.shape[by_axis], f\"both arrays must have the same size along axis {by_axis}\"\n stats, p_values = [], []\n for i in range(data1.shape[by_axis]):\n d1 = np.take(data1, indices=i, axis=by_axis)\n d2 = np.take(data2, indices=i, axis=by_axis)\n stat, p_value = ks_2samp(d1.flatten(), d2.flatten(), alternative=alternative, mode=mode)\n stats.append(stat) \n p_values.append(np.sign(stat) * p_value)\n return stats, p_values \n \n\ndef tscore(o): \n if o.std() == 0: return 0\n else: return np.sqrt(len(o)) * o.mean() / o.std()", "_____no_output_____" ], [ "a = np.random.normal(0.5, 1, 100)\nb = np.random.normal(0.15, .5, 50)\nplt.hist(a, 50)\nplt.hist(b, 50)\nplt.show()\nttest(a,b)", "_____no_output_____" ], [ "a = np.random.normal(0.5, 1, (100,3))\nb = np.random.normal(0.5, 1, (50,))\nkstest(a,b)", "_____no_output_____" ], [ "a = np.random.normal(0.5, 1, (100,3))\nb = np.random.normal(0.15, .5, (50,))\nkstest(a,b)", "_____no_output_____" ], [ "data1 = np.random.normal(0,1,(100, 5, 3))\ndata2 = np.random.normal(0,2,(100, 5, 3))\nkstest(data1, data2, by_axis=1)", "_____no_output_____" ], [ "a = np.random.normal(0.5, 1, 100)\nt = torch.normal(0.5, 1, (100, ))\ntscore(a), tscore(t)", "_____no_output_____" ], [ "#export\ndef ttest_tensor(a, b):\n \"differentiable pytorch function equivalent to scipy.stats.ttest_ind with equal_var=False\"\n # calculate standard errors\n se1, se2 = torch.std(a)/np.sqrt(len(a)), torch.std(b)/np.sqrt(len(b))\n # standard error on the difference between the samples\n sed = torch.sqrt(se1**2.0 + se2**2.0)\n # calculate the t statistic\n t_stat = (torch.mean(a) - torch.mean(b)) / sed\n return t_stat", "_____no_output_____" ], [ "a = torch.rand(100).requires_grad_(True) + .1\nb = torch.rand(100).requires_grad_(True)\nttest_tensor(a, b)", "_____no_output_____" ], [ "#export\ndef pcc(a, b):\n return pearsonr(a, b)[0]\n\ndef scc(a, b):\n return spearmanr(a, b)[0]\n\na = np.random.normal(0.5, 1, 100)\nb = np.random.normal(0.15, .5, 100)\npcc(a, b), scc(a, b)", "_____no_output_____" ], [ "#export\ndef remove_fn(fn, verbose=False):\n \"Removes a file (fn) if exists\"\n try: \n os.remove(fn)\n pv(f'{fn} file removed', verbose)\n except OSError: \n pv(f'{fn} does not exist', verbose)\n pass", "_____no_output_____" ], [ "#export\ndef npsave(array_fn, array, verbose=True):\n remove_fn(array_fn, verbose)\n pv(f'saving {array_fn}...', verbose)\n np.save(array_fn, array)\n pv(f'...{array_fn} saved', verbose)\n \nnp_save = npsave", "_____no_output_____" ], [ "fn = 'data/remove_fn_test.npy'\na = np.zeros(1)\nnpsave(fn, a)\ndel a\nnp.load(fn, mmap_mode='r+')\nremove_fn(fn, True)\nremove_fn(fn, True)", "data/remove_fn_test.npy does not exist\nsaving data/remove_fn_test.npy...\n...data/remove_fn_test.npy saved\ndata/remove_fn_test.npy file removed\ndata/remove_fn_test.npy does not exist\n" ], [ "#export\ndef permute_2D(array, axis=None):\n \"Permute rows or columns in an array. This can be used, for example, in feature permutation\"\n if axis == 0: return array[np.random.randn(*array.shape).argsort(axis=0), np.arange(array.shape[-1])[None, :]] \n elif axis == 1 or axis == -1: return array[np.arange(len(array))[:,None], np.random.randn(*array.shape).argsort(axis=1)] \n return array[np.random.randn(*array.shape).argsort(axis=0), np.random.randn(*array.shape).argsort(axis=1)] ", "_____no_output_____" ], [ "s = np.arange(100 * 50).reshape(100, 50) \ntest_eq(permute_2D(s, axis=0).mean(0), s.mean(0))\ntest_ne(permute_2D(s, axis=0), s)\ntest_eq(permute_2D(s, axis=1).mean(1), s.mean(1))\ntest_ne(permute_2D(s, axis=1), s)\ntest_ne(permute_2D(s), s)", "_____no_output_____" ], [ "#export\ndef random_normal():\n \"Returns a number between -1 and 1 with a normal distribution\"\n while True:\n o = np.random.normal(loc=0., scale=1/3)\n if abs(o) <= 1: break\n return o\n\ndef random_half_normal():\n \"Returns a number between 0 and 1 with a half-normal distribution\"\n while True:\n o = abs(np.random.normal(loc=0., scale=1/3))\n if o <= 1: break\n return o\n\ndef random_normal_tensor(shape=1, device=None):\n \"Returns a tensor of a predefined shape between -1 and 1 with a normal distribution\"\n return torch.empty(shape, device=device).normal_(mean=0, std=1/3).clamp_(-1, 1)\n\ndef random_half_normal_tensor(shape=1, device=None):\n \"Returns a tensor of a predefined shape between 0 and 1 with a half-normal distribution\"\n return abs(torch.empty(shape, device=device).normal_(mean=0, std=1/3)).clamp_(0, 1)", "_____no_output_____" ], [ "#export\nfrom matplotlib.backends.backend_agg import FigureCanvasAgg\n\ndef default_dpi():\n DPI = plt.gcf().get_dpi()\n plt.close()\n return int(DPI)\n\ndef get_plot_fig(size=None, dpi=default_dpi()):\n fig = plt.figure(figsize=(size / dpi, size / dpi), dpi=dpi, frameon=False) if size else plt.figure()\n ax = fig.add_axes([0,0,1,1])\n ax.spines['top'].set_visible(False)\n ax.spines['right'].set_visible(False)\n ax.spines['bottom'].set_visible(False)\n ax.spines['left'].set_visible(False)\n ax.get_xaxis().set_visible(False)\n ax.get_yaxis().set_visible(False)\n config = plt.gcf()\n plt.close('all')\n return config\n\ndef fig2buf(fig):\n canvas = FigureCanvasAgg(fig)\n fig.canvas.draw()\n return np.asarray(canvas.buffer_rgba())[..., :3]", "_____no_output_____" ], [ "default_dpi()", "_____no_output_____" ], [ "#export\ndef plot_scatter(x, y, deg=1):\n linreg = linregress(x, y)\n plt.scatter(x, y, label=f'R2:{linreg.rvalue:.2f}', color='lime', edgecolor='black', alpha=.5)\n plt.plot(np.unique(x), np.poly1d(np.polyfit(x, y, deg))(np.unique(x)), color='r')\n plt.legend(loc='best')\n plt.show()", "_____no_output_____" ], [ "a = np.random.rand(100)\nb = np.random.rand(100)**2\nplot_scatter(a, b)", "_____no_output_____" ], [ "#export\ndef get_idxs(o, aList): return array([o.tolist().index(v) for v in aList])", "_____no_output_____" ], [ "a = random_shuffle(np.arange(100, 200))\nb = np.random.choice(a, 10, False)\nidxs = get_idxs(a, b)\ntest_eq(a[idxs], b)", "_____no_output_____" ], [ "#export\ndef apply_cmap(o, cmap):\n o = toarray(o)\n out = plt.get_cmap(cmap)(o)[..., :3]\n out = tensor(out).squeeze(1)\n return out.permute(0, 3, 1, 2)", "_____no_output_____" ], [ "a = np.random.rand(16, 1, 40, 50)\ns = L(a.shape)\ns[1] = 3\ntest_eq(L(apply_cmap(a, 'viridis').shape), s)\n\ns[0] = 1\na = np.random.rand(1, 40, 50)\ntest_eq(L(apply_cmap(a, 'viridis').shape), s)", "_____no_output_____" ], [ "#export\ndef torch_tile(a, n_tile, dim=0):\n if ismin_torch(\"1.10\") and dim == 0:\n if isinstance(n_tile, tuple): \n return torch.tile(a, n_tile)\n return torch.tile(a, (n_tile,))\n init_dim = a.size(dim)\n repeat_idx = [1] * a.dim()\n repeat_idx[dim] = n_tile\n a = a.repeat(*(repeat_idx))\n order_index = torch.cat([init_dim * torch.arange(n_tile) + i for i in range(init_dim)]).to(device=a.device)\n return torch.index_select(a, dim, order_index)", "_____no_output_____" ], [ "test_eq(torch_tile(torch.arange(2), 3), tensor([0, 1, 0, 1, 0, 1]))", "_____no_output_____" ], [ "#export\ndef to_tsfresh_df(ts):\n r\"\"\"Prepares a time series (Tensor/ np.ndarray) to be used as a tsfresh dataset to allow feature extraction\"\"\"\n ts = to3d(ts)\n if isinstance(ts, np.ndarray):\n ids = np.repeat(np.arange(len(ts)), ts.shape[-1]).reshape(-1,1)\n joint_ts = ts.transpose(0,2,1).reshape(-1, ts.shape[1])\n cols = ['id'] + np.arange(ts.shape[1]).tolist()\n df = pd.DataFrame(np.concatenate([ids, joint_ts], axis=1), columns=cols)\n elif isinstance(ts, torch.Tensor):\n ids = torch_tile(torch.arange(len(ts)), ts.shape[-1]).reshape(-1,1)\n joint_ts = ts.transpose(1,2).reshape(-1, ts.shape[1])\n cols = ['id']+np.arange(ts.shape[1]).tolist()\n df = pd.DataFrame(torch.cat([ids, joint_ts], dim=1).numpy(), columns=cols)\n df['id'] = df['id'].astype(int)\n df.reset_index(drop=True, inplace=True)\n return df", "_____no_output_____" ], [ "ts = torch.rand(16, 3, 20)\na = to_tsfresh_df(ts)\nts = ts.numpy()\nb = to_tsfresh_df(ts)", "_____no_output_____" ], [ "#export\ndef pcorr(a, b): \n return pearsonr(a, b)\n\ndef scorr(a, b): \n corr = spearmanr(a, b)\n return corr[0], corr[1]", "_____no_output_____" ], [ "#export\ndef torch_diff(t, lag=1, pad=True, append=0):\n import torch.nn.functional as F\n diff = t[..., lag:] - t[..., :-lag]\n if pad: \n return F.pad(diff, (lag, append))\n else: \n return diff", "_____no_output_____" ], [ "t = torch.arange(24).reshape(2,3,4)\ntest_eq(torch_diff(t, 1)[..., 1:].float().mean(), 1.)\ntest_eq(torch_diff(t, 2)[..., 2:].float().mean(), 2.)", "_____no_output_____" ], [ "#export\ndef get_outliers_IQR(o, axis=None, quantile_range=(25.0, 75.0)):\n if isinstance(o, torch.Tensor):\n Q1 = torch.nanquantile(o, quantile_range[0]/100, axis=axis, keepdims=axis is not None)\n Q3 = torch.nanquantile(o, quantile_range[1]/100, axis=axis, keepdims=axis is not None)\n else:\n Q1 = np.nanpercentile(o, quantile_range[0], axis=axis, keepdims=axis is not None)\n Q3 = np.nanpercentile(o, quantile_range[1], axis=axis, keepdims=axis is not None)\n IQR = Q3 - Q1\n return Q1 - 1.5 * IQR, Q3 + 1.5 * IQR\n\ndef clip_outliers(o, axis=None):\n min_outliers, max_outliers = get_outliers_IQR(o, axis=axis)\n if isinstance(o, (np.ndarray, pd.core.series.Series)):\n return np.clip(o, min_outliers, max_outliers)\n elif isinstance(o, torch.Tensor):\n return torch.clamp(o, min_outliers, max_outliers)\n\ndef get_percentile(o, percentile, axis=None):\n if isinstance(o, torch.Tensor): \n return torch.nanquantile(o, percentile/100, axis=axis, keepdims=axis is not None)\n else: \n return np.nanpercentile(o, percentile, axis=axis, keepdims=axis is not None)\n\ndef torch_clamp(o, min=None, max=None):\n r\"\"\"Clamp torch.Tensor using 1 or multiple dimensions\"\"\"\n if min is not None: o = torch.max(o, min)\n if max is not None: o = torch.min(o, max)\n return o", "_____no_output_____" ], [ "t = torch.randn(2,3,100)\ntest_eq(type(get_outliers_IQR(t, -1)[0]), torch.Tensor)\na = t.numpy()\ntest_eq(type(get_outliers_IQR(a, -1)[0]), np.ndarray)\ntest_close(get_percentile(t, 25).numpy(), get_percentile(a, 25))", "_____no_output_____" ], [ "#export\ndef get_robustscale_params(o, by_var=True, percentiles=(25, 75), eps=1e-6):\n assert o.ndim == 3\n if by_var: \n median = np.nanpercentile(o, 50, axis=(0,2), keepdims=True)\n Q1 = np.nanpercentile(o, percentiles[0], axis=(0,2), keepdims=True)\n Q3 = np.nanpercentile(o, percentiles[1], axis=(0,2), keepdims=True)\n IQR = Q3 - Q1\n else:\n median = np.nanpercentile(o, .50)\n Q1 = np.nanpercentile(o, percentiles[0])\n Q3 = np.nanpercentile(o, percentiles[1])\n IQR = Q3 - Q1\n if eps is not None: IQR = np.maximum(IQR, eps)\n return median, IQR", "_____no_output_____" ], [ "a = np.random.rand(16, 3, 100)\na[a>.8] = np.nan\nmedian, IQR = get_robustscale_params(a, by_var=True, percentiles=(25, 75))\na_scaled = (a - median) / IQR\ntest_eq(a.shape, a_scaled.shape)\ntest_eq(np.isnan(median).sum(),0)\ntest_eq(np.isnan(IQR).sum(),0)\ntest_eq(np.isnan(a), np.isnan(a_scaled))", "_____no_output_____" ], [ "#export\ndef torch_slice_by_dim(t, index, dim=-1, **kwargs):\n if not isinstance(index, torch.Tensor): index = torch.Tensor(index)\n assert t.ndim == index.ndim, \"t and index must have the same ndim\"\n index = index.long()\n return torch.gather(t, dim, index, **kwargs)", "_____no_output_____" ], [ "t = torch.rand(5, 3)\nindex = torch.randint(0, 3, (5, 1))\n# index = [[0, 2], [0, 1], [1, 2], [0, 2], [0, 1]]\ntorch_slice_by_dim(t, index)", "_____no_output_____" ], [ "#export\ndef torch_nanmean(o, dim=None, keepdim=False):\n \"\"\"There's currently no torch.nanmean function\"\"\"\n mask = torch.isnan(o)\n if mask.any():\n output = torch.from_numpy(np.asarray(np.nanmean(o.cpu().numpy(), axis=dim, keepdims=keepdim))).to(o.device)\n if output.shape == mask.shape:\n output[mask] = 0\n return output\n else:\n return torch.mean(o, dim=dim, keepdim=keepdim) if dim is not None else torch.mean(o)\n\n\ndef torch_nanstd(o, dim=None, keepdim=False):\n \"\"\"There's currently no torch.nanstd function\"\"\"\n mask = torch.isnan(o)\n if mask.any():\n output = torch.from_numpy(np.asarray(np.nanstd(o.cpu().numpy(), axis=dim, keepdims=keepdim))).to(o.device)\n if output.shape == mask.shape:\n output[mask] = 1\n return output\n else:\n return torch.std(o, dim=dim, keepdim=keepdim) if dim is not None else torch.std(o)", "_____no_output_____" ], [ "t = torch.rand(1000)\nt[:100] = float('nan')\nassert torch_nanmean(t).item() > 0", "_____no_output_____" ], [ "#export\ndef concat(*ls, dim=0):\n \"Concatenate tensors, arrays, lists, or tuples by a dimension\"\n if not len(ls): return []\n it = ls[0]\n if isinstance(it, torch.Tensor): return torch.cat(ls, dim=dim)\n elif isinstance(it, np.ndarray): return np.concatenate(ls, axis=dim)\n else:\n res = np.concatenate(ls, axis=dim).tolist()\n return retain_type(res, typ=type(it))", "_____no_output_____" ], [ "#export\ndef reduce_memory_usage(df):\n \n start_memory = df.memory_usage().sum() / 1024**2\n print(f\"Memory usage of dataframe is {start_memory} MB\")\n \n for col in df.columns:\n col_type = df[col].dtype\n \n if col_type != 'object':\n c_min = df[col].min()\n c_max = df[col].max()\n \n if str(col_type)[:3] == 'int':\n if c_min > np.iinfo(np.int8).min and c_max < np.iinfo(np.int8).max:\n df[col] = df[col].astype(np.int8)\n elif c_min > np.iinfo(np.int16).min and c_max < np.iinfo(np.int16).max:\n df[col] = df[col].astype(np.int16)\n elif c_min > np.iinfo(np.int32).min and c_max < np.iinfo(np.int32).max:\n df[col] = df[col].astype(np.int32)\n elif c_min > np.iinfo(np.int64).min and c_max < np.iinfo(np.int64).max:\n df[col] = df[col].astype(np.int64)\n \n else:\n if c_min > np.finfo(np.float16).min and c_max < np.finfo(np.float16).max:\n df[col] = df[col].astype(np.float16)\n elif c_min > np.finfo(np.float32).min and c_max < np.finfo(np.float32).max:\n df[col] = df[col].astype(np.float32)\n else:\n pass\n else:\n df[col] = df[col].astype('category')\n \n end_memory = df.memory_usage().sum() / 1024**2\n print(f\"Memory usage of dataframe after reduction {end_memory} MB\")\n print(f\"Reduced by {100 * (start_memory - end_memory) / start_memory} % \")\n return df", "_____no_output_____" ], [ "#export\ndef cls_name(o): return o.__class__.__name__", "_____no_output_____" ], [ "test_eq(cls_name(timer), 'Timer')", "_____no_output_____" ], [ "#export\ndef roll2d(o, roll1: Union[None, list, int] = None, roll2: Union[None, list, int] = None):\n \"\"\"Rolls a 2D object on the indicated axis\n This solution is based on https://stackoverflow.com/questions/20360675/roll-rows-of-a-matrix-independently\n \"\"\"\n \n assert o.ndim == 2, \"roll2D can only be applied to 2d objects\"\n axis1, axis2 = np.ogrid[:o.shape[0], :o.shape[1]]\n if roll1 is not None:\n if isinstance(roll1, int): axis1 = axis1 - np.array(roll1).reshape(1,1)\n else: axis1 = np.array(roll1).reshape(o.shape[0],1)\n if roll2 is not None:\n if isinstance(roll2, int): axis2 = axis2 - np.array(roll2).reshape(1,1)\n else: axis2 = np.array(roll2).reshape(1,o.shape[1])\n return o[axis1, axis2]\n\n\ndef roll3d(o, roll1: Union[None, list, int] = None, roll2: Union[None, list, int] = None, roll3: Union[None, list, int] = None):\n \"\"\"Rolls a 3D object on the indicated axis\n This solution is based on https://stackoverflow.com/questions/20360675/roll-rows-of-a-matrix-independently\n \"\"\"\n \n assert o.ndim == 3, \"roll3D can only be applied to 3d objects\"\n axis1, axis2, axis3 = np.ogrid[:o.shape[0], :o.shape[1], :o.shape[2]]\n if roll1 is not None:\n if isinstance(roll1, int): axis1 = axis1 - np.array(roll1).reshape(1,1,1)\n else: axis1 = np.array(roll1).reshape(o.shape[0],1,1)\n if roll2 is not None:\n if isinstance(roll2, int): axis2 = axis2 - np.array(roll2).reshape(1,1,1)\n else: axis2 = np.array(roll2).reshape(1,o.shape[1],1)\n if roll3 is not None:\n if isinstance(roll3, int): axis3 = axis3 - np.array(roll3).reshape(1,1,1)\n else: axis3 = np.array(roll3).reshape(1,1,o.shape[2])\n return o[axis1, axis2, axis3]\n\n\ndef random_roll2d(o, axis=(), replace=False):\n \"\"\"Rolls a 2D object on the indicated axis\n This solution is based on https://stackoverflow.com/questions/20360675/roll-rows-of-a-matrix-independently\n \"\"\"\n \n assert o.ndim == 2, \"roll2D can only be applied to 2d objects\"\n axis1, axis2 = np.ogrid[:o.shape[0], :o.shape[1]]\n if 0 in axis:\n axis1 = np.random.choice(np.arange(o.shape[0]), o.shape[0], replace).reshape(-1, 1)\n if 1 in axis:\n axis2 = np.random.choice(np.arange(o.shape[1]), o.shape[1], replace).reshape(1, -1)\n return o[axis1, axis2]\n\n\ndef random_roll3d(o, axis=(), replace=False):\n \"\"\"Randomly rolls a 3D object along the indicated axes\n This solution is based on https://stackoverflow.com/questions/20360675/roll-rows-of-a-matrix-independently\n \"\"\"\n \n assert o.ndim == 3, \"random_roll3d can only be applied to 3d objects\"\n axis1, axis2, axis3 = np.ogrid[:o.shape[0], :o.shape[1], :o.shape[2]]\n if 0 in axis:\n axis1 = np.random.choice(np.arange(o.shape[0]), o.shape[0], replace).reshape(-1, 1, 1)\n if 1 in axis:\n axis2 = np.random.choice(np.arange(o.shape[1]), o.shape[1], replace).reshape(1, -1, 1)\n if 2 in axis:\n axis3 = np.random.choice(np.arange(o.shape[2]), o.shape[2], replace).reshape(1, 1, -1)\n return o[axis1, axis2, axis3]\n\ndef rotate_axis0(o, steps=1):\n return o[np.arange(o.shape[0]) - steps]\n\ndef rotate_axis1(o, steps=1):\n return o[:, np.arange(o.shape[1]) - steps]\n\ndef rotate_axis2(o, steps=1):\n return o[:, :, np.arange(o.shape[2]) - steps]", "_____no_output_____" ], [ "a = np.tile(np.arange(10), 3).reshape(3, 10) * np.array([1, 10, 100]).reshape(-1, 1)\na", "_____no_output_____" ], [ "roll2d(a, roll1=[2, 1, 0])", "_____no_output_____" ], [ "roll2d(a, roll2=3)", "_____no_output_____" ], [ "o = torch.arange(24).reshape(2,3,4)\ntest_eq(rotate_axis0(o)[1], o[0])\ntest_eq(rotate_axis1(o)[:,1], o[:,0])\ntest_eq(rotate_axis2(o)[...,1], o[...,0])", "_____no_output_____" ], [ "#export\ndef chunks_calculator(shape, dtype='float32', n_bytes=1024**3):\n \"\"\"Function to calculate chunks for a given size of n_bytes (default = 1024**3 == 1GB). \n It guarantees > 50% of the chunk will be filled\"\"\"\n \n X = np.random.rand(1, *shape[1:]).astype(dtype)\n byts = get_size(X, return_str=False)\n n = n_bytes // byts\n if shape[0] / n <= 1: return False\n remainder = shape[0] % n\n if remainder / n < .5: \n n_chunks = shape[0] // n\n n += np.ceil(remainder / n_chunks).astype(int)\n return (n, -1, -1)", "_____no_output_____" ], [ "shape = (1_000, 10, 1000)\ndtype = 'float32'\ntest_eq(chunks_calculator(shape, dtype), False)\n\nshape = (54684, 10, 1000)\ndtype = 'float32'\ntest_eq(chunks_calculator(shape, dtype), (27342, -1, -1))", "_____no_output_____" ], [ "#export\ndef is_memory_shared(a, b):\n r\"\"\"Test function to check if 2 array-like object share memory. \n Be careful because it changes their values!!!)\"\"\"\n \n try: \n a[:] = 1\n except: \n try: \n b[:] = 1\n except: \n print('unknown')\n return \n return torch.equal(tensor(a), tensor(b))", "_____no_output_____" ], [ "a = np.random.rand(2,3,4)\nt1 = torch.from_numpy(a)\ntest_eq(is_memory_shared(a, t1), True)\na = np.random.rand(2,3,4)\nt2 = torch.as_tensor(a)\ntest_eq(is_memory_shared(a, t2), True)\na = np.random.rand(2,3,4)\nt3 = torch.tensor(a)\ntest_eq(is_memory_shared(a, t3), False)", "_____no_output_____" ], [ "#export\ndef assign_in_chunks(a, b, chunksize='auto', inplace=True, verbose=True):\n \"\"\"Assigns values in b to an array-like object a using chunks to avoid memory overload.\n \n The resulting a retains it's dtype and share it's memory.\n a: array-like object\n b: may be an integer, float, str, 'rand' (for random data), or another array like object.\n chunksize: is the size of chunks. If 'auto' chunks will have around 1GB each. \n \"\"\"\n \n if b != 'rand' and not isinstance(b, (Iterable, Generator)):\n a[:] = b\n else:\n shape = a.shape\n dtype = a.dtype\n if chunksize == \"auto\": \n chunksize = chunks_calculator(shape, dtype)\n chunksize = shape[0] if not chunksize else chunksize[0]\n for i in progress_bar(range((shape[0] - 1) // chunksize + 1), display=verbose, leave=False):\n start, end = i * chunksize, min(shape[0], (i + 1) * chunksize)\n if start >= shape[0]: break\n if b == 'rand': \n a[start:end] = np.random.rand(end - start, *shape[1:])\n else: \n a[start:end] = b[start:end]\n if not inplace: return a", "_____no_output_____" ], [ "a = np.random.rand(10,3,4).astype('float32')\na_dtype = a.dtype\na_id = id(a)\nb = np.random.rand(10,3,4).astype('float64')\nassign_in_chunks(a, b, chunksize=2, inplace=True, verbose=True)\ntest_close(a, b)\ntest_eq(a.dtype, a_dtype)\ntest_eq(id(a), a_id)\n\na = np.random.rand(10,3,4).astype('float32')\na_dtype = a.dtype\na_id = id(a)\nb = 1\nassign_in_chunks(a, b, chunksize=2, inplace=True, verbose=True)\ntest_eq(a, np.ones_like(a).astype(a.dtype))\ntest_eq(a.dtype, a_dtype)\ntest_eq(id(a), a_id)\n\na = np.random.rand(10,3,4).astype('float32')\na_dtype = a.dtype\na_id = id(a)\nb = 0.5\nassign_in_chunks(a, b, chunksize=2, inplace=True, verbose=True)\ntest_eq(a.dtype, a_dtype)\ntest_eq(id(a), a_id)\n\na = np.random.rand(10,3,4).astype('float32')\na_dtype = a.dtype\na_id = id(a)\nb = 'rand'\nassign_in_chunks(a, b, chunksize=2, inplace=True, verbose=True)\ntest_eq(a.dtype, a_dtype)\ntest_eq(id(a), a_id)", "/Users/nacho/opt/anaconda3/envs/py37torch110/lib/python3.7/site-packages/ipykernel_launcher.py:11: FutureWarning: elementwise comparison failed; returning scalar instead, but in the future will perform elementwise comparison\n # This is added back by InteractiveShellApp.init_path()\n" ], [ "a = np.random.rand(10,3,4).astype('float32')\nb = np.random.rand(10,3,4).astype('float64')\nc = assign_in_chunks(a, b, chunksize=2, inplace=False, verbose=True)\ntest_close(c, b)\ntest_eq(a.dtype, c.dtype)\ntest_eq(is_memory_shared(a, c), True)\n\na = np.random.rand(10,3,4).astype('float32')\nb = 1\nc = assign_in_chunks(a, b, chunksize=2, inplace=False, verbose=True)\ntest_eq(a, np.ones_like(a).astype(a.dtype))\ntest_eq(a.dtype, c.dtype)\ntest_eq(is_memory_shared(a, c), True)\n\na = np.random.rand(10,3,4).astype('float32')\nb = 0.5\nc = assign_in_chunks(a, b, chunksize=2, inplace=False, verbose=True)\ntest_eq(a.dtype, c.dtype)\ntest_eq(is_memory_shared(a, c), True)\n\na = np.random.rand(10,3,4).astype('float32')\nb = 'rand'\nc = assign_in_chunks(a, b, chunksize=2, inplace=False, verbose=True)\ntest_eq(a.dtype, c.dtype)\ntest_eq(is_memory_shared(a, c), True)", "/Users/nacho/opt/anaconda3/envs/py37torch110/lib/python3.7/site-packages/ipykernel_launcher.py:11: FutureWarning: elementwise comparison failed; returning scalar instead, but in the future will perform elementwise comparison\n # This is added back by InteractiveShellApp.init_path()\n" ], [ "#export\ndef create_array(shape, fname=None, path='./data', on_disk=True, dtype='float32', mode='r+', fill_value='rand', chunksize='auto', verbose=True, **kwargs):\n \"\"\"\n mode:\n ‘r’: Open existing file for reading only.\n ‘r+’: Open existing file for reading and writing.\n ‘w+’: Create or overwrite existing file for reading and writing.\n ‘c’: Copy-on-write: assignments affect data in memory, but changes are not saved to disk. The file on disk is read-only.\n fill_value: 'rand' (for random numbers), int or float\n chunksize = 'auto' to calculate chunks of 1GB, or any integer (for a given number of samples)\n \"\"\"\n if on_disk:\n assert fname is not None, 'you must provide a fname (filename)'\n path = Path(path)\n if not fname.endswith('npy'): fname = f'{fname}.npy'\n filename = path/fname\n filename.parent.mkdir(parents=True, exist_ok=True)\n # Save a small empty array\n _temp_fn = path/'temp_X.npy'\n np.save(_temp_fn, np.empty(0))\n # Create & save file\n arr = np.memmap(_temp_fn, dtype=dtype, mode='w+', shape=shape, **kwargs)\n np.save(filename, arr)\n del arr\n os.remove(_temp_fn)\n # Open file in selected mode\n arr = np.load(filename, mmap_mode=mode)\n else:\n arr = np.empty(shape, dtype=dtype, **kwargs)\n if fill_value != 0:\n assign_in_chunks(arr, fill_value, chunksize=chunksize, inplace=True, verbose=verbose)\n return arr\n\ncreate_empty_array = partial(create_array, fill_value=0)", "_____no_output_____" ], [ "fname = 'X_on_disk'\nshape = (100, 10, 10)\nX = create_array(shape, fname, on_disk=True, mode='r+')\ntest_ne(abs(X).sum(), 0)\nos.remove(X.filename)\ndel X", "_____no_output_____" ], [ "fname = 'X_on_disk'\nshape = (100, 10, 10)\nX = create_empty_array(shape, fname, on_disk=True, mode='r+')\ntest_eq(abs(X).sum(), 0)\n\nchunksize = 10\npbar = progress_bar(range(math.ceil(len(X) / chunksize)), leave=False)\nstart = 0\nfor i in pbar: \n end = min(start + chunksize, len(X))\n partial_data = np.random.rand(end - start, X.shape[1] , X.shape[2])\n X[start:end] = partial_data\n start = end\n del partial_data\n gc.collect()\nfilename = X.filename\ndel X\nX = np.load(filename, mmap_mode='r+')\ntest_eq((X == 0).sum(), 0)\ntest_eq(X.shape, shape)\nos.remove(X.filename)\ndel X", "_____no_output_____" ], [ "#export\nimport gzip\n\ndef np_save_compressed(arr, fname=None, path='./data', verbose=False, **kwargs):\n assert fname is not None, 'you must provide a fname (filename)'\n if fname.endswith('npy'): fname = f'{fname}.gz'\n elif not fname.endswith('npy.gz'): fname = f'{fname}.npy.gz'\n filename = Path(path)/fname\n filename.parent.mkdir(parents=True, exist_ok=True)\n f = gzip.GzipFile(filename, 'w', **kwargs)\n np.save(file=f, arr=arr)\n f.close()\n pv(f'array saved to {filename}', verbose)\n \ndef np_load_compressed(fname=None, path='./data', **kwargs):\n assert fname is not None, 'you must provide a fname (filename)'\n if fname.endswith('npy'): fname = f'{fname}.gz'\n elif not fname.endswith('npy.gz'): fname = f'{fname}.npy.gz'\n filename = Path(path)/fname\n f = gzip.GzipFile(filename, 'r', **kwargs)\n arr = np.load(f)\n f.close()\n return arr", "_____no_output_____" ], [ "X1 = np.random.rand(10)\nnp_save_compressed(X1, 'X_comp', path='./data')\nX2 = np_load_compressed('X_comp')\ntest_eq(X1, X2)", "_____no_output_____" ], [ "#export\ndef np2memmap(arr, fname=None, path='./data', dtype='float32', mode='c', **kwargs):\n \"\"\" Function that turns an ndarray into a memmap ndarray\n mode:\n ‘r’: Open existing file for reading only.\n ‘r+’: Open existing file for reading and writing.\n ‘w+’: Create or overwrite existing file for reading and writing.\n ‘c’: Copy-on-write: assignments affect data in memory, but changes are not saved to disk. The file on disk is read-only.\n \"\"\"\n assert fname is not None, 'you must provide a fname (filename)'\n if not fname.endswith('npy'): fname = f'{fname}.npy'\n filename = Path(path)/fname\n filename.parent.mkdir(parents=True, exist_ok=True)\n # Save file\n np.save(filename, arr)\n # Open file in selected mode\n arr = np.load(filename, mmap_mode=mode)\n return arr", "_____no_output_____" ], [ "X1 = np.random.rand(10)\nX2 = np2memmap(X1, 'X1_test')\ntest_eq(X1, X2)\ntest_ne(type(X1), type(X2))", "_____no_output_____" ], [ "#export \ndef torch_mean_groupby(o, idxs):\n \"\"\"Computes torch mean along axis 0 grouped by the idxs. \n Need to ensure that idxs have the same order as o\"\"\"\n if is_listy(idxs[0]): idxs = flatten_list(idxs)\n flattened_idxs = torch.tensor(idxs)\n idxs, vals = torch.unique(flattened_idxs, return_counts=True)\n vs = torch.split_with_sizes(o, tuple(vals))\n return torch.cat([v.mean(0).unsqueeze(0) for k,v in zip(idxs, vs)])", "_____no_output_____" ], [ "o = torch.arange(6*2*3).reshape(6, 2, 3).float()\nidxs = np.array([[0,1,2,3], [2,3]], dtype=object)\noutput = torch_mean_groupby(o, idxs)\ntest_eq(o[:2], output[:2])\ntest_eq(o[2:4].mean(0), output[2])\ntest_eq(o[4:6].mean(0), output[3])", "_____no_output_____" ], [ "#export\ndef torch_flip(t, dims=-1):\n if dims == -1: return t[..., np.arange(t.shape[dims])[::-1].copy()]\n elif dims == 0: return t[np.arange(t.shape[dims])[::-1].copy()]\n elif dims == 1: return t[:, np.arange(t.shape[dims])[::-1].copy()]\n elif dims == 2: return t[:, :, np.arange(t.shape[dims])[::-1].copy()]", "_____no_output_____" ], [ "t = torch.randn(2, 3, 4)\ntest_eq(torch.flip(t, (2,)), torch_flip(t, dims=-1))", "_____no_output_____" ], [ "#export \ndef torch_nan_to_num(o, num=0, inplace=False):\n if ismin_torch(\"1.8\") and not inplace: \n return torch.nan_to_num(o, num)\n mask = torch.isnan(o)\n return torch_masked_to_num(o, mask, num=num, inplace=inplace)\n\ndef torch_masked_to_num(o, mask, num=0, inplace=False):\n if inplace: \n o[:] = o.masked_fill(mask, num)\n else: \n return o.masked_fill(mask, num)", "_____no_output_____" ], [ "x = torch.rand(2, 4, 6)\nx[:, :3][x[:, :3] < .5] = np.nan\nnan_values = torch.isnan(x).sum()\ny = torch_nan_to_num(x[:, :3], inplace=False)\ntest_eq(torch.isnan(y).sum(), 0)\ntest_eq(torch.isnan(x).sum(), nan_values)\ntorch_nan_to_num(x[:, :3], inplace=True)\ntest_eq(torch.isnan(x).sum(), 0)", "_____no_output_____" ], [ "x = torch.rand(2, 4, 6)\nmask = x[:, :3] > .5\nx[:, :3] = torch_masked_to_num(x[:, :3], mask, num=0, inplace=False)\ntest_eq(x[:, :3][mask].sum(), 0)", "_____no_output_____" ], [ "x = torch.rand(2, 4, 6)\nmask = x[:, :3] > .5\ntorch_masked_to_num(x[:, :3], mask, num=0, inplace=True)\ntest_eq(x[:, :3][mask].sum(), 0)", "_____no_output_____" ], [ "#export\ndef mpl_trend(x, y, deg=1): \n return np.poly1d(np.polyfit(x, y, deg))(x)", "_____no_output_____" ], [ "x = np.sort(np.random.randint(0, 100, 100)/10)\ny = np.random.rand(100) + np.linspace(0, 10, 100)\ntrend = mpl_trend(x, y)\nplt.scatter(x, y)\nplt.plot(x, trend, 'r')\nplt.show()", "_____no_output_____" ], [ "#export\ndef int2digits(o, n_digits=None, normalize=True):\n if n_digits is not None:\n iterable = '0' * (n_digits - len(str(abs(o)))) + str(abs(o))\n else:\n iterable = str(abs(o))\n sign = np.sign(o)\n digits = np.array([sign * int(d) for d in iterable])\n if normalize:\n digits = digits / 10\n return digits\n\n\ndef array2digits(o, n_digits=None, normalize=True):\n output = np.array(list(map(partial(int2digits, n_digits=n_digits), o)))\n if normalize:\n output = output / 10\n return output", "_____no_output_____" ], [ "o = -9645\ntest_eq(int2digits(o, 6), np.array([ 0, 0, -.9, -.6, -.4, -.5]))\n\na = np.random.randint(-1000, 1000, 10)\ntest_eq(array2digits(a,5).shape, (10,5))", "_____no_output_____" ], [ "#export\ndef sincos_encoding(seq_len, device=None, to_np=False):\n if to_np:\n sin = np.sin(np.arange(seq_len) / seq_len * 2 * np.pi)\n cos = np.cos(np.arange(seq_len) / seq_len * 2 * np.pi)\n else:\n if device is None: device = default_device()\n sin = torch.sin(torch.arange(seq_len, device=device) / seq_len * 2 * np.pi)\n cos = torch.cos(torch.arange(seq_len, device=device) / seq_len * 2 * np.pi)\n return sin, cos", "_____no_output_____" ], [ "sin, cos = sincos_encoding(100)\nplt.plot(sin.cpu().numpy())\nplt.plot(cos.cpu().numpy())\nplt.show()", "_____no_output_____" ], [ "#export\ndef linear_encoding(seq_len, device=None, to_np=False, lin_range=(-1,1)):\n if to_np:\n enc = np.linspace(lin_range[0], lin_range[1], seq_len)\n else:\n if device is None: device = default_device()\n enc = torch.linspace(lin_range[0], lin_range[1], seq_len, device=device)\n return enc", "_____no_output_____" ], [ "lin = linear_encoding(100)\nplt.plot(lin.cpu().numpy())\nplt.show()", "_____no_output_____" ], [ "#export\ndef encode_positions(pos_arr, min_val=None, max_val=None, linear=False, lin_range=(-1,1)):\n \"\"\" Encodes an array with positions using a linear or sincos methods\n \"\"\"\n \n if min_val is None:\n min_val = np.nanmin(pos_arr)\n if max_val is None:\n max_val = np.nanmax(pos_arr)\n \n if linear: \n return (((pos_arr - min_val)/(max_val - min_val)) * (lin_range[1] - lin_range[0]) + lin_range[0])\n else:\n sin = np.sin((pos_arr - min_val)/(max_val - min_val) * 2 * np.pi)\n cos = np.cos((pos_arr - min_val)/(max_val - min_val) * 2 * np.pi)\n return sin, cos", "_____no_output_____" ], [ "n_samples = 10\nlength = 500\n_a = []\nfor i in range(n_samples):\n a = np.arange(-4000, 4000, 10)\n mask = np.random.rand(len(a)) > .5\n a = a[mask]\n a = np.concatenate([a, np.array([np.nan] * (length - len(a)))])\n _a.append(a.reshape(-1,1))\na = np.concatenate(_a, -1).transpose(1,0)\nsin, cos = encode_positions(a, linear=False)\ntest_eq(a.shape, (n_samples, length))\ntest_eq(sin.shape, (n_samples, length))\ntest_eq(cos.shape, (n_samples, length))\nplt.plot(sin.T)\nplt.plot(cos.T)\nplt.xlim(0, 500)\nplt.show()", "_____no_output_____" ], [ "n_samples = 10\nlength = 500\n_a = []\nfor i in range(n_samples):\n a = np.arange(-4000, 4000, 10)\n mask = np.random.rand(len(a)) > .5\n a = a[mask]\n a = np.concatenate([a, np.array([np.nan] * (length - len(a)))])\n _a.append(a.reshape(-1,1))\na = np.concatenate(_a, -1).transpose(1,0)\nlin = encode_positions(a, linear=True)\ntest_eq(a.shape, (n_samples, length))\ntest_eq(lin.shape, (n_samples, length))\nplt.plot(lin.T)\nplt.xlim(0, 500)\nplt.show()", "_____no_output_____" ], [ "#export\ndef sort_generator(generator, bs):\n g = list(generator)\n for i in range(len(g)//bs + 1): g[bs*i:bs*(i+1)] = np.sort(g[bs*i:bs*(i+1)])\n return (i for i in g)", "_____no_output_____" ], [ "generator = (i for i in np.random.permutation(np.arange(1000000)).tolist())\nl = list(sort_generator(generator, 512))\ntest_eq(l[:512], sorted(l[:512]))", "_____no_output_____" ], [ "#export\ndef get_subset_dict(d, keys):\n return dict((k,d[k]) for k in listify(keys) if k in d)", "_____no_output_____" ], [ "keys = string.ascii_lowercase\nvalues = np.arange(len(keys))\nd = {k:v for k,v in zip(keys,values)}\ntest_eq(get_subset_dict(d, ['a', 'k', 'j', 'e']), {'a': 0, 'k': 10, 'j': 9, 'e': 4})", "_____no_output_____" ], [ "#export\ndef create_dir(directory, verbose=True): \n if not is_listy(directory): directory = [directory]\n for d in directory:\n d = Path(d)\n if d.exists():\n if verbose: print(f\"{d} directory already exists.\")\n else: \n d.mkdir(parents=True, exist_ok=True)\n assert d.exists(), f\"a problem has occurred while creating {d}\"\n if verbose: print(f\"{d} directory created.\")\n\n\ndef remove_dir(directory, verbose=True):\n import shutil\n if not is_listy(directory): directory = [directory]\n for d in directory:\n d = Path(d)\n if d.is_file(): d = d.parent\n if not d.exists():\n if verbose: print(f\"{d} directory doesn't exist.\")\n else:\n shutil.rmtree(d)\n assert not d.exists(), f\"a problem has occurred while deleting {d}\"\n if verbose: print(f\"{d} directory removed.\")", "_____no_output_____" ], [ "path = \"wandb3/wandb2/wandb\"\ncreate_dir(path)\nassert Path(path).exists()\n\npaths = [\"wandb3/wandb2/wandb\", \"wandb3/wandb2\", \"wandb\"]\nremove_dir(paths)\nfor p in paths: \n assert not Path(p).exists()\n\npath = \"wandb3\"\nassert Path(path).exists()\nremove_dir(path)\nassert not Path(path).exists()", "wandb3/wandb2/wandb directory created.\nwandb3/wandb2/wandb directory removed.\nwandb3/wandb2 directory removed.\nwandb directory doesn't exist.\nwandb3 directory removed.\n" ], [ "create_dir('./test')", "test directory created.\n" ], [ "%%file ./test/mod_dev.py\na = 5\ndef fn(b): return a + b", "Writing ./test/mod_dev.py\n" ], [ "fname = \"./test/mod_dev.py\"\nwhile True: \n if fname[0] in \"/ .\": fname = fname.split(fname[0], 1)[1]\n else: break\nif '/' in fname and fname.rsplit('/', 1)[0] not in sys.path: sys.path.append(fname.rsplit('/', 1)[0])\nmod = import_file_as_module(fname)\ntest_eq(mod.fn(3), 8)\nsys.path = sys.path[:-1]\nremove_dir('./test/')", "test directory removed.\n" ], [ "#export\nclass named_partial(object):\n \"\"\"Create a partial function with a __name__\"\"\"\n \n def __init__(self, name, func, *args, **kwargs):\n self._func = partial(func, *args, **kwargs)\n self.__name__ = name\n def __call__(self, *args, **kwargs):\n return self._func(*args, **kwargs)\n def __repr__(self):\n return self.__name__", "_____no_output_____" ], [ "def add_1(x, add=1): return x+add\ntest_eq(add_1(1), 2)\nadd_2 = partial(add_1, add=2)\ntest_eq(add_2(2), 4)\ntest_ne(str(add_2), \"add_2\")\nadd_2 = named_partial('add_2', add_1, add=2)\ntest_eq(add_2(2), 4)\ntest_eq(str(add_2), \"add_2\")\n\nclass _A():\n def __init__(self, add=1): self.add = add\n def __call__(self, x): return x + self.add\n \ntest_eq(_A()(1), 2)\n_A2 = partial(_A, add=2)\ntest_eq(_A2()(1), 3)\ntest_ne(str(_A2), '_A2')\n_A2 = named_partial('_A2', _A, add=2)\ntest_eq(_A2()(1), 3)\ntest_eq(str(_A2), '_A2')", "_____no_output_____" ], [ "# export\ndef yaml2dict(fname):\n import yaml\n with maybe_open(fname, 'r') as f:\n dictionary = yaml.safe_load(f)\n return AttrDict(dictionary)", "_____no_output_____" ], [ "%%file sweep_config.yaml\n\nprogram: wandb_scripts/train_script.py # (required) Path to training script.\nmethod: bayes # (required) Specify the search strategy: grid, random or bayes\nparameters: # (required) Specify parameters bounds to search.\n bs:\n values: [32, 64, 128]\n depth:\n values: [3, 6, 9, 12]\n fc_dropout:\n distribution: uniform\n min: 0.\n max: 0.5\n lr_max:\n values: [0.001, 0.003, 0.01, 0.03, 0.1]\n n_epoch:\n values: [10, 15, 20]\n nb_filters:\n values: [32, 64, 128]\nname: LSST_sweep_01\nmetric: \n name: accuracy # This must match one of the metrics in the training script\n goal: maximize\nearly_terminate: \n type: hyperband\n min_iter: 3\nproject: LSST_wandb_hpo", "Writing sweep_config.yaml\n" ], [ "fname = \"sweep_config.yaml\"\nsweep_config = yaml2dict(fname)\nprint(sweep_config)\ntest_eq(sweep_config.method, 'bayes')\ntest_eq(sweep_config['metric'], {'name': 'accuracy', 'goal': 'maximize'})\nos.remove(fname)", "- program: wandb_scripts/train_script.py\n- method: bayes\n- parameters: \n - bs: \n - values: \n - 32\n - 64\n - 128\n - depth: \n - values: \n - 3\n - 6\n - 9\n - 12\n - fc_dropout: \n - distribution: uniform\n - min: 0.0\n - max: 0.5\n - lr_max: \n - values: \n - 0.001\n - 0.003\n - 0.01\n - 0.03\n - 0.1\n - n_epoch: \n - values: \n - 10\n - 15\n - 20\n - nb_filters: \n - values: \n - 32\n - 64\n - 128\n- name: LSST_sweep_01\n- metric: \n - name: accuracy\n - goal: maximize\n- early_terminate: \n - type: hyperband\n - min_iter: 3\n- project: LSST_wandb_hpo\n" ], [ "#export\ndef str2list(o):\n if o is None: return []\n elif o is not None and not isinstance(o, (list, L)):\n if isinstance(o, pd.core.indexes.base.Index): o = o.tolist()\n else: o = [o]\n return o\n\ndef str2index(o):\n if o is None: return o\n o = str2list(o)\n if len(o) == 1: return o[0]\n return o\n\ndef get_cont_cols(df):\n return df._get_numeric_data().columns.tolist()\n\ndef get_cat_cols(df):\n cols = df.columns.tolist()\n cont_cols = df._get_numeric_data().columns.tolist()\n return [col for col in cols if col not in cont_cols]", "_____no_output_____" ], [ "#export\nalphabet = L(list(string.ascii_lowercase))\nALPHABET = L(list(string.ascii_uppercase))", "_____no_output_____" ], [ "#export\ndef get_mapping(arr, dim=1, return_counts=False):\n maps = [L(np.unique(np.take(arr, i, dim)).tolist()) for i in range(arr.shape[dim])]\n if return_counts:\n counts = [len(m) for m in maps]\n return maps, counts\n return maps\n\ndef map_array(arr, dim=1):\n out = stack([np.unique(np.take(arr, i, dim), return_inverse=True)[1] for i in range(arr.shape[dim])])\n if dim == 1: out = out.T\n return out", "_____no_output_____" ], [ "a = np.asarray(alphabet[np.random.randint(0,15,30)]).reshape(10,3)\nb = np.asarray(ALPHABET[np.random.randint(6,10,30)]).reshape(10,3)\nx = concat(a,b,dim=1)\nmaps, counts = get_mapping(x, dim=1, return_counts=True)\nx, maps, counts", "_____no_output_____" ], [ "x = np.asarray(alphabet[np.random.randint(0,15,30)]).reshape(10,3)\nx, map_array(x), map_array(x, 1)", "_____no_output_____" ], [ "# export\ndef log_tfm(o, inplace=False):\n \"Log transforms an array-like object with positive and/or negative values\"\n if isinstance(o, torch.Tensor):\n pos_o = torch.log1p(o[o > 0])\n neg_o = -torch.log1p(torch.abs(o[o < 0]))\n else: \n pos_o = np.log1p(o[o > 0])\n neg_o = -np.log1p(np.abs(o[o < 0]))\n if inplace:\n o[o > 0] = pos_o\n o[o < 0] = neg_o\n return o\n else:\n if hasattr(o, \"clone\"): output = o.clone()\n elif hasattr(o, \"copy\"): output = o.copy()\n output[output > 0] = pos_o\n output[output < 0] = neg_o\n return output", "_____no_output_____" ], [ "arr = np.asarray([-1000, -100, -10, -1, 0, 1, 10, 100, 1000]).astype(float)\nplt.plot(arr, log_tfm(arr, False))\nplt.show()", "_____no_output_____" ], [ "t = tensor([-1000, -100, -10, -1, 0, 1, 10, 100, 1000]).float()\nplt.plot(t, log_tfm(t, False))\nplt.show()", "_____no_output_____" ], [ "#export\ndef to_sincos_time(arr, max_value):\n sin = np.sin(arr / max_value * 2 * np.pi)\n cos = np.cos(arr / max_value * 2 * np.pi)\n return sin, cos", "_____no_output_____" ], [ "arr = np.sort(np.random.rand(100) * 5)\narr_sin, arr_cos = to_sincos_time(arr, 5)\nplt.scatter(arr, arr_sin)\nplt.scatter(arr, arr_cos)\nplt.show()", "_____no_output_____" ], [ "#export\ndef plot_feature_dist(X, percentiles=[0,0.1,0.5,1,5,10,25,50,75,90,95,99,99.5,99.9,100]):\n for i in range(X.shape[1]):\n ys = []\n for p in percentiles:\n ys.append(np.percentile(X[:, i].flatten(), p))\n plt.plot(percentiles, ys)\n plt.xticks(percentiles, rotation='vertical')\n plt.grid(color='gainsboro', linewidth=.5)\n plt.title(f\"var_{i}\")\n plt.show()", "_____no_output_____" ], [ "arr = np.random.rand(10, 3, 100)\nplot_feature_dist(arr, percentiles=[0,0.1,0.5,1,5,10,25,50,75,90,95,99,99.5,99.9,100])", "_____no_output_____" ], [ "#export\ndef rolling_moving_average(o, window=2):\n if isinstance(o, torch.Tensor):\n cunsum = torch.cumsum(o, axis=-1) # nancumsum not available (can't be used with missing data!)\n lag_cunsum = torch.cat([torch.zeros((o.shape[0], o.shape[1], window), device=o.device), torch.cumsum(o[..., :-window], axis=-1)], -1)\n count = torch.clip(torch.ones_like(o).cumsum(-1), max=window)\n return (cunsum - lag_cunsum) / count\n else:\n cunsum = np.nancumsum(o, axis=-1)\n lag_cunsum = np.concatenate([np.zeros((o.shape[0], o.shape[1], window)), np.nancumsum(o[..., :-window], axis=-1)], -1)\n count = np.minimum(np.ones_like(o).cumsum(-1), window)\n return (cunsum - lag_cunsum) / count", "_____no_output_____" ], [ "a = np.arange(60).reshape(2,3,10).astype(float)\nt = torch.arange(60).reshape(2,3,10).float()\ntest_close(rolling_moving_average(a, window=3), rolling_moving_average(t, window=3).numpy())\nprint(t)\nprint(rolling_moving_average(t, window=3))", "tensor([[[ 0., 1., 2., 3., 4., 5., 6., 7., 8., 9.],\n [10., 11., 12., 13., 14., 15., 16., 17., 18., 19.],\n [20., 21., 22., 23., 24., 25., 26., 27., 28., 29.]],\n\n [[30., 31., 32., 33., 34., 35., 36., 37., 38., 39.],\n [40., 41., 42., 43., 44., 45., 46., 47., 48., 49.],\n [50., 51., 52., 53., 54., 55., 56., 57., 58., 59.]]])\ntensor([[[ 0.0000, 0.5000, 1.0000, 2.0000, 3.0000, 4.0000, 5.0000,\n 6.0000, 7.0000, 8.0000],\n [10.0000, 10.5000, 11.0000, 12.0000, 13.0000, 14.0000, 15.0000,\n 16.0000, 17.0000, 18.0000],\n [20.0000, 20.5000, 21.0000, 22.0000, 23.0000, 24.0000, 25.0000,\n 26.0000, 27.0000, 28.0000]],\n\n [[30.0000, 30.5000, 31.0000, 32.0000, 33.0000, 34.0000, 35.0000,\n 36.0000, 37.0000, 38.0000],\n [40.0000, 40.5000, 41.0000, 42.0000, 43.0000, 44.0000, 45.0000,\n 46.0000, 47.0000, 48.0000],\n [50.0000, 50.5000, 51.0000, 52.0000, 53.0000, 54.0000, 55.0000,\n 56.0000, 57.0000, 58.0000]]])\n" ], [ "#export\ndef ffill_sequence(o):\n \"\"\"Forward fills an array-like object alongside sequence dimension\"\"\"\n if isinstance(o, torch.Tensor):\n mask = torch.isnan(o)\n idx = torch.where(~mask, torch.arange(mask.shape[-1], device=o.device), 0)\n idx = torch.cummax(idx, dim=-1).values\n return o[torch.arange(o.shape[0], device=o.device)[:,None,None], torch.arange(o.shape[1], device=o.device)[None,:,None], idx]\n else:\n mask = np.isnan(o)\n idx = np.where(~mask, np.arange(mask.shape[-1]), 0)\n idx = np.maximum.accumulate(idx, axis=-1)\n return o[np.arange(o.shape[0])[:,None,None], np.arange(o.shape[1])[None,:,None], idx]\n\ndef bfill_sequence(o):\n \"\"\"Backward fills an array-like object alongside sequence dimension\"\"\"\n if isinstance(o, torch.Tensor):\n o = torch.flip(o, (-1,))\n o = ffill_sequence(o)\n return torch.flip(o, (-1,))\n else:\n o = o[..., ::-1]\n o = ffill_sequence(o)\n return o[..., ::-1]\n\ndef fbfill_sequence(o):\n \"\"\"Forward and backward fills an array-like object alongside sequence dimension\"\"\"\n o = ffill_sequence(o)\n o = bfill_sequence(o)\n return o", "_____no_output_____" ], [ "a = np.arange(80).reshape(2, 4, 10).astype(float)\nmask = np.random.rand(*a.shape)\na[mask > .8] = np.nan\nt = torch.from_numpy(a)\nt", "_____no_output_____" ], [ "# forward fill\nfilled_a = ffill_sequence(a)\nprint(filled_a)\nm = np.isnan(filled_a)\ntest_eq(filled_a[~m], ffill_sequence(t).numpy()[~m])", "[[[ 0. 1. 2. 3. 3. 5. 6. 7. 8. 9.]\n [10. 11. 12. 13. 14. 14. 16. 17. 17. 19.]\n [20. 21. 22. 23. 23. 25. 26. 26. 28. 29.]\n [30. 31. 32. 33. 34. 35. 36. 37. 38. 39.]]\n\n [[40. 41. 42. 42. 42. 45. 45. 47. 48. 48.]\n [50. 51. 52. 53. 54. 55. 56. 56. 56. 59.]\n [60. 61. 62. 63. 64. 65. 65. 67. 68. 69.]\n [70. 71. 72. 72. 74. 75. 76. 77. 77. 79.]]]\n" ], [ "# backward fill\nfilled_a = bfill_sequence(a)\nprint(filled_a)\nm = np.isnan(filled_a)\ntest_eq(filled_a[~m], bfill_sequence(t).numpy()[~m])", "[[[ 0. 1. 2. 3. 5. 5. 6. 7. 8. 9.]\n [10. 11. 12. 13. 14. 16. 16. 17. 19. 19.]\n [20. 21. 22. 23. 25. 25. 26. 28. 28. 29.]\n [30. 31. 32. 33. 34. 35. 36. 37. 38. 39.]]\n\n [[40. 41. 42. 45. 45. 45. 47. 47. 48. nan]\n [50. 51. 52. 53. 54. 55. 56. 59. 59. 59.]\n [60. 61. 62. 63. 64. 65. 67. 67. 68. 69.]\n [70. 71. 72. 74. 74. 75. 76. 77. 79. 79.]]]\n" ], [ "# forward & backward fill\nfilled_a = fbfill_sequence(a)\nprint(filled_a)\nm = np.isnan(filled_a)\ntest_eq(filled_a[~m], fbfill_sequence(t).numpy()[~m])", "[[[ 0. 1. 2. 3. 3. 5. 6. 7. 8. 9.]\n [10. 11. 12. 13. 14. 14. 16. 17. 17. 19.]\n [20. 21. 22. 23. 23. 25. 26. 26. 28. 29.]\n [30. 31. 32. 33. 34. 35. 36. 37. 38. 39.]]\n\n [[40. 41. 42. 42. 42. 45. 45. 47. 48. 48.]\n [50. 51. 52. 53. 54. 55. 56. 56. 56. 59.]\n [60. 61. 62. 63. 64. 65. 65. 67. 68. 69.]\n [70. 71. 72. 72. 74. 75. 76. 77. 77. 79.]]]\n" ], [ "#export\ndef dummify(o:Union[np.ndarray, torch.Tensor], by_var:bool=True, inplace:bool=False, skip:Optional[list]=None, random_state=None):\n \"\"\"Shuffles an array-like object along all dimensions or dimension 1 (variables) if by_var is True.\"\"\"\n if not inplace: \n if isinstance(o, np.ndarray): o_dummy = o.copy()\n elif isinstance(o, torch.Tensor): o_dummy = o.clone()\n else: o_dummy = o\n if by_var:\n for k in progress_bar(range(o.shape[1]), leave=False):\n if skip is not None and k in listify(skip): continue\n o_dummy[:, k] = random_shuffle(o[:, k].flatten(), random_state=random_state).reshape(o[:, k].shape)\n else:\n o_dummy[:] = random_shuffle(o.flatten(), random_state=random_state).reshape(o.shape)\n if not inplace: \n return o_dummy", "_____no_output_____" ], [ "arr = np.random.rand(2,3,10)\narr_original = arr.copy()\ndummy_arr = dummify(arr)\ntest_ne(arr_original, dummy_arr)\ntest_eq(arr_original, arr)\ndummify(arr, inplace=True)\ntest_ne(arr_original, arr)", "_____no_output_____" ], [ "t = torch.rand(2,3,10)\nt_original = t.clone()\ndummy_t = dummify(t)\ntest_ne(t_original, dummy_t)\ntest_eq(t_original, t)\ndummify(t, inplace=True)\ntest_ne(t_original, t)", "_____no_output_____" ], [ "#export\ndef shuffle_along_axis(o, axis=-1, random_state=None):\n if isinstance(o, torch.Tensor): size = o.numel()\n else: size = np.size(o)\n for ax in listify(axis):\n idx = random_shuffle(np.arange(size), random_state=random_state).reshape(*o.shape).argsort(axis=ax)\n o = np.take_along_axis(o, idx, axis=ax)\n return o", "_____no_output_____" ], [ "X = np.arange(60).reshape(2,3,10) + 10\nX_shuffled = shuffle_along_axis(X,(0, -1), random_state=23)\ntest_eq(X_shuffled, np.array([[[13, 15, 41, 14, 40, 49, 18, 42, 47, 46],\n [28, 56, 53, 50, 52, 25, 24, 57, 51, 59],\n [34, 30, 38, 35, 69, 66, 63, 67, 61, 62]],\n\n [[19, 10, 11, 16, 43, 12, 17, 48, 45, 44],\n [23, 20, 26, 22, 21, 27, 58, 29, 54, 55],\n [36, 31, 39, 60, 33, 68, 37, 32, 65, 64]]]))", "_____no_output_____" ], [ "# export\ndef analyze_feature(feature, bins=100, density=False, feature_name=None, clip_outliers_plot=False, quantile_range=(25.0, 75.0), \n percentiles=[1, 25, 50, 75, 99], text_len=12, figsize=(10,6)):\n non_nan_feature = feature[~np.isnan(feature)]\n nan_perc = np.isnan(feature).mean()\n print(f\"{'dtype':>{text_len}}: {feature.dtype}\")\n print(f\"{'nan values':>{text_len}}: {nan_perc:.1%}\")\n print(f\"{'max':>{text_len}}: {np.nanmax(feature)}\")\n for p in percentiles:\n print(f\"{p:>{text_len}.0f}: {get_percentile(feature, p)}\")\n print(f\"{'min':>{text_len}}: {np.nanmin(feature)}\")\n min_outliers, max_outliers = get_outliers_IQR(feature, quantile_range=quantile_range)\n print(f\"{'outlier min':>{text_len}}: {min_outliers}\")\n print(f\"{'outlier max':>{text_len}}: {max_outliers}\")\n print(f\"{'outliers':>{text_len}}: {((non_nan_feature < min_outliers) | (non_nan_feature > max_outliers)).mean():.1%}\")\n print(f\"{'mean':>{text_len}}: {np.nanmean(feature)}\")\n print(f\"{'std':>{text_len}}: {np.nanstd(feature)}\")\n print(f\"{'normal dist':>{text_len}}: {normaltest(non_nan_feature, axis=0, nan_policy='propagate')[1] > .05}\")\n plt.figure(figsize=figsize)\n if clip_outliers_plot:\n plt.hist(np.clip(non_nan_feature, min_outliers, max_outliers), bins, density=density, color='lime', edgecolor='black')\n else: \n plt.hist(non_nan_feature, bins, density=density, color='lime', edgecolor='black')\n plt.axvline(min_outliers, lw=1, ls='--', color='red')\n plt.axvline(max_outliers, lw=1, ls='--', color='red')\n plt.title(f\"feature: {feature_name}\")\n plt.show()\n \ndef analyze_array(o, bins=100, density=False, feature_names=None, clip_outliers_plot=False, quantile_range=(25.0, 75.0), \n percentiles=[1, 25, 50, 75, 99], text_len=12, figsize=(10,6)):\n if percentiles:\n percentiles = np.sort(percentiles)[::-1]\n print(f\"{'array shape':>{text_len}}: {o.shape}\")\n if o.ndim > 1:\n for f in range(o.shape[1]):\n feature_name = f\"{feature_names[f]}\" if feature_names is not None else f\n print(f\"\\n{f:3} {'feature':>{text_len - 4}}: {feature_name}\\n\")\n analyze_feature(o[:, f].flatten(), feature_name=feature_name)\n else:\n analyze_feature(o.flatten(), feature_name=feature_names) ", "_____no_output_____" ], [ "x = np.random.normal(size=(1000))\nanalyze_array(x)", " array shape: (1000,)\n dtype: float64\n nan values: 0.0%\n max: 2.8962792292643007\n 1: -2.120764247253952\n 25: -0.6776624311119962\n 50: 0.03797148808519438\n 75: 0.6837275374444655\n 99: 2.116961348657088\n min: -3.23432018653886\n outlier min: -2.7197473839466886\n outlier max: 2.725812490279158\n outliers: 0.5%\n mean: 0.030734294784013245\n std: 0.9799586630237788\n normal dist: True\n" ], [ "x1 = np.random.normal(size=(1000,2))\nx2 = np.random.normal(3, 5, size=(1000,2))\nx = x1 + x2\nanalyze_array(x)", " array shape: (1000, 2)\n\n 0 feature: 0\n\n dtype: float64\n nan values: 0.0%\n max: 22.193197926165894\n 1: -8.938044186311068\n 25: -0.6161653903943344\n 50: 2.64602402345483\n 75: 6.355462078190231\n 99: 15.520739262450803\n min: -11.708551852753047\n outlier min: -11.073606593271183\n outlier max: 16.81290328106708\n outliers: 0.7%\n mean: 2.899962696597163\n std: 5.23084401171235\n normal dist: True\n" ], [ "#export\ndef get_relpath(path):\n current_path = os.getcwd()\n if is_listy(path):\n relpaths = []\n for p in path:\n relpaths.append(os.path.relpath(p, current_path))\n return relpaths\n else:\n return os.path.relpath(path, current_path)", "_____no_output_____" ], [ "#hide\nfrom tsai.imports import *\nfrom tsai.export import *\nnb_name = get_nb_name()\n# nb_name = \"001_utils.ipynb\"\ncreate_scripts(nb_name);", "_____no_output_____" ] ] ]
[ "code", "markdown", "code", "markdown", "code" ]
[ [ "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code" ] ]
ece8632579b72fdd8b67dbffb497668f2e5c717d
6,824
ipynb
Jupyter Notebook
.ipynb_checkpoints/google_translate_api-checkpoint.ipynb
anir16293/MarcoSearch
6c1f3932057e8b55f43bb61ee32d94fc149ad82e
[ "MIT" ]
1
2019-09-16T22:33:49.000Z
2019-09-16T22:33:49.000Z
google_translate_api.ipynb
anir16293/MarcoSearch
6c1f3932057e8b55f43bb61ee32d94fc149ad82e
[ "MIT" ]
null
null
null
google_translate_api.ipynb
anir16293/MarcoSearch
6c1f3932057e8b55f43bb61ee32d94fc149ad82e
[ "MIT" ]
null
null
null
89.789474
1,808
0.672333
[ [ [ "from googletrans import Translator", "_____no_output_____" ], [ "translator = Translator()", "_____no_output_____" ], [ "translator.detect('kk')", "_____no_output_____" ] ] ]
[ "code" ]
[ [ "code", "code", "code" ] ]
ece863ecba647303b8f0182b06f45b0d755f909c
7,360
ipynb
Jupyter Notebook
Notebook 2, Graphical Error Modelling.ipynb
pranav-ust/cognates
c1924e6daa7fb4367ca72e37f613be60ea2fbca5
[ "MIT" ]
4
2018-11-25T18:50:13.000Z
2022-02-05T01:32:59.000Z
Notebook 2, Graphical Error Modelling.ipynb
pranav-ust/cognates
c1924e6daa7fb4367ca72e37f613be60ea2fbca5
[ "MIT" ]
null
null
null
Notebook 2, Graphical Error Modelling.ipynb
pranav-ust/cognates
c1924e6daa7fb4367ca72e37f613be60ea2fbca5
[ "MIT" ]
null
null
null
32.566372
182
0.550815
[ [ [ "# Graphical Error Modelling", "_____no_output_____" ], [ "This notebook details about algorithms discussed in section 3 of the paper, \"Alignment Analysis of Sequential Segmentation of Lexicons to Improve Automatic Cognate Detection\"", "_____no_output_____" ], [ "## Imports", "_____no_output_____" ], [ "The function `common_elements` will carry out the operation of set intersection and `uncommon_elements` will carry out the operation of set difference.", "_____no_output_____" ] ], [ [ "from math import ceil, floor\n\ndef common_elements(list1, list2):\n ''' Carries out set intersection '''\n return [element for element in list1 if element in list2]\n\ndef uncommon_elements(list1, list2):\n ''' Carries out set difference '''\n return [element for element in list1 if element not in list2]", "_____no_output_____" ] ], [ [ "We import the functions from the previous notebook which are `shingle` and `two_ends`.", "_____no_output_____" ] ], [ [ "def shingle(input, k):\n ''' Shingles the input into k-grams '''\n k = min(len(input), k)\n start_combinations = [input[:i] for i in range(1, k)]\n kgrams = [input[i:i + k] for i in range(len(input) - k + 1)]\n end_combinations = [input[-i:] for i in range(k - 1, 0, -1)]\n return start_combinations + kgrams + end_combinations\n\ndef two_ends(input, k):\n ''' Shingles the input into k-grams but encodes numbers from two ends '''\n basic = shingle(input, k)\n result =[]\n for i in range(1, len(basic) + 1):\n if i <= (len(input) - i + 2):\n result.append(str(i) + basic[i - 1]) # Append numbers from start\n else:\n result.append(basic[i - 1] + str(len(basic) - i + 1)) # Append numbers from end\n return result", "_____no_output_____" ] ], [ [ "## Graphical Modelling Algorithm", "_____no_output_____" ], [ "The algorithm consists of three parts:\n1. Initialization\n2. Equalization of the set cardinalities\n3. Inserting the mappings of the set members into the graph\n\nThis algorithm returns the graphical structure between two shingle sets.", "_____no_output_____" ] ], [ [ "def graph_model(first, second):\n ''' Constructs the graphical structure between two shingle sets. '''\n \n # Step 1: Initialization\n # If the given sets first and second are empty, we initialize \n # them by inserting an empty token, (nun), into those sets.\n \n if len(first) == 0:\n first.append(\"nun\") #insert empty token if found empty\n if len(second) == 0:\n second.append(\"nun\") #insert empty token if found empty\n \n # Step 2: Equalization of the set cardinalities\n # The cardinalities of the sets first and second made\n # equal by inserting empty tokens (nun) into the\n # middle of the sets.\n \n # While loops to equalize the sizes\n while(len(first) < len(second)):\n pos = ceil(len(first) / 2)\n first.insert(pos, \"nun\")\n \n # While loops to equalize the sizes\n while(len(first) > len(second)):\n pos = floor(len(second) / 2)\n second.insert(pos, \"nun\")\n \n # Step 3: Inserting the mappings of the set members into the graph\n # The empty graph is initialized as graph = {}.\n # The directed edges are generated, originating from every set member\n # of first to every set member of second. This results in a complete \n # directed bipartite graph between first and second sets.\n \n # Pairs in tuples\n graph = set() #Graph in sets to avoid duplicates\n \n for i in range(len(first)):\n pair = (first[i], second[i]) # One to one mapping with same index\n graph.add(pair)\n for i in range(len(first) - 1):\n pair = (first[i], second[i + 1]) # One to one mapping with an index ahead\n graph.add(pair)\n if len(first) > 1:\n for i in range(1, len(first)):\n pair = (first[i], second[i - 1]) # One to one mapping with an index before\n graph.add(pair)\n return graph", "_____no_output_____" ] ], [ [ "## Playing with the functions ", "_____no_output_____" ], [ "Let the source cognate be *mesia* and target cognate be *messia*.\n\nFirstly, we will shingle them using `two_ends` function.\n\nUsing the `common_elements` function, we would find $S \\cap T$.\n\nUsing the `uncommon_elements` function, we would find $S - (S \\cap T)$, which would form the *top*.\n\nUsing the `uncommon_elements` function, we would find $T - (S \\cap T)$, which would form the *bottom*.\n\nUsing the `graph_model` function, the graph would be outputted.", "_____no_output_____" ] ], [ [ "source = two_ends(\"mesia\", 2) # Source cognate\ntarget = two_ends(\"messia\", 2) # Target cognate\nprint(source, target)\nst = common_elements(source, target) # s cap t\ntop = uncommon_elements(source, st) # s - (s cap t)\nbottom = uncommon_elements(target, st) # t - (s cap t)\nprint(\"Top and bottom are\", top, bottom)\nprint(\"Graph is \", graph_model(top, bottom))", "_____no_output_____" ] ], [ [ "We get the graph as `{('nun', '4ss')}`, which means as $\\phi \\rightarrow 4ss$.\nIntuitively, $\\phi \\rightarrow 4ss$ means that if the letter s is added at position 4 of the word of the source *mesia*, then one could get the target word *messia*.", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown" ]
[ [ "markdown", "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown" ] ]
ece864b317a34ddf08f32b30a7cdee8233a2cc37
108,318
ipynb
Jupyter Notebook
Multiple_sclerosis_detection_K_fold_3D_intersection_data_preparation.ipynb
abr-98/segmentation_for_multiple_Scleorisis_Detection
ec147d662d5184df3afe2a810bb3619438a52ee0
[ "MIT" ]
null
null
null
Multiple_sclerosis_detection_K_fold_3D_intersection_data_preparation.ipynb
abr-98/segmentation_for_multiple_Scleorisis_Detection
ec147d662d5184df3afe2a810bb3619438a52ee0
[ "MIT" ]
null
null
null
Multiple_sclerosis_detection_K_fold_3D_intersection_data_preparation.ipynb
abr-98/segmentation_for_multiple_Scleorisis_Detection
ec147d662d5184df3afe2a810bb3619438a52ee0
[ "MIT" ]
null
null
null
23.795694
483
0.300144
[ [ [ "## Mounter", "_____no_output_____" ] ], [ [ "from google.colab import drive\ndrive.mount('/content/drive')", "Go to this URL in a browser: https://accounts.google.com/o/oauth2/auth?client_id=947318989803-6bn6qk8qdgf4n4g3pfee6491hc0brc4i.apps.googleusercontent.com&redirect_uri=urn%3aietf%3awg%3aoauth%3a2.0%3aoob&response_type=code&scope=email%20https%3a%2f%2fwww.googleapis.com%2fauth%2fdocs.test%20https%3a%2f%2fwww.googleapis.com%2fauth%2fdrive%20https%3a%2f%2fwww.googleapis.com%2fauth%2fdrive.photos.readonly%20https%3a%2f%2fwww.googleapis.com%2fauth%2fpeopleapi.readonly\n\nEnter your authorization code:\n··········\nMounted at /content/drive\n" ] ], [ [ "## Function Declaration", "_____no_output_____" ] ], [ [ "import tensorflow as tf\nimport numpy as np\nimport os\nimport nibabel as nib\nimport cv2 as cv\nimport matplotlib.pyplot as plt\n \ndef modality(Path,index):\n X = []\n X_per_paitent = []\n p=os.listdir(Path) \n recs_in=[]\n counter=0\n counter_2=0\n\n for i in p[:14]: # Loading all the folders in the given path\n q = os.listdir(os.path.join(Path,i)) \n\n x = nib.load(os.path.join(Path,i,q[index])) \n f = x.get_fdata()\n f = np.asarray(f,'float32')\n \n ct=0\n recs_in.append(f.shape[2])\n #print(counter_2)\n counter_2+=1\n for j in range(f.shape[2]): # Processing the MRI Scan in the axial view\n _slice = cv.resize(f[:,:,j],(256,256),interpolation=cv.INTER_NEAREST) # Resizing the slice to the shape(256,256)\n if(index not in [3,4,5,6,7,8,9] and np.sum(_slice) != 0 ): \n if index==1:\n ct+=1 \n counter+=1 # To check whether the slice is null or not\n # _slice = _slice / (np.max(_slice) + 0.00001) # Normalization\n _slice = (_slice - np.mean(_slice) + 0.00001) / (np.std(_slice) + 0.00001)\n # Standardization\n else:\n # To check whether the slice is null or not\n # _slice = _slice / (np.max(_slice) + 0.00001) # Normalization\n _slice = (_slice - np.mean(_slice) + 0.00001) / (np.std(_slice) + 0.00001) # Standardization\n elif(index in [3,4,5,6,7,8,9]): # if index = 3, Then it is output mask and we don't normalize or standardize it \n _slice = np.array(_slice)\n _slice[_slice > 0] = 1.0\n _slice[_slice < 0] = 0.0\n _slice = _slice.T\n _slice = _slice[:,:,np.newaxis]\n X.append(_slice)\n # X=np.array(X,dtype='float32')\n return X,recs_in", "_____no_output_____" ], [ "def remove_null_samples(X_Dp, X_Flair, X_Gado, X_T1, X_T2, Y_Manual,recs): \n \n X=[]\n Y=[]\n counter=0\n counter_2=0\n mult=0;\n count=0\n rec=[]\n keep_count=[]\n keep=[]\n print(recs)\n r=np.array(recs,dtype='float32')\n print(np.sum(r))\n print(len(X_Dp))\n\n for i in range(len(X_Dp)): \n if counter==(recs[mult]):\n print(counter)\n mult+=1\n rec.append(count)\n counter=0\n print(counter_2)\n count=0\n final_slice = np.concatenate((X_Dp[i],X_Flair[i],X_Gado[i],X_T1[i],X_T2[i]), axis = -1)\n if(np.sum(final_slice) != 0): # checking whether the final slice is empty or not \n X.append(final_slice)\n Y.append(Y_Manual[i])\n \n count+=1\n counter+=1\n counter_2+=1\n\n \n rec.append(count)\n# Converting the list into array \n X=np.array(X,dtype='float32')\n Y=np.array(Y,dtype='float32')\n rec=np.array(rec,dtype='float32')\n \n return X,Y,rec\n", "_____no_output_____" ], [ "def store_data(X,Y,rec):\n np.save(\"drive/My Drive/MS_data/X_intersection_new.npy\",X)\n np.save(\"drive/My Drive/MS_data/Y_intersection_new.npy\",Y)\n np.save(\"drive/My Drive/MS_data/intersection_new_rec_after_removal.npy\",rec)", "_____no_output_____" ], [ "import math\ndef intersection(Y_1,Y_2,Y_3,Y_4,Y_5,Y_6,Y_7):\n Y=[]\n sum2=[]\n flag=0\n #y=np.array()\n print(\"A\")\n for i in range (len(Y_1)):\n #print(Y_1[i])\n \n f=np.concatenate((Y_1[i],Y_2[i],Y_3[i],Y_4[i],Y_5[i],Y_6[i],Y_7[i]),axis=-1)\n sum=np.sum(f,axis=2)\n # print(sum)\n \n #print(j)\n sum_1=np.divide(sum,7)\n sum_1=np.floor(sum_1)\n sum2.append(sum_1)\n\n #sum2=np.array(sum2,dtype='float32')\n \n \n return sum2\n ", "_____no_output_____" ] ], [ [ "## Data Preprocessing", "_____no_output_____" ] ], [ [ "import numpy as np\nimport matplotlib.pyplot as plt\nimport tensorflow as tf\nfrom sklearn.model_selection import train_test_split\n\n#import keras\nfrom tensorflow.keras.models import Model, load_model\nfrom tensorflow.keras.layers import Input ,BatchNormalization , Activation \nfrom tensorflow.keras.layers import Conv2D, UpSampling2D\nfrom tensorflow.keras.layers import MaxPooling2D\nfrom tensorflow.keras.layers import concatenate\nfrom tensorflow.keras.callbacks import EarlyStopping, ModelCheckpoint\nfrom tensorflow.keras.optimizers import Adam\nfrom sklearn.model_selection import train_test_split\n\n\n#import dataPrepare as process\n# Loading all the 5 different modalities of each MRI Scan of all 15 different patients and 1st rater Manual SegmentationX_Dp = modality(Path,0)\n#import Modified_UNet \n#import plots\n#import Metrics\n\n# Setting the path\nPath='drive/My Drive/Pre-processed'\n\n\n\n# Loading all the 5 different modalities of each MRI Scan of all 15 different patients and 1st rater Manual Segmentation\nX_Dp_t,rec = modality(Path,0)\nX_Flair_t,rec_1 = modality(Path,1)\nX_Gado_t,rec = modality(Path,2)\nX_T1_t,rec = modality(Path,10)\nX_T2_t,rec = modality(Path,11)\nrec=np.array(rec_1,dtype='float32')\nnp.save(\"drive/My Drive/MS_data/intersection_new_rec_before_removal.npy\",rec)", "_____no_output_____" ], [ "import numpy as np\nimport matplotlib.pyplot as plt\nimport tensorflow as tf\nfrom sklearn.model_selection import train_test_split\n\n#import keras\nfrom tensorflow.keras.models import Model, load_model\nfrom tensorflow.keras.layers import Input ,BatchNormalization , Activation \nfrom tensorflow.keras.layers import Conv2D, UpSampling2D\nfrom tensorflow.keras.layers import MaxPooling2D\nfrom tensorflow.keras.layers import concatenate\nfrom tensorflow.keras.callbacks import EarlyStopping, ModelCheckpoint\nfrom tensorflow.keras.optimizers import Adam\nfrom sklearn.model_selection import train_test_split\n\n\n#import dataPrepare as process\n# Loading all the 5 different modalities of each MRI Scan of all 15 different patients and 1st rater Manual SegmentationX_Dp = modality(Path,0)\n#import Modified_UNet \n#import plots\n#import Metrics\n\n# Setting the path\nPath='drive/My Drive/Pre-processed'\nY_1,rec = modality(Path,3)\nY_2,rec = modality(Path,4)\nY_3,rec = modality(Path,5)\nY_4,rec = modality(Path,6)\nY_5,rec = modality(Path,7)\nY_6,rec = modality(Path,8)\nY_7,rec = modality(Path,9)", "_____no_output_____" ], [ "Y_Manual=intersection(Y_1,Y_2,Y_3,Y_4,Y_5,Y_6,Y_7)", "A\n" ], [ "import numpy as np\n\nnp.save(\"drive/My Drive/MS_data/Y_manual_new.npy\",Y_Manual)", "_____no_output_____" ], [ "Y_manual=list(Y_Manual)", "_____no_output_____" ], [ "Y_Manual=np.load(\"drive/My Drive/MS_data/Y_manual_new.npy\")", "_____no_output_____" ], [ "X, Y,rec = remove_null_samples(X_Dp_t, X_Flair_t, X_Gado_t, X_T1_t, X_T2_t, Y_manual,rec_1)", "[256, 256, 512, 512, 256, 512, 512, 256, 512, 256, 336, 336, 336, 336]\n5184.0\n5184\n256\n256\n256\n512\n512\n1024\n512\n1536\n256\n1792\n512\n2304\n512\n2816\n256\n3072\n512\n3584\n256\n3840\n336\n4176\n336\n4512\n336\n4848\n" ], [ "store_data(X,Y,rec)", "_____no_output_____" ], [ "import numpy as np\nY=np.load(\"drive/My Drive/MS_data/Y_intersection_new.npy\")", "_____no_output_____" ], [ "Y_Manual=list(Y)", "_____no_output_____" ], [ "Y_1=Y_Manual[:1401]\nY_2=Y_Manual[1401:]", "_____no_output_____" ], [ "p=[]\ny=[]\nz=[]\na=[]\nn=1\nfor i in Y_1:\n print(n)\n for j in i:\n for k in j:\n #print(k)\n a.append(k)\n # print(a)\n z.append(a)\n a=[]\n #print(z)\n y.append(z)\n z=[]\n p.append(y)\n y=[]\n n+=1\n\n\n#print(p)\n", "_____no_output_____" ], [ "temp=np.array(p)\nnp.save(\"drive/My Drive/MS_data/temp_1_new.npy\",temp)", "_____no_output_____" ], [ "q=[]\ny=[]\nz=[]\na=[]\nn=1401\nfor i in Y_2:\n print(n)\n for j in i:\n for k in j:\n #print(k)\n a.append(k)\n # print(a)\n z.append(a)\n a=[]\n #print(z)\n y.append(z)\n z=[]\n q.append(y)\n y=[]\n n+=1\n", "1401\n1402\n1403\n1404\n1405\n1406\n1407\n1408\n1409\n1410\n1411\n1412\n1413\n1414\n1415\n1416\n1417\n1418\n1419\n1420\n1421\n1422\n1423\n1424\n1425\n1426\n1427\n1428\n1429\n1430\n1431\n1432\n1433\n1434\n1435\n1436\n1437\n1438\n1439\n1440\n1441\n1442\n1443\n1444\n1445\n1446\n1447\n1448\n1449\n1450\n1451\n1452\n1453\n1454\n1455\n1456\n1457\n1458\n1459\n1460\n1461\n1462\n1463\n1464\n1465\n1466\n1467\n1468\n1469\n1470\n1471\n1472\n1473\n1474\n1475\n1476\n1477\n1478\n1479\n1480\n1481\n1482\n1483\n1484\n1485\n1486\n1487\n1488\n1489\n1490\n1491\n1492\n1493\n1494\n1495\n1496\n1497\n1498\n1499\n1500\n1501\n1502\n1503\n1504\n1505\n1506\n1507\n1508\n1509\n1510\n1511\n1512\n1513\n1514\n1515\n1516\n1517\n1518\n1519\n1520\n1521\n1522\n1523\n1524\n1525\n1526\n1527\n1528\n1529\n1530\n1531\n1532\n1533\n1534\n1535\n1536\n1537\n1538\n1539\n1540\n1541\n1542\n1543\n1544\n1545\n1546\n1547\n1548\n1549\n1550\n1551\n1552\n1553\n1554\n1555\n1556\n1557\n1558\n1559\n1560\n1561\n1562\n1563\n1564\n1565\n1566\n1567\n1568\n1569\n1570\n1571\n1572\n1573\n1574\n1575\n1576\n1577\n1578\n1579\n1580\n1581\n1582\n1583\n1584\n1585\n1586\n1587\n1588\n1589\n1590\n1591\n1592\n1593\n1594\n1595\n1596\n1597\n1598\n1599\n1600\n1601\n1602\n1603\n1604\n1605\n1606\n1607\n1608\n1609\n1610\n1611\n1612\n1613\n1614\n1615\n1616\n1617\n1618\n1619\n1620\n1621\n1622\n1623\n1624\n1625\n1626\n1627\n1628\n1629\n1630\n1631\n1632\n1633\n1634\n1635\n1636\n1637\n1638\n1639\n1640\n1641\n1642\n1643\n1644\n1645\n1646\n1647\n1648\n1649\n1650\n1651\n1652\n1653\n1654\n1655\n1656\n1657\n1658\n1659\n1660\n1661\n1662\n1663\n1664\n1665\n1666\n1667\n1668\n1669\n1670\n1671\n1672\n1673\n1674\n1675\n1676\n1677\n1678\n1679\n1680\n1681\n1682\n1683\n1684\n1685\n1686\n1687\n1688\n1689\n1690\n1691\n1692\n1693\n1694\n1695\n1696\n1697\n1698\n1699\n1700\n1701\n1702\n1703\n1704\n1705\n1706\n1707\n1708\n1709\n1710\n1711\n1712\n1713\n1714\n1715\n1716\n1717\n1718\n1719\n1720\n1721\n1722\n1723\n1724\n1725\n1726\n1727\n1728\n1729\n1730\n1731\n1732\n1733\n1734\n1735\n1736\n1737\n1738\n1739\n1740\n1741\n1742\n1743\n1744\n1745\n1746\n1747\n1748\n1749\n1750\n1751\n1752\n1753\n1754\n1755\n1756\n1757\n1758\n1759\n1760\n1761\n1762\n1763\n1764\n1765\n1766\n1767\n1768\n1769\n1770\n1771\n1772\n1773\n1774\n1775\n1776\n1777\n1778\n1779\n1780\n1781\n1782\n1783\n1784\n1785\n1786\n1787\n1788\n1789\n1790\n1791\n1792\n1793\n1794\n1795\n1796\n1797\n1798\n1799\n1800\n1801\n1802\n1803\n1804\n1805\n1806\n1807\n1808\n1809\n1810\n1811\n1812\n1813\n1814\n1815\n1816\n1817\n1818\n1819\n1820\n1821\n1822\n1823\n1824\n1825\n1826\n1827\n1828\n1829\n1830\n1831\n1832\n1833\n1834\n1835\n1836\n1837\n1838\n1839\n1840\n1841\n1842\n1843\n1844\n1845\n1846\n1847\n1848\n1849\n1850\n1851\n1852\n1853\n1854\n1855\n1856\n1857\n1858\n1859\n1860\n1861\n1862\n1863\n1864\n1865\n1866\n1867\n1868\n1869\n1870\n1871\n1872\n1873\n1874\n1875\n1876\n1877\n1878\n1879\n1880\n1881\n1882\n1883\n1884\n1885\n1886\n1887\n1888\n1889\n1890\n1891\n1892\n1893\n1894\n1895\n1896\n1897\n1898\n1899\n1900\n1901\n1902\n1903\n1904\n1905\n1906\n1907\n1908\n1909\n1910\n1911\n1912\n1913\n1914\n1915\n1916\n1917\n1918\n1919\n1920\n1921\n1922\n1923\n1924\n1925\n1926\n1927\n1928\n1929\n1930\n1931\n1932\n1933\n1934\n1935\n1936\n1937\n1938\n1939\n1940\n1941\n1942\n1943\n1944\n1945\n1946\n1947\n1948\n1949\n1950\n1951\n1952\n1953\n1954\n1955\n1956\n1957\n1958\n1959\n1960\n1961\n1962\n1963\n1964\n1965\n1966\n1967\n1968\n1969\n1970\n1971\n1972\n1973\n1974\n1975\n1976\n1977\n1978\n1979\n1980\n1981\n1982\n1983\n1984\n1985\n1986\n1987\n1988\n1989\n1990\n1991\n1992\n1993\n1994\n1995\n1996\n1997\n1998\n1999\n2000\n2001\n2002\n2003\n2004\n2005\n2006\n2007\n2008\n2009\n2010\n2011\n2012\n2013\n2014\n2015\n2016\n2017\n2018\n2019\n2020\n2021\n2022\n2023\n2024\n2025\n2026\n2027\n2028\n2029\n2030\n2031\n2032\n2033\n2034\n2035\n2036\n2037\n2038\n2039\n2040\n2041\n2042\n2043\n2044\n2045\n2046\n2047\n2048\n2049\n2050\n2051\n2052\n2053\n2054\n2055\n2056\n2057\n2058\n2059\n2060\n2061\n2062\n2063\n2064\n2065\n2066\n2067\n2068\n2069\n2070\n2071\n2072\n2073\n2074\n2075\n2076\n2077\n2078\n2079\n2080\n2081\n2082\n2083\n2084\n2085\n2086\n2087\n2088\n2089\n2090\n2091\n2092\n2093\n2094\n2095\n2096\n2097\n2098\n2099\n2100\n2101\n2102\n2103\n2104\n2105\n2106\n2107\n2108\n2109\n2110\n2111\n2112\n2113\n2114\n2115\n2116\n2117\n2118\n2119\n2120\n2121\n2122\n2123\n2124\n2125\n2126\n2127\n2128\n2129\n2130\n2131\n2132\n2133\n2134\n2135\n2136\n2137\n2138\n2139\n2140\n2141\n2142\n2143\n2144\n2145\n2146\n2147\n2148\n2149\n2150\n2151\n2152\n2153\n2154\n2155\n2156\n2157\n2158\n2159\n2160\n2161\n2162\n2163\n2164\n2165\n2166\n2167\n2168\n2169\n2170\n2171\n2172\n2173\n2174\n2175\n2176\n2177\n2178\n2179\n2180\n2181\n2182\n2183\n2184\n2185\n2186\n2187\n2188\n2189\n2190\n2191\n2192\n2193\n2194\n2195\n2196\n2197\n2198\n2199\n2200\n2201\n2202\n2203\n2204\n2205\n2206\n2207\n2208\n2209\n2210\n2211\n2212\n2213\n2214\n2215\n2216\n2217\n2218\n2219\n2220\n2221\n2222\n2223\n2224\n2225\n2226\n2227\n2228\n2229\n2230\n2231\n2232\n2233\n2234\n2235\n2236\n2237\n2238\n2239\n2240\n2241\n2242\n2243\n2244\n2245\n2246\n2247\n2248\n2249\n2250\n2251\n2252\n2253\n2254\n2255\n2256\n2257\n2258\n2259\n2260\n2261\n2262\n2263\n2264\n2265\n2266\n2267\n2268\n2269\n2270\n2271\n2272\n2273\n2274\n2275\n2276\n2277\n2278\n2279\n2280\n2281\n2282\n2283\n2284\n2285\n2286\n2287\n2288\n2289\n2290\n2291\n2292\n2293\n2294\n2295\n2296\n2297\n2298\n2299\n2300\n2301\n2302\n2303\n2304\n2305\n2306\n2307\n2308\n2309\n2310\n2311\n2312\n2313\n2314\n2315\n2316\n2317\n2318\n2319\n2320\n2321\n2322\n2323\n2324\n2325\n2326\n2327\n2328\n2329\n2330\n2331\n2332\n2333\n2334\n2335\n2336\n2337\n2338\n2339\n2340\n2341\n2342\n2343\n2344\n2345\n2346\n2347\n2348\n2349\n2350\n2351\n2352\n2353\n2354\n2355\n2356\n2357\n2358\n2359\n2360\n2361\n2362\n2363\n2364\n2365\n2366\n2367\n2368\n2369\n2370\n2371\n2372\n2373\n2374\n2375\n2376\n2377\n2378\n2379\n2380\n2381\n2382\n2383\n2384\n2385\n2386\n2387\n2388\n2389\n2390\n2391\n2392\n2393\n2394\n2395\n2396\n2397\n2398\n2399\n2400\n2401\n2402\n2403\n2404\n2405\n2406\n2407\n2408\n2409\n2410\n2411\n2412\n2413\n2414\n2415\n2416\n2417\n2418\n2419\n2420\n2421\n2422\n2423\n2424\n2425\n2426\n2427\n2428\n2429\n2430\n2431\n2432\n2433\n2434\n2435\n2436\n2437\n2438\n2439\n2440\n2441\n2442\n2443\n2444\n2445\n2446\n2447\n2448\n2449\n2450\n2451\n2452\n2453\n2454\n2455\n2456\n2457\n2458\n2459\n2460\n2461\n2462\n2463\n2464\n2465\n2466\n2467\n2468\n2469\n2470\n2471\n2472\n2473\n2474\n2475\n2476\n2477\n2478\n2479\n2480\n2481\n2482\n2483\n2484\n2485\n2486\n2487\n2488\n2489\n2490\n2491\n2492\n2493\n2494\n2495\n2496\n2497\n2498\n2499\n2500\n2501\n2502\n2503\n2504\n2505\n2506\n2507\n2508\n2509\n2510\n2511\n2512\n2513\n2514\n2515\n2516\n2517\n2518\n2519\n2520\n2521\n2522\n2523\n2524\n2525\n2526\n2527\n2528\n2529\n2530\n2531\n2532\n2533\n2534\n2535\n2536\n2537\n2538\n2539\n2540\n2541\n2542\n2543\n2544\n2545\n2546\n2547\n2548\n2549\n2550\n2551\n2552\n2553\n2554\n2555\n2556\n2557\n2558\n2559\n2560\n2561\n2562\n2563\n2564\n2565\n2566\n2567\n2568\n2569\n2570\n2571\n2572\n2573\n2574\n2575\n2576\n2577\n2578\n2579\n2580\n2581\n2582\n2583\n2584\n2585\n2586\n2587\n2588\n2589\n2590\n2591\n2592\n2593\n2594\n2595\n2596\n2597\n2598\n2599\n2600\n2601\n2602\n2603\n2604\n2605\n2606\n2607\n2608\n2609\n2610\n2611\n2612\n2613\n2614\n2615\n2616\n2617\n2618\n2619\n2620\n2621\n2622\n2623\n2624\n2625\n2626\n2627\n2628\n2629\n2630\n2631\n2632\n2633\n2634\n2635\n2636\n2637\n2638\n2639\n2640\n2641\n2642\n2643\n2644\n2645\n2646\n2647\n2648\n2649\n2650\n2651\n2652\n2653\n2654\n2655\n2656\n2657\n2658\n2659\n2660\n2661\n2662\n2663\n2664\n2665\n2666\n2667\n2668\n2669\n2670\n2671\n2672\n2673\n2674\n2675\n2676\n2677\n2678\n2679\n2680\n2681\n2682\n2683\n2684\n2685\n2686\n2687\n2688\n2689\n2690\n2691\n2692\n2693\n2694\n2695\n2696\n2697\n2698\n2699\n2700\n2701\n2702\n2703\n2704\n2705\n2706\n2707\n2708\n2709\n2710\n2711\n2712\n2713\n2714\n2715\n2716\n2717\n2718\n2719\n2720\n2721\n2722\n2723\n2724\n2725\n2726\n2727\n2728\n2729\n2730\n2731\n2732\n2733\n2734\n2735\n2736\n2737\n2738\n2739\n2740\n2741\n2742\n2743\n2744\n2745\n2746\n2747\n2748\n2749\n2750\n2751\n2752\n2753\n2754\n2755\n2756\n2757\n2758\n2759\n2760\n2761\n2762\n2763\n2764\n2765\n2766\n2767\n2768\n2769\n2770\n2771\n2772\n2773\n2774\n2775\n2776\n2777\n2778\n2779\n2780\n2781\n2782\n2783\n2784\n2785\n2786\n2787\n2788\n2789\n2790\n2791\n2792\n2793\n2794\n2795\n2796\n2797\n2798\n2799\n2800\n2801\n2802\n2803\n2804\n2805\n2806\n2807\n2808\n2809\n2810\n2811\n2812\n2813\n2814\n2815\n2816\n2817\n2818\n2819\n2820\n2821\n2822\n2823\n2824\n2825\n2826\n2827\n2828\n2829\n2830\n2831\n2832\n2833\n2834\n2835\n2836\n2837\n2838\n2839\n2840\n2841\n2842\n2843\n2844\n2845\n2846\n2847\n2848\n2849\n2850\n2851\n2852\n2853\n2854\n2855\n2856\n2857\n2858\n2859\n2860\n2861\n2862\n2863\n2864\n2865\n2866\n2867\n2868\n2869\n2870\n2871\n2872\n2873\n2874\n2875\n2876\n2877\n2878\n2879\n2880\n2881\n2882\n2883\n2884\n2885\n2886\n2887\n2888\n2889\n2890\n2891\n2892\n2893\n2894\n2895\n2896\n2897\n2898\n2899\n2900\n2901\n2902\n2903\n2904\n2905\n2906\n2907\n2908\n2909\n2910\n2911\n2912\n2913\n2914\n2915\n2916\n2917\n2918\n2919\n2920\n2921\n2922\n2923\n2924\n2925\n2926\n2927\n2928\n2929\n2930\n2931\n2932\n2933\n2934\n2935\n2936\n2937\n2938\n2939\n" ], [ "temp2=np.array(q)\nnp.save(\"drive/My Drive/MS_data/temp_2_new.npy\",temp2)", "_____no_output_____" ], [ "temp=np.load(\"drive/My Drive/MS_data/temp_1_new.npy\")\np=list(temp)", "_____no_output_____" ], [ "temp2=np.load(\"drive/My Drive/MS_data/temp_2_new.npy\")\nq=list(temp2)", "_____no_output_____" ], [ "j=1401\nfor i in q:\n print(j)\n p.append(i)\n j+=1", "1401\n1402\n1403\n1404\n1405\n1406\n1407\n1408\n1409\n1410\n1411\n1412\n1413\n1414\n1415\n1416\n1417\n1418\n1419\n1420\n1421\n1422\n1423\n1424\n1425\n1426\n1427\n1428\n1429\n1430\n1431\n1432\n1433\n1434\n1435\n1436\n1437\n1438\n1439\n1440\n1441\n1442\n1443\n1444\n1445\n1446\n1447\n1448\n1449\n1450\n1451\n1452\n1453\n1454\n1455\n1456\n1457\n1458\n1459\n1460\n1461\n1462\n1463\n1464\n1465\n1466\n1467\n1468\n1469\n1470\n1471\n1472\n1473\n1474\n1475\n1476\n1477\n1478\n1479\n1480\n1481\n1482\n1483\n1484\n1485\n1486\n1487\n1488\n1489\n1490\n1491\n1492\n1493\n1494\n1495\n1496\n1497\n1498\n1499\n1500\n1501\n1502\n1503\n1504\n1505\n1506\n1507\n1508\n1509\n1510\n1511\n1512\n1513\n1514\n1515\n1516\n1517\n1518\n1519\n1520\n1521\n1522\n1523\n1524\n1525\n1526\n1527\n1528\n1529\n1530\n1531\n1532\n1533\n1534\n1535\n1536\n1537\n1538\n1539\n1540\n1541\n1542\n1543\n1544\n1545\n1546\n1547\n1548\n1549\n1550\n1551\n1552\n1553\n1554\n1555\n1556\n1557\n1558\n1559\n1560\n1561\n1562\n1563\n1564\n1565\n1566\n1567\n1568\n1569\n1570\n1571\n1572\n1573\n1574\n1575\n1576\n1577\n1578\n1579\n1580\n1581\n1582\n1583\n1584\n1585\n1586\n1587\n1588\n1589\n1590\n1591\n1592\n1593\n1594\n1595\n1596\n1597\n1598\n1599\n1600\n1601\n1602\n1603\n1604\n1605\n1606\n1607\n1608\n1609\n1610\n1611\n1612\n1613\n1614\n1615\n1616\n1617\n1618\n1619\n1620\n1621\n1622\n1623\n1624\n1625\n1626\n1627\n1628\n1629\n1630\n1631\n1632\n1633\n1634\n1635\n1636\n1637\n1638\n1639\n1640\n1641\n1642\n1643\n1644\n1645\n1646\n1647\n1648\n1649\n1650\n1651\n1652\n1653\n1654\n1655\n1656\n1657\n1658\n1659\n1660\n1661\n1662\n1663\n1664\n1665\n1666\n1667\n1668\n1669\n1670\n1671\n1672\n1673\n1674\n1675\n1676\n1677\n1678\n1679\n1680\n1681\n1682\n1683\n1684\n1685\n1686\n1687\n1688\n1689\n1690\n1691\n1692\n1693\n1694\n1695\n1696\n1697\n1698\n1699\n1700\n1701\n1702\n1703\n1704\n1705\n1706\n1707\n1708\n1709\n1710\n1711\n1712\n1713\n1714\n1715\n1716\n1717\n1718\n1719\n1720\n1721\n1722\n1723\n1724\n1725\n1726\n1727\n1728\n1729\n1730\n1731\n1732\n1733\n1734\n1735\n1736\n1737\n1738\n1739\n1740\n1741\n1742\n1743\n1744\n1745\n1746\n1747\n1748\n1749\n1750\n1751\n1752\n1753\n1754\n1755\n1756\n1757\n1758\n1759\n1760\n1761\n1762\n1763\n1764\n1765\n1766\n1767\n1768\n1769\n1770\n1771\n1772\n1773\n1774\n1775\n1776\n1777\n1778\n1779\n1780\n1781\n1782\n1783\n1784\n1785\n1786\n1787\n1788\n1789\n1790\n1791\n1792\n1793\n1794\n1795\n1796\n1797\n1798\n1799\n1800\n1801\n1802\n1803\n1804\n1805\n1806\n1807\n1808\n1809\n1810\n1811\n1812\n1813\n1814\n1815\n1816\n1817\n1818\n1819\n1820\n1821\n1822\n1823\n1824\n1825\n1826\n1827\n1828\n1829\n1830\n1831\n1832\n1833\n1834\n1835\n1836\n1837\n1838\n1839\n1840\n1841\n1842\n1843\n1844\n1845\n1846\n1847\n1848\n1849\n1850\n1851\n1852\n1853\n1854\n1855\n1856\n1857\n1858\n1859\n1860\n1861\n1862\n1863\n1864\n1865\n1866\n1867\n1868\n1869\n1870\n1871\n1872\n1873\n1874\n1875\n1876\n1877\n1878\n1879\n1880\n1881\n1882\n1883\n1884\n1885\n1886\n1887\n1888\n1889\n1890\n1891\n1892\n1893\n1894\n1895\n1896\n1897\n1898\n1899\n1900\n1901\n1902\n1903\n1904\n1905\n1906\n1907\n1908\n1909\n1910\n1911\n1912\n1913\n1914\n1915\n1916\n1917\n1918\n1919\n1920\n1921\n1922\n1923\n1924\n1925\n1926\n1927\n1928\n1929\n1930\n1931\n1932\n1933\n1934\n1935\n1936\n1937\n1938\n1939\n1940\n1941\n1942\n1943\n1944\n1945\n1946\n1947\n1948\n1949\n1950\n1951\n1952\n1953\n1954\n1955\n1956\n1957\n1958\n1959\n1960\n1961\n1962\n1963\n1964\n1965\n1966\n1967\n1968\n1969\n1970\n1971\n1972\n1973\n1974\n1975\n1976\n1977\n1978\n1979\n1980\n1981\n1982\n1983\n1984\n1985\n1986\n1987\n1988\n1989\n1990\n1991\n1992\n1993\n1994\n1995\n1996\n1997\n1998\n1999\n2000\n2001\n2002\n2003\n2004\n2005\n2006\n2007\n2008\n2009\n2010\n2011\n2012\n2013\n2014\n2015\n2016\n2017\n2018\n2019\n2020\n2021\n2022\n2023\n2024\n2025\n2026\n2027\n2028\n2029\n2030\n2031\n2032\n2033\n2034\n2035\n2036\n2037\n2038\n2039\n2040\n2041\n2042\n2043\n2044\n2045\n2046\n2047\n2048\n2049\n2050\n2051\n2052\n2053\n2054\n2055\n2056\n2057\n2058\n2059\n2060\n2061\n2062\n2063\n2064\n2065\n2066\n2067\n2068\n2069\n2070\n2071\n2072\n2073\n2074\n2075\n2076\n2077\n2078\n2079\n2080\n2081\n2082\n2083\n2084\n2085\n2086\n2087\n2088\n2089\n2090\n2091\n2092\n2093\n2094\n2095\n2096\n2097\n2098\n2099\n2100\n2101\n2102\n2103\n2104\n2105\n2106\n2107\n2108\n2109\n2110\n2111\n2112\n2113\n2114\n2115\n2116\n2117\n2118\n2119\n2120\n2121\n2122\n2123\n2124\n2125\n2126\n2127\n2128\n2129\n2130\n2131\n2132\n2133\n2134\n2135\n2136\n2137\n2138\n2139\n2140\n2141\n2142\n2143\n2144\n2145\n2146\n2147\n2148\n2149\n2150\n2151\n2152\n2153\n2154\n2155\n2156\n2157\n2158\n2159\n2160\n2161\n2162\n2163\n2164\n2165\n2166\n2167\n2168\n2169\n2170\n2171\n2172\n2173\n2174\n2175\n2176\n2177\n2178\n2179\n2180\n2181\n2182\n2183\n2184\n2185\n2186\n2187\n2188\n2189\n2190\n2191\n2192\n2193\n2194\n2195\n2196\n2197\n2198\n2199\n2200\n2201\n2202\n2203\n2204\n2205\n2206\n2207\n2208\n2209\n2210\n2211\n2212\n2213\n2214\n2215\n2216\n2217\n2218\n2219\n2220\n2221\n2222\n2223\n2224\n2225\n2226\n2227\n2228\n2229\n2230\n2231\n2232\n2233\n2234\n2235\n2236\n2237\n2238\n2239\n2240\n2241\n2242\n2243\n2244\n2245\n2246\n2247\n2248\n2249\n2250\n2251\n2252\n2253\n2254\n2255\n2256\n2257\n2258\n2259\n2260\n2261\n2262\n2263\n2264\n2265\n2266\n2267\n2268\n2269\n2270\n2271\n2272\n2273\n2274\n2275\n2276\n2277\n2278\n2279\n2280\n2281\n2282\n2283\n2284\n2285\n2286\n2287\n2288\n2289\n2290\n2291\n2292\n2293\n2294\n2295\n2296\n2297\n2298\n2299\n2300\n2301\n2302\n2303\n2304\n2305\n2306\n2307\n2308\n2309\n2310\n2311\n2312\n2313\n2314\n2315\n2316\n2317\n2318\n2319\n2320\n2321\n2322\n2323\n2324\n2325\n2326\n2327\n2328\n2329\n2330\n2331\n2332\n2333\n2334\n2335\n2336\n2337\n2338\n2339\n2340\n2341\n2342\n2343\n2344\n2345\n2346\n2347\n2348\n2349\n2350\n2351\n2352\n2353\n2354\n2355\n2356\n2357\n2358\n2359\n2360\n2361\n2362\n2363\n2364\n2365\n2366\n2367\n2368\n2369\n2370\n2371\n2372\n2373\n2374\n2375\n2376\n2377\n2378\n2379\n2380\n2381\n2382\n2383\n2384\n2385\n2386\n2387\n2388\n2389\n2390\n2391\n2392\n2393\n2394\n2395\n2396\n2397\n2398\n2399\n2400\n2401\n2402\n2403\n2404\n2405\n2406\n2407\n2408\n2409\n2410\n2411\n2412\n2413\n2414\n2415\n2416\n2417\n2418\n2419\n2420\n2421\n2422\n2423\n2424\n2425\n2426\n2427\n2428\n2429\n2430\n2431\n2432\n2433\n2434\n2435\n2436\n2437\n2438\n2439\n2440\n2441\n2442\n2443\n2444\n2445\n2446\n2447\n2448\n2449\n2450\n2451\n2452\n2453\n2454\n2455\n2456\n2457\n2458\n2459\n2460\n2461\n2462\n2463\n2464\n2465\n2466\n2467\n2468\n2469\n2470\n2471\n2472\n2473\n2474\n2475\n2476\n2477\n2478\n2479\n2480\n2481\n2482\n2483\n2484\n2485\n2486\n2487\n2488\n2489\n2490\n2491\n2492\n2493\n2494\n2495\n2496\n2497\n2498\n2499\n2500\n2501\n2502\n2503\n2504\n2505\n2506\n2507\n2508\n2509\n2510\n2511\n2512\n2513\n2514\n2515\n2516\n2517\n2518\n2519\n2520\n2521\n2522\n2523\n2524\n2525\n2526\n2527\n2528\n2529\n2530\n2531\n2532\n2533\n2534\n2535\n2536\n2537\n2538\n2539\n2540\n2541\n2542\n2543\n2544\n2545\n2546\n2547\n2548\n2549\n2550\n2551\n2552\n2553\n2554\n2555\n2556\n2557\n2558\n2559\n2560\n2561\n2562\n2563\n2564\n2565\n2566\n2567\n2568\n2569\n2570\n2571\n2572\n2573\n2574\n2575\n2576\n2577\n2578\n2579\n2580\n2581\n2582\n2583\n2584\n2585\n2586\n2587\n2588\n2589\n2590\n2591\n2592\n2593\n2594\n2595\n2596\n2597\n2598\n2599\n2600\n2601\n2602\n2603\n2604\n2605\n2606\n2607\n2608\n2609\n2610\n2611\n2612\n2613\n2614\n2615\n2616\n2617\n2618\n2619\n2620\n2621\n2622\n2623\n2624\n2625\n2626\n2627\n2628\n2629\n2630\n2631\n2632\n2633\n2634\n2635\n2636\n2637\n2638\n2639\n2640\n2641\n2642\n2643\n2644\n2645\n2646\n2647\n2648\n2649\n2650\n2651\n2652\n2653\n2654\n2655\n2656\n2657\n2658\n2659\n2660\n2661\n2662\n2663\n2664\n2665\n2666\n2667\n2668\n2669\n2670\n2671\n2672\n2673\n2674\n2675\n2676\n2677\n2678\n2679\n2680\n2681\n2682\n2683\n2684\n2685\n2686\n2687\n2688\n2689\n2690\n2691\n2692\n2693\n2694\n2695\n2696\n2697\n2698\n2699\n2700\n2701\n2702\n2703\n2704\n2705\n2706\n2707\n2708\n2709\n2710\n2711\n2712\n2713\n2714\n2715\n2716\n2717\n2718\n2719\n2720\n2721\n2722\n2723\n2724\n2725\n2726\n2727\n2728\n2729\n2730\n2731\n2732\n2733\n2734\n2735\n2736\n2737\n2738\n2739\n2740\n2741\n2742\n2743\n2744\n2745\n2746\n2747\n2748\n2749\n2750\n2751\n2752\n2753\n2754\n2755\n2756\n2757\n2758\n2759\n2760\n2761\n2762\n2763\n2764\n2765\n2766\n2767\n2768\n2769\n2770\n2771\n2772\n2773\n2774\n2775\n2776\n2777\n2778\n2779\n2780\n2781\n2782\n2783\n2784\n2785\n2786\n2787\n2788\n2789\n2790\n2791\n2792\n2793\n2794\n2795\n2796\n2797\n2798\n2799\n2800\n2801\n2802\n2803\n2804\n2805\n2806\n2807\n2808\n2809\n2810\n2811\n2812\n2813\n2814\n2815\n2816\n2817\n2818\n2819\n2820\n2821\n2822\n2823\n2824\n2825\n2826\n2827\n2828\n2829\n2830\n2831\n2832\n2833\n2834\n2835\n2836\n2837\n2838\n2839\n2840\n2841\n2842\n2843\n2844\n2845\n2846\n2847\n2848\n2849\n2850\n2851\n2852\n2853\n2854\n2855\n2856\n2857\n2858\n2859\n2860\n2861\n2862\n2863\n2864\n2865\n2866\n2867\n2868\n2869\n2870\n2871\n2872\n2873\n2874\n2875\n2876\n2877\n2878\n2879\n2880\n2881\n2882\n2883\n2884\n2885\n2886\n2887\n2888\n2889\n2890\n2891\n2892\n2893\n2894\n2895\n2896\n2897\n2898\n2899\n2900\n2901\n2902\n2903\n2904\n2905\n2906\n2907\n2908\n2909\n2910\n2911\n2912\n2913\n2914\n2915\n2916\n2917\n2918\n2919\n2920\n2921\n2922\n2923\n2924\n2925\n2926\n2927\n2928\n2929\n2930\n2931\n2932\n2933\n2934\n2935\n2936\n2937\n2938\n2939\n" ], [ "np.save(\"drive/My Drive/MS_data/Y_Manual_2_new.npy\",p)", "_____no_output_____" ], [ "import numpy as np\n\nP=np.load(\"drive/My Drive/MS_data/Y_Manual_2_new.npy\")", "_____no_output_____" ], [ "Y=p", "_____no_output_____" ], [ "Y=list(Y)", "_____no_output_____" ], [ "dummy_y=[]\nd_temp_1=[]\nd_temp_2=[]\nfor i in range(len(Y[0])):\n d_temp_1=[]\n for j in range(len(Y[0][0])):\n d_temp_2=[]\n for k in range(len(Y[0][0][0])):\n d_temp_2.append(0.)\n d_temp_1.append(d_temp_2)\n dummy_y.append(d_temp_1)\nY_dummy=np.array(dummy_y, dtype='float32')\n\n\n\n ", "_____no_output_____" ], [ "X=np.load(\"drive/My Drive/MS_data/X_intersection_new.npy\")", "_____no_output_____" ], [ "X=list(X)", "_____no_output_____" ], [ "dummy_x=[]\nd_temp_1=[]\nd_temp_2=[]\nfor i in range(len(X[0])):\n d_temp_1=[]\n for j in range(len(X[0][0])):\n d_temp_2=[]\n for k in range(len(X[0][0][0])):\n d_temp_2.append(0.)\n d_temp_1.append(d_temp_2)\n dummy_x.append(d_temp_1)\nX_dummy=np.array(dummy_x, dtype='float32')", "_____no_output_____" ], [ "import numpy as np\n\n#Y=np.load(\"drive/My Drive/MS_data/Y_Manual_2_new.npy\")\nX=np.load(\"drive/My Drive/MS_data/X_intersection_new.npy\")", "_____no_output_____" ], [ "X=list(X)", "_____no_output_____" ], [ "extra=[0,1,6,4,6,0,5,0,4,3,1,3,3,0]", "_____no_output_____" ], [ "ext=np.array(extra,dtype='float32')", "_____no_output_____" ], [ "r=np.load(\"drive/My Drive/MS_data/intersection_new_rec_after_removal.npy\")", "_____no_output_____" ], [ "rec=list(r)", "_____no_output_____" ], [ "j=0\nindex=0\nfor i in extra:\n k=0\n \n print(i)\n index=index+rec[j]\n j+=1\n while(k<i):\n index=index+k\n print(index)\n index=int(index)\n Y.insert(index,Y_dummy)\n X.insert(index,X_dummy)\n\n k+=1", "0\n1\n279.0\n6\n561.0\n562\n564\n567\n571\n576\n4\n844.0\n845\n847\n850\n6\n988.0\n989\n991\n994\n998\n1003\n0\n5\n1590.0\n1591\n1593\n1596\n1600\n0\n4\n2036.0\n2037\n2039\n2042\n3\n2183.0\n2184\n2186\n1\n2377.0\n3\n2598.0\n2599\n2601\n3\n2782.0\n2783\n2785\n0\n" ], [ "rec=list(r)\nX=list(X)\nY=list(Y)", "_____no_output_____" ], [ "X=np.array(X,dtype='float32')\nY=np.array(Y,dtype='float32')\n", "_____no_output_____" ], [ "np.save(\"drive/My Drive/MS_data/Actual_X.npy\",X)\nnp.save(\"drive/My Drive/MS_data/Actual_Y.npy\",Y)", "_____no_output_____" ], [ "import numpy as np\n\nY=np.load(\"drive/My Drive/MS_data/Actual_X.npy\")\nX=np.load(\"drive/My Drive/MS_data/Actual_Y.npy\")", "_____no_output_____" ], [ "X_1=[]\nY_1=[]", "_____no_output_____" ], [ "X=list(X)\nY=list(Y)", "_____no_output_____" ], [ "index_x_p=0\nlength=8\nindex_x=length\nindex_y_p=0\nindex_y=length\nx_temp=[]\ny_temp=[]\n\n", "_____no_output_____" ], [ "while index_x<=len(X) and index_y<=len(Y):\n x_temp=X[index_x_p:index_x]\n y_temp=Y[index_y_p:index_y]\n x_temp_1=np.array(x_temp,dtype='float32')\n y_temp_1=np.array(y_temp,dtype='float32')\n X_1.append(x_temp_1)\n Y_1.append(y_temp_1)\n index_y_p=index_y_p+length\n index_x_p=index_x_p+length\n index_x=index_x+length\n index_y=index_y+length\n", "_____no_output_____" ], [ "X_1=np.array(X_1,dtype='float32')\nY_1=np.array(Y_1,dtype='float32')\n", "_____no_output_____" ], [ "np.save(\"drive/My Drive/MS_data/train_X.npy\",X_1)\nnp.save(\"drive/My Drive/MS_data/train_Y.npy\",Y_1)", "_____no_output_____" ], [ "import numpy as np\n\nX_1=np.load(\"drive/My Drive/MS_data/train_X.npy\")\nY_1=np.load(\"drive/My Drive/MS_data/train_Y.npy\")", "_____no_output_____" ] ], [ [ "## creating train test and validation sets", "_____no_output_____" ] ], [ [ "extra=[0,1,6,4,6,0,5,0,4,3,1,3,3,0]", "_____no_output_____" ], [ "r=np.load(\"drive/My Drive/MS_data/intersection_new_rec_after_removal.npy\")", "_____no_output_____" ], [ "rec=list(r)", "_____no_output_____" ], [ "j=0\nfinal=[]\nwhile j<len(extra):\n leng=rec[j]+extra[j]\n final.append(leng/8)\n j+=1\n\nnp.save(\"drive/My Drive/MS_data/intersection_final_size.npy\",final)", "_____no_output_____" ], [ "r=np.load(\"drive/My Drive/MS_data/intersection_final_size.npy\")", "_____no_output_____" ], [ "rec=list(r)", "_____no_output_____" ], [ "import json\n\ni=0\nfile_w=\"drive/My Drive/MS_data/per_paitent_records.json\"\nprev=0\ntotal=0\ndict_k={}\nwhile i<len(r):\n current=r[i]\n j=i+1\n dict_k[str(j)]={}\n dict_k[str(j)][\"Starting\"]=prev\n dict_k[str(j)][\"Ending\"]=total+current-1\n total+=current\n prev=total\n i+=1\n\nwith open(file_w, \"w\") as outfile:\n\t\t\tjson.dump(dict_k, outfile) \n", "_____no_output_____" ], [ "import random\nimport json\ndict_l={}\nlist_val=[]\nlist_train=[]\nfile_w=\"drive/My Drive/MS_data/set_divisions.json\"\nlist_test=[]\nfor i in range(5):\n patient_list=['1','2','3','4','5','6','7','8','9','10','11','12','13','14']\n for j in range(7):\n flag=0\n while(flag!=1):\n num=random.randint(1,14)\n if patient_list.count(str(num))> 0:\n list_train.append(str(num))\n patient_list.remove(str(num))\n flag=1\n if dict_l.get(\"train\") is None:\n dict_l[\"train\"]=[list_train]\n else:\n dict_l[\"train\"].append(list_train)\n list_train=[]\n for j in range(3):\n flag=0\n while(flag!=1):\n num=random.randint(1,14)\n if patient_list.count(str(num))> 0:\n list_val.append(str(num))\n patient_list.remove(str(num))\n flag=1\n if dict_l.get(\"validate\") is None:\n dict_l[\"validate\"]=[list_val]\n else:\n dict_l[\"validate\"].append(list_val)\n list_val=[]\n for j in range(4):\n flag=0\n while(flag!=1):\n num=random.randint(1,14)\n if patient_list.count(str(num))> 0:\n list_test.append(str(num))\n patient_list.remove(str(num))\n flag=1\n if dict_l.get(\"test\") is None:\n dict_l[\"test\"]=[list_test]\n else:\n dict_l[\"test\"].append(list_test)\n list_test=[] \n\nwith open(file_w, \"w\") as outfile:\n\t\t\tjson.dump(dict_l, outfile) \n\n ", "_____no_output_____" ], [ "import pandas as pd\nimport numpy as np\nimport json\nimport os\n\nY=np.load(\"drive/My Drive/MS_data/train_Y.npy\")\nprint(len(Y))\nX=np.load(\"drive/My Drive/MS_data/train_X.npy\")\nprint(len(X))\nf1=open(\"drive/My Drive/MS_data/per_paitent_records.json\")\ndata_1=json.load(f1)\nf2=open(\"drive/My Drive/MS_data/set_divisions.json\")\ndata_2=json.load(f2)\ni=1\nwhile i<=5:\n file_1=\"drive/My Drive/MS_data/\"+str(i)\n print(file_1)\n os.mkdir(file_1)\n train_list=(data_2['train'])[i-1]\n file_1_ext=file_1+'/train'\n print(file_1_ext)\n os.mkdir(file_1_ext)\n X_temp=[]\n Y_temp=[]\n for j in train_list:\n st_index=data_1[str(j)][\"Starting\"]\n end_index=data_1[str(j)][\"Ending\"]\n X_temp.append(X[int(st_index):int(end_index)])\n Y_temp.append(Y[int(st_index):int(end_index)])\n file_1_X=file_1_ext+'/'+'X.npy'\n file_1_Y=file_1_ext+'/'+'Y.npy'\n X_temp_np=np.concatenate((X_temp[0],X_temp[1],X_temp[2],X_temp[3],X_temp[4],X_temp[5],X_temp[6]),axis=0)\n Y_temp_np=np.concatenate((Y_temp[0],Y_temp[1],Y_temp[2],Y_temp[3],Y_temp[4],Y_temp[5],Y_temp[6]),axis=0)\n print(len(X_temp_np))\n print(len(Y_temp_np))\n np.save(file_1_X,X_temp_np)\n np.save(file_1_Y,Y_temp_np)\n train_list=(data_2['validate'])[i-1]\n file_1_ext=file_1+'/validate'\n os.mkdir(file_1_ext)\n X_temp=[]\n Y_temp=[]\n for j in train_list:\n st_index=data_1[str(j)][\"Starting\"]\n end_index=data_1[str(j)][\"Ending\"]\n X_temp.append(X[int(st_index):int(end_index)])\n Y_temp.append(Y[int(st_index):int(end_index)])\n file_1_X=file_1_ext+'/'+'X.npy'\n file_1_Y=file_1_ext+'/'+'Y.npy'\n X_temp_np=np.concatenate((X_temp[0],X_temp[1],X_temp[2]),axis=0)\n Y_temp_np=np.concatenate((Y_temp[0],Y_temp[1],Y_temp[2]),axis=0)\n print(len(X_temp_np))\n print(len(Y_temp_np))\n np.save(file_1_X,X_temp_np)\n np.save(file_1_Y,Y_temp_np)\n train_list=(data_2['test'])[i-1]\n file_1_ext=file_1+'/test'\n os.mkdir(file_1_ext)\n X_temp=[]\n Y_temp=[]\n for j in train_list:\n st_index=data_1[str(j)][\"Starting\"]\n end_index=data_1[str(j)][\"Ending\"]\n X_temp.append(X[int(st_index):int(end_index)])\n Y_temp.append(Y[int(st_index):int(end_index)])\n file_1_X=file_1_ext+'/'+'X.npy'\n file_1_Y=file_1_ext+'/'+'Y.npy'\n X_temp_np=np.concatenate((X_temp[0],X_temp[1],X_temp[2],X_temp[3]),axis=0)\n Y_temp_np=np.concatenate((Y_temp[0],Y_temp[1],Y_temp[2],Y_temp[3]),axis=0)\n print(len(X_temp_np))\n print(len(Y_temp_np))\n np.save(file_1_X,X_temp_np)\n np.save(file_1_Y,Y_temp_np)\n i+=1\n\n", "372\n372\ndrive/My Drive/MS_data/1\ndrive/My Drive/MS_data/1/train\n198\n198\n76\n76\n84\n84\ndrive/My Drive/MS_data/2\ndrive/My Drive/MS_data/2/train\n201\n201\n60\n60\n97\n97\ndrive/My Drive/MS_data/3\ndrive/My Drive/MS_data/3/train\n182\n182\n51\n51\n125\n125\ndrive/My Drive/MS_data/4\ndrive/My Drive/MS_data/4/train\n179\n179\n89\n89\n90\n90\ndrive/My Drive/MS_data/5\ndrive/My Drive/MS_data/5/train\n157\n157\n78\n78\n123\n123\n" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code" ]
[ [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code", "code", "code", "code", "code" ] ]
ece86cca6e6d2e73327141181a1a88dcca87f90c
552,853
ipynb
Jupyter Notebook
1. Load and Visualize Data.ipynb
joeysantana3/udacity-nano-degree
6fe322d4933f093be2dd48dbedc2326aeb25782b
[ "MIT" ]
null
null
null
1. Load and Visualize Data.ipynb
joeysantana3/udacity-nano-degree
6fe322d4933f093be2dd48dbedc2326aeb25782b
[ "MIT" ]
null
null
null
1. Load and Visualize Data.ipynb
joeysantana3/udacity-nano-degree
6fe322d4933f093be2dd48dbedc2326aeb25782b
[ "MIT" ]
null
null
null
845.340979
160,488
0.948397
[ [ [ "# Facial Keypoint Detection\n \nThis project will be all about defining and training a convolutional neural network to perform facial keypoint detection and using computer vision techniques to transform images of faces. The first step in any challenge like this will be to load and visualize the data you'll be working with. \n\nLet's take a look at some examples of images and corresponding facial keypoints.\n\n<img src='images/key_pts_example.png' width=50% height=50%/>\n\nFacial keypoints (also called facial landmarks) are the small magenta dots shown on each of the faces in the image above. In each training and test image, there is a single face and **68 keypoints, with coordinates (x, y), for that face**. These keypoints mark important areas of the face: the eyes, corners of the mouth, the nose, etc. These keypoints are relevant for a variety of tasks, such as face filters, emotion recognition, pose recognition, and so on. Here they are, numbered, and you can see that specific ranges of points match different portions of the face.\n\n<img src='images/landmarks_numbered.jpg' width=30% height=30%/>\n\n---", "_____no_output_____" ], [ "## Load and Visualize Data\n\nThe first step in working with any dataset is to become familiar with your data; you'll need to load in the images of faces and their keypoints and visualize them! This set of image data has been extracted from the [YouTube Faces Dataset](https://www.cs.tau.ac.il/~wolf/ytfaces/), which includes videos of people in YouTube videos. These videos have been fed through some processing steps and turned into sets of image frames containing one face and the associated keypoints.\n\n#### Training and Testing Data\n\nThis facial keypoints dataset consists of 5770 color images. All of these images are separated into either a training or a test set of data.\n\n* 3462 of these images are training images, for you to use as you create a model to predict keypoints.\n* 2308 are test images, which will be used to test the accuracy of your model.\n\nThe information about the images and keypoints in this dataset are summarized in CSV files, which we can read in using `pandas`. Let's read the training CSV and get the annotations in an (N, 2) array where N is the number of keypoints and 2 is the dimension of the keypoint coordinates (x, y).\n\n---", "_____no_output_____" ] ], [ [ "# import the required libraries\nimport glob\nimport os\nimport numpy as np\nimport pandas as pd\nimport matplotlib.pyplot as plt\nimport matplotlib.image as mpimg\n\nimport cv2", "_____no_output_____" ], [ "key_pts_frame = pd.read_csv('data/training_frames_keypoints.csv')\n\nn = 0\nimage_name = key_pts_frame.iloc[n, 0]\nkey_pts = key_pts_frame.iloc[n, 1:].as_matrix()\nkey_pts = key_pts.astype('float').reshape(-1, 2)\n\nprint('Image name: ', image_name)\nprint('Landmarks shape: ', key_pts.shape)\nprint('First 4 key pts: {}'.format(key_pts[:4]))", "_____no_output_____" ], [ "# print out some stats about the data\nprint('Number of images: ', key_pts_frame.shape[0])", "_____no_output_____" ] ], [ [ "## Look at some images\n\nBelow, is a function `show_keypoints` that takes in an image and keypoints and displays them. As you look at this data, **note that these images are not all of the same size**, and neither are the faces! To eventually train a neural network on these images, we'll need to standardize their shape.", "_____no_output_____" ] ], [ [ "def show_keypoints(image, key_pts):\n \"\"\"Show image with keypoints\"\"\"\n plt.imshow(image)\n plt.scatter(key_pts[:, 0], key_pts[:, 1], s=20, marker='.', c='m')\n", "_____no_output_____" ], [ "# Display a few different types of images by changing the index n\n\n# select an image by index in our data frame\nn = 0\nimage_name = key_pts_frame.iloc[n, 0]\nkey_pts = key_pts_frame.iloc[n, 1:].values\nkey_pts = key_pts.astype('float').reshape(-1, 2)\n\nplt.figure(figsize=(5, 5))\nshow_keypoints(mpimg.imread(os.path.join('data/training/', image_name)), key_pts)\nplt.show()", "_____no_output_____" ] ], [ [ "## Dataset class and Transformations\n\nTo prepare our data for training, we'll be using PyTorch's Dataset class. Much of this this code is a modified version of what can be found in the [PyTorch data loading tutorial](http://pytorch.org/tutorials/beginner/data_loading_tutorial.html).\n\n#### Dataset class\n\n``torch.utils.data.Dataset`` is an abstract class representing a\ndataset. This class will allow us to load batches of image/keypoint data, and uniformly apply transformations to our data, such as rescaling and normalizing images for training a neural network.\n\n\nYour custom dataset should inherit ``Dataset`` and override the following\nmethods:\n\n- ``__len__`` so that ``len(dataset)`` returns the size of the dataset.\n- ``__getitem__`` to support the indexing such that ``dataset[i]`` can\n be used to get the i-th sample of image/keypoint data.\n\nLet's create a dataset class for our face keypoints dataset. We will\nread the CSV file in ``__init__`` but leave the reading of images to\n``__getitem__``. This is memory efficient because all the images are not\nstored in the memory at once but read as required.\n\nA sample of our dataset will be a dictionary\n``{'image': image, 'keypoints': key_pts}``. Our dataset will take an\noptional argument ``transform`` so that any required processing can be\napplied on the sample. We will see the usefulness of ``transform`` in the\nnext section.\n", "_____no_output_____" ] ], [ [ "from torch.utils.data import Dataset, DataLoader\n\nclass FacialKeypointsDataset(Dataset):\n \"\"\"Face Landmarks dataset.\"\"\"\n\n def __init__(self, csv_file, root_dir, transform=None):\n \"\"\"\n Args:\n csv_file (string): Path to the csv file with annotations.\n root_dir (string): Directory with all the images.\n transform (callable, optional): Optional transform to be applied\n on a sample.\n \"\"\"\n self.key_pts_frame = pd.read_csv(csv_file)\n self.root_dir = root_dir\n self.transform = transform\n\n def __len__(self):\n return len(self.key_pts_frame)\n\n def __getitem__(self, idx):\n image_name = os.path.join(self.root_dir,\n self.key_pts_frame.iloc[idx, 0])\n \n image = mpimg.imread(image_name)\n \n # if image has an alpha color channel, get rid of it\n if(image.shape[2] == 4):\n image = image[:,:,0:3]\n \n key_pts = self.key_pts_frame.iloc[idx, 1:].values\n key_pts = key_pts.astype('float').reshape(-1, 2)\n sample = {'image': image, 'keypoints': key_pts}\n\n if self.transform:\n sample = self.transform(sample)\n\n return sample", "_____no_output_____" ] ], [ [ "Now that we've defined this class, let's instantiate the dataset and display some images.", "_____no_output_____" ] ], [ [ "# Construct the dataset\nface_dataset = FacialKeypointsDataset(csv_file='data/training_frames_keypoints.csv',\n root_dir='data/training/')\n\n# print some stats about the dataset\nprint('Length of dataset: ', len(face_dataset))", "Length of dataset: 3462\n" ], [ "# Display a few of the images from the dataset\nnum_to_display = 3\n\nfor i in range(num_to_display):\n \n # define the size of images\n fig = plt.figure(figsize=(20,10))\n \n # randomly select a sample\n rand_i = np.random.randint(0, len(face_dataset))\n sample = face_dataset[rand_i]\n\n # print the shape of the image and keypoints\n print(i, sample['image'].shape, sample['keypoints'].shape)\n\n ax = plt.subplot(1, num_to_display, i + 1)\n ax.set_title('Sample #{}'.format(i))\n \n # Using the same display function, defined earlier\n show_keypoints(sample['image'], sample['keypoints'])\n", "0 (187, 164, 3) (68, 2)\n1 (169, 211, 3) (68, 2)\n2 (123, 101, 3) (68, 2)\n" ] ], [ [ "## Transforms\n\nNow, the images above are not of the same size, and neural networks often expect images that are standardized; a fixed size, with a normalized range for color ranges and coordinates, and (for PyTorch) converted from numpy lists and arrays to Tensors.\n\nTherefore, we will need to write some pre-processing code.\nLet's create four transforms:\n\n- ``Normalize``: to convert a color image to grayscale values with a range of [0,1] and normalize the keypoints to be in a range of about [-1, 1]\n- ``Rescale``: to rescale an image to a desired size.\n- ``RandomCrop``: to crop an image randomly.\n- ``ToTensor``: to convert numpy images to torch images.\n\n\nWe will write them as callable classes instead of simple functions so\nthat parameters of the transform need not be passed every time it's\ncalled. For this, we just need to implement ``__call__`` method and \n(if we require parameters to be passed in), the ``__init__`` method. \nWe can then use a transform like this:\n\n tx = Transform(params)\n transformed_sample = tx(sample)\n\nObserve below how these transforms are generally applied to both the image and its keypoints.\n\n", "_____no_output_____" ] ], [ [ "import torch\nfrom torchvision import transforms, utils\n# tranforms\n\nclass Normalize(object):\n \"\"\"Convert a color image to grayscale and normalize the color range to [0,1].\"\"\" \n\n def __call__(self, sample):\n image, key_pts = sample['image'], sample['keypoints']\n \n image_copy = np.copy(image)\n key_pts_copy = np.copy(key_pts)\n\n # convert image to grayscale\n image_copy = cv2.cvtColor(image, cv2.COLOR_RGB2GRAY)\n \n # scale color range from [0, 255] to [0, 1]\n image_copy= image_copy/255.0\n \n # scale keypoints to be centered around 0 with a range of [-1, 1]\n # mean = 100, sqrt = 50, so, pts should be (pts - 100)/50\n key_pts_copy = (key_pts_copy - 100)/50.0\n\n\n return {'image': image_copy, 'keypoints': key_pts_copy}\n\n\nclass Rescale(object):\n \"\"\"Rescale the image in a sample to a given size.\n\n Args:\n output_size (tuple or int): Desired output size. If tuple, output is\n matched to output_size. If int, smaller of image edges is matched\n to output_size keeping aspect ratio the same.\n \"\"\"\n\n def __init__(self, output_size):\n assert isinstance(output_size, (int, tuple))\n self.output_size = output_size\n\n def __call__(self, sample):\n image, key_pts = sample['image'], sample['keypoints']\n\n h, w = image.shape[:2]\n if isinstance(self.output_size, int):\n if h > w:\n new_h, new_w = self.output_size * h / w, self.output_size\n else:\n new_h, new_w = self.output_size, self.output_size * w / h\n else:\n new_h, new_w = self.output_size\n\n new_h, new_w = int(new_h), int(new_w)\n\n img = cv2.resize(image, (new_w, new_h))\n \n # scale the pts, too\n key_pts = key_pts * [new_w / w, new_h / h]\n\n return {'image': img, 'keypoints': key_pts}\n\n\nclass RandomCrop(object):\n \"\"\"Crop randomly the image in a sample.\n\n Args:\n output_size (tuple or int): Desired output size. If int, square crop\n is made.\n \"\"\"\n\n def __init__(self, output_size):\n assert isinstance(output_size, (int, tuple))\n if isinstance(output_size, int):\n self.output_size = (output_size, output_size)\n else:\n assert len(output_size) == 2\n self.output_size = output_size\n\n def __call__(self, sample):\n image, key_pts = sample['image'], sample['keypoints']\n\n h, w = image.shape[:2]\n new_h, new_w = self.output_size\n\n top = np.random.randint(0, h - new_h)\n left = np.random.randint(0, w - new_w)\n\n image = image[top: top + new_h,\n left: left + new_w]\n\n key_pts = key_pts - [left, top]\n\n return {'image': image, 'keypoints': key_pts}\n\n\nclass ToTensor(object):\n \"\"\"Convert ndarrays in sample to Tensors.\"\"\"\n\n def __call__(self, sample):\n image, key_pts = sample['image'], sample['keypoints']\n \n # if image has no grayscale color channel, add one\n if(len(image.shape) == 2):\n # add that third color dim\n image = image.reshape(image.shape[0], image.shape[1], 1)\n \n # swap color axis because\n # numpy image: H x W x C\n # torch image: C X H X W\n image = image.transpose((2, 0, 1))\n \n return {'image': torch.from_numpy(image),\n 'keypoints': torch.from_numpy(key_pts)}", "_____no_output_____" ] ], [ [ "## Test out the transforms\n\nLet's test these transforms out to make sure they behave as expected. As you look at each transform, note that, in this case, **order does matter**. For example, you cannot crop an image using a value smaller than the original image (and the orginal images vary in size!), but, if you first rescale the original image, you can then crop it to any size smaller than the rescaled size.", "_____no_output_____" ] ], [ [ "# test out some of these transforms\nrescale = Rescale(100)\ncrop = RandomCrop(50)\ncomposed = transforms.Compose([Rescale(250),\n RandomCrop(224)])\n\n# apply the transforms to a sample image\ntest_num = 400\nsample = face_dataset[test_num]\n\nfig = plt.figure()\nfor i, tx in enumerate([rescale, crop, composed]):\n transformed_sample = tx(sample)\n\n ax = plt.subplot(1, 3, i + 1)\n plt.tight_layout()\n ax.set_title(type(tx).__name__)\n show_keypoints(transformed_sample['image'], transformed_sample['keypoints'])\n\nplt.show()", "_____no_output_____" ] ], [ [ "## Create the transformed dataset\n\nApply the transforms in order to get grayscale images of the same shape. Verify that your transform works by printing out the shape of the resulting data (printing out a few examples should show you a consistent tensor size).", "_____no_output_____" ] ], [ [ "# define the data tranform\n# order matters! i.e. rescaling should come before a smaller crop\ndata_transform = transforms.Compose([Rescale(250),\n RandomCrop(224),\n Normalize(),\n ToTensor()])\n\n# create the transformed dataset\ntransformed_dataset = FacialKeypointsDataset(csv_file='data/training_frames_keypoints.csv',\n root_dir='data/training/',\n transform=data_transform)\n", "_____no_output_____" ], [ "# print some stats about the transformed data\nprint('Number of images: ', len(transformed_dataset))\n\n# make sure the sample tensors are the expected size\nfor i in range(5):\n sample = transformed_dataset[i]\n print(i, sample['image'].size(), sample['keypoints'].size())\n", "Number of images: 3462\n0 torch.Size([1, 224, 224]) torch.Size([68, 2])\n1 torch.Size([1, 224, 224]) torch.Size([68, 2])\n2 torch.Size([1, 224, 224]) torch.Size([68, 2])\n3 torch.Size([1, 224, 224]) torch.Size([68, 2])\n4 torch.Size([1, 224, 224]) torch.Size([68, 2])\n" ] ], [ [ "## Data Iteration and Batching\n\nRight now, we are iterating over this data using a ``for`` loop, but we are missing out on a lot of PyTorch's dataset capabilities, specifically the abilities to:\n\n- Batch the data\n- Shuffle the data\n- Load the data in parallel using ``multiprocessing`` workers.\n\n``torch.utils.data.DataLoader`` is an iterator which provides all these\nfeatures, and we'll see this in use in the *next* notebook, Notebook 2, when we load data in batches to train a neural network!\n\n---\n\n", "_____no_output_____" ], [ "## Ready to Train!\n\nNow that you've seen how to load and transform our data, you're ready to build a neural network to train on this data.\n\nIn the next notebook, you'll be tasked with creating a CNN for facial keypoint detection.", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown" ]
[ [ "markdown", "markdown" ], [ "code", "code", "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown", "markdown" ] ]
ece875388faa615033008ec80227fb389f8f5531
68,272
ipynb
Jupyter Notebook
notebooks/generate_observations.ipynb
natbolon/weather_prediction
f3a18264e5e642cf366e7d181e6e081ab7f21eab
[ "MIT" ]
2
2020-07-22T15:07:25.000Z
2020-09-22T12:32:51.000Z
notebooks/generate_observations.ipynb
natbolon/weather_prediction
f3a18264e5e642cf366e7d181e6e081ab7f21eab
[ "MIT" ]
null
null
null
notebooks/generate_observations.ipynb
natbolon/weather_prediction
f3a18264e5e642cf366e7d181e6e081ab7f21eab
[ "MIT" ]
2
2020-09-27T20:20:03.000Z
2021-01-15T12:15:21.000Z
48.248763
4,193
0.512992
[ [ [ "## Generate data ", "_____no_output_____" ] ], [ [ "import sys\nsys.path.append('/'.join(sys.path[0].split('/')[:-1]))\n\nimport os\nimport xarray as xr\nimport numpy as np\nimport time\nimport matplotlib.pyplot as plt\nimport healpy as hp\n\nimport torch\nfrom torch import nn, optim\nfrom torch.utils.data import Dataset, DataLoader\n", "_____no_output_____" ], [ "datadir = \"../data/healpix/5.625deg_nearest/\"\n\nlr=1e-4\ndr=0\nbatch_size=128\npatience=3\ntrain_years=('1979', '2015')\nvalid_years=('2016', '2016')\ntest_years=('2017', '2018')\ngpu=1\niterative=False\n\nvars = ['z', 't']\nkernel_size = 5", "_____no_output_____" ], [ "z = xr.open_mfdataset(f'{datadir}geopotential_500/*.nc', combine='by_coords')\nt = xr.open_mfdataset(f'{datadir}temperature_850/*.nc', combine='by_coords')\nds = xr.merge([z, t], compat='override')", "_____no_output_____" ], [ "def create_iterative_observations_healpix(ds, lead_time, max_lead_time, nb_timesteps, test_years, nodes):\n \n lead_times = np.arange(lead_time, max_lead_time + lead_time, lead_time)\n\n data = ds.to_array(dim='level', name='Dataset').transpose('time', 'node', 'level')\n n_samples = data.isel(time=slice(0, -nb_timesteps*lead_time)).shape[0] - max_lead_time\n\n obs_list = []\n \n print('Generating observations list...')\n for lead in lead_times:\n obs_list.append(data.isel(time=slice(lead, lead + n_samples)).isel(level=slice(0, 2)))\n\n #observations_numpy = np.array(obs_list)\n observations_joint = xr.concat(obs_list, dim='lead_time')\n #return observations_numpy\n print('Obtaining coordinates...')\n # Lat lon coordinates\n nside = int(np.sqrt(nodes/12))\n out_lon, out_lat = hp.pix2ang(nside, np.arange(nodes), lonlat=True)\n \n print('Generate set of times to study', end='\\n')\n # Actual times\n start = np.datetime64(test_years[0], 'h') + np.timedelta64(lead_time, 'h')\n stop = start + np.timedelta64(n_samples, 'h')\n times = np.arange(start, stop)\n\n # Variables\n var_dict_out = {var: None for var in ['z', 't']}\n\n das = [];\n lev_idx = 0\n \n for var, levels in var_dict_out.items():\n das.append(observations_joint.isel(level=lev_idx).rename(var))\n lev_idx +=1\n \n \"\"\"\n \n \n if levels is None: \n das.append(xr.DataArray(\n observations_numpy[:, :, :, lev_idx],\n dims=['lead_time', 'time', 'node'],\n coords={'lead_time': lead_times, 'time': times, 'node': np.arange(nodes)},\n name=var\n ))\n lev_idx += 1\n # never this case...\n else:\n nlevs = len(levels)\n das.append(xr.DataArray(\n observations_numpy[:, :, :, lev_idx:lev_idx+nlevs],\n dims=['lead_time', 'time', 'node', 'level'],\n coords={'lead_time': lead_times, 'time': valid_time, 'node': nodes, 'level': nlevs},\n name=var\n ))\n lev_idx += nlevs\n \"\"\"\n \n print('\\nGenerate observation...')\n observation_ds = xr.merge(das, compat='override').reset_coords(names='level', drop=True)\n observation_ds = observation_ds.assign_coords({'lat': out_lat, 'lon': out_lon})\n return observation_ds\n", "_____no_output_____" ], [ "nodes = 12*16*16\nmax_lead_time = 5*24\nlead_time = 6\nout_features = 2\nnb_timesteps = 2\nnside = int(np.sqrt(nodes/12))", "_____no_output_____" ], [ "#obs = create_iterative_observations_hp(\"../data/equiangular/5.625deg/\", test_years, lead_time, max_lead_time, nside, nb_timesteps=2)", "_____no_output_____" ], [ "obs = create_iterative_observations_healpix(ds, lead_time, max_lead_time, nb_timesteps, test_years, nodes)", "Generating observations list...\nObtaining coordinates...\nGenerate set of times to study\n\nGenerate observation...\n" ], [ "obs", "_____no_output_____" ], [ "obs.to_netcdf(datadir + 'observations.nc')", "_____no_output_____" ] ], [ [ "### Use new functions from Iciar", "_____no_output_____" ] ], [ [ "from modules.test import create_iterative_observations_hp", "_____no_output_____" ], [ "nodes = 12*16*16\nmax_lead_time = 5*24\nlead_time = 6\nout_features = 2\nnb_timesteps = 2\nnside = int(np.sqrt(nodes/12))", "_____no_output_____" ], [ "def create_iterative_observations_hp(input_dir, test_years, lead_time, max_lead_time, nside, nb_timesteps=2):\n z500 = xr.open_mfdataset(f'{input_dir}geopotential_500/*.nc', combine='by_coords').sel(time=slice(*test_years))\n t850 = xr.open_mfdataset(f'{input_dir}temperature_850/*.nc', combine='by_coords').sel(time=slice(*test_years))\n\n test_data = xr.merge([z500, t850], compat='override')\n\n\n n_samples = test_data.isel(time=slice(0, -nb_timesteps*lead_time)).dims['time'] - max_lead_time\n nb_iter = max_lead_time // lead_time\n n_pixels = 12*(nside**2)\n print(n_samples)\n\n # Lead times\n lead_times = np.arange(lead_time, max_lead_time + lead_time, lead_time)\n\n # Lat lon coordinates\n out_lon, out_lat = hp.pix2ang(nside, np.arange(n_pixels), lonlat=True)\n\n # Actual times\n start = np.datetime64(test_years[0], 'h') + np.timedelta64(lead_time, 'h')\n stop = start + np.timedelta64(n_samples, 'h')\n times = np.arange(start, stop, 1)\n\n # Variables\n data_vars = ['z', 't']\n var_dict_out = {var: None for var in data_vars}\n \n data = np.zeros((2, nb_iter, n_samples, 3072))\n \n #data = np.zeros((2, nb_iter, n_samples, 32, 64))\n for i in range(nb_iter):\n #data[0, i, :, :, :] = test_data.z.isel(time=slice(lead_time*(i+1), lead_time*(i+1) + n_samples)).values\n #data[1, i, :, :, :] = test_data.t.isel(time=slice(lead_time*(i+1), lead_time*(i+1) + n_samples)).values\n \n data[0, i, :, :] = test_data.z.isel(time=slice(lead_time*(i+1), lead_time*(i+1) + n_samples)).values\n data[1, i, :, :] = test_data.t.isel(time=slice(lead_time*(i+1), lead_time*(i+1) + n_samples)).values\n\n\n das = [];\n lev_idx = 0\n for var in data_vars: \n das.append(xr.DataArray(\n data[lev_idx, :, :, :],\n dims=['lead_time', 'time', 'node'],\n coords={'lead_time': lead_times, 'time': times, 'node': np.arange(n_pixels)},\n name=var\n ))\n lev_idx += 1\n observations = xr.merge(das)\n #observations = observations.assign_coords({'lat': out_lat, 'lon': out_lon})\n\n return observations", "_____no_output_____" ], [ "observations = create_iterative_observations_hp('../data/healpix/5.625deg_nearest/', test_years, lead_time, max_lead_time, nside, nb_timesteps=2)", "17388\n" ], [ "observations", "_____no_output_____" ], [ "n_pixels = 12*(nside**2)\nout_lon, out_lat = hp.pix2ang(nside, np.arange(n_pixels), lonlat=True)", "_____no_output_____" ], [ "observations = observations.assign_coords({'lat': ('node', out_lat), 'lon': ('node', out_lon)})", "_____no_output_____" ], [ "observations.to_netcdf('../data/healpix/predictions/' + 'observations_nearest.nc')", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code" ]
[ [ "markdown" ], [ "code", "code", "code", "code", "code", "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code", "code", "code", "code" ] ]
ece8775709d66e5c0dc6a822c43e3eb3c3346e96
5,044
ipynb
Jupyter Notebook
src/20grid_random_walk_01.ipynb
songqsh/Computational-finance
8314838e1486c3b894726bef214f8e7694e76059
[ "MIT" ]
2
2020-02-07T23:11:44.000Z
2020-02-08T14:39:45.000Z
src/Lecture notes/20grid_random_walk_01.ipynb
JiaminJIAN/20MA573
a9bc42a769b112c9beb9b1fcafb742a768a2a18e
[ "MIT" ]
null
null
null
src/Lecture notes/20grid_random_walk_01.ipynb
JiaminJIAN/20MA573
a9bc42a769b112c9beb9b1fcafb742a768a2a18e
[ "MIT" ]
8
2020-01-21T08:39:31.000Z
2020-09-16T00:44:16.000Z
25.474747
421
0.45797
[ [ [ "# Escaping from a Gridworld \n\nWe will take n dimensional gridworld of any shape as our experimental environment. Read and understand the code in the following shell. Then proceed to next questions.", "_____no_output_____" ] ], [ [ "import random\n\nclass grid:\n def __init__(self, shape = (5, 5)):\n self.n_dim = len(shape)\n self.shape = shape\n print('>>> grid world shape is: '+str(shape))\n\n def is_interior(self,ix):\n return all([0.<a<b-1 for a,b in zip(ix,list(self.shape))])\n \n #input: lists of index\n #return: running cost, list of next index, list of probability\n def step(self, ix):\n run_cost = 0. \n ix_next = []; pr_next= []\n if self.is_interior(ix):\n run_cost = 1.\n for i in range(self.n_dim):\n ix1 = ix.copy(); ix1[i]+=1; ix_next += [ix1,]\n pr1 = 1./(self.n_dim*2.0) \n pr_next += [pr1,]\n for i in range(self.n_dim):\n ix1 = ix.copy(); ix1[i]-=1; ix_next += [ix1,]\n pr1 = 1./(self.n_dim*2.0) \n pr_next += [pr1,]\n \n return run_cost, ix_next, pr_next\n \n def step_random(self, ix):\n run_cost, ix_next, pr_next = self.step(ix)\n ix_next_rd = random.choices(ix_next, pr_next, k = 1)\n return run_cost, ix_next_rd[0]\n", "_____no_output_____" ] ], [ [ "- Explain the meaning of each output from the next shell\n- (Your answer)\n", "_____no_output_____" ] ], [ [ "\n#####check\ng1 = grid(shape=(5,5))\nix1 = [2,3]\nprint('>>>', g1.is_interior(ix1))\no1, o2, o3 = g1.step(ix1)\nprint('>>>', o1, '\\n', o2, '\\n', o3)\n\nprint('>>>>>>>>><<<<<<<<<<')\nix2 = [1,4]\nprint('>>>', g1.is_interior(ix2))\no1, o2, o3 = g1.step(ix2)\nprint('>>>', o1, '\\n', o2, '\\n', o3)\n", ">>> grid world shape is: (5, 5)\n>>> True\n>>> 1.0 \n [[3, 3], [2, 4], [1, 3], [2, 2]] \n [0.25, 0.25, 0.25, 0.25]\n>>>>>>>>><<<<<<<<<<\n>>> False\n>>> 0.0 \n [] \n []\n" ] ], [ [ "- Explain the meaning of each output from the next shell\n- (Your answer)", "_____no_output_____" ] ], [ [ "#import ipdb\n\nix = [2,2]\ntot_cost = 0.\nwhile g1.is_interior(ix):\n #ipdb.set_trace()\n run_cost, ix = g1.step_random(ix)\n print('>>>', ix)\n tot_cost+=run_cost\n \nprint('>>>', tot_cost)", ">>> [1, 2]\n>>> [1, 3]\n>>> [1, 4]\n>>> 3.0\n" ] ], [ [ "- Consider a grid world of shape (4,4). This means, the agent's state belongs to the state space of $\\{0,1,2,3,4\\}\\times \\{0,1,2,3,4\\}$. A state $(a,b)$ is called a boundary if one of its coordinate is either $0$ or $4$. Otherwise, the state is called interior. An agent moves a random walk in the grid world from initial state $(2,2)$. Use your math to find the expected number of steps to reach the boundary.", "_____no_output_____" ], [ "- (your answer)", "_____no_output_____" ], [ "- Use your code to verify your answer.", "_____no_output_____" ] ], [ [ "#your code \npass", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code" ]
[ [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown", "markdown" ], [ "code" ] ]
ece8811b32d5d11097757442db9d95a3e2a2655c
50,981
ipynb
Jupyter Notebook
src/model/pytorch/21-RL/DeepRL-Tutorials/08.Rainbow.ipynb
inessus/ai-skills
527f32d49887f06eee357c83bb6a9a21edc69bc5
[ "MIT" ]
5
2018-10-30T01:36:04.000Z
2020-11-26T02:38:09.000Z
src/model/pytorch/21-RL/DeepRL-Tutorials/08.Rainbow.ipynb
inessus/ai-skills
527f32d49887f06eee357c83bb6a9a21edc69bc5
[ "MIT" ]
4
2020-11-18T21:34:06.000Z
2022-03-11T23:32:44.000Z
src/model/pytorch/21-RL/DeepRL-Tutorials/08.Rainbow.ipynb
inessus/ai-skills
527f32d49887f06eee357c83bb6a9a21edc69bc5
[ "MIT" ]
1
2018-11-27T09:19:38.000Z
2018-11-27T09:19:38.000Z
140.057692
38,724
0.864067
[ [ [ "# Rainbow: Combining Improvements in Deep Reinforcement Learning", "_____no_output_____" ], [ "## Imports", "_____no_output_____" ] ], [ [ "import gym\nimport numpy as np\n\nimport torch\nimport torch.optim as optim\nimport torch.nn as nn\nimport torch.nn.functional as F\n\nfrom IPython.display import clear_output\nfrom matplotlib import pyplot as plt\n%matplotlib inline\n\nfrom timeit import default_timer as timer\nfrom datetime import timedelta\nimport math\n\nfrom utils.wrappers import *\nfrom utils.ReplayMemory import PrioritizedReplayMemory\nfrom networks.layers import NoisyLinear\nfrom agents.DQN import Model as DQN_Agent\n\nfrom utils.hyperparameters import Config", "_____no_output_____" ] ], [ [ "## Hyperparameters", "_____no_output_____" ] ], [ [ "config = Config()\n\nconfig.device = torch.device(\"cuda\" if torch.cuda.is_available() else \"cpu\")\n\n#Multi-step returns\nconfig.N_STEPS = 3\n\n#misc agent variables\nconfig.GAMMA=0.99\nconfig.LR=1e-4\n\n#memory\nconfig.TARGET_NET_UPDATE_FREQ = 1000\nconfig.EXP_REPLAY_SIZE = 100000\nconfig.BATCH_SIZE = 32\nconfig.PRIORITY_ALPHA=0.6\nconfig.PRIORITY_BETA_START=0.4\nconfig.PRIORITY_BETA_FRAMES = 100000\n\n#epsilon variables\nconfig.SIGMA_INIT=0.5\n\n#Learning control variables\nconfig.LEARN_START = 10000\nconfig.MAX_FRAMES=700000\n\n#Categorical Params\nconfig.ATOMS = 51\nconfig.V_MAX = 10\nconfig.V_MIN = -10", "_____no_output_____" ] ], [ [ "## Network", "_____no_output_____" ] ], [ [ "class CategoricalDuelingDQN(nn.Module):\n def __init__(self, input_shape, num_actions, sigma_init=0.5, atoms=51):\n super(CategoricalDuelingDQN, self).__init__()\n \n self.input_shape = input_shape\n self.num_actions = num_actions\n self.atoms = atoms\n\n self.conv1 = nn.Conv2d(self.input_shape[0], 32, kernel_size=8, stride=4)\n self.conv2 = nn.Conv2d(32, 64, kernel_size=4, stride=2)\n self.conv3 = nn.Conv2d(64, 64, kernel_size=3, stride=1)\n\n self.adv1 = NoisyLinear(self.feature_size(), 512, sigma_init)\n self.adv2 = NoisyLinear(512, self.num_actions*self.atoms, sigma_init)\n\n self.val1 = NoisyLinear(self.feature_size(), 512, sigma_init)\n self.val2 = NoisyLinear(512, 1*self.atoms, sigma_init)\n \n def forward(self, x):\n x = F.relu(self.conv1(x))\n x = F.relu(self.conv2(x))\n x = F.relu(self.conv3(x))\n x = x.view(x.size(0), -1)\n adv = F.relu(self.adv1(x))\n adv = self.adv2(adv).view(-1, self.num_actions, self.atoms)\n\n val = F.relu(self.val1(x))\n val = self.val2(val).view(-1, 1, self.atoms)\n\n final = val + adv - adv.mean(dim=1).view(-1, 1, self.atoms)\n\n return F.softmax(final, dim=2)\n \n def feature_size(self):\n return self.conv3(self.conv2(self.conv1(torch.zeros(1, *self.input_shape)))).view(1, -1).size(1)\n\n def sample_noise(self):\n self.adv1.sample_noise()\n self.adv2.sample_noise()\n self.val1.sample_noise()\n self.val2.sample_noise()", "_____no_output_____" ] ], [ [ "## Agent", "_____no_output_____" ] ], [ [ "class Model(DQN_Agent):\n def __init__(self, static_policy=False, env=None, config=None):\n self.atoms=config.ATOMS\n self.v_max=config.V_MAX\n self.v_min=config.V_MIN\n self.supports = torch.linspace(self.v_min, self.v_max, self.atoms).view(1, 1, self.atoms).to(config.device)\n self.delta = (self.v_max - self.v_min) / (self.atoms - 1)\n\n super(Model, self).__init__(static_policy, env, config)\n\n self.nsteps=max(self.nsteps,3)\n \n def declare_networks(self):\n self.model = CategoricalDuelingDQN(self.num_feats, self.num_actions, sigma_init=self.sigma_init, atoms=self.atoms)\n self.target_model = CategoricalDuelingDQN(self.num_feats, self.num_actions, sigma_init=self.sigma_init, atoms=self.atoms)\n \n def declare_memory(self):\n self.memory = PrioritizedReplayMemory(self.experience_replay_size, self.priority_alpha, self.priority_beta_start, self.priority_beta_frames)\n \n def projection_distribution(self, batch_vars):\n batch_state, batch_action, batch_reward, non_final_next_states, non_final_mask, empty_next_state_values, indices, weights = batch_vars\n\n with torch.no_grad():\n max_next_dist = torch.zeros((self.batch_size, 1, self.atoms), device=self.device, dtype=torch.float) + 1./self.atoms\n if not empty_next_state_values:\n max_next_action = self.get_max_next_state_action(non_final_next_states)\n self.target_model.sample_noise()\n max_next_dist[non_final_mask] = self.target_model(non_final_next_states).gather(1, max_next_action)\n max_next_dist = max_next_dist.squeeze()\n\n\n Tz = batch_reward.view(-1, 1) + (self.gamma**self.nsteps)*self.supports.view(1, -1) * non_final_mask.to(torch.float).view(-1, 1)\n Tz = Tz.clamp(self.v_min, self.v_max)\n b = (Tz - self.v_min) / self.delta\n l = b.floor().to(torch.int64)\n u = b.ceil().to(torch.int64)\n l[(u > 0) * (l == u)] -= 1\n u[(l < (self.atoms - 1)) * (l == u)] += 1\n \n\n offset = torch.linspace(0, (self.batch_size - 1) * self.atoms, self.batch_size).unsqueeze(dim=1).expand(self.batch_size, self.atoms).to(batch_action)\n m = batch_state.new_zeros(self.batch_size, self.atoms)\n m.view(-1).index_add_(0, (l + offset).view(-1), (max_next_dist * (u.float() - b)).view(-1)) # m_l = m_l + p(s_t+n, a*)(u - b)\n m.view(-1).index_add_(0, (u + offset).view(-1), (max_next_dist * (b - l.float())).view(-1)) # m_u = m_u + p(s_t+n, a*)(b - l)\n\n return m\n \n def compute_loss(self, batch_vars):\n batch_state, batch_action, batch_reward, non_final_next_states, non_final_mask, empty_next_state_values, indices, weights = batch_vars\n\n batch_action = batch_action.unsqueeze(dim=-1).expand(-1, -1, self.atoms)\n batch_reward = batch_reward.view(-1, 1, 1)\n\n #estimate\n self.model.sample_noise()\n current_dist = self.model(batch_state).gather(1, batch_action).squeeze()\n\n target_prob = self.projection_distribution(batch_vars)\n \n loss = -(target_prob * current_dist.log()).sum(-1)\n self.memory.update_priorities(indices, loss.detach().squeeze().abs().cpu().numpy().tolist())\n loss = loss * weights\n loss = loss.mean()\n\n return loss\n \n def get_action(self, s):\n with torch.no_grad():\n X = torch.tensor([s], device=self.device, dtype=torch.float)\n self.model.sample_noise()\n a = self.model(X) * self.supports\n a = a.sum(dim=2).max(1)[1].view(1, 1)\n return a.item()\n \n def get_max_next_state_action(self, next_states):\n next_dist = self.model(next_states) * self.supports\n return next_dist.sum(dim=2).max(1)[1].view(next_states.size(0), 1, 1).expand(-1, -1, self.atoms)", "_____no_output_____" ] ], [ [ "## Plot Results", "_____no_output_____" ] ], [ [ "def plot(frame_idx, rewards, losses, sigma, elapsed_time):\n clear_output(True)\n plt.figure(figsize=(20,5))\n plt.subplot(131)\n plt.title('frame %s. reward: %s. time: %s' % (frame_idx, np.mean(rewards[-10:]), elapsed_time))\n plt.plot(rewards)\n if losses:\n plt.subplot(132)\n plt.title('loss')\n plt.plot(losses)\n if sigma:\n plt.subplot(133)\n plt.title('noisy param magnitude')\n plt.plot(sigma)\n plt.show()", "_____no_output_____" ] ], [ [ "## Training Loop", "_____no_output_____" ] ], [ [ "start=timer()\n\nenv_id = \"PongNoFrameskip-v4\"\nenv = make_atari(env_id)\nenv = wrap_deepmind(env, frame_stack=False)\nenv = wrap_pytorch(env)\nmodel = Model(env=env, config=config)\n\nepisode_reward = 0\n\nobservation = env.reset()\nfor frame_idx in range(1, config.MAX_FRAMES + 1):\n action = model.get_action(observation)\n prev_observation=observation\n observation, reward, done, _ = env.step(action)\n observation = None if done else observation\n\n model.update(prev_observation, action, reward, observation, frame_idx)\n episode_reward += reward\n\n if done:\n model.finish_nstep()\n model.reset_hx()\n observation = env.reset()\n model.save_reward(episode_reward)\n episode_reward = 0\n \n if np.mean(model.rewards[-10:]) > 19:\n plot(frame_idx, all_rewards, losses, timedelta(seconds=int(timer()-start)))\n break\n\n if frame_idx % 10000 == 0:\n plot(frame_idx, model.rewards, model.losses, model.sigma_parameter_mag, timedelta(seconds=int(timer()-start)))\n\nmodel.save_w()\nenv.close()", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code" ]
[ [ "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ] ]
ece8976b5c46532e228d496587cd047b139c0148
941,110
ipynb
Jupyter Notebook
tensorflow_probability/examples/jupyter_notebooks/Gaussian_Copula.ipynb
Tensorflow-Devs/probability
719eafd6a052d774b364e47e5a2142e486c33231
[ "Apache-2.0" ]
3,670
2018-02-14T03:29:40.000Z
2022-03-30T01:19:52.000Z
tensorflow_probability/examples/jupyter_notebooks/Gaussian_Copula.ipynb
emilyfertig/probability
bb15d65fafb9311475878f518a328c8f9bdcf31e
[ "Apache-2.0" ]
1,395
2018-02-24T02:28:49.000Z
2022-03-31T16:12:06.000Z
tensorflow_probability/examples/jupyter_notebooks/Gaussian_Copula.ipynb
emilyfertig/probability
bb15d65fafb9311475878f518a328c8f9bdcf31e
[ "Apache-2.0" ]
1,135
2018-02-14T01:51:10.000Z
2022-03-28T02:24:11.000Z
1,893.581489
231,460
0.947479
[ [ [ "##### Copyright 2018 The TensorFlow Probability Authors.\n\nLicensed under the Apache License, Version 2.0 (the \"License\");", "_____no_output_____" ] ], [ [ "#@title Licensed under the Apache License, Version 2.0 (the \"License\"); { display-mode: \"form\" }\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# https://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.", "_____no_output_____" ] ], [ [ "# Copulas Primer\n\n<table class=\"tfo-notebook-buttons\" align=\"left\">\n <td>\n <a target=\"_blank\" href=\"https://www.tensorflow.org/probability/examples/Gaussian_Copula\"><img src=\"https://www.tensorflow.org/images/tf_logo_32px.png\" />View on TensorFlow.org</a>\n </td>\n <td>\n <a target=\"_blank\" href=\"https://colab.research.google.com/github/tensorflow/probability/blob/main/tensorflow_probability/examples/jupyter_notebooks/Gaussian_Copula.ipynb\"><img src=\"https://www.tensorflow.org/images/colab_logo_32px.png\" />Run in Google Colab</a>\n </td>\n <td>\n <a target=\"_blank\" href=\"https://github.com/tensorflow/probability/blob/main/tensorflow_probability/examples/jupyter_notebooks/Gaussian_Copula.ipynb\"><img src=\"https://www.tensorflow.org/images/GitHub-Mark-32px.png\" />View source on GitHub</a>\n </td>\n <td>\n <a href=\"https://storage.googleapis.com/tensorflow_docs/probability/tensorflow_probability/examples/jupyter_notebooks/Gaussian_Copula.ipynb\"><img src=\"https://www.tensorflow.org/images/download_logo_32px.png\" />Download notebook</a>\n </td>\n</table>", "_____no_output_____" ] ], [ [ "import numpy as np\nimport matplotlib.pyplot as plt\nimport tensorflow.compat.v2 as tf\ntf.enable_v2_behavior()\n\nimport tensorflow_probability as tfp\ntfd = tfp.distributions\ntfb = tfp.bijectors", "_____no_output_____" ] ], [ [ "A [copula](https://en.wikipedia.org/wiki/Copula_(probability_theory%29) is a classical approach for capturing the dependence between random variables. More formally, a copula is a multivariate distribution $C(U_1, U_2, ...., U_n)$ such that marginalizing gives $U_i \\sim \\text{Uniform}(0, 1)$.\n\n\nCopulas are interesting because we can use them to create multivariate distributions with arbitrary marginals. This is the recipe:\n\n* Using the [Probability Integral Transform](https://en.wikipedia.org/wiki/Probability_integral_transform) turns an arbitrary continuous R.V. $X$ into a uniform one $F_X(X)$, where $F_X$ is the CDF of $X$.\n* Given a copula (say bivariate) $C(U, V)$, we have that $U$ and $V$ have uniform marginal distributions.\n* Now given our R.V's of interest $X, Y$, create a new distribution $C'(X, Y) = C(F_X(X), F_Y(Y))$. The marginals for $X$ and $Y$ are the ones we desired.\n\nMarginals are univariate and thus may be easier to measure and/or model. A copula enables starting from marginals yet also achieving arbitrary correlation between dimensions.\n", "_____no_output_____" ], [ "# Gaussian Copula\n\nTo illustrate how copulas are constructed, consider the case of capturing dependence according to multivariate Gaussian correlations. A Gaussian Copula is one given by $C(u_1, u_2, ...u_n) = \\Phi_\\Sigma(\\Phi^{-1}(u_1), \\Phi^{-1}(u_2), ... \\Phi^{-1}(u_n))$ where $\\Phi_\\Sigma$ represents the CDF of a MultivariateNormal, with covariance $\\Sigma$ and mean 0, and $\\Phi^{-1}$ is the inverse CDF for the standard normal.\n\n\nApplying the normal's inverse CDF warps the uniform dimensions to be normally distributed. Applying the multivariate normal's CDF then squashes the distribution to be marginally uniform and with Gaussian correlations.\n\nThus, what we get is that the Gaussian Copula is a distribution over the unit hypercube $[0, 1]^n$ with uniform marginals.\n\nDefined as such, the Gaussian Copula can be implemented with `tfd.TransformedDistribution` and appropriate `Bijector`. That is, we are transforming a MultivariateNormal, via the use of the Normal distribution's inverse CDF, implemented by the [`tfb.NormalCDF`](https://www.tensorflow.org/probability/api_docs/python/tfp/bijectors/NormalCDF) bijector.", "_____no_output_____" ], [ "Below, we implement a Gaussian Copula with one simplifying assumption: that the covariance is parameterized\nby a Cholesky factor (hence a covariance for `MultivariateNormalTriL`). (One could use other `tf.linalg.LinearOperators` to encode different matrix-free assumptions.).", "_____no_output_____" ] ], [ [ "class GaussianCopulaTriL(tfd.TransformedDistribution):\n \"\"\"Takes a location, and lower triangular matrix for the Cholesky factor.\"\"\"\n def __init__(self, loc, scale_tril):\n super(GaussianCopulaTriL, self).__init__(\n distribution=tfd.MultivariateNormalTriL(\n loc=loc,\n scale_tril=scale_tril),\n bijector=tfb.NormalCDF(),\n validate_args=False,\n name=\"GaussianCopulaTriLUniform\")\n\n\n# Plot an example of this.\nunit_interval = np.linspace(0.01, 0.99, num=200, dtype=np.float32)\nx_grid, y_grid = np.meshgrid(unit_interval, unit_interval)\ncoordinates = np.concatenate(\n [x_grid[..., np.newaxis],\n y_grid[..., np.newaxis]], axis=-1)\n\npdf = GaussianCopulaTriL(\n loc=[0., 0.],\n scale_tril=[[1., 0.8], [0., 0.6]],\n).prob(coordinates)\n\n# Plot its density.\n\nplt.contour(x_grid, y_grid, pdf, 100, cmap=plt.cm.jet);", "_____no_output_____" ] ], [ [ "The power, however, from such a model is using the Probability Integral Transform, to use the copula on arbitrary R.V.s. In this way, we can specify arbitrary marginals, and use the copula to stitch them together.\n\nWe start with a model:\n\n$$\\begin{align*}\nX &\\sim \\text{Kumaraswamy}(a, b) \\\\\nY &\\sim \\text{Gumbel}(\\mu, \\beta)\n\\end{align*}$$\n\nand use the copula to get a bivariate R.V. $Z$, which has marginals [Kumaraswamy](https://en.wikipedia.org/wiki/Kumaraswamy_distribution) and [Gumbel](https://en.wikipedia.org/wiki/Gumbel_distribution).\n\n\nWe'll start by plotting the product distribution generated by those two R.V.s. This is just to serve as a comparison point to when we apply the Copula.", "_____no_output_____" ] ], [ [ "a = 2.0\nb = 2.0\ngloc = 0.\ngscale = 1.\n\nx = tfd.Kumaraswamy(a, b)\ny = tfd.Gumbel(loc=gloc, scale=gscale)\n\n# Plot the distributions, assuming independence\nx_axis_interval = np.linspace(0.01, 0.99, num=200, dtype=np.float32)\ny_axis_interval = np.linspace(-2., 3., num=200, dtype=np.float32)\nx_grid, y_grid = np.meshgrid(x_axis_interval, y_axis_interval)\n\npdf = x.prob(x_grid) * y.prob(y_grid)\n\n# Plot its density\n\nplt.contour(x_grid, y_grid, pdf, 100, cmap=plt.cm.jet);", "_____no_output_____" ] ], [ [ "# Joint Distribution with Different Marginals", "_____no_output_____" ], [ "Now we use a Gaussian copula to couple the distributions together, and plot that. Again our tool of choice is `TransformedDistribution` applying the appropriate `Bijector` to obtain the chosen marginals.\n\nSpecifically, we use a [`Blockwise`](https://www.tensorflow.org/probability/api_docs/python/tfp/bijectors/Blockwise) bijector which applies different bijectors at different parts of the vector (which is still a bijective transformation).", "_____no_output_____" ], [ "Now we can define the Copula we want. Given a list of target marginals (encoded as bijectors), we can easily construct\na new distribution that uses the copula and has the specified marginals.\n", "_____no_output_____" ] ], [ [ "class WarpedGaussianCopula(tfd.TransformedDistribution):\n \"\"\"Application of a Gaussian Copula on a list of target marginals.\n\n This implements an application of a Gaussian Copula. Given [x_0, ... x_n]\n which are distributed marginally (with CDF) [F_0, ... F_n],\n `GaussianCopula` represents an application of the Copula, such that the\n resulting multivariate distribution has the above specified marginals.\n\n The marginals are specified by `marginal_bijectors`: These are\n bijectors whose `inverse` encodes the CDF and `forward` the inverse CDF.\n\n block_sizes is a 1-D Tensor to determine splits for `marginal_bijectors`\n length should be same as length of `marginal_bijectors`.\n See tfb.Blockwise for details\n \"\"\"\n def __init__(self, loc, scale_tril, marginal_bijectors, block_sizes=None):\n super(WarpedGaussianCopula, self).__init__(\n distribution=GaussianCopulaTriL(loc=loc, scale_tril=scale_tril),\n bijector=tfb.Blockwise(bijectors=marginal_bijectors,\n block_sizes=block_sizes),\n validate_args=False,\n name=\"GaussianCopula\")", "_____no_output_____" ] ], [ [ "Finally, let's actually use this Gaussian Copula. We'll use a Cholesky of $\\begin{bmatrix}1 & 0\\\\\\rho & \\sqrt{(1-\\rho^2)}\\end{bmatrix}$, which will correspond to variances 1, and correlation $\\rho$ for the multivariate normal.\n\n\nWe'll look at a few cases: ", "_____no_output_____" ] ], [ [ "# Create our coordinates:\ncoordinates = np.concatenate(\n [x_grid[..., np.newaxis], y_grid[..., np.newaxis]], -1)\n\n\ndef create_gaussian_copula(correlation):\n # Use Gaussian Copula to add dependence.\n return WarpedGaussianCopula(\n loc=[0., 0.],\n scale_tril=[[1., 0.], [correlation, tf.sqrt(1. - correlation ** 2)]],\n # These encode the marginals we want. In this case we want X_0 has\n # Kumaraswamy marginal, and X_1 has Gumbel marginal.\n\n marginal_bijectors=[\n tfb.Invert(tfb.KumaraswamyCDF(a, b)),\n tfb.Invert(tfb.GumbelCDF(loc=0., scale=1.))])\n\n\n# Note that the zero case will correspond to independent marginals!\ncorrelations = [0., -0.8, 0.8]\ncopulas = []\nprobs = []\nfor correlation in correlations:\n copula = create_gaussian_copula(correlation)\n copulas.append(copula)\n probs.append(copula.prob(coordinates))\n\n\n# Plot it's density\n\nfor correlation, copula_prob in zip(correlations, probs):\n plt.figure()\n plt.contour(x_grid, y_grid, copula_prob, 100, cmap=plt.cm.jet)\n plt.title('Correlation {}'.format(correlation))", "_____no_output_____" ] ], [ [ "Finally, let's verify that we actually get the marginals we want.", "_____no_output_____" ] ], [ [ "def kumaraswamy_pdf(x):\n return tfd.Kumaraswamy(a, b).prob(np.float32(x))\n\ndef gumbel_pdf(x):\n return tfd.Gumbel(gloc, gscale).prob(np.float32(x))\n\n\ncopula_samples = []\nfor copula in copulas:\n copula_samples.append(copula.sample(10000))\n\nplot_rows = len(correlations)\nplot_cols = 2 # for 2 densities [kumarswamy, gumbel]\nfig, axes = plt.subplots(plot_rows, plot_cols, sharex='col', figsize=(18,12))\n\n# Let's marginalize out on each, and plot the samples.\n\nfor i, (correlation, copula_sample) in enumerate(zip(correlations, copula_samples)):\n k = copula_sample[..., 0].numpy()\n g = copula_sample[..., 1].numpy()\n\n\n _, bins, _ = axes[i, 0].hist(k, bins=100, density=True)\n axes[i, 0].plot(bins, kumaraswamy_pdf(bins), 'r--')\n axes[i, 0].set_title('Kumaraswamy from Copula with correlation {}'.format(correlation))\n\n _, bins, _ = axes[i, 1].hist(g, bins=100, density=True)\n axes[i, 1].plot(bins, gumbel_pdf(bins), 'r--')\n axes[i, 1].set_title('Gumbel from Copula with correlation {}'.format(correlation))\n ", "_____no_output_____" ] ], [ [ "# Conclusion\nAnd there we go! We've demonstrated that we can construct Gaussian Copulas using the `Bijector` API.\n\nMore generally, writing bijectors using the `Bijector` API and composing them with a distribution, can create rich families of distributions for flexible modelling.", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown" ]
[ [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ] ]
ece8a02392739a9877674436ed3af1c830f27aa8
8,253
ipynb
Jupyter Notebook
s5/codes/wordnet.ipynb
gebakx/ihlt
10c22f6e6c38a84772457d8c4093878ef917d9cf
[ "MIT" ]
null
null
null
s5/codes/wordnet.ipynb
gebakx/ihlt
10c22f6e6c38a84772457d8c4093878ef917d9cf
[ "MIT" ]
null
null
null
s5/codes/wordnet.ipynb
gebakx/ihlt
10c22f6e6c38a84772457d8c4093878ef917d9cf
[ "MIT" ]
1
2019-11-20T22:58:32.000Z
2019-11-20T22:58:32.000Z
20.428218
101
0.469284
[ [ [ "## WordNet reader", "_____no_output_____" ] ], [ [ "from nltk.corpus import wordnet as wn", "_____no_output_____" ] ], [ [ "### synsets", "_____no_output_____" ] ], [ [ "wn.synsets('age','n')", "_____no_output_____" ], [ "age = wn.synset('age.n.1')\nage", "_____no_output_____" ] ], [ [ "### definitions, examples and lemmas", "_____no_output_____" ] ], [ [ "age.definition()", "_____no_output_____" ], [ "age.examples()", "_____no_output_____" ], [ "ls = wn.synsets('age','n')\nll = ls[1].lemmas()\n[lemma.name() for lemma in ll]", "_____no_output_____" ] ], [ [ "### hyponyms and hypernyms", "_____no_output_____" ] ], [ [ "age.hyponyms()", "_____no_output_____" ], [ "age.hypernyms()", "_____no_output_____" ], [ "age.root_hypernyms()", "_____no_output_____" ], [ "hyper = lambda s: s.hypernyms()\nlist(age.closure(hyper))", "_____no_output_____" ], [ "age.tree(hyper)", "_____no_output_____" ] ], [ [ "### antonyms", "_____no_output_____" ] ], [ [ "good = wn.synset('good.a.01')\ngood.lemmas()[0].antonyms()", "_____no_output_____" ] ], [ [ "### all lexical relations", "_____no_output_____" ] ], [ [ "def getRelValue(name):\n method = getattr(age, rel)\n return method()\n \nlexRels = ['hypernyms', 'instance_hypernyms', 'hyponyms', 'instance_hyponyms', \\\n 'member_holonyms', 'substance_holonyms', 'part_holonyms', \\\n 'member_meronyms', 'substance_meronyms', 'part_meronyms', \\\n 'attributes', 'entailments', 'causes', 'also_sees', 'verb_groups', 'similar_tos']\nage = wn.synset('age.n.01')\n\nresults = {}\nfor rel in lexRels:\n val = getRelValue(rel)\n if val != []:\n results[rel] = val\nresults", "_____no_output_____" ], [ "for rel in results:\n for s in results[rel]:\n print(rel, s.name())", "hyponyms bone_age.n.01\nhyponyms chronological_age.n.01\nhyponyms developmental_age.n.01\nhyponyms fetal_age.n.01\nhyponyms mental_age.n.01\nhyponyms newness.n.01\nhyponyms oldness.n.01\nhyponyms oldness.n.02\nhyponyms youngness.n.01\nhypernyms property.n.02\nattributes immature.a.04\nattributes mature.a.03\nattributes new.a.01\nattributes old.a.01\nattributes old.a.02\nattributes young.a.01\n" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code" ]
[ [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ] ]
ece8ab5884ac11aa0921c01c624547a1be880c09
657,840
ipynb
Jupyter Notebook
Seaborn/Seaborn.ipynb
jimit105/Python-Libraries
7e45e996454d323dad3aaa54875eef8eb865739b
[ "MIT" ]
1
2018-10-02T13:36:56.000Z
2018-10-02T13:36:56.000Z
Seaborn/Seaborn.ipynb
shubhanshu1995/Python-Libraries
ac4df5e28ac3bce2da4376b7989ffc96f2848fd9
[ "MIT" ]
null
null
null
Seaborn/Seaborn.ipynb
shubhanshu1995/Python-Libraries
ac4df5e28ac3bce2da4376b7989ffc96f2848fd9
[ "MIT" ]
2
2019-07-15T06:57:54.000Z
2021-11-17T05:00:21.000Z
496.483019
88,512
0.939402
[ [ [ "# Seaborn", "_____no_output_____" ], [ "Seaborn provides a high-level interface to Matplotlib, a powerful but sometimes unwieldy Python visualization library.\n\n> Seaborn is a complement, not a substitute, for Matplotlib. There are some tweaks that still require Matplotlib", "_____no_output_____" ] ], [ [ "# Importing Libraries\n\n# Pandas for managing datasets\nimport pandas as pd\n\n# Matplotlib for additional customization\nfrom matplotlib import pyplot as plt\n%matplotlib inline\n\n# Seaborn for plotting and styling\nimport seaborn as sns", "_____no_output_____" ], [ "# Importing dataset\n\ndf = pd.read_csv('Pokemon.csv', index_col=0)\n# index_col=0 simply means we'll treat the first column of the dataset as the ID column.", "_____no_output_____" ], [ "# Display first 5 observations\ndf.head()", "_____no_output_____" ] ], [ [ "## Seaborn's plotting functions", "_____no_output_____" ], [ "One of Seaborn's greatest strengths is its diversity of plotting functions. For instance, making a **scatter plot** is just one line of code using the ```lmplot()``` function.\n\nThere are two ways you can do so.\n\n* The first way (recommended) is to pass your DataFrame to the ```data=``` argument, while passing column names to the axes arguments, ```x=``` and ```y=```.\n* The second way is to directly pass in Series of data to the axes arguments", "_____no_output_____" ] ], [ [ "# Recommended way\nsns.lmplot(x='Attack', y='Defense', data=df)", "_____no_output_____" ] ], [ [ "Seaborn doesn't have a dedicated scatter plot function, which is why you see a diagonal line. We actually used Seaborn's function for fitting and plotting a **regression line**.", "_____no_output_____" ], [ "Here's how we can tweak the ```lmplot()```:\n\n* First, we'll set ```fit_reg=False``` to remove the regression line, since we only want a scatter plot.\n* Then, we'll set ```hue='Stage'``` to color our points by the Pokémon's evolution stage. \nThis hue argument is very useful because it allows you to express a third dimension of information using color.", "_____no_output_____" ] ], [ [ "# Scatterplot arguments\nsns.lmplot(x='Attack', y='Defense', data=df,\n fit_reg=False, # No regression line\n hue='Stage') # Color by evolution stage", "_____no_output_____" ] ], [ [ "## Customizing with Matplotlib", "_____no_output_____" ], [ "Setting axes limits:\n\n1. First, invoke your Seaborn plotting function as normal.\n2. Then, invoke Matplotlib's customization functions. In this case, we'll use its ```ylim()``` and ```xlim()``` functions.", "_____no_output_____" ] ], [ [ "# Plot using Seaborn\nsns.lmplot(x='Attack', y='Defense', data=df,\n fit_reg=False,\n hue='Stage')\n\n# Tweak using Matplotlib\nplt.ylim(0, None)\nplt.xlim(0, None)", "_____no_output_____" ] ], [ [ "## Role of Pandas", "_____no_output_____" ], [ "Pandas actually plays a very important role \nSeaborn's plotting functions benefit from a base DataFrame that's reasonably formatted.", "_____no_output_____" ] ], [ [ "sns.boxplot(data=df)", "_____no_output_____" ], [ "# Pre-format DataFrame\nstats_df = df.drop(['Total', 'Stage', 'Legendary'], axis=1)\n\n# New boxplot using stats_df\nsns.boxplot(data=stats_df)", "_____no_output_____" ] ], [ [ "## Seaborn themes", "_____no_output_____" ], [ "The default theme is called 'darkgrid'. \n\nNext, we'll change the theme to 'whitegrid' while making a violin plot. \n\n* Violin plots are useful alternatives to box plots.\n* They show the distribution (through the thickness of the violin) instead of only the summary statistics.", "_____no_output_____" ] ], [ [ "# Set theme\nsns.set_style('whitegrid')\n\n# Violin plot\nsns.violinplot(x='Type 1', y='Attack', data=df)", "_____no_output_____" ] ], [ [ "## Color palattes", "_____no_output_____" ], [ "Seaborn allows us to set custom color palettes. We can simply create an ordered Python list of color hex values.", "_____no_output_____" ] ], [ [ "pkmn_type_colors = ['#78C850', # Grass\n '#F08030', # Fire\n '#6890F0', # Water\n '#A8B820', # Bug\n '#A8A878', # Normal\n '#A040A0', # Poison\n '#F8D030', # Electric\n '#E0C068', # Ground\n '#EE99AC', # Fairy\n '#C03028', # Fighting\n '#F85888', # Psychic\n '#B8A038', # Rock\n '#705898', # Ghost\n '#98D8D8', # Ice\n '#7038F8', # Dragon\n ]", "_____no_output_____" ], [ "# Violin plot with Pokemon color palatte\nsns.violinplot(x='Type 1', y='Attack', data=df,\n palette=pkmn_type_colors) # Set color palatte", "_____no_output_____" ], [ "# Swarm plot with Pokemon color palatte\nsns.swarmplot(x='Type 1', y='Attack', data=df,\n palette=pkmn_type_colors)", "_____no_output_____" ] ], [ [ "## Overlaying plots", "_____no_output_____" ], [ "It's pretty straightforward to overlay plots using Seaborn, and it works the same way as with Matplotlib. Here's what we'll do:\n\n1. First, we'll make our figure larger using Matplotlib.\n2. Then, we'll plot the violin plot. However, we'll set ```inner=None``` to remove the bars inside the violins.\n3. Next, we'll plot the swarm plot. This time, we'll make the points black so they pop out more.\n4. Finally, we'll set a title using Matplotlib.", "_____no_output_____" ] ], [ [ "# Set figure size with Matplotlib\nplt.figure(figsize=(10,6))\n\n# Create plot\nsns.violinplot(x='Type 1', y='Attack', data=df,\n inner=None, #Remove the bars inside the violins\n palette=pkmn_type_colors)\n\nsns.swarmplot(x='Type 1', y='Attack', data=df,\n color='k', # Make points black\n alpha=0.7) # and slightly transparent\n\n# Set title with matplotlib\nplt.title('Attack by Type')", "_____no_output_____" ] ], [ [ "## Putting it all together", "_____no_output_____" ], [ "we could certainly repeat that chart for each stat. \nBut we can also combine the information into one chart... we just have to do some **data wrangling** with Pandas beforehand.", "_____no_output_____" ] ], [ [ "stats_df.head()", "_____no_output_____" ] ], [ [ "As you can see, all of our stats are in separate columns. Instead, we want to \"melt\" them into one column.\n\nTo do so, we'll use Pandas's ```melt()``` function. It takes 3 arguments:\n\n* First, the DataFrame to melt.\n* Second, ID variables to keep (Pandas will melt all of the other ones).\n* Finally, a name for the new, melted variable.", "_____no_output_____" ] ], [ [ "# Melt DataFrames\nmelted_df = pd.melt(stats_df,\n id_vars=['Name', 'Type 1', 'Type 2'], #Variables to keep\n var_name='Stat') # Name of melted variable\n\n#melted_df.head()\nmelted_df.tail()", "_____no_output_____" ] ], [ [ "All 6 of the stat columns have been \"melted\" into one, and the new Stat column indicates the original stat (HP, Attack, Defense, Sp. Attack, Sp. Defense, or Speed). \n\nFor example, it's hard to see here, but Bulbasaur now has 6 rows of data.", "_____no_output_____" ] ], [ [ "# Shape comparison\nprint(stats_df.shape)\nprint(melted_df.shape)", "(151, 9)\n(906, 5)\n" ], [ "# Swarm plot with melted_df\nsns.swarmplot(x='Stat', y='value', data=melted_df,\n hue='Type 1')", "_____no_output_____" ] ], [ [ "Finally, let's make a few final tweaks for a more readable chart:\n\n1. Enlarge the plot.\n2. Separate points by hue using the argument `split=True` .\n> `split` parameter has been renamed to `dodge`\n3. Use our custom Pokemon color palette.\n4. Adjust the y-axis limits to end at 0.\n5. Place the legend to the right.", "_____no_output_____" ] ], [ [ "# 1. Enlarge the plot\nplt.Figure(figsize=(10,6))\n\nsns.swarmplot(x='Stat', y='value', data=melted_df,\n hue='Type 1',\n dodge=True, # 2. Separate points by hue\n # split parameter has been renamed to dodge\n palette=pkmn_type_colors) # 3. Use Pokemon palatte\n\n# 4. Adjust the y-axis\nplt.ylim(0, 260)\n\n# 5. Place legend to the right\nplt.legend(bbox_to_anchor=(1,1), loc=2)", "_____no_output_____" ] ], [ [ "## Pokedex (mini-gallery)", "_____no_output_____" ], [ "### Heatmap\n> Heatmaps help you visualize matrix-like data.", "_____no_output_____" ] ], [ [ "# Calculate correlations\ncorr = stats_df.corr()\n\n# Heatmap\nsns.heatmap(corr)", "_____no_output_____" ] ], [ [ "### Histogram\n> Histograms allow you to plot the distributions of numeric variables.", "_____no_output_____" ] ], [ [ "#Distribution Plot (a.k.a. Histogram)\nsns.distplot(df.Attack)", "_____no_output_____" ] ], [ [ "### Bar Plot\n> Bar plots help you visualize the distributions of categorical variables.", "_____no_output_____" ] ], [ [ "# Count Plot (a.k.a. Bar Plot)\nsns.countplot(x='Type 1', data=df, palette=pkmn_type_colors)\n\n# Rotate x-labels\nplt.xticks(rotation=-45)", "_____no_output_____" ] ], [ [ "### Factor Plot\n> Factor plots make it easy to separate plots by categorical classes.", "_____no_output_____" ] ], [ [ "# Factor plot\ng = sns.factorplot(x='Type 1', y='Attack', data=df,\n hue='Stage', # Color by stage\n col='Stage', # Separate by stage\n kind='swarm')# Swarmplot\n\n# Rotate x-axis labels\ng.set_xticklabels(rotation=-45)\n\n#Doesn't work because only rotates last plot\n# plt.xticks(rotation=-45)", "_____no_output_____" ] ], [ [ "### Density Plot\n> Density plots display the distribution between two variables.", "_____no_output_____" ] ], [ [ "# Density plot\nsns.kdeplot(df.Attack, df.Defense)", "_____no_output_____" ] ], [ [ "### Joint DIstribution Plot\n> Joint distribution plots combine information from scatter plots and histograms to give you detailed information for bi-variate distributions.", "_____no_output_____" ] ], [ [ "# Joint Distribution Plot\nsns.jointplot(x='Attack', y='Defense', data=df)", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code" ]
[ [ "markdown", "markdown" ], [ "code", "code", "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code", "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code", "code", "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ] ]
ece8ac882bc5777ec7ee52f326ae940c49b55b0a
746,606
ipynb
Jupyter Notebook
SHBank/FinDataAnalysis_SHBank_ing.ipynb
waverDeep/FinDataAnalysis
65d4b0987194317780078bf1c5c63e96385a0f8f
[ "MIT" ]
null
null
null
SHBank/FinDataAnalysis_SHBank_ing.ipynb
waverDeep/FinDataAnalysis
65d4b0987194317780078bf1c5c63e96385a0f8f
[ "MIT" ]
5
2021-08-31T08:29:50.000Z
2021-09-04T09:28:50.000Z
SHBank/FinDataAnalysis_SHBank_ing.ipynb
waverDeep/FinDataAnalysis
65d4b0987194317780078bf1c5c63e96385a0f8f
[ "MIT" ]
null
null
null
363.135214
116,216
0.93389
[ [ [ "# 3️⃣ 신한은행 - 서울시 지역단위 '소득', '지출', '금융자산' 정보", "_____no_output_____" ], [ "---", "_____no_output_____" ], [ "## 전체 데이터 파악하기", "_____no_output_____" ] ], [ [ "%matplotlib inline\n\nimport pandas as pd\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport seaborn as sns\n\n# 한글 폰트 사용을 위해서 세팅\nfrom matplotlib import font_manager, rc\nfont_path = \"C:/Windows/Fonts/NanumBarunpenR.ttf\"\nfont = font_manager.FontProperties(fname=font_path).get_name()\nrc('font', family=font)", "_____no_output_____" ], [ "file_path = './신한은행_서울시 지역단위 \\'소득\\', \\'지출\\', \\'금융자산\\' 정보.csv'\ndf = pd.read_csv(file_path, encoding='cp949')\n\nprint(df.shape)\nprint(df.info())\ndf.head()", "(1039568, 20)\n<class 'pandas.core.frame.DataFrame'>\nRangeIndex: 1039568 entries, 0 to 1039567\nData columns (total 20 columns):\n # Column Non-Null Count Dtype \n--- ------ -------------- ----- \n 0 기준년월 1039568 non-null int64 \n 1 지역구 1039568 non-null object \n 2 법정동코드 1039568 non-null int64 \n 3 법정동 1039568 non-null object \n 4 집계구코드 1039568 non-null float64\n 5 나이 1039568 non-null int64 \n 6 성별 1039568 non-null int64 \n 7 직장인여부 1039568 non-null int64 \n 8 급여입금 1039568 non-null int64 \n 9 가맹점매출입금 1039568 non-null int64 \n 10 연금입금 1039568 non-null int64 \n 11 총소비금액 1039568 non-null int64 \n 12 총수신금액 1039568 non-null int64 \n 13 예적금금액 1039568 non-null int64 \n 14 신탁금액 1039568 non-null int64 \n 15 수익증권금액 1039568 non-null int64 \n 16 신용대출금액 1039568 non-null int64 \n 17 담보대출금액 1039568 non-null int64 \n 18 주택대출금액 1039568 non-null int64 \n 19 전세자금대출금액 1039568 non-null int64 \ndtypes: float64(1), int64(17), object(2)\nmemory usage: 158.6+ MB\nNone\n" ], [ "df.columns", "_____no_output_____" ] ], [ [ "##### 결측 데이터 조회 - 없음", "_____no_output_____" ] ], [ [ "#df.isnull().sum()", "_____no_output_____" ] ], [ [ "---", "_____no_output_____" ], [ "#### 자치구 ('지역구') 확인하기", "_____no_output_____" ] ], [ [ "county = df['지역구'].unique()\n\n# 각 지역구가 맞는지 판별하기 위한 True/False\niscounty = []\nfor i in range(len(county)):\n iscounty.append(df['지역구'] == county[i])\n \n# 각 지역구의 데이터프레임\ncounties = []\nfor j in range(25):\n counties.append(df[iscounty[j]])", "_____no_output_____" ] ], [ [ "---", "_____no_output_____" ], [ "# ✍🏻", "_____no_output_____" ], [ "## 성별과 나이", "_____no_output_____" ], [ "### 📌 서울특별시 전체 인구 대상", "_____no_output_____" ] ], [ [ "### 서울특별시 전체 대상으로 1(남성)과 2(여성)의 인구가 매우 비슷함", "_____no_output_____" ], [ "print(df['성별'].value_counts())", "1 525568\n2 514000\nName: 성별, dtype: int64\n" ], [ "### 전체적인 인구로 보았을 때, 30대가 약 20만명으로 가장 많고, 70대가 14만명으로 가장 적다.\n## 중장년층인 4,5,60대에 비해 2,30대 청년층의 인구가 비교적 적다.", "_____no_output_____" ], [ "plt.figure(figsize=(5,5))\ndf['나이'].value_counts().plot.bar()\nplt.show()\n\nfor i in range(6):\n print(str(i+2) + '0대 :', len(df[df['나이'] == (i+2)]))", "_____no_output_____" ], [ "### 각 연령대 기준으로 성별 비율이 유사함\n## 20대, 30대는 여성이 더 많고 그 외는 모두 남성이 더 많음", "_____no_output_____" ], [ "ax = sns.countplot(x='나이', hue = '성별', data = df)", "_____no_output_____" ], [ "thirty = df['나이'] == 3\nman = df['성별'] == 1\nwoman = df['성별'] == 2\n\nprint(len(df[thirty & man]))\nprint(len(df[thirty & woman]))", "97221\n98183\n" ] ], [ [ "---", "_____no_output_____" ], [ "### 📌 각 자치구 개별적 대상", "_____no_output_____" ] ], [ [ "###", "_____no_output_____" ], [ "plt.figure(figsize=(20,20))\n\nfor i in range(25):\n plt.subplot(5, 5, i+1)\n counties[i]['나이'].value_counts().plot.bar()\n plt.ylim([0, 14000])\n plt.xticks(rotation=0)\n plt.title(county[i] + ' 연령대 분포')\n\nplt.show()", "_____no_output_____" ], [ "###", "_____no_output_____" ], [ "plt.figure(figsize=(20,20))\n\nfor i in range(25):\n plt.subplot(5, 5, i+1)\n sns.countplot(x='나이', hue = '성별', data = counties[i])\n plt.title(county[i] + ' 나이대 성별')\n\nplt.show()", "_____no_output_____" ] ], [ [ "---", "_____no_output_____" ], [ "# 🔔 총소비금액", "_____no_output_____" ], [ "### ⚪각 자치구 별 총소비금액이 0 이하인 데이터 확인하기", "_____no_output_____" ] ], [ [ "iszero = []\nsum = 0\nfor i in range(25):\n iszero.append(counties[i]['총소비금액'] <= 0)\n #print(county[i] + ' : ' + str(len(counties[i][iszero[i]])))\n sum += len(counties[i][iszero[i]])\n\nprint(\"\\n\\n총합 : \" + str(sum))", "\n\n총합 : 246787\n" ] ], [ [ "#### ⚫ 각 자치구 별 총소비금액 평균 (원본)", "_____no_output_____" ] ], [ [ "# 각 자치구 별 총소비금액 평균\nmean = []\nfor i in range(25):\n mean.append(counties[i]['총소비금액'].mean())\n #print(county[i] + ' : '+ str(counties[i]['총소비금액'].mean()))", "_____no_output_____" ] ], [ [ "---", "_____no_output_____" ], [ "#### ⚫ 각 자치구 별로 '총소비금액'이 0 이하인 데이터에 평균값(자치구 단위)으로 넣어주기", "_____no_output_____" ] ], [ [ "for i in range(25):\n counties[i].loc[iszero[i], '총소비금액'] = counties[i]['총소비금액'].mean()", "C:\\Users\\serak\\anaconda3\\lib\\site-packages\\pandas\\core\\indexing.py:966: SettingWithCopyWarning: \nA value is trying to be set on a copy of a slice from a DataFrame.\nTry using .loc[row_indexer,col_indexer] = value instead\n\nSee the caveats in the documentation: https://pandas.pydata.org/pandas-docs/stable/user_guide/indexing.html#returning-a-view-versus-a-copy\n self.obj[item] = s\n" ] ], [ [ "---", "_____no_output_____" ], [ "#### ⚫ 각 자치구 별 총소비금액 평균 (가공 이후)", "_____no_output_____" ] ], [ [ "for i in range(25):\n mean.append(counties[i]['총소비금액'].mean())\n #print(county[i] + ' : '+ str(counties[i]['총소비금액'].mean()))", "_____no_output_____" ] ], [ [ "---", "_____no_output_____" ], [ "# 🔔 총수신금액", "_____no_output_____" ], [ "### ⚪ '총수신금액'이 0 이하인 데이터 확인하기", "_____no_output_____" ] ], [ [ "iszero = []\nsum = 0\nfor i in range(25):\n iszero.append(counties[i]['총수신금액'] <= 0)\n #print(county[i] + ' : ' + str(len(counties[i][iszero[i]])))\n sum += len(counties[i][iszero[i]])\n\nprint(\"\\n\\n총합 : \" + str(sum))", "\n\n총합 : 3\n" ] ], [ [ "#### ⚫ 각 자치구 별 총수신금액", "_____no_output_____" ] ], [ [ "# 각 자치구 별 총수신금액 평균\nmean = []\nfor i in range(25):\n mean.append(counties[i]['총수신금액'].mean())\n #print(county[i] + ' : '+ str(counties[i]['총수신금액'].mean()))", "_____no_output_____" ] ], [ [ "---", "_____no_output_____" ], [ "# 🔔 직장인여부 / 급여입금", "_____no_output_____" ], [ "### ⚪ 자치구 별 직장인 수(급여입금 받는 인원수) 확인하기", "_____no_output_____" ] ], [ [ "# 자치구 별로 급여입금 받는 인원수 확인하기\nisworker = []\nsum = 0\nfor i in range(25):\n isworker.append(counties[i]['급여입금'] > 0)\n #print(county[i] + ' : ' + str(len(counties[i][isworker[i]])))\n sum += len(counties[i][isworker[i]])\n\nprint(\"\\n\\n총합 : \" + str(sum))", "\n\n총합 : 36009\n" ] ], [ [ "#### ⚫ 각 자치구 별 직장인 대상 급여입금 평균\n** 급여입금이 없는 사람들, 즉, 직장인에 해당하지 않는 사람들 제외", "_____no_output_____" ] ], [ [ "# 각 자치구 별 직장인 대상 급여입금 평균 \nmean = []\nfor i in range(25):\n mean.append(counties[i][isworker[i]]['급여입금'].mean())\n #print(county[i] + ' : '+ str(counties[i][isworker[i]]['급여입금'].mean()))", "_____no_output_____" ] ], [ [ "---", "_____no_output_____" ], [ "## 🟡 총소비금액 & 총수신금액 & 직장인 수 & 급여입금 - 막대 그래프, 꺾은선 그래프\n- 직장인 수\n- 총소비금액 & 급여입금 모두 0 데이터 제외 이후 평균 값으로 적용 !!\n- 급여입금은 직장인 대상 평균", "_____no_output_____" ] ], [ [ "worker = []\nfor i in range(25):\n worker.append(len(counties[i][isworker[i]]))", "_____no_output_____" ], [ "consumption = []\nfor i in range(25):\n consumption.append(counties[i]['총소비금액'].mean())", "_____no_output_____" ], [ "asset = []\nfor i in range(25):\n asset.append(counties[i]['총수신금액'].mean())", "_____no_output_____" ], [ "pay = []\nfor i in range(25):\n pay.append(counties[i][isworker[i]]['급여입금'].mean())", "_____no_output_____" ] ], [ [ "---", "_____no_output_____" ], [ "# ✍🏻 - 여기 정리하기", "_____no_output_____" ], [ "### 급여입금/총소비/총자산 순위", "_____no_output_____" ] ], [ [ "### 급여입금 순위\ndic_pay = {}\nfor i in range(25):\n dic_pay[county[i]] = int(pay[i])\n\ndic_pay = dict(sorted(dic_pay.items(), key=lambda x:x[1], reverse=True))\nfor i in range(25):\n print(str(i+1) +'위 : ' +list(dic_pay.keys())[i] + '\\t' + str(list(dic_pay.values())[i]) + ' 원')", "1위 : 서초구\t4434539 원\n2위 : 강남구\t4397614 원\n3위 : 양천구\t4096984 원\n4위 : 용산구\t3993082 원\n5위 : 송파구\t3898295 원\n6위 : 중구\t3735404 원\n7위 : 종로구\t3684768 원\n8위 : 성북구\t3668333 원\n9위 : 강동구\t3586378 원\n10위 : 마포구\t3571755 원\n11위 : 성동구\t3542189 원\n12위 : 서대문구\t3522997 원\n13위 : 강북구\t3506619 원\n14위 : 영등포구\t3492250 원\n15위 : 광진구\t3398319 원\n16위 : 동작구\t3362681 원\n17위 : 구로구\t3261096 원\n18위 : 노원구\t3170992 원\n19위 : 강서구\t3162227 원\n20위 : 동대문구\t3139883 원\n21위 : 도봉구\t3057016 원\n22위 : 은평구\t3049636 원\n23위 : 관악구\t2925671 원\n24위 : 금천구\t2893865 원\n25위 : 중랑구\t2878703 원\n" ], [ "### 총소비금액 순위\ndic_cons = {}\nfor i in range(25):\n dic_cons[county[i]] = int(consumption[i])\n\ndic_cons = dict(sorted(dic_cons.items(), key=lambda x:x[1], reverse=True))\nfor i in range(25):\n print(str(i+1) +'위 : ' +list(dic_cons.keys())[i] + '\\t' + str(list(dic_cons.values())[i]) + ' 원')", "1위 : 강남구\t1914132 원\n2위 : 서초구\t1870463 원\n3위 : 용산구\t1708576 원\n4위 : 송파구\t1587059 원\n5위 : 성동구\t1581830 원\n6위 : 중구\t1576351 원\n7위 : 마포구\t1541429 원\n8위 : 종로구\t1504799 원\n9위 : 양천구\t1462725 원\n10위 : 강서구\t1452815 원\n11위 : 영등포구\t1451760 원\n12위 : 동작구\t1438760 원\n13위 : 광진구\t1431412 원\n14위 : 강동구\t1408536 원\n15위 : 성북구\t1403073 원\n16위 : 은평구\t1381959 원\n17위 : 서대문구\t1379245 원\n18위 : 관악구\t1371172 원\n19위 : 도봉구\t1366498 원\n20위 : 구로구\t1364992 원\n21위 : 중랑구\t1356135 원\n22위 : 동대문구\t1334980 원\n23위 : 금천구\t1333950 원\n24위 : 노원구\t1330321 원\n25위 : 강북구\t1273048 원\n" ], [ "### 총수신금액 순위\ndic_ass = {}\nfor i in range(25):\n dic_ass[county[i]] = int(asset[i])\n\ndic_ass = dict(sorted(dic_ass.items(), key=lambda x:x[1], reverse=True))\nfor i in range(25):\n print(str(i+1) +'위 : ' +list(dic_ass.keys())[i] + '\\t' + str(list(dic_ass.values())[i]) + ' 원')", "1위 : 서초구\t6559954 원\n2위 : 강남구\t6541343 원\n3위 : 송파구\t5233921 원\n4위 : 용산구\t5016620 원\n5위 : 성동구\t4701616 원\n6위 : 마포구\t4656763 원\n7위 : 중구\t4401267 원\n8위 : 종로구\t4385111 원\n9위 : 동작구\t4373611 원\n10위 : 양천구\t4336122 원\n11위 : 영등포구\t4263833 원\n12위 : 강서구\t4084911 원\n13위 : 성북구\t3974936 원\n14위 : 강동구\t3965753 원\n15위 : 광진구\t3886307 원\n16위 : 서대문구\t3821235 원\n17위 : 도봉구\t3765938 원\n18위 : 구로구\t3743025 원\n19위 : 노원구\t3682922 원\n20위 : 은평구\t3681698 원\n21위 : 관악구\t3555417 원\n22위 : 동대문구\t3440613 원\n23위 : 중랑구\t3341291 원\n24위 : 금천구\t3193268 원\n25위 : 강북구\t2945819 원\n" ] ], [ [ "### 자치구 별 급여입금에 따른 총소비금액과 총수신금액", "_____no_output_____" ] ], [ [ "# x축 : 자치구 = county\n# y축 막대 : 급여입금 = pay\n# y축 선 green : 총소비금액\n# y축 선 skyblue : 총수신금액\n\nfig, ax1 = plt.subplots(figsize=(15,7))\n\nax1.bar(county, pay, color='lightgray', label='급여입금', alpha=0.5, width=0.7)\nax1.set_xlabel('자치구')\nax1.set_ylabel('급여입금')\n\nax2 = ax1.twinx()\nax2.plot(county, consumption,'-s', color='green', label='총소비금액', markersize=6, linewidth=3)\nax2.set_xlabel('자치구')\nax2.set_yticks([])\n\nax3 = ax1.twinx()\nax3.plot(county, asset,'-s', color='skyblue', label='총수신금액', markersize=6, linewidth=3)\nax3.set_xlabel('자치구')\nax3.set_yticks([])\n\nplt.show()", "_____no_output_____" ] ], [ [ "---", "_____no_output_____" ], [ "### (급여입금 - 총소비금액) 차이 비교하기", "_____no_output_____" ] ], [ [ "money = {}\nfor i in range(25):\n money[county[i]] = int(pay[i] - consumption[i])\n\ndic_money = dict(sorted(money.items(), key=lambda x:x[1], reverse=True))\nfor i in range(25):\n print(str(i+1) +'위 : ' +list(dic_money.keys())[i] + '\\t' + str(list(dic_money.values())[i]) + ' 원')", "1위 : 양천구\t2634258 원\n2위 : 서초구\t2564075 원\n3위 : 강남구\t2483481 원\n4위 : 송파구\t2311235 원\n5위 : 용산구\t2284506 원\n6위 : 성북구\t2265259 원\n7위 : 강북구\t2233571 원\n8위 : 종로구\t2179968 원\n9위 : 강동구\t2177842 원\n10위 : 중구\t2159053 원\n11위 : 서대문구\t2143752 원\n12위 : 영등포구\t2040490 원\n13위 : 마포구\t2030326 원\n14위 : 광진구\t1966907 원\n15위 : 성동구\t1960358 원\n16위 : 동작구\t1923920 원\n17위 : 구로구\t1896103 원\n18위 : 노원구\t1840671 원\n19위 : 동대문구\t1804903 원\n20위 : 강서구\t1709412 원\n21위 : 도봉구\t1690518 원\n22위 : 은평구\t1667676 원\n23위 : 금천구\t1559915 원\n24위 : 관악구\t1554498 원\n25위 : 중랑구\t1522568 원\n" ] ], [ [ "### 자치구 별 급여 대비 소비 경향 비교 (총소비금액 / 급여입금)\n- 급여, 총소비금액 모두 자치구 평균 값을 이용\n- 값이 높을 수록 소비 경향이 강함\n\n- 급여 대비 소비 경향이 있어야 할 것 같은 이유 \n: 강남구, 서초구와 같은 경우 자산이 많은 상위권에 속하고 그 자산에 비해 소비금액이 적은거인데 다른 구들에 비해 소비 경향이 너무 적게 나옴", "_____no_output_____" ] ], [ [ "money = {}\nfor i in range(25):\n money[county[i]] = round(consumption[i] / pay[i], 4)\n\nplt.figure(figsize=(15,5))\nplt.plot(list(money.keys()), list(money.values()))\nplt.show()\n\ndic_money = dict(sorted(money.items(), key=lambda x:x[1], reverse=True))\nfor i in range(25):\n print(str(i+1) +'위 : ' +list(dic_money.keys())[i] + '\\t' + str(list(dic_money.values())[i]))", "_____no_output_____" ] ], [ [ "### 자치구 별 자산 대비 소비 경향 비교 (총소비금액 / 총수신금액)\n- 총수신금액, 총소비금액 모두 자치구 평균 값을 이용\n- 값이 높을 수록 소비 경향이 강함", "_____no_output_____" ] ], [ [ "money = {}\nfor i in range(25):\n money[county[i]] = round(consumption[i] / asset[i], 4)\n\nplt.figure(figsize=(15,5))\nplt.plot(list(money.keys()), list(money.values()))\nplt.show()\n\ndic_money = dict(sorted(money.items(), key=lambda x:x[1], reverse=True))\nfor i in range(25):\n print(str(i+1) +'위 : ' +list(dic_money.keys())[i] + '\\t' + str(list(dic_money.values())[i]))", "_____no_output_____" ] ], [ [ "### 연령대 별 자산 대비 소비 경향 비교\n- 총수신금액 대비 총소비금액 비율 (총소비금액 / 총수신금액)\n- 값이 높을수록 소비경향 강함", "_____no_output_____" ] ], [ [ "### 강남구와 서초구에서만 연령대가 증가함에 따라 일정하게 감소함 (30대의 소비경향이 40대보다 큼)\n## 이외 다른 자치구는 모두 20대 -> 30대는 감소하였다가 30대->40대에서 소비경향이 증가하였고, \n# 그 이후에는 마찬가지로 일정하게 감소함", "_____no_output_____" ], [ "age_group = []\nfor i in range(2, 8):\n age_group.append(df['나이'] == i)\n\nman = df['성별'] == 1\nwoman = df['성별'] == 2 \n\ncounty_group = []\nfor i in range(25):\n county_group.append(df['지역구'] == county[i])", "_____no_output_____" ], [ "group = {}\nfor i in range(25):\n dic = {}\n for j in range(6):\n value = (df[age_group[j] & county_group[i]]['총소비금액']).mean() / (df[age_group[j] & county_group[i]]['총수신금액']).mean()\n dic[j+2] = round(value, 4) \n group[county[i]] = dic\n\nfor i in range(25):\n print(county[i], group[county[i]])\n plt.plot(list(group[county[i]].keys()), list(group[county[i]].values()))\n plt.show()", "강남구 {2: 0.2961, 3: 0.2874, 4: 0.277, 5: 0.2523, 6: 0.1997, 7: 0.1549}\n" ] ], [ [ "---", "_____no_output_____" ], [ "# ✍🏻", "_____no_output_____" ], [ "## 클ㄹㅓ스터링 - 총소비금액,총수신금액", "_____no_output_____" ] ], [ [ "from sklearn.cluster import KMeans\nfrom sklearn import preprocessing", "_____no_output_____" ], [ "clug = df.groupby('지역구')[['총수신금액', '총소비금액']].mean()\n\n# k=2,3,4\nfor i in range(2,5):\n estimator = KMeans(n_clusters = i)\n cluster_ids = estimator.fit_predict(clug)\n\n plt.scatter(clug['총소비금액'], clug['총수신금액'], c=cluster_ids)\n plt.xlabel(\"총소비금액\")\n plt.ylabel(\"총수신금액\")\n plt.show()", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code" ]
[ [ "markdown", "markdown", "markdown" ], [ "code", "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown", "markdown", "markdown", "markdown" ], [ "code", "code", "code", "code", "code", "code", "code" ], [ "markdown", "markdown" ], [ "code", "code", "code", "code" ], [ "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code", "code", "code", "code" ], [ "markdown", "markdown", "markdown" ], [ "code", "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown", "markdown", "markdown" ], [ "code", "code" ] ]
ece8baabef84f3086995fadc2123bb235c096682
9,997
ipynb
Jupyter Notebook
site/en/r2/tutorials/estimators/keras_model_to_estimator.ipynb
crypdra/docs
41ab06fd14b3a3dff933bb80b19ce46c7c5781cf
[ "Apache-2.0" ]
2
2019-10-25T18:51:16.000Z
2019-10-25T18:51:18.000Z
site/en/r2/tutorials/estimators/keras_model_to_estimator.ipynb
crypdra/docs
41ab06fd14b3a3dff933bb80b19ce46c7c5781cf
[ "Apache-2.0" ]
null
null
null
site/en/r2/tutorials/estimators/keras_model_to_estimator.ipynb
crypdra/docs
41ab06fd14b3a3dff933bb80b19ce46c7c5781cf
[ "Apache-2.0" ]
null
null
null
29.576923
301
0.517455
[ [ [ "##### Copyright 2019 The TensorFlow Authors.", "_____no_output_____" ] ], [ [ "#@title Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# https://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.", "_____no_output_____" ] ], [ [ "# How-to create an Estimator from a Keras model", "_____no_output_____" ], [ "<table class=\"tfo-notebook-buttons\" align=\"left\">\n <td>\n <a target=\"_blank\" href=\"https://www.tensorflow.org/beta/tutorials/estimators/keras_model_to_estimator\"><img src=\"https://www.tensorflow.org/images/tf_logo_32px.png\" />View on TensorFlow.org</a>\n </td>\n <td>\n <a target=\"_blank\" href=\"https://colab.research.google.com/github/tensorflow/docs/blob/master/site/en/r2/tutorials/estimators/keras_model_to_estimator.ipynb\"><img src=\"https://www.tensorflow.org/images/colab_logo_32px.png\" />Run in Google Colab</a>\n </td>\n <td>\n <a target=\"_blank\" href=\"https://github.com/tensorflow/docs/blob/master/site/en/r2/tutorials/estimators/keras_model_to_estimator.ipynb\"><img src=\"https://www.tensorflow.org/images/GitHub-Mark-32px.png\" />View source on GitHub</a>\n </td>\n <td>\n <a href=\"https://storage.googleapis.com/tensorflow_docs/docs/site/en/r2/tutorials/estimators/keras_model_to_estimator.ipynb\"><img src=\"https://www.tensorflow.org/images/download_logo_32px.png\" />Download notebook</a>\n </td>\n</table>", "_____no_output_____" ], [ "## Overview\n\nTensorFlow Estimators are fully supported in TensorFlow, and can be created from new and existing `tf.keras` models. This tutorial contains a complete, minimal example of that process.", "_____no_output_____" ], [ "## Setup", "_____no_output_____" ] ], [ [ "from __future__ import absolute_import, division, print_function, unicode_literals", "_____no_output_____" ], [ "try:\n # %tensorflow_version only exists in Colab.\n %tensorflow_version 2.x\nexcept Exception:\n pass", "_____no_output_____" ], [ "import tensorflow as tf\n\nimport numpy as np\nimport tensorflow_datasets as tfds", "_____no_output_____" ] ], [ [ "### Create a simple Keras model.", "_____no_output_____" ], [ "In Keras, you assemble *layers* to build *models*. A model is (usually) a graph\nof layers. The most common type of model is a stack of layers: the\n`tf.keras.Sequential` model.\n\nTo build a simple, fully-connected network (i.e. multi-layer perceptron):", "_____no_output_____" ] ], [ [ "model = tf.keras.models.Sequential([\n tf.keras.layers.Dense(16, activation='relu', input_shape=(4,)),\n tf.keras.layers.Dropout(0.2),\n tf.keras.layers.Dense(1, activation='sigmoid')\n])", "_____no_output_____" ] ], [ [ "Compile the model and get a summary.", "_____no_output_____" ] ], [ [ "model.compile(loss='categorical_crossentropy', optimizer='adam')\nmodel.summary()", "_____no_output_____" ] ], [ [ "### Create an input function\n\nUse the [Datasets API](../../guide/data.md) to scale to large datasets\nor multi-device training.\n\nEstimators need control of when and how their input pipeline is built. To allow this, they require an \"Input function\" or `input_fn`. The `Estimator` will call this function with no arguments. The `input_fn` must return a `tf.data.Dataset`.", "_____no_output_____" ] ], [ [ "def input_fn():\n split = tfds.Split.TRAIN\n dataset = tfds.load('iris', split=split, as_supervised=True)\n dataset = dataset.map(lambda features, labels: ({'dense_input':features}, labels))\n dataset = dataset.batch(32).repeat()\n return dataset", "_____no_output_____" ] ], [ [ "Test out your `input_fn`", "_____no_output_____" ] ], [ [ "for features_batch, labels_batch in input_fn().take(1):\n print(features_batch)\n print(labels_batch)", "_____no_output_____" ] ], [ [ "### Create an Estimator from the tf.keras model.\n\nA `tf.keras.Model` can be trained with the `tf.estimator` API by converting the\nmodel to an `tf.estimator.Estimator` object with\n`tf.keras.estimator.model_to_estimator`.", "_____no_output_____" ] ], [ [ "model_dir = \"/tmp/tfkeras_example/\"\nkeras_estimator = tf.keras.estimator.model_to_estimator(\n keras_model=model, model_dir=model_dir)", "_____no_output_____" ] ], [ [ "Train and evaluate the estimator.", "_____no_output_____" ] ], [ [ "keras_estimator.train(input_fn=input_fn, steps=25)\neval_result = keras_estimator.evaluate(input_fn=input_fn, steps=10)\nprint('Eval result: {}'.format(eval_result))", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code" ]
[ [ "markdown" ], [ "code" ], [ "markdown", "markdown", "markdown", "markdown" ], [ "code", "code", "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ] ]
ece8c301514eb56ad0bf94bf3e4397821122ac36
7,289
ipynb
Jupyter Notebook
Week2-GraphTheory/Pset2_GraphTheory.ipynb
sarpcelikel/brain-networks-course
e9474b0dc9b2b70a9f2a1fb9176a869a14718940
[ "MIT" ]
null
null
null
Week2-GraphTheory/Pset2_GraphTheory.ipynb
sarpcelikel/brain-networks-course
e9474b0dc9b2b70a9f2a1fb9176a869a14718940
[ "MIT" ]
null
null
null
Week2-GraphTheory/Pset2_GraphTheory.ipynb
sarpcelikel/brain-networks-course
e9474b0dc9b2b70a9f2a1fb9176a869a14718940
[ "MIT" ]
null
null
null
28.69685
376
0.592262
[ [ [ "### Week 2 Problem Set: Graph Theory\n\nIn this exercise we will work with real data from the C. Elegans nervous system, using data shared by the [WormAtlas](http://www.wormatlas.org/) database. We will treat it as an undirected connectome for the purposes of this exercise.\n\nFor some problems you will be provided with skeleton code - fill in the lines marked ```...``` with appropriate code to solve problem.", "_____no_output_____" ] ], [ [ "import pandas,numpy\nimport os,sys\nimport networkx as nx\nimport matplotlib.pyplot as plt\nimport scipy.stats\n\nfrom brainnetworks.utils import mk_random_graph\n\n%matplotlib inline\n\n# read the data from Wormatlas.org: see section 2.1 of http://www.wormatlas.org/neuronalwiring.html for details\n\ncelegans_connectome=pandas.read_excel('http://www.wormatlas.org/images/NeuronConnect.xls')\n\n# set up the graph\nGd = nx.DiGraph()\nfor i in celegans_connectome.index:\n Gd.add_edge(celegans_connectome.loc[i]['Neuron 1'],celegans_connectome.loc[i]['Neuron 2'])\n \nGu=Gd.to_undirected()\n\n# the graph has two connected components, so we will just keep the giant component\ncomponents=nx.connected_component_subgraphs(Gu)\nG=next(components)\n", "_____no_output_____" ] ], [ [ "### Problem 1:\n\na. Plot a histogram of the degree distribution, and print out the mean and maximum degree\n", "_____no_output_____" ], [ "b. Compute the average clustering coefficient for the connectome", "_____no_output_____" ] ], [ [ "ce_clustering = ...\nprint(ce_clustering)", "_____no_output_____" ] ], [ [ "c. Compute the average path length for the connectome", "_____no_output_____" ] ], [ [ "ce_avgpathlength = ...\nprint(ce_avgpathlength)", "_____no_output_____" ] ], [ [ "#### Problem 2\n\nFirst, use the provided code to create 100 random graphs with the same size as the C. Elegans connectome. For each graph, we compute its average clustering, average shortest path length, and maximum degree, and store those to a numpy array. (This may take a few minutes to run.)", "_____no_output_____" ] ], [ [ "# PROVIDED CODE\nnruns=100\nmeasures=['clustering','avgpathlength','maxdegree']\nresults=pandas.DataFrame(numpy.zeros((nruns,len(measures))),\n columns=measures)\n\nfor i in range(nruns):\n G_rand = mk_random_graph(G)\n results.iloc[i]['clustering']=nx.average_clustering(G_rand)\n results.iloc[i]['avgpathlength']=nx.average_shortest_path_length(G_rand)\n results.iloc[i]['maxdegree']=numpy.max([G_rand.degree[i] for i in G_rand.nodes])\n ", "_____no_output_____" ] ], [ [ "a. Compute the tail probability of the observed values of cluster and path length for the C. elegans network, in comparison to the random network results. That is, what is the probability of a value as or more extreme than the observed value wihtin the random graph distribution? You may find the function ```scipy.stats.percentileofscore()``` useful for this purpose.", "_____no_output_____" ] ], [ [ "### PROVIDED CODE: \n\np_clustering=...\nprint('Observed:',ce_clustering,'Random:',results['clustering'].mean(),'P<',p_clustering)\n\np_avgpathlength=...\nprint('Observed:',ce_avgpathlength,'Random:',results['avgpathlength'].mean(),'P<',p_avgpathlength)\n\np_maxdegree=...\nprint('Observed:',numpy.max(degree_vals),'Random:',results['maxdegree'].mean(),'P<',p_maxdegree)\n", "_____no_output_____" ] ], [ [ "Based on the results of this analysis, do you think that the C. Elegans connectome is a \"small world\" network\"? explain your answer.", "_____no_output_____" ], [ "ANSWER:\n ", "_____no_output_____" ], [ "### Problem 3:\n\nDetermine which neuron is the most important based on these four criteria:\n\n- degree centrality\n- betweenness centrality\n- closeness centrality\n- eigenvector centrality\n\nBecause networkx returns a dictionary when it computes centrality measures, we have provided a utility function to return the entry with the largest value.", "_____no_output_____" ] ], [ [ "### PROVIDED CODE\n\ndef get_max_from_dict(d):\n \"\"\"\n return the dict entry with the max value\n after https://stackoverflow.com/questions/268272/getting-key-with-maximum-value-in-dictionary\n \"\"\"\n return max(d, key=lambda key: d[key])\n\ncentral_nodes={}\n# compute degree centrality\ncentral_nodes['degree']=get_max_from_dict(...)\n# compute eigenvector centrality\ncentral_nodes['eigenvector']=get_max_from_dict(...)\n# compute betweenness centrality\ncentral_nodes['betweenness']=get_max_from_dict(...)\n# compute closeness centrality\ncentral_nodes['closeness']=get_max_from_dict(...)\n\nprint(central_nodes)", "_____no_output_____" ] ], [ [ "You should see that the most central node differs between the different measures. Given what you know about those measures, explain why this might be the case.", "_____no_output_____" ], [ "ANSWER:\n\n", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown" ]
[ [ "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown", "markdown" ] ]
ece8de4bbf4fe1fd7bd0d911e14311a8adc1af85
30,363
ipynb
Jupyter Notebook
notebooks/Introduction.ipynb
lwcook/horsetail-matching
f3d5f8d01249debbca978f412ce4eae017458119
[ "MIT" ]
2
2017-05-17T17:07:08.000Z
2018-03-29T12:42:36.000Z
notebooks/Introduction.ipynb
lwcook/horsetail-matching
f3d5f8d01249debbca978f412ce4eae017458119
[ "MIT" ]
null
null
null
notebooks/Introduction.ipynb
lwcook/horsetail-matching
f3d5f8d01249debbca978f412ce4eae017458119
[ "MIT" ]
null
null
null
112.873606
12,618
0.859105
[ [ [ "Horsetail matching is a method for optimizing under uncertainty. \n\nWe are looking to minimize a quantity of interest, q, which is a function of design variables (that we can control), x, and uncertain variables (which are uncontrollable), u. Since u is uncertain, q at a given design x is also uncertain. We therefore need to define some measure of the behaviour of q under uncertainty to minimize instead. With horseail matching this measure is the difference between the bounds on the CDF (the horsetail plot) of q for a given design and targets for these bounds. \n\nFor further details see the website: http://www-edc.eng.cam.ac.uk/aerotools/horsetailmatching/ which has links to the relevant publications. \n\n**This tutorial illustrates how to setup and run a simple horsetail matching optimization.** \nNote that for clarification on how any of the module works, see the documentation available at: http://www-edc.eng.cam.ac.uk/aerotools/horsetailmatching/documentation/\n\nTo begin with, we will only consider probabilistic uncertainties. This requires a probability distribution to be assigned to all of our uncertain parameters, and this is propagated to give a probability distribution of our quantity of interest for a given design x. Therefore the horsetail plot is the CDF itself and we are essentially doing CDF matching.\n\nFirstly we need to import the functions and classes we will use from the horsetail matching module...", "_____no_output_____" ] ], [ [ "from horsetailmatching import UncertainParameter, UniformParameter, GaussianParameter, HorsetailMatching\nfrom horsetailmatching.demoproblems import TP1", "_____no_output_____" ] ], [ [ "A horsetail matching object requires a function that takes two arguments: the value of the design variables, and the value of the uncertainties; it should return the value quantity of interest. Here we will use a test problem that comes that is part of the horsetail matching module. In reality, this function is likely to be a computationally expensive simulation, for example a 3D computational fluid dynamics simulation of a wing.", "_____no_output_____" ] ], [ [ "my_func = TP1\nprint TP1(x=[0, 1], u=[1, 0])", "0.1\n" ] ], [ [ "Next we must define the input uncertainties to the problem, by creating horsetail matching parameters. We can assign the parameters a distribution by using the base UncertainParameter class and defining a pdf function, or we can use the specific distribution subclasses such as UniformParameter and GaussianParameter. \n\nHere we create a uniformly distributed uncertainty and a gaussian uncertainty.\n\nThen we create the HorsetailMatching object which will do the evalation of the horsetail matching metric. We can specify the target inverse CDF using the ftarget argument (by default a standard target of t(h) = 0 is used). ", "_____no_output_____" ] ], [ [ "def myPDF(u):\n if u > 1 or u < -1:\n return 0\n else:\n return 0.5\n\nu_uniform = UncertainParameter(myPDF, lower_bound=-1, upper_bound=1)\nu_gaussian = GaussianParameter(mean=0, standard_deviation=1)\n\ndef my_target(h): \n return 0\n\ntheHM = HorsetailMatching(my_func, [u_uniform, u_gaussian], ftarget=my_target)", "_____no_output_____" ] ], [ [ "Now we use the evalMetric method to evaluate the horsetail matching metric at a design point:", "_____no_output_____" ] ], [ [ "print(theHM.evalMetric(x=[1,1]))", "1.5954675519106958\n" ] ], [ [ "We can use the getHorsetail() method to visualize the behaviour under uncertainty. It returns plottable lists of values for the two horsetail curves (the upper and lower bounds on the CDF) as the first two returned tuples. In this case the two bounds are coincidental and the horsetail plot is the CDF. \n\nWe can then plot this using the matplotlib module for plotting in python. ", "_____no_output_____" ] ], [ [ "import matplotlib.pyplot as plt\n(x1, y1, t1), (x2, y2, t2), _ = theHM.getHorsetail()\nplt.plot(x1, y1, 'b', label='CDF')\nplt.plot(t1, y1, 'k--', label='Target')\nplt.xlim([-1, 6])\nplt.xlabel('Quantity of Interest')\nplt.legend(loc='lower right')\nplt.show()", "_____no_output_____" ] ], [ [ "Now if we want to use this within an optimization, its a simple as passing theHM.evalMetric to whichever optimizer we like. For example, using the scipy optimize module's minimize function:", "_____no_output_____" ] ], [ [ "from scipy.optimize import minimize\n\nsolution = minimize(theHM.evalMetric, x0=[3,2], method='Nelder-Mead')\nprint(solution)", " status: 0\n nfev: 107\n success: True\n fun: 0.0\n x: array([-2.46257592, 2.46257594])\n message: 'Optimization terminated successfully.'\n nit: 54\n" ], [ "(x1, y1, t1), (x2, y2, t2), _ = theHM.getHorsetail()\nplt.plot(x1, y1, 'b', label='CDF')\nplt.plot([theHM.ftarget(y) for y in y1], y1, 'k--', label='Target')\nplt.xlim([-1, 6])\nplt.xlabel('Quantity of Interest')\nplt.legend(loc='lower right')\nplt.show()", "_____no_output_____" ] ], [ [ "For this simple test problem the minimum is where the CDF is a step function at 0, and so the horsetail matching metric is also 0. ", "_____no_output_____" ], [ "This concludes our introduction to horsetail matching. \n\nIn the next tutorial, we do horsetail matching when not all of our uncertainties can be represented probabilistically: http://nbviewer.jupyter.org/github/lwcook/horsetail-matching/blob/master/notebooks/MixedUncertainties.ipynb\n\nFor other tutorials, please visit http://www-edc.eng.cam.ac.uk/aerotools/horsetailmatching/", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown" ]
[ [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown", "markdown" ] ]
ece9045c6df82ed6b86742f43d9cf8f2aab311fe
1,573
ipynb
Jupyter Notebook
Python basics practice/Python 3 (23)/While Loops and Incrementing - Exercise_Py3.ipynb
rachithh/data-science
1f7c5678094fc3acfda30cb00f9de93a2974f505
[ "MIT" ]
null
null
null
Python basics practice/Python 3 (23)/While Loops and Incrementing - Exercise_Py3.ipynb
rachithh/data-science
1f7c5678094fc3acfda30cb00f9de93a2974f505
[ "MIT" ]
null
null
null
Python basics practice/Python 3 (23)/While Loops and Incrementing - Exercise_Py3.ipynb
rachithh/data-science
1f7c5678094fc3acfda30cb00f9de93a2974f505
[ "MIT" ]
null
null
null
18.72619
96
0.506039
[ [ [ "## While Loops and Incrementing", "_____no_output_____" ], [ "*Suggested Answers follow (usually there are multiple ways to solve a problem in Python).*", "_____no_output_____" ], [ "Create a while loop that will print all odd numbers from 0 to 30 on the same row. \n<br />\n*Hint: There are two ways in which you can create the odd values!*", "_____no_output_____" ] ], [ [ "x=1\nwhile x<=30:\n print(x,end=' ')\n x=x+2", "1 3 5 7 9 11 13 15 17 19 21 23 25 27 29 " ] ] ]
[ "markdown", "code" ]
[ [ "markdown", "markdown", "markdown" ], [ "code" ] ]
ece90673266378e1016498d5a6008e43c5d5dfae
92,034
ipynb
Jupyter Notebook
Writing Functions in Python.ipynb
Yuleii/programming_notebook
5d4ba849fcff1ea6dcdd0e2259193b2509b17026
[ "MIT" ]
2
2021-08-31T20:59:00.000Z
2022-03-13T09:59:38.000Z
Writing Functions in Python.ipynb
Yuleii/programming_notebook
5d4ba849fcff1ea6dcdd0e2259193b2509b17026
[ "MIT" ]
null
null
null
Writing Functions in Python.ipynb
Yuleii/programming_notebook
5d4ba849fcff1ea6dcdd0e2259193b2509b17026
[ "MIT" ]
2
2021-09-25T12:50:24.000Z
2022-03-21T08:27:03.000Z
28.007912
13,548
0.512517
[ [ [ "<h1>Table of Contents<span class=\"tocSkip\"></span></h1>\n<div class=\"toc\"><ul class=\"toc-item\"><li><span><a href=\"#Best-Practices\" data-toc-modified-id=\"Best-Practices-1\"><span class=\"toc-item-num\">1&nbsp;&nbsp;</span>Best Practices</a></span><ul class=\"toc-item\"><li><span><a href=\"#Docstrings\" data-toc-modified-id=\"Docstrings-1.1\"><span class=\"toc-item-num\">1.1&nbsp;&nbsp;</span>Docstrings</a></span><ul class=\"toc-item\"><li><span><a href=\"#Anatomy-of-a-docstring\" data-toc-modified-id=\"Anatomy-of-a-docstring-1.1.1\"><span class=\"toc-item-num\">1.1.1&nbsp;&nbsp;</span>Anatomy of a docstring</a></span></li><li><span><a href=\"#Docstring-formats\" data-toc-modified-id=\"Docstring-formats-1.1.2\"><span class=\"toc-item-num\">1.1.2&nbsp;&nbsp;</span>Docstring formats</a></span><ul class=\"toc-item\"><li><span><a href=\"#Google-style\" data-toc-modified-id=\"Google-style-1.1.2.1\"><span class=\"toc-item-num\">1.1.2.1&nbsp;&nbsp;</span>Google style</a></span></li><li><span><a href=\"#Numpydoc\" data-toc-modified-id=\"Numpydoc-1.1.2.2\"><span class=\"toc-item-num\">1.1.2.2&nbsp;&nbsp;</span><a href=\"https://numpydoc.readthedocs.io/en/latest/format.html\" target=\"_blank\">Numpydoc</a></a></span></li></ul></li><li><span><a href=\"#Retrieving-docstrings\" data-toc-modified-id=\"Retrieving-docstrings-1.1.3\"><span class=\"toc-item-num\">1.1.3&nbsp;&nbsp;</span>Retrieving docstrings</a></span></li><li><span><a href=\"#Exercise-1\" data-toc-modified-id=\"Exercise-1-1.1.4\"><span class=\"toc-item-num\">1.1.4&nbsp;&nbsp;</span>Exercise 1</a></span></li><li><span><a href=\"#Exercise-2\" data-toc-modified-id=\"Exercise-2-1.1.5\"><span class=\"toc-item-num\">1.1.5&nbsp;&nbsp;</span>Exercise 2</a></span></li></ul></li><li><span><a href=\"#DRY-and-&quot;Do-One-Thing&quot;\" data-toc-modified-id=\"DRY-and-&quot;Do-One-Thing&quot;-1.2\"><span class=\"toc-item-num\">1.2&nbsp;&nbsp;</span>DRY and \"Do One Thing\"</a></span><ul class=\"toc-item\"><li><span><a href=\"#Don't-repeat-yourself-(DRY)\" data-toc-modified-id=\"Don't-repeat-yourself-(DRY)-1.2.1\"><span class=\"toc-item-num\">1.2.1&nbsp;&nbsp;</span>Don't repeat yourself (DRY)</a></span></li><li><span><a href=\"#Do-One-Thing\" data-toc-modified-id=\"Do-One-Thing-1.2.2\"><span class=\"toc-item-num\">1.2.2&nbsp;&nbsp;</span>Do One Thing</a></span></li></ul></li><li><span><a href=\"#Pass-by-assignment\" data-toc-modified-id=\"Pass-by-assignment-1.3\"><span class=\"toc-item-num\">1.3&nbsp;&nbsp;</span>Pass by assignment</a></span><ul class=\"toc-item\"><li><span><a href=\"#surprising-example\" data-toc-modified-id=\"surprising-example-1.3.1\"><span class=\"toc-item-num\">1.3.1&nbsp;&nbsp;</span>surprising example</a></span></li><li><span><a href=\"#Mutable-default-arguments-are-dangerous\" data-toc-modified-id=\"Mutable-default-arguments-are-dangerous-1.3.2\"><span class=\"toc-item-num\">1.3.2&nbsp;&nbsp;</span>Mutable default arguments are dangerous</a></span></li><li><span><a href=\"#Best-practice-for-default-arguments\" data-toc-modified-id=\"Best-practice-for-default-arguments-1.3.3\"><span class=\"toc-item-num\">1.3.3&nbsp;&nbsp;</span>Best practice for default arguments</a></span></li></ul></li></ul></li><li><span><a href=\"#Context-Managers\" data-toc-modified-id=\"Context-Managers-2\"><span class=\"toc-item-num\">2&nbsp;&nbsp;</span>Context Managers</a></span><ul class=\"toc-item\"><li><span><a href=\"#Using-context-managers\" data-toc-modified-id=\"Using-context-managers-2.1\"><span class=\"toc-item-num\">2.1&nbsp;&nbsp;</span>Using context managers</a></span><ul class=\"toc-item\"><li><span><a href=\"#A-real-world-example\" data-toc-modified-id=\"A-real-world-example-2.1.1\"><span class=\"toc-item-num\">2.1.1&nbsp;&nbsp;</span>A real-world example</a></span></li><li><span><a href=\"#Exercise:-The-speed-of-cats\" data-toc-modified-id=\"Exercise:-The-speed-of-cats-2.1.2\"><span class=\"toc-item-num\">2.1.2&nbsp;&nbsp;</span>Exercise: The speed of cats</a></span></li></ul></li><li><span><a href=\"#Writing-context-managers\" data-toc-modified-id=\"Writing-context-managers-2.2\"><span class=\"toc-item-num\">2.2&nbsp;&nbsp;</span>Writing context managers</a></span><ul class=\"toc-item\"><li><span><a href=\"#How-to-create-a-context-manager\" data-toc-modified-id=\"How-to-create-a-context-manager-2.2.1\"><span class=\"toc-item-num\">2.2.1&nbsp;&nbsp;</span>How to create a context manager</a></span></li><li><span><a href=\"#The-&quot;yield&quot;-keyword\" data-toc-modified-id=\"The-&quot;yield&quot;-keyword-2.2.2\"><span class=\"toc-item-num\">2.2.2&nbsp;&nbsp;</span>The \"yield\" keyword</a></span></li><li><span><a href=\"#Setup-and-teardown\" data-toc-modified-id=\"Setup-and-teardown-2.2.3\"><span class=\"toc-item-num\">2.2.3&nbsp;&nbsp;</span>Setup and teardown</a></span></li><li><span><a href=\"#Yielding-a-value-or-None\" data-toc-modified-id=\"Yielding-a-value-or-None-2.2.4\"><span class=\"toc-item-num\">2.2.4&nbsp;&nbsp;</span>Yielding a value or None</a></span></li><li><span><a href=\"#Exercise:-The-timer()-context-manager\" data-toc-modified-id=\"Exercise:-The-timer()-context-manager-2.2.5\"><span class=\"toc-item-num\">2.2.5&nbsp;&nbsp;</span>Exercise: The timer() context manager</a></span></li><li><span><a href=\"#Exercise:-A-read-only-open()-context-manager\" data-toc-modified-id=\"Exercise:-A-read-only-open()-context-manager-2.2.6\"><span class=\"toc-item-num\">2.2.6&nbsp;&nbsp;</span>Exercise: A read-only open() context manager</a></span></li></ul></li><li><span><a href=\"#Advanced-topics\" data-toc-modified-id=\"Advanced-topics-2.3\"><span class=\"toc-item-num\">2.3&nbsp;&nbsp;</span>Advanced topics</a></span><ul class=\"toc-item\"><li><span><a href=\"#Nested-contexts\" data-toc-modified-id=\"Nested-contexts-2.3.1\"><span class=\"toc-item-num\">2.3.1&nbsp;&nbsp;</span>Nested contexts</a></span></li><li><span><a href=\"#Handling-errors\" data-toc-modified-id=\"Handling-errors-2.3.2\"><span class=\"toc-item-num\">2.3.2&nbsp;&nbsp;</span>Handling errors</a></span></li><li><span><a href=\"#Context-manager-patterns\" data-toc-modified-id=\"Context-manager-patterns-2.3.3\"><span class=\"toc-item-num\">2.3.3&nbsp;&nbsp;</span>Context manager patterns</a></span></li><li><span><a href=\"#Exercise:-Scraping-the-NASDAQ\" data-toc-modified-id=\"Exercise:-Scraping-the-NASDAQ-2.3.4\"><span class=\"toc-item-num\">2.3.4&nbsp;&nbsp;</span>Exercise: Scraping the NASDAQ</a></span></li><li><span><a href=\"#Exercise:-Changing-the-working-directory\" data-toc-modified-id=\"Exercise:-Changing-the-working-directory-2.3.5\"><span class=\"toc-item-num\">2.3.5&nbsp;&nbsp;</span>Exercise: Changing the working directory</a></span></li></ul></li></ul></li><li><span><a href=\"#Decorators\" data-toc-modified-id=\"Decorators-3\"><span class=\"toc-item-num\">3&nbsp;&nbsp;</span>Decorators</a></span><ul class=\"toc-item\"><li><span><a href=\"#Functions-are-objects\" data-toc-modified-id=\"Functions-are-objects-3.1\"><span class=\"toc-item-num\">3.1&nbsp;&nbsp;</span>Functions are objects</a></span><ul class=\"toc-item\"><li><span><a href=\"#Functions-are-just-another-type-of-object\" data-toc-modified-id=\"Functions-are-just-another-type-of-object-3.1.1\"><span class=\"toc-item-num\">3.1.1&nbsp;&nbsp;</span>Functions are just another type of object</a></span></li><li><span><a href=\"#Functions-as-variables\" data-toc-modified-id=\"Functions-as-variables-3.1.2\"><span class=\"toc-item-num\">3.1.2&nbsp;&nbsp;</span>Functions as variables</a></span></li><li><span><a href=\"#Lists-and-dictionaries-of-functions\" data-toc-modified-id=\"Lists-and-dictionaries-of-functions-3.1.3\"><span class=\"toc-item-num\">3.1.3&nbsp;&nbsp;</span>Lists and dictionaries of functions</a></span></li><li><span><a href=\"#Referencing-a-function\" data-toc-modified-id=\"Referencing-a-function-3.1.4\"><span class=\"toc-item-num\">3.1.4&nbsp;&nbsp;</span>Referencing a function</a></span></li><li><span><a href=\"#Functions-as-arguments\" data-toc-modified-id=\"Functions-as-arguments-3.1.5\"><span class=\"toc-item-num\">3.1.5&nbsp;&nbsp;</span>Functions as arguments</a></span></li><li><span><a href=\"#Defining-a-function-inside-another-function\" data-toc-modified-id=\"Defining-a-function-inside-another-function-3.1.6\"><span class=\"toc-item-num\">3.1.6&nbsp;&nbsp;</span>Defining a function inside another function</a></span></li></ul></li><li><span><a href=\"#Scope\" data-toc-modified-id=\"Scope-3.2\"><span class=\"toc-item-num\">3.2&nbsp;&nbsp;</span>Scope</a></span><ul class=\"toc-item\"><li><span><a href=\"#The-global-keyword\" data-toc-modified-id=\"The-global-keyword-3.2.1\"><span class=\"toc-item-num\">3.2.1&nbsp;&nbsp;</span>The global keyword</a></span></li><li><span><a href=\"#The-nonlocal-keyword\" data-toc-modified-id=\"The-nonlocal-keyword-3.2.2\"><span class=\"toc-item-num\">3.2.2&nbsp;&nbsp;</span>The nonlocal keyword</a></span></li></ul></li><li><span><a href=\"#Closures\" data-toc-modified-id=\"Closures-3.3\"><span class=\"toc-item-num\">3.3&nbsp;&nbsp;</span>Closures</a></span><ul class=\"toc-item\"><li><span><a href=\"#Attaching-nonlocal-variables-to-nested-functions\" data-toc-modified-id=\"Attaching-nonlocal-variables-to-nested-functions-3.3.1\"><span class=\"toc-item-num\">3.3.1&nbsp;&nbsp;</span>Attaching nonlocal variables to nested functions</a></span></li><li><span><a href=\"#Closures-and-deletion\" data-toc-modified-id=\"Closures-and-deletion-3.3.2\"><span class=\"toc-item-num\">3.3.2&nbsp;&nbsp;</span>Closures and deletion</a></span></li><li><span><a href=\"#Closures-and-overwriting\" data-toc-modified-id=\"Closures-and-overwriting-3.3.3\"><span class=\"toc-item-num\">3.3.3&nbsp;&nbsp;</span>Closures and overwriting</a></span></li></ul></li><li><span><a href=\"#Why-does-all-of-this-matter?\" data-toc-modified-id=\"Why-does-all-of-this-matter?-3.4\"><span class=\"toc-item-num\">3.4&nbsp;&nbsp;</span>Why does all of this matter?</a></span></li><li><span><a href=\"#Decorators\" data-toc-modified-id=\"Decorators-3.5\"><span class=\"toc-item-num\">3.5&nbsp;&nbsp;</span>Decorators</a></span><ul class=\"toc-item\"><li><span><a href=\"#The-double_args-decorator\" data-toc-modified-id=\"The-double_args-decorator-3.5.1\"><span class=\"toc-item-num\">3.5.1&nbsp;&nbsp;</span>The double_args decorator</a></span></li><li><span><a href=\"#Decorator-syntax\" data-toc-modified-id=\"Decorator-syntax-3.5.2\"><span class=\"toc-item-num\">3.5.2&nbsp;&nbsp;</span>Decorator syntax</a></span></li><li><span><a href=\"#Exercise:-Defining-a-decorator\" data-toc-modified-id=\"Exercise:-Defining-a-decorator-3.5.3\"><span class=\"toc-item-num\">3.5.3&nbsp;&nbsp;</span>Exercise: Defining a decorator</a></span></li></ul></li><li><span><a href=\"#Real-world-examples\" data-toc-modified-id=\"Real-world-examples-3.6\"><span class=\"toc-item-num\">3.6&nbsp;&nbsp;</span>Real-world examples</a></span><ul class=\"toc-item\"><li><span><a href=\"#Time-a-function\" data-toc-modified-id=\"Time-a-function-3.6.1\"><span class=\"toc-item-num\">3.6.1&nbsp;&nbsp;</span>Time a function</a></span></li><li><span><a href=\"#Using-timer()\" data-toc-modified-id=\"Using-timer()-3.6.2\"><span class=\"toc-item-num\">3.6.2&nbsp;&nbsp;</span>Using timer()</a></span></li><li><span><a href=\"#When-to-use-decorators\" data-toc-modified-id=\"When-to-use-decorators-3.6.3\"><span class=\"toc-item-num\">3.6.3&nbsp;&nbsp;</span>When to use decorators</a></span></li><li><span><a href=\"#Exercise:-Print-the-return-type\" data-toc-modified-id=\"Exercise:-Print-the-return-type-3.6.4\"><span class=\"toc-item-num\">3.6.4&nbsp;&nbsp;</span>Exercise: Print the return type</a></span></li><li><span><a href=\"#Exercise:-Counter\" data-toc-modified-id=\"Exercise:-Counter-3.6.5\"><span class=\"toc-item-num\">3.6.5&nbsp;&nbsp;</span>Exercise: Counter</a></span></li></ul></li><li><span><a href=\"#Decorators-and-metadata\" data-toc-modified-id=\"Decorators-and-metadata-3.7\"><span class=\"toc-item-num\">3.7&nbsp;&nbsp;</span>Decorators and metadata</a></span><ul class=\"toc-item\"><li><span><a href=\"#fix:-The-timer-decorator\" data-toc-modified-id=\"fix:-The-timer-decorator-3.7.1\"><span class=\"toc-item-num\">3.7.1&nbsp;&nbsp;</span>fix: The timer decorator</a></span></li><li><span><a href=\"#Access-to-the-original-function\" data-toc-modified-id=\"Access-to-the-original-function-3.7.2\"><span class=\"toc-item-num\">3.7.2&nbsp;&nbsp;</span>Access to the original function</a></span></li></ul></li><li><span><a href=\"#Decorators-that-take-arguments\" data-toc-modified-id=\"Decorators-that-take-arguments-3.8\"><span class=\"toc-item-num\">3.8&nbsp;&nbsp;</span>Decorators that take arguments</a></span><ul class=\"toc-item\"><li><span><a href=\"#A-decorator-factory\" data-toc-modified-id=\"A-decorator-factory-3.8.1\"><span class=\"toc-item-num\">3.8.1&nbsp;&nbsp;</span>A decorator factory</a></span></li></ul></li><li><span><a href=\"#Timeout():-a-real-world-example\" data-toc-modified-id=\"Timeout():-a-real-world-example-3.9\"><span class=\"toc-item-num\">3.9&nbsp;&nbsp;</span>Timeout(): a real world example</a></span><ul class=\"toc-item\"><li><span><a href=\"#Exercise:-Tag-your-functions\" data-toc-modified-id=\"Exercise:-Tag-your-functions-3.9.1\"><span class=\"toc-item-num\">3.9.1&nbsp;&nbsp;</span>Exercise: Tag your functions</a></span></li><li><span><a href=\"#Exercise:-Check-the-return-type\" data-toc-modified-id=\"Exercise:-Check-the-return-type-3.9.2\"><span class=\"toc-item-num\">3.9.2&nbsp;&nbsp;</span>Exercise: Check the return type</a></span></li></ul></li></ul></li></ul></div>", "_____no_output_____" ], [ "# Best Practices", "_____no_output_____" ], [ "## Docstrings", "_____no_output_____" ], [ "### Anatomy of a docstring", "_____no_output_____" ] ], [ [ "def function_name(arguments):\n \"\"\" \n Description of what the function does.\n \n Description of the arguments, if any. \n \n Description of the return value(s), if any. \n \n Description of errors raised, if any. \n \n Optional extra notes or examples of usage. \n \"\"\"", "_____no_output_____" ] ], [ [ "### Docstring formats\n\n- Google Style\n\n- Numpydoc\n\n- reStructuredText\n\n- EpyText", "_____no_output_____" ], [ "#### Google style ", "_____no_output_____" ] ], [ [ "def function(arg_1, arg_2=42):\n \"\"\"Description of what the function does. \n \n Args: arg_1 (str): Description of arg_1 that can break onto the next line \n if needed. \n arg_2 (int, optional): Write optional when an argument has a default \n value. \n \n Returns: \n bool: Optional description of the return value \n Extra lines are not indented. \n \n Raises: \n ValueError: Include any error types that the function intentionally\n raises. \n \n Notes: \n See https://www.datacamp.com/community/tutorials/docstrings-python \n for more info. \n \"\"\"\n", "_____no_output_____" ] ], [ [ "#### [Numpydoc](https://numpydoc.readthedocs.io/en/latest/format.html)", "_____no_output_____" ] ], [ [ "def function(arg_1, arg_2=42):\n \"\"\" Description of what the function does.\n \n Parameters \n ---------- \n arg_1 : expected type of arg_1 \n Description of arg_1. \n arg_2 : int, optional \n Write optional when an argument has a default value.\n Default=42.\n \n Returns \n ------- \n The type of the return value\n Can include a description of the return value.\n Replace \"Returns\" with \"Yields\" if this function is a generator. \n \"\"\"", "_____no_output_____" ] ], [ [ "### Retrieving docstrings", "_____no_output_____" ] ], [ [ "def the_answer():\n \"\"\"Return the answer to life, the universe, and everything.\n \n Returns: \n int\n \"\"\"\n return 42\nprint(the_answer.__doc__)", "Return the answer to life, the universe, and everything.\n \n Returns: \n int\n \n" ], [ "import inspect \nprint(inspect.getdoc(the_answer))", "Return the answer to life, the universe, and everything.\n\nReturns: \n int\n" ] ], [ [ "### Exercise 1", "_____no_output_____" ] ], [ [ "def count_letter(content, letter):\n \"\"\"Count the number of times `letter` appears in `content`.\n\n Args:\n content (str): The string to search.\n letter (str): The letter to search for.\n\n Returns:\n int\n\n # Add a section detailing what errors might be raised\n Raises:\n ValueError: If `letter` is not a one-character string.\n \"\"\"\n if (not isinstance(letter, str)) or len(letter) != 1:\n raise ValueError('`letter` must be a single character string.')\n return len([char for char in content if char == letter])", "_____no_output_____" ] ], [ [ "### Exercise 2", "_____no_output_____" ] ], [ [ "# Get the docstring with an attribute of count_letter()\ndocstring = count_letter.__doc__\n\nborder = '#' * 28\nprint('{}\\n{}\\n{}'.format(border, docstring, border))", "############################\nCount the number of times `letter` appears in `content`.\n\n Args:\n content (str): The string to search.\n letter (str): The letter to search for.\n\n Returns:\n int\n\n # Add a section detailing what errors might be raised\n Raises:\n ValueError: If `letter` is not a one-character string.\n \n############################\n" ], [ "import inspect\n\n# Get the docstring with a function from the inspect module\ndocstring = inspect.getdoc(count_letter)\n\nborder = '#' * 28\nprint('{}\\n{}\\n{}'.format(border, docstring, border))", "############################\nCount the number of times `letter` appears in `content`.\n\nArgs:\n content (str): The string to search.\n letter (str): The letter to search for.\n\nReturns:\n int\n\n# Add a section detailing what errors might be raised\nRaises:\n ValueError: If `letter` is not a one-character string.\n############################\n" ], [ "# Use the inspect module again to get the docstring for any function being passed to the build_tooltip() function.\n\ndef build_tooltip(function):\n \"\"\"Create a tooltip for any function that shows the \n function's docstring.\n\n Args:\n function (callable): The function we want a tooltip for.\n\n Returns:\n str\n \"\"\"\n # Use 'inspect' to get the docstring\n docstring = inspect.getdoc(function)\n border = '#' * 28\n return '{}\\n{}\\n{}'.format(border, docstring, border)\n\nprint(build_tooltip(count_letter))\nprint(build_tooltip(range))\nprint(build_tooltip(print))", "############################\nCount the number of times `letter` appears in `content`.\n\nArgs:\n content (str): The string to search.\n letter (str): The letter to search for.\n\nReturns:\n int\n\n# Add a section detailing what errors might be raised\nRaises:\n ValueError: If `letter` is not a one-character string.\n############################\n############################\nrange(stop) -> range object\nrange(start, stop[, step]) -> range object\n\nReturn an object that produces a sequence of integers from start (inclusive)\nto stop (exclusive) by step. range(i, j) produces i, i+1, i+2, ..., j-1.\nstart defaults to 0, and stop is omitted! range(4) produces 0, 1, 2, 3.\nThese are exactly the valid indices for a list of 4 elements.\nWhen step is given, it specifies the increment (or decrement).\n############################\n############################\nprint(value, ..., sep=' ', end='\\n', file=sys.stdout, flush=False)\n\nPrints the values to a stream, or to sys.stdout by default.\nOptional keyword arguments:\nfile: a file-like object (stream); defaults to the current sys.stdout.\nsep: string inserted between values, default a space.\nend: string appended after the last value, default a newline.\nflush: whether to forcibly flush the stream.\n############################\n" ] ], [ [ "## DRY and \"Do One Thing\"", "_____no_output_____" ], [ "### Don't repeat yourself (DRY)", "_____no_output_____" ] ], [ [ "def load_and_plot(path):\n \"\"\"Load a data set and plot the first two principal components. \n \n Args: \n path (str): The location of a CSV file. \n \n Returns: \n tuple of ndarray: (features, labels) \n \"\"\"\n # load the data\n data = pd.read_csv(path) \n y = data['label'].values \n X = data[col for col in train.columns if col != 'label'].values \n \n # plot the first two principal components\n pca = PCA(n_components=2).fit_transform(X) \n plt.scatter(pca[:,0], pca[:,1])\n \n # return loaded dat\n return X, y\n\n\ntrain_X, train_y = load_and_plot('train.csv')\nval_X, val_y = load_and_plot('validation.csv')\ntest_X, test_y = load_and_plot('test.csv')", "_____no_output_____" ] ], [ [ "### Do One Thing", "_____no_output_____" ] ], [ [ "def load_data(path):\n \"\"\"Load a data set. \n Args: \n path (str): The location of a CSV file. \n \n Returns: \n tuple of ndarray: (features, labels) \n \"\"\" \n data = pd.read_csv(path) \n y = data['labels'].values \n X = data[col for col in data.columns if col != 'labels'].values\n \n return X, y\n\n\ndef plot_data(X):\n \"\"\"Plot the first two principal components of a matrix.\n \n Args: \n X (numpy.ndarray): The data to plot. \n \"\"\" \n pca = PCA(n_components=2).fit_transform(X) \n plt.scatter(pca[:,0], pca[:,1])", "_____no_output_____" ] ], [ [ "## Pass by assignment", "_____no_output_____" ], [ "### surprising example", "_____no_output_____" ], [ "Immutable\n- int\n- float\n- bool\n- string\n- bytes\n- tuple\n- frozenset\n- None\n\nMutable\n- list\n- dict\n- set\n- bytearray\n- objects\n- functions\n- almost everything else!", "_____no_output_____" ] ], [ [ "def foo(x): \n x[0] = 99\n \nmy_list = [1, 2, 3]\nfoo(my_list)\nprint(my_list)", "[99, 2, 3]\n" ], [ "# intergers are Immutable\ndef bar(x):\n x = x + 90\n \nmy_var = 3\nbar(my_var)\nprint(my_var)", "3\n" ] ], [ [ "### Mutable default arguments are dangerous", "_____no_output_____" ] ], [ [ "def foo(var=[]):\n var.append(1)\n return var\n\nfoo()", "_____no_output_____" ], [ "foo()", "_____no_output_____" ], [ "def foo(var=None):\n if var is None:\n var = []\n var.append(1)\n return var\n\nfoo()", "_____no_output_____" ], [ "foo()", "_____no_output_____" ] ], [ [ "### Best practice for default arguments", "_____no_output_____" ], [ "Bad way to assign mutable default arguments", "_____no_output_____" ] ], [ [ "def add_column(values, df=pandas.DataFrame()):\n \"\"\"Add a column of `values` to a DataFrame `df`.\n The column will be named \"col_<n>\" where \"n\" is\n the numerical index of the column.\n\n Args:\n values (iterable): The values of the new column\n df (DataFrame, optional): The DataFrame to update.\n If no DataFrame is passed, one is created by default.\n\n Returns:\n DataFrame\n \"\"\"\n df['col_{}'.format(len(df.columns))] = values\n return df", "_____no_output_____" ] ], [ [ "Good way to assign mutable default arguments", "_____no_output_____" ] ], [ [ "# Use an immutable variable for the default argument \ndef better_add_column(values, df=None):\n \"\"\"Add a column of `values` to a DataFrame `df`.\n The column will be named \"col_<n>\" where \"n\" is\n the numerical index of the column.\n\n Args:\n values (iterable): The values of the new column\n df (DataFrame, optional): The DataFrame to update.\n If no DataFrame is passed, one is created by default.\n\n Returns:\n DataFrame\n \"\"\"\n # Update the function to create a default DataFrame\n if df is None:\n df = pandas.DataFrame()\n df['col_{}'.format(len(df.columns))] = values\n return df", "_____no_output_____" ] ], [ [ "# Context Managers", "_____no_output_____" ], [ "## Using context managers\n\nA context manager:\n- Sets up a context\n- Runs your code \n- Removes the context", "_____no_output_____" ] ], [ [ "with <context-manager>(<args>) as <variable-name>:\n # Run your code here\n # This code is running \"inside the context\"\n \n# This code runs after the context is removed", "_____no_output_____" ] ], [ [ "### A real-world example\n\n`open()` does three things: \n- Sets up a context by opening a file\n- Lets you run any code you want on that file\n- Removes the context by closing thele", "_____no_output_____" ] ], [ [ "with open('my_file.txt') as my_file:\n text = my_file.read()\n length = len(text)\n \nprint('The file is {} characters long'.format(length))", "_____no_output_____" ] ], [ [ "### Exercise: The speed of cats", "_____no_output_____" ] ], [ [ "image = get_image_from_instagram()\n\n# Time how long process_with_numpy(image) takes to run\nwith timer():\n print('Numpy version')\n process_with_numpy(image)\n\n# Time how long process_with_pytorch(image) takes to run\nwith timer():\n print('Pytorch version')\n process_with_pytorch(image)\n \nNumpy version\nProcessing..........done!\nElapsed: 1.52 seconds\nPytorch version\nProcessing..........done!\nElapsed: 0.33 seconds", "_____no_output_____" ] ], [ [ "## Writing context managers", "_____no_output_____" ], [ "Two ways to define a context manager\n- Class-based\n- **Function-based**", "_____no_output_____" ], [ "### How to create a context manager\n\n1. Define a function.\n2. (optional) Add any set up code your context needs.\n3. Use the \"yield\" keyword.\n4. (optional) Add any teardown code your context needs.\n5. Add the @contextlib.contextmanager decorator.", "_____no_output_____" ] ], [ [ "import contextlib\nfrom contextlib import contextmanager\n\[email protected]\ndef my_context():\n # Add any set up code you need\n yield\n # Add any teardown code you need", "_____no_output_____" ] ], [ [ "### The \"yield\" keyword", "_____no_output_____" ] ], [ [ "@contextlib.contextmanager\ndef my_context(): \n print('hello')\n yield 42 # The value that your contex manager yields can be assigned to a variable in the \"with\" statement by adding\"as\" variable name.\n print('goodbye')\n \nwith my_context() as foo: \n print('foo is {}'.format(foo))", "hello\nfoo is 42\ngoodbye\n" ] ], [ [ "### Setup and teardown", "_____no_output_____" ] ], [ [ "@contextlib.contextmanager\ndef database(url):\n # set up database connection \n db = postgres.connect(url)\n yield db # yields the database connection\n \n # tear down database connection \n db.disconnect()\n \n\nurl = 'http://datacamp.com/data'\nwith database(url) as my_db: \n course_list = my_db.execute('SELECT * FROM courses')", "_____no_output_____" ] ], [ [ "### Yielding a value or None", "_____no_output_____" ] ], [ [ "# context manager that changes the current working directory to a specific path and the change it back after the context block is done.\n# It does not need to return anything with \"yield\" statement\[email protected]\ndef in_dir(path): \n # save current working directory \n old_dir = os.getcwd()\n \n # switch to new working directory \n os.chdir(path)\n \n yield # not yield explicit value\n \n # change back to previous\n # working directory \n os.chdir(old_dir)\n \n \nwith in_dir('/data/project_1/'): # not yield a value\n project_files = os.listdir()", "_____no_output_____" ] ], [ [ "### Exercise: The timer() context manager", "_____no_output_____" ] ], [ [ "# Add a decorator that will make timer() a context manager\[email protected]\ndef timer():\n \"\"\"Time the execution of a context block.\n\n Yields:\n None\n \"\"\"\n start = time.time()\n # Send control back to the context block\n yield\n end = time.time()\n print('Elapsed: {:.2f}s'.format(end - start))\n\nwith timer():\n print('This should take approximately 0.25 seconds')\n time.sleep(0.25)\n \nThis should take approximately 0.25 seconds\nElapsed: 0.25s", "_____no_output_____" ] ], [ [ "### Exercise: A read-only open() context manager\n\nThe regular `open()` context manager:\n\n- takes a filename and a mode (`'r'` for read, `'w'` for write, or `'a'` for append)\n- opens the file for reading, writing, or appending\n- sends control back to the context, along with a reference to the file\n- waits for the context to finish\n- and then closes the file before exiting", "_____no_output_____" ] ], [ [ "@contextlib.contextmanager\ndef open_read_only(filename):\n \"\"\"Open a file in read-only mode.\n\n Args:\n filename (str): The location of the file to read\n\n Yields:\n file object\n \"\"\"\n read_only_file = open(filename, mode='r')\n # Yield read_only_file so it can be assigned to my_file\n yield read_only_file\n # Close read_only_file\n read_only_file.close()\n\nwith open_read_only('my_file.txt') as my_file:\n print(my_file.read())", "_____no_output_____" ] ], [ [ "## Advanced topics", "_____no_output_____" ], [ "### Nested contexts", "_____no_output_____" ], [ "This approach works fine until you try to copy a file that is too large to fit in memory", "_____no_output_____" ] ], [ [ "def copy(src, dst):\n \"\"\"Copy the contents of one file to another. \n \n Args: \n src (str): File name of the file to be copied. \n dst (str): Where to write the new file. \n \"\"\"\n \n # Open the source file and read in the contents\n with open(src) as f_src:\n contents = f_src.read()\n \n # Open the destination file and write out the contents\n with open(dst, 'w') as f_dst: \n f_dst.write(contents)", "_____no_output_____" ] ], [ [ "Ideal: open both files at once and copy over oneline at a time.", "_____no_output_____" ] ], [ [ "with open('my_file.txt') as my_file:\n for line in my_file:\n # do something", "_____no_output_____" ], [ "def copy(src, dst):\n \"\"\"Copy the contents of one file to another. \n \n Args: \n src (str): File name of the file to be copied. \n dst (str): Where to write the new file. \n \"\"\"\n # Open both files\n with open(src) as f_src:\n with open(dst, 'w') as f_dst:\n # Read and write each line, one at a time\n for line in f_src: \n f_dst.write(line)", "_____no_output_____" ] ], [ [ "### Handling errors", "_____no_output_____" ], [ "No handling errors", "_____no_output_____" ] ], [ [ "def get_printer(ip): \n p = connect_to_printer(ip)\n \n yield\n \n # This MUST be called or no one else will\n # be able to connect to the printer \n p.disconnect() \n print('disconnected from printer')\n \ndoc = {'text': 'This is my text.'}\n\nwith get_printer('10.0.34.111') as printer: \n printer.print_page(doc['txt'])\n \n\nTraceback (most recent call last): \n File \"<stdin>\", line 1, in <module> \n printer.print_page(doc['txt'])\nKeyError: 'txt'", "_____no_output_____" ] ], [ [ "with handling errors", "_____no_output_____" ] ], [ [ "try:\n # code that might raise an error\nexcept:\n # do something about the error\nfinally:\n # this code runs no matter what", "_____no_output_____" ], [ "def get_printer(ip): \n p = connect_to_printer(ip)\n \n try:\n yield\n finally: \n p.disconnect() \n print('disconnected from printer')\n \n \ndoc = {'text': 'This is my text.'}\n\nwith get_printer('10.0.34.111') as printer: \n printer.print_page(doc['txt'])\n \n \ndisconnected from printer\nTraceback (most recent call last): \n File \"<stdin>\", line 1, in <module> \n printer.print_page(doc['txt'])\nKeyError: 'txt'", "_____no_output_____" ] ], [ [ "### Context manager patterns\n\n| | |\n|---------|------------|\n| Open | Close |\n| Lock | Release |\n| Change | Reset |\n| Enter | Exit |\n| Start | Stop |\n| Setup | Teardown |\n| Connect | Disconnect |\n\n\n", "_____no_output_____" ], [ "### Exercise: Scraping the NASDAQ\n\nThe context manager `stock('NVDA')` will connect to the NASDAQ and return an object that you can use to get the latest price by calling its `.price()` method.\n\nYou want to connect to `stock('NVDA')` and record 10 timesteps of price data by writing it to the file `NVDA.txt`.", "_____no_output_____" ] ], [ [ "# Use the \"stock('NVDA')\" context manager\n# and assign the result to the variable \"nvda\"\nwith stock('NVDA') as nvda:\n # Open \"NVDA.txt\" for writing as f_out\n with open('NVDA.txt', 'w') as f_out:\n for price in range(10):\n value = nvda.price()\n print('Logging ${:.2f} for NVDA'.format(value))\n f_out.write('{:.2f}\\n'.format(value))", "_____no_output_____" ] ], [ [ "### Exercise: Changing the working directory", "_____no_output_____" ] ], [ [ "def in_dir(directory):\n \"\"\"Change current working directory to `directory`,\n allow the user to run some code, and change back.\n\n Args:\n directory (str): The path to a directory to work in.\n \"\"\"\n current_dir = os.getcwd()\n os.chdir(directory)\n\n # Add code that lets you handle errors\n try:\n yield\n # Ensure the directory is reset,\n # whether there was an error or not\n finally:\n os.chdir(current_dir)", "_____no_output_____" ] ], [ [ "# Decorators\n[Python 函数装饰器](https://www.runoob.com/w3cnote/python-func-decorators.html)", "_____no_output_____" ], [ "## Functions are objects", "_____no_output_____" ], [ "### Functions are just another type of object", "_____no_output_____" ], [ "Python objects:", "_____no_output_____" ] ], [ [ "def x():\n pass\nx = [1, 2, 3]\nx = {'foo': 42}\nx = pandas.DataFrame()\nx = 'This is a sentence.'\nx = 3\nx = 71.2\nimport x", "_____no_output_____" ] ], [ [ "### Functions as variables", "_____no_output_____" ] ], [ [ "def my_function(): \n print('Hello')\nx = my_function\ntype(x)", "_____no_output_____" ], [ "x", "_____no_output_____" ], [ "x()", "Hello\n" ], [ "PrintyMcPrintface = print\nPrintyMcPrintface('Python is awesome!')", "Python is awesome!\n" ] ], [ [ "### Lists and dictionaries of functions", "_____no_output_____" ] ], [ [ "list_of_functions = [my_function, open, print]\nlist_of_functions[2]('I am printing with an element of a list!')", "I am printing with an element of a list!\n" ], [ "dict_of_functions = {\n 'func1': my_function,\n 'func2': open,\n 'func3': print\n}\ndict_of_functions['func3']('I am printing with a value of a dict!')", "I am printing with a value of a dict!\n" ] ], [ [ "### Referencing a function", "_____no_output_____" ] ], [ [ "def my_function():\n return 42\n\nx = my_function\nmy_function()", "_____no_output_____" ], [ "my_function", "_____no_output_____" ] ], [ [ "### Functions as arguments", "_____no_output_____" ] ], [ [ "def has_docstring(func):\n \"\"\"Check to see if the function `func` has a docstring. \n \n Args: \n func (callable): A function. \n \n Returns: \n bool \n \"\"\"\n return func.__doc__ is not None", "_____no_output_____" ], [ "def no():\n return 42\n\ndef yes():\n \"\"\"Return the value 42 \"\"\"\n return 42\n\nhas_docstring(no)", "_____no_output_____" ], [ "has_docstring(yes)", "_____no_output_____" ] ], [ [ "### Defining a function inside another function", "_____no_output_____" ] ], [ [ "def foo():\n x = [3, 6, 9]\n \n def bar(y): \n print(y)\n \n for value in x:\n bar(x)", "_____no_output_____" ], [ "def foo(x, y):\n if x > 4 and x < 10 and y > 4 and y < 10: \n print(x * y)\n \n \n \ndef foo(x, y):\n def in_range(v):\n return v > 4 and v < 10 \n \n if in_range(x) and in_range(y): \n print(x * y)", "_____no_output_____" ], [ "def get_function():\n def print_me(s): \n print(s)\n \n return print_me\n\nnew_func = get_function()\nnew_func('This is a sentence.')", "This is a sentence.\n" ] ], [ [ "## Scope", "_____no_output_____" ] ], [ [ "x = 7\ny = 200\nprint(x)", "7\n" ], [ "def foo(): \n x = 42 \n print(x) \n print(y)\n \nfoo()", "42\n200\n" ], [ "print(x)", "7\n" ] ], [ [ "### The global keyword", "_____no_output_____" ] ], [ [ "x = 7\n\ndef foo(): \n x = 42 \n print(x)\n \nfoo()", "42\n" ], [ "print(x)", "7\n" ], [ "x = 7\n\ndef foo():\n global x \n x = 42 \n print(x)\n \nfoo()", "42\n" ], [ "print(x)", "42\n" ] ], [ [ "### The nonlocal keyword\nYou should try to avoid using `global` variables if possible, because it can make testing and debugging harder.\nThe `nonlocal` keyword works exactly the same as the `global` key word, but it is used whenyou are inside a nested function, and you want to update a variable that is defined inside your parent function.", "_____no_output_____" ] ], [ [ "def foo(): \n x = 10\n \n def bar(): \n x = 200 \n print(x) \n \n bar() \n print(x)\n \nfoo()", "200\n10\n" ], [ "def foo(): \n x = 10\n \n def bar():\n nonlocal x\n x = 200 \n print(x) \n \n bar() \n print(x)\n \nfoo()", "200\n200\n" ], [ "x = 50\n\ndef one():\n x = 10\n\ndef two():\n global x\n x = 30\n\ndef three():\n x = 100\n print(x)\n\nfor func in [one, two, three]:\n func()\n print(x)", "50\n30\n100\n30\n" ] ], [ [ "## Closures\n\nA closure in Python is a tuple of variables that are no longer in scope, but that a function needs in order to run.", "_____no_output_____" ], [ "### Attaching nonlocal variables to nested functions", "_____no_output_____" ] ], [ [ "def foo(): \n a = 5\n def bar():\n print(a)\n return bar\n\nfunc = foo()\n\nfunc()", "5\n" ], [ "type(func.__closure__)", "_____no_output_____" ], [ "len(func.__closure__)", "_____no_output_____" ], [ "func.__closure__[0].cell_contents", "_____no_output_____" ] ], [ [ "### Closures and deletion", "_____no_output_____" ] ], [ [ "x = 25 # x is defined in the gloabl scope\n\ndef foo(value):\n def bar(): \n print(value)\n return bar\n\nmy_func = foo(x)\nmy_func() # print the value x", "25\n" ], [ "# delete x and call my_func again, it would still print 25. \n# Because foo()'s value argument gets added to the closure attached to the new my_func function.\n# So even though x doesn't exist anymore, thevalue persists in its closure\ndel(x)\nmy_func() ", "25\n" ], [ "len(my_func.__closure__)", "_____no_output_____" ], [ "my_func.__closure__[0].cell_contents", "_____no_output_____" ] ], [ [ "### Closures and overwriting", "_____no_output_____" ] ], [ [ "# pass x into foo() and then assigned the new function to the variable x. \n# The old value of x 25 is still stored in the new function's closure. \n# even though the new function is now stores in the x variable\nx = 25\n\ndef foo(value):\n def bar(): \n print(value)\n return bar\n\nx = foo(x)\nx() ", "25\n" ], [ "len(x.__closure__)", "_____no_output_____" ], [ "x.__closure__[0].cell_contents", "_____no_output_____" ] ], [ [ "## Why does all of this matter?", "_____no_output_____" ], [ "Decorators use:\n- Functions as objects\n- Nested functions\n- Nonlocal scope\n- Closures", "_____no_output_____" ], [ "Nested function: A function dened inside another function", "_____no_output_____" ] ], [ [ "# outer function\ndef parent():\n # nested function\n def child():\n pass\n return child", "_____no_output_____" ] ], [ [ "Nonlocal variables: Variables dened in the parent function that are used by the child function", "_____no_output_____" ] ], [ [ "def parent(arg_1, arg_2):\n # From child()'s point of view,\n # `value` and `my_dict` are nonlocal variables,\n # as are `arg_1` and `arg_2`. \n value = 22 \n my_dict = {'chocolate': 'yummy'}\n \n def child(): \n print(2 * value) \n print(my_dict['chocolate']) \n print(arg_1 + arg_2)\n \n return child", "_____no_output_____" ] ], [ [ "Closure: Nonlocal variables attached to a returned function", "_____no_output_____" ] ], [ [ "def parent(arg_1, arg_2): \n value = 22 \n my_dict = {'chocolate': 'yummy'}\n def child(): \n print(2 * value) \n print(my_dict['chocolate']) \n print(arg_1 + arg_2)\n \n return child\n\nnew_function = parent(3, 4)\n\nprint([cell.cell_contents for cell in new_function.__closure__])", "[3, 4, {'chocolate': 'yummy'}, 22]\n" ] ], [ [ "## Decorators", "_____no_output_____" ], [ "### The double_args decorator", "_____no_output_____" ] ], [ [ "def multiply(a, b):\n return a * b\ndef double_args(func):\n def wrapper(a, b):\n # Call the passed in function, but double each argument\n return func(a * 2, b * 2)\n return wrapper\n\n# assign the new function to \"new_multiply\"\nnew_multiply = double_args(multiply)\nnew_multiply(1, 5)", "_____no_output_____" ], [ "multiply(1, 5)", "_____no_output_____" ], [ "def multiply(a, b):\n return a * b\ndef double_args(func):\n def wrapper(a, b):\n # Call the passed in function, but double each argument\n return func(a * 2, b * 2)\n return wrapper\n\n# instead of assign the new function to \"new_multiply\", we're going to overwrite the \"multiply\" varable\n# we can do this because Python stores the orignial multipy function in the new function's closure\nmultiply = double_args(multiply)\nmultiply(1, 5)", "_____no_output_____" ] ], [ [ "### Decorator syntax", "_____no_output_____" ] ], [ [ "def double_args(func):\n def wrapper(a, b):\n return func(a * 2, b * 2)\n return wrapper\n\n@double_args\ndef multiply(a, b):\n return a * b\n\nmultiply(1, 5)", "_____no_output_____" ] ], [ [ "### Exercise: Defining a decorator", "_____no_output_____" ] ], [ [ "def print_before_and_after(func):\n def wrapper(*args):\n print('Before {}'.format(func.__name__))\n # Call the function being decorated with *args\n func(*args)\n print('After {}'.format(func.__name__))\n # Return the nested function\n return wrapper\n\n@print_before_and_after\ndef multiply(a, b):\n print(a * b)\n\nmultiply(5, 10)", "Before multiply\n50\nAfter multiply\n" ] ], [ [ "## Real-world examples", "_____no_output_____" ], [ "### Time a function", "_____no_output_____" ] ], [ [ "import time\n\ndef timer(func):\n \"\"\"A decorator that prints how long a function took to run. \n \n Args: \n func (callable): The function being decorated. \n \n Returns: \n callable: The decorated function. \n \"\"\"\n # Define the wrapper function to return.\n def wrapper(*args, **kwargs):\n # When wrapper() is called, get the current time. \n t_start = time.time()\n # Call the decorated function and store the result. \n result = func(*args, **kwargs)\n # Get the total time it took to run, and print it. \n t_total = time.time() - t_start \n print('{} took {}s'.format(func.__name__, t_total))\n return result\n return wrapper", "_____no_output_____" ] ], [ [ "### Using timer()", "_____no_output_____" ] ], [ [ "@timer\ndef sleep_n_seconds(n): \n time.sleep(n)", "_____no_output_____" ], [ "sleep_n_seconds(5)", "sleep_n_seconds took 5.001962184906006s\n" ], [ "sleep_n_seconds(10)", "sleep_n_seconds took 10.00467300415039s\n" ] ], [ [ "### When to use decorators\n\nAdd common behavior to multiple functions", "_____no_output_____" ] ], [ [ "@timer\ndef foo():\n # do some computation\n \n@timer\ndef bar():\n # do some other computation\n \n@timer\ndef baz():\n # do something else", "_____no_output_____" ] ], [ [ "### Exercise: Print the return type", "_____no_output_____" ] ], [ [ "def print_return_type(func):\n # Define wrapper(), the decorated function\n def wrapper(*args, **kwargs):\n # Call the function being decorated\n result = func(*args, **kwargs)\n print('{}() returned type {}'.format(\n func.__name__, type(result)\n ))\n return result\n # Return the decorated function\n return wrapper\n \n@print_return_type\ndef foo(value):\n return value\n \nprint(foo(42))\nprint(foo([1, 2, 3]))\nprint(foo({'a': 42}))", "foo() returned type <class 'int'>\n42\nfoo() returned type <class 'list'>\n[1, 2, 3]\nfoo() returned type <class 'dict'>\n{'a': 42}\n" ] ], [ [ "### Exercise: Counter", "_____no_output_____" ] ], [ [ "def counter(func):\n def wrapper(*args, **kwargs):\n wrapper.count += 1\n # Call the function being decorated and return the result\n return wrapper\n wrapper.count = 0\n # Return the new decorated function\n return wrapper\n\n# Decorate foo() with the counter() decorator\n@counter\ndef foo():\n print('calling foo()')\n\n \nfoo()\nfoo()\nfoo()\n\nprint('foo() was called {} times.'.format(foo.count))", "foo() was called 3 times.\n" ] ], [ [ "## Decorators and metadata", "_____no_output_____" ] ], [ [ "def sleep_n_seconds(n=10):\n \"\"\"Pause processing for n seconds. \n \n Args: \n n (int): The number of seconds to pause for. \n \"\"\" \n time.sleep(n)\n \nprint(sleep_n_seconds.__doc__)", "Pause processing for n seconds. \n \n Args: \n n (int): The number of seconds to pause for. \n \n" ], [ "print(sleep_n_seconds.__name__)", "sleep_n_seconds\n" ], [ "print(sleep_n_seconds.__defaults__)", "(10,)\n" ] ], [ [ "problem: Ouput输出应该是\"a_function_requiring_decoration\"。这里的函数被warpTheFunction替代了。它重写了我们函数的名字和注释文档(docstring)。", "_____no_output_____" ] ], [ [ "@timer\ndef sleep_n_seconds(n=10):\n \"\"\"Pause processing for n seconds. \n \n Args: \n n (int): The number of seconds to pause for. \n \"\"\" \n time.sleep(n)\n \nprint(sleep_n_seconds.__doc__)", "None\n" ], [ "print(sleep_n_seconds.__name__)", "wrapper\n" ] ], [ [ "### fix: The timer decorator", "_____no_output_____" ] ], [ [ "from functools import wraps\ndef timer(func):\n \"\"\"A decorator that prints how long a function took to run. \"\"\"\n @wraps(func)\n def wrapper(*args, **kwargs):\n # When wrapper() is called, get the current time. \n t_start = time.time()\n # Call the decorated function and store the result. \n result = func(*args, **kwargs)\n # Get the total time it took to run, and print it. \n t_total = time.time() - t_start \n print('{} took {}s'.format(func.__name__, t_total))\n return result\n return wrapper", "_____no_output_____" ], [ "@timer\ndef sleep_n_seconds(n=10):\n \"\"\"Pause processing for n seconds. \n \n Args: \n n (int): The number of seconds to pause for. \n \"\"\" \n time.sleep(n)\n \nprint(sleep_n_seconds.__doc__)", "Pause processing for n seconds. \n \n Args: \n n (int): The number of seconds to pause for. \n \n" ], [ "print(sleep_n_seconds.__name__)", "sleep_n_seconds\n" ], [ "print(sleep_n_seconds.__defaults__)", "None\n" ] ], [ [ "### Access to the original function", "_____no_output_____" ] ], [ [ "sleep_n_seconds.__wrapped__", "_____no_output_____" ] ], [ [ "## Decorators that take arguments", "_____no_output_____" ] ], [ [ "def run_three_times(func):\n def wrapper(*args, **kwargs):\n for i in range(3): \n func(*args, **kwargs)\n return wrapper\n\n@run_three_times\ndef print_sum(a, b): \n print(a + b)\n \nprint_sum(3, 5)", "8\n8\n8\n" ] ], [ [ "### A decorator factory", "_____no_output_____" ] ], [ [ "def run_n_times(n):\n \"\"\"Define and return a decorator\"\"\"\n def decorator(func):\n def wrapper(*args, **kwargs):\n for i in range(n): \n func(*args, **kwargs)\n return wrapper\n return decorator\n \nrun_three_times = run_n_times(3)\n\n@run_three_times\ndef print_sum(a, b): \n print(a + b)\n \n@run_n_times(3)\ndef print_sum(a, b): \n print(a + b)", "_____no_output_____" ] ], [ [ "Using run_n_times()", "_____no_output_____" ] ], [ [ "@run_n_times(3)\ndef print_sum(a, b): \n print(a + b)\n\nprint_sum(3, 5)", "8\n8\n8\n" ], [ "@run_n_times(5)\ndef print_sum(a, b): \n print(\"Hello!\")\n\nprint_sum(3, 5)", "Hello!\nHello!\nHello!\nHello!\nHello!\n" ] ], [ [ "## Timeout(): a real world example", "_____no_output_____" ] ], [ [ "import signal\n\ndef raise_timeout(*args, **kwargs):\n raise TimeoutError()\n\n# When an \"alarm\" signal goes off, call raise_timeout()\nsignal.signal(signalnum=signal.SIGALRM, handler=raise_timeout)\n# Set off an alarm in 5 seconds\nsignal.alarm(5)\n# Cancel the alarm\nsignal.alarm(0)", "_____no_output_____" ], [ "def timeout_in_5s(func): \n @wraps(func)\n def wrapper(*args, **kwargs):\n # Set an alarm for 5 seconds \n signal.alarm(5)\n try:\n # Call the decorated func\n return func(*args, **kwargs)\n finally:\n # Cancel alarm \n signal.alarm(0)\n return wrapper", "_____no_output_____" ], [ "@timeout_in_5s\ndef foo(): \n time.sleep(10) \n print('foo!')\n \nfoo()", "_____no_output_____" ], [ "def timeout(n_seconds): \n def decorator(func): \n @wraps(func)\n def wrapper(*args, **kwargs):\n # Set an alarm for 5 seconds \n signal.alarm(n_seconds)\n try:\n # Call the decorated func\n return func(*args, **kwargs)\n finally:\n # Cancel alarm \n signal.alarm(0)\n return wrapper\n return decorator\n\n\n@timeout(5)\ndef foo(): \n time.sleep(10) \n print('foo!')\n \nfoo()", "_____no_output_____" ], [ "@timeout(20)\ndef bar(): \n time.sleep(10) \n print('bar!')\n \nbar()", "bar!\n" ] ], [ [ "### Exercise: Tag your functions\n\nYou've decided to write a decorator that will let you tag your functions with an arbitrary list of tags. You could use these tags for many things:\n\n- Adding information about who has worked on the function, so a user can look up who to ask if they run into trouble using it.\n- Labeling functions as \"experimental\" so that users know that the inputs and outputs might change in the future.\n- Marking any functions that you plan to remove in a future version of the code.\n- tc.", "_____no_output_____" ] ], [ [ "def tag(*tags):\n # Define a new decorator, named \"decorator\", to return\n def decorator(func):\n # Ensure the decorated function keeps its metadata\n @wraps(func)\n def wrapper(*args, **kwargs):\n # Call the function being decorated and return the result\n return func(*args, **kwargs)\n wrapper.tags = tags\n return wrapper\n # Return the new decorator\n return decorator\n\n@tag('test', 'this is a tag')\ndef foo():\n pass\n\nprint(foo.tags)", "('test', 'this is a tag')\n" ] ], [ [ "### Exercise: Check the return type", "_____no_output_____" ] ], [ [ "def returns_dict(func):\n # Complete the returns_dict() decorator\n def wrapper(*args, **kwargs):\n result = func(*args, **kwargs)\n assert(type(result) == dict)\n return result\n return wrapper\n\n@returns_dict\ndef foo(value):\n return value\n\ntry:\n print(foo([1,2,3]))\nexcept AssertionError:\n print('foo() did not return a dict!')", "foo() did not return a dict!\n" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code" ]
[ [ "markdown", "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown", "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown", "markdown" ], [ "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code", "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ] ]
ece90e28ce7be9bed4dfec4cb818c3e546063628
200,547
ipynb
Jupyter Notebook
03.power_consumption/08.energy-forecasting-XGBoost.ipynb
mkumar73/time-series
6a19c69a34002981ae29b9c667486445a7a83c58
[ "MIT" ]
2
2021-10-30T19:08:19.000Z
2021-11-17T10:21:16.000Z
03.power_consumption/08.energy-forecasting-XGBoost.ipynb
mkumar73/time-series
6a19c69a34002981ae29b9c667486445a7a83c58
[ "MIT" ]
14
2020-01-28T22:53:33.000Z
2022-02-10T00:18:47.000Z
03.power_consumption/.ipynb_checkpoints/08.energy-forecasting-XGBoost-checkpoint.ipynb
mkumar73/time-series
6a19c69a34002981ae29b9c667486445a7a83c58
[ "MIT" ]
1
2020-05-02T02:49:33.000Z
2020-05-02T02:49:33.000Z
389.41165
124,848
0.928924
[ [ [ "# Forecasting using XGBoost\n\n- One-step recursive forecasting", "_____no_output_____" ] ], [ [ "import pandas as pd\nimport numpy as np\nimport os\n\nimport matplotlib.pyplot as plt\nimport seaborn as sns\n%matplotlib inline\n\n\n# accuracy measures and data spliting\nfrom sklearn.metrics import mean_squared_error\nfrom sklearn.metrics import mean_absolute_error\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.model_selection import GridSearchCV\n\n# deep learning libraries\nimport xgboost as xgb\nimport graphviz", "_____no_output_____" ], [ "plt.style.use('seaborn')\nplt.rcParams['figure.figsize'] = 15, 7", "_____no_output_____" ] ], [ [ "## 1. Data import", "_____no_output_____" ] ], [ [ "DATADIR = '../data/power-consumption/'\nMODELDIR = '../checkpoints/power/xgb/model/'", "_____no_output_____" ], [ "# directory for saving model \nif os.path.exists(MODELDIR):\n pass\nelse:\n os.makedirs(MODELDIR)", "_____no_output_____" ], [ "data = pd.read_csv(os.path.join(DATADIR, 'processed_data.csv'))", "_____no_output_____" ], [ "data.head()", "_____no_output_____" ] ], [ [ "## 2. Train test split", "_____no_output_____" ] ], [ [ "y = data[['Global_active_power']].copy()\n\n# let xgboost decide the most important features\nX = data.drop(columns=['date', 'Global_active_power', 'Median_active_power', 'Lagged_active_power', 'median_residual'], axis=1)\n\n# last 40 weeks data for testing\ntest_size = np.int16(7*1)\ntrain_size = X.shape[0] - test_size\n\n\nX_train, X_test = X.loc[:train_size-1, :], X.loc[train_size:, :]\ny_train, y_test = y.loc[:train_size-1, :], y.loc[train_size:, :]\n\nX_train.shape, X_test.shape, y_train.shape, y_test.shape", "_____no_output_____" ] ], [ [ "## 5. Model fitting", "_____no_output_____" ] ], [ [ "params = {'objective' :'reg:linear', \n 'colsample_bytree' : 1, \n 'learning_rate' : 0.1,\n 'max_depth': 5,\n 'alpha' : 10, \n 'n_estimators' : 200}", "_____no_output_____" ], [ "X_train_v = X_train.values\nX_test_v = X_test.values\ny_train_v = y_train.values\ny_test_v = y_test.values", "_____no_output_____" ], [ "def xgb_model(X_train_v, y_train_v, params=dict()):\n model = xgb.XGBRegressor(**params)\n \n model.fit(X_train_v, y_train_v, verbose=False)\n y_train_pred = model.predict(X_train_v)\n \n onestep_pred = []\n for i in range(X_test.shape[0]):\n model.fit(X_train_v, y_train_v, verbose=False)\n\n pred = model.predict(X_test_v[i, :].reshape(1, -1))\n onestep_pred.append(pred)\n\n tempX = np.vstack((X_train_v, X_test_v[i, :])) \n X_train_v = tempX.copy()\n\n arr_pred = np.array([pred]).reshape(-1, 1)\n tempY = np.vstack((y_train_v, arr_pred))\n y_train_v = tempY.copy()\n \n y_test_pred = np.array(onestep_pred)\n \n return y_train_pred, y_test_pred", "_____no_output_____" ], [ "y_train_pred, y_test_pred = xgb_model(X_train_v, y_train_v, params=params)", "_____no_output_____" ] ], [ [ "## 6. Model evaluation", "_____no_output_____" ] ], [ [ "def model_evaluation(y_train, y_test, y_train_pred, y_test_pred):\n\n # MAE and NRMSE calculation\n train_rmse = np.sqrt(mean_squared_error(y_train, y_train_pred))\n train_mae = np.round(mean_absolute_error(y_train, y_train_pred), 3)\n train_nrmse = np.round(train_rmse/np.std(y_train), 3)\n\n test_rmse = np.sqrt(mean_squared_error(y_test, y_test_pred))\n test_mae = np.round(mean_absolute_error(y_test, y_test_pred), 3)\n test_nrmse = np.round(test_rmse/np.std(y_test), 3)\n\n print(f'Training MAE: {train_mae}')\n print(f'Trainig NRMSE: {train_nrmse}')\n\n print(f'Test MAE: {test_mae}')\n print(f'Test NRMSE: {test_nrmse}')\n ", "_____no_output_____" ], [ "model_evaluation(y_train_v, y_test_v, y_train_pred, y_test_pred)", "Training MAE: 4.162\nTrainig NRMSE: 0.009\nTest MAE: 8.907\nTest NRMSE: 0.026\n" ], [ "plt.plot(y_train_v, label='actual')\nplt.plot(y_train_pred, label='predicted')\nplt.ylabel('kWatt')\nplt.xlabel('index')\nplt.title('Actual vs Predicted on Training data', fontsize=14)\nplt.legend()\nplt.tight_layout()\nplt.show()", "_____no_output_____" ], [ "plt.plot(y_test_v, label='actual')\nplt.plot(y_test_pred, label='predicted')\nplt.ylabel('kWatt')\nplt.xlabel('index')\nplt.title('Actual vs predicted on test data using recursive XGBoost', fontsize=14)\nplt.legend()\nplt.tight_layout()\nplt.show()", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code" ]
[ [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code" ] ]
ece922f56c123945651af9e89857c0eda154f5f3
11,587
ipynb
Jupyter Notebook
02_introduccion_a_flask.ipynb
PythonistaMX/py231
734f5972948f7d23d581ec54ee7af581eca651ce
[ "MIT" ]
3
2020-05-05T23:38:45.000Z
2021-11-05T01:03:38.000Z
02_introduccion_a_flask.ipynb
PythonistaMX/py221
734f5972948f7d23d581ec54ee7af581eca651ce
[ "MIT" ]
null
null
null
02_introduccion_a_flask.ipynb
PythonistaMX/py221
734f5972948f7d23d581ec54ee7af581eca651ce
[ "MIT" ]
4
2020-10-20T01:15:23.000Z
2022-02-08T02:27:35.000Z
30.412073
405
0.579184
[ [ [ "[![img/pythonista.png](img/pythonista.png)](https://www.pythonista.io)", "_____no_output_____" ], [ "# Introducción a *Flask*.", "_____no_output_____" ], [ "[*Flask*](https://flask.palletsprojects.com) es conocido como un \"*microframework*\". Es decir, que a diferencia de proyectos como *Django* que viene \"con las pilas incluidas\", *Flask* solamente contiene funcionalidades básicas tales como:\n\n* Un servidor de aplicaciones basado en la biblioteca [*Werkzeug*](https://werkzeug.palletsprojects.com/).\n* Soporte de Plantillas por medio de [*Jinja*](https://jinja.palletsprojects.com).\n* Una herramienta de depuración.\n* Soporte para pruebas unitarias.\n* Soporte para cookies seguras.\n* Soporte para desarrollo de instrucciones por medio de la interfaz de línea de comandos (*CLI*) mediante [*Click*](https://click.palletsprojects.com/).", "_____no_output_____" ], [ "## Las extensiones de *Flask*.\n\nAún cuando la instalación básica de *Flask* contiene componentes mínimos, el proyecto cuenta con un amplio catáĺogo de extensiones disponibles en la siguiente liga: \n\nhttps://flask.palletsprojects.com/en/2.0.x/extensions/", "_____no_output_____" ], [ "## Instalación del paquete ```flask```.\n\nEl paquete ```flask``` se puede encontrar dentro del catalogo de [*pypi*](https://pypi.org/project/Flask/) y puede ser instalado mediante ```pip```.", "_____no_output_____" ] ], [ [ "!pip install flask", "_____no_output_____" ] ], [ [ "## La clase ```flask.Flask```.\n\nLa clase ```flask.Flask``` es el componente principal del framework. Los objetos instanciados a partir de esta clase realizarán todas las funciones del servidor de aplicaciones.\n\nEl único parámetro requerido obligatoriamente al instanciar un objeto de tipo ```Flask``` es el nombre de la aplicación, el cual de principio corresponde al objeto asignado al nombre del entorno global ```__name__```.\n\nSolamente si se piensa en utilizar un objeto de tipo ```Flask``` dentro de un paquete, el nombre deber de ser cambiado por el nombre del paquete.\n\n**Sintaxis:**\n\n``` python\n<nombre> = flask.Flask(<nombre de la aplicación>)\n```\n\n**Nota:** aún cuando puede asignársele cualquier nombre, se utiliza el nombre ```app``` por convención. En adelante se usará dicha convención para hacer referencia a un objeto instanciado de la clase ```flask.Flask```. ", "_____no_output_____" ], [ "**Ejemplo:**\n\nLas siguientes líneas de código asigna el nombre ```app``` al objeto instanciado de la clase ```Flask```, ingresando el nombre ```_name__``` como argumento.", "_____no_output_____" ] ], [ [ "from flask import Flask\napp = Flask(__name__)", "_____no_output_____" ], [ "app", "_____no_output_____" ] ], [ [ "## El atributo ```app.config```. ", "_____no_output_____" ], [ "El atributo ```app.config``` es un objeto instanciado de una subclase de ```dict``` que sirve como el recurso usado para consultar, añadir o modificar los parámetros de configuración de la aplicación, incluyendo aquellas que son utulizadas por las extensiones de *Flask*.\n\nLa documentación de ```app.config``` puede ser consultada en:\n\nhttps://flask.palletsprojects.com/en/2.0.x/config/", "_____no_output_____" ] ], [ [ "app.config", "_____no_output_____" ] ], [ [ "### Opciones comunes de ```app.config```.\n\n* ```app.config['ENV']```, la cual define el entorno en el que se ejecutará la aplicación y que corresponde a:\n * ```'production'``` para entornos en producción.\n * ```'development'``` para entornos de desarrollo.\n* ```app.config['DEBUG']```, la cual indica si la aplicación se ejecutará en modo de depuración en caso de que sea ```True```.\n* ```app.config['TESTING']```, la cual indica si la aplicación está corriendo pruebas, en caso de que sea ```True```. \n* ```app.config['SECRET_KEY']```, la cual define la clave base a partir de la cual se construirán los *hash* de cifrado. ", "_____no_output_____" ], [ "**Ejemplos:**", "_____no_output_____" ] ], [ [ "app.config['ENV']", "_____no_output_____" ], [ "app.config['DEBUG']", "_____no_output_____" ], [ "app.config['TESTING']", "_____no_output_____" ], [ "app.config['SECRET_KEY']", "_____no_output_____" ] ], [ [ "### Métodos para la carga de configuraciones de ```app.config``` .\n\nAlgunos valores de configuración de ```app.config``` pueden contener contraseñas o datos sensibles que por seguridad no deben de ser incluidos en el código de la aplicación.\n\nEs por ello que los siguientes métodos permiten cargar los parámetros de configuración de ```app.config``` desde un archivo u objeto.\n\n```\n<método>(<ruta>)\n```\nDonde: \n* <método> es uno de los siguientes métodos:\n * ```app.config.from_file()```\n * ```app.config.from_json()```\n * ```app.config.from_pyfile()```", "_____no_output_____" ], [ "**Ejemplo:**", "_____no_output_____" ], [ "* La siguiente celda importará la configuración del archivo ```settings.cfg```. ", "_____no_output_____" ] ], [ [ "app.config.from_pyfile('settings.cfg')", "_____no_output_____" ], [ "%pycat settings.cfg", "_____no_output_____" ], [ "app.config['SECRET_KEY']", "_____no_output_____" ] ], [ [ "## Rutas y funciones de vista.\n\nCuando un cliente realiza una petición al servidor de aplicaciones de *Flask*, ésta debe de corresponder a una ruta válida capaz de gestionar dicha petición. De lo contrario, el servidor de *Flask* regresará un estado ```404```.\n\n**Nota:** las respuestas que envía el servidor de aplicaciones de *Flask* son documentos *HTML* por defecto.", "_____no_output_____" ], [ "## El método ```app.route()```.\n\n \nEl método ```app.route()``` es una función de orden superior que al ser usada como decorador, permite ligar a una ruta con una función.\n\n\n```\[email protected](<regla de url>, methods=<métodos>)\ndef <vista>():\n ...\n ...\n```\n\nDonde:\n\n* ```<regla de url>``` es un objeto ```str``` que define la ruta a la que se asociará una función de vista. \n* ```<métodos>``` es un objeto de tipo ```list``` que contiene los nombres en formato ```str``` de los métodos *HTTP*. El valor por defecto es ```[\"GET\"]```.\n* ```<función de vista>``` es la definición de una función cuyo resultado será enviado al cliente quen realice una petición a la ruta indicada.", "_____no_output_____" ], [ "**Ejemplo:**", "_____no_output_____" ], [ "La siguiente celda creará una regla de *URL* para el directorio raíz del servidor de aplicaciones, la cual enviará como mensaje de respuesta un documento *HTML*, el cual contendrá el texto ```<p>Hola, Mundo.</p>```", "_____no_output_____" ] ], [ [ "@app.route('/')\ndef inicio():\n return '<p>Hola, Mundo.</p>'", "_____no_output_____" ] ], [ [ "## El método ```app.run()```.\n\nEl método ```app.run()``` es el encargado de levantar el servidor web. Es posible ingresar algunos parámetros iniciales tales como:\n\n\n```\napp.run(<parámetros>)\n```\n\n* El parámetro ```host``` al cual se le asignaría un objeto de tipo ```str``` que definiría el rango de las direcciones *IP* que puede escuchar. El valor inicial es ```'localhost'```, pero se puede especificar el rango de direcciones *IP* al que se desee atender. Por ejemplo, ```'0.0.0.0'``` indica que el servidor de aplicaciones escuchará a todas las direcciones *IP*.\n* El parámetro ```port``` corresponde a un número entero cuyo valor predetermiando es ```5000``` y defiene el puerto al que el servicio será asignado.\n* El parámetro ```debug``` permite a *Flask* entrar en modo de depuración y su valor es un boolenado que por defecto es ```False```.\n", "_____no_output_____" ], [ "**Ejemplo:**", "_____no_output_____" ] ], [ [ "app.run(host=\"0.0.0.0\", port=5000)", "_____no_output_____" ] ], [ [ "<p style=\"text-align: center\"><a rel=\"license\" href=\"http://creativecommons.org/licenses/by/4.0/\"><img alt=\"Licencia Creative Commons\" style=\"border-width:0\" src=\"https://i.creativecommons.org/l/by/4.0/80x15.png\" /></a><br />Esta obra está bajo una <a rel=\"license\" href=\"http://creativecommons.org/licenses/by/4.0/\">Licencia Creative Commons Atribución 4.0 Internacional</a>.</p>\n<p style=\"text-align: center\">&copy; José Luis Chiquete Valdivieso. 2022.</p>", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown" ]
[ [ "markdown", "markdown", "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code", "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code", "code", "code", "code" ], [ "markdown", "markdown", "markdown" ], [ "code", "code", "code" ], [ "markdown", "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown" ] ]
ece9235e59621c92d2aa50147982204f71e534a3
812,222
ipynb
Jupyter Notebook
WordClouds/Plot_Wordclouds.ipynb
ashishsarout/AI_for_social_good
d9e04aae58a8e67e52d17a15375f55d1af1e8e88
[ "MIT" ]
6
2021-08-06T14:15:20.000Z
2022-03-19T09:58:15.000Z
WordClouds/Plot_Wordclouds.ipynb
Zubrah/AI_For_Social_Good
da80066b53545893d85b152d0fe3bef7beea2684
[ "MIT" ]
null
null
null
WordClouds/Plot_Wordclouds.ipynb
Zubrah/AI_For_Social_Good
da80066b53545893d85b152d0fe3bef7beea2684
[ "MIT" ]
5
2021-07-26T07:25:07.000Z
2022-03-29T23:14:55.000Z
4,186.71134
404,084
0.964443
[ [ [ "import pandas as pd \nfrom wordcloud import WordCloud\nimport matplotlib.pyplot as plt\nfrom nltk import FreqDist\nfrom nltk import ngrams\nimport numpy as np\nimport matplotlib as mpl\n%matplotlib inline", "_____no_output_____" ], [ "def generate_frequency(text_list):\n fdist = FreqDist()\n for i in text_list:\n words = i.split(' ')\n words = [word for word in words if word != '']\n words = ngrams(words,1)\n for x in words:\n fdist[x[0]]+=1\n return fdist", "_____no_output_____" ], [ "df = pd.read_csv('../Dataset/cleanedRedditSuicide.csv')", "_____no_output_____" ], [ "fdist = generate_frequency(df['cleaned'])\ntop_words = fdist.most_common(n=150)", "_____no_output_____" ], [ "word_dict = {}\nfor i in range(len(top_words)):\n word_dict[top_words[i][0]] = top_words[i][1]", "_____no_output_____" ], [ "cmap = plt.cm.Blues(np.linspace(0,1,20)) \ncmap = mpl.colors.ListedColormap(cmap[-9:,:-1]) ", "_____no_output_____" ], [ "\nwordcloud = WordCloud(background_color=\"white\",max_font_size=40,colormap=cmap).generate_from_frequencies(word_dict)\n\nplt.figure(figsize = (14, 8)) \nplt.imshow(wordcloud, interpolation='bilinear')\nplt.axis(\"off\")\nplt.show()", "_____no_output_____" ], [ "cmap = plt.cm.Oranges(np.linspace(0,1,20)) \ncmap = mpl.colors.ListedColormap(cmap[-9:,:-1]) ", "_____no_output_____" ], [ "df = pd.read_csv('../Dataset/cleanedTwitterSuicide.csv')", "_____no_output_____" ], [ "fdist = generate_frequency(df['cleaned'])\ntop_words = fdist.most_common(n=150)", "_____no_output_____" ], [ "word_dict = {}\nfor i in range(len(top_words)):\n word_dict[top_words[i][0]] = top_words[i][1]", "_____no_output_____" ], [ "wordcloud = WordCloud(background_color=\"white\",max_font_size=40, colormap=cmap).generate_from_frequencies(word_dict)\n\nplt.figure(figsize = (14, 8)) \nplt.imshow(wordcloud, interpolation='bilinear')\nplt.axis(\"off\")\nplt.show()\n", "_____no_output_____" ] ] ]
[ "code" ]
[ [ "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code" ] ]
ece929c4443344a39c52fced3a061e6587738b44
1,606
ipynb
Jupyter Notebook
剑指offer/.ipynb_checkpoints/#59-checkpoint.ipynb
yang-233/algorithm-notebook
f225933bc61e01f911c5280867741d46d6542211
[ "MIT" ]
null
null
null
剑指offer/.ipynb_checkpoints/#59-checkpoint.ipynb
yang-233/algorithm-notebook
f225933bc61e01f911c5280867741d46d6542211
[ "MIT" ]
null
null
null
剑指offer/.ipynb_checkpoints/#59-checkpoint.ipynb
yang-233/algorithm-notebook
f225933bc61e01f911c5280867741d46d6542211
[ "MIT" ]
null
null
null
22.305556
78
0.433998
[ [ [ "from typing import *\nclass Solution:\n def maxSlidingWindow(self, nums: List[int], k: int) -> List[int]:\n que, max_que, res = [], [], []\n\n for i in nums:\n que.append(i)\n while len(max_que) > 0 and max_que[-1] < i:\n max_que.pop()\n max_que.append(i)\n\n if len(que) == k:\n res.append(max_que[0])\n v = que.pop(0)\n if max_que[0] == v:\n max_que.pop(0)\n\n if not res and max_que: # 如果k > n\n res.append(max_que[0])\n\n return res\n \n", "_____no_output_____" ], [ "# 实际上就是那个可以获得最大值的队列的应用", "_____no_output_____" ] ] ]
[ "code" ]
[ [ "code", "code" ] ]
ece92c49a4284e5a42bbf55e97f32b81979a3094
5,455
ipynb
Jupyter Notebook
NotebookExamples/polyglot/d3js.ipynb
nstone1012/interactive
6f996e0ff070c4d6b5f3fcb10a0bb7bc760c8252
[ "MIT" ]
null
null
null
NotebookExamples/polyglot/d3js.ipynb
nstone1012/interactive
6f996e0ff070c4d6b5f3fcb10a0bb7bc760c8252
[ "MIT" ]
null
null
null
NotebookExamples/polyglot/d3js.ipynb
nstone1012/interactive
6f996e0ff070c4d6b5f3fcb10a0bb7bc760c8252
[ "MIT" ]
null
null
null
36.125828
305
0.448579
[ [ [ "# Visualizing data using d3js\n\n**This is a work in progress.** It doesn't work yet in [Binder](https://mybinder.org/v2/gh/dotnet/interactive/master?urlpath=lab) because it relies on HTTP communication between the kernel and the Jupyter frontend.\n\nThis notebooks uses directly [d3.js](https://d3js.org/) library to perform custom data visualisation.", "_____no_output_____" ] ], [ [ "var rnd = new Random();\nvar a = Enumerable.Range(1,rnd.Next(4,12)).Select( t => rnd.Next(t, t*10)).ToArray();", "_____no_output_____" ] ], [ [ "Using [RequireJS](https://requirejs.org/) we import [d3.js](https://d3js.org/). We setup the rendering code, some SVG filter inspiredy by [Visual Cinnamon](https://www.visualcinnamon.com/) article on [gooey effect](https://www.visualcinnamon.com/2016/06/fun-data-visualizations-svg-gooey-effect).\n\nUsing `interactive.csharp.getVariable` we fetch the variable `a` value.", "_____no_output_____" ] ], [ [ "#!html\n<svg id=\"dataPlot\" width=\"100%\"></svg>\n<script>\n dtree_require = require.config({\n paths: {\n d3: \"https://d3js.org/d3.v5.min\"\n }\n });\n dtree_require([\"d3\"], function (d3) {\n let svg = d3.\n select(\"svg#dataPlot\");\n svg.selectAll(\"defs\").remove();\n svg.selectAll(\"g\").remove();\n \n let defs = svg.append(\"defs\");\n\n let filter = defs.append(\"filter\").attr(\"id\", \"gooeyCodeFilter\");\n\n filter.append(\"feGaussianBlur\")\n .attr(\"in\", \"SourceGraphic\")\n .attr(\"stdDeviation\", \"10\")\n .attr(\"color-interpolation-filters\", \"sRGB\")\n .attr(\"result\", \"blur\");\n\n filter.append(\"feColorMatrix\")\n .attr(\"in\", \"blur\")\n .attr(\"mode\", \"matrix\")\n .attr(\"values\", \"1 0 0 0 0 0 1 0 0 0 0 0 1 0 0 0 0 0 19 -9\")\n .attr(\"result\", \"gooey\");\n\n let container = d3\n .select(\"svg#dataPlot\")\n .append(\"g\")\n .style(\"filter\", \"url(#gooeyCodeFilter)\");\n\n if (typeof (interval) !== 'undefined') {\n clearInterval(interval);\n }\n updateD3Rendering = () => interactive.csharp.getVariable(\"a\")\n .then(data => {\n container\n .selectAll(\".points\")\n .data(data, (d,i) => i)\n .join(\n enter => enter.append(\"circle\")\n .attr(\"class\", \"points\")\n .attr(\"cy\", 80)\n .attr(\"cx\", (d, i) => ((i + 1) * 60))\n .transition()\n .duration(2000)\n .style(\"fill\", d => d3.interpolateTurbo(d / 80))\n .ease(d3.easeElastic.period(0.1))\n .attr(\"r\", d => Math.max(0,d)),\n update => update\n .transition()\n .duration(2000)\n .ease(d3.easeElastic.period(0.1))\n .style(\"fill\", d => d3.interpolateTurbo(d / 80))\n .attr(\"r\", d => Math.max(0,d)),\n exit => exit.remove());\n });\n interval = setInterval(() => updateD3Rendering(), 1000);\n });\n</script>", "_____no_output_____" ] ], [ [ "Notice the `setInterval` call near the end of the previous cell. This rechecks the data in the kernel and updates the plot.\n\nBack on the kernel, we can now update the data so that the kernel can see it.\n\nYes, this is a contrived example, and we're planning to support true streaming data, but it's a start.", "_____no_output_____" ] ], [ [ "for(var i = 0; i < 10; i++){\n await Task.Delay(2000);\n var limit = rnd.Next(4,12);\n a = Enumerable.Range(1,limit).Select( t => rnd.Next(15, 80)).ToArray();\n}", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code" ]
[ [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ] ]
ece92e612bd669e75bd7076d158b93c3f9c86c53
82,775
ipynb
Jupyter Notebook
examples/conll-2003.ipynb
sloth2012/ner-bert
52097650ffad0d2ee964d4bc8a64a901d6045cdd
[ "MIT" ]
1
2019-09-24T04:23:50.000Z
2019-09-24T04:23:50.000Z
examples/conll-2003.ipynb
672425265/ner-bert
730cd1700513dbd51dc9736290db36855f27a3e0
[ "MIT" ]
null
null
null
examples/conll-2003.ipynb
672425265/ner-bert
730cd1700513dbd51dc9736290db36855f27a3e0
[ "MIT" ]
1
2019-11-07T07:48:08.000Z
2019-11-07T07:48:08.000Z
22.305309
154
0.467786
[ [ [ "### Conll 2003 evaluation\n\nData downloaded from [here](https://github.com/kyzhouhzau/BERT-NER/tree/master/NERdata).", "_____no_output_____" ] ], [ [ "%reload_ext autoreload\n%autoreload 2\n%matplotlib inline\n\nimport pandas as pd\nimport warnings\nimport os\nimport sys\n\nsys.path.append(\"../\")\n\nwarnings.filterwarnings(\"ignore\")", "_____no_output_____" ], [ "data_path = \"/datadrive/conll-2003/\"\n\ntrain_path = data_path + \"train.txt\"\ndev_path = data_path + \"dev.txt\"\ntest_path = data_path + \"test.txt\"", "_____no_output_____" ] ], [ [ "### 0. Prc data for csv format", "_____no_output_____" ] ], [ [ "import codecs\n\n\ndef read_data(input_file):\n \"\"\"Reads a BIO data.\"\"\"\n with codecs.open(input_file, \"r\", encoding=\"utf-8\") as f:\n lines = []\n words = []\n labels = []\n for line in f:\n contends = line.strip()\n word = line.strip().split(' ')[0]\n label = line.strip().split(' ')[-1]\n if contends.startswith(\"-DOCSTART-\"):\n words.append('')\n continue\n \n if len(contends) == 0 and not len(words):\n words.append(\"\")\n \n if len(contends) == 0 and words[-1] == '.':\n l = ' '.join([label for label in labels if len(label) > 0])\n w = ' '.join([word for word in words if len(word) > 0])\n lines.append([l, w])\n words = []\n labels = []\n continue\n words.append(word)\n labels.append(label.replace(\"-\", \"_\"))\n return lines\n", "_____no_output_____" ], [ "train_f = read_data(train_path)\ndev_f = read_data(dev_path)\ntest_f = read_data(test_path)", "_____no_output_____" ], [ "[l for l in train_f]", "_____no_output_____" ], [ "len(train_f), len(dev_f), len(test_f)", "_____no_output_____" ], [ "train_f[0]", "_____no_output_____" ], [ "import pandas as pd", "_____no_output_____" ], [ "train_df = pd.DataFrame(train_f, columns=[\"0\", \"1\"])\ntrain_df.to_csv(data_path + \"train.csv\", index=False)", "_____no_output_____" ], [ "valid_df = pd.DataFrame(dev_f, columns=[\"0\", \"1\"])\nvalid_df.to_csv(data_path + \"valid.csv\", index=False)", "_____no_output_____" ], [ "test_df = pd.DataFrame(test_f, columns=[\"0\", \"1\"])\ntest_df.to_csv(data_path + \"test.csv\", index=False)", "_____no_output_____" ] ], [ [ "### 1. Create data loaders", "_____no_output_____" ] ], [ [ "import os\n\ndata_path = \"/datadrive/conll-2003/\"\ntrain_path = data_path + \"train.csv\"\nvalid_path = data_path + \"valid.csv\"\ntest_path = data_path + \"test.csv\"\n\nmodel_dir = \" /datadrive/models/multi_cased_L-12_H-768_A-12/\"\ninit_checkpoint_pt = os.path.join(\"/datadrive/models/multi_cased_L-12_H-768_A-12/\", \"pytorch_model.bin\")\nbert_config_file = os.path.join(\"/datadrive/bert/multi_cased_L-12_H-768_A-12/\", \"bert_config.json\")\nvocab_file = os.path.join(\"/datadrive/bert/multi_cased_L-12_H-768_A-12/\", \"vocab.txt\")", "_____no_output_____" ], [ "import torch\ntorch.cuda.set_device(0)\ntorch.cuda.is_available(), torch.cuda.current_device()", "_____no_output_____" ], [ "from modules import BertNerData as NerData", "INFO:summarizer.preprocessing.cleaner:'pattern' package not found; tag filters are not available for English\n" ], [ "data = NerData.create(train_path, valid_path, vocab_file)", "_____no_output_____" ], [ "len(data.train_dl.dataset), len(data.valid_dl.dataset)", "_____no_output_____" ], [ "print(data.id2label)", "['<pad>', '[CLS]', '[SEP]', 'B_ORG', 'B_O', 'I_O', 'B_MISC', 'B_PER', 'I_PER', 'B_LOC', 'I_LOC', 'I_ORG', 'I_MISC']\n" ], [ "sup_labels = ['B_ORG', 'B_MISC', 'B_PER', 'I_PER', 'B_LOC', 'I_LOC', 'I_ORG', 'I_MISC']", "_____no_output_____" ], [ "max([len(f.labels_ids) for f in data.train_dl.dataset])", "_____no_output_____" ] ], [ [ "### 2. Create model", "_____no_output_____" ] ], [ [ "from modules.models.bert_models import BertBiLSTMAttnCRF", "_____no_output_____" ], [ "model = BertBiLSTMAttnCRF.create(len(data.label2idx), bert_config_file, init_checkpoint_pt, enc_hidden_dim=256)", "_____no_output_____" ], [ "model.get_n_trainable_params()", "_____no_output_____" ] ], [ [ "#### TODO: fix bug with len", "_____no_output_____" ], [ "### 3. Create Learner", "_____no_output_____" ] ], [ [ "from modules import NerLearner", "_____no_output_____" ], [ "num_epochs = 100\nlearner = NerLearner(model, data,\n best_model_path=\"/datadrive/models/conll-2003/bilstm_attn_cased.cpt\",\n lr=0.001, clip=1.0, sup_labels=data.id2label[5:],\n t_total=num_epochs * len(data.train_dl))", "_____no_output_____" ] ], [ [ "### 4. Start learning", "_____no_output_____" ] ], [ [ "learner.fit(num_epochs, target_metric='f1')", "_____no_output_____" ] ], [ [ "### 5. Evaluate dev set", "_____no_output_____" ] ], [ [ "from modules.data.bert_data import get_bert_data_loader_for_predict\ndl = get_bert_data_loader_for_predict(data_path + \"valid.csv\", learner)", "_____no_output_____" ], [ "learner.load_model()", "_____no_output_____" ], [ "preds = learner.predict(dl)", "_____no_output_____" ] ], [ [ "IOB precision", "_____no_output_____" ] ], [ [ "from modules.train.train import validate_step\nprint(validate_step(learner.data.valid_dl, learner.model, learner.data.id2label, learner.sup_labels))", "_____no_output_____" ] ], [ [ "Span precision", "_____no_output_____" ] ], [ [ "from modules.utils.plot_metrics import get_bert_span_report\nclf_report = get_bert_span_report(dl, preds, [])\nprint(clf_report)", " precision recall f1-score support\n\n MISC 0.902 0.886 0.894 905\n LOC 0.925 0.914 0.920 1669\n O 0.992 0.992 0.992 41803\n ORG 0.849 0.875 0.862 1282\n PER 0.936 0.950 0.943 1686\n\n micro avg 0.982 0.982 0.982 47345\n macro avg 0.921 0.923 0.922 47345\nweighted avg 0.982 0.982 0.982 47345\n\n" ] ], [ [ "### 6. Evaluate test set", "_____no_output_____" ] ], [ [ "from modules.data.bert_data import get_bert_data_loader_for_predict\ndl = get_bert_data_loader_for_predict(data_path + \"test.csv\", learner)", "_____no_output_____" ], [ "preds = learner.predict(dl)", "_____no_output_____" ], [ "data = NerData.create(train_path, data_path + \"test.csv\", vocab_file)", "_____no_output_____" ] ], [ [ "IOB precision", "_____no_output_____" ] ], [ [ "from modules.train.train import validate_step\nprint(validate_step(data.valid_dl, learner.model, learner.data.id2label, learner.sup_labels))", "_____no_output_____" ] ], [ [ "Span precision", "_____no_output_____" ] ], [ [ "from modules.utils.plot_metrics import get_bert_span_report\nclf_report = get_bert_span_report(dl, preds, [])\nprint(clf_report)", " precision recall f1-score support\n\n MISC 0.759 0.794 0.776 688\n LOC 0.865 0.852 0.859 1570\n O 0.979 0.977 0.978 37693\n ORG 0.635 0.665 0.650 1533\n PER 0.878 0.879 0.878 1566\n\n micro avg 0.955 0.955 0.955 43050\n macro avg 0.823 0.833 0.828 43050\nweighted avg 0.955 0.955 0.955 43050\n\n" ] ], [ [ "### 7. Get mean and stdv on 10 runs", "_____no_output_____" ] ], [ [ "from modules.utils.plot_metrics import *\nfrom modules import NerLearner\n\n\nnum_runs = 10\nbest_reports = []\nnum_epochs = 100\nfor i in range(num_runs):\n model = BertBiLSTMAttnCRF.create(len(data.label2idx), bert_config_file, init_checkpoint_pt, enc_hidden_dim=256)\n best_model_path = \"/datadrive/models/conll-2003/exp_{}_attn_cased.cpt\".format(i)\n learner = NerLearner(model, data,\n best_model_path=best_model_path, verbose=False,\n lr=0.001, clip=5.0, sup_labels=data.id2label[5:], t_total=num_epochs * len(data.train_dl))\n learner.fit(num_epochs, target_metric='f1')\n idx, res = get_mean_max_metric(learner.history, \"f1\", True)\n best_reports.append(learner.history[idx])", "_____no_output_____" ], [ "import numpy as np", "_____no_output_____" ] ], [ [ "#### f1", "_____no_output_____" ], [ "Mean and std", "_____no_output_____" ] ], [ [ "np.mean([get_mean_max_metric([r]) for r in best_reports]), np.round(np.std([get_mean_max_metric([r]) for r in best_reports]), 3)", "_____no_output_____" ] ], [ [ "Best", "_____no_output_____" ] ], [ [ "get_mean_max_metric(best_reports)", "_____no_output_____" ] ], [ [ "#### precision", "_____no_output_____" ], [ "Mean and std", "_____no_output_____" ] ], [ [ "np.mean([get_mean_max_metric([r], \"prec\") for r in best_reports]), np.round(np.std([get_mean_max_metric([r], \"prec\") for r in best_reports]), 3)", "_____no_output_____" ] ], [ [ "Best", "_____no_output_____" ] ], [ [ "get_mean_max_metric(best_reports, \"prec\")", "_____no_output_____" ] ], [ [ "#### Test set", "_____no_output_____" ] ], [ [ "idx = np.array([get_mean_max_metric([r]) for r in best_reports]).argmax()", "_____no_output_____" ], [ "learner.load_model(\"/datadrive/models/conll-2003/exp_{}_attn_cased.cpt\".format(idx))", "_____no_output_____" ], [ "from modules.data.bert_data import get_bert_data_loader_for_predict\ndl = get_bert_data_loader_for_predict(data_path + \"test.csv\", learner)", "_____no_output_____" ], [ "from modules.train.train import validate_step\nprint(validate_step(dl, learner.model, learner.data.id2label, learner.sup_labels))", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code" ]
[ [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code", "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown", "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code", "code", "code" ] ]
ece980881118f44979cbe79d672fae731fe0d4ed
39,503
ipynb
Jupyter Notebook
preprocessing.ipynb
princessivy/course
1f567398ebce4ab49ac5cfbf318fb68ab03c6c93
[ "MIT" ]
null
null
null
preprocessing.ipynb
princessivy/course
1f567398ebce4ab49ac5cfbf318fb68ab03c6c93
[ "MIT" ]
null
null
null
preprocessing.ipynb
princessivy/course
1f567398ebce4ab49ac5cfbf318fb68ab03c6c93
[ "MIT" ]
null
null
null
32.647107
228
0.481482
[ [ [ "<a href=\"https://colab.research.google.com/github/princessivy/course/blob/main/preprocessing.ipynb\" target=\"_parent\"><img src=\"https://colab.research.google.com/assets/colab-badge.svg\" alt=\"Open In Colab\"/></a>", "_____no_output_____" ], [ "#Installs & Imports", "_____no_output_____" ] ], [ [ "!pip install ndjson --quiet\n!pip install beautifulsoup4 --quiet\n!pip install html2text --quiet\n!pip install nltk --quiet\n!pip install HanTa --quiet\n!pip install langdetect --quiet", "_____no_output_____" ], [ "import ndjson\nimport requests\nimport pandas as pd\nfrom bs4 import BeautifulSoup\nimport gzip\nfrom pathlib import Path\n# Uncomment the follwoing line if working in Google Colab \n# from google.colab import drive\nfrom collections import Counter, OrderedDict\nimport html2text\nimport numpy as np\n\nimport re\nimport nltk\nfrom nltk.corpus import stopwords\nfrom sklearn.feature_extraction import text\nfrom HanTa import HanoverTagger as ht\nfrom langdetect import detect\nimport gc\ngc.enable()\n\n#for sentiment analysis\nfrom textblob import TextBlob\n", "_____no_output_____" ] ], [ [ "# Dataloader \n### Test-Dataset", "_____no_output_____" ] ], [ [ "# Mount Google Drive\ndrive.mount('/gdrive')\ndata_path = Path('/gdrive/MyDrive/industry_data/')\nfile_name = 'test_small.ndjson.gz'\n\nwith gzip.open(data_path/file_name, \"rt\", encoding='UTF-8') as file:\n data = ndjson.load(file)\ndf_test = pd.DataFrame(data)", "Mounted at /gdrive\n" ], [ " \n # check for null entries\nif df_test.isnull().any(axis=None):\n print('\\nPreview of data with null values:\\nxxxxxxxxxxxxx')\n print(df_test[df_test.isnull().any(axis=1)].head(3))\n #missingno.matrix(df_test)\n #plt.show()\nelse:\n print('No null entries found')", "_____no_output_____" ], [ "# check for null entries\nif df_test.isnull().any(axis=None):\n print('\\nPreview of data with null values:\\nxxxxxxxxxxxxx')\n print(df_test[df_test.isnull().any(axis=1)].head(3))\n #missingno.matrix(df_test)\n #plt.show()\nelse:\n print('No null entries found')", "No null entries found\n" ], [ "# generate count statistics of duplicate entries\nif len(df_test[df_test.duplicated()]) > 0:\n print('Number of duplicated entries: ', len(df_test[df_test.duplicated()]))\n print(df_test[df_test.duplicated(keep=False)].sort_values(by=list(df_test.columns)).head())\nelse:\n print('No duplicated entries found')", "No duplicated entries found\n" ] ], [ [ "### Train-Dataset", "_____no_output_____" ] ], [ [ "drive.mount('/gdrive')\ndata_path = Path('/gdrive/MyDrive/industry_data/')\nfile_name = 'train_small.ndjson.gz'\nwith gzip.open(data_path/file_name, \"rt\", encoding='UTF-8') as file:\n data = []\n data = [ndjson.loads(line) for line in file]", "_____no_output_____" ], [ "# Nested List rausholen, Flat-List erzeugen, um Daten in DataFrame zu bekommen\nflat_list = [item for sublist in data for item in sublist]\ndf_train = pd.DataFrame(flat_list)", "_____no_output_____" ], [ "# check for null entries\nif df_train.isnull().any(axis=None):\n print('\\nPreview of data with null values:\\nxxxxxxxxxxxxx')\n print(df_train[df_train.isnull().any(axis=1)].head(3))\n #missingno.matrix(df_train)\n #plt.show()\nelse:\n print('No null entries found')", "No null entries found\n" ], [ "# generate count statistics of duplicate entries\nif len(df_train[df_train.duplicated()]) > 0:\n print('Number of duplicated entries: ', len(df_train[df_train.duplicated()]))\n print(df_train[df_train.duplicated(keep=False)].sort_values(by=list(df_train.columns)).head())\nelse:\n print(\"No duplicated entries found\")", "_____no_output_____" ] ], [ [ "# \"Datasaver\" & \"data-reloader\"\n### To ndjson and ndjson.gz\n\nIn order to save preprocessed data. If it is not saved row by row the run time crashes.", "_____no_output_____" ] ], [ [ "# save to ndjson (either regular or gzip)\n\ndef datasaver_to_zip(df, name):\n # create flat list in dict form: {'col1': 'value', 'col2': 'value', ...} from df\n flat_list_back = []\n for i in range(len(df)):\n line = df.loc[i].to_dict()\n #line['industry'] = str(line['industry']) # use if idustry number (e.g. 13) should be enclosed in '' (e.g. '13')\n flat_list_back.append([line])\n\n filename_zip = str(name) + '.ndjson.gz'\n\n with gzip.open(filename_zip, 'wt', encoding='UTF-8') as z:\n for item in flat_list_back:\n z.write('{}\\n'.format(ndjson.dumps(item)))\n\ndef datasaver_to_ndjson(df, name):\n # create flat list in dict form: {'col1': 'value', 'col2': 'value', ...} from df\n flat_list_back = []\n for i in range(len(df)):\n line = df.loc[i].to_dict()\n #line['industry'] = str(line['industry']) # use if idustry number (e.g. 13) should be enclosed in '' (e.g. '13')\n flat_list_back.append([line])\n\n filename = str(name) + '.ndjson'\n\n # https://stackoverflow.com/questions/21058935/python-json-loads-shows-valueerror-extra-data\n with open(filename, mode='w') as f:\n for item in flat_list_back:\n f.write('{}\\n'.format(ndjson.dumps(item))) ", "_____no_output_____" ], [ "def data_reloader_from_zip(file_name):\n with gzip.open(file_name, 'rt', encoding='UTF-8') as file:\n data = []\n data = [ndjson.loads(line.strip()) for line in file]\n\n flat_list = [item for sublist in data for item in sublist]\n df = pd.DataFrame(flat_list)\n\n return df\n\ndef data_reloader_from_ndjson(file_name):\n with open(file_name, 'rt', encoding='UTF-8') as file:\n data = []\n data = [ndjson.loads(line.strip()) for line in file]\n\n flat_list = [item for sublist in data for item in sublist]\n df = pd.DataFrame(flat_list)\n\n return df", "_____no_output_____" ], [ "# save file to drive \n# make sure to have folder connected (use url for access and create link for your own drive)\n# Acess: https://drive.google.com/drive/folders/1qR-9z3uFmp5Nvsb_1QrR9lU8yNE8hi6l?usp=sharing\ndrive.mount('/gdrive')\n\n!cp test_html_to_text.ndjson.gz \"/gdrive/MyDrive/industry_data_processed/\" # exchange file name", "Drive already mounted at /gdrive; to attempt to forcibly remount, call drive.mount(\"/gdrive\", force_remount=True).\n" ] ], [ [ "# HTML Feature-Checkout", "_____no_output_____" ], [ "## Checking out Tag-Occurence", "_____no_output_____" ] ], [ [ "all_tags = [] #der ersten 1000 Datensätze\n\nfor i in range(1000): #len(df_train)\n soup = BeautifulSoup(data[i][0]['html'], 'html.parser')\n #for tag in soup.findAll(True):\n #print(tag.name)\n tags = set(tag.name for tag in BeautifulSoup(data[i][0]['html'], 'html.parser').find_all())\n all_tags.extend(tags)\n\n #print(soup.get_text()[:1024])\n #print(tags)", "_____no_output_____" ], [ "# count all text, print sorted by most occurences\ncounted = Counter(all_tags)\nOrderedDict(counted.most_common())", "_____no_output_____" ] ], [ [ "## Get the whole Text between Tags", "_____no_output_____" ] ], [ [ "pd.options.mode.chained_assignment = None", "_____no_output_____" ], [ "def parse_to_text(html):\n soup = BeautifulSoup(html, features=\"html.parser\")\n\n # kill all script and style elements\n for script in soup([\"script\", \"style\"]):\n script.extract() # rip it out\n\n # get text\n text = soup.get_text()\n\n # break into lines and remove leading and trailing space on each\n lines = (line.strip() for line in text.splitlines())\n # break multi-headlines into a line each\n chunks = (phrase.strip() for line in lines for phrase in line.split(\" \"))\n # drop blank lines\n text = '\\n'.join(chunk for chunk in chunks if chunk)\n\n return text", "_____no_output_____" ], [ "# eliminate html elements from text, return text elements\n# 21 minutes for train dataset\n\n# assign column for new text\ndf_train = df_train.assign(html_to_text='')\n\nfor line in range(0, len(df_train)):\n content = parse_to_text(df_train.html[line])\n df_train.html_to_text[line] = content", "_____no_output_____" ], [ "# on test dataset\n# 7 minutes\n\ndf_test['html_to_text'] = ''\n\nfor line in range(0, len(df_test)):\n content = parse_to_text(df_test.html[line])\n df_test.html_to_text[line] = content", "_____no_output_____" ], [ "# duplicate to keep working with original df (if necessary)\ndf_test_html_to_text = df_test.copy()\ndf_test_html_to_text=df_test_html_to_text.drop(columns='html')", "_____no_output_____" ], [ "gc.collect()", "_____no_output_____" ] ], [ [ "### Preprocessing", "_____no_output_____" ] ], [ [ "nltk.download('stopwords')\nnltk.download('wordnet')", "[nltk_data] Downloading package stopwords to /root/nltk_data...\n[nltk_data] Unzipping corpora/stopwords.zip.\n[nltk_data] Downloading package wordnet to /root/nltk_data...\n[nltk_data] Unzipping corpora/wordnet.zip.\n" ], [ "# https://medium.com/analytics-vidhya/applying-text-classification-using-logistic-regression-a-comparison-between-bow-and-tf-idf-1f1ed1b83640\n\n# hier könnten wir sprachenabhängig arbeiten: clean_text_german(), clean_text_english() und anhand des lang-tags anwenden\n \ndef clean_text(mixed_text):\n '''Text Preprocessing '''\n \n # convert words to lower case\n content = mixed_text.lower()\n \n # ENGLISH use this for english text\n # Expand contractions (you've -> you have)\n #if True:\n # text = text.split()\n # new_text = []\n # for word in text:\n # if word in contractions:\n # new_text.append(contractions[word])\n # else:\n # new_text.append(word)\n # text = \" \".join(new_text)\n \n # Format words and remove unwanted characters\n #content = re.sub(r'https?:\\/\\/.*[\\r\\n]*', '', content, flags=re.MULTILINE) # brauchen wir nicht mehr, da schon geparst\n #content = re.sub(r'\\<a href', ' ', content)\n content = re.sub(r'&amp;', '', content) \n content = re.sub(r'[_\"\\-;%()|+&=*%.,!?:#$@\\[\\]/]', ' ', content)\n content = re.sub(r'<br />', ' ', content)\n content = re.sub(r'\\'', ' ', content)\n \n # remove stopwords\n content = content.split()\n stops = set(stopwords.words('german'))\n content = [w for w in content if not w in stops]\n content = ' '.join(content)\n\n # tokenize each word\n content = nltk.WordPunctTokenizer().tokenize(content)\n \n # lemmatize each token in German (reduce words to stem)\n tagger = ht.HanoverTagger('morphmodel_ger.pgz')\n word_list = []\n for w in content:\n lemma = [lemma for (word,lemma,pos) in tagger.tag_sent(w.split())]\n word_list.append(' '.join(lemma))\n\n # ENGLISH use this for english text\n # lemmatize each token\n #lemm = nltk.stem.WordNetLemmatizer()\n #content = list(map(lambda word:list(map(lemm.lemmatize, word)), content))\n \n return word_list", "_____no_output_____" ], [ "df_test_html_to_text = data_reloader_from_zip('test_html_to_text.ndjson.gz')", "_____no_output_____" ], [ "df_test_html_to_text['html_cleaned'] = ''", "_____no_output_____" ], [ "len(df_test_html_to_text)", "_____no_output_____" ], [ "for line in range(4000):\n content = clean_text(df_test_html_to_text['html_to_text'][line])\n print(line) #debugging\n df_test_html_to_text.html_cleaned[line] = content", "_____no_output_____" ], [ "datasaver_to_zip(df_test_html_to_text, 'df_test_cleaned_text')", "_____no_output_____" ] ], [ [ "## Specific Tags for additional Features", "_____no_output_____" ] ], [ [ "def getHTML(url):\n # später anpassen, wenn wir live-url abfragen! evtl. Fallunterscheidung?!\n ''' r = requests.get(url)\n r.text'''\n return BeautifulSoup(url, 'html.parser')\n\n\n\n## Img-Description from IMG-Tag\ndef getImgDescriptionHTMLtag(url):\n soup = getHTML(url)\n\n results = soup.find_all('img', alt = True)\n img_description = []\n for x in range(0,len(results)):\n first_result = results[x]\n img_description.append(first_result['alt'])\n \n return list(filter(None, img_description))\n\n\n## Title\ndef getTitleHTMLtag(url):\n soup = getHTML(url)\n\n if (soup.title is not None):\n return str(soup.title.string)\n else:\n return \"\"\n\n\n## h1\ndef getH1HTMLtag(url):\n soup = getHTML(url)\n\n heading = soup.findAll('h1')\n n = len(heading)\n\n liste = []\n for x in range(n):\n liste.append(str.strip(heading[x].text))\n\n return list(filter(None, liste))\n\n\n## h2\ndef getH2HTMLtag(url):\n soup = getHTML(url)\n\n heading = soup.findAll('h2')\n n = len(heading)\n\n liste = []\n for x in range(n):\n liste.append(str.strip(heading[x].text))\n\n return list(filter(None, liste))\n\n\n## h3\ndef getH3HTMLtag(url):\n soup = getHTML(url)\n\n heading = soup.findAll('h3')\n n = len(heading)\n\n liste = []\n for x in range(n):\n liste.append(str.strip(heading[x].text))\n\n return list(filter(None, liste))\n\n\n## strong - fragwürdig\ndef getStrongHTMLtag(url):\n soup = getHTML(url)\n\n heading = soup.findAll('strong')\n n = len(heading)\n\n liste = []\n for x in range(n):\n liste.append(str.strip(heading[x].text))\n\n return list(filter(None, liste))\n\n\n## bold\ndef getBoldHTMLtag(url):\n soup = getHTML(url)\n\n heading = soup.findAll('bold')\n n = len(heading)\n\n liste = []\n for x in range(n):\n liste.append(str.strip(heading[x].text))\n\n return list(filter(None, liste))\n \n\n## language code\ndef getLangHTMLtag(url):\n \n try:\n soup = getHTML(url)\n body_text = soup.body.get_text()\n return detect(body_text)\n \n except:\n return str(\"NaN\")\n \n \n## figcaption\ndef getFigCaptionHTMLtag(url):\n soup = getHTML(url)\n\n heading = soup.findAll('figcaption')\n n = len(heading)\n\n liste = []\n for x in range(n):\n liste.append(str.strip(heading[x].text))\n\n return list(filter(None, liste))", "_____no_output_____" ], [ "df_test = df_test.assign(img_alt='', title='', h1='', h2='', h3='', strong='', bold='', lang_code='', figcaption='')", "_____no_output_____" ], [ "# Befüllen der Extra-Features\n\ndef retrieve_features(df):\n for i in range (4, 13):\n for j in range(0, len(df)):\n if i == 4:\n df.iloc[:, i][j] = getImgDescriptionHTMLtag(df.html[j])\n elif i == 5:\n df.iloc[:, i][j] = getTitleHTMLtag(df.html[j])\n elif i == 6:\n df.iloc[:, i][j] = getH1HTMLtag(df.html[j])\n elif i == 7:\n df.iloc[:, i][j] = getH2HTMLtag(df.html[j])\n elif i == 8:\n df.iloc[:, i][j] = getH3HTMLtag(df.html[j])\n elif i == 9:\n df.iloc[:, i][j] = getStrongHTMLtag(df.html[j])\n elif i == 10:\n df.iloc[:, i][j] = getBoldHTMLtag(df.html[j])\n elif i == 11:\n df.iloc[:, i][j] = getLangHTMLtag(df.html[j])\n elif i == 12:\n df.iloc[:, i][j] = getFigCaptionHTMLtag(df.html[j])", "_____no_output_____" ], [ "# aus List-Elementen in DataFrame einfache Strings machen, um besser cleanen zu können\ndef convert_features_toString(df):\n # img\n for x in range(len(df)):\n df.img_alt[x] = ' '.join(df.img_alt[x])\n\n # h1\n for x in range(len(df)):\n df.h1[x] = ' '.join(df.h1[x])\n\n # h2\n for x in range(len(df)):\n df.h2[x] = ' '.join(df.h2[x])\n\n # h3\n for x in range(len(df)):\n df.h3[x] = ' '.join(df.h3[x])\n\n # strong\n for x in range(len(df)):\n df.strong[x] = ' '.join(df.strong[x])\n\n # bold\n for x in range(len(df)):\n df.bold[x] = ' '.join(df.bold[x])\n\n # figcaption\n for x in range(len(df)):\n df.figcaption[x] = ' '.join(df.figcaption[x])", "_____no_output_____" ], [ "# alle extra-features durchgehen um zu gucken, welche wir noch durch clean_text laufen lassen müssen\ncolumns = df_train_reloaded.columns.tolist()\ncolumns = columns[4:13]", "_____no_output_____" ], [ "# lassen lang-code weg\ncolumns = ['img_alt','title','h1','h2','h3','strong','bold','figcaption']", "_____no_output_____" ], [ "# alle extra-features durch clean_text schicken, ohne lang-feature\ndef clean_dataframe(df):\n columns = ['img_alt','title','h1','h2','h3','strong','bold','figcaption']\n for x in columns:\n print(x)\n for y in range(len(df)):\n df[x][y] = clean_text(df[x][y])", "_____no_output_____" ], [ "# perform preprocessing\ndf_test = df_test.assign(img_alt='', title='', h1='', h2='', h3='', strong='', bold='', lang_code='', figcaption='')\nretrieve_features(df_test)\nconvert_features_toString(df_test)\nclean_dataframe(df_test)\ndf_test = df_test.drop(columns=['html'], axis=1)\ndatasaver_to_ndjson(df=df_test, name='df_test_preprocessed')\n", "/var/folders/0w/wzymnpfd4rjdkg4k8khhk6dc0000gn/T/ipykernel_44039/2078399981.py:7: SettingWithCopyWarning: \nA value is trying to be set on a copy of a slice from a DataFrame\n\nSee the caveats in the documentation: https://pandas.pydata.org/pandas-docs/stable/user_guide/indexing.html#returning-a-view-versus-a-copy\n df.iloc[:, i][j] = getImgDescriptionHTMLtag(df.html[j])\n/var/folders/0w/wzymnpfd4rjdkg4k8khhk6dc0000gn/T/ipykernel_44039/2078399981.py:9: SettingWithCopyWarning: \nA value is trying to be set on a copy of a slice from a DataFrame\n\nSee the caveats in the documentation: https://pandas.pydata.org/pandas-docs/stable/user_guide/indexing.html#returning-a-view-versus-a-copy\n df.iloc[:, i][j] = getTitleHTMLtag(df.html[j])\n/var/folders/0w/wzymnpfd4rjdkg4k8khhk6dc0000gn/T/ipykernel_44039/2078399981.py:11: SettingWithCopyWarning: \nA value is trying to be set on a copy of a slice from a DataFrame\n\nSee the caveats in the documentation: https://pandas.pydata.org/pandas-docs/stable/user_guide/indexing.html#returning-a-view-versus-a-copy\n df.iloc[:, i][j] = getH1HTMLtag(df.html[j])\n/var/folders/0w/wzymnpfd4rjdkg4k8khhk6dc0000gn/T/ipykernel_44039/2078399981.py:13: SettingWithCopyWarning: \nA value is trying to be set on a copy of a slice from a DataFrame\n\nSee the caveats in the documentation: https://pandas.pydata.org/pandas-docs/stable/user_guide/indexing.html#returning-a-view-versus-a-copy\n df.iloc[:, i][j] = getH2HTMLtag(df.html[j])\n/var/folders/0w/wzymnpfd4rjdkg4k8khhk6dc0000gn/T/ipykernel_44039/2078399981.py:15: SettingWithCopyWarning: \nA value is trying to be set on a copy of a slice from a DataFrame\n\nSee the caveats in the documentation: https://pandas.pydata.org/pandas-docs/stable/user_guide/indexing.html#returning-a-view-versus-a-copy\n df.iloc[:, i][j] = getH3HTMLtag(df.html[j])\n/var/folders/0w/wzymnpfd4rjdkg4k8khhk6dc0000gn/T/ipykernel_44039/2078399981.py:17: SettingWithCopyWarning: \nA value is trying to be set on a copy of a slice from a DataFrame\n\nSee the caveats in the documentation: https://pandas.pydata.org/pandas-docs/stable/user_guide/indexing.html#returning-a-view-versus-a-copy\n df.iloc[:, i][j] = getStrongHTMLtag(df.html[j])\n/var/folders/0w/wzymnpfd4rjdkg4k8khhk6dc0000gn/T/ipykernel_44039/2078399981.py:19: SettingWithCopyWarning: \nA value is trying to be set on a copy of a slice from a DataFrame\n\nSee the caveats in the documentation: https://pandas.pydata.org/pandas-docs/stable/user_guide/indexing.html#returning-a-view-versus-a-copy\n df.iloc[:, i][j] = getBoldHTMLtag(df.html[j])\n/var/folders/0w/wzymnpfd4rjdkg4k8khhk6dc0000gn/T/ipykernel_44039/2078399981.py:21: SettingWithCopyWarning: \nA value is trying to be set on a copy of a slice from a DataFrame\n\nSee the caveats in the documentation: https://pandas.pydata.org/pandas-docs/stable/user_guide/indexing.html#returning-a-view-versus-a-copy\n df.iloc[:, i][j] = getLangHTMLtag(df.html[j])\n/var/folders/0w/wzymnpfd4rjdkg4k8khhk6dc0000gn/T/ipykernel_44039/2078399981.py:23: SettingWithCopyWarning: \nA value is trying to be set on a copy of a slice from a DataFrame\n\nSee the caveats in the documentation: https://pandas.pydata.org/pandas-docs/stable/user_guide/indexing.html#returning-a-view-versus-a-copy\n df.iloc[:, i][j] = getFigCaptionHTMLtag(df.html[j])\n/var/folders/0w/wzymnpfd4rjdkg4k8khhk6dc0000gn/T/ipykernel_44039/41399377.py:5: SettingWithCopyWarning: \nA value is trying to be set on a copy of a slice from a DataFrame\n\nSee the caveats in the documentation: https://pandas.pydata.org/pandas-docs/stable/user_guide/indexing.html#returning-a-view-versus-a-copy\n df.img_alt[x] = ' '.join(df.img_alt[x])\n/var/folders/0w/wzymnpfd4rjdkg4k8khhk6dc0000gn/T/ipykernel_44039/41399377.py:25: SettingWithCopyWarning: \nA value is trying to be set on a copy of a slice from a DataFrame\n\nSee the caveats in the documentation: https://pandas.pydata.org/pandas-docs/stable/user_guide/indexing.html#returning-a-view-versus-a-copy\n df.bold[x] = ' '.join(df.bold[x])\n/var/folders/0w/wzymnpfd4rjdkg4k8khhk6dc0000gn/T/ipykernel_44039/41399377.py:29: SettingWithCopyWarning: \nA value is trying to be set on a copy of a slice from a DataFrame\n\nSee the caveats in the documentation: https://pandas.pydata.org/pandas-docs/stable/user_guide/indexing.html#returning-a-view-versus-a-copy\n df.figcaption[x] = ' '.join(df.figcaption[x])\n" ] ], [ [ "# Datei(en) exportieren", "_____no_output_____" ] ], [ [ "df_train_work.to_json('train_preprocessed.json')", "_____no_output_____" ] ], [ [ "#Sentiment-Analysis", "_____no_output_____" ] ], [ [ "df_train['sentiment_analysis'] =''", "_____no_output_____" ], [ "# spalte aus welcher sentiment-analysis gemacht wird, in string casten\ndf_train['pure_text'] = df_train['pure_text'].astype(str)", "_____no_output_____" ], [ "for x in range(0, len(df_train)):\n df_train.sentiment_analysis[x] = round(TextBlob(df_train['pure_text'][x]).sentiment.polarity,2)", "_____no_output_____" ], [ "# Überprüfung des means\ndf_train.groupby('industry_label')['sentiment_analysis'].mean()", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code" ]
[ [ "markdown", "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown", "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code", "code", "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code", "code", "code" ] ]
ece98d1248928ce7f953664722b7deb2bb678314
17,896
ipynb
Jupyter Notebook
wrappers/Sklearn-wrapper-plus-scalers.ipynb
gverbock/feature-engine-examples
c5f6bcbe3844f51481bd102a6c21e5891468dead
[ "BSD-3-Clause" ]
null
null
null
wrappers/Sklearn-wrapper-plus-scalers.ipynb
gverbock/feature-engine-examples
c5f6bcbe3844f51481bd102a6c21e5891468dead
[ "BSD-3-Clause" ]
4
2021-09-13T07:22:51.000Z
2021-12-28T12:18:05.000Z
examples/wrappers/Sklearn-wrapper-plus-scalers.ipynb
TremaMiguel/feature_engine
117ea3061ec9cf65f9d012aff4875d2b88e8cf71
[ "BSD-3-Clause" ]
5
2021-09-10T15:34:08.000Z
2022-03-06T07:28:07.000Z
30.749141
102
0.413835
[ [ [ "import pandas as pd\nimport numpy as np\n\nimport matplotlib.pyplot as plt\n\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.preprocessing import StandardScaler\n\nfrom feature_engine.wrappers import SklearnTransformerWrapper", "_____no_output_____" ], [ "data = pd.read_csv('houseprice.csv')\ndata.head()", "_____no_output_____" ], [ "# let's separate into training and testing set\n\nX_train, X_test, y_train, y_test = train_test_split(\n data.drop(['Id', 'SalePrice'], axis=1), data['SalePrice'], test_size=0.3, random_state=0)\n\nX_train.shape, X_test.shape", "_____no_output_____" ] ], [ [ "## Scaling", "_____no_output_____" ] ], [ [ "cols = [var for var in X_train.columns if X_train[var].dtypes !='O']\n\ncols", "_____no_output_____" ], [ "# let's apply the standard scaler on the above variables\n\nscaler = SklearnTransformerWrapper(transformer = StandardScaler(),\n variables = cols)\n\nscaler.fit(X_train.fillna(0))", "_____no_output_____" ], [ "X_train = scaler.transform(X_train.fillna(0))\nX_test = scaler.transform(X_test.fillna(0))", "_____no_output_____" ], [ "# mean values, learnt by the StandardScaler\nscaler.transformer_.mean_", "_____no_output_____" ], [ "# std values, learnt by the StandardScaler\nscaler.transformer_.scale_", "_____no_output_____" ], [ "# the mean of the scaled variables is 0\nX_train[cols].mean()", "_____no_output_____" ], [ "# the std of the scaled variables is ~1\n\nX_train[cols].std()", "_____no_output_____" ] ] ]
[ "code", "markdown", "code" ]
[ [ "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code", "code", "code" ] ]
ece9aba7673ea5c50c7c5fa6e82388d857daeb10
21,492
ipynb
Jupyter Notebook
testing/sahel_cropmask/5_Accuracy_assessment.ipynb
digitalearthafrica/crop-mask
18ae773c4d5eb71c0add765260a1032c46c68a0e
[ "Apache-2.0" ]
11
2020-12-15T04:09:41.000Z
2022-01-19T11:07:21.000Z
testing/sahel_cropmask/5_Accuracy_assessment.ipynb
digitalearthafrica/crop-mask
18ae773c4d5eb71c0add765260a1032c46c68a0e
[ "Apache-2.0" ]
15
2021-03-15T02:17:32.000Z
2022-02-24T02:50:01.000Z
testing/sahel_cropmask/5_Accuracy_assessment.ipynb
digitalearthafrica/crop-mask
18ae773c4d5eb71c0add765260a1032c46c68a0e
[ "Apache-2.0" ]
4
2020-12-16T04:48:36.000Z
2021-03-30T16:51:37.000Z
29.400821
378
0.468407
[ [ [ "# Accuracy assessment of the Sahel Africa Cropland Mask\n\n", "_____no_output_____" ], [ "## Description\n\nNow that we have run classifications for the Sahel Africa AEZ, its time to conduct an accuracy assessment. The data used for assessing the accuracy was collected previously and set aside. Its stored in the data/ folder: `data/Validation_samples.shp` \n\nThis notebook will output a `confusion error matrix` containing Overall, Producer's, and User's accuracy, along with the F1 score for each class.", "_____no_output_____" ], [ "***\n## Getting started\n\nTo run this analysis, run all the cells in the notebook, starting with the \"Load packages\" cell. ", "_____no_output_____" ], [ "### Load Packages", "_____no_output_____" ] ], [ [ "import os\nimport sys\nimport glob\nimport rasterio\nimport pandas as pd\nimport numpy as np\nimport seaborn as sn\nimport matplotlib.pyplot as plt\nimport geopandas as gpd\nfrom sklearn.metrics import f1_score\nfrom odc.io.cgroups import get_cpu_quota\n\nfrom deafrica_tools.spatial import zonal_stats_parallel", "_____no_output_____" ] ], [ [ "## Analysis Parameters\n\n* `pred_tif` : a binary classification of crop/no-crop output by the ML script.\n* `grd_truth` : a shapefile containing crop/no-crop points to serve as the \"ground-truth\" dataset\n* `aez_region` : a shapefile used to limit the ground truth points to the region where the model has classified crop/non-crop\n", "_____no_output_____" ] ], [ [ "pred_tif = \"results/classifications/20211110/Sahel_gm_mads_two_seasons_20211110_mosaic_clipped.tif\"\n# pred_tif = 'results/classifications/gfsad_sahel_masked_clipped.tif'\ngrd_truth = 'data/validation_samples.shp'", "_____no_output_____" ] ], [ [ "### Load the datasets\n\n`Ground truth points`", "_____no_output_____" ] ], [ [ "#ground truth shapefile\nground_truth = gpd.read_file(grd_truth).to_crs('EPSG:6933')", "_____no_output_____" ], [ "# rename the class column to 'actual'\nground_truth = ground_truth.rename(columns={'Class':'Actual'})\nground_truth.head()", "_____no_output_____" ] ], [ [ "Reclassify 'Actual' column to match the raster values", "_____no_output_____" ] ], [ [ "ground_truth['Actual'] = np.where(ground_truth['Actual']=='non-crop', 0, ground_truth['Actual'])\nground_truth['Actual'] = np.where(ground_truth['Actual']=='crop', 1, ground_truth['Actual'])", "_____no_output_____" ] ], [ [ "### This cell if point sampling", "_____no_output_____" ] ], [ [ "# #Point sampling of raster for validation purpose\n# prediction = rasterio.open(pred_tif)\n# coords = [(x,y) for x, y in zip(ground_truth.geometry.x, ground_truth.geometry.y)]\n# # Sample the raster at every point location and store values in DataFrame\n# ground_truth['Prediction'] = [int(x[0]) for x in prediction.sample(coords)]", "_____no_output_____" ] ], [ [ "### The next two cells if polygon sampling\n#### Convert points into polygons\n\nWhen the validation data was collected, 40x40m polygons were evaluated as either crop/non-crop rather than points, so we want to sample the raster using the same small polygons. We'll find the majority or 'mode' statistic within the polygon and use that to compare with the validation dataset.\n", "_____no_output_____" ] ], [ [ "#set radius (in metres) around points\nradius = 20\n\n#create circle buffer around points, then find envelope\nground_truth['geometry'] = ground_truth['geometry'].buffer(radius).envelope\n\n#export to file for use in zonal-stats\nground_truth.to_file(grd_truth[:-4]+\"_poly.shp\")", "_____no_output_____" ] ], [ [ "### Calculate zonal statistics\n\nWe want to know what the majority pixel value is inside each validation polygon.", "_____no_output_____" ] ], [ [ "zonal_stats_parallel(shp=grd_truth[:-4]+\"_poly.shp\",\n raster=pred_tif,\n statistics=['majority'],\n out_shp=grd_truth[:-4]+\"_poly.shp\",\n ncpus=round(get_cpu_quota()),\n nodata=-999)\n\n#read in the results\nx=gpd.read_file(grd_truth[:-4]+\"_poly.shp\")\n\n#add result to original ground truth array\nground_truth['Prediction'] = x['majority'].astype(np.int16)\n\n#Remove the temporary shapefile we made\n[os.remove(i) for i in glob.glob(grd_truth[:-4]+\"_poly\"+'*')]", "/env/lib/python3.8/site-packages/fiona/collection.py:350: FionaDeprecationWarning: Collection slicing is deprecated and will be disabled in a future version.\n return self.session.__getitem__(item)\n" ] ], [ [ "***\n\n## Create a confusion matrix", "_____no_output_____" ] ], [ [ "confusion_matrix = pd.crosstab(ground_truth['Actual'],\n ground_truth['Prediction'],\n rownames=['Actual'],\n colnames=['Prediction'],\n margins=True)\n\nconfusion_matrix", "_____no_output_____" ] ], [ [ "### Calculate User's and Producer's Accuracy", "_____no_output_____" ], [ "`Producer's Accuracy`", "_____no_output_____" ] ], [ [ "confusion_matrix[\"Producer's\"] = [confusion_matrix.loc[0, 0] / confusion_matrix.loc[0, 'All'] * 100,\n confusion_matrix.loc[1, 1] / confusion_matrix.loc[1, 'All'] * 100,\n np.nan]", "_____no_output_____" ] ], [ [ "`User's Accuracy`", "_____no_output_____" ] ], [ [ "users_accuracy = pd.Series([confusion_matrix[0][0] / confusion_matrix[0]['All'] * 100,\n confusion_matrix[1][1] / confusion_matrix[1]['All'] * 100]\n ).rename(\"User's\")\n\nconfusion_matrix = confusion_matrix.append(users_accuracy)", "_____no_output_____" ] ], [ [ "`Overall Accuracy`", "_____no_output_____" ] ], [ [ "confusion_matrix.loc[\"User's\",\"Producer's\"] = (confusion_matrix.loc[0, 0] + \n confusion_matrix.loc[1, 1]) / confusion_matrix.loc['All', 'All'] * 100", "_____no_output_____" ] ], [ [ "`F1 Score`\n\nThe F1 score is the harmonic mean of the precision and recall, where an F1 score reaches its best value at 1 (perfect precision and recall), and is calculated as:\n\n$$\n\\begin{aligned}\n\\text{Fscore} = 2 \\times \\frac{\\text{UA} \\times \\text{PA}}{\\text{UA} + \\text{PA}}.\n\\end{aligned}\n$$\n\nWhere UA = Users Accuracy, and PA = Producer's Accuracy", "_____no_output_____" ] ], [ [ "fscore = pd.Series([(2*(confusion_matrix.loc[\"User's\", 0]*confusion_matrix.loc[0, \"Producer's\"]) / (confusion_matrix.loc[\"User's\", 0]+confusion_matrix.loc[0, \"Producer's\"])) / 100,\n f1_score(ground_truth['Actual'].astype(np.int8), ground_truth['Prediction'].astype(np.int8), average='binary')]\n ).rename(\"F-score\")\n\nconfusion_matrix = confusion_matrix.append(fscore)", "_____no_output_____" ] ], [ [ "### Tidy Confusion Matrix\n\n* Limit decimal places,\n* Add readable class names\n* Remove non-sensical values ", "_____no_output_____" ] ], [ [ "# round numbers\nconfusion_matrix = confusion_matrix.round(decimals=2)", "_____no_output_____" ], [ "# rename booleans to class names\nconfusion_matrix = confusion_matrix.rename(columns={0:'Non-crop', 1:'Crop', 'All':'Total'},\n index={0:'Non-crop', 1:'Crop', 'All':'Total'})", "_____no_output_____" ], [ "#remove the nonsensical values in the table\nconfusion_matrix.loc[\"User's\", 'Total'] = '--'\nconfusion_matrix.loc['Total', \"Producer's\"] = '--'\nconfusion_matrix.loc[\"F-score\", 'Total'] = '--'\nconfusion_matrix.loc[\"F-score\", \"Producer's\"] = '--'", "_____no_output_____" ], [ "confusion_matrix", "_____no_output_____" ] ], [ [ "### Export csv", "_____no_output_____" ] ], [ [ "# confusion_matrix.to_csv('results/Sahel_confusion_matrix.csv')", "_____no_output_____" ] ], [ [ "## Next steps\n\nThis is the last notebook in the `Southern Africa Cropland Mask` workflow! To revist any of the other notebooks, use the links below.\n\n1. [Extract_training_data](1_Extract_training_data.ipynb) \n2. [Inspect_training_data](2_Inspect_training_data.ipynb)\n3. [Train_fit_evaluate_classifier](3_Train_fit_evaluate_classifier.ipynb)\n4. [Predict](4_Predict.ipynb)\n5. [Object-based_filtering](5_Object-based_filtering.ipynb)\n6. **Accuracy_assessment (this notebook)**", "_____no_output_____" ], [ "***\n\n## Additional information\n\n**License:** The code in this notebook is licensed under the [Apache License, Version 2.0](https://www.apache.org/licenses/LICENSE-2.0). \nDigital Earth Africa data is licensed under the [Creative Commons by Attribution 4.0](https://creativecommons.org/licenses/by/4.0/) license.\n\n**Contact:** If you need assistance, please post a question on the [Open Data Cube Slack channel](http://slack.opendatacube.org/) or on the [GIS Stack Exchange](https://gis.stackexchange.com/questions/ask?tags=open-data-cube) using the `open-data-cube` tag (you can view previously asked questions [here](https://gis.stackexchange.com/questions/tagged/open-data-cube)).\nIf you would like to report an issue with this notebook, you can file one on [Github](https://github.com/digitalearthafrica/deafrica-sandbox-notebooks).\n\n**Last modified:** Dec 2020\n", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown" ]
[ [ "markdown", "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code", "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown" ] ]
ece9be8f645f715989fc6f3f9bd5f5351e4e2a13
75,535
ipynb
Jupyter Notebook
ex6/Exercise6.ipynb
doyeonkp/BigData_2017
51e71afb364d70e30e52b546331e0a44f1265230
[ "MIT" ]
null
null
null
ex6/Exercise6.ipynb
doyeonkp/BigData_2017
51e71afb364d70e30e52b546331e0a44f1265230
[ "MIT" ]
null
null
null
ex6/Exercise6.ipynb
doyeonkp/BigData_2017
51e71afb364d70e30e52b546331e0a44f1265230
[ "MIT" ]
null
null
null
114.273828
15,590
0.856662
[ [ [ "import pandas as p\ndata = p.read_csv('train.csv')\ndata.head()", "_____no_output_____" ], [ "X = data.drop('date',axis=1)\nX = X.drop('Occupancy', axis=1)\nY = data['Occupancy']", "_____no_output_____" ], [ "maxdepth = [2, 5, 10, 15, 20]\nk = [1, 2, 3, 4, 5, 10,15]\nc = [0.001, 0.01, 0.1, 0.5, 1]", "_____no_output_____" ], [ "from sklearn import tree\nfrom sklearn.metrics import accuracy_score\nfrom sklearn.model_selection import cross_val_score\naccuracy = []\nfor i in maxdepth:\n clf = tree.DecisionTreeClassifier(max_depth=i)\n scores = cross_val_score(clf, X, Y, cv=5)\n accuracy.append(scores.mean())\n", "_____no_output_____" ], [ "accuracy", "_____no_output_____" ], [ "import pandas as p\ndata = p.read_csv('test.csv')\ndata.head()\n\nX_test = data.drop('date',axis=1)\nX_test = X_test.drop('Occupancy', axis=1)\nY_test = data['Occupancy']", "_____no_output_____" ], [ "%matplotlib inline\nimport matplotlib\nimport matplotlib.pyplot as plt \nplt.plot(maxdepth,accuracy,'ro-')", "_____no_output_____" ], [ "from sklearn import tree\nfrom sklearn.neighbors import KNeighborsClassifier\nfrom sklearn.model_selection import cross_val_score\naccuracy = []\nfor i in k:\n clf = KNeighborsClassifier(n_neighbors=i)\n scores = cross_val_score(clf, X, Y, cv=5)\n accuracy.append(scores.mean())", "_____no_output_____" ], [ "accuracy", "_____no_output_____" ], [ "%matplotlib inline\nimport matplotlib\nimport matplotlib.pyplot as plt \nplt.plot(k,accuracy,'ro-')", "_____no_output_____" ], [ "from sklearn import tree\nfrom sklearn.metrics import accuracy_score\nfrom sklearn.model_selection import cross_val_score\nfrom sklearn.linear_model import LogisticRegression\naccuracy = []\nfor i in c:\n clf = LogisticRegression(C=i)\n scores = cross_val_score(clf, X, Y, cv=5)\n accuracy.append(scores.mean())\n ", "_____no_output_____" ], [ "accuracy", "_____no_output_____" ], [ "%matplotlib inline\nimport matplotlib\nimport matplotlib.pyplot as plt \nplt.plot(c,accuracy,'ro-')", "_____no_output_____" ], [ "clf = tree.DecisionTreeClassifier(max_depth=5.0)\nclf = clf.fit(X, Y)\nY_pred = clf.predict(X_test)\naccuracy_test_Decision = accuracy_score(Y_test,Y_pred)\n\naccuracy_test_Decision", "_____no_output_____" ], [ "clf = KNeighborsClassifier(n_neighbors=3)\nclf = clf.fit(X, Y)\nY_pred = clf.predict(X_test)\naccuracy_test_Knearest = accuracy_score(Y_test,Y_pred)\n\naccuracy_test_Knearest", "_____no_output_____" ], [ "clf = LogisticRegression(C=1.0)\nclf = clf.fit(X, Y)\nY_pred = clf.predict(X_test)\naccuracy_test_logistic = accuracy_score(Y_test,Y_pred)\n\naccuracy_test_logistic", "_____no_output_____" ], [ "accuracy_test_Decision=[]\naccuracy_score_list=[]\nfor i in maxdepth:\n clf = tree.DecisionTreeClassifier(max_depth=i)\n clf = clf.fit(X, Y)\n Y_pred_train = clf.predict(X)\n Y_pred = clf.predict(X_test)\n accuracy_score_list.append(accuracy_score(Y,Y_pred_train))\n accuracy_test_Decision.append(accuracy_score(Y_test,Y_pred))", "_____no_output_____" ], [ "import matplotlib.pyplot as plt \nplt.plot(maxdepth,accuracy_score_list,'ro-',maxdepth,accuracy_test_Decision,'bv--')", "_____no_output_____" ], [ "accuracy_test_Knearest=[]\naccuracy_score_list=[]\nfor i in maxdepth:\n clf = KNeighborsClassifier(n_neighbors=i)\n clf = clf.fit(X, Y)\n Y_pred_train = clf.predict(X)\n Y_pred = clf.predict(X_test)\n accuracy_score_list.append(accuracy_score(Y,Y_pred_train))\n accuracy_test_Knearest.append(accuracy_score(Y_test,Y_pred))", "_____no_output_____" ], [ "from sklearn import linear_model\nfrom sklearn.svm import SVC\n\nDesicion_acc_list = []\nKNeighbors_acc_list=[]\nlogistic_acc_list = []\n\nclf = tree.DecisionTreeClassifier(max_depth=5.0)\nclf = clf.fit(X, Y)\nY_pred = clf.predict(X_test)\naccuracy_test_Decision = accuracy_score(Y_test,Y_pred)\n\nclf = KNeighborsClassifier(n_neighbors=3)\nclf = clf.fit(X, Y)\nY_pred = clf.predict(X_test)\naccuracy_test_Knearest = accuracy_score(Y_test,Y_pred)\n\n \nclf = LogisticRegression(C=1.0)\nclf = clf.fit(X, Y)\nY_pred = clf.predict(X_test)\naccuracy_test_logistic = accuracy_score(Y_test,Y_pred)\n\nmethods = ['Dtree','Kneighbor','logistic']\nacc = [accuracy_test_Decision,accuracy_test_Knearest,accuracy_test_logistic]\nplt.bar([1, 2, 3], acc)\nplt.xticks([1,2,3], methods)\n\n# ystick = [0.1,0.2,0.3,0.4,0.5,0.6,0.7,0.8,0.85,0.97,0.98,0.9]\n\nprint accuracy_test_Decision\n\nprint accuracy_test_Knearest\n\nprint accuracy_test_logistic\n\n\n", "0.984753607405\n0.988429077049\n0.987612306017\n" ], [ "#######################################\n# Show differency to re-scale the y-axis by yscale('log')\n#######################################\nmethods = ['Dtree','Kneighbor','logistic']\nacc = [accuracy_test_Decision,accuracy_test_Knearest,accuracy_test_logistic]\nplt.bar([1, 2, 3], acc)\nplt.xticks([1,2,3], methods)\nplt.yscale(\"log\")", "_____no_output_____" ] ] ]
[ "code" ]
[ [ "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code" ] ]
ece9c0fa4a9d92cf49cc96a40eeccf114caca44c
24,547
ipynb
Jupyter Notebook
site/en/r1/tutorials/images/hub_with_keras.ipynb
BlazerYoo/docs
37094bf40829f515e1acf683fdc452cdb02738e9
[ "Apache-2.0" ]
5,672
2018-08-27T18:49:33.000Z
2022-03-31T07:52:12.000Z
site/en/r1/tutorials/images/hub_with_keras.ipynb
agentdavidjoseph/docs
9fbb00bd50f962d5edfba09f426b761ae9283aec
[ "Apache-2.0" ]
1,635
2018-08-28T15:27:17.000Z
2022-03-23T23:15:14.000Z
site/en/r1/tutorials/images/hub_with_keras.ipynb
agentdavidjoseph/docs
9fbb00bd50f962d5edfba09f426b761ae9283aec
[ "Apache-2.0" ]
6,035
2018-08-27T19:13:09.000Z
2022-03-31T08:55:13.000Z
26.710555
257
0.498595
[ [ [ "##### Copyright 2018 The TensorFlow Authors.", "_____no_output_____" ] ], [ [ "#@title Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# https://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.", "_____no_output_____" ] ], [ [ "# Hub with Keras\n\n<table class=\"tfo-notebook-buttons\" align=\"left\">\n <td>\n <a target=\"_blank\" href=\"https://colab.research.google.com/github/tensorflow/docs/blob/master/site/en/r1/tutorials/images/hub_with_keras.ipynb\"><img src=\"https://www.tensorflow.org/images/colab_logo_32px.png\" />Run in Google Colab</a>\n </td>\n <td>\n <a target=\"_blank\" href=\"https://github.com/tensorflow/docs/blob/master/site/en/r1/tutorials/images/hub_with_keras.ipynb\"><img src=\"https://www.tensorflow.org/images/GitHub-Mark-32px.png\" />View source on GitHub</a>\n </td>\n <td>\n <a href=\"https://tfhub.dev/s?module-type=image-classification\"><img src=\"https://www.tensorflow.org/images/hub_logo_32px.png\" />See TF Hub model</a>\n </td>\n</table>", "_____no_output_____" ], [ "> Note: This is an archived TF1 notebook. These are configured\nto run in TF2's \n[compatbility mode](https://www.tensorflow.org/guide/migrate)\nbut will run in TF1 as well. To use TF1 in Colab, use the\n[%tensorflow_version 1.x](https://colab.research.google.com/notebooks/tensorflow_version.ipynb)\nmagic.", "_____no_output_____" ], [ "[TensorFlow Hub](http://tensorflow.org/hub) is a way to share pretrained model components. See the [TensorFlow Module Hub](https://tfhub.dev/) for a searchable listing of pre-trained models.\n\nThis tutorial demonstrates:\n\n1. How to use TensorFlow Hub with `tf.keras`.\n1. How to do image classification using TensorFlow Hub.\n1. How to do simple transfer learning.", "_____no_output_____" ], [ "## Setup", "_____no_output_____" ], [ "### Imports", "_____no_output_____" ] ], [ [ "!pip install -U tensorflow_hub", "_____no_output_____" ], [ "import matplotlib.pylab as plt\n\nimport tensorflow.compat.v1 as tf\n", "_____no_output_____" ], [ "import os\nimport tensorflow_hub as hub\nfrom tensorflow.keras import layers\n\n# Load compressed models from tensorflow_hub\nos.environ[\"TFHUB_MODEL_LOAD_FORMAT\"] = \"COMPRESSED\"", "_____no_output_____" ] ], [ [ "## An ImageNet classifier", "_____no_output_____" ], [ "### Download the classifier\n\nUse `hub.module` to load a mobilenet, and `tf.keras.layers.Lambda` to wrap it up as a keras layer.\n\nThe URL of any [TF2-compatible image classification module](https://tfhub.dev/s?module-type=image-classification&q=tf2) from tfhub.dev will work here.", "_____no_output_____" ] ], [ [ "classifier_url =\"https://tfhub.dev/google/tf2-preview/mobilenet_v2/classification/2\" #@param {type:\"string\"}", "_____no_output_____" ], [ "IMAGE_SHAPE = (224, 224)\n\nclassifier = tf.keras.Sequential([\n hub.KerasLayer(classifier_url, input_shape=IMAGE_SHAPE+(3,))\n])", "_____no_output_____" ] ], [ [ "### Run it on a single image", "_____no_output_____" ], [ "Download a single image to try the model on.", "_____no_output_____" ] ], [ [ "import numpy as np\nimport PIL.Image as Image\n\ngrace_hopper = tf.keras.utils.get_file('image.jpg','https://storage.googleapis.com/download.tensorflow.org/example_images/grace_hopper.jpg')\ngrace_hopper = Image.open(grace_hopper).resize(IMAGE_SHAPE)\ngrace_hopper", "_____no_output_____" ], [ "grace_hopper = np.array(grace_hopper)/255.0\ngrace_hopper.shape", "_____no_output_____" ] ], [ [ "Add a batch dimension, and pass the image to the model.", "_____no_output_____" ] ], [ [ "result = classifier.predict(grace_hopper[np.newaxis, ...])\nresult.shape", "_____no_output_____" ] ], [ [ "The result is a 1001 element vector of logits, rating the probability of each class for the image.\n\nSo the top class ID can be found with argmax:", "_____no_output_____" ] ], [ [ "predicted_class = np.argmax(result[0], axis=-1)\npredicted_class", "_____no_output_____" ] ], [ [ "### Decode the predictions\n\nWe have the predicted class ID,\nFetch the `ImageNet` labels, and decode the predictions", "_____no_output_____" ] ], [ [ "labels_path = tf.keras.utils.get_file('ImageNetLabels.txt','https://storage.googleapis.com/download.tensorflow.org/data/ImageNetLabels.txt')\nimagenet_labels = np.array(open(labels_path).read().splitlines())", "_____no_output_____" ], [ "plt.imshow(grace_hopper)\nplt.axis('off')\npredicted_class_name = imagenet_labels[predicted_class]\n_ = plt.title(\"Prediction: \" + predicted_class_name.title())", "_____no_output_____" ] ], [ [ "## Simple transfer learning", "_____no_output_____" ], [ "Using TF Hub it is simple to retrain the top layer of the model to recognize the classes in our dataset.", "_____no_output_____" ], [ "### Dataset\n\n For this example you will use the TensorFlow flowers dataset:", "_____no_output_____" ] ], [ [ "data_root = tf.keras.utils.get_file(\n 'flower_photos','https://storage.googleapis.com/download.tensorflow.org/example_images/flower_photos.tgz',\n untar=True)", "_____no_output_____" ] ], [ [ "The simplest way to load this data into our model is using `tf.keras.preprocessing.image.ImageDataGenerator`,\n\nAll of TensorFlow Hub's image modules expect float inputs in the `[0, 1]` range. Use the `ImageDataGenerator`'s `rescale` parameter to achieve this.\n\nThe image size will be handled later.", "_____no_output_____" ] ], [ [ "image_generator = tf.keras.preprocessing.image.ImageDataGenerator(rescale=1/255)\nimage_data = image_generator.flow_from_directory(str(data_root), target_size=IMAGE_SHAPE)", "_____no_output_____" ] ], [ [ "The resulting object is an iterator that returns `image_batch, label_batch` pairs.", "_____no_output_____" ] ], [ [ "for image_batch, label_batch in image_data:\n print(\"Image batch shape: \", image_batch.shape)\n print(\"Label batch shape: \", label_batch.shape)\n break", "_____no_output_____" ] ], [ [ "### Run the classifier on a batch of images", "_____no_output_____" ], [ "Now run the classifier on the image batch.", "_____no_output_____" ] ], [ [ "result_batch = classifier.predict(image_batch)\nresult_batch.shape", "_____no_output_____" ], [ "predicted_class_names = imagenet_labels[np.argmax(result_batch, axis=-1)]\npredicted_class_names", "_____no_output_____" ] ], [ [ "Now check how these predictions line up with the images:", "_____no_output_____" ] ], [ [ "plt.figure(figsize=(10,9))\nplt.subplots_adjust(hspace=0.5)\nfor n in range(30):\n plt.subplot(6,5,n+1)\n plt.imshow(image_batch[n])\n plt.title(predicted_class_names[n])\n plt.axis('off')\n_ = plt.suptitle(\"ImageNet predictions\")", "_____no_output_____" ] ], [ [ "See the `LICENSE.txt` file for image attributions.\n\nThe results are far from perfect, but reasonable considering that these are not the classes the model was trained for (except \"daisy\").", "_____no_output_____" ], [ "### Download the headless model\n\nTensorFlow Hub also distributes models without the top classification layer. These can be used to easily do transfer learning.\n\nThe URL of any [TF2-compatible image feature vector module](https://tfhub.dev/s?module-type=image-feature-vector&q=tf2) from tfhub.dev will work here.", "_____no_output_____" ] ], [ [ "feature_extractor_url = \"https://tfhub.dev/google/tf2-preview/mobilenet_v2/feature_vector/2\" #@param {type:\"string\"}", "_____no_output_____" ] ], [ [ "Create the module, and check the expected image size:", "_____no_output_____" ] ], [ [ "feature_extractor_layer = hub.KerasLayer(feature_extractor_url,\n input_shape=(224,224,3))", "_____no_output_____" ] ], [ [ "The feature extractor returns a 1280-element vector for each image.", "_____no_output_____" ] ], [ [ "feature_batch = feature_extractor_layer(image_batch)\nprint(feature_batch.shape)", "_____no_output_____" ] ], [ [ "Freeze the variables in the feature extractor layer, so that the training only modifies the new classifier layer.", "_____no_output_____" ] ], [ [ "feature_extractor_layer.trainable = False", "_____no_output_____" ] ], [ [ "### Attach a classification head\n\nNow wrap the hub layer in a `tf.keras.Sequential` model, and add a new classification layer.", "_____no_output_____" ] ], [ [ "model = tf.keras.Sequential([\n feature_extractor_layer,\n layers.Dense(image_data.num_classes, activation='softmax')\n])\n\nmodel.summary()", "_____no_output_____" ], [ "predictions = model(image_batch)", "_____no_output_____" ], [ "predictions.shape", "_____no_output_____" ] ], [ [ "### Train the model\n\nUse compile to configure the training process:", "_____no_output_____" ] ], [ [ "model.compile(\n optimizer=tf.keras.optimizers.Adam(),\n loss='categorical_crossentropy',\n metrics=['acc'])", "_____no_output_____" ] ], [ [ "Now use the `.fit` method to train the model.\n\nTo keep this example short train just 2 epochs. To visualize the training progress, use a custom callback to log the loss and accuracy of each batch individually, instead of the epoch average.", "_____no_output_____" ] ], [ [ "class CollectBatchStats(tf.keras.callbacks.Callback):\n def __init__(self):\n self.batch_losses = []\n self.batch_acc = []\n\n def on_train_batch_end(self, batch, logs=None):\n self.batch_losses.append(logs['loss'])\n self.batch_acc.append(logs['acc'])\n self.model.reset_metrics()", "_____no_output_____" ], [ "steps_per_epoch = np.ceil(image_data.samples/image_data.batch_size)\n\nbatch_stats_callback = CollectBatchStats()\n\nhistory = model.fit(image_data, epochs=2,\n steps_per_epoch=steps_per_epoch,\n callbacks = [batch_stats_callback])", "_____no_output_____" ] ], [ [ "Now after, even just a few training iterations, we can already see that the model is making progress on the task.", "_____no_output_____" ] ], [ [ "plt.figure()\nplt.ylabel(\"Loss\")\nplt.xlabel(\"Training Steps\")\nplt.ylim([0,2])\nplt.plot(batch_stats_callback.batch_losses)", "_____no_output_____" ], [ "plt.figure()\nplt.ylabel(\"Accuracy\")\nplt.xlabel(\"Training Steps\")\nplt.ylim([0,1])\nplt.plot(batch_stats_callback.batch_acc)", "_____no_output_____" ] ], [ [ "### Check the predictions\n\nTo redo the plot from before, first get the ordered list of class names:", "_____no_output_____" ] ], [ [ "class_names = sorted(image_data.class_indices.items(), key=lambda pair:pair[1])\nclass_names = np.array([key.title() for key, value in class_names])\nclass_names", "_____no_output_____" ] ], [ [ "Run the image batch through the model and convert the indices to class names.", "_____no_output_____" ] ], [ [ "predicted_batch = model.predict(image_batch)\npredicted_id = np.argmax(predicted_batch, axis=-1)\npredicted_label_batch = class_names[predicted_id]", "_____no_output_____" ] ], [ [ "Plot the result", "_____no_output_____" ] ], [ [ "label_id = np.argmax(label_batch, axis=-1)", "_____no_output_____" ], [ "plt.figure(figsize=(10,9))\nplt.subplots_adjust(hspace=0.5)\nfor n in range(30):\n plt.subplot(6,5,n+1)\n plt.imshow(image_batch[n])\n color = \"green\" if predicted_id[n] == label_id[n] else \"red\"\n plt.title(predicted_label_batch[n].title(), color=color)\n plt.axis('off')\n_ = plt.suptitle(\"Model predictions (green: correct, red: incorrect)\")", "_____no_output_____" ] ], [ [ "## Export your model\n\nNow that you've trained the model, export it as a saved model:", "_____no_output_____" ] ], [ [ "import time\nt = time.time()\n\nexport_path = \"/tmp/saved_models/{}\".format(int(t))\ntf.keras.experimental.export_saved_model(model, export_path)\n\nexport_path", "_____no_output_____" ] ], [ [ "Now confirm that we can reload it, and it still gives the same results:", "_____no_output_____" ] ], [ [ "reloaded = tf.keras.experimental.load_from_saved_model(export_path, custom_objects={'KerasLayer':hub.KerasLayer})", "_____no_output_____" ], [ "result_batch = model.predict(image_batch)\nreloaded_result_batch = reloaded.predict(image_batch)", "_____no_output_____" ], [ "abs(reloaded_result_batch - result_batch).max()", "_____no_output_____" ] ], [ [ "This saved model can be loaded for inference later, or converted to [TFLite](https://www.tensorflow.org/lite/convert/) or [TFjs](https://github.com/tensorflow/tfjs-converter).\n", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown" ]
[ [ "markdown" ], [ "code" ], [ "markdown", "markdown", "markdown", "markdown", "markdown" ], [ "code", "code", "code" ], [ "markdown", "markdown" ], [ "code", "code" ], [ "markdown", "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown" ] ]
ece9c21c5d681500710961bc7900894bfe61761a
9,325
ipynb
Jupyter Notebook
tutorials/using-dask/1.intro-to-dask.ipynb
kseager/azureml-examples
02e575cc19a1dec55e42fbb00352dc06b54eb8e7
[ "MIT" ]
1
2021-04-28T11:26:59.000Z
2021-04-28T11:26:59.000Z
tutorials/using-dask/1.intro-to-dask.ipynb
kseager/azureml-examples
02e575cc19a1dec55e42fbb00352dc06b54eb8e7
[ "MIT" ]
null
null
null
tutorials/using-dask/1.intro-to-dask.ipynb
kseager/azureml-examples
02e575cc19a1dec55e42fbb00352dc06b54eb8e7
[ "MIT" ]
null
null
null
22.96798
377
0.544129
[ [ [ "# Introduction to Dask\n\nIn this notebook, we'll learn how to use [Dask](https://dask.org) for reading and processing data from Azure.", "_____no_output_____" ], [ "## Install required packages", "_____no_output_____" ] ], [ [ "!pip install --upgrade dask distributed dask-sql bokeh adlfs fsspec fastparquet pyarrow python-snappy lz4 \"pandas>=1.2.0\"", "_____no_output_____" ] ], [ [ "## Get AML Workspace\n\nYou can use the AML workspace to retrieve datastores and keyvaults for accessing data credentials securely.", "_____no_output_____" ] ], [ [ "from azureml.core import Workspace\n\nws = Workspace.from_config()\nws", "_____no_output_____" ] ], [ [ "## Create a distributed client\n\nThe [client](https://distributed.dask.org/en/latest/client.html) is the primary entrypoint for parallel processing with Dask. Calling it without inputs will create a local distributed scheduler, utilizing all the CPUs and cores on your machine. This can be useful for faster processing of larger in memory dataframes, or even computations on out of memory (OOM) data. \n\nWhen your local machine isn't powerful enough, you can provision a larger VM in Azure - the M series has 100+ CPUs and TBs of RAM. If this still isn't powerful enough, you can create a distributed Dask cluster on most hardware - see [the Dask setup guide](https://docs.dask.org/en/latest/setup.html) for details.\n\nIf you still need acceleration, [RAPIDSAI](https://github.com/rapidsai) further extends the PyData APIs on GPUs.\n\n**Make sure you check out the dashboard!**", "_____no_output_____" ] ], [ [ "from distributed import Client\n\n# initialize local client\nc = Client()\n\n# print Python objects\nprint(c)\nprint(c.dashboard_link)\n\n# print notebook widget widget\nc", "_____no_output_____" ] ], [ [ "## Reading cloud data\n\nReading data from the cloud is as easy! Python implements various cloud protocols, including ``az`` for Blob and ADLSv2 and ``adl`` for ADLSv1.\n\n### Public Data\n\nPublic data can simply be read via ``https``.\n", "_____no_output_____" ] ], [ [ "account_name = \"azuremlexamples\"\ncontainer_name = \"datasets\"\n\nstorage_options = {\"account_name\": account_name}", "_____no_output_____" ], [ "data_uri = f\"https://{account_name}.blob.core.windows.net/{container_name}/iris.csv\"\ndata_uri", "_____no_output_____" ], [ "import pandas as pd\n\ndf = pd.read_csv(data_uri)\ndf", "_____no_output_____" ] ], [ [ "alternatively, we can use the ``az`` protocol and pass in ``storage_options``:", "_____no_output_____" ] ], [ [ "data_uri = f\"az://{container_name}/iris.csv\"\ndata_uri", "_____no_output_____" ], [ "import pandas as pd\n\ndf = pd.read_csv(data_uri, storage_options=storage_options)\ndf", "_____no_output_____" ] ], [ [ "## Private Data \n\nPassing in storage options allows for reading private data. For instance, you can easily retrieve the information from an Azure ML Datastore:\n\n```python\nfrom azureml.core import Workspace\n\nws = Workspace.from_config()\nds = ws.get_default_datastore() # ws.datastores[\"my-datastore-name\"]\n\ncontainer_name = ds.container_name\nstorage_options = {\n \"account_name\": ds.account_name,\n \"account_key\": ds.account_key,\n}\n```", "_____no_output_____" ] ], [ [ "from adlfs import AzureBlobFileSystem\n\nds = ws.get_default_datastore()\n\ncontainer_name = ds.container_name\nstorage_options = {\"account_name\": ds.account_name, \"account_key\": ds.account_key}\n\nfs = AzureBlobFileSystem(**storage_options)\nfs", "_____no_output_____" ], [ "fs.ls(f\"{container_name}\")", "_____no_output_____" ] ], [ [ "## Pythonic Filesystem\n\nIn the previous section, we used [ADLFS](https://github.com/dask/adlfs) to initialize a Pythonic filesystem and perform operations.\n\nThe below cell demonstrate some basic operations to raed and manipulate data in Python.", "_____no_output_____" ] ], [ [ "color = \"green\"\ncontainer_name = \"nyctlc\"\nstorage_options = {\"account_name\": \"azureopendatastorage\"}", "_____no_output_____" ], [ "fs = AzureBlobFileSystem(**storage_options)\nfs", "_____no_output_____" ], [ "fs.ls(f\"{container_name}\")", "_____no_output_____" ], [ "fs.ls(f\"{container_name}/{color}\")", "_____no_output_____" ], [ "fs.ls(f\"{container_name}/{color}/puYear=2016/\")", "_____no_output_____" ], [ "files = fs.glob(f\"{container_name}/{color}/puYear=2016/puMonth=12/*.parquet\")\nfiles = [f\"az://{file}\" for file in files]\nfiles[-5:]", "_____no_output_____" ], [ "import dask.dataframe as dd\n\nddf = (\n dd.read_parquet(files, storage_options=storage_options)\n .repartition(npartitions=8)\n .persist()\n)\nddf", "_____no_output_____" ], [ "%%time\nlen(ddf)", "_____no_output_____" ], [ "%%time\nlen(ddf)", "_____no_output_____" ], [ "ddf.info()", "_____no_output_____" ], [ "import matplotlib.pyplot as plt\n\nplt.style.use(\"dark_background\")\n\nddf[\"tipAmount\"].compute().hist(\n figsize=(16, 8),\n bins=256,\n range=(0.1, 20),\n)", "_____no_output_____" ], [ "df = ddf.compute()\ndf.info()", "_____no_output_____" ], [ "%%time\ndf.describe()", "_____no_output_____" ], [ "%%time\ngbs = round(df.memory_usage(index=True, deep=True).sum() / 1e9, 2)\nprint(f\"df is {gbs} GBs\")", "_____no_output_____" ], [ "%%time\ngbs = round(ddf.memory_usage(index=True, deep=True).sum().compute() / 1e9, 2)\nprint(f\"ddf is {gbs} GBs\")", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code" ]
[ [ "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code" ] ]
ece9cd3dadb72555ca08fc050f9335b79ff12424
516,388
ipynb
Jupyter Notebook
beginner-python/002-plots.ipynb
OHBM-Seattle/brain-hacking-101
6dabcd1018e50dac6dcba04831d4314920b1d225
[ "Apache-2.0" ]
1
2018-08-24T02:57:05.000Z
2018-08-24T02:57:05.000Z
beginner-python/002-plots.ipynb
sinodanish/brain-hacking-101
6dabcd1018e50dac6dcba04831d4314920b1d225
[ "Apache-2.0" ]
null
null
null
beginner-python/002-plots.ipynb
sinodanish/brain-hacking-101
6dabcd1018e50dac6dcba04831d4314920b1d225
[ "Apache-2.0" ]
null
null
null
1,426.486188
159,988
0.94949
[ [ [ "### Brain-hacking 101\n\nAuthor: [**Ariel Rokem**](http://arokem.org), [**The University of Washington eScience Institute**](http://escience.washington.edu)", "_____no_output_____" ], [ "### Hack 2: Look at your data\n\nA picture is worth a thousand words. Data visualization allows you to look directly at different aspects of the data that are not readily available to you by just looking at the numbers. \n\nIn this tutorial, we will look at the FIAC data using the [Matplotlib](http://matplotlib.org) software library. Matplotlib is an open-source software library that can be used to create beautiful 2-d data visualizations, such as lines, scatter-plots, and images. It can be used to produce publication-quality figures in a variety of file-formats.\n\nLet's ", "_____no_output_____" ] ], [ [ "import numpy as np\nimport nibabel as nib", "_____no_output_____" ], [ "import matplotlib as mpl\nimport matplotlib.pyplot as plt", "_____no_output_____" ], [ "%matplotlib inline", "_____no_output_____" ], [ "img = nib.load('./data/run1.nii.gz')\ndata = img.get_data()", "_____no_output_____" ], [ "fig, ax = plt.subplots(1)\nax.plot(data[32, 32, 15, :])", "_____no_output_____" ] ], [ [ "Congratulations! You first MPL plot. Let's make this a little bit larger, use a style to make it look better, and add some annotations. ", "_____no_output_____" ] ], [ [ "mpl.style.use('bmh')\nfig, ax = plt.subplots(1)\nax.plot(data[32, 32, 15, :])\nax.set_xlabel('Time (TR)')\nax.set_ylabel('MRI signal (a.u.)')\nax.set_title('Time-series from voxel [32, 32, 15]')\nfig.set_size_inches([12, 6])", "_____no_output_____" ] ], [ [ "## Impressions about the data?", "_____no_output_____" ], [ "If we want to compare several voxels side by side we can plot them on the same axis:", "_____no_output_____" ] ], [ [ "fig, ax = plt.subplots(1)\nax.plot(data[32, 32, 15, :])\nax.plot(data[32, 32, 14, :])\nax.plot(data[32, 32, 13, :])\nax.plot(data[32, 32, 12, :])\nax.set_xlabel('Time (TR)')\nax.set_ylabel('MRI signal (a.u.)')\nax.set_title('Time-series from a few voxels')\nfig.set_size_inches([12, 6])", "_____no_output_____" ] ], [ [ "Alternatively, we can create different subplots for each time-series", "_____no_output_____" ] ], [ [ "fig, ax = plt.subplots(2, 2)\n# ax is now an array!\nax[0, 0].plot(data[32, 32, 15, :])\nax[0, 1].plot(data[32, 32, 14, :])\nax[1, 0].plot(data[32, 32, 13, :])\nax[1, 1].plot(data[32, 32, 12, :])\nax[1, 0].set_xlabel('Time (TR)')\nax[1, 1].set_xlabel('Time (TR)')\nax[0, 0].set_ylabel('MRI signal (a.u.)')\nax[1, 0].set_ylabel('MRI signal (a.u.)')\n# Note that we now set the title through the fig object!\nfig.suptitle('Time-series from a few voxels')\nfig.set_size_inches([12, 6])", "_____no_output_____" ] ], [ [ "Another kind of plot is an image. For example, we can take a look at the mean and standard deviation of the time-series for one entire slice:", "_____no_output_____" ] ], [ [ "fig, ax = plt.subplots(1, 2)\n# We'll use a reasonable colormap, and no smoothing:\nax[0].matshow(np.mean(data[:, :, 15], -1), cmap=mpl.cm.hot)\nax[0].axis('off')\nax[1].matshow(np.std(data[:, :, 15], -1), cmap=mpl.cm.hot)\nax[1].axis('off')\nfig.set_size_inches([12, 6])\n# You can save the figure to file:\nfig.savefig('mean_and_std.png')", "_____no_output_____" ] ], [ [ "There are many other kinds of figures you could create:", "_____no_output_____" ] ], [ [ "fig, ax = plt.subplots(2, 2)\n# Note the use of `ravel` to create a 1D array:\nax[0, 0].hist(np.ravel(data))\nax[0, 0].set_xlabel(\"fMRI signal\")\nax[0, 0].set_ylabel(\"# voxels\")\n\n# Bars are 0.8 wide:\nax[0, 1].bar([0.6, 1.6, 2.6, 3.6], [np.mean(data[:, :, 15]), np.mean(data[:, :, 14]), np.mean(data[:, :, 13]), np.mean(data[:, :, 12])])\nax[0, 1].set_ylabel(\"Average signal in the slice\")\nax[0, 1].set_xticks([1,2,3,4])\nax[0, 1].set_xticklabels([\"15\", \"14\", \"13\", \"12\"])\nax[0, 1].set_xlabel(\"Slice #\")\n\n# Compares subsequent time-points:\nax[1, 0].scatter(data[:, :, 15, 0], data[:, :, 15, 1])\nax[1, 0].set_xlabel(\"fMRI signal (time-point 0)\")\nax[1, 0].set_ylabel(\"fMRI signal (time-point 1)\")\n\n# `.T` denotes a transposition\nax[1, 1].boxplot(data[32, 32].T)\nfig.set_size_inches([12, 12])\nax[1, 1].set_xlabel(\"Position\")\nax[1, 1].set_ylabel(\"fMRI signal\")", "_____no_output_____" ] ], [ [ "One way to learn more about all the different options available in Matplotlib is to visit the [gallery](http://matplotlib.org/gallery.html)", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown" ]
[ [ "markdown", "markdown" ], [ "code", "code", "code", "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ] ]
ece9d2907cabd70bd6acc447f3a2ad9078a30a50
45,173
ipynb
Jupyter Notebook
notebooks/seti_cnn_tf_training.ipynb
IBM/powerai_seti_signal_classification
e40b6064fd24ce146ef731bf9a0c0e76263fa61c
[ "Apache-2.0" ]
16
2018-12-04T13:52:31.000Z
2022-02-05T22:57:06.000Z
notebooks/seti_cnn_tf_training.ipynb
IBM/powerai_seti_signal_classification
e40b6064fd24ce146ef731bf9a0c0e76263fa61c
[ "Apache-2.0" ]
1
2018-12-03T22:42:55.000Z
2018-12-05T22:37:32.000Z
notebooks/seti_cnn_tf_training.ipynb
IBM/powerai_seti_signal_classification
e40b6064fd24ce146ef731bf9a0c0e76263fa61c
[ "Apache-2.0" ]
10
2018-12-02T03:54:12.000Z
2020-09-01T06:41:54.000Z
34.668457
775
0.596418
[ [ [ "<a href=\"https://www.bigdatauniversity.com\"><img src = \"https://ibm.box.com/shared/static/cw2c7r3o20w9zn8gkecaeyjhgw3xdgbj.png\" width = 200, align = \"left\"></a>\n<br>\n<br>\n\n--------------------\n# Search for Extra Terrestrial Intelligence (SETI)\n### SETI Signal Classification on PowerAI with Single GPU\n<hr>\n<br>", "_____no_output_____" ], [ "### Introduction\nIn this notebook, we will use the famous [SETI Dataset](https://github.com/setiQuest/ML4SETI/) to build a convolutional neural network (CNN) able to perform signal classification. The CNN will determine, with some associated error, what type of signal is presented.\n\n### Project overview\nEach night, using the Allen Telescope Array (ATA) in northern California, the SETI Institute scans the sky at various radio frequencies, observing star systems with known exoplanets, searching for faint but persistent signals. The current signal detection system is programmed to search only for particular kinds of signals: narrow-band carrier waves. However, the detection system sometimes triggers on signals that are not narrow-band signals (with unknown efficiency) and are also not explicitly-known radio frequency interference (RFI). Various categories of these kinds of events have been observed in the past.\n\nOur goal is to classify these accurately in real-time. This may allow the signal detection system to make better observational decisions, increase the efficiency of the nightly scans, and allow for explicit detection of these other signal types.\n\nFor more information refer to [SETI hackathon page](https://github.com/setiQuest/ML4SETI/).\n\n### Performance\nConvolutional neural networks involve a lot of matrix and vector multiplications that can be parallelized. GPUs can improve performance, because GPUs were designed to handle these operations in parallel!\n\n### GPU vs. CPU\nA single core CPU takes a matrix operation in serial, one element at a time, but a single GPU could have hundreds or thousands of cores, while a CPU typically has no more than a few cores.\n\n### How to use GPUs with TensorFlow?\nIt is important to notice that if both CPU and GPU are available on the machine that you are running the notebook, and if a TensorFlow operation has both CPU and GPU implementations, the GPU devices will be given priority when the operation is assigned to a device.\n\nIn our case, as we are running this notebook on [IBM PowerAI](http://cocl.us/SETI-NIMBIX-PowerAI), you may have access to multiple GPUs, but lets use one of the GPUs in this notebook, for the sake of simplicity.\n\n> Note: If you are running the **free trial**, you would expect to have zero GPUs. This notebook will work, but the training will be slow.", "_____no_output_____" ] ], [ [ "import sys\nsys.path.insert(0, \"/opt/DL/tensorflow/lib/python2.7/site-packages/\")\nimport requests\nimport json\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport tensorflow as tf\nimport pickle\nimport time\nimport os\nfrom sklearn.metrics import confusion_matrix\nfrom sklearn import metrics\nfrom six.moves import urllib\nimport tarfile\nimport numpy as np\nfrom PIL import Image\nimport math\nimport datetime\n%matplotlib inline", "_____no_output_____" ] ], [ [ "### Set the parameters", "_____no_output_____" ] ], [ [ "### Set your working space here. Use this folder to save intermediate results.\ndataset_name = 'SETI_ds_64x128'\ndata_dir = \"tmp/SETI1_data/\"\ntrain_dir = 'tmp/SETI1_train/'\nlog_dir = train_dir + '1GPU_4'\n# check point directory\nchk_directory = train_dir + '/save1/'\ncheckpoint_path = chk_directory + 'model.ckpt'", "_____no_output_____" ] ], [ [ "### Create the necessary folders", "_____no_output_____" ] ], [ [ "if os.path.exists(data_dir) is False:\n os.makedirs(data_dir)\nprint data_dir\nprint os.popen(\"ls -lrt \"+ data_dir).read()\nif os.path.exists(train_dir) is False:\n os.makedirs(train_dir)\nprint train_dir", "_____no_output_____" ] ], [ [ "### Import the dataset reader\nThe signals have been converted into spectogram images and stored as 4 files.\nThe following cell will download and import python code to help us decode the binary files, and read the SETI dataset.", "_____no_output_____" ] ], [ [ "!wget -q --output-document SETI.zip https://ibm.box.com/shared/static/jhqdhcblhua5dx2t7ixwm88okitjrl6l.zip\n!unzip -o SETI.zip\nimport SETI", "_____no_output_____" ] ], [ [ "### Download data\nThe following cell downloads and extracts the dataset.", "_____no_output_____" ] ], [ [ "def maybe_download_and_extract():\n DATA_URL = 'https://ibm.box.com/shared/static/qz33lcio9ip2j8qi2atxqs62gn3bnu2s.gz'\n dest_directory = data_dir\n if not os.path.exists(dest_directory):\n os.makedirs(dest_directory)\n filename = DATA_URL.split('/')[-1]\n filepath = os.path.join(dest_directory, filename)\n if not os.path.exists(filepath):\n # print 'No zip file exists in ', filepath\n def _progress(count, block_size, total_size):\n sys.stdout.write('\\r>> Downloading %s %.1f%%' % (filename, float(count * block_size) / float(total_size) * 100.0))\n sys.stdout.flush()\n filepath, _ = urllib.request.urlretrieve(DATA_URL, filepath, _progress)\n\n statinfo = os.stat(filepath)\n print('Successfully downloaded', filename, statinfo.st_size, 'bytes.')\n extracted_dir_path = os.path.join(dest_directory, dataset_name)\n if not os.path.exists(extracted_dir_path):\n print 'Extracting to', extracted_dir_path\n tarfile.open(filepath, 'r:gz').extractall(dest_directory)\nmaybe_download_and_extract()", "_____no_output_____" ] ], [ [ "### Load data SETI", "_____no_output_____" ] ], [ [ "ds_directory = data_dir + dataset_name\ndataset = SETI.read_data_sets(ds_directory, one_hot=True, validation_size=0)\ndataset.train.images.shape", "_____no_output_____" ] ], [ [ "### Understanding the imported data\n\nThe imported data can be divided as follows:\n\n- Training (dataset.train) >> Use the given dataset with inputs and related outputs for training of NN. In our case, if you give an image that you know represents \"class1\", this set will tell the neural network that we expect \"class1\" as the output. \n - 694 signals (images)\n - dataset.train.images for inputs\n - dataset.train.labels for outputs\n \n \n- Test (mnist.test) >> The model does not have access to this information prior to the test phase. It is used to evaluate the performance and accuracy of the model against \"real life situations\". No further optimization beyond this point. \n - 10,000 data points\n - dataset.test.images for inputs\n - dataset.test.labels for outputs\n \n \n#### Labels\n- Each image (spectrum of signal) in the dataset has been labeled from 1 to 4, representing:\n - squiggle\n - narrowband\n - noise\n - narrowbanddrd", "_____no_output_____" ], [ "## Network parameters", "_____no_output_____" ] ], [ [ "# Parameters\ndecay_rate=0.96\ndecay_steps=500\nlearning_rate = 0.005\ntraining_epochs = 50\nbatch_size = 128\ndisplay_step = 1\n\nn_classes = 4 # number of possible classifications for the problem\ndropout = 0.60 # Dropout, probability to keep units\n\nheight = 64 # height of the image in pixels \nwidth = 128 # width of the image in pixels \nn_input = width * height # number of pixels in one image \n", "_____no_output_____" ] ], [ [ "### Inputs\n\nIt's a best practice to create placeholders before variable assignments when using TensorFlow. Here we'll create placeholders for inputs (\"x\") and outputs (\"y_\"). \n\n__Placeholder 'x':__ represents the \"space\" allocated input or the images. \n * Each input has 8192 pixels distributed by a 64 width x 128 height matrix.\n * The 'shape' argument defines the tensor size by its dimensions.\n * 1st dimension = None. Indicates that the batch size can be of any size.\n * 2nd dimension = 8192. Indicates the number of pixels on a single flattened spectogram image.\n \n__Placeholder 'y_':__ represents the final output or the labels.\n * 4 possible classes (0,1,2,3).\n * The 'shape' argument defines the tensor size by its dimensions.\n * 1st dimension = None. Indicates that the batch size can be of any size.\n * 2nd dimension = 4. Indicates the number of targets/outcomes.\n\n__dtype for both placeholders:__ If you are not sure, use tf.float32. The limitation here is that the later presented softmax function only accepts float32 or float64 dtypes. For more dtypes, check TensorFlow's documentation <a href=\"https://www.tensorflow.org/versions/r0.9/api_docs/python/framework.html#tensor-types\">here</a>", "_____no_output_____" ] ], [ [ "x = tf.placeholder(tf.float32, shape=[None, n_input], name = 'x')\ny_ = tf.placeholder(tf.float32, shape=[None, n_classes], name = 'y_')", "_____no_output_____" ] ], [ [ "The input image is a 64 pixels by 128 pixels, 1 channel (grayscale). In this case, the first dimension is the __batch number__ of the image and can be of any size (so we set it to -1). The second and third dimensions are width and height. The last one is the image channels.", "_____no_output_____" ] ], [ [ "x_image = tf.reshape(x, [-1,height,width,1]) ", "_____no_output_____" ] ], [ [ "### Convolutional neural networks (CNNs)\n\nConvolutional neural networks (CNNs) are a type of feed-forward neural network, consisting of multiple layers of neurons that have learnable weights and biases. Each neuron in a layer that receives some input, processes it, and optionally follows it with non-linearity. The network has multiple layers such as convolution, max pool, drop out and fully connected layers. In each layer, small neurons process portions of the input image. The outputs of these collections are then tiled, so that their input regions overlap, to obtain a higher-resolution representation of the original image. This is repeated for every such layer. The important point here is: CNNs are able to break the complex patterns down into a series of simpler patterns, through multiple layers.\n\n### CNN architecture\nNow we are going to use a Deep Neural Network. \n\nArchitecture of our network is:\n \n- [Input] >>>> (batch_size, 64, 128, 1) \n- [Convolutional layer 1] >>>> (batch_size, 64, 128, 32)\n- [ReLU 1] >>>> (batch_size, 64, 128, 32)\n- [Max pooling 1] >>>> (batch_size, 32, 64, 32)\n- [Convolutional layer 2] >>>> (batch_size, 32, 64, 64)\n- [ReLU 2] >>>> (batch_size, 32, 64, 64)\n- [Max pooling 2] >>>> (batch_size, 8, 16, 64) \n- [fully connected layer 1] >>>> (batch_size, 1024)\n- [ReLU] >>>> (batch_size, 1024)\n- [Drop out] >>>> (batch_size, 1024)\n- [Readout layer] >>>> (batch_size, 1024)\n", "_____no_output_____" ], [ "#### Convolutional Layer 1\nWe define a kernel here. The size of the filter/kernel is 5x5; Input channels is 1 (greyscale); and we need 32 different feature maps (here, 32 feature maps means 32 different filters are applied on each image. So, the output of the convolution layer would be 64x128x32). In this step, we create a filter / kernel tensor of shape `[filter_height, filter_width, in_channels, out_channels]`.\n\n- To create a convolutional layer, we use __tf.nn.conv2d__. It computes a 2-D convolution given 4-D input and filter tensors.\n- The convolutional layer slides the \"kernel window\" across the input tensor. As the input tensor has 4 dimensions: [batch, height, width, channels], then the convolution operates on a 2D window on the height and width dimensions. __strides__ determines how much the window shifts in each of the dimensions. As the first and last dimensions are related to batch and channels, we set the stride to 1. But for second and third dimension, we could set other values, e.g. [1, 2, 2, 1]\n- __max pooling__ is a form of non-linear down-sampling. It partitions the input image into a set of rectangles and, and then finds the maximum value for that region. Let's use __tf.nn.max_pool__ function to perform max pooling. ", "_____no_output_____" ] ], [ [ "with tf.variable_scope('layer1') as scope:\n W_conv1 = tf.get_variable( initializer= tf.truncated_normal([5, 5, 1, 32], stddev=0.1), name= 'w1')\n b_conv1 = tf.get_variable(initializer=tf.constant(0.1, shape=[32]), name='b1') # need 32 biases for 32 outputs\n convolve1 = tf.nn.conv2d(x_image, W_conv1, strides=[1, 1, 1, 1], padding='SAME', name='convolv1') + b_conv1\n h_conv1 = tf.nn.relu(convolve1, name='relu1')\n conv1 = tf.nn.max_pool(h_conv1, ksize=[1, 2, 2, 1], strides=[1, 2, 2, 1], padding='SAME', name='pool1') # max_pool_2x2", "_____no_output_____" ] ], [ [ "#### Convolutional Layer 2\nWe apply the convolution again in this layer. Let's look at the second layer kernel: \n- Filter/kernel: 5x5x32\n- Input channels: 32 (from the 1st Conv layer, we had 32 feature maps), so we use 64 output feature maps ", "_____no_output_____" ] ], [ [ "with tf.variable_scope('layer2') as scope:\n W_conv2 = tf.get_variable(initializer=tf.truncated_normal([5, 5, 32, 64], stddev=0.1), name= 'w2')\n b_conv2 = tf.get_variable(initializer=tf.constant(0.1, shape=[64]), name= 'b2') # need 64 biases for 64 outputs\n convolve2= tf.nn.conv2d(conv1, W_conv2, strides=[1, 1, 1, 1], padding='SAME', name='convolv2')+ b_conv2\n h_conv2 = tf.nn.relu(convolve2, name='relu2')\n conv2 = tf.nn.max_pool(h_conv2, ksize=[1, 2, 2, 1], strides=[1, 4, 4, 1], padding='SAME', name='pool2') # max_pool_2x2", "_____no_output_____" ] ], [ [ "#### Fully Connected Layer\n\nYou need a fully connected layer to use the Softmax and create the probabilities in the end. Fully connected layers take the high-level filtered images from the previous layer, that is all 64 metrics, and converts them to a flat array.\n\nSo, the matrix (8x16)x64 will be converted to a matrix of size (1x1024).", "_____no_output_____" ] ], [ [ "dim = conv2.get_shape().as_list()\ndims = dim[1]*dim[2]*dim[3]\nnodes1 = 1024\nwith tf.variable_scope('FullyCon1') as scope:\n prv_layer_matrix = tf.reshape(conv2, [-1, dims], name='rshaped_inp')\n W_fc1 = tf.get_variable(initializer= tf.truncated_normal([dims, nodes1], stddev=0.1), name='W_FC1')\n b_fc1 = tf.get_variable(initializer= tf.constant(0.1, shape=[nodes1]), name='b_FC1') # need 1024 biases for 1024 outputs\n h_fcl1 = tf.matmul(prv_layer_matrix, W_fc1) + b_fc1\n fc_layer1 = tf.nn.relu(h_fcl1, name='relu_FC1') ", "_____no_output_____" ] ], [ [ "#### Droupout 1\nThis is a phase where the network \"forgets\" some features. At each training step in a mini-batch, some units get switched off randomly so that they will not interact with the network. That is, its weights cannot be updated, nor affect the learning of the other network nodes. This can be very useful for very large neural networks to prevent overfitting.", "_____no_output_____" ] ], [ [ "keep_prob = tf.placeholder(tf.float32, name= 'keep_prob')\nlayer_drop1 = tf.nn.dropout(fc_layer1, keep_prob, name='dropout')", "_____no_output_____" ] ], [ [ "#### Readout Layer\nIn the last layer, CNN takes the high-level filtered images and translates them into votes using softmax. __softmax__ allows us to interpret the outputs of __fcl4__ as probabilities. So, __y_conv__ is a tensor of probabilities.\nInput channels in this layer is 1024 (neurons from the 3rd layer) pixels, and output features are 4 classes. ", "_____no_output_____" ] ], [ [ "with tf.variable_scope('ReadoutLayer') as scope:\n W_fc = tf.get_variable(initializer = tf.truncated_normal([nodes1, n_classes], stddev=0.1), name ='W_FC2') # 1024 neurons\n b_fc = tf.get_variable(initializer =tf.constant(0.1, shape=[n_classes]), name = 'b_FC2') # 10 possibilities for classes [0,1,2,3]\n fc = tf.matmul(layer_drop1, W_fc) + b_fc\n y_CNN = tf.nn.softmax(fc, name = 'softmax_linear')", "_____no_output_____" ] ], [ [ "#### Loss function\nWe need to compare our output, layer4 tensor, with ground truth for all mini_batch. We can use __cross entropy__ to see how bad our CNN is working - to measure the error at a softmax layer. __softmax_cross_entropy_with_logits__ \ncomputes softmax cross entropy between logits and labels, and __reduce_mean__ computes the mean of all elements in the tensor.", "_____no_output_____" ] ], [ [ "with tf.name_scope('cross_entropy'):\n cross_entropy = tf.reduce_mean(tf.nn.softmax_cross_entropy_with_logits(logits=y_CNN, labels=y_))\n tf.summary.scalar('cross_entropy', cross_entropy)", "_____no_output_____" ] ], [ [ "#### Training\nIt is obvious that we want to minimize the error of our network which is calculated by the cross_entropy metric. To solve the problem, we have to compute gradients for the loss (which is minimizing the cross-entropy) and apply gradients to variables. This will be done by an optimizer: GradientDescent. ", "_____no_output_____" ] ], [ [ "# Create a variable to track the global step.\nglobal_step = tf.Variable(0, trainable=False)\n\n# create learning_decay\nlr = tf.train.exponential_decay( learning_rate,\n global_step,\n decay_steps,\n decay_rate, staircase=True )\ntf.summary.scalar('learning_rate', lr)", "_____no_output_____" ], [ "# Use the optimizer to apply the gradients that minimize the loss\n# (and also increment the global step counter) as a single training step.\n\nwith tf.name_scope('train'):\n optimizer = tf.train.GradientDescentOptimizer(lr)\n train_op = optimizer.minimize(cross_entropy, global_step=global_step)", "_____no_output_____" ] ], [ [ "#### Evaluation\nDo you want to know how many of the cases in a mini-batch have been classified correctly? Let's count them.", "_____no_output_____" ] ], [ [ "with tf.name_scope('accuracy_all'):\n pred_lb = tf.argmax(y_CNN,1)\n true_lb = tf.argmax(y_,1)\n correct_prediction = tf.equal(pred_lb,true_lb )\n accuracy = tf.reduce_mean(tf.cast(correct_prediction, tf.float32))\n tf.summary.scalar('accuracy', accuracy)", "_____no_output_____" ], [ "# Calculate accuracy for test images\ndef evaluate():\n with tf.variable_scope(tf.get_variable_scope()):\n num_iter = int(math.ceil(dataset.test.num_examples / batch_size))\n true_count = 0 # Counts the number of correct predictions.\n total_sample_count = num_iter * batch_size\n step = 0\n while step < num_iter:\n x_batch, y_batch = dataset.test.next_batch(batch_size)\n predictions = sess.run([correct_prediction], feed_dict={x: x_batch, y_: y_batch, keep_prob: 1.})\n true_count += np.sum(predictions)\n step += 1\n\n precision = true_count*1.0 / total_sample_count\n tf.summary.scalar('precision', precision)\n return precision", "_____no_output_____" ] ], [ [ "### Create checkpoint directory", "_____no_output_____" ] ], [ [ "directory = os.path.dirname(chk_directory)\ntry:\n os.stat(directory)\n ckpt = tf.train.get_checkpoint_state(chk_directory)\n print ckpt\nexcept:\n os.mkdir(directory) ", "_____no_output_____" ] ], [ [ "<div class=\"alert alert-success alertsuccess\" style=\"margin-top: 20px\">\n<font size = 3><strong>*You can run this cell if you REALLY have time to wait, or you are running it using PowerAI </strong></font>\n<br>\n\n<b> What is PowerAI? </b>\n\nRunning deep learning programs usually needs a high performance platform. PowerAI speeds up deep learning and AI. Built on IBM's Power Systems, PowerAI is a scalable software platform that accelerates deep learning and AI with blazing performance for individual users or enterprises. The PowerAI platform supports popular machine learning libraries and dependencies including Tensorflow, Caffe, Torch, and Theano. You can download a [free version of PowerAI](https://cocl.us/DX0108EN-PowerAI).\n\n<br>\n__Notice:__ If you are running this notebook on PowerAI, it will automatically run on a GPU, otherwise it will use a CPU. Also, you can change the number of epochs to gain a higher accuracy.\n</div>", "_____no_output_____" ], [ "## Training", "_____no_output_____" ] ], [ [ "num_examples = dataset.train.num_examples\ntotal_batch = int(num_examples / batch_size)\nprint 'Training dataset size:',num_examples, ' Signals'\nprint 'Testing dataset size:',dataset.test.num_examples, ' Signals'\nprint 'Total epochs:', training_epochs\nprint 'Total steps:', training_epochs * total_batch\nprint 'Batch size:', batch_size\nprint 'Total batchs per epoch:',total_batch\nprint str(datetime.datetime.now())\n# Initializing the variables\ninit = tf.global_variables_initializer()\nloss_values = []\nsess = tf.Session(config=tf.ConfigProto(allow_soft_placement=True))\nsess.run(init)\n\nmerged = tf.summary.merge_all()\ntrain_writer = tf.summary.FileWriter(log_dir, sess.graph)\n \nX_test = dataset.test.images\ny_test = dataset.test.labels\nsess.run(init)\nsaver = tf.train.Saver(tf.global_variables())\n\n# load previously trained model if appilcable\nckpt = tf.train.get_checkpoint_state(chk_directory)\nif ckpt:\n print (\"loading model: \",ckpt.model_checkpoint_path)\n # saver.restore(sess, ckpt.model_checkpoint_path)\n\n# Training cycle\ntr_start = time.time()\nfor epoch in range(training_epochs):\n avg_loss = 0.\n avg_train_acc = 0.\n total_batch = int(num_examples / batch_size)\n\n # Loop over all batches\n start = time.time()\n for step in range(total_batch):\n x_batch, y_batch = dataset.train.next_batch(batch_size ,shuffle=True)\n b_start_time = time.time()\n loss, acc = sess.run([cross_entropy, accuracy], feed_dict={x: x_batch,y_: y_batch,keep_prob: 1.})\n sess.run(train_op,feed_dict={x: x_batch, y_: y_batch, keep_prob: dropout})\n b_end_time = time.time() - b_start_time\n avg_loss += loss / total_batch\n avg_train_acc += acc / total_batch\n \n \n # save model every x epochs\n if epoch >= 0 and epoch % 50 == 0:\n # Save model\n # print (\"model saved to {}\".format(checkpoint_path))\n saver.save(sess, checkpoint_path, global_step = epoch)\n \n \n ## Display model every 1 epochs\n if epoch >= 0 and epoch % display_step == 0:\n end = time.time()\n plr = sess.run(lr)\n loss_values.append(avg_loss) \n test_acc = evaluate()\n print(\"Epoch:\"+ '%04d' % (epoch+1) + \\\n \", Ep_time=\" + \"{:.2f}\".format(end - start) + \\\n \", lr=\" + \"{:.9f}\".format(plr) + \\\n \", avg_cost=\" + \"{:.3f}\".format(avg_loss) + \\\n \", Train_Acc=\" + \"{:.2f}\".format(avg_train_acc) + \\\n \", Test_Acc=\" + \"{:.2f}\".format(test_acc) + \n \", Batch (sec)=\" + \"{:.3f}\".format(b_end_time) ) \n \n # Summarize model every x epochs\n if epoch >= 0 and epoch % 1 == 0: \n summary= sess.run(merged,feed_dict={x: x_batch,y_: y_batch,keep_prob: 1.})\n train_writer.add_summary(summary, epoch)\nprint(\"Wall Time:\",\"{:.1f}\".format((time.time() - tr_start)/60.0, \"Min\")) \nprint(\"Optimization Finished!\")\nprint (\"model saved to {}\".format(checkpoint_path))\nsaver.save(sess, checkpoint_path, global_step = (epoch+1)*step)\ntrain_writer.close()", "_____no_output_____" ], [ "# Find the labels of test set\ny_pred_lb = sess.run(tf.argmax(y_CNN,1), feed_dict={x: X_test, y_: y_test, keep_prob: 1.})\ny_pred = sess.run(y_CNN, feed_dict={x: X_test, y_: y_test, keep_prob: 1.})\n\n# lets save kernels\nkernels_l1 = sess.run(tf.reshape(tf.transpose(W_conv1, perm=[2, 3, 0, 1]),[32,-1]))\nkernels_l2 = sess.run(tf.reshape(tf.transpose(W_conv2, perm=[2, 3, 0, 1]),[32*64,-1]))", "_____no_output_____" ], [ "plt.plot([np.mean(loss_values[i:i+5]) for i in range(len(loss_values))])\nplt.show()", "_____no_output_____" ] ], [ [ "## Evaluation", "_____no_output_____" ], [ "Accuracy depends on the number of epochs that you set in the parameters part.", "_____no_output_____" ] ], [ [ "y_ = np.argmax(y_test,1) # ground truth\nprint metrics.classification_report(y_true= y_, y_pred= y_pred_lb)\nprint metrics.confusion_matrix(y_true= y_, y_pred= y_pred_lb)\nprint(\"Classification accuracy: %0.6f\" % metrics.accuracy_score(y_true= y_, y_pred= y_pred_lb) )\nprint(\"Log Loss: %0.6f\" % metrics.log_loss(y_true= y_, y_pred= y_pred) )", "_____no_output_____" ] ], [ [ "### Viz\nDo you want to look at all the filters? Let's use __utils__ to visualize them.", "_____no_output_____" ] ], [ [ "!wget --output-document utils1.py http://deeplearning.net/tutorial/code/utils.py\nimport utils1\nfrom utils1 import tile_raster_images", "_____no_output_____" ] ], [ [ "Here you can plot the 32 filters in the first convolutional layer.", "_____no_output_____" ] ], [ [ "image = Image.fromarray(tile_raster_images(kernels_l1, img_shape=(5, 5) ,tile_shape=(4, 8), tile_spacing=(1, 1)))\n### Plot image\nplt.rcParams['figure.figsize'] = (18.0, 18.0)\nimgplot = plt.imshow(image)\nimgplot.set_cmap('gray') ", "_____no_output_____" ] ], [ [ "Also, you can plot and take a look at some filters from the second convolutional layers:", "_____no_output_____" ] ], [ [ "image = Image.fromarray(tile_raster_images(kernels_l2, img_shape=(5, 5) ,tile_shape=(4, 12), tile_spacing=(1, 1)))\n### Plot image\nplt.rcParams['figure.figsize'] = (18.0, 18.0)\nimgplot = plt.imshow(image)\nimgplot.set_cmap('gray') ", "_____no_output_____" ] ], [ [ "To understand better, let's apply one of these filters on a sample signal (spectogram image). First, let's plot a sample spectogram:", "_____no_output_____" ] ], [ [ "plt.rcParams['figure.figsize'] = (5.0, 5.0)\nsampleimage1 = X_test[3]\nplt.imshow(np.reshape(sampleimage1,[64,128]), cmap=\"gray\")", "_____no_output_____" ] ], [ [ "Now we apply different filters on them, and plot the result:", "_____no_output_____" ] ], [ [ "# Launch the graph\nwith tf.Session() as sess:\n sess.run(init)\n saver = tf.train.Saver(tf.all_variables())\n \n # load previously trained model if appilcable\n ckpt = tf.train.get_checkpoint_state(chk_directory)\n if ckpt:\n print \"loading model: \",ckpt.model_checkpoint_path\n saver.restore(sess, ckpt.model_checkpoint_path)\n ActivatedUnits1 = sess.run(convolve1,feed_dict={x:np.reshape(sampleimage1,[1,64*128],order='F'),keep_prob:1.0})\n plt.figure(1, figsize=(20,20))\n n_columns = 3\n n_rows = 3\n for i in range(9):\n plt.subplot(n_rows, n_columns, i+1)\n plt.title('Filter ' + str(i))\n plt.imshow(ActivatedUnits1[0,:,:,i], interpolation=\"nearest\", cmap=\"gray\")", "_____no_output_____" ] ], [ [ "### Benchmark:\n- SETI_multi_gpu_train.py achieves ~72% accuracy after 3k epochs of data (75K steps).\n- Speed: With batch_size 128. \n- __Notice:__ The model is not optimized to reach its highest accuracy. You can achieve better results by tuning the parameters.\n\n<table border=\"1\" style=\"box-sizing: border-box; border-spacing: 30px; background-color: transparent; color: #333333; font-family: 'Helvetica Neue', Helvetica, Arial, sans-serif; font-size: 12px;\">\n<tbody style=\"box-sizing: border-box;\">\n<tr style=\"box-sizing: border-box;\">\n<td style=\"box-sizing: border-box; padding: 3px; text-align:left;\"><span style=\"box-sizing: border-box; font-weight: bold;\">CPU Architecture</span></td>\n<td style=\"box-sizing: border-box; padding: 3px; text-align:left;\"><span style=\"box-sizing: border-box; font-weight: bold;\">CPU cores&nbsp;</span></td>\n<td style=\"box-sizing: border-box; padding: 3px; text-align:left;\"><span style=\"box-sizing: border-box; font-weight: bold;\">Memory&nbsp;</span></td>\n<td style=\"box-sizing: border-box; padding: 3px; text-align:left;\"><span style=\"box-sizing: border-box; font-weight: bold;\">GPU&nbsp;</span></td>\n<td style=\"box-sizing: border-box; padding: 3px; text-align:left;\"><span style=\"box-sizing: border-box; font-weight: bold;\">Step time (sec/batch)&nbsp;</span></td>\n<td style=\"box-sizing: border-box; padding: 3px; text-align:left;\"><span style=\"box-sizing: border-box; font-weight: bold;\">&nbsp;Accuracy</span></td>\n</tr>\n<tr style=\"box-sizing: border-box;\">\n<td style=\"box-sizing: border-box; padding: 3px; text-align:left;\">POWER8</td>\n<td style=\"box-sizing: border-box; padding: 3px; text-align:center;\">40</td>\n<td style=\"box-sizing: border-box; padding: 3px; text-align:center;\">256 GB</td>\n<td style=\"box-sizing: border-box; padding: 3px; text-align:left;\">1 x Tesla K80</td>\n<td style=\"box-sizing: border-box; padding: 3px; text-align:center;\">~0.127 </td>\n<td style=\"box-sizing: border-box; padding: 3px; text-align:left;\">~72% at 75K steps (3 hours)</td>\n</tr>\n<tr style=\"box-sizing: border-box;\">\n<td style=\"box-sizing: border-box; padding: 3px; text-align:left;\" >POWER8</td>\n<td style=\"box-sizing: border-box; padding: 3px; text-align:center;\">32</td>\n<td style=\"box-sizing: border-box; padding: 3px; text-align:center;\">128 GB</td>\n<td style=\"box-sizing: border-box; padding: 3px; text-align:left;\">1 x Tesla P100 w/NVLink np8g4</td>\n<td style=\"box-sizing: border-box; padding: 3px; text-align:center;\">~0.035 </td>\n<td style=\"box-sizing: border-box; padding: 3px; text-align:left;\">~72% at 75K steps (1 hour)</td>\n</tr>\n\n\n</tbody>\n</table>\n\n\n", "_____no_output_____" ], [ "## Want to learn more?\n\n[Deep Learning with TensorFlow](https://cocl.us/SETI_CodePattern_ML0120EN) is a free course in __cognitiveclass.ia__ where you can learn TensorFlow and Deep Learning together.\n\nRunning deep learning programs usually needs a high performance platform. PowerAI speeds up deep learning and AI. Built on IBM's Power Systems, PowerAI is a scalable software platform that accelerates deep learning and AI with blazing performance for individual users or enterprises. The PowerAI platform supports popular machine learning libraries and dependencies including Tensorflow, Caffe, Torch, and Theano. You can use [PowerAI on IBM Cloud](https://cocl.us/ML0120EN_PAI).", "_____no_output_____" ], [ "### Authors\n\n<div class=\"teacher-image\" style=\" float: left;\n width: 115px;\n height: 115px;\n margin-right: 10px;\n margin-bottom: 10px;\n border: 1px solid #CCC;\n padding: 3px;\n border-radius: 3px;\n text-align: center;\"><img class=\"alignnone wp-image-2258 \" src=\"https://ibm.box.com/shared/static/tyd41rlrnmfrrk78jx521eb73fljwvv0.jpg\" alt=\"Saeed Aghabozorgi\" width=\"178\" height=\"178\" /></div>\n\n#### Saeed Aghabozorgi\n\n[Saeed Aghabozorgi](https://ca.linkedin.com/in/saeedaghabozorgi), PhD is Sr. Data Scientist in IBM with a track record of developing enterprise level applications that substantially increases clients’ ability to turn data into actionable knowledge. He is a researcher in data mining field and expert in developing advanced analytic methods like machine learning and statistical modelling on large datasets.</p>\n\n\n", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown" ]
[ [ "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code", "code", "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown", "markdown" ] ]
ece9ddd08cec906df591192965d45f3c363588d2
111,668
ipynb
Jupyter Notebook
nb/PyIntro.ipynb
sybenzvi/PHY403
159f1203b5fc92ffc1f2b7dc9fef3c2f78207dd7
[ "BSD-3-Clause" ]
3
2020-05-27T23:51:39.000Z
2021-02-03T03:34:53.000Z
nb/PyIntro.ipynb
sybenzvi/PHY403
159f1203b5fc92ffc1f2b7dc9fef3c2f78207dd7
[ "BSD-3-Clause" ]
null
null
null
nb/PyIntro.ipynb
sybenzvi/PHY403
159f1203b5fc92ffc1f2b7dc9fef3c2f78207dd7
[ "BSD-3-Clause" ]
7
2020-05-06T16:01:09.000Z
2022-02-04T18:47:26.000Z
43.131711
23,476
0.690251
[ [ [ "# Basic Introduction to Python\n\nThis is a very basic introduction to Python. It is not exhaustive, but is meant to give you a starting point.\n\nThis notebook was written for PHY 403 by Segev BenZvi, University of Rochester, (Spring 2021).\n\nIt is based on a similar (longer) Python guide written by Kyle Jero (UW-Madison) for the IceCube Programming Bootcamp in June 2015, and includes elements from older guides by Jakob van Santen and Nathan Whitehorn.\n\n\n## What is Python?\n\nPython is an **imperative**, **interpreted** programming language with **strong** **dynamic** typing.\n\n- **Imperative**: programs are built around one or more subroutines known as \"functions\" and \"classes\"\n- **Interpreted**: program instructions are executed on the fly rather than being pre-compiled into machine code\n- **Dynamic Typing**: data types of *variables* (`int`, `float`, `string`, etc.) are determined on the fly as the program runs\n- **Strong Typing**: converting a variable from one type to another (e.g., `int` to `string`) is not always done automatically\n\nPython offers fast and flexible development and can be used to glue together many different analysis packages which have \"Python bindings.\"\n\nAs a rule, Python programs are slower than compiled programs written in Fortran, C, and C++. But it's a much more forgiving programming language.\n\n## Why Use Python?\n\nPython is one of the most popular scripting languages in the world, with a huge community of users and support on all major platforms (Windows, OS X, Linux).\n\n![Languages](intro/languages.jpg)\n\n![Github](intro/github.png)\n\nPretty much every time I've run into a problem programming in Python, I've found a solution after a couple of minutes of searching on google or stackoverflow.com!\n\n## Key Third-Party Packages\n\n### Must-Haves\n\n- <a href=\"http://www.numpy.org/\">NumPy</a>: random number generation, transcendental functions, vectorized math, linear algebra.\n- <a href=\"http://www.scipy.org/\">SciPy</a>: statistical tests, special functions, numerical integration, curve fitting and minimization.\n- <a href=\"http://matplotlib.org/\">Matplotlib</a>: plotting: xy plots, error bars, contour plots, histograms, etc.\n- <a href=\"http://ipython.org/\">IPython</a>: an interactive python shell, which can be used to run Mathematica-style analysis notebooks.\n\n### Worth Using\n\n- <a href=\"https://scikits.appspot.com/\">SciKits</a>: data analysis add-ons to SciPy, including machine learning algorithms.\n- <a href=\"http://pandas.pydata.org/\">Pandas</a>: functions and classes for specialized data analysis.\n- <a href=\"http://www.astropy.org/\">AstroPy</a>: statistical methods useful for time series analysis and data reduction in astronomy.\n- <a href=\"http://dan.iel.fm/emcee/current/\">Emcee</a>: great implementation of Markov Chain Monte Carlo; nice to combine with the package <a href=\"https://pypi.python.org/pypi/corner\">Corner</a>.\n- <a href=\"https://pytorch.org/\">PyTorch</a>: currently the most popular deep learning framework.\n\n### Specialized Bindings\n\nMany C and C++ packages used in high energy physics come with bindings to Python. For example, the <a href=\"https://root.cern.ch/\">ROOT</a> package distributed by CERN can be run completely from Python.\n\n### Online Tools\n\nIf you don't want to install all these packages on your own computer, you can create a free account with many cloud services:\n- <a href=\"https://colab.research.google.com/\">Google Colab</a>\n- <a href=\"https://mybinder.org/\">Binder</a>\n- <a href=\"https://www.kaggle.com/\">Kaggle</a>\n- <a href=\"https://notebooks.azure.com\">Microsoft Azure</a>\n- <a href=\"https://cocalc.com/\">CoCalc</a>\n- <a href=\"https://datalore.io/\">Datalore</a>\n\nThe screen capture below shows an Azure notebook.\n\n![Azure](intro/azure.png)\n\nAzure gives you access to jupyter notebooks running on remote servers. Recent versions of SciPy, NumPy, and Matplotlib are provided.\n\n## Programming Basics\n\nWe will go through the following topics, and then do some simple exercises.\n\n- Arithmetic Operators\n- Variables and Lists\n- Conditional Statements\n- Loops (`for` and `while`)\n- Functions\n- Importing Modules", "_____no_output_____" ], [ "### Arithmetic Operators", "_____no_output_____" ], [ "#### Addition", "_____no_output_____" ] ], [ [ "1+2", "_____no_output_____" ] ], [ [ "#### Subtraction", "_____no_output_____" ] ], [ [ "19993 - 7743", "_____no_output_____" ] ], [ [ "#### Multiplication", "_____no_output_____" ] ], [ [ "3*8", "_____no_output_____" ] ], [ [ "#### Division", "_____no_output_____" ] ], [ [ "50 / 2", "_____no_output_____" ], [ "1 / 3", "_____no_output_____" ] ], [ [ "Note: in Python 2, division of two integers is always **floor division**. In Python 3, 1/2 automatically evaluates to the *floating point number* 0.5. To use floor division in Python 3, you'll have to run `1 // 2`.", "_____no_output_____" ] ], [ [ "1 // 2 # floor division: will give you zero (int), not 0.5 (float)", "_____no_output_____" ] ], [ [ "#### Modulo/Remainder", "_____no_output_____" ] ], [ [ "30 % 4", "_____no_output_____" ], [ "3.14159265359 % 1.", "_____no_output_____" ] ], [ [ "#### Exponentiation", "_____no_output_____" ] ], [ [ "4**2", "_____no_output_____" ] ], [ [ "### Variables\n\nVariables are extremely useful for storing values and using them later. One can declare a variable to contain the output of any variable, function call, etc. However, variable names must follow certain rules:\n\n1. Variable names must start with a letter (upper or lower case) or underscore\n2. Variable names may contain only letters, numbers, and underscores _\n3. The following names are **reserved keywords** in Python and cannot be used as variable names:\n\n ` and del from not while`\n\n ` as elif global or with`\n \n ` assert else if pass yield`\n \n ` break except import print`\n \n ` class exec in raise`\n \n ` continue finally is return`\n \n ` def for lambda try`", "_____no_output_____" ] ], [ [ "x = 5 + 6", "_____no_output_____" ] ], [ [ "This time nothing printed out because the output of the expression was stored in the variable `x`. To see the value we have to just evaluate `x` in a cell...", "_____no_output_____" ] ], [ [ "x", "_____no_output_____" ] ], [ [ "...or we explicitly call the `print` function:", "_____no_output_____" ] ], [ [ "print(x)", "11\n" ] ], [ [ "Recall that we don't have to explicitly declare what type something is in python, something that is not true in many other languages, we simply have to name our variable and specify what we want it to store. However, it is still nice to know the types of things sometimes and learn what types python has available for our use.", "_____no_output_____" ], [ "#### Data Types", "_____no_output_____" ] ], [ [ "type(x)", "_____no_output_____" ], [ "y = 2\ntype(x/y)", "_____no_output_____" ], [ "z = 1.\ntype(z/y)", "_____no_output_____" ], [ "w = True\ntype(w)", "_____no_output_____" ], [ "h = 'Hello'\ntype(h)", "_____no_output_____" ] ], [ [ "##### Strings", "_____no_output_____" ], [ "Strings are collections of characters between pairs of single or double quotes. (Note: in other languages like C/C++, Java, etc., you must use double quotes for strings.) The ability to mix/match single and double quote pairs makes it easy to put a literal quotation mark inside a string. That is, the quotation mark won't be treated as a delimiting character that indicates the start or end of a string.\n\nString manipulation is an important part of managing certain kinds of data, such as records in text files. Python lets you do nice things like combine strings using simple arithmetic operators.", "_____no_output_____" ] ], [ [ "s = \" \" # Create a string with double quotes.\nw = 'World!' # Create a string with single quotes.\n\nprint(h + s + w) # Concatenate strings with + and print the result.", "Hello World!\n" ], [ "mystring1 = 'mystring1'\nmystring2 = \"mystring2\"\n\napostrophes=\"They're \"\nquotes='\"hypothetically\" ' # Add literal \" marks inside the string.\nsaying=apostrophes + quotes + \"good for you to know.\"\n\nprint(saying)", "They're \"hypothetically\" good for you to know.\n" ] ], [ [ "C-style formatted printing is also allowed:", "_____no_output_____" ] ], [ [ "p = 'Pi'\nprint('%s = %.3f' % (p, 3.14159265359)) # old-style string formatting\nprint('{:s} = {:.6f}'.format(p, 3.14159265359)) # new-style string formatting", "Pi = 3.142\nPi = 3.141593\n" ] ], [ [ "### Lists\n\nImagine that we are storing the heights of people or the results of a random process. We could imagine taking and making a new variable for each piece of information but this becomes convoluted very quickly. In instances like this it is best to store the collection of information together in one place. In python this collection is called a list and can be defined by enclosing data separated by commas in square brackets. A empty list can also be specified by square brackets with nothing between them and filled later in the program.", "_____no_output_____" ] ], [ [ "blanklist = []\nblanklist", "_____no_output_____" ], [ "alist=[1, 2.5, '3']\nprint(alist)\nprint(type(alist))", "[1, 2.5, '3']\n<class 'list'>\n" ] ], [ [ "Notice that the type of our list is `list` and no mention of the data type it contains is made. This is because python does not fuss about what type of thing is in a list or even mixing of types in lists. If you have worked with nearly any other language this is different then you are used to since the type of your list must be homogeneous.", "_____no_output_____" ] ], [ [ "blist=[1, 'two', 3.0]\nblist", "_____no_output_____" ], [ "print(type(blist))", "<class 'list'>\n" ] ], [ [ "You can check the current length of a list by calling the `len` function with the list as the argument:", "_____no_output_____" ] ], [ [ "len(blist)", "_____no_output_____" ] ], [ [ "In addition, you can add objects to the list or remove them from the list in several ways:", "_____no_output_____" ] ], [ [ "blist.append(\"4\") # Add the string '4' to the end.\nblist", "_____no_output_____" ], [ "blist.insert(0, \"0\") # Add the string '0' before the first element.\nblist", "_____no_output_____" ], [ "blist.extend([5,6]) # Insert one list into another.\nprint(blist)\nprint(len(blist))", "['0', 1, 'two', 3.0, '4', 5, 6]\n7\n" ], [ "blist.append(7)\nblist", "_____no_output_____" ], [ "# Multiplication doubles the elements in the list (not term-by-term multiplication).\n\nblist = blist*2\nblist", "_____no_output_____" ], [ "# Search for and remove the first element matching '4'.\n\nblist.remove(\"4\")\nblist", "_____no_output_____" ], [ "# Search for and remove the first element matching '4'.\n# There are now no '4' elements left.\n\nblist.remove('4')\nblist", "_____no_output_____" ] ], [ [ "##### List Element Access\n\nIndividual elements (or ranges of elements) in the list can be accessed using the square bracket operators [ ]. For example:", "_____no_output_____" ] ], [ [ "# Access elements with respect to the front of the list.\n# The first element has index 0.\n\nprint(blist[0])\nprint(blist[4])", "0\n5\n" ], [ "# Access elements with respect to the end of the list.\n# The last element has index -1.\n\nprint(blist[-1])\nprint(blist[-2])\nprint(blist[-3])", "7\n6\n5\n" ], [ "blist[0:4]", "_____no_output_____" ], [ "print(blist) # list slicing example:\nblist[0:6:2] # sytax: start, stop, stride", "['0', 1, 'two', 3.0, 5, 6, 7, '0', 1, 'two', 3.0, 5, 6, 7]\n" ] ], [ [ "This is an example of a slice, where we grab a subset of the list and also decide to step through the list by skipping every other element. The syntax is\n\n`listname[start:stop:stride]`\n\nNote that if start and stop are left blank, the full list is used in the slice by default.", "_____no_output_____" ] ], [ [ "blist[::2]", "_____no_output_____" ], [ "print(blist[::-1]) # An easy way to reverse the order of elements\nprint(blist)", "[7, 6, 5, 3.0, 'two', 1, '0', 7, 6, 5, 3.0, 'two', 1, '0']\n['0', 1, 'two', 3.0, 5, 6, 7, '0', 1, 'two', 3.0, 5, 6, 7]\n" ] ], [ [ "A simple built-in function that is used a lot is the range function. It is not a list but returns one so we will discuss it here briefly. The syntax of the function is range(starting number, ending number, step size ). All three function arguments are required to be integers with the ending number not being included in the list. Additionally the step size does not have to be specified, and if it is not the value is assumed to be 1.", "_____no_output_____" ] ], [ [ "for i in range(0,10):\n print(i)", "0\n1\n2\n3\n4\n5\n6\n7\n8\n9\n" ], [ "for batman in range(0,10,2):\n print(batman)", "0\n2\n4\n6\n8\n" ] ], [ [ "### Conditional Statements\n\nConditionals are useful for **altering the flow of control** in your programs. For example, you can execute blocks of code (or skip them entirely) if certain conditions are met.\n\nConditions are created using `if/elif/else` blocks.\n\nFor those of you familiar with C, C++, Java, and similar languages, you are probably used to code blocks being marked off with curly braces: { }\n\nIn Python braces are not used. Code blocks are *indented*, and the Python interpreter decides what's in a block depending on the indentation. Good practice (for readability) is to use 4 spaces per indentation. If you are programming in a jupyter notebook, the notebook will automatically indent conditional blocks for you.", "_____no_output_____" ] ], [ [ "x = 54\n\nif x > 10:\n print(\"x > 10\")\nelif x > 5:\n print(\"x > 5\")\nelse:\n print(\"x <= 5\")", "x > 10\n" ], [ "isEven = (x % 2 == 0) # Store a boolean value\nprint(isEven)\n\n# Note the perverse double negative in this boolean expression...\nif not isEven:\n print(\"x is odd\")\nelse:\n print(\"x is even\")", "True\nx is even\n" ] ], [ [ "#### Comparison Operators\n\nThere are several predefined operators used to make boolean comparisons in Python. They are similar to operators used in C, C++, and Java:\n\n`==` ... test for equality\n\n`!=` ... test for not equal\n\n`>` ... greater than\n\n`>=` ... greater than or equal to\n\n`<` ... less than\n\n`<=` ... less than or equal to\n\n#### Combining Boolean Values\n\nFollowing the usual rules of boolean algebra, boolean values can be negated or combined in several ways:\n\n##### Logical AND\n\nYou can combine two boolean variables using the operator `&&` or the keyword `and`:", "_____no_output_____" ] ], [ [ "print('x y | x && y')\nprint('---------------')\n\nfor x in [True, False]:\n for y in [True, False]:\n print('{:d} {:d} | {:d}'.format(x, y, x and y))", "x y | x && y\n---------------\n1 1 | 1\n1 0 | 0\n0 1 | 0\n0 0 | 0\n" ], [ "# Will evaluate to True and print the output.\n\nx = 10\nif x > 2 and x < 20:\n print(x)", "10\n" ], [ "# Will evaluate to False and not print the output.\n\nif x < 2 and x > 20:\n print(x)", "_____no_output_____" ] ], [ [ "##### Logical OR\n\nYou can also combine two boolean variables using the operator `||` or the keyword `or`:", "_____no_output_____" ] ], [ [ "print('x y | x || y')\nprint('---------------')\n\nfor x in [True, False]:\n for y in [True, False]:\n\n print('{:d} {:d} | {:d}'.format(x, y, x or y))", "x y | x || y\n---------------\n1 1 | 1\n1 0 | 1\n0 1 | 1\n0 0 | 0\n" ], [ "# Will evaluate to True and print the output.\n\nx = 10\nif x > 2 or x < 0:\n print(x)", "10\n" ], [ "# Will evaluate to False and print the output.\n\nif x < 2 or x > 20:\n print(x)", "_____no_output_____" ] ], [ [ "##### Logical NOT\n\nIt's possible to negate a boolean expression using the keyword `not`:", "_____no_output_____" ] ], [ [ "print('x | not x')\nprint('----------')\nfor x in [True, False]:\n print('{:d} | {:d}'.format(x, not x))", "x | not x\n----------\n1 | 0\n0 | 1\n" ] ], [ [ "A more complex truth table demonstrating the duality\n\n$\\overline{AB} = \\overline{A}+\\overline{B}$:", "_____no_output_____" ] ], [ [ "print('A B | A and B | !(A and B) | !A or !B')\nprint('-------------------------------------------')\nfor A in [True, False]:\n for B in [True, False]:\n print('{:d} {:d} | {:d} | {:d} | {:d}'.format(\n A, B, A and B, not (A and B), not A or not B)) ", "A B | A and B | !(A and B) | !A or !B\n-------------------------------------------\n1 1 | 1 | 0 | 0\n1 0 | 0 | 1 | 1\n0 1 | 0 | 1 | 1\n0 0 | 0 | 1 | 1\n" ] ], [ [ "### Loops\n\nLoops are useful for executing blocks of code as long as a logical condition is satisfied.\n\nOnce the loop condition is no longer satisfied, the flow of control is returned to the main body of the program. Note that **infinite loops**, a serious runtime bug where the loop condition never evaluates to `False`, are possible, so you have to be careful.\n\n#### While Loop\n\nThe `while` loop evaluates until a condition is false. Note that loops can be nested inside each other, and can also contain nested conditional statements.", "_____no_output_____" ] ], [ [ "i = 0\nwhile i < 10: # Loop condition: i < 10\n i += 1 # Increment the value of i\n if i % 2 == 0: # Print i if it's even\n print(i)", "2\n4\n6\n8\n10\n" ] ], [ [ "#### For Loop\n\nThe `for` loop provides the same basic functionality as the `while` loop, but allows for a simpler syntax in certain cases.\n\nFor example, if we wanted to access all the elements inside a list one by one, we could write a while loop with a variable index `i` and access the list elements as `listname[i]`, incrementing `i` until it's the same size as the length of the list.\n\nHowever, the `for` loop lets us avoid the need to declare an index variable. For example:", "_____no_output_____" ] ], [ [ "for x in range(1,11): # Loop through a list of values [1..10]\n if x % 2 == 0: # Print the list value if it's even\n print(x)", "2\n4\n6\n8\n10\n" ], [ "for i, x in enumerate(['a', 'b', 'c', 'd', 'e']):\n print('{:d} {:s}'.format(i+1, x))", "1 a\n2 b\n3 c\n4 d\n5 e\n" ] ], [ [ "##### List Comprehension and Zipping Lists in a For Loop", "_____no_output_____" ], [ "If we are interested in building lists we can start from a blank list and append things to it in a for loop, or use a **list comprehension** which combines for loops and list creation into line. The syntax is a set of square brackets that contains formula and a for loop.", "_____no_output_____" ] ], [ [ "squaredrange = [e**2 for e in range(1,11)]\n\nprint(squaredrange)", "[1, 4, 9, 16, 25, 36, 49, 64, 81, 100]\n" ] ], [ [ "You can also loop through **two lists simultaneously** using the `zip` function:", "_____no_output_____" ] ], [ [ "mylist = range(1,11)\nmylist2 = [e**2 for e in mylist]\n\nfor x, y in zip(mylist, mylist2):\n print('{:2d} {:4d}'.format(x, y))", " 1 1\n 2 4\n 3 9\n 4 16\n 5 25\n 6 36\n 7 49\n 8 64\n 9 81\n10 100\n" ] ], [ [ "### Functions\n\nFunctions are subroutines that accept some input and produce zero or more outputs. They are typically used to define common tasks in a program.\n\nRule of thumb: if you find that you are copying a piece of code over and over inside your script, it should probably go into a function.\n\n#### Example: Rounding\n\nThe following function will round integers to the nearest 10:", "_____no_output_____" ] ], [ [ "def round_int(x):\n # Note that we are using floor division (assumes Python 3).\n return 10 * ((x + 5)//10)\n\nfor x in range(2, 50, 5):\n print('{:5d} {:5d}'.format(x, round_int(x)))", " 2 0\n 7 10\n 12 10\n 17 20\n 22 20\n 27 30\n 32 30\n 37 40\n 42 40\n 47 50\n" ] ], [ [ "# In-Class Exercise\n\nWith the small amount we've gone through, you can already write reasonably sophisticated programs. For example, we can write a loop that generates the Fibonacci sequence.\n\nJust to remind you, the Fibonacci sequence is the list of numbers\n\n1, 1, 2, 3, 5, 8, 13, 21, 34, 55, ...\n\nIt is defined by the linear homogeneous recurrence relation\n\n$F_{n} = F_{n-1} + F_{n-2}$, where $F_0=F_1=1$.\n\nThe exercise is:\n1. Write a Python function that generate $F_n$ given $n$.\n2. Use your function to generate the first 100 numbers in the Fibonacci sequence.", "_____no_output_____" ] ], [ [ "# Easy implementation: recursive function\n\ndef fib(n):\n \"\"\"Generate term n of the Fibonacci sequence\"\"\"\n if n <= 1:\n # if n==0 or n==1: return 1\n return 1\n else:\n return fib(n-1) + fib(n-2)", "_____no_output_____" ], [ "# Only generate up to element 35 in the sequence, because this function is s-l-o-w.\n\nfor n in range(0, 35):\n Fn = fib(n)\n print('{:3d}{:25d}'.format(n, Fn))", " 0 1\n 1 1\n 2 2\n 3 3\n 4 5\n 5 8\n 6 13\n 7 21\n 8 34\n 9 55\n 10 89\n 11 144\n 12 233\n 13 377\n 14 610\n 15 987\n 16 1597\n 17 2584\n 18 4181\n 19 6765\n 20 10946\n 21 17711\n 22 28657\n 23 46368\n 24 75025\n 25 121393\n 26 196418\n 27 317811\n 28 514229\n 29 832040\n 30 1346269\n 31 2178309\n 32 3524578\n 33 5702887\n 34 9227465\n" ] ], [ [ "This function will work just fine for small n. Unfortunately, the recursive calls to `fib` cause the **function call stack** to grow rapidly with n. When n gets sufficiently large, you may hit the Python call stack limit. At that point your program will crash.\n\nEven worse, the call is s-l-o-w. Using the builtin `timeit` function available in the notebook, we can time the length of the function call to evaluate Fibonacci element 20. It will take several milliseconds, which doesn't sound bad, but it's actually very slow in computer terms. Imagine calling this function millions of times in a loop; it will be a huge bottleneck.", "_____no_output_____" ] ], [ [ "timeit(fib(20))", "2 ms ± 39.6 µs per loop (mean ± std. dev. of 7 runs, 100 loops each)\n" ] ], [ [ "This next implementation uses two \"state variables\" `a` and `b` to build up the Fibonacci sequence. Instead of computing the series recursively, which builds up a huge stack of function calls in memory, all of the work is done by an internal `while` loop.", "_____no_output_____" ] ], [ [ "# Better implementation which uses two state variables to compute the sequence.\n\ndef fibBetter(n):\n \"\"\"Generate the Fibonacci series at position n\"\"\"\n a, b = 0, 1 # initial values\n while n > 0: # build up the series from n=0\n a, b, n = b, a+b, n-1 # store results in loop variables\n return b", "_____no_output_____" ] ], [ [ "While this doesn't seem like a major change, the time it takes to run the function is of order a few microseconds, or 1000 times faster than the recursive version. Quite a nice optimization!", "_____no_output_____" ] ], [ [ "timeit(fibBetter(20))", "1.31 µs ± 16.9 ns per loop (mean ± std. dev. of 7 runs, 1000000 loops each)\n" ] ], [ [ "As a result, it's trivial to compute the first 100 Fibonacci numbers; `fibBetter` runs almost instantly, while `fib` probably would have crashed the notebook due to its memory requirements before getting even halfway through.", "_____no_output_____" ] ], [ [ "for n in range(0, 100):\n Fn = fibBetter(n)\n print(\"%3d%25d\" % (n, Fn))", " 0 1\n 1 1\n 2 2\n 3 3\n 4 5\n 5 8\n 6 13\n 7 21\n 8 34\n 9 55\n 10 89\n 11 144\n 12 233\n 13 377\n 14 610\n 15 987\n 16 1597\n 17 2584\n 18 4181\n 19 6765\n 20 10946\n 21 17711\n 22 28657\n 23 46368\n 24 75025\n 25 121393\n 26 196418\n 27 317811\n 28 514229\n 29 832040\n 30 1346269\n 31 2178309\n 32 3524578\n 33 5702887\n 34 9227465\n 35 14930352\n 36 24157817\n 37 39088169\n 38 63245986\n 39 102334155\n 40 165580141\n 41 267914296\n 42 433494437\n 43 701408733\n 44 1134903170\n 45 1836311903\n 46 2971215073\n 47 4807526976\n 48 7778742049\n 49 12586269025\n 50 20365011074\n 51 32951280099\n 52 53316291173\n 53 86267571272\n 54 139583862445\n 55 225851433717\n 56 365435296162\n 57 591286729879\n 58 956722026041\n 59 1548008755920\n 60 2504730781961\n 61 4052739537881\n 62 6557470319842\n 63 10610209857723\n 64 17167680177565\n 65 27777890035288\n 66 44945570212853\n 67 72723460248141\n 68 117669030460994\n 69 190392490709135\n 70 308061521170129\n 71 498454011879264\n 72 806515533049393\n 73 1304969544928657\n 74 2111485077978050\n 75 3416454622906707\n 76 5527939700884757\n 77 8944394323791464\n 78 14472334024676221\n 79 23416728348467685\n 80 37889062373143906\n 81 61305790721611591\n 82 99194853094755497\n 83 160500643816367088\n 84 259695496911122585\n 85 420196140727489673\n 86 679891637638612258\n 87 1100087778366101931\n 88 1779979416004714189\n 89 2880067194370816120\n 90 4660046610375530309\n 91 7540113804746346429\n 92 12200160415121876738\n 93 19740274219868223167\n 94 31940434634990099905\n 95 51680708854858323072\n 96 83621143489848422977\n 97 135301852344706746049\n 98 218922995834555169026\n 99 354224848179261915075\n" ] ], [ [ "## Accessing Functions Beyond the Built-In Functions\n\nIf we want to use libraries and modules not defined within the built-in functionality of python we have to import them. There are a number of ways to do this.", "_____no_output_____" ] ], [ [ "import numpy as np\nimport scipy as sp", "_____no_output_____" ] ], [ [ "This imports the module `numpy` and the module `scipy`, and creates a reference to that modules in the current namespace. After you’ve run this statement, you can use `np.name` and `sp.name` to refer to constants, functions, and classes defined in module numpy and scipy.", "_____no_output_____" ] ], [ [ "np.pi", "_____no_output_____" ], [ "# Evaluate the sine and cosine of 120 degrees.\nnp.sin(2*np.pi/3), np.cos(2*np.pi/3)", "_____no_output_____" ], [ "# Exponentiation and logarithms.\n\na = np.exp(-1.) # 1/e = 0.368\nb = np.log(a) # ln(1/e) = -1\nc = np.log2(a) # base-2 log of 1/e\nd = np.log10(a) # base-10 log of 1/e\n\na, b, c, d", "_____no_output_____" ], [ "from numpy import *", "_____no_output_____" ] ], [ [ "This imports the module numpy, and creates references in the current namespace to all public objects defined by that module (that is, everything that doesn’t have a name starting with “_”).\n\nOr in other words, after you’ve run this statement, you can simply use a plain name to refer to things defined in module numpy. Here, numpy itself is not defined, so numpy.name doesn’t work. If name was already defined, it is replaced by the new version. Also, if name in numpy is changed to point to some other object, your module won’t notice.", "_____no_output_____" ] ], [ [ "pi", "_____no_output_____" ] ], [ [ "#### Importing Submodules\n\nYou can also import submodules from within a module.\n\nFor example, `scipy` has a submodule called `special` that contains a number of useful transcendental functions beyond the basic exponentiation and trigonometric functions available in `numpy`.\n\nIn the example below, we make three function calls to the Error function `Erf` using this module.", "_____no_output_____" ] ], [ [ "from scipy import special\n\n# The error function is the cumulative distribution of a Gaussian with mean 0 and width 1\n# (a.k.a., the normal distribution).\n\nprint(special.erf(0),\n special.erf(1),\n special.erf(2))", "0.0 0.8427007929497148 0.9953222650189527\n" ] ], [ [ "## NumPy Tips and Tricks\n\nNumPy is optimized for numerical work. The `array` type inside of the module behaves a lot like a list, but it is *vectorized* so that you can apply arithmetic operations and other functions to the array without having to loop through it.\n\nFor example, when we wanted to square every element inside a python list we used a list comprehension:", "_____no_output_____" ] ], [ [ "mylist = range(1,11)\n[x**2 for x in mylist]", "_____no_output_____" ] ], [ [ "This isn't that hard, but the syntax is a little ugly and we do have to explicitly loop through the list. In contrast, to square all the elements in the NumPy array you just apply the operator to the array variable itself:", "_____no_output_____" ] ], [ [ "myarray = np.arange(1,11)\nmyarray**2", "_____no_output_____" ] ], [ [ "### Evenly Spaced Numbers\n\nNumPy provides two functions to give evenly spaced numbers on linear or logarithmic scales.", "_____no_output_____" ] ], [ [ "np.linspace(1, 10, 21) # gives 21 evenly spaced numbers in [1..10]", "_____no_output_____" ], [ "np.logspace(1, 6, 6) # gives 6 logarithmically spaced numbers\n # between 1e1=10 and 1e6=1000000", "_____no_output_____" ], [ "np.logspace(1, 6, 6, base=2) # same as above, but using base-2 logarithm", "_____no_output_____" ] ], [ [ "### Slicing Arrays with Boolean Masks\n\nAn extremely useful feature in NumPy is the ability to create a \"mask\" array which can select values satisfying a logical condition:", "_____no_output_____" ] ], [ [ "x = np.arange(0, 8) # [0, 1, 2, 3, 4, 5, 6, 7]\ny = 3*x # [0, 3, 6, 9, 12, 15, 18, 21]\n\nc = x < 3\n\n# Print whether or not each element is < 3.\nprint(c)", "[ True True True False False False False False]\n" ], [ "# Select out only the elements of x for which the corresponding elements of c are < 3.\nprint(x[c])", "[0 1 2]\n" ], [ "# Select out only the elements of y for which the corresponding elements of c are < 3.\nprint(y[c])\n\n# Select out only the elements of y for which the corresponding elements of x are >= 3.\nprint(y[x >= 3])", "[0 3 6]\n[ 9 12 15 18 21]\n" ], [ "# Combine cuts on x with bitwise OR (| symbol) or AND (& symbol).\nc = (x<3) | (x>5)\nprint(y[c])", "[ 0 3 6 18 21]\n" ] ], [ [ "This is the type of selection used *all the time* in data analysis.", "_____no_output_____" ], [ "### File Input/Output\n\nStandard Python has functions to read basic text and binary files from disk.\n\nHowever, for numerical analysis your files will usually be nicely formatted into numerical columns separated by spaces, commas, etc. For reading such files, NumPy has a nice function called `genfromtxt`:", "_____no_output_____" ] ], [ [ "# Load data from file into a multidimensional array\ndata = np.genfromtxt('intro/data.txt')\n\nx = data[:,0] # x is the first column (numbering starts @ 0)\ny = data[:,1] # y is the second column\n\nprint(x)\nprint(y)", "[ 1. 2. 3. 4. 5. 6. 7. 8. 9. 10.]\n[ 1. 4. 9. 16. 25. 36. 49. 64. 81. 100.]\n" ] ], [ [ "## Plotting with Matplotlib\n\nMatplotlib is used to plot data and can be used to produce the usual xy scatter plots, contour plots, histograms, etc. that you're used to making for all basic data analyses.\n\nI strongly recommend that you go to the Matplotlib website and check out the huge <a href=\"http://matplotlib.org/gallery.html\">plot gallery</a>. This is the easiest way to learn how to make a particular kind of plot.", "_____no_output_____" ] ], [ [ "import matplotlib.pyplot as plt", "_____no_output_____" ], [ "plt.plot(x, y, \"k.\")\nplt.xlabel(\"x [arb. units]\")\nplt.ylabel(\"y [arb. units]\")\nplt.title(\"Some XY data\");", "_____no_output_____" ] ], [ [ "Here is an example of how to change the default formatting of the text in your plot. Also note how LaTeX is supported!", "_____no_output_____" ] ], [ [ "import matplotlib as mpl\nmpl.rc('font', size=16)\n\nplt.plot(x, y, \"k.\")\nplt.xlabel(r\"$\\sin({x)}$ [arb. units]\")\nplt.ylabel(r\"$\\zeta(y)$ [arb. units]\")\nplt.title(\"Some XY data\")", "_____no_output_____" ] ], [ [ "### Using NumPy and Matplotlib Together\n\nHere we create some fake data with NumPy and plot it, including a legend.", "_____no_output_____" ] ], [ [ "x = np.linspace(-np.pi, np.pi, 1000, endpoint=True)\nc = np.cos(x)\ns = np.sin(x)\n\nplt.plot(x,c,label=\"Cosine\",color=\"r\",linestyle=\"--\",linewidth=2)\nplt.plot(x,s,label=\"Sine\",color=\"b\",linestyle=\"-.\",linewidth=2)\nplt.xlabel(\"$x$\",fontsize=14)\nplt.xlim(-np.pi,np.pi)\n\n# Override default ticks and labels\nxticks = [-np.pi, -0.5*np.pi, 0, 0.5*np.pi, np.pi]\nlabels = [\"$-\\pi$\", \"$-\\pi/2$\", \"$0$\", \"$\\pi/2$\", \"$\\pi$\"]\nplt.xticks(xticks, labels)\n\nplt.ylabel(\"$y$\",fontsize=14)\nplt.ylim(-1,1)\nplt.legend(fontsize=14, loc=\"best\", numpoints=1)", "_____no_output_____" ] ], [ [ "## Help Manual and Inspection\n\nWhen running interactive sessions, you can use the built-in help function to view module and function documentation.\n\nFor example, here is how to view the internal documentation for the built-in function that calculates the greatest common divisor of two numbers:", "_____no_output_____" ] ], [ [ "from fractions import gcd\n\nhelp(gcd)", "Help on function gcd in module fractions:\n\ngcd(a, b)\n Calculate the Greatest Common Divisor of a and b.\n \n Unless b==0, the result will have the same sign as b (so that when\n b is divided by it, the result comes out positive).\n\n" ] ], [ [ "The `inspect` module is nice if you actually want to look at the **source code** of a function. Just import inspect and call the `getsource` function for the code you want to see:", "_____no_output_____" ] ], [ [ "from inspect import getsource\n\nprint(getsource(gcd))", "def gcd(a, b):\n \"\"\"Calculate the Greatest Common Divisor of a and b.\n\n Unless b==0, the result will have the same sign as b (so that when\n b is divided by it, the result comes out positive).\n \"\"\"\n import warnings\n warnings.warn('fractions.gcd() is deprecated. Use math.gcd() instead.',\n DeprecationWarning, 2)\n if type(a) is int is type(b):\n if (b or a) < 0:\n return -math.gcd(a, b)\n return math.gcd(a, b)\n return _gcd(a, b)\n\n" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code" ]
[ [ "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code", "code", "code", "code", "code" ], [ "markdown", "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code", "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ] ]
ece9f0bbbb70a081c59b7f450b7ace2879f6aeec
35,253
ipynb
Jupyter Notebook
original_code/ch04.ipynb
jieunjeon/CatchingDudeoji
6cbb7b8d484ff21eac2d2640a862decfb5e3b9be
[ "MIT" ]
6
2021-07-16T11:16:00.000Z
2021-08-31T09:23:27.000Z
original_code/ch04.ipynb
jieunjeon/CatchingDudeoji
6cbb7b8d484ff21eac2d2640a862decfb5e3b9be
[ "MIT" ]
3
2021-07-27T03:14:17.000Z
2021-08-02T10:44:42.000Z
original_code/ch04.ipynb
jieunjeon/CatchingDudeoji
6cbb7b8d484ff21eac2d2640a862decfb5e3b9be
[ "MIT" ]
22
2021-07-18T01:25:01.000Z
2021-08-20T10:28:31.000Z
18.983845
82
0.460812
[ [ [ "# NumPy Basics: Arrays and Vectorized Computation", "_____no_output_____" ] ], [ [ "import numpy as np\nnp.random.seed(12345)\nimport matplotlib.pyplot as plt\nplt.rc('figure', figsize=(10, 6))\nnp.set_printoptions(precision=4, suppress=True)", "_____no_output_____" ], [ "import numpy as np\nmy_arr = np.arange(1000000)\nmy_list = list(range(1000000))", "_____no_output_____" ], [ "%time for _ in range(10): my_arr2 = my_arr * 2\n%time for _ in range(10): my_list2 = [x * 2 for x in my_list]", "_____no_output_____" ] ], [ [ "## The NumPy ndarray: A Multidimensional Array Object", "_____no_output_____" ] ], [ [ "import numpy as np\n# Generate some random data\ndata = np.random.randn(2, 3)\ndata", "_____no_output_____" ], [ "data * 10\ndata + data", "_____no_output_____" ], [ "data.shape\ndata.dtype", "_____no_output_____" ] ], [ [ "### Creating ndarrays", "_____no_output_____" ] ], [ [ "data1 = [6, 7.5, 8, 0, 1]\narr1 = np.array(data1)\narr1", "_____no_output_____" ], [ "data2 = [[1, 2, 3, 4], [5, 6, 7, 8]]\narr2 = np.array(data2)\narr2", "_____no_output_____" ], [ "arr2.ndim\narr2.shape", "_____no_output_____" ], [ "arr1.dtype\narr2.dtype", "_____no_output_____" ], [ "np.zeros(10)\nnp.zeros((3, 6))\nnp.empty((2, 3, 2))", "_____no_output_____" ], [ "np.arange(15)", "_____no_output_____" ] ], [ [ "### Data Types for ndarrays", "_____no_output_____" ] ], [ [ "arr1 = np.array([1, 2, 3], dtype=np.float64)\narr2 = np.array([1, 2, 3], dtype=np.int32)\narr1.dtype\narr2.dtype", "_____no_output_____" ], [ "arr = np.array([1, 2, 3, 4, 5])\narr.dtype\nfloat_arr = arr.astype(np.float64)\nfloat_arr.dtype", "_____no_output_____" ], [ "arr = np.array([3.7, -1.2, -2.6, 0.5, 12.9, 10.1])\narr\narr.astype(np.int32)", "_____no_output_____" ], [ "numeric_strings = np.array(['1.25', '-9.6', '42'], dtype=np.string_)\nnumeric_strings.astype(float)", "_____no_output_____" ], [ "int_array = np.arange(10)\ncalibers = np.array([.22, .270, .357, .380, .44, .50], dtype=np.float64)\nint_array.astype(calibers.dtype)", "_____no_output_____" ], [ "empty_uint32 = np.empty(8, dtype='u4')\nempty_uint32", "_____no_output_____" ] ], [ [ "### Arithmetic with NumPy Arrays", "_____no_output_____" ] ], [ [ "arr = np.array([[1., 2., 3.], [4., 5., 6.]])\narr\narr * arr\narr - arr", "_____no_output_____" ], [ "1 / arr\narr ** 0.5", "_____no_output_____" ], [ "arr2 = np.array([[0., 4., 1.], [7., 2., 12.]])\narr2\narr2 > arr", "_____no_output_____" ] ], [ [ "### Basic Indexing and Slicing", "_____no_output_____" ] ], [ [ "arr = np.arange(10)\narr\narr[5]\narr[5:8]\narr[5:8] = 12\narr", "_____no_output_____" ], [ "arr_slice = arr[5:8]\narr_slice", "_____no_output_____" ], [ "arr_slice[1] = 12345\narr", "_____no_output_____" ], [ "arr_slice[:] = 64\narr", "_____no_output_____" ], [ "arr2d = np.array([[1, 2, 3], [4, 5, 6], [7, 8, 9]])\narr2d[2]", "_____no_output_____" ], [ "arr2d[0][2]\narr2d[0, 2]", "_____no_output_____" ], [ "arr3d = np.array([[[1, 2, 3], [4, 5, 6]], [[7, 8, 9], [10, 11, 12]]])\narr3d", "_____no_output_____" ], [ "arr3d[0]", "_____no_output_____" ], [ "old_values = arr3d[0].copy()\narr3d[0] = 42\narr3d\narr3d[0] = old_values\narr3d", "_____no_output_____" ], [ "arr3d[1, 0]", "_____no_output_____" ], [ "x = arr3d[1]\nx\nx[0]", "_____no_output_____" ] ], [ [ "#### Indexing with slices", "_____no_output_____" ] ], [ [ "arr\narr[1:6]", "_____no_output_____" ], [ "arr2d\narr2d[:2]", "_____no_output_____" ], [ "arr2d[:2, 1:]", "_____no_output_____" ], [ "arr2d[1, :2]", "_____no_output_____" ], [ "arr2d[:2, 2]", "_____no_output_____" ], [ "arr2d[:, :1]", "_____no_output_____" ], [ "arr2d[:2, 1:] = 0\narr2d", "_____no_output_____" ] ], [ [ "### Boolean Indexing", "_____no_output_____" ] ], [ [ "names = np.array(['Bob', 'Joe', 'Will', 'Bob', 'Will', 'Joe', 'Joe'])\ndata = np.random.randn(7, 4)\nnames\ndata", "_____no_output_____" ], [ "names == 'Bob'", "_____no_output_____" ], [ "data[names == 'Bob']", "_____no_output_____" ], [ "data[names == 'Bob', 2:]\ndata[names == 'Bob', 3]", "_____no_output_____" ], [ "names != 'Bob'\ndata[~(names == 'Bob')]", "_____no_output_____" ], [ "cond = names == 'Bob'\ndata[~cond]", "_____no_output_____" ], [ "mask = (names == 'Bob') | (names == 'Will')\nmask\ndata[mask]", "_____no_output_____" ], [ "data[data < 0] = 0\ndata", "_____no_output_____" ], [ "data[names != 'Joe'] = 7\ndata", "_____no_output_____" ] ], [ [ "### Fancy Indexing", "_____no_output_____" ] ], [ [ "arr = np.empty((8, 4))\nfor i in range(8):\n arr[i] = i\narr", "_____no_output_____" ], [ "arr[[4, 3, 0, 6]]", "_____no_output_____" ], [ "arr[[-3, -5, -7]]", "_____no_output_____" ], [ "arr = np.arange(32).reshape((8, 4))\narr\narr[[1, 5, 7, 2], [0, 3, 1, 2]]", "_____no_output_____" ], [ "arr[[1, 5, 7, 2]][:, [0, 3, 1, 2]]", "_____no_output_____" ] ], [ [ "### Transposing Arrays and Swapping Axes", "_____no_output_____" ] ], [ [ "arr = np.arange(15).reshape((3, 5))\narr\narr.T", "_____no_output_____" ], [ "arr = np.random.randn(6, 3)\narr\nnp.dot(arr.T, arr)", "_____no_output_____" ], [ "arr = np.arange(16).reshape((2, 2, 4))\narr\narr.transpose((1, 0, 2))", "_____no_output_____" ], [ "arr\narr.swapaxes(1, 2)", "_____no_output_____" ] ], [ [ "## Universal Functions: Fast Element-Wise Array Functions", "_____no_output_____" ] ], [ [ "arr = np.arange(10)\narr\nnp.sqrt(arr)\nnp.exp(arr)", "_____no_output_____" ], [ "x = np.random.randn(8)\ny = np.random.randn(8)\nx\ny\nnp.maximum(x, y)", "_____no_output_____" ], [ "arr = np.random.randn(7) * 5\narr\nremainder, whole_part = np.modf(arr)\nremainder\nwhole_part", "_____no_output_____" ], [ "arr\nnp.sqrt(arr)\nnp.sqrt(arr, arr)\narr", "_____no_output_____" ] ], [ [ "## Array-Oriented Programming with Arrays", "_____no_output_____" ] ], [ [ "points = np.arange(-5, 5, 0.01) # 1000 equally spaced points\nxs, ys = np.meshgrid(points, points)\nys", "_____no_output_____" ], [ "z = np.sqrt(xs ** 2 + ys ** 2)\nz", "_____no_output_____" ], [ "import matplotlib.pyplot as plt\nplt.imshow(z, cmap=plt.cm.gray); plt.colorbar()\nplt.title(\"Image plot of $\\sqrt{x^2 + y^2}$ for a grid of values\")", "_____no_output_____" ], [ "plt.draw()", "_____no_output_____" ], [ "plt.close('all')", "_____no_output_____" ] ], [ [ "### Expressing Conditional Logic as Array Operations", "_____no_output_____" ] ], [ [ "xarr = np.array([1.1, 1.2, 1.3, 1.4, 1.5])\nyarr = np.array([2.1, 2.2, 2.3, 2.4, 2.5])\ncond = np.array([True, False, True, True, False])", "_____no_output_____" ], [ "result = [(x if c else y)\n for x, y, c in zip(xarr, yarr, cond)]\nresult", "_____no_output_____" ], [ "result = np.where(cond, xarr, yarr)\nresult", "_____no_output_____" ], [ "arr = np.random.randn(4, 4)\narr\narr > 0\nnp.where(arr > 0, 2, -2)", "_____no_output_____" ], [ "np.where(arr > 0, 2, arr) # set only positive values to 2", "_____no_output_____" ] ], [ [ "### Mathematical and Statistical Methods", "_____no_output_____" ] ], [ [ "arr = np.random.randn(5, 4)\narr\narr.mean()\nnp.mean(arr)\narr.sum()", "_____no_output_____" ], [ "arr.mean(axis=1)\narr.sum(axis=0)", "_____no_output_____" ], [ "arr = np.array([0, 1, 2, 3, 4, 5, 6, 7])\narr.cumsum()", "_____no_output_____" ], [ "arr = np.array([[0, 1, 2], [3, 4, 5], [6, 7, 8]])\narr\narr.cumsum(axis=0)\narr.cumprod(axis=1)", "_____no_output_____" ] ], [ [ "### Methods for Boolean Arrays", "_____no_output_____" ] ], [ [ "arr = np.random.randn(100)\n(arr > 0).sum() # Number of positive values", "_____no_output_____" ], [ "bools = np.array([False, False, True, False])\nbools.any()\nbools.all()", "_____no_output_____" ] ], [ [ "### Sorting", "_____no_output_____" ] ], [ [ "arr = np.random.randn(6)\narr\narr.sort()\narr", "_____no_output_____" ], [ "arr = np.random.randn(5, 3)\narr\narr.sort(1)\narr", "_____no_output_____" ], [ "large_arr = np.random.randn(1000)\nlarge_arr.sort()\nlarge_arr[int(0.05 * len(large_arr))] # 5% quantile", "_____no_output_____" ] ], [ [ "### Unique and Other Set Logic", "_____no_output_____" ] ], [ [ "names = np.array(['Bob', 'Joe', 'Will', 'Bob', 'Will', 'Joe', 'Joe'])\nnp.unique(names)\nints = np.array([3, 3, 3, 2, 2, 1, 1, 4, 4])\nnp.unique(ints)", "_____no_output_____" ], [ "sorted(set(names))", "_____no_output_____" ], [ "values = np.array([6, 0, 0, 3, 2, 5, 6])\nnp.in1d(values, [2, 3, 6])", "_____no_output_____" ] ], [ [ "## File Input and Output with Arrays", "_____no_output_____" ] ], [ [ "arr = np.arange(10)\nnp.save('some_array', arr)", "_____no_output_____" ], [ "np.load('some_array.npy')", "_____no_output_____" ], [ "np.savez('array_archive.npz', a=arr, b=arr)", "_____no_output_____" ], [ "arch = np.load('array_archive.npz')\narch['b']", "_____no_output_____" ], [ "np.savez_compressed('arrays_compressed.npz', a=arr, b=arr)", "_____no_output_____" ], [ "!rm some_array.npy\n!rm array_archive.npz\n!rm arrays_compressed.npz", "_____no_output_____" ] ], [ [ "## Linear Algebra", "_____no_output_____" ] ], [ [ "x = np.array([[1., 2., 3.], [4., 5., 6.]])\ny = np.array([[6., 23.], [-1, 7], [8, 9]])\nx\ny\nx.dot(y)", "_____no_output_____" ], [ "np.dot(x, y)", "_____no_output_____" ], [ "np.dot(x, np.ones(3))", "_____no_output_____" ], [ "x @ np.ones(3)", "_____no_output_____" ], [ "from numpy.linalg import inv, qr\nX = np.random.randn(5, 5)\nmat = X.T.dot(X)\ninv(mat)\nmat.dot(inv(mat))\nq, r = qr(mat)\nr", "_____no_output_____" ] ], [ [ "## Pseudorandom Number Generation", "_____no_output_____" ] ], [ [ "samples = np.random.normal(size=(4, 4))\nsamples", "_____no_output_____" ], [ "from random import normalvariate\nN = 1000000\n%timeit samples = [normalvariate(0, 1) for _ in range(N)]\n%timeit np.random.normal(size=N)", "_____no_output_____" ], [ "np.random.seed(1234)", "_____no_output_____" ], [ "rng = np.random.RandomState(1234)\nrng.randn(10)", "_____no_output_____" ] ], [ [ "## Example: Random Walks", "_____no_output_____" ] ], [ [ "import random\nposition = 0\nwalk = [position]\nsteps = 1000\nfor i in range(steps):\n step = 1 if random.randint(0, 1) else -1\n position += step\n walk.append(position)", "_____no_output_____" ], [ "plt.figure()", "_____no_output_____" ], [ "plt.plot(walk[:100])", "_____no_output_____" ], [ "np.random.seed(12345)", "_____no_output_____" ], [ "nsteps = 1000\ndraws = np.random.randint(0, 2, size=nsteps)\nsteps = np.where(draws > 0, 1, -1)\nwalk = steps.cumsum()", "_____no_output_____" ], [ "walk.min()\nwalk.max()", "_____no_output_____" ], [ "(np.abs(walk) >= 10).argmax()", "_____no_output_____" ] ], [ [ "### Simulating Many Random Walks at Once", "_____no_output_____" ] ], [ [ "nwalks = 5000\nnsteps = 1000\ndraws = np.random.randint(0, 2, size=(nwalks, nsteps)) # 0 or 1\nsteps = np.where(draws > 0, 1, -1)\nwalks = steps.cumsum(1)\nwalks", "_____no_output_____" ], [ "walks.max()\nwalks.min()", "_____no_output_____" ], [ "hits30 = (np.abs(walks) >= 30).any(1)\nhits30\nhits30.sum() # Number that hit 30 or -30", "_____no_output_____" ], [ "crossing_times = (np.abs(walks[hits30]) >= 30).argmax(1)\ncrossing_times.mean()", "_____no_output_____" ], [ "steps = np.random.normal(loc=0, scale=0.25,\n size=(nwalks, nsteps))", "_____no_output_____" ] ], [ [ "## Conclusion", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown" ]
[ [ "markdown" ], [ "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code", "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code" ], [ "markdown" ] ]
ece9f5860a8a4a18be5490fc13c6f2dadcfcddc8
826
ipynb
Jupyter Notebook
Function and Circulation/fiveTimes.ipynb
npukujui11/Automate_with_Pyhton
d1d810da53222b9642674d8dd2ed2dbe5afae326
[ "MIT" ]
null
null
null
Function and Circulation/fiveTimes.ipynb
npukujui11/Automate_with_Pyhton
d1d810da53222b9642674d8dd2ed2dbe5afae326
[ "MIT" ]
null
null
null
Function and Circulation/fiveTimes.ipynb
npukujui11/Automate_with_Pyhton
d1d810da53222b9642674d8dd2ed2dbe5afae326
[ "MIT" ]
null
null
null
20.65
137
0.529056
[ [ [ "print('My name is')\nfor i in range(5):\n print('Jimmy Five Times (' + str(i) + ')')", "My name is\nJimmy Five Times (0)\nJimmy Five Times (1)\nJimmy Five Times (2)\nJimmy Five Times (3)\nJimmy Five Times (4)\n" ] ] ]
[ "code" ]
[ [ "code" ] ]
ece9fa1ffba09a43de54a26bc916a848802e0f95
12,417
ipynb
Jupyter Notebook
Lecture7_RNN_intro.ipynb
CharlesPoletowin/YCBS-273
28678a5e693cad6d29ac3cc7a9f53acb551d4470
[ "MIT" ]
null
null
null
Lecture7_RNN_intro.ipynb
CharlesPoletowin/YCBS-273
28678a5e693cad6d29ac3cc7a9f53acb551d4470
[ "MIT" ]
null
null
null
Lecture7_RNN_intro.ipynb
CharlesPoletowin/YCBS-273
28678a5e693cad6d29ac3cc7a9f53acb551d4470
[ "MIT" ]
1
2020-07-28T14:25:16.000Z
2020-07-28T14:25:16.000Z
32.762533
242
0.472417
[ [ [ "<a href=\"https://colab.research.google.com/github/CharlesPoletowin/YCBS-273/blob/master/Lecture7_RNN_intro.ipynb\" target=\"_parent\"><img src=\"https://colab.research.google.com/assets/colab-badge.svg\" alt=\"Open In Colab\"/></a>", "_____no_output_____" ] ], [ [ "import time\nfrom sklearn.metrics import accuracy_score", "_____no_output_____" ] ], [ [ "This notebook was inspired from https://github.com/bentrevett/pytorch-sentiment-analysis. Great thanks to the authors!", "_____no_output_____" ], [ "# Data setup", "_____no_output_____" ] ], [ [ "import torch\nfrom torch.utils.data import TensorDataset\nfrom torch.utils.data import DataLoader\n\ndef get_data(seq_len, inp_dim, device, data_size=25000):\n\n data = torch.randint(low=0, high=inp_dim, size=(data_size, seq_len), out=None, device=device)\n labels = torch.abs(data[:, 0])\n\n train_data = TensorDataset(data[:int(0.7*data_size)], labels[:int(0.7*data_size)])\n valid_data = TensorDataset(data[int(0.7*data_size): int(0.85*data_size)], labels[int(0.7*data_size): int(0.85*data_size)])\n test_data = TensorDataset(data[int(0.85*data_size): int(data_size)], labels[int(0.85*data_size): int(data_size)])\n\n train_data_loader = DataLoader(train_data, batch_size=64)\n valid_data_loader = DataLoader(valid_data, batch_size=64)\n test_data_loader = DataLoader(test_data, batch_size=64)\n \n return train_data_loader, valid_data_loader, test_data_loader", "_____no_output_____" ] ], [ [ "# Model definition", "_____no_output_____" ] ], [ [ "import torch.nn as nn\n\nclass RNN(nn.Module):\n def __init__(self, input_dim, embedding_dim, hidden_dim, output_dim):\n \n super().__init__()\n \n self.embedding = nn.Embedding(input_dim, embedding_dim)\n \n self.rnn = nn.RNN(embedding_dim, hidden_dim)\n \n self.fc = nn.Linear(hidden_dim, output_dim)\n \n def forward(self, text):\n\n #text = [seq len, batch size]\n \n embedded = self.embedding(text)\n \n #embedded = [seq len, batch size, emb dim]\n \n output, hidden = self.rnn(embedded)\n \n #output = [seq len, batch size, hid dim]\n #hidden = [1, batch size, hid dim]\n \n assert torch.equal(output[-1,:,:], hidden.squeeze(0))\n \n \n \n return self.fc(hidden.squeeze(0))", "_____no_output_____" ], [ "def count_parameters(model):\n return sum(p.numel() for p in model.parameters() if p.requires_grad)", "_____no_output_____" ] ], [ [ "# Loss function", "_____no_output_____" ] ], [ [ "import torch.nn.functional as F\n\nloss_func = F.cross_entropy", "_____no_output_____" ] ], [ [ "# Optimizer", "_____no_output_____" ] ], [ [ "import torch.optim as optim", "_____no_output_____" ], [ "def evaluate(model, data_iterator, loss_func):\n \n epoch_loss = 0\n epoch_acc = 0\n \n model.eval()\n \n with torch.no_grad():\n \n for inp, label in data_iterator:\n\n predictions = model(inp.t()).squeeze(1)\n \n loss = loss_func(predictions, label)\n \n acc = accuracy_score(torch.argmax(predictions, dim=1).cpu().detach().numpy(), label.cpu().numpy())\n\n epoch_loss += loss.item()\n epoch_acc += acc\n \n return epoch_acc / len(data_iterator), epoch_loss / len(data_iterator)", "_____no_output_____" ] ], [ [ "# Training", "_____no_output_____" ] ], [ [ "def train_model(model, train_data, valid_data, loss_func, optimizer, epochs=5):\n\n for epoch in range(epochs):\n \n model.train()\n epoch_loss = 0\n epoch_acc = 0\n \n tic = time.time()\n for inp, label in train_data:\n \n predictions = model(inp.t()).squeeze(1)\n loss = loss_func(predictions, label)\n loss.backward()\n optimizer.step()\n optimizer.zero_grad()\n\n epoch_loss += loss.item()\n \n toc = time.time()\n \n train_acc, _ = evaluate(model, train_data, loss_func)\n acc, _ = evaluate(model, valid_data, loss_func)\n toe = time.time()\n print(len(train_data))\n print('Loss at epoch %d : %f, train acc : %f, valid acc : %f | train time : %d sec, eval time : %d sec' % (epoch, epoch_loss / len(train_data), train_acc, acc, toc-tic, toe - toc))", "_____no_output_____" ], [ "SEQ_LEN = 10\nINPUT_DIM = 50\nOUTPUT_DIM = INPUT_DIM\n\nEMBEDDING_DIM = 32\nHIDDEN_DIM = 256\n\nN_LAYERS = 1\nBIDIRECTIONAL = False\nDROPOUT = 0\n\ndevice = torch.device('cuda' if torch.cuda.is_available() else 'cpu')\n\ntrain_data_loader, valid_data_loader, test_data_loader = get_data(SEQ_LEN, INPUT_DIM, device=device, data_size=100000)\nmodel = RNN(INPUT_DIM, EMBEDDING_DIM, HIDDEN_DIM, OUTPUT_DIM) #, N_LAYERS, BIDIRECTIONAL, DROPOUT)\nmodel = model.to(device)\n\noptimizer = optim.Adam(model.parameters(), weight_decay=0.00001)\n\nprint(f'The model has {count_parameters(model):,} trainable parameters')\n\ntrain_model(model, train_data_loader, valid_data_loader, loss_func, optimizer, epochs=15)", "The model has 88,690 trainable parameters\n1094\nLoss at epoch 0 : 0.448821, train acc : 1.000000, valid acc : 1.000000 | train time : 4 sec, eval time : 2 sec\n1094\nLoss at epoch 1 : 0.001877, train acc : 1.000000, valid acc : 1.000000 | train time : 4 sec, eval time : 2 sec\n1094\nLoss at epoch 2 : 0.000670, train acc : 1.000000, valid acc : 1.000000 | train time : 4 sec, eval time : 2 sec\n1094\nLoss at epoch 3 : 0.000333, train acc : 1.000000, valid acc : 1.000000 | train time : 4 sec, eval time : 2 sec\n1094\nLoss at epoch 4 : 0.032066, train acc : 0.999986, valid acc : 1.000000 | train time : 4 sec, eval time : 2 sec\n1094\nLoss at epoch 5 : 0.003594, train acc : 1.000000, valid acc : 1.000000 | train time : 4 sec, eval time : 2 sec\n1094\nLoss at epoch 6 : 0.000300, train acc : 1.000000, valid acc : 1.000000 | train time : 4 sec, eval time : 2 sec\n1094\nLoss at epoch 7 : 0.000162, train acc : 1.000000, valid acc : 1.000000 | train time : 4 sec, eval time : 2 sec\n1094\nLoss at epoch 8 : 0.000113, train acc : 1.000000, valid acc : 1.000000 | train time : 4 sec, eval time : 2 sec\n1094\nLoss at epoch 9 : 0.000094, train acc : 1.000000, valid acc : 1.000000 | train time : 4 sec, eval time : 2 sec\n1094\nLoss at epoch 10 : 0.027022, train acc : 0.999871, valid acc : 0.999867 | train time : 4 sec, eval time : 2 sec\n1094\nLoss at epoch 11 : 0.000511, train acc : 1.000000, valid acc : 1.000000 | train time : 4 sec, eval time : 2 sec\n1094\nLoss at epoch 12 : 0.000166, train acc : 1.000000, valid acc : 1.000000 | train time : 4 sec, eval time : 2 sec\n1094\nLoss at epoch 13 : 0.000109, train acc : 1.000000, valid acc : 1.000000 | train time : 4 sec, eval time : 2 sec\n1094\nLoss at epoch 14 : 0.000083, train acc : 1.000000, valid acc : 1.000000 | train time : 5 sec, eval time : 2 sec\n" ], [ "", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code" ]
[ [ "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code", "code" ] ]
ecea0281b9041148b3f78fe3d7e947d010b821f6
38,183
ipynb
Jupyter Notebook
Pandas_CombiningDataFrames.ipynb
MIDAS-26/Python_NumPy
fd5ba281981b6c92883707e1469b053318c13d24
[ "Apache-2.0" ]
null
null
null
Pandas_CombiningDataFrames.ipynb
MIDAS-26/Python_NumPy
fd5ba281981b6c92883707e1469b053318c13d24
[ "Apache-2.0" ]
null
null
null
Pandas_CombiningDataFrames.ipynb
MIDAS-26/Python_NumPy
fd5ba281981b6c92883707e1469b053318c13d24
[ "Apache-2.0" ]
null
null
null
25.88678
223
0.295734
[ [ [ "# Merging, Joining, and Concatenating\n\nThere are 3 main ways of combining DataFrames together: Merging, Joining and Concatenating. In this lecture we will discuss these 3 methods with examples.\n\n____", "_____no_output_____" ], [ "### Example DataFrames", "_____no_output_____" ] ], [ [ "import pandas as pd", "_____no_output_____" ], [ "df1 = pd.DataFrame({'A': ['A0', 'A1', 'A2', 'A3'],\r\n 'B': ['B0', 'B1', 'B2', 'B3'],\r\n 'C': ['C0', 'C1', 'C2', 'C3'],\r\n 'D': ['D0', 'D1', 'D2', 'D3']},\r\n index=[0, 1, 2, 3])", "_____no_output_____" ], [ "df2 = pd.DataFrame({'A': ['A4', 'A5', 'A6', 'A7'],\r\n 'B': ['B4', 'B5', 'B6', 'B7'],\r\n 'C': ['C4', 'C5', 'C6', 'C7'],\r\n 'D': ['D4', 'D5', 'D6', 'D7']},\r\n index=[4, 5, 6, 7]) ", "_____no_output_____" ], [ "df3 = pd.DataFrame({'A': ['A8', 'A9', 'A10', 'A11'],\r\n 'B': ['B8', 'B9', 'B10', 'B11'],\r\n 'C': ['C8', 'C9', 'C10', 'C11'],\r\n 'D': ['D8', 'D9', 'D10', 'D11']},\r\n index=[8, 9, 10, 11])", "_____no_output_____" ], [ "df1", "_____no_output_____" ], [ "df2", "_____no_output_____" ], [ "df3", "_____no_output_____" ] ], [ [ "## Concatenation\n\nConcatenation basically glues together DataFrames. Keep in mind that dimensions should match along the axis you are concatenating on. You can use **pd.concat** and pass in a list of DataFrames to concatenate together:", "_____no_output_____" ] ], [ [ "pd.concat([df1,df2,df3])", "_____no_output_____" ], [ "pd.concat([df1,df2,df3],axis=1)", "_____no_output_____" ] ], [ [ "_____\n## Example DataFrames", "_____no_output_____" ] ], [ [ "left = pd.DataFrame({'key': ['K0', 'K1', 'K2', 'K3'],\r\n 'A': ['A0', 'A1', 'A2', 'A3'],\r\n 'B': ['B0', 'B1', 'B2', 'B3']})\r\n \r\nright = pd.DataFrame({'key': ['K0', 'K1', 'K2', 'K3'],\r\n 'C': ['C0', 'C1', 'C2', 'C3'],\r\n 'D': ['D0', 'D1', 'D2', 'D3']}) ", "_____no_output_____" ], [ "left", "_____no_output_____" ], [ "right", "_____no_output_____" ] ], [ [ "___", "_____no_output_____" ], [ "## Merging\n\nThe **merge** function allows you to merge DataFrames together using a similar logic as merging SQL Tables together. For example:", "_____no_output_____" ] ], [ [ "pd.merge(left,right,how='inner',on='key')", "_____no_output_____" ] ], [ [ "Or to show a more complicated example:", "_____no_output_____" ] ], [ [ "left = pd.DataFrame({'key1': ['K0', 'K0', 'K1', 'K2'],\r\n 'key2': ['K0', 'K1', 'K0', 'K1'],\r\n 'A': ['A0', 'A1', 'A2', 'A3'],\r\n 'B': ['B0', 'B1', 'B2', 'B3']})\r\n \r\nright = pd.DataFrame({'key1': ['K0', 'K1', 'K1', 'K2'],\r\n 'key2': ['K0', 'K0', 'K0', 'K0'],\r\n 'C': ['C0', 'C1', 'C2', 'C3'],\r\n 'D': ['D0', 'D1', 'D2', 'D3']})", "_____no_output_____" ], [ "pd.merge(left, right, on=['key1', 'key2'])", "_____no_output_____" ], [ "pd.merge(left, right, how='outer', on=['key1', 'key2'])", "_____no_output_____" ], [ "pd.merge(left, right, how='right', on=['key1', 'key2'])", "_____no_output_____" ], [ "pd.merge(left, right, how='left', on=['key1', 'key2'])", "_____no_output_____" ] ], [ [ "## Joining\nJoining is a convenient method for combining the columns of two potentially differently-indexed DataFrames into a single result DataFrame.", "_____no_output_____" ] ], [ [ "left = pd.DataFrame({'A': ['A0', 'A1', 'A2'],\r\n 'B': ['B0', 'B1', 'B2']},\r\n index=['K0', 'K1', 'K2']) \r\n\r\nright = pd.DataFrame({'C': ['C0', 'C2', 'C3'],\r\n 'D': ['D0', 'D2', 'D3']},\r\n index=['K0', 'K2', 'K3'])", "_____no_output_____" ], [ "left.join(right)", "_____no_output_____" ], [ "left.join(right, how='outer')", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code" ]
[ [ "markdown", "markdown" ], [ "code", "code", "code", "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code" ] ]
ecea04d20dc327aed83177f0f181dd61cadcd9a5
3,525
ipynb
Jupyter Notebook
Project5.ipynb
akashgh0sh2807/Project-Euler
de0542a59c56918408b263885dcb6d84d25339d1
[ "Apache-2.0" ]
null
null
null
Project5.ipynb
akashgh0sh2807/Project-Euler
de0542a59c56918408b263885dcb6d84d25339d1
[ "Apache-2.0" ]
null
null
null
Project5.ipynb
akashgh0sh2807/Project-Euler
de0542a59c56918408b263885dcb6d84d25339d1
[ "Apache-2.0" ]
null
null
null
23.5
119
0.455887
[ [ [ "# Project 5\n## Smallest multiple\n2520 is the smallest number that can be divided by each of the numbers from 1 to 10 without any remainder.\n\nWhat is the smallest positive number that is evenly divisible by all of the numbers from 1 to 20?", "_____no_output_____" ], [ "## To do list\n\n* Create a function to check divisibility from 11 to 20.\n* Give it a starting point.\n* Create a loop to run the function.\n* Printing the result.\n\n\n", "_____no_output_____" ], [ "## Divisibility checker\nCreated a function to check if a number is divisible by 11 to 20.", "_____no_output_____" ] ], [ [ "def dividend(number):\n for x in range(11, 21, 1):\n if number % x != 0:\n return False\n return True", "_____no_output_____" ] ], [ [ "## Starting point\nCreated a value which is the smallest number divisible by all number from 1 to 10.", "_____no_output_____" ] ], [ [ "value = 2520", "_____no_output_____" ] ], [ [ "## While loop\nCreated a loop in which if function fails, then adds 20 in it.", "_____no_output_____" ] ], [ [ "while not dividend(value):\n value += 20", "_____no_output_____" ] ], [ [ "## Printing result\nPrinting the smallest no. which is divisible by 1 to 20.", "_____no_output_____" ] ], [ [ "print(\"Smallest number which is divisible by all number from 1 to 20 is:\", value)", "Smallest number which is divisible by all number from 1 to 20 is: 232792560\n" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code" ]
[ [ "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ] ]
ecea05a83178d6a0d879bec60d2e8237517c565d
6,151
ipynb
Jupyter Notebook
content/python/basics/indexing_and_slicing_numpy_arrays.ipynb
vedraiyani/notes-1
85b86787e5bdb9c5b4160438c026391c7c8c3a48
[ "CC0-1.0" ]
1
2019-06-17T19:46:34.000Z
2019-06-17T19:46:34.000Z
content/python/basics/indexing_and_slicing_numpy_arrays.ipynb
vedraiyani/notes-1
85b86787e5bdb9c5b4160438c026391c7c8c3a48
[ "CC0-1.0" ]
null
null
null
content/python/basics/indexing_and_slicing_numpy_arrays.ipynb
vedraiyani/notes-1
85b86787e5bdb9c5b4160438c026391c7c8c3a48
[ "CC0-1.0" ]
null
null
null
23.299242
269
0.52463
[ [ [ "---\ntitle: \"Indexing And Slicing NumPy Arrays\"\nauthor: \"Chris Albon\"\ndate: 2017-12-20T11:53:49-07:00\ndescription: \"Indexing and slicing NumPy arrays in Python.\"\ntype: technical_note\ndraft: false\n---", "_____no_output_____" ] ], [ [ "## Slicing Arrays", "_____no_output_____" ], [ "### Explanation Of Broadcasting", "_____no_output_____" ], [ "Unlike many other data types, slicing an array into a new variable means that any chances to that new variable are broadcasted to the original variable. Put other way, a slice is a hotlink to the original array variable, not a seperate and independent copy of it.", "_____no_output_____" ] ], [ [ "# Import Modules\nimport numpy as np", "_____no_output_____" ], [ "# Create an array of battle casualties from the first to the last battle\nbattleDeaths = np.array([1245, 2732, 3853, 4824, 5292, 6184, 7282, 81393, 932, 10834])", "_____no_output_____" ], [ "# Divide the array of battle deaths into start, middle, and end of the war\nwarStart = battleDeaths[0:3]; print('Death from battles at the start of war:', warStart)\nwarMiddle = battleDeaths[3:7]; print('Death from battles at the middle of war:', warMiddle)\nwarEnd = battleDeaths[7:10]; print('Death from battles at the end of war:', warEnd)", "Death from battles at the start of war: [1245 2732 3853]\nDeath from battles at the middle of war: [4824 5292 6184 7282]\nDeath from battles at the end of war: [81393 932 10834]\n" ], [ "# Change the battle death numbers from the first battle\nwarStart[0] = 11101", "_____no_output_____" ], [ "# View that change reflected in the warStart slice of the battleDeaths array\nwarStart", "_____no_output_____" ], [ "# View that change reflected in (i.e. \"broadcasted to) the original battleDeaths array\nbattleDeaths", "_____no_output_____" ] ], [ [ "## Indexing Arrays", "_____no_output_____" ], [ "Note: This multidimensional array behaves like a dataframe or matrix (i.e. columns and rows)", "_____no_output_____" ] ], [ [ "# Create an array of regiment information\nregimentNames = ['Nighthawks', 'Sky Warriors', 'Rough Riders', 'New Birds']\nregimentNumber = [1, 2, 3, 4]\nregimentSize = [1092, 2039, 3011, 4099]\nregimentCommander = ['Mitchell', 'Blackthorn', 'Baker', 'Miller']\n\nregiments = np.array([regimentNames, regimentNumber, regimentSize, regimentCommander])\nregiments", "_____no_output_____" ], [ "# View the first column of the matrix\nregiments[:,0]", "_____no_output_____" ], [ "# View the second row of the matrix\nregiments[1,]", "_____no_output_____" ], [ "# View the top-right quarter of the matrix\nregiments[:2,2:]", "_____no_output_____" ] ] ]
[ "raw", "markdown", "code", "markdown", "code" ]
[ [ "raw" ], [ "markdown", "markdown", "markdown" ], [ "code", "code", "code", "code", "code", "code" ], [ "markdown", "markdown" ], [ "code", "code", "code", "code" ] ]
ecea0a34e4622e106827027ad3328e06cb2ac0c5
1,034,959
ipynb
Jupyter Notebook
Tarea7/.ipynb_checkpoints/T07_MC_Esteban_Reyes-checkpoint.ipynb
esteban-rs/Machine-Learning-I-CIMAT-2021
118ff1f4031c4ebe0ecd90b5585c953bd9113d3c
[ "MIT" ]
null
null
null
Tarea7/.ipynb_checkpoints/T07_MC_Esteban_Reyes-checkpoint.ipynb
esteban-rs/Machine-Learning-I-CIMAT-2021
118ff1f4031c4ebe0ecd90b5585c953bd9113d3c
[ "MIT" ]
null
null
null
Tarea7/.ipynb_checkpoints/T07_MC_Esteban_Reyes-checkpoint.ipynb
esteban-rs/Machine-Learning-I-CIMAT-2021
118ff1f4031c4ebe0ecd90b5585c953bd9113d3c
[ "MIT" ]
1
2022-03-19T19:58:47.000Z
2022-03-19T19:58:47.000Z
647.658949
482,012
0.939134
[ [ [ "# WGAN: Redes Generadoras Antagónicas Convolucionales Profundas con Métrica Wasserstein\n\n## Aprendizaje Profundo I\n\n### Esteban Reyes Saldaña", "_____no_output_____" ], [ "El modelo de redes profundas Redes Antagónicas Generadoras (Generative Adversarial Networks, GANs) fué propuesto por Goodfellow et al., 2014b. Las GANs son un modelo generador (de los muchos que existen) que se basa en redes neuronales profundas.\n\n\nEn esta sección reformularemos el modelo de la GAN basadas en Perceptrones Multicapa a Convolucionales (Radford and Metz, 2015). La siguiente figura ilustra (a manera de recordatorio) la arquitectura de una red convolucional. Detalles de implementación en Keras pueden encontrarse en Dumoulin and Visin (2016), notas convnets 1, notes convnets 2 y Chollet (2018)", "_____no_output_____" ], [ "El modelo general seguirá siendo el de una GAN: un generador y un discriminador. El propósito del Generador $ G $ es generar datos sintéticos $ \\tilde{x} $ que sean indistinguibles de los datos reales para el Discriminador $ D $. Por lo que el discriminador es una red que clasifica datos que se le presentan entre reales o falsos (fakes). La siguiente figura representa una GAN.", "_____no_output_____" ], [ "![gan.png](data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAABKgAAAOwCAYAAAAa/0xdAAABfGlDQ1BJQ0MgUHJvZmlsZQAAKJFjYGAqSSwoyGFhYGDIzSspCnJ3UoiIjFJgv8PAzcDDIMRgxSCemFxc4BgQ4MOAE3y7xsAIoi/rgsxK8/x506a1fP4WNq+ZclYlOrj1gQF3SmpxMgMDIweQnZxSnJwLZOcA2TrJBUUlQPYMIFu3vKQAxD4BZIsUAR0IZN8BsdMh7A8gdhKYzcQCVhMS5AxkSwDZAkkQtgaInQ5hW4DYyRmJKUC2B8guiBvAgNPDRcHcwFLXkYC7SQa5OaUwO0ChxZOaFxoMcgcQyzB4MLgwKDCYMxgwWDLoMjiWpFaUgBQ65xdUFmWmZ5QoOAJDNlXBOT+3oLQktUhHwTMvWU9HwcjA0ACkDhRnEKM/B4FNZxQ7jxDLX8jAYKnMwMDcgxBLmsbAsH0PA4PEKYSYyjwGBn5rBoZt5woSixLhDmf8xkKIX5xmbARh8zgxMLDe+///sxoDA/skBoa/E////73o//+/i4H2A+PsQA4AJHdp4IxrEg8AAAAJcEhZcwAAFiUAABYlAUlSJPAAAAIFaVRYdFhNTDpjb20uYWRvYmUueG1wAAAAAAA8eDp4bXBtZXRhIHhtbG5zOng9ImFkb2JlOm5zOm1ldGEvIiB4OnhtcHRrPSJYTVAgQ29yZSA1LjQuMCI+CiAgIDxyZGY6UkRGIHhtbG5zOnJkZj0iaHR0cDovL3d3dy53My5vcmcvMTk5OS8wMi8yMi1yZGYtc3ludGF4LW5zIyI+CiAgICAgIDxyZGY6RGVzY3JpcHRpb24gcmRmOmFib3V0PSIiCiAgICAgICAgICAgIHhtbG5zOmV4aWY9Imh0dHA6Ly9ucy5hZG9iZS5jb20vZXhpZi8xLjAvIgogICAgICAgICAgICB4bWxuczp0aWZmPSJodHRwOi8vbnMuYWRvYmUuY29tL3RpZmYvMS4wLyI+CiAgICAgICAgIDxleGlmOlBpeGVsWURpbWVuc2lvbj45NDQ8L2V4aWY6UGl4ZWxZRGltZW5zaW9uPgogICAgICAgICA8ZXhpZjpQaXhlbFhEaW1lbnNpb24+MTE5MjwvZXhpZjpQaXhlbFhEaW1lbnNpb24+CiAgICAgICAgIDx0aWZmOk9yaWVudGF0aW9uPjE8L3RpZmY6T3JpZW50YXRpb24+CiAgICAgIDwvcmRmOkRlc2NyaXB0aW9uPgogICA8L3JkZjpSREY+CjwveDp4bXBtZXRhPgoM3Q+zAABAAElEQVR4AezdCYAcRb348d9ce2+S3SSEkIuQcBgSjiRAAig3csl9mOcfFRUkBlF5ihJEfSI8H4g8EAQeKuiTh4BcghBQOZRAlCSEJFwh9wW5dpO9Z+f6169me3Zmd2ezs5k9uvdb2unq6uru6k8FSH5bVe1LmCQkBBBAAAEEEEAAAQQQQAABBBBAAAEE+kjA30fP5bEIIIAAAggggAACCCCAAAIIIIAAAghYAQJU/EZAAAEEEEAAAQQQQAABBBBAAAEEEOhTAQJUfcrPwxFAAAEEEEAAAQQQQAABBBBAAAEECFDxewABBBBAAAEEEEAAAQQQQAABBBBAoE8FCFD1KT8PRwABBBBAAAEEEEAAAQQQQAABBBAgQMXvAQQQQAABBBBAAAEEEEAAAQQQQACBPhUgQNWn/DwcAQQQQAABBBBAAAEEEEAAAQQQQIAAFb8HEEAAAQQQQAABBBBAAAEEEEAAAQT6VIAAVZ/y83AEEEAAAQQQQAABBBBAAAEEEEAAAQJU/B5AAAEEEEAAAQQQQAABBBBAAAEEEOhTAQJUfcrPwxFAAAEEEEAAAQQQQAABBBBAAAEECFDxewABBBBAAAEEEEAAAQQQQAABBBBAoE8FCFD1KT8PRwABBBBAAAEEEEAAAQQQQAABBBAgQMXvAQQQQAABBBBAAAEEEEAAAQQQQACBPhUgQNWn/DwcAQQQQAABBBBAAAEEEEAAAQQQQIAAFb8HEEAAAQQQQAABBBBAAAEEEEAAAQT6VIAAVZ/y83AEEEAAAQQQQAABBBBAAAEEEEAAAQJU/B5AAAEEEEAAAQQQQAABBBBAAAEEEOhTAQJUfcrPwxFAAAEEEEAAAQQQQAABBBBAAAEEghAggAACCCCAQP4FJp1+0x7d9N3nr9+j67kYAQQQQAABBBBAAAE3CTCCyk29RVsRQAABBBBAAAEEEEAAAQQQQAABDwoQoPJgp/JKCCCAAAIIIIAAAggggAACCCCAgJsECFC5qbdoKwIIIIAAAggggAACCCCAAAIIIOBBAdag8mCn8koIIIAAAv1PoHLcjE4bVbVuQafnOYkAAggggAACCCCAgJcFGEHl5d7l3RBAAAEEEEAAAQQQQAABBBBAAAEXCBCgckEn0UQEEEAAAQQQQAABBBBAAAEEEEDAywIEqLzcu7wbAggggAACCCCAAAIIIIAAAggg4AIBAlQu6CSaiAACCCCAAAIIIIAAAggggAACCHhZgACVl3uXd0MAAQQQQAABBBBAAAEEEEAAAQRcIECAygWdRBMRQAABBBBAAAEEEEAAAQQQQAABLwsQoPJy7/JuCCCAAAIIIIAAAggggAACCCCAgAsECFC5oJNoIgIIIIAAAggggAACCCCAAAIIIOBlAQJUXu5d3g0BBBBAAAEEEEAAAQQQQAABBBBwgQABKhd0Ek1EAAEEEEAAAQQQQAABBBBAAAEEvCxAgMrLvcu7IYAAAggggAACCCCAAAIIIIAAAi4QIEDlgk6iiQgggAACCCCAAAIIIIAAAggggICXBQhQebl3eTcEEEAAAQQQQAABBBBAAAEEEEDABQIEqFzQSTQRAQQQQAABBBBAAAEEEEAAAQQQ8LIAASov9y7vhgACCCCAAAIIIIAAAggggAACCLhAgACVCzqJJiKAAAIIIIAAAggggAACCCCAAAJeFiBA5eXe5d0QQAABBBBAAAEEEEAAAQQQQAABFwgQoHJBJ9FEBBBAAAEEEEAAAQQQQAABBBBAwMsCBKi83Lu8GwIIIIAAAggggAACCCCAAAIIIOACAQJULugkmogAAggggAACCCCAAAIIIIAAAgh4WYAAlZd7l3dDAAEEEEAAAQQQQAABBBBAAAEEXCBAgMoFnUQTEUAAAQQQQAABBBBAAAEEEEAAAS8LEKDycu/ybggggAACCCCAAAIIIIAAAggggIALBAhQuaCTaCICCCCAAAIIIIAAAggggAACCCDgZQECVF7uXd4NAQQQQAABBBBAAAEEEEAAAQQQcIEAASoXdBJNRAABBBBAAAEEEEAAAQQQQAABBLwsQIDKy73LuyGAAAIIIIAAAggggAACCCCAAAIuECBA5YJOookIIIAAAggggAACCCCAAAIIIICAlwUIUHm5d3k3BBBAAAEEEEAAAQQQQAABBBBAwAUCBKhc0Ek0EQEEEEAAAQQQQAABBBBAAAEEEPCyAAEqL/cu74YAAggggAACCCCAAAIIIIAAAgi4QIAAlQs6iSYigAACCCCAAAIIIIAAAggggAACXhYgQOXl3uXdEEAAAQQQQAABBBBAAAEEEEAAARcIEKByQSfRRAQQQAABBBBAAAEEEEAAAQQQQMDLAgSovNy7vBsCCCCAAAIIIIAAAggggAACCCDgAgECVC7oJJqIAAIIIIAAAggggAACCCCAAAIIeFmAAJWXe5d3QwABBBBAAAEEEEAAAQQQQAABBFwgQIDKBZ1EExFAAAEEEEAAAQQQQAABBBBAAAEvCxCg8nLv8m4IIIAAAggggAACCCCAAAIIIICACwQIULmgk2giAggggAACCCCAAAIIIIAAAggg4GUBAlRe7l3eDQEEEEAAAQQQQAABBBBAAAEEEHCBAAEqF3QSTUQAAQQQQAABBBBAAAEEEEAAAQS8LECAysu9y7shgAACCCCAAAIIIIAAAggggAACLhAgQOWCTqKJCCCAAAIIIIAAAggggAACCCCAgJcFCFB5uXd5NwQQQAABBBBAAAEEEEAAAQQQQMAFAgSoXNBJNBEBBBBAAAEEEEAAAQQQQAABBBDwsgABKi/3Lu+GAAIIIIAAAggggAACCCCAAAIIuECAAJULOokmIoAAAggggAACCCCAAAIIIIAAAl4WIEDl5d7l3RBAAAEEEEAAAQQQQAABBBBAAAEXCBCgckEn0UQEEEAAAQQQQAABBBBAAAEEEEDAywIEqLzcu7wbAggggAACCCCAAAIIIIAAAggg4AIBAlQu6CSaiAACCCCAAAIIIIAAAggggAACCHhZgACVl3uXd0MAAQQQQAABBBBAAAEEEEAAAQRcIECAygWdRBMRQAABBBBAAAEEEEAAAQQQQAABLwsQoPJy7/JuCCCAAAIIIIAAAggggAACCCCAgAsECFC5oJNoIgIIIIAAAggggAACCCCAAAIIIOBlAQJUXu5d3g0BBBBAAAEEEEAAAQQQQAABBBBwgQABKhd0Ek1EAAEEEEAAAQQQQAABBBBAAAEEvCxAgMrLvcu7IYAAAggggAACCCCAAAIIIIAAAi4QIEDlgk6iiQgggAACCCCAAAIIIIAAAggggICXBQhQebl3eTcEEEAAAQQQQAABBBBAAAEEEEDABQIEqFzQSTQRAQQQQAABBBBAAAEEEEAAAQQQ8LIAASov9y7vhgACCCCAAAIIIIAAAggggAACCLhAgACVCzqJJiKAAAIIIIAAAggggAACCCCAAAJeFiBA5eXe5d0QQAABBBBAAAEEEEAAAQQQQAABFwgQoHJBJ9FEBBBAAAEEEEAAAQQQQAABBBBAwMsCBKi83Lu8GwIIIIAAAggggAACCCCAAAIIIOACAQJULugkmogAAggggAACCCCAAAIIIIAAAgh4WYAAlZd7l3dDAAEEEEAAAQQQQAABBBBAAAEEXCBAgMoFnUQTEUAAAQQQQAABBBBAAAEEEEAAAS8LEKDycu/ybggggAACCCCAAAIIIIAAAggggIALBAhQuaCTaCICCCCAAAIIIIAAAggggAACCCDgZQECVF7uXd4NAQQQQAABBBBAAAEEEEAAAQQQcIEAASoXdBJNRAABBBBAAAEEEEAAAQQQQAABBLwsQIDKy73LuyGAAAIIIIAAAggggAACCCCAAAIuECBA5YJOookIIIAAAggggAACCCCAAAIIIICAlwUIUHm5d3k3BBBAAAEEEEAAAQQQQAABBBBAwAUCBKhc0Ek0EQEEEEAAAQQQQAABBBBAAAEEEPCyAAEqL/cu74YAAggggAACCCCAAAIIIIAAAgi4QIAAlQs6iSYigAACCCCAAAIIIIAAAggggAACXhYgQOXl3uXdEEAAAQQQQAABBBBAAAEEEEAAARcIEKByQSfRRAQQQAABBBBAAAEEEEAAAQQQQMDLAgSovNy7vBsCCCCAAAIIIIAAAggggAACCCDgAgECVC7oJJqIAAIIIIAAAggggAACCCCAAAIIeFmAAJWXe5d3QwABBBBAAAEEEEAAAQQQQAABBFwgQIDKBZ1EExFAAAEEEEAAAQQQQAABBBBAAAEvCxCg8nLv8m4IIIAAAggggAACCCCAAAIIIICACwQIULmgk2giAggggAACCCCAAAIIIIAAAggg4GUBAlRe7l3eDQEEEEAAAQQQQAABBBBAAAEEEHCBAAEqF3QSTUQAAQQQQAABBBBAAAEEEEAAAQS8LECAysu9y7shgAACCCCAAAIIIIAAAggggAACLhAgQOWCTqKJCCCAAAIIIIAAAggggAACCCCAgJcFCFB5uXd5NwQQQAABBBBAAAEEEEAAAQQQQMAFAgSoXNBJNBEBBBBAAAEEEEAAAQQQQAABBBDwsgABKi/3Lu+GAAIIIIAAAggggAACCCCAAAIIuECAAJULOokmIoAAAggggAACCCCAAAIIIIAAAl4WIEDl5d7l3RBAAAEEEEAAAQQQQAABBBBAAAEXCBCgckEn0UQEEEAAAQQQQAABBBBAAAEEEEDAywIEqLzcu7wbAggggAACCCCAAAIIIIAAAggg4AIBAlQu6CSaiAACCCCAAAIIIIAAAggggAACCHhZgACVl3uXd0MAAQQQQAABBBBAAAEEEEAAAQRcIECAygWdRBMRQAABBBBAAAEEEEAAAQQQQAABLwsQoPJy7/JuCCCAAAIIIIAAAggggAACCCCAgAsECFC5oJNoIgIIIIAAAggggAACCCCAAAIIIOBlAQJUXu5d3g0BBBBAAAEEEEAAAQQQQAABBBBwgUDQBW30dBPDzVGpqw9LXUNYas1Wr/uW4/R8uDkm0VhMIhHdxyUaNfloy94cJ8vNOVMWMed0b+u3PW65XutoCgUDEgz4JRTSvdmCJm/KQmaffqzleqz1QqZ+0KnT5vrCgoCUlRQmt9LkvtzsS01ZeVp5YQG/9Tz9G5uXQwABBBBAAAEEEEAAAQQQQCAHAaIEOWB1tWqTCTp9vK1GtmyvkW1VdbJjZ71U7WxI7ne17E3ZDlPWFI509bY9Uk8DZGG9c2OP3D7rTYsKQzJ0SIlUDik1+1KpHFyS3JsyPR5eWSYjhg2SvYcPkiKCWVkdOYEAAggggAACCCCAAAIIIICAFwQIUHWzFxubIrJ+c5Ws21xttiqTN/tNelwl26vrJJHo5o0HyGUamNu0ZZfdOntln09kWEWZjNunUsaNqpSx+1Qk82Y/1pQVF4U6u5xzCCCAAAIIIIAAAggggAACCCDgAgECVF3sJA06vblsvSxavkEWmv3KddskThSqi3rdr6bEOgpNt4XL12fcyG+iVxPHDZfpU8bKtMlj5Aiz12AWCQEEEEAAAQQQQAABBBBAAAEE3CVAgGo3/fXWuxvlgccXyEtvrOiRgJTfr+s/haQgFLR7zSe3oClz8iEJmPWf/H6f2fwtW0ve5xefLUs/Z+qY4E1rXa1jjk1dv1k/yjmnrx6Px+17xc06VvFEXBLxRLJMy51zTt7uE6ZOsq49n1E/ITG7TlbErIkVkWazRSJRm9fjZFnyWK/d06QBwhVrt9rt/55ZaN/rxJkHyGUXzJDDJ43e09tzPQIIIIAAAggggAACCCCAAAII9JIAAaos0LpQ+Y13z5NnX16epUb2Yp8JDpUUF0tJSXIrKiqUoqIiKSrUvdnMvlCPTT4U7Nsu0MBXQF+ll5sRiUalqSks4aYmsw5X2OaTe3NsyhsaGpNbY6OZLtm1+ZIasPrr6x/Y7awTJssNc04TXaCdhAACCCCAAAIIIIAAAggggAAC/Vugl8MS/RsjvXXX3PyEzF+8Or2oXb64uEgGlZdLeXmZDBpUbvJlJl8uZaUldvRSuwsoSAloYC5UFpTystJUWUcZHWlVV98gtbW1UlNbJzU1tSZv9ua4sbGpo0tsmQYWq82C9PffNCtrHU4ggAACCCCAAAIIIIAAAggggED/ECBA1UE/RKKxrMGpiRPGy+hRI2XY0KFSWFjQwdUU5VNApylq4E+3UW1uHA43y/YdO2Tjpo9k5ao1bc6K7UPty1DQjhFrd54CBBBAAAEEEEAAAQQQQAABBBDoHwIEqDroh4BZpylbWrV6rQmKVEllxRCpMJvdDxls143Kdg3l+RPQdayqd+6SquqdUm023e/aVZP1AZ31ZdaLOIEAAggggAACCCCAAAIIIIAAAr0qQICqA25dRDxb0vWQdpoAiW6yZl2qmq4xpVP7SktLW/YlUmamr5WaMl2PKtjHa02lGtrPM1GzNlWDWXeq3kzrq6urT+5Nvr6+3k71azJrVuWSOuvLXO5DXQQQQAABBBBAAAEEEEAAAQQQ6DkBAlR5stXAiW46uqqjpAGqggLztb6CAvt1Pt3rFEH9Up8tc87p+VQ+JIXmWBddd1PSIF64uVmam82X/DL2aWVmJJRO0bPnTd6ppwEqEgIIIIAAAggggAACCCCAAAIIDCwBAlRd7O/C8hESaaiWeKy5i1dkVtPAix0dZL5Ol2sK6pf2zDpKfp9ffH6fBMy6TD6z6fpMdjMBrFTenLfnTN1kWeY5vYdTV9uhi5DbLaH7ROuxkzflCVsn7ZwJQDnX6bmY2RJa39SNmTWforFYrq+4x/X9gQIJlVRIuHbLHt+LGyCAAAIIIIAAAggggAACCCCAQO8KEKDqondp5XgRs8WjYYlFGiTa3CAx3Uw+FtFpZ4ku3in3ahrw6YugT+4t7Y0rTIAuVGS2EgkUlEjQbJr3BwvtwwlQ9UYf8AwEEEAAAQQQQAABBBBAAAEE8itAgCpHTw2E6BYqrmi9UkcOmZFVGrxq3ZpNWctxLGLiV/HW+uSyC+gIr0DIGgcCal1g8457wIyUMkPEsl/PGQQQQAABBBBAAAEEEEAAAQQQcJ0AAap8dJkJmASCZlSP2bKlRDxmYlRRu8XT8k55vOWcrWNGTOk+WabT5XpudFa29u5ZuU4zNFMS/UGzN5uZomj3Jp8qM+dby1rzeh0JAQQQQAABBBBAAAEEEEAAAQQGlgABql7qbw28JIMvhZJrCCaho6/MpouPm19MixNmpyOydJ9epvnMclMjWWbqJfPJ+jZv7mBCSeaX5NaaN2tctZQnF2jX86bMLtbeUrcl77OjmdLKbD1GOPXSbysegwACCCCAAAIIIIAAAggggIAnBAhQuaAbbRBIAz8uaCtNRAABBBBAAAEEEEAAAQQQQAABBHIVYKhLrmLURwABBBBAAAEEEEAAAQQQQAABBBDIqwABqrxycjMEEEAAAQQQQAABBBBAAAEEEEAAgVwFCFDlKkZ9BBBAAAEEEEAAAQQQQAABBBBAAIG8ChCgyisnN0MAAQQQQAABBBBAAAEEEEAAAQQQyFWAAFWuYtRHAAEEEEAAAQQQQAABBBBAAAEEEMirAAGqvHJyMwQQQAABBBBAAAEEEEAAAQQQQACBXAUIUOUqRn0EEEAAAQQQQAABBBBAAAEEEEAAgbwKEKDKKyc3QwABBBBAAAEEEEAAAQQQQAABBBDIVYAAVa5i1EcAAQQQQAABBBBAAAEEEEAAAQQQyKsAAaq8cnIzBBBAAAEEEEAAAQQQQAABBBBAAIFcBQhQ5SpGfQQQQAABBBBAAAEEEEAAAQQQQACBvAoQoMorJzdDAAEEEEAAAQQQQAABBBBAAAEEEMhVgABVrmLURwABBBBAAAEEEEAAAQQQQAABBBDIqwABqrxycjMEEEAAAQQQQAABBBBAAAEEEEAAgVwFCFDlKkZ9BBBAAAEEEEAAAQQQQAABBBBAAIG8ChCgyisnN0MAAQQQQAABBBBAAAEEEEAAAQQQyFWAAFWuYtRHAAEEEEAAAQQQQAABBBBAAAEEEMirAAGqvHJyMwQQQAABBBBAAAEEEEAAAQQQQACBXAUIUOUqRn0EEEAAAQQQQAABBBBAAAEEEEAAgbwKEKDKKyc3QwABBBBAAAEEEEAAAQQQQAABBBDIVSCY6wXURwABBBBAAAHvCXzve9+TxYsXZ7xYUVGRHHTQQXLooYfK0UcfLePHj884zwECCCCAAAIIIIAAAvkSIECVL0nugwACCCCAgIsF7rjjDmlqamr3Bs8884wtC4VCcu2118oNN9wghYWF7erlWvDcc8/JwoULZcaMGXLqqafmejn1EUAAAQQQQAABBDwmQIDKYx3K6yCAAAIIILAnAmeeeaacd955Eg6H5YMPPpB33nlH/vnPf0pdXZ3cdNNN8vjjj8v8+fOlsrJyTx4jd911lzz//PP2WQSo9oiSixFAAAEEEEAAAU8IEKDyRDfyEggggAACCORH4LDDDpMvf/nLGTdbu3atzJo1SxYsWCDvv/++XHfddXLfffdl1OEAAQQQQAABBBBAAIE9ESBA1UW9WKRBAqGSLtamWl8IaB+REEAAAQTyL7DvvvvKP/7xD5k2bZosXbpUfvWrX8lXvvIVOeKIIzIeplP2Xn75ZVm3bp00NzfbNat0dJRe5yQdjaVBLh2dpUnr/va3v7X5kpISueiii2ze+aW2ttaO2Hr77bdlzZo19p6TJ0+W448/XkpLS51q7BFAAAEEEEAAAQRcLkCAqoMOjMXi7Up3bV5qAlTFEiqukFBRuQQLy8Xnh68dVC8WJOJRiYZrJdJktsZqiUUa2z1d+zIQ4GOV7WAoQAABBHIUCAaD8pOf/ETOPvtsicfj8thjj6UCVBpkuvDCC+2aUm1ve/3118vXv/510TWuNP3oRz+SefPmparpwuxf/OIXU8ca2NL1rjRpUOxzn/ucbNiwIXXeyUycOFF+//vfy1FHHeUUsUcAAQQQQAABBBBwsQARlg46TwMax0zdT+YvXp1xVgMgujXVJIv9gQITtCoSf7DIBq+S+WIJBM3isT5fxrUcdFMgkZBYNCzxqNo3Wf94VPdNEo81d3pT7UOCU50ScRIBBBDISeAzn/mMDBs2TLZv325HQTkXOwuel5eX2zWldIRTfX29PPjgg3aE1J133il67cknn2zPaz0NPn388ccyatQo+4VAvVdFhfkhUEtwSte5OuGEEyQWi9mRUhoYGzdunCxZskReeOEFWblypR1FtWLFChkzZozTFPYIIIAAAggggAACLhUgQJWl434+93y58e558uzLy7PUEBsgSQZJWiJWTk0TnNLgld2CZu8PiS8QMsdmb/L+QDCV9/kDzlUDap+IxyQRjxhD3aKpfEKPtTza3OprglS5prNOmCw3zDkt18uojwACCCCwG4EJEybYAJUzRU+rH3vssfLTn/5UvvrVr8qQIUNSd5g9e7ZofQ1W6agpDVBdccUVdjvjjDPsIulHHnmkPProo6lrnMx3v/tdG5zSgNhf/vIX0bWxnPTQQw/JpZdear86+MMf/lB+85vfOKfYI4AAAggggAACCLhUgABVlo4rLy2UW649R2adNU0eeHyBvPTGCol3NVBi6sXtqJ+wiPl/p8kEs3SqoM8XMIGsgMm3bObYlrccmwMxNZMjs9LyPjtSS8vN+VTe1GzJ+0y5OWmvc8p0dFcyL5LQd7LvlUjLx5N5SS9z8jr90cmbvfmfiS7Zezj5ZPBJA1Am8JTQfXKL694eR1ue2alMzif95r1OnHmAXHbBDDl80uicr+cCBBBAAIHdCzhf79uxY0eq8pQpU0S3tmnEiBEyffp0efXVV+1Iqrbnsx3rOlc6gkrTN7/5zYzglJbptD9dB+uVV16xXxUkQKUqJAQQQAABBBBAwN0CBKh2038a6Dh80oWyvbpO3ly2XhYt3yALzX7lum1dD1h19gwTHNJRQwkxo4Y6q8e5dgIakJo4brhMnzJWpk0eI0eY/bCKsnb1KEAAAQQQyJ9AdXW1vdnIkSM7vOnmzZtl48aNdvpeOByWhobkByyampo6rN9RoU7bc9IFF1zgZDP2OuVPA1Q1NTV2RJeOtCIhgAACCCCAAAIIuFeAAJV7+46WI4AAAggg0OsCa9eutc/cf//9M5791FNPiU6309FPe5pWrVqVukW29aWGDh2aqqNrWRGgSnGQQQABBBBAAAEEXClAgGo33fbWuxtzn+K3m3tyOj8COuVyxdqtdvu/ZxYKU/zy48pdEEAAgWwCuu7Uli1b7OkDDjggVU2DU/oVP13QXKcA6ugmDSwVFRXJAw88YBc0T1XuQka/GOgkHYVVWlrqHKb20aiZLt6SdNF1EgIIIIAAAggggIC7BVr/BOju98h762vrw7tdJD3bQ3V9p5LiYikpSW5FRYX2D+lFhbo3m9kXmj+0a16DKpFIxGxRabb7ZN4pS+4jEjPrN8XjZm0r82nv5NaSN+s/JWxZ+jlTxwRvtJ6ei9m9OTZ147HWc9p+v99v2+A3Xy706zpWfp8ETJlPy51zTt7udc2sZF173tRP7rW+XhuwX2DSrzCFQsGWfHJfkFam7WtqCkvYTPloMn/50Hxyb45NvqGhMbk1Nrash5VNu7Vc7/nX1z+wm7NIuq4lRkIAAQQQyI/A97///dS/k3XBcyfdfvvtNjh10EEHyYIFC2Tw4MHOKVm4cGGnASr9b1XbtN9++6WKNCg2c+bM1LGT0XJNhea/qdlGWTl12SOAAAIIIIAAAgj0fwECVFn66Jqbn5D5i1dnOZssLi4ukkHmp7bl5WUyaFC5yZeZfLmUlZbYoE2nF6edDAQCJliVVjAAsvrtwlBZUMrL2v9UPP319S8udfUNUltbKzW1dWatkVqTN3tz3NiYfT0T/fpi9a4Guf+mWem3I48AAggg0A0B/XexBqH++Mc/2qvPP/98+0U+51aLFy+22eOPPz4jOKXXbd261amWsS8oKLDHH374YUa5Hhx44IGpMl0AvW2AStezevDBB22dadOm5fTf3NSNySCAAAIIIIAAAgj0KwECVB10RyQayxqcmjhhvIweNVKGmbUvCguTf7ju4BYU5UlAR2dp4E+3UW3uGQ43y3bzFamNmz6SlavWtDkrtg+1L0NBDYeREEAAAQS6IqDrOa1Zs8aO7l29erXoguWPPPKIvP766/ZynW6nwar0NHr0aHn//ffl6aeflh//+McyfPhwu3j5F7/4xdTX+NLra95ZZF1HQr388sty3HHHybJly+TQQw+VSZMmyWmnnSbz5s2zX+sbP368XH311VJWVibr16+XL3zhC6nAlz6PhAACCCCAAAIIIOB+AV/CJPe/Rn7fQKeKTT7j5g5vqtP3Bg8eJJUVQ6TCbHY/ZLCdytbhBRTmVUCnPFbv3CVV1Tul2my637WrJjXlpO3Dlj83105hbFvOMQIIINDTApNOvynjEZXjZmQctz2oWrcgo+jd56/POO7pg2IzNX13X9r79Kc/LXfddZdMnDgxozk33nij/OAHP7BlOipYp9xt2rTJBrmcimeddZY888wzzqHoulXnnXde6lif39zcLLrmlN5Dg1UatHK+Gqg/sND1rbZv3566Zvbs2fLLX/4ydUwGAQQQQAABBBBAwL0CjKDqoO90XahsSeN5O02ARDdZsy5VTReC1al9+pPl5L7E/KS31ByX2PWo0hd8TV1Epp2ALnrbYNadqjfT+urq6pN7k6+vr7dT/Xb3l6e2N+ysL9vW5RgBBBAYyAJTp05NjZJSBw0IjR071gaj9It9p556qpx77rkdEs2dO9dOxb7zzjttgEm/9Kejna644go7GmrOnDntfpBzzjnn2FFRv/jFL+wPGRrNv/v33XdfG5zSh0yZMsUGqa688kr529/+ZqZ1N9rglLZLF2jXgNisWUzj7rBDKEQAAQQQQAABBFwowAiqLJ3W9iffWap1uVgDVAUFIbMViC4WrnudIujkU+f0vFPP7AvNsY7aclPSIF7Y/BS8uTlifxreuk8rMyOhdIqe/rRcF4e3e1M//atM+Xjn3h6BkI82cw8EEPCGQNv/jvT3EVT5UNcfIqxcudJ+pOPggw+2wabkD3Z22jUaO/phjf4AQq/Rj2to4KmjOrqWlU431FG048aNMx8hKclHc7kHAggggAACCCCAQD8SYARVFzujsHyERBqqzVfwmrt4RWY1DbzY0UHm63S5pqCZ6hAw6yh1+JU985NkHSWkP1FObp1/ZU/v4dTVdugf+u2mX/jr6lcCTQDKua7tVwJjZs2nqPnMeG8nf6BAQiUVEq5Nfv68t5/P8xBAAAEExH6xdvLkyRkU+kOWioqKjLL0Ax15rOtOdZb0v1ttpxV2Vp9zCCCAAAIIIIAAAu4TIEDVxT4rrRwvYrZ4NCyxSINEmxskppvJxyL6NbmeW8pLAz59EfTpIk0vV/NJIFRkthIJFJRI0Gya9wcLbTsIUPVyd/A4BBBAAAEEEEAAAQQQQAABBPIgQIAqR0QNhOgWKk77abAZfRQzI6s0eNW6NZuyluNYxMSv4jk+aYBW1xFegZA1DgTUusDmHfeAGSklpg4JAQQQQAABBBBAAAEEEEAAAQS8I0CAKh99aQImgaAZ1WO2bCkRj5kYVdRu8bS8Ux5vOWfrmBFTuk+W6XS5nhudla29e1au0wzNlER/0OzNZqYo2r3Jp8rM+day1rxeR0IAAQQQQAABBBBAAAEEEEAAgYElQICql/pbAy/J4Euh5BqCSejoK7PpQrPmF9PilrwJXLWWOXkdqeXkzV6DW/Z6J5+8hy03NU0oyfyS3Frz/lR5coF2PW/K7GLtLXVb8skyM6LJHNu8rccIp176bcVjEEAAAQQQQAABBBBAAAEEEPCEAAEqF3SjT6e0aeDHBW2liQgggAACCCCAAAIIIIAAAggggECuAgx1yVWM+ggggAACCCCAAAIIIIAAAggggAACeRUgQJVXTm6GAAIIIIAAAggggAACCCCAAAIIIJCrAAGqXMWojwACCCCAAAIIIIAAAggggAACCCCQVwECVHnl5GYIIIAAAggggAACCCCAAAIIIIAAArkKEKDKVYz6CCCAAAIIIIAAAggggAACCCCAAAJ5FSBAlVdOboaAuwVuvPFGOfXUU+XTn/60zJ07V6LRaJde6JlnnpEzzzzTXjtr1ixpbGzs0nVUQgABBLorcOmll0ooFJIvf/nL3b0F1yGAAAIIIIAAAgj0I4FgP2oLTUEAgT4WuPnmm6Wpqcm24sUXX5Tx48fL5Zdf3mmrdu3aJZ/73OektrY2Ve/aa6+Vww8/PHXc3zI7d+6Ue++91zbrqquukrKysv7WRNqDQL8QeOutt+T73/++RCKRVHv8fr8MGjRIZsyYIeeff77su+++qXO9mamqqrJB9B07dvTmY3kWAggggAACCCCAQA8JEKDqIVhui4AXBO65557dBqgefPDBjOCUvncikejXr79ixQq57rrrbBvPOOMMOeSQQ/p1e2kcAn0loIHq5557rsPHP/bYY/afo1tvvVWuvvrqDutQiAACCCCAAAIIIIBAVwWY4tdVKeohMIAEDjzwQPu2Onpi4cKFWd9cA1G//OUv7flDDz00az1OIICA+wV+/etfywMPPCB33XWXnHfeefaFmpub5Rvf+Ib8/e9/d/8L8gYIIIAAAggggAACfSrACKo+5efhCPRPgUsuuUR+/vOfS11dndx3330yffr0Dhv617/+VXQ0kqYrr7xSZs+e3a6eTr/5wx/+IMFgUL74xS9KYWFhRp3ly5fLSy+9JKNHj7bThTJOmgO9/pVXXhENltXU1NjRTjNnzpSDDz64bVV7/NFHH8lTTz0lH3zwgZ2WpMG2Cy64QEaNGmWnLz766KOydOnS1LVPP/20vbcWHH300bL//vvb9ixbtsyuqaXtUgO935gxY+Sb3/ymnd6k9XWq4Lx582TJkiW2nTrtacqUKaJ+xcXFWoWEgGcEPv/5z9t/jvWF5syZI7/5zW9S6z/df//98qlPfcoz78qLIIAAAggggAACCPS+AAGq3jfniQj0e4Hy8nIbZNEREw8//LANVmlZ23T33XfbIg0YZRtBtXjxYtF1njSdfPLJMmHCBJt3fnnooYfkpz/9qYwdO7ZdgEqDP1/4whdk69atTnW79/l8dhH3//iP/5BAIJA6p1OOvvSlL9nAWqrQZL73ve/JqlWrZP369fZ+6ed+8IMfpA51YfibbrpJbrvtNjut6ZprrhENwqUHtI4//nj7F3Gd1vSjH/1IGhoaUtc7mR/+8IeiC8czddARYe9FAQ1YaaBK16378MMPO3zFXAPMOmLz5ZdflnXr1omOztJ18PTDDdOmTevw/hQigAACCCCAAAIIeEeAAJV3+pI3QSCvAl/5yldEA1T19fWiQSQdIZWeNNjz7LPP2qKORk6l1+1O/o033hBdH0qnEeqoJg1UaZDsiSeekFdffdUGkoYPH26nF+n99S/CTnBK/zL71a9+VWKxmA00aTurq6tl4sSJ8tnPfla2bdsmf/vb32yzTjnlFBkyZIjNn3DCCXbv/KKjyDRVVFTYvyC/+eab9qthWvazn/3MBqemTp0qp512mgwdOtROh9TRYmrzta99TV577TWtSkLAkwI6KlJHCmqAavDgwe3eMZcAswakLrzwwg6nFF9//fXy9a9/Xe644452z6AAAQQQQAABBBBAwDsCBKi805e8CQJ5FdAvdOk0unfeeUf+53/+p12ASr+CpwGgyspKueiii1LT5PLViG9/+9s2ODV58mSZP39+alqdjsY699xz7Qgl/eqgBs502qAGj3RKoiYNSO299942r+f1L786QktHXumIsH/9619y1FFH2fMaaOpspJMGtXR9nZEjR9r6zi/f+c537NRHHVGVnnQqod5T26zt4QuB6TrkvSSgH1HQwK+mY445JuPVcg0w60LsOnpKg9C6vpX+c6/Bcf0Ig/7ze+edd8pnPvMZOwoz40EcIIAAAggggAACCHhGgEXSPdOVvAgC+Rf48pe/bG+q6z9pAMhJ4XBYfvWrX9lDXVeqqKjIOZWXva5r9frrr9t76RQ8XdvJSfqJew1eadKpfzp1T1MoFLJ7/UWDQ+lp3LhxNjiVXtaVvAa+nnzyyXbBKb1W29A2OKXlZ555pu5s0r9YkxDwioCOqNR/7nUarI5u1FGCmg444AC7Nlv6e6YHmDXwpKOg9Et/ut6cBpo0aYBZ/12i6dhjj7VTfXX04W9/+1vRALBOof3nP/8ppaWlto6OyCIhgAACCCCAAAIIeFeAEVTe7VveDIE9Frj00kvt+k26FoyOojriiCPsPXWtJ50mpyOSdCpdvlP6ejb6F9QNGzZkPOLjjz9OHWuAatKkSbZtw4YNk+3bt9tpfLo+jv6FONvaWKkbdJLRtW90JEdnSac3rV69WnRxdh1NsmbNmlR1PUdCwCsCbaf56ntddtlldsRgehC5KwFmXaPNCTDrP7/6cQHd2qYRI0bYkYo6rZeAb1sdjhFAAAEEEEAAAW8JEKDyVn/yNgjkVUADPuecc45oQCp9sXRncfQTTzzRjp7I60PNzZxRUXpfXbC8sxSPx+1p/QvyH//4R7n44ovtX3z1C2O66ZfFdO2aww47rLPbdHhOR2tlS7t27bKjRvSZztTCbHUpR8ALAjr1rra2VvTDB1VVVfaV9CMBukZbeupOgDn9+s2bN8vGjRtFA9E6wsr5EAEB33Ql8ggggAACCCCAgPcECFB5r095IwTyKqCLpWuAylks/cgjj5QFCxbYZ3Q0oqKzhzvBpM7q6DmdWuekW265xa5z5Ryn70tKSuT0009PFR133HH2a2I62uu+++6TlStX2vWjdD0tHYm1J6OpUg8xGf1Ls05xcqYh6vSk6dOniwb09C/Wuj4XCQGvCTz66KOiC6NrkEo/WqDTXx955BE56aST5PLLL0+9bncCzHrxU089JfoFzPSvZqZuSgYBBBBAAAEEEEDA8wIEqDzfxbwgAnsmcPLJJ4uu4aTTazTwo+vJaNJFyHV01e5S+vpUzqiL3V2z3377parol/U0+NPVpCOpdP2ba665xn6FcM6cOTagpIss6xo6bVNXg2bp12lgyglOaTAqfZqj/uWaAFW6FnmvCehC5rpOlAaqdWqrrkmlX9AMBAL2VbsTYNbglH7Fz/nwwtlnny1jxoyx69s98MADNtjsNUfeBwEEEEAAAQQQQCBTgABVpgdHCCDQRkCnuek6M7pgsS6WrpsmXUA9fWHyNpelDtO/frdo0aLU1/O0wrJlyzoMGumiy0763e9+l1OAyrlO262jOh5//HF54YUXRKcNOamgoMDJ2hFXuU7/0ylOTtJpT+kpfX2s9HLyCHhJQINUusbbddddZ4PXuqaUfl1TU3cCzLfffrsNTh100EE28DV48OAUlwbFdTQkCQEEEEAAAQQQQMDbAtkXWPH2e/N2CCCQg4AGqNLXY9L8FVdc0aU76CiIyspKW/c///M/5Z133hFddP2JJ54QXcNKF1tvm3TE1gUXXGCLdaqejn5KH+mki5Hr+lI/+clPUpfqVEANmqVPL9IFy5csWWLr7L///qm66UGzP/zhD3Zxcx0J0tXg0ujRo1P30q+aJRIJe/zcc8/JRRddlDpHBgEvC+ioKSdInT5qsG2AuSsGTtD3+OOPl/TglP5zr4upkxBAAAEEEEAAAQS8L0CAyvt9zBsisMcCY8eOFf2inZN0/SUt60rS6T5XXXWVrarrM+lX8XT0hQag9It7Bx98cIe3+a//+i8ZOnSoDWZ94xvfkLKyMjnkkENEA0261pMGo3R0lZM0MKVBKz1/4IEH2lFXut+yZYsUFxfb+k5d/TKYM2pKA2UasBo1apRdt8qp09le19zRe2i6/vrrZa+99rLbmWeeadfq6uxaziHgFQH9fa9T8TS9+OKL9muWmu9OgNkJ+j799NOpoHVNTY2d9jd//ny9LQkBBBBAAAEEEEDA4wIEqDzewbweArkIOGvHOPv0a9MXRHcCTunn069Jz2sdXfj42muvtYEiPdYRVDqySteW0VFVmtKn3enxhAkT5N1335XPfvazoouhNzY22imBzlSfI444Qm688UatatPXvvY1OeaYY2xeP3Ov0wkjkYhMmzZNdGRT2wXSdeSTMxVJFz3X5ATdnLa0fQ9byfyiAbI//elPqSCXBtp0JJgG21555ZXUfZzRJc517BHwmoB+REGTjiK8//77U6+Xa4D53/7t3+y1OpJRA8bjx4+3/5zpQuwkBBBAAAEEEEAAgYEhwBpUA6OfeUsEuiSgU9z0U+5DhgxpV18XRNevd2nS0Uxt09SpU+15n88npaWlGad1SqD+hVUXU37//fftwscTJ05M1dEpezrKqW3SERoPP/yw/cvv+vXrZefOnXZKkY7QaPsMDUC99tprogux69Q+nRrk/CW37X31WANXGuxau3at7Nixw476GD58uK2qXy3U0RvpU43a3kO/ZqjrcelfqHUBeX0fDVxp0nfU4Jgu2E5CwM0CzkcONGir/2y3TTqyUoPJOoLR+YCC1nECzDr6UYO5DQ0NNsCs5/TfBxpg/vd//3c9tGnu3Ln23x86nVcDxvrPpf57RqcST5o0SfRjB20DvrsLJDv3Zo8AAggggAACCCDgDgGf+alncvEUd7S311o56fSbMp5VOW5GxjEH/VOgat2CjIa9+/z1GcccIIAAAp0JbK0Oy5aqRpkyoX2QtrPrOjqX639H+uu/vzQwrIEqJ1jV9l01GKvBa5262zaIpHX1jxm7CzA799QAuQaONcCsIxL1y4B6vbZB7x8Mtv5cLRqNpgLJzhcEnfuwRwABBBBAAAEEEHCfQOuf9NzX9l5tcSzSIIFQSa8+k4flJqB9REIAAQT2RGDHrrDMvnWRHPGJSvnSWePzEqjak/b0h2s7GlGZ3i4NSjkfQkgvd/I68kpHPeq2u6RBMF2nLj3p9RUVFelFNq/Bqs6e2+4CChBAAAEEEEAAAQT6tQABqg66JxaLtyvdtXmpCVAVS6i4QkJF5qe4heXi88PXDqoXCxLxqETDtRJpMltjtcQije2ern0ZCLDUWjsYChBAoFOBN9+rEt0IVHXKxEkEEEAAAQQQQAABBPImQISlA0oNaBwzdT+Zv3h1xlkNgOjWVJMs9gcKTNCqSPzBIhu8SuaLJRAsFLNYR8a1HHRTwEztiEXDEo+qfZP1j0d13yTxWHOnN9U+JDjVKREnEUBgNwIEqnYDxGkEEEAAAQQQQAABBPIkQIAqC+TP554vN949T559eXmWGmIDJMkgSUvEyqlpglMavLJb0Oz9IfEFQubY7E3eHwim8j5/wLlqQO0T8Zgk4hFjqFs0lU/osZZHm1t9u7FM2lknTJYb5pw2oEx5WQQQ6DkBAlU9Z8udEUAAAQQQQAABBBBQAQJUWX4flJcWyi3XniOzzpomDzy+QF56Y4XEuxooMfXidtSP+XR98uv1WZ5iik0wS6cK+nwBE8gKmHzLZo5tecuxORBTMzkyKy2f/KqSlpvzdtRWso6T95lyczL5nHbnk4vXmhVoTZ2EXYi2Na/THNuWmXqJzHJTI1lm7uHkk8EnDUCZwFNC98ktrnt7HDXX6DPzm/zm/U6ceYBcdsEMOXzS6PzenLshgAACRoBAFb8NEEAAAQQQQAABBBDoGQECVD3j2vW7amDHjBpKiBk11PWrqIkAAggg0IcCBKr6EJ9HI4AAAggggAACCHhSgABVlm6trQ/vdopflksp7iMBHeH219c/sJszxU9HwpEQQACBnhIgUNVTstwXAQQQQAABBBBAYKAJ+BImDbSX7sr7Xn79w+0WSW97XXFxkQwqL5fy8jIZNKjc5MtMvlyKiwqlsSksDQ0N0tTUZLZwcgtn7sPmXDQWa3vbAXEcDASk0HxOvMhYFRWaLX2veXOupKQkZVlbWys1tXVSU1Mrtbo3x42NTZ1a6SLp9980q9M6bU8ee+Xf2hZxjAACCHRZIP2rf5NOvynjuspxMzKO2x5UrVtgi3ZXr+11HLtb4LGbjpGRQ4vc/RK0HgEEEEAAAQQQyIMAI6g6QIxEY1mDUxMnjJfRo0bKsKFDpbCwoIOrk0WhUMgGrLJWaDkRMwGqSCRitqg0230ynyxrzcfM+k3xuFnbKh5v2VryZk2ohC1LP2fqmLij1tVzMbs3x6ZuPNZ6Tpvg9/tF127ymy8X+nUdK7/Ja5mz6Tknb/e6ZlaybrI8/bxPAmbNLH335BbM2BfY8mRZwASoupocy1FtLgiHm2X7jh2ycdNHsnLVmjZnxfah9mUo2PVntbsJBQgggEAOAukjqoKF5RIN1+ZwNVURQAABBBBAAAEEEBi4AgSoOuj7gAnWZEurVq81QZEqqawYIhVms/shg20gJts1nZVroEY3M2CI1AUBDdxV79wlVdU7pdpsut+1q81XFNPu01lfplUjiwACCCCAAAIIIIAAAggggAACfShAgKoDfB1RlC3pjMidJkCim6xZl6qmU9LKSkuktLS0ZV8iZWWl5rhESoqLJRiEOoXVSSYajUpDY6PU1zdIXV19cm/y9fX1Umf2OmUyl9RZX+ZyH+oigAACXRHImOI377muXEIdBBBAAAEEEEAAAQQQMAJETfL02yC51lSTHV3V0S01QFVQEDJbgehUN93rFEEnnzqn5516Zl9ojn2dBMw6elZfl2kQL9zcLM3NEbOl79PyZiSUTtGz503eqacBqr5Mr917Ul8+nmcjgEAfC7y3tkYu/+mbObciPTCV88VtLnDWonr3+evbnOHQCwIXXT9fPtqR2w9bvPDevAMCCCCAAAIIILA7AQJUuxNqOV9YPkIiDdVmDafmLl6RWU0DL3Z0UENj5okuHOmC4gGzjpKzRlTArAFl14Fy1obKWCeq8zWi9B7OmlL66NSaVro+VVfXuGpZ36qjNa5iZs2nvlj43R8okFBJhYRrt3RBlCoIIIBAfgTyGZjKT4u4CwIIIIAAAggggAAC7hQgQNXFfiutHC9itng0LLFIg0SbGySmm8nHIvqT0J77GKIGfPoi6NNFml6uZhZiDxWZrUQCBSUSNJvm/cFC2w4CVL3cHTwOgQEqQGBqgHY8r40AAggggAACCCDQYwIEqHKk1UCIbqHiitYrzeijmBlZpcGr1q3ZlLUcxyImfhVvrU8uu4CO8AqErHEgoNYFNu+4B8xIKTN8LPv1nEEAAQR6UIDAVA/icmsEEEAAAQQQQACBAS1AgCof3W8CJoGgGdVjtmwpEY+ZGFXUbvG0vFMebzln65gRU7pPlsXMLXtudFa29u5ZuU4zNFMS/UGzN5uZomj3Jp8qM+dby1rzeh0JAQQQ6G8CBKb6W4/QHgQQQAABBBBAAAGvCRCg6qUe1cBLMvhSKLmGYBI6+spsuvi4+cW0OGF2OiJL9+llms8sNzWSZaZeMp+sb/PmDiaUZH5Jbq15s8ZVS3lygXY9b8rsYu0tdVvyPjuaKa3M1mOEUy/9tuIxCCDQwwIEpnoYmNsjgAACCCCAAAIIINAiQIDKBb8VbBBIAz8uaCtNRAABBLwgQGCq+71YU9ckC5etl7fe2yhbttfI1h11sq2qzuxr7U33GlouwyvLZK+hZTJi2CA5/BOjZfqUsTKoLPso5O63hisRQAABBBBAAAEE3CJAgMotPUU7EUAAAQR6XGDo4EK55zvTZMqEIT3+LC89INwclb/Mf1/+9NJyeeOtNWYNxuzrLq7ZuEN0S0+BgF9mHj5ezj5xspxyzEFSWMAfT9J9yCOAAAIIIIAAAgNBgD8BDoRe5h0RQAABBLoksFdFoehG6ppANBqXJ15cIvc8/JoZLZUcIdW1KzNraUDrtYWr7DZi2Esye9axcv6ph0kwyJTxTCmOEEAAAQQQQAAB7woQoPJu3/JmCCCAAAII9JjA/MWr5eZ7Xmw3GmpPH6iBrh/94nn57ZP/krmzT5Vjpu63p7fkegQQQAABBBBAAAEXCBCgckEn0UQEEEAAAQT6i0B9Y7PcfO+L8uSLb2dt0uAx22Tvw1bK0P03SfHQWimuMFtlcoRVY1W5NFabbUe57PhwlHy8ZKLs2jC83b10GuDl1z8s5516qMy98lQpLS5oV4cCBBBAAAEEEEAAAe8IEKDyTl/yJggggAACCPSowIaPd8rsHzwiqzdsb/ecYFGzTPz0Qtn/tDelfJ/MNabSK4dGhWXQqOT1+x631J6q3TxUPpx3hKx8YbpEmzIDURoIe/u9TXLPjy+RMXuzNli6JXkEEEAAAQQQQMBLAizu4KXe5F0QQAABBBDoIYENH1XLpd/+XbvglM8ft4Gps+/9b5n6pXmdBqeyNU0DWnqt3kODXHrP9KQBMX22toGEAAIIIIAAAggg4E0BAlTe7FfeCgEEEEAAgbwJ1NQ1yZVm5NTWHZkLoQ8eu1U+fev9cuTX/iRFFXV7/Dy9h95L76n3Tk/6bG2DtoWEAAIIIIAAAggg4D0BAlTe61PeCAEEEEAAgbwJxBMJ+fZ/PdVuMfRxxy6X0267Tyonbsrbs5wb6T313vqM9KTrUmlbtE0kBBBAAAEEEEAAAW8JEKDyVn/yNggggAACCORV4P5HXpfXFq7KuOd+Jy6RY779mAQKIhnl+TzQe+sz9FnpSduibSIhgAACCCCAAAIIeEuAAJW3+pO3QQABBBBAIG8C7678WO5+6O8Z99vr4LVy5FVPi/h6YRSTeYY+S5+ZnrRN2jYSAggggAACCCCAgHcECFB5py95EwQQQAABBPImEIsn5Ad3/Fmi0dYFy4srauXYax8VfyCWt+fs7kb6LH2mPttJ2iZtm7aRhAACCCCAAAIIIOANAQJU3uhH3gIBBBBAH7Xq6QAAQABJREFUAIG8Cjzx4pLMUUpmNNOMq5+SoiF7vhh6rg3VZ874xpMZo7Z0BJW2kYQAAggggAACCCDgDQECVN7oR94CAQQQQACBvAk0R2Lyy4f+kXG/8ce/LSOnfphR1psHIw9fKdqG9KRt1LaSEEAAAQQQQAABBNwvQIDK/X3IGyCAAAIIIJBXgWdfXi5btrdOqQsWRuSwz/8lr8/ozs20DdoWJ2kbta0kBBBAAAEEEEAAAfcLEKByfx/yBggggAACCORV4JHnFmfcb+Jpb0pxZWvAKuNkLx5oG7Qt6altW9PPkUcAAQQQQAABBBBwjwABKvf0FS1FAAEEEECgxwXWb66WZR9sTj3HZ9aeOuDMf6aO+zqjbdE2OUnbqm0mIYAAAggggAACCLhbgACVu/uP1iOAAAIIIJBXgX8sXJlxv+GT1knZiP4TANK2aJvSU9s2p58jjwACCCCAAAIIIOAOAQJU7ugnWokAAggggECvCCx6Z0PGc/aZviLjuD8ctG1T2zb3hzbSBgQQQAABBBBAAIHcBAhQ5eZFbQQQQAABBDwt8OHabRnvN+ygzIBVxsk+OmjbprZt7qNm8VgEEEAAAQQQQACBPRAgQLUHeFyKAAIIIICA1wQ+2laT8UqDRm3POO4PB23b1LbN/aGNtAEBBBBAAAEEEEAgNwECVLl5URsBBBBAAAHPCsTiCWlobG59P7MYeeGghtbjfpKzbUpbKF3brG0nIYAAAggggAACCLhXgACVe/uOliOAAAIIIJBfgURmkEe/lpf+xbz8Pqz7d+uwXW3a3v27cyUCCCCAAAIIIIBAXwgQoOoLdZ6JAAIIIIAAAggggAACCCCAAAIIIJASIECVoiCDAAIIIIDAABfw+TIAEgmf6NbfUoftatP2/tZm2oMAAggggAACCCDQuQABqs59Umdjkf63BkeqcWSsAH3EbwQEEEBgzwQCfp+UFBe03sQEp8I1Ja3H/SRn25QWONM2a9tJCCCAAAIIIIAAAu4VCLq36T3X8lgs3u7muzYvlUCoWELFFRIqKpdgYbn4/PC1g+rFgkQ8KtFwrUSazNZYLbFIY7una18GAsRh28FQgAACCGQRGDl8kKxa3/rlvppNw6RocH2W2n1TrG1KT9pmEgIIIIAAAggggIC7BYiwdNB/GtA4Zup+Mn/x6oyzGgDRranlC9z+gPmJbahI/MEiG7xK5oslECwUs6psxrUcdFPALHobi4YlHlX7Jusfj+q+SeKxtC9NdXB77UOCUx3AUIQAAgh0IrD/vsMzAlTb3x8je01a18kVvX9K25SetM0kBBBAAAEEEEAAAXcLEKDK0n8/n3u+3Hj3PHn25eVZaogNkCSDJC0RK6emCU5p8MpuQbP3h8QXCJljszd5fyCYyvv8AeeqAbVPxGOSiEeMoW7RVD6hx1oebW717caXmc46YbLcMOe0AWXKyyKAAAL5EJh28BiZ9/f3UrfavPAAmXT+a6nj/pDRNqUnbTMJAQQQQAABBBBAwN0CBKiy9F95aaHccu05MuusafLA4wvkpTdWSLyrgRJTL25H/YRFzP87TSaYpVMFfb6ACWQFTL5lM8e2vOXYHIipmRyZlZb32ZFaWm7Op/KmZkveZ8rNSXudU6aju5J5MYvfmk+K2/dKpOXjybyklzl5nf7o5M3e/M9El+w9nHwy+KQBKBN4Sug+ucV1b4+jLc/sVCbnk37zXifOPEAuu2CGHD5pdM7XcwECCCCAgMgnp080DC+mKLa9O07qtlRI2YjqVFlfZrQt2qb0lGxzegl5BBBAAAEEEEAAAbcJEKDaTY9poOPwSRfK9uo6eXPZelm0fIMsNPuV67Z1PWDV2TNMcEhHDSXEjBrqrB7n2gloQGriuOEyfcpYmTZ5jBxh9sMqytrVowABBBBAoOsCY/epkCkH7iPLPthsL9Iv5q3481Ey9Uvzun6THqypbUn/sqC2VdtMQgABBBBAAAEEEHC3AKtHu7v/aD0CCCCAAAJ5F7jkjKkZ91w57whprCrPKOuLA23DyhemZzy6bVszTnKAAAIIIIAAAggg4BoBRlDtpqveendj7lP8dnNPTudHQKdcrli71W7/98xCYYpffly5CwIIIKDr+P3if1+VLdtrLUY0HJIlvztFZn7ziT7F0TZEmwpSbRgxrFy0rSQEEEAAAQQQQAAB9wsQoMrSh7X14d0ukp7lUru+U0lxsZSUJLeiokIpKiqSokLdm83sC/XY5DWoEolEzBaVZrtP5p2y5D4iMbN+Uzxu1raKx1u2lrxZ/ylhy9LPmTomeKN19VzM7s2xqRuPtZ7T9vv9ftsGv/lyoV/XsfL7JGDKfFrunHPydq9rZiXr2vOmfnKv9fXagIRCoZYtmLEvsOXJMm1fU1NYwk1N0hQO23xyb45NeUNDY3JrbGxZDyubdmu53vOvr39gN2eRdF1LjIQAAgggkJtAQSggX/vcJ+WHdzyXunDNK4fKvp9aJiOnfpgq683MR29NFG1DetI2altJCCCAAAIIIIAAAu4XIECVpQ+vufkJmb94dZazyeLi4iIZVF4u5eVlMmhQucmXmXy5lJWW2KBNpxennQwEAiZYlVYwALL614lQWVDKy0o7fVsNstXVN0htba3U1NZJTU2tyZu9OW5sbMp6rX59sXpXg9x/06ysdTiBAAIIIJBd4PxTD5NH/rxY3l35cbKSWYtqwZ3nyun/fY8UDanLfmEPnGnaWSYL7jjPfGDDfPSjJU2auLdoG0kIIIAAAggggAAC3hAgQNVBP0aisazBqYkTxsvoUSNl2NChUljYOs2gg9tQlAcBHZ2lgT/dRrW5XzjcLNt37JCNmz6SlavWtDkrtg+1L0NBfrreDocCBBBAYDcCATMq9sffOFM++60HJBpNfsajsbpcXrvlYjnxxt+KPxDbzR3yczoeC9hn6rOdFAz6bdu0jSQEEMhdYNO2Rhk1vDj3C7kCAQQQQACBHhRgkfQOcANmulu2tGr1Wlmy9B1ZvGSpvL9ipWzdtt1O0ctWn/L8CuiURzVXe+0D7Qvtk2yps77Mdg3lCCCAAAJJAR2lNOdzn8rg2PrOvvKvu87JGM2UUSGfB2bElD5Ln5metE3aNhICCHRPYPatC2XF+uQac927A1chgAACCCCQfwFGUHVgqutCZUsJs87Rzp277CZr1qWq6RpTOrWvtLS0ZV8iZWb6Wqkp0/WogkGoU1idZKLRqDSYdafqzbS+urr65N7k6+vr7VS/JrNmVS6ps77M5T7URQABBAaqwOWXHC2L3tkgry1clSJY/dJhEmsOyoxvPCmBgkiqPJ+ZWHPITutb91rmIujHTp8g2iYSAgh0X6Cqplmu+vkiufmrh8j0T1R2/0ZciQACCCCAQB4FiJrkCVMDJ7pt31HV4R01QFVQEDJbgVnQNbnXKYJOPnVOzzv1zL7QHPs6CZh1+LA+LtQgXri5WZqbI2ZL36flzUgonaJnz5u8U08DVCQEEEAAgf4joIH+n333XJn1rQdlzcYdqYZp4Gjn+r1kpglSVU7clCrPR6Zq5Sh5w6w5tcvcPz2NHz3UtoUfPqSrkPeywKTTb9qj13v3+euzXt/QFJPv3P22XP+FSXLyESOy1uMEAggggAACvSVAgKqL0oXlIyTSUG2+gtfcxSsyq2ngxY4OMl+nyzUFzSLqAbOOUodf2bNf1Mv8kl5nX9nTezhf3dN2pL4KqF/46+pXAk0Ayrmu7VcCY2bNp2isd9YlSXf0BwokVFIh4dot6cXkEUAAAQTyIDCorEju/fElcul3/le27midFqQBpBe+c7lMOGWxHDLrJSmq2LPF05uqy2TpwyfKqr9MNV+hzZxuv9fQctsGbQsJAQTyIxAx68v9+DfviI6ouvikMfm5KXdBAAEEEECgmwIEqLoIV1o5XsRs8WhYYpEGiTY3mOkNZjP5WESnnSW6eKfcq2nApy+CPrm3tDeu8EkgVGS2EjOtpESCZtO8P1hoH06Aqjf6gGcggMBAFBgzskL+92efl9k/eERWb9ieItBA0soXpsvaVw+RiZ9eKPuf9qaU79M60ipVsZNM7eah8uG8I+x9ok3tP0Cy35hhco8JkI3Ze0gnd+EUAgh0RyBufvB452MrZEdNWK48d6IZud+du3ANAggggAACey5AgCpHQw2E6BYqrmi90ow+ipmRVRq8at2aTVnLccysz2HqkLogoCO8AiFrHAiodYHNO+4BM1JKTB0SAggggEDvC2iA6JE7LpOb731Rnnzx7YwGaGDp/aePttvgMdtk78NWytD9N0nx0FoprjBbZXLkVWNVuegX+Rp3lMuOD0fJx0smyq4NwzPulX5w3qmHytwrT5XS4vaBq/R65BFAYM8EHnphnR1J9d3/9wkJBohS7ZkmVyOAAAIIdEeAAFV31NpeYwImgaAZ1WO2bCkRj5kYVdRu8bS8Ux5vOWfrmBFTuk+W6XS5nhudla29e1buMzEkMyXRHzR7s5kpinZv8qkyc761rDWv15EQQAABBPqvgAaKbvrWWXLGcZPk5ntezFiXymm1Bpw6Czo59Trb63pTc2efKsdM3a+zapxDYEAJVI6b0en7Vq1b0On53Z18/o2PZGdts9x4+RQpKuTPZLvz4jwCCCCAQH4FCFDl1zPr3TTwkgy+FEqu/7lP6Ogrs+ni4+YX84yWvAlctZY5eR2p5eTNXoNb9nonn7yHLTc1TSjJ/JLcWvP+VHlygXY9b8rsmO+Wui35ZJkZ0WSObd7WY4RT1t8InEAAAQQ8IqCBo6fvuUKeeHGJ3PPwa7Jle+vaVHvyiiOGlcvsWcfK+aceZr6Ay39P9sSSaxHojsAby3fI1f+9WG6dc5gMLgt15xZcgwACCCCAQLcECFB1i613L/LplDYN/PTuY3kaAggggAACnQpoAOniM6bKOScfIn+Z/7786aXl8sZba8wU99ymtQcCfpl5+Hg5+8TJcsoxB5kv2PLHk07hOYlAngU0ELWrzixJ0ZLeXVMjs29dKLddfbiMHJp9hoBTnz0CCCCAAAL5EOBPgPlQ5B4IIIAAAggMYAENKJ11wmS71dQ1ycJl6+Wt9zaaUVU15qt/dbKtqi719T/9Gt/wyjLZa2iZjBg2SA7/xGiZPmWs8HW+AfwbiFfvc4FLThkrT72ySbZW64d/kmn9lgaZfYsGqQ6TCaPKnGL2CCCAAAII9JgAAaoeo+XGCCCAAAIIDDwBDTSdOPMAuw28t+eNEXCnQElRUC48aYw8+9omWf9xQ+oltu8Ky5yfLZKfzj5EDjsg7QNBqRpkEEAAAQQQyJ8Aizvkz5I7IYAAAggggAACCCDgSoGCkF/OOW60HDhuUEb76xqjcs0vlsirb23NKOcAAQQQQACBfAsQoMq3KPdDAAEEEEAAAQQQQMCFAgG/T06bOVIOPzBztFRzJC433L9cnnx1owvfiiYjgAACCLhFgACVW3qKdiKAAAIIIIAAAggg0MMC+qHm46buJcceOjzjSfF4Qm57+AP59TOrM8o5QAABBBBAIF8CBKjyJcl9EEAAAQQQQAABBBDwiMD0SZVy6oyR4jejqtLTA39eI7f8/j3RgBUJAQQQQACBfAoQoMqnJvdCAAEEEEAAAQQQQMAjApPGD5KzPzlKQsHMvzL86bXNcv19yyRspv6REEAAAQQQyJdA5n9t8nVX7oMAAggggAACCCCAAAKuF9h3n1K54MQxUlwYyHiXf7y9Tb51x1tS2xDNKOcAAQQQQACB7goQoOquHNchgAACCCCAAAIIIDAABPYeWiQXnzxWBpWGMt526cqdMudni2RrdTijnAMEEEAAAQS6I0CAqjtqXIMAAggggAACCCCAwAASqBhUIJecMlaGDSnMeOvVm+tk9q0LZe3H9RnlHCCAAAIIIJCrAAGqXMWojwACCCCAAAIIIIDAABQoLQ7KRSeNldF7lWS8/ZaqJvnarYtk2apdGeUcIIAAAgggkIsAAapctKiLAAIIIIAAAggggMAAFigs8Mt5x4+WiWPKMxRq6iN2Tar5S7dnlHOAAAIIIIBAVwUIUHVVinoIIIAAAggggAACCCAggYBPzjxmHzlk/yEZGk3NMZl771L58+ubM8o5QAABBBBAoCsCBKi6okQdBBBAAAEEEEAAAQQQSAn4fCInTh8hM6cMS5VpJhZPyE//9z353fNrM8o5QAABBBBAYHcCBKh2J8R5BBBAAAEEEEAAAQQQ6FDgqMlD5eQj9xafRqxaUiIh8j9Pr5Lb//CBxPWAhAACCCCAQBcECFB1AYkqCCCAAAIIIIAAAggg0LHA5AmD5axj95GgmfqXnh5/ZaP88FfLJRKNpxeTRwABBBBAoEMBAlQdslCIAAIIIIAAAggggAACXRWYMLpMzj9hjBQVBDIueXnRVvn3XyyR+qZoRjkHCCCAAAIItBUgQNVWhGMEEEAAAQQQQAABBBDIWWCf4cVy0cljpKwkmHHt4g+q5arbFsuOXeGMcg4QQAABBBBIFyBAla5BHgEEEEAAAQQQQAABBLotMHRwoVxyylipHFSQcY8PN9TKlbcuko1bGzLKOUAAAQQQQMARIEDlSLBHAAEEEEAAAQQQQACBPRYoLwnJxSZINXJYcca9PtreaIJUC+W9tTUZ5RwggAACCCCgAgSo+H2AAAIIIIAAAggggAACeRXQtaguOHG07DeqLOO+O2sjcvXti+Wf7+7IKOcAAQQQQAABAlT8HkAAAQQQQAABBBBAAIG8CwQDfjnrk/vIwfsNzrh3Yzgm3737bXnhnx9nlHOAAAIIIDCwBQhQDez+5+0RQAABBBBAAAEEEOgxAb/PJ6cctbccefDQjGdEYwn5yYPvyB/+sj6jnAMEEEAAgYErQIBq4PY9b44AAggggAACCCCAQK8IHH3IMDlh2l5i4lWplEiI3PX4h3bTPAkBBBBAYGALEKDqYv/HInxxpItUfVaNPuozeh6MAAIIIIAAAgjsVuDQAyrkjKP3kUAgLUplrtJRVDqaSkdVkRBAAAEEBq5AcOC+evY3j8Xi7U7u2rxUAqFiCRVXSKioXIKF5eLzw9cOqhcLEvGoRMO1EmkyW2O1xCKN7Z6ufRkw6x+QEEAAAQQQQAABBPpeYP+x5VJUGJBn/rFJmiOtf+bW9ah0AfWffHWKFJvzJAQQQACBgSdAhKWDPteAxjFT95P5i1dnnNUAiG5NLV/G9QcKTNCqSPzBIhu8SuaLJRAslIzxyxl34SAnATPeOxYNSzyq9k3WPx7VfZPEY82d3kr7kOBUp0ScRAABBBBAAAEEel1gzIgSueiksfLUqxulvjGaer5+2U+/8HfrnENlSHlBqpwMAggggMDAECBAlaWffz73fLnx7nny7MvLs9QQGyBJBklaIlZOTTO5XoNXdguavT8kvkDIHJu9yfsDwVTe5x+YPyFKxGOSiEeMoW7RVD6hx1oebW717caiBGedMFlumHOa0yPsEUAAAQQQQAABBPqRwPCKQrnk5LHyxCsbzcip1h86vre2Rq68dZHcfvVhMnJYcT9qMU1BAAEEEOhpAQJUWYTLSwvllmvPkVlnTZMHHl8gL72xQuJdDZSYenE76icsYv7faTLBLJ0q6PMFTCArYPItmzm25S3H5kBMzeTIrLS8z640qeXmfCpvarbkfabcnLTXOWU6uiuZF0noO9n3SrTJ65DrtmVaN7Pc1EiWmXs4+WTwSQNQJvCU0H1yi+veHpuflNlndiqT80n9SsyJMw+Qyy6YIYdPGp3z9VyAAAIIIIAAAggg0HsCg8pCcskpZiSVCVJtqWpKPXjj1gYTpFooP7vqMNl/THmqnAwCCCCAgLcFCFDtpn810HH4pAtle3WdvLlsvSxavkEWmv3Kddu6HrDq7Bka2DGjhhJiRg11Vo9z7QQ0IDVx3HCZPmWsTJs8Ro4w+2EVZe3qUYAAAggggAACCCDQPwV0vakLTxojz762WdZ9VJ9q5I5dzXLVbYvlP2cfIlMPrEiVk0EAAQQQ8K4AAaou9q0GPk7/1CS76SWNTRFZv7lK1m2uNluVyZv9Jj2ussGsHhgg1MWWequaDgRT+3H7VMq4UZUydp+KZN7sx5qy4qKQt16Yt0EAAQQQQAABBAaYQCjol3M+NUpeNAulv2+m+Dmpvikq//6LJfKDyw6WE6bt5RSzRwABBBDwqAABqm52rAZGDtxvhN3a3qKpOSofb6uRLdtrZFtVnezYWS9VOxuS+10te1O2w5Q1hSNtLx8Qx0WFIRk6pEQqh5SafalUDi5J7k2ZHg+vLJMRwwbJ3sMHSVEBv00HxG8KXhIBBBBAAAEEBqyA3++T02aOlNKioCx6vyrlEInG5Ye/Wi5VtQfIBcezhEMKhgwCCCDgQQH+5t8DnaoBlX3NaB/ddpfCJphVVx+Wuoaw1JqtXvctx065ngs3xyQai0kkovu4RKMmb/6DbffmOFluzpmyiDmne1u/7XHL9VpHUygYkKD5amEopHuzmZ9ghdoeax1True1np4PmjL9aZfu068vLAhIWUmh6BpepWafni9vOS4z5woJOu3utwbnEUAAAQQQQACBASfwycOHS0lxQP7x1rbUu+s6sLf/4QPZsSssl589QZdTJSGAAAIIeFCAAFUfd6oGanQbWlHaxy3h8QgggAACCCCAAAII9L3AtIMq7UgqnfIXj5sP8rSk3z2/1gSpmuXa/3eQBMyIKxICCCCAgLcE9BNvJAQQQAABBBBAAAEEEECg3wgctO8guy6VjthPT39+fbPMvXepNJnZBSQEEEAAAW8JZP4b31vvxtsggAACCCCAAAIIIICASwXGjSy1X/jTL/2lp/lLt8u37nhLauoH5lqu6RbkEUAAAS8JEKDyUm/yLggggAACCCCAAAIIeEhgRGWRXHLKWBlclvnl5mWrdsnsWxfJlqomD70tr4IAAggMbAECVAO7/3l7BBBAAAEEEEAAAQT6tcCQ8gK5+OSxMryiMKOd6z6uN0GqhbJ6c11GOQcIIIAAAu4UIEDlzn6j1QgggAACCCCAAAIIDBiB0uKgXHTSWBkzoiTjnbdWh2XOzxbJ0pU7M8o5QAABBBBwnwABKvf1GS1GAAEEEEAAAQQQQGDACRSE/HLu8aNl/7HlGe9e2xC1a1L94+1tGeUcIIAAAgi4S4AAlbv6i9YigAACCCCAAAIIIDBgBQJ+n5xx9D5y6AEVGQbhSFyuv2+Z/Om1TRnlHCCAAAIIuEeAAJV7+oqWIoAAAggggAACCCAw4AV8PpETpu0lRx8yLMMiHk/ILb9/Xx7485qMcg4QQAABBNwhQIDKHf1EKxFAAAEEEEAAAQQQQCBN4MiDh8opR+0tfo1YpaVfP7Nabnv4A9GAFQkBBBBAwD0CBKjc01e0FAEEEEAAAQQQQAABBNIEDt5vsJz1yX0kGMz8a82Tr26UG+5fLs1m6h8JAQQQQMAdApn/JndHm2klAggggAACCCCAAAIIIGAF9htVJhecMFqKCgIZIq++tVWu+cVbUmcWUSchgAACCPR/AQJU/b+PaCECCCCAAAIIIIAAAgh0IjByWLFcfMpYKS8JZdRasmKnzLltkWzfGc4o5wABBBBAoP8JEKDqf31CixBAAAEEEEAAAQQQQCBHgcpBBXKJCVINHVyYceWqTXVy5S0LZf2WhoxyDhBAAAEE+pcAAar+1R+0BgEEEEAAAQQQQAABBLopUFYSlItPHiP7DC/OuMPHVU0y+9aF8u6amoxyDhBAAAEE+o8AAar+0xe0BAEEEEAAAQQQQAABBPZQoNCsRXX+CWNkwuiyjDvtqovI1f+9WN5YviOjnAMEEEAAgf4hQICqf/QDrUAAAQQQQAABBBBAAIE8CQQDPjnr2FEyecLgjDs2hWNy3T1vy/NvfJRRzgECCCCAQN8LEKDq+z6gBQgggAACCCCAAAIIIJBnAZ9P5OQj95ajJg/NuHM0lpCbf/euPPTCuoxyDhBAAAEE+laAAFXf+vN0BBBAAAEEEEAAAQQQ6EGBmVOGyYnTR4hPI1YtKZEQuefJlXLnYyskrgckBBBAAIE+FyBA1eddQAMQQAABBBBAAAEEEECgJwUO2X+InHnMSNGpf+np0b9tkB//5h2JROPpxeQRQAABBP4/e+cBX0WxhfFDeicJvYVO6CAdRXpVwEJTsPtULIhdn9gLKsWuDxQ72BGRriC919A7odf0hPTw9puwm92bQhJucm+S7/x+NzszOzsz+5+IyZdzzjiAAAUqB0DnlCRAAiRAAiRAAiRAAiRAAsVLoEEtf7m5e03x9LD+CrR441l59rMwuZiUXrwL4mwkQAIkQAIWAtZ/nS23WCEBEiABEiABEiABEiABEiCB0kOgZmUfGdYrRHy93SwvtWlPpDz2/maJjE2xtLNCAiRAAiRQfAQoUBUf60LPlJaWJnXq1JFKlSrJ3LlzCz1OYR/8+OOPpUKFCtK9e/fCDsHnSIAESIAESIAESIAESMApCFQM9JQRfUIkKMDDsp79x+Lk4Ymb5OT5REs7KyRAAiRAAsVDwPqng+KZs9CzbN26VV566SVJTU01xnBxcZGAgABp1qyZ3H///VKzZk3jXmkppKeny9GjmaeMnD9/vthf6/Tp0xIZGSlHjhwp9rk5IQmQAAmQAAmQAAmQAAnYm0CAr7sM7x0is5efkDMRScbwEKcgUk16rLU0CvE32lkgARIgARIoegIlSqD6+++/Zf78+TlS+e233+Sdd96RDz/8UEaPHp1jHzY6ngD2b9OmTdKpUyfp27ev4xfEFZAACZAACZAACZAACZRJAt6erjKkZy2Zv/qUHDmVYDBAmB/C/cY/1FLaNQk22lkgARIgARIoWgIlMsQPR8T++OOP6vPZZ5/JQw89JJ6enpKcnCyPP/647Ny5s2ipcfRCE/j000/l1VdflSlTphR6DD5IAiRAAiRAAiRAAiRAAvYg4O7mIoOuryFN6wZYhkPCdCRORwJ1GgmQAAmQQPEQKFEeVGYkt99+u7kqffr0kaFDh6rwv4ULF0rz5s0t91khARIgARIgARIgARIgARIgAVsCLi7lpG+nauKjJU7ftDvSuJ2aliFvfL1LJU4f3quW0c4CCZAACZBA0RAosQKVLY6bbrrJaNq/f79R1gvR0dEC4Wrbtm0SERGh8la1aNFCRowYId7e3no3yxX9N27cKMePH5eKFStK165dVVga8l7Z2oYNG1TfXbt2SfXq1aV169bSo0cP8fX1te2aZ33Pnj2yYsUK2b59u9SvX1969+4toaGheT6Dm/aY/+LFi/Lvv/+qseLj4+Xaa6+VAQMGXHFuhOwtXbpU5clKSUmRunXrKk5t27Y1nl2/fr3s3btX9u3bp9qQU+u7775TZR8f7TSVYcOMvijExcXJ6tWrJSwsTOW+wpgQHZGoPTemBdkvy2SskAAJkAAJkAAJkAAJlHkCXVpVEl8vN1mx9ZxcupSJI0MrfPzbfomITZbRNzcQLZCDRgIkQAIkUEQESo1AFRUVZSDCaXdmmzhxorz22msCAcbWEG42Z84cadmypXErMTFRbrjhBlm2bJnRhgJyXI0dO1bludJvQEh55JFHZPr06XqTcYWwhNxYEMLyY8if9cwzzwiSouuGcMYXXnhBr2a72mt+iHCDBg1SgpA+yQcffCBNmjTJVSCDyASvNQhUtjZu3DgZM2aMfPTRR+oW+ENA0m3Lli1yzz336FWBsOXu7q7qK1eulFGjRilh0OhwudCgQQPFumPHjsatguyX8RALJEACJEACJEACJEACJGBD4JrQIPHxcpW/152R9IzLKpXWZ8aio8qT6vk7moibK1UqG2yskgAJkIBdCJQKgQqn+j377LMGkFtuucUoozBp0iQlTrVp00b69+8vFSpUUKLKzz//LMeOHVMC06pVq4xn3nzzTSVOQTB56qmn1AmB8Pz55ptv5NSpU0Y/FOD5s2jRIu2vKeXkrrvuUsm/w8PDBbmW8Ey/fv3k0KFDuXpp6YNhLU8++aSqwlsL68Q48CKCMJab2WP+S9pfhrBOeG/BrrvuOqlXr54SlNCmt9uuQU947u/vL2AOD6eEhAT59ttvlTfVxx9/rEQveIHhPvpBfDpz5ozUqFFDeWhhzKCgIEOcwvvC8wwiHTylBg8eLLVr11aeb+B88OBB5UUFL7latTJdrQuyX7bvwDoJkAAJkAAJkAAJkAAJmAmE1g4Qb083mbvqpKSkZhi3Fqw9LdFxKfLmAy3ES0uwTiMBEiABErAvgRIpUEFQufvuuxWJ2NhYgTcOhCbYhAkTpF27dqqsf4F4hTaEh5kNIgnEK4giCGnz8/NTtxHmBrv33nvl3XffVWV8eemllwShgrrhVEGIJrBp06bJfffdp9+SW2+9VYlVp0+flqlTp8oTTzxh3Mup8N5776nmypUry9q1a5VAhAZ4ccGzCeuzNXvNP3v2bEOEevrppxUTzAWvJuT6+uOPP2ynVvUuXbooPkhSHxgYaPR5+OGHVXgixCp4TUGgevDBB9UHnmkLFiyQDh06yK+//mo8oxeef/55JU5BpPvnn39UqKR+b8aMGXLnnXdKUlKSSrT+9ddfq1v53S99HF5JgARIgARIgARIgARIIC8CIVV9ZKiWd+rPZScECdN1W7szQh7/cItMfLS1lPfL9P7X7/FKAiRAAiRwdQRcru5xxz39/fffCz5//vmnIU5hNch1dOHCBcvCEDZnK06hw4033mj0Q7iabnqo2datWwXhY7p5eXlJ1apV9aryFEKlWbNmFnEKbRBg9DnXrFmDplwNnlbIjQXDKYTwXtINY8ybN0+vWq7wVIJd7fy//PKLGicgIEDgjaSbh4eH4F7fvn31JssVoYsQlMziFDpUqVLFEAnNXC0P51BB3i2IhTAIesjjZTaE/XXr1k01zZw507iV3/0yHmCBBEiABEiABEiABEiABK5AoHKQl4zoUzubELX7SKw8PHGTnIlMusIIvE0CJEACJFAQAiXSgwrhdJ9//rl6z+TkZDlx4oTMnTtXJeGGcIFk3BCXdOFCBwLPm8OHDwu8mpCz6siRI/ot5ZWjV5AYHCF/SJAOkeS5556T2267LVty7gMHDqhH4NH1/vvv648b1/Pnz6syQvzyMrOIYxbN9GfM+Zb0NlztPT9C62wTxru5uQlCI+GtlZch9BH7gPA97Ime7wvM82vm5PZDhgzJ8TGE/MGrDJ5zECLhaZXf/cpxQDaSAAmQAAmQAAmQAAmQQC4E4CU1ok+I5kl1Us5FZf1ce+zsRRn93iaZ/HhrqV8jMwojlyHYTAIkQAIkkE8CJVKgwruNHj3a8ooIxUNI3g8//CA4SQ/hYHoS7piYGOWR8/vvv+cYKmcZSKsgJHD37t1qDIgm//nPf1TycuSjQsJyXfjShSf0RWhcbpaRkRW7nlMfPTwR93ACYH7N3vMXZG59jfBgQ6J5eD9drenvg3H0/FK2YyJ/mG4QwyBQ5Xe/9Od4JQESIAESIAESIAESIIH8EvDRTvZDuB9yUh07k3Xo0oWYZHl00mZ59+GW0rpRUH6HYz8SIAESIIFcCJRYgcr2fVxdXVUycQhUMHg/QaCCNw/yHulhdsibhHxUEDbg8TNlyhTboZQAhVP5kEsJib5nzZqlck+98sorSvxCQnOYp6enumJMzJWbIdwvL4NHmG7I+5Rfs/f8BZkba4Q4hVP8kNA8ODhYJTSHsIRQSCSUR0Lzghi8tXTDviFJuq2lpaUZTUi6DoNgmJ/9Mh5kgQRIgARIgARIgARIgAQKQMDD3UVu6lZTO93vtOw7Gmc8GZ+YJk99sk1eva+ZdLumstHOAgmQAAmQQMEJZCkCBX/W6Z4we9foYguEKV2cghiFhN66wesnJ4FKv4/T7PCBhxPyHyHsDzmZJk6cqDx8kCsKXjzly5eX+++/X3+swFcka9cNYXs1a9bUq+qKEMKczJ7zQ6zTQwZt58pt/g8++ECJU40bN5Z169YpDvqzmzZtylOgysmrzJx7C3m5OnfurA9nXNEOgzhn62V1pf0yBmGBBEiABEiABEiABEiABApIwNWlnPTvXF18vM7J1n1RxtM46e/lL3fKkyMayS2aiEUjARIgARIoHIESmyQ9p9eFR49uderUUUWc8KfbLbfcohfVFeJSfiwkJESJUnpf5FuChYaGquvixYtVXitVKcSXRo0aGU8hDNHWxo0bZ9uk6vaeHwnKbZkg7A7eUDmZzhaJ3CHS6Qbx6dy5c3rVckXidVhOYpj+Privn9CHsm7IZ6Unhm/btq24uOT87Zvbfunj8EoCJEACJEACJEACJEAChSGAwIdubSpLl1aVLI9nZFySyT/tk6/mHLa0s0ICJEACJJB/Ajn/hp//5x3SEx49EInwOXnypDoBb8KECfLAAw+o9SDk684771RlszfStGnTRPcGmj9/vgwbNizb+pGvCp44CBlLTU1V9/HM7Nmzjb4NGjRQ5bFjxyqRBOFoEL+QgF03PAMvIuRHulJ+JngOde3aVT0Kjy7kz0IoGxK5I/9VTgnY0dle8yN3FwyhemCiC3BIjI7wxdzEJp0t2OgJ4ZG8HGF/+ml8amDTl2rVqqkaPKGWLl0qELPCwsJUW9OmTaV///6qjL0aP368kTMMXmxIhq6v5Y033lD9CrJf6gF+IQESIAESIAESIAESIIGrJNCuabD07VRN+10gK1UHhvxm3hGZMH2P9jNuzhEQVzktHycBEiCBUk2gxIb4mcPizDsEcerbb78VeNHAevXqJVWqVJGzZ88KPJEQloacTxBUkLfK1nD6nB4WOGbMGIEYBa8ihMDBRo4cKXooYatWreTxxx+XDz/8UNavXy/169dX4XnIx4T+kZGR6plrrrlGWrZsqcq5fUF+K5zgB7HrjjvuUB94GyFUEcnLExMTlWBlft5e8+P0vt69ews8wRDGCLbgqAt0zZo1U7m3zHOjDBZYN05FhPCEkDsIhvpztv1R79evnwqrhBjWs2dPdWog3hHvjf2A0AiWEOewXy+//LLKb4UT+3RDbjDsK6wg+6U/zysJkAAJkAAJkAAJkAAJXC2BpnUDxMfTVeatPiWpaRnGcH+tOiVRcany2n+ai6eWu4pGAiRAAiSQPwIl6l/M5s2bGyfomV8PybQbNmyoTvbbsWOHEk70+0iG/tdff0nr1q1VE4QOiFMQXZYtW2YIWRBkYBBaXnrpJSWKREdHKy8oiE0+Pj7qpL4vv/xS9dO/QPBC0nR4QUH4Ql94TEGcgpCF3FR9+/bVu+d6heACry5zuB8Em0GDBqk14N1h+jr1gew1/9y5c+Xuu+8WP7/MY3IhMkFwmzp1qnz00Uc5zv3iiy8qDzHkg4LgFB4ernJDPfroo/LZZ5/l+MxNN92kRD09MTyENwhbuljYokULwR4OHDhQiVfwsMKeIZwPua5+/PFH+fzzz/XXL/B+GQ+yQAIkQAIkQAIkQAIkQAJXSaBOdV8Z0rOWeGtCldlWhp2XJz/aKnEXsw74Md9nmQRIgARIIDuBclooWpnxP4Wnz9GjR5VXFIQrGAQSiDEBAQEWOhBGcAodxJFKlSpJ7dq1Rc+fZOloqsTHx6uE6vAGwvi2SbxNXfMsYk6IPfC60ueEl1FCQoIEBQXl+qw95ofQtGvXLgkMDDTEO0yIUDoIUTihz9aQGwqswAzCH8QmfFtB4MNJe+bT+fRn8S54BoIbRLmc+mA8hE1if8AfImFuVpj9ym0stpMACZCAPQg0HfC2ZZjg2p0sddtK5NF1lqbdC3LOP2jpxEqJIzBs3Go5HZFkrPu3t6+TahWy/7/V6MBCmSZQVP+OdBm9xML1idsz86paGlkpEIGo2BSZteyExCZkpgjRH65X3U8mjWktlYM89SZeSYAESIAEciFQYkP8cnmfPJvhHaXnQNI7ent7K08dva5f4bED4cTs0aTfy+0K7yPkUbpag7ilC2j6WBCqdLFKb7O92mN+iEs5hSOak6DbzgvRSvfw0u/BQyovMQ1ebwhRzMuwB3q+r7z64V5h9utKY/I+CZAACZAACZAACZAACeSHQFCAh4zoE6JEqgvRycYjh0/Fy8MTN8nkx1tLnaq+RjsLJEACJEAC2QmUqBC/7MtnCwmQAAmQAAmQAAmQAAmQAAk4noCvt5sM6xUiNStbvf7PRibJIxM3y45DMY5fJFdAAiRAAk5MgAKVE28Ol0YCJEACJEACJEACJEACJFByCHh6uMgt3WtKg1r+lkUj9A85qVZvzzr4x9KBFRIgARIgAaFAxW8CEiABEiABEiABEiABEiABErATAVfXcnLjddWlZcNAy4hJKeny4pTtMm/NKUs7KyRAAiRAApkEKFDxO4EESIAESIAESIAESIAESIAE7EhAS8cqPdtVkc4tMg9m0odOz7gk7/6wR75fEK438UoCJEACJHCZAAUqfiuQAAmQAAmQAAmQAAmQAAmQQBEQ6Ni8gvTuUFVwgJBuOEP9i9mH5IOf90lG2TlQXX99XkmABEggVwIUqHJFwxskQAIkQAIkQAIkQAIkQAIkcHUEmtcvLwO7VBc3LfTPbDOXnZBXp+2U1LQMczPLJEACJFBmCVCgKrNbzxcnARIgARIgARIgARIgARIoDgL1a/rJrT1qiZeHq2W6pZvPydOfbJOEpDRLOyskQAIkUBYJUKAqi7vOdyYBEiABEiABEiABEiABEihWAtUrecuw3rXEz8fNMu+WfVHy2OQtEhGTbGlnhQRIgATKGgEKVGVtx/m+JEACJEACJEACJEACJEACDiFQobynjOgTIsEBHpb5DxyPk9ETN8uJcxct7ayQAAmQQFkiQIHKTrudnJwsZ86ckZSUFDuNyGFIgARIgARIgARIgARIgARKGwF/H3cZrolU1Sp6W17t9IVETaTaJHvCYy3trJAACZBAWSFQLAJVWlqa1KlTRypVqiRz584tEraXtBMwwsPDJSkpqUjGz2vQs2fPSkiI9j+ZatXkueeey6sr7zkBgd27d4uPj4/4+vrKgQMHnGBFXAIJkAAJkAAJkAAJkEBZIoBcVEN61pR6Nfwsrx0dlyqPf7BF1u+OsLSzQgIkQAJlgUCxCFTp6ely9OhRuXDhgpw/f75IuHbp0kXq1q0r/fv3L5Lx8xr0ySeflHPnzkmtWrXkpZdeyqsr7zkBgYSEBElMTJSLFy8KyjQSIAESIAESIAESIAESKG4Cbq4uMvD66tKsXnnL1InJ6fL8Z2GyaP0ZSzsrJEACJFDaCRSLQFUcEHVPmBMnThTHdMYcixYtkp9++kk8PDzk999/l4oVKxr3WCABEiABEiABEiABEiABEiCB3Ai4lCsnfTpWlQ7NKli6pKVfkre+3SU//3PM0s4KCZAACZRmAtYjJErwm7755puycuVKGTx4cLG+xfPPP6/m++ijj6RDhw7FOjcnIwESIAESIAESIAESIAESKPkErm1ZUXy9XGXZlnOiZS5RhuunMw/IhdhkefTWhqJpWTQSIAESKNUESo1A9dBDDwk+xW0LFiwQ5NhCeB+NBEiABEiABEiABEiABEiABApDoFWjIPHxcpOF605LuuZBpRu8qKJiU+S/dzUVN1eqVDoXXkmABEofAbsLVHv27JEVK1bI9u3bpX79+tK7d28JDQ29IrkNGzbIxo0bZdeuXVK9enVp3bq19OjRQyWyvuLDWofly5dLWFiYtGnTRpCPSrd///1XduzYIX379pUmTZrIqlWr5J9//pGoqChp3ry5jBo1ypjj2LFj8sMPP8jx48eladOm0qdPH/WMPpb5umnTJlm6dKnKrYWT+5D/CnO0bdvW3M0oZ2RkyLx582Tbtm3qtD+EAiJxvItLZpSlp6enDB8+3KjjwYiICFm2bJls3bpVYmNjpWXLltK5c2dp1qyZMS4KuId1I+n3PffcI5GRkfL333/L6tWrpWrVqtKxY0e1D5aHTJX8zqM/cvr0afnzzz9l3759kpqaqvZ3yJAhUqNGDb1LjlfbdWIvfv31VwGbm266SfHWHyzomqKjo2XhwoWKL54NCAiQFi1ayIgRI8Tb23pCij5HXteCfD8Wlkde8/MeCZAACZAACZAACZBA2SPQMMRfvDxdZc7Kk5KSmmEAQD4qJFB/66EW4q3dp5EACZBAaSRgV4Hqww8/lGeeeUZT/NMNVuU0X9QXXnjBqNsW4uLi5JFHHpHp06fb3lLCx2+//aaEhmw3bRomTJgg8+fPV0KHWaCaPHmyakfy8iNHjsiMGTMsT06aNEkJanj2scceU8mz9Q4Qjb777jslcuhtSPY+dOhQgUBla+PGjZMxY8YIwv3MdvLkSbWuzZs3m5uzlbt3764EJdyA2HL33Xer5OvmjuD54osvyuuvvy6urpn/c4JQhLWj3rBhQ7n11luzPYexvvrqK+MZfcyCzINnsB/33XefxMfH60OoK/b40KFD6iRDyw1TxbxOCFujR49W4hS64GQ9CIKwgq5p4sSJ8tprr6mk52oA05dXX31V5syZo8Q9U3OuxYJ+P14Nj1wXwRskQAIkQAIkQAIkQAJllkCtKj4yrFeI/Ln8hCQkphkccLIfTvib+GgrCfT3MNpZIAESIIHSQsBuAtXPP/8sOM0OBu8gnKYXHh6uvHjeeeedXHkNGzZMkGgcwstdd90lnTp1Us99+umnykOnX79+SvgojBeMedK33npLVcuXLy8DBw6UnTt3Ko8rJFe/9tprlXiFDhC3atasqYSY5ORkFTZ44403ip9f5hGwELIgTvn7+8stt9yivLBwEty3336rvKk+/vhjGTRokMVj6d577xWIU3iHp556So0FXvD4giF3FTyvgoODVX3t2rVyww03aPHnl5TgBHEJ8/3xxx/KU+ztt9+WSpUqydixY1V//QuEwZ49ewo8uuAdhnGx3jNnziihDXty22236d2loPPAM0kXp7BehFRiTswxd+5c5ZVWrVo1Y/zcCngGoiQ8p9q1a6e8sJBkHlbQNeEZiIw4kQ/ec3jHChUqqD0CY3jFYS54a+XHCvL9aC8e+VkX+5AACZAACZAACZAACZQdApWCPGVE7xD5Y9kJzXMqxXjxPeGxMnriZvng8dZSrWLBowSMgVggARIgASckYDeB6r333lOvV7lyZSUy1KtXT9URogbBxtbjBjcRhgZxCjZt2jQlfqiK9gVeQBCrED41depUeeKJJ/Rbhb5CDEF4X2BgoEB8QugbRCJ4Vrm7uytRCqFmsG7dusnDDz8sMTExSnzRhR0IWO+++64SZzCObuiLkEaIVfAAQmgj7OzZs2pOlPEed955J4qC5Oq1a9dW4YRVqlSRzz//XLXjC7zQIE5BZEKYHsLVYPCSuvnmm5VH0Pjx45UHEry8zAbPpClTphj5uPRwRfCHiKa/B54p6DwIwdT3EYIUwgdh8ISCZ1lISIiq5+cL8nbB8wlrMFtB14Rnn332WSV0wQPNbAg5hHgFhli3LjKa+5jLBf1+tCcP8zpYJgESIAESIAESIAESIIEAP3cZ0UfzpNJEqrORSQaQE+cuaiLVJpn0WGtpWMvfaGeBBEiABEo6gcwESFf5FgjdQm4l2OOPPy66OIU6RAPkXsrJIJjAkFMJnjlmg/ePLjisWbPGfKtQ5QYNGihBTBeVIOwgX5NuWIsuTqEN9yBawcI1TzDdkNcI4pI+jt4OkQkCGAxijW7IqaUbclTpBo8xeHLBwE+3/fv3i/6+r7zyiiFO4T7yVemCzrlz55Rnmf6cfkWYpTlZPJK3Q9SCmddVmHl0HhgLoo/ZILbhnfJrI0eONN5Ff6Ywa8KzYKJ/r+hj4QrPN93M76632V4L+v1oTx62a2GdBEiABEiABEiABEiABJBvamivWlK7mq8FRkRMijw2eYts2RdlaWeFBEiABEoyAbt4UJl/+TeLAjoYeCrlZAivg8Fb6P3338/W5fz586oNuY2u1iCCBQUFWYZBYnPdrr/+er2orl5eXiqM7tSpUyrpuOXm5QrunThxQoXQwSMLYWawpKSsv3CY54Q3GZJ264Zk5jCEFOqmM0F9/fr1ysNKv4crwvV0Axckczdb165dzVVVrqMlY4fpPFEuzDzt27dX4ZsXLlxQnlgIyYQg2apVKwxZIHv66aez9S/MmsyDgPvhw4eV1x2S4MMzTjfznuhttld9/vx+P9qTh+1aWCcBEiABEiABEiABEiABEHB3c5GbutaQv7VE6Xu1ED/dEpLS5OlPtskr9zaTHm0r6828kgAJkECJJWAXgQp5fnTDCXz5NV14QoLsnAQLfRzkKioKsw2Ps50jt/s4wQ7Jt3FS4ZUMAhLyMiFUEYnNkX8JYYKzZ89WIYV43iwq6UzQjgTveVl+uSB/FczcvzDzINTw999/V6cNwoPr66+/Vh+sH4nhcfJifk0/vdDcvzBrwvMIw0QIKNamhyCax81vWZ8/v9+P9uSR3zWyHwmQAAmQAAmQAAmQQNkj4OJSTvp3ria+Xm6yeW/mH7lBITUtQ16dtlMi4xrJkO5Zf/Que4T4xiRAAqWBgF0EKnNoFxJ059d0AQiCjTnczvZ5hPs5i0Gcwil+EJqQ1Hzw4MGCMDp4XH3zzTdy8OBBy1IRBob7EKj27Nkjo0aNstyH9xFCBnXTmaCOkwn1xOn6ff3q4+MjAwYM0KsFvhZ2HuTmgqfRF198oXJq4X1XrFih8oXB46sw3lT64guzJniuIaG8HhaJ7yWEWiJRP7zbkI8rv6bPX5Dvx6Lkkd91sx8JkAAJkAAJkAAJkEDZIHD9NZXEx9tVVm7NjDTBW2do0Sgf/LxPImKS5YHB9bW0G2WDBd+SBEig9BGwi0CFZNS6Qbwwh6yhHSFTORlyVSFkDSfr3X///Tl1cbq2Dz74QIlTjRs3lnXr1qm164vE6X62AhXyb23YsEGQPH748OHqNLno6GiVUB0nFCJEThdGMI45f1ePHj2MvFb6HPa6Xs088BxC3iecSPjVV1/Jo48+qpLO4wRD1AtrhVkThCldnDInh8ca4OFWEIGqsN+PRcWjsBz5HAmQAAmQAAmQAAmQQOkl0LZxsPKkQshfRkbW71nfLwjXRKoUee6OxuKqeVzRSIAESKCkEbBLkvRGjRoZ740wK1sbN26cbZOqh4aGquvixYuVh1GOnZysccuWLWpFSMoNYU03hM8h7M3WcNIbrEKFCupEua1bt6rcSHhnnD5nFqfQz8zy+++/R1ORmD3mQZjeAw88ID179lRrRE6uq7HCrEnfD8x7yy23WKY35+uy3MilcrXfj/bmkcsy2UwCJEACJEACJEACJFDGCTSuE6DyUiE/ldnmrTklL07ZLkkp6eZmlkmABEigRBCw/otWyCXD80TPowSPlRkzZkhaWpogUfV//vOfHBOgY6qxY8eqk+kQpgVxAQmudYPXFTySIOLkJ9eT/lxRX3XvMOSQ0pOOx8bGqrA/25PtsJYmTZqoJSG8DyIVwhUh6AwaNEidXIg8U/o46IjT8IYMGaKemTp1qsAryZw7CkyR++mtt95SfQr7pTDzIOQQnm56ribMjUTk+gmODRs2LOxy1HOFWZO+Hxhg2rRphrfe/PnzZdiwYQVaT0G/H4uaR4EWz84kQAIkQAIkQAIkQAJligBO9sMJfzjpz2yrt1+QJz/aKrEJqeZmlkmABEjA6QnYRaDCW77yyivKGwhiyh133CHIvVS1alUV8oXE6ebT7HQqyFeEEDcY8hfVr19f5WtCO3II4ZS0SZMmyc6dO/VHHH4dOXKkWgNySiH5OU4CxFpnzZqV49puu+02adOmjbqXkJAg8KhaunSpzJ07V+WsQqgcxjCLW++9954Ss5DPC6KJn5+ftGzZUiAAYS6IRPbwriroPBCmII5hHfA2Qq4nXM+ePSve3t52CdMs6Jp69eolVapUUXzhqSzWDXsAAEAASURBVIdQSnxwmiR4F8QK+v1YHDwKsn72JQESIAESIAESIAESKFsEqgR7yYg+IVLez93y4jsOxcjDEzfL2cis08UtHVghARIgASckYDeBCkIBvFbMYVqurq7KUwieUM2bN1evD+HKbMjp9PPPP6vcS0i2jsTW8JiKjIxUIg3EmL59+5ofybHs4eGh2m1D5nJrR2e9L+a1XZf5vj4G2l588UUjNA+J0sPDw9U4yMP02WefoYtlLHj1IAwNc3z66adKiFu+fLnMnDnTyD8FIWXMmDHqWXyBUIeT5CBuIRl6YmKi7Nixw8hvBeHuzTffNPqb16e/k3FTK+j39at+r6DzPPLII3Ldddepx/fv3y+bN2+W1NRUadu2rdr7KyVIN8+f0zoxcEHXBMHur7/+Mk4QvHDhgvJIa9asmSxbtkxCQkLUes37m9c6CvL9eLU81ML4hQRIgARIgARIgARIgASugkCgv4cM7x0ilYI8LaMcPZOgiVSb5PCpeEs7KyRAAiTgrATKaaF0WZn17LRKiAQQbuD1o4sB8AaCEJOTJ5U+bXx8vBw7dkwl3IbwgNPv8msIKUSoHfJCQRjTLbd2/X5MTIy4ubmJr6+v3mRcEXp48eJFCQwMVAKTcUMrJCUlKcEIHmMQQzAnUCIBur+/vxoT/SHYIXH8Qw89lGPCbghw8EqCgIL5zCci4nmMCSYYF30QBpfTWuPi4tSz8LayNX1dELtyE4byOw/GhniI0D68u+5BZjtnbvW81mn7TEHWhGfh1Xb06FFp0KCB8jRDG8Q9iGhIZG62/Kwjv9+PV8PDvCaWSYAESheBpgPetrxQcO1OlrptJfLoOkvT7gU552+0dGKlxBEYNm61nI7I8mj47e3rpFoFrxL3Hlxw8RAoqn9HuoxeYnmBJ27PzAtraWSlxBFISc2QOStPyvGzFy1r9/dxk/ceaSUtGwRa2lkhARIgAWcjYJdT/GxfCuISPmaDUKWLVeZ2cxniStOmTc1N+S5DZAoODs7WP7d2vaM50bnepl8h5uQm6Hh5eRleYXp/iEu2AhxEE1hu8yA8DgbPIVtxCu1ogyiFT14GUSw3y2ldtn3zOw+eA+ecWNuOmVM9r3Xa9i/ImvAsQi7xMRtCD/GxtfysI7/fj1fDw3ZdrJMACZAACZAACZAACZBAYQh4uLvIzd1rysK1p+XAsThjiLiLaSon1Wv/aS7Xt6pktLNAAiRAAs5GoEgEKmd7SUeuB3maEGqGXFrw7kE4HLyOcOLfTz/9JKtWrVLLe/DBBx25TM5NAiRAAiRAAiRAAiRAAiRQwgm4upSTG66tLsu8zknY/ijjbZI176pxU3fIMyNDZXCXGkY7CyRAAiTgTAQoUBXxbnz11VfSr18/FQ74yy+/CD5mg1cZEsU/+eST5maWSYAESIAESIAESKDICNx5550qB+hdd92lDrQpsok4MAmQQLET0AIwpEfbyuLr5SprtBP9dMvIuCQTpu+ViJgUuffGunozryRAAiTgNAQoUBXxVtSrV0/27dsnixcvVif1nTp1SuVuQn6tOnXqSP/+/dVph0W8DA5PAiRAAiRAAiRQwgjg4Jh3331X1qxZI4cPH1b5LZs0aSI333yzym2ZU97J/L4i8iciT2dERER+H2E/EiCBEkagQ7MK4uvtJks2nJUMU9rhr+YclsjYFHlyRCNx0TyuaCRAAiTgLAQoUBXDTri4uKiTCPNzGmExLIdTkAAJkAAJkAAJODmBdevWSY8ePdShLOalIjUAPjghGakCSprh0JcpU6aoZT/22GNyNSJbSXt3rpcEHEGgWb3y4u3pKvPXnNZE6QxjCbOWn1Ai1av3NRPkrqKRAAmQgDMQ4L9GzrALXAMJkAAJkAAJkAAJXCaQnp4uw4cPV+IUBJyJEycKBKtFixbJmDFj1Gm+UVFZuWVKErj9+/fLf//7X/WBVxiNBEig6AnUq+EnQ3rUFC+PrJPOMevyrefkqU+2SryWRJ1GAiRAAs5AgB5UzrALXAMJkAAJkAAJkAAJXCawZ88eOX78uKqNHz9eiVI6HHhjjxs3Ts6cOaM38UoCJEACVyRQraK3DO8TIrOWnpC4i6lG/237o+XRyZtl8pjWUjHQ02hngQRIgAQcQYAClSOoc04SIAESIAESIAESyIXAgQMHjDs4+dfWqlSpIvjkZBs2bJCNGzfKrl27pHr16tK6dWsVKujr65tT91zbkJsKpxBv3bpVYmNjpWXLltK5c2dp1qxZrs+cPHlSlixZIjt37pTk5GRBvqxbb71VKleurLzBfv31V9m+fbvx/OzZs9X4aLj22mulYcOGxr24uDiVuzMsLEyOHDmiTkBu3ry5dO/eXXmQGR21Atb3ww8/qPZ77rlHhUBiroyMDLnpppukT58+5u4sk0CZJRAc4CEjIFItO6ElSk82OBw6GS+jJ2yS98deIyFVfIx2FkiABEiguAlQoCpu4pyPBEiABEiABEiABPIgUL9+fePu1KlTZcCAAeLqag3NMTpcLkDQeeSRR2T69Om2tyQ0NFR+++03adGiRbZ7OTUsXLhQ7r77bjl37pzldjntaLAXX3xRXn/99Wzr+fLLL+Xpp58WrMNsTz31lHz77bdSu3ZtNab53iuvvGJUMe7bb7+t6itXrpRRo0YZXmRGJ63QoEED9Y4dO3Y0mnEYDfJZgVFqaqqMHj1aiVPosHv3bgpUBikWSEDEz8dNhveuJbNXnJRT5xMNJGcik+ThiZtk4qOtpWndAKOdBRIgARIoTgIUqIqTNuciARIgARIgARIggSsQgOcRTvoNDw+XuXPnSr9+/eSTTz5RHkm5PTps2DCVowoi0l133SWdOnVSz3/66afqNGGMcejQIfH29s5tCNW+du1aueGGG+SSduIXPJogVPn7+8sff/why5cvVyJSpUqVZOzYscY48+bNkwcffFDVq1WrJoMHD1YnDs6fP195PyEpOgSy2267Tc6fP6+8rNAZnk2BgYHqOSSEh61evVp5fCEPF7y+MBbErW3btqn3O3jwoPKiQi4rnIhsNjwDkQ6eU+3atVNilYeHh7kLyyRAAhoBTy0X1a09asmCNafk0Il4g0lMfKo8/uEWefOBFtK5eQWjnQUSIAESKC4CFKiKizTnIQESIAESIAESIIF8EHB3d5fvvvtOeU5dvHhRCToIbxsyZIi8/PLL2Tyh/v77byXeYOhp06bJfffdZ8yCEDuIVadPnxZ4Yz3xxBPGvZwKzzzzjBKnMB/EooCATE8KeCjdfPPNMmfOHEFeLHgpeXp6aqeCpSnPKYyFcEScMIjQQhjuvfPOOyq8r0KFCurUQYQg6t5PkyZNUqGDqvPlL88//7xAaKpYsaL8888/KkRRvz9jxgy588471XivvvqqfP311/ot44o5kVQe70EjARLInYCbazkZ2KWGLNl4RnYeijE6JiWny3//FybP39FEBnSuZrSzQAIkQALFQYCn+BUHZc5BAiRAAiRAAiRAAgUg0LVrV3VyX/v27dVT8AqCF1KrVq2UlxByPOmGEDoY8kOZxSm0dejQQXkcobxmzRpccjV4Jel9EH6ni1N4wMXFxRB9EPoHbywYQugQYgd79tlnDXEKdTc3NyWo6aF7aMvLkJ8KohgMQhryZ5kNYX/dunVTTTNnzjTfMsojR4401mk0skACJJAjAc3hUnp3qCodbbyl0tIvyfjvd8uMRUdzfI6NJEACJFBUBOhBVVRkOS4JkAAJkAAJkAAJXAUB5IyCxxGSib/77rtKsELo3f/+9z8VvocQOpieVB333n///WwzIqwOpotK2TpcbtDHQXX9+vXZckCZTw7EWE2bNhWE3OnWv39/vVioKwQy3eAtlpMh5A/J25EY/cKFC8rTytwPebBoJEACBSPQuUVF8fVyk6WbzykPSjyt/XMi/5t1UCJik+WxoQ3FBWoWjQRIgASKmAAFqiIGzOFJgARIgARIgARI4GoI4CQ6fH7//Xe5//77lTizYMEC+fPPP1XYnS48wZspL4EGXlh5mT4O+kyePDmvrkYScrNApYf25flgHjfN89vml9IfQ6igbhDMEApoNnh60UiABApOoGXDQPHxcpWFa08LPKh0+3XJcYmMTZFxdzcVdzf+96Vz4ZUESKBoCFCgKhquHJUESIAESIAESIAE7Epg6NCh4uXlJYMGDVLjIhwOeaGQCwrWpUsXueeee1Q5py8I98vL9HHQZ8KECRIcHJxjdx8fH5UfCzeRlF23xMREYy16W0GuCAnUDSGMSJJua8gxpRuSt9NIgATsR6BBLX+52dNV5qw8KckpWYL24o1nJVpLoD7+oZZKxLLfjByJBEiABKwEsn4SsLazRgIkQAIkQAIkQAIk4GQEcPKdbnFxcapYr149gTdR+fLllYeVfr+gV4yjG07Vw0l4VzIkRtft8OHD0qZNG716xautR5d5fuS16ty5c7Yx9HxXENNy87LK9hAbSIAE8k2gZmUfGdYrRGYtOyEJiVmC8KY9kfLY+5tl0mOtJTiAp2PmGyg7kgAJFIgA/TQLhIudSYAESIAESIAESKBoCSBZuDnfk3m2jRs3GlUkRYeFhoaq6+LFi9VpfapSiC+NGjUynvr++++Ncl6FBg0aGLf1ZO1Gg1aAiGZes4dH1i+25pxXeEZ/D5RzOqEvKSlJ9Dnatm2rErejL40ESMC+BCoGesqIPiESZCNE7T8WJw9P3CQnzyfad0KORgIkQAKXCVCg4rcCCZAACZAACZAACTgRgaeeekrgTYR8UmFhYZKSkiIIeZs3b57gJDsYhB7dm2rs2LFKrEGfW265ReDJpBsSp2/atEmdsAfhKy+rXbu26MnJp06dKh9//LGRawrPRUVFKeHorbfeMobBSXudOnVS9SlTpqgE7lgHQvGQ3B2nDuJEPt2qVcs6tv7nn39WY54+fVoJcki6ridanzZtmowfP17i4+PVo8eOHVNhhThBEPbGG2+oK7+QAAkUDYEAX3cZ3jtEqlbwskwAcQoiFcQqGgmQAAnYmwAFKnsT5XgkQAIkQAIkQAIkcBUEKleuLMjnhBP5IAB5e3sL8j4NHDhQINTAcK9x48aqDBHo8ccfV2Wcvle/fn0V/oZ2JBFv3769TJo0SXbu3Kn65PXlvffeEyQihygG4cvPz09atmwpDRs2VGMhSbutdxVOGETIXWpqqjzyyCPqGeSPQn6sI0eOWHJJValSRb0T1vDHH38IBKsaNWoIBDEYcl8FBQWp8rhx41TYYqVKlQTiGU7vgz388MPSq1cvVeYXEiCBoiPgreWjGtKzltStbs0Hh6TpCPdD2B+NBEiABOxJgAKVPWlyLBIgARIgARIgARK4SgKfffaZ8njST8VDriZ8kJAcYtPff/8tjz76qGWWDz74QOCRBM8r9Dtx4oTAYyoyMlIJThCW+vbtazyjh9qZE6PjJsQtnAZ42223KVEMQtmOHTtEP60P87/55pvGOCh069ZN1q5dq4QnzA3vKQhcENbuvfdemT59uqU/vKP0fFPwtoKFhISoa4sWLdR8EOPwPN77woULykMMgtyPP/4on3/+ueqrf9HfBXXb99H78EoCJFA4Aji5b9D1NaRp3QDLABeT0uXZz8Jk8aazlnZWSIAESOBqCJTTXL+zzhG9mpH4LAmQAAmQAAmQgEGg6YC3jTIKwbUzw6AsjaZK5NF1pprI7gXjLHVWSgeBYeNWy+mIJONlfnv7OqlmE0Jj3NQK8JhCPiqINRCP4El1JUNYHJ6D+AMPqpySiUNEio2NVR5Krq6uOQ6JHxExTnR0tLi7uysvppxO1jM/jLn37NmjhCLktMKpgzkZxg4PD5eIiAg1LrykbA3iFMIV4ZkFD6q83h25riCOweOrNFlR/TvSZfQSC6Ynbs/MY2ZpZIUEbAisCjsvm3ZbvaZctP/uHhvaUIb3qmXTm1USIAESKDgBnuJXcGZ8ggRIgARIgARIgASKhQA8i3TvovxOCJEG+ZzyMjc3NwkODs6rixJ8IAzhk1/D3PCyupJBTMIJgOZTAG2fcXFxEXMSdtv75rq/v7+5yjIJkEAREOjSqpL4ernJiq3nRHdxyNAKH/+2XyJik2X0zQ20fzeKYGIOSQIkUGYIMMSvzGw1X5QESIAESIAESIAESIAESIAECk/gmtAg6d+5mri6WJWoGYuOyvjvd0taOoNzCk+XT5IACVCg4vcACZAACZAACZAACZAACZAACZBAvgiE1g6Qm7rVFA9366+SC9aelv/+L0ySktPzNQ47kQAJkIAtAeu/KrZ3WScBEiABEiABEiABEiABEiABEiABE4GQqj4yVMs75eNlzWG3dmeEPP7hFomJTzX1ZpEESIAE8keAAlX+OLEXCZAACZAACZAACZAACZAACZDAZQKVg7xkRJ/aUt7P3cJk95FYeXjiJjkTmXUghKUDKyRAAiSQCwEKVLmAYTMJkAAJkAAJkAAJkAAJkAAJkEDuBCBOjegTIhCrzHbs7EUZ/d4mOXQy3tzMMgmQAAnkSYACVZ54eJMESIAESIAESIAESIAESIAESCA3Aj7ayX4I90PYn9kuxCTLo5M2y7b9UeZmlkmABEggVwIUqHJFwxskQAIkQAIkQAIkQAIkQAIkQAJXIoCE6UicHlrb39I1PjFNnvpkmyzfes7SzgoJkAAJ5ESAAlVOVNhGAiRAAiRAAiRAAiRAAiRAAiSQbwKuLuWkf+fqck1okOWZlNQMefnLnZY2VkiABEggJwIUqHKiwjYSIAESIAESIAESIAESIAESIIECEShXTqRbm8rSpVUly3MZGZcsdVZIgARIICcCFKhyosI2EiABEiABEiABEiABEiABEiCBQhFo1zS4UM/xIRIggbJNgAJV2d5/vj0JkAAJkAAJkAAJkAAJkAAJkAAJkAAJOJwABSqHbwEXQAIkQAIkQAIkQAIkQAIkQAIkQAIkQAJlmwAFqrK9/3x7EiABEiABEiABEiABEiABEiABEiABEnA4ATeHr4ALIAESIAESIAESIIEyTCAjI0Pi4hMkOiZOomPj1TUmNk5iULdt09rRBgss7y/lAzI/KAcG+GW2XW43t/n7+YqLC/8uWYa/zfjqJEACJEACJOD0BChQOf0WcYEkQAIkQAIkQAKli4B2mlVqtEjyOflkyn7Ztn2nEqgK+o5nzkUIPvkxCFSd27eU5k0aSIumjaRe7RoUrPIDjn1IgARIgARIgASKjQAFqmJDzYlIgARIgARIgATKKoFjJ87Ipm27JOrISikXd0okI0WhWLmmeIjAQ+vvpWvVBzP6+fpIm5ZNpG3rptKudTMJqVm1eBbCWUiABEiABEiABEggFwIUqHIBw2YSIAESIAESIAESKCyBS5cuyd4D4bJ89SZZvmaTnDh1trBDFclz8QkXZcXazeqDCWpWryLdrm0n3a5rJ40b1pFy5coVybwclARIgARIgARIgARyI0CBKjcybCcBEiABEiABEiCBAhKAp9K8v1fKXwuXyfGTZ/L9tKeHhwQFanmkygeoPFJB+jUws67ySWltQcg1pX1gyEUVpfJUxWaWozOvme16OVaiouMkOSXTYyu3BUFAm/H7PPWpVaOqDO7fXW7se70gNJBGAiRAAiRwZQJNB7x95U6loMfuBeNKwVvwFZyVAAUqZ90ZrosESIAESIAESKDEEDh3PlJ++XORzFm4XBKTkq68bhcvEc9Kcsfg9nJd+6YqN9SVH7L2qOrlKVWrVLQ25lLbueegbN+1X3Zo1517DhiJ1nPqDmHts69+lq9n/CmD+neTETf3k8qVgnPqyjYSIAESIAESIAESsBsBClR2Q8mBSIAESIAESCB3ApFH1+V+k3dKLAF4Hk3/ba4s+neNpKWl5/oeHh7u0qpZI9l71k9iMyqJuAepvoNvvE6qVdDEqiI2JEfHB4bww8PhJ2TD1p2ycctOCdOEq5SU1GwrgND2qya6/TF3sfTrea3cMWygCgXM1pENJEACJEACJEACJGAHAhSo7ACRQ5AACZAACZAACZQtAqfOnJdvf5qtko6np+csTLm7u0nHti2lV9cO0qXjNeKleTwNG7daYiPy4WFVhDiRX6p+3Vrqc/utAyQpKVlWrd8qS1ZskPWbt0tqappldghvCFtcuGSN9O3eWe4ZeZNUr6qJbDQSIAESIAESIAESsCMBClR2hMmhSIAESIAESIAESjeBSC3P03eaMIUcU7l5TFWqECRDBveWgf26SXl/P6cHAuGsd7dO6hMTFy9zFy2XmX8tlvMRUZa1Q4hbsGSV/LN8rcpRdfftN0mwliOLRgIkQAIkkJ1AacnVVFZya2XfQbY4ggAFKkdQ55wkQAIkQAIkQAIligBC4JBjavqvc+ViYs4eULVrVpORQ2+QPj06i7tbyfwRC4LaqKE3ynAt79Q/S9fKj7/Pl6MnTlv2CsLcH3OXaB5Vq+WO4QNVjiqEMNJIgARIgARIgARI4GoIlMyfnq7mjfksCZAACZAACRQDgdLyl9NiQOX0U6zduF0+mjpdTp4+l+Na69WuKXffPli6d2kvLlr4XGkwCGw39Lle+vfuIstWbdS8xv6Sw0dPWF4NQt0X3/2uhf+tkLEP3SGd27e03GeFBEiABEjg6gnExifKxh3HZcuuY3L4+AUJPxklMXGJkpCYrAb39fbUvHW9pU6NIKlXq6K0aRYi7VvUkgA/76ufnCOQQDEToEBVzMA5HQmQAAmQAAmQQMkggHC+j6ZMl39XbshxwTWrV5H777hFenbtWGqEKdsXheDW8/oOSnz7d8V6+Wr6LEFieLNBuHvutfdVv7Gj72DYnxkOyyRAAiRQCAIpWi7ARav2yl+Lt8u6sHBJz7iU6yjRmliFz9FTkbJ84yH55o/14upSTjq1qiODe7eUfl0ai4eWE5FGAiWBAL9TS8IucY0kQAIkQAIkQALFSmD56k0y8dPvJCY2Ltu8geX95d6RN6s8TG5urtnul8YGCFXIU9X9uvYq/9Y3P/4p0TFWNhDyNoftkWcfu1u6XdeuNGLgO5EACZBAkRJITkmTn+Zukm9mrpPzUQmFnguC1uqtR9Rn0rTFcu+QTnL7wHbi6cFf/wsNlQ8WCwF+hxYLZk5CAiRAAiRAAiRQEggkp6Ro4XwzZM7C5dmW6+rqKrcO7CX3jbpZ/Hx9st0vCw0Q5MCgr5Zn6+sZf6pcVOZTDCHovTT+UxnUv5sW9jdK+2XIoyxg4TuSAAmQwFUTWLHhoLw9ZZEcPxN91WOZB4DQNWHaEk342izjRveTrh0amG+zTAJORYAClVNtBxdDAiRAAiRAAiTgKAJnz0fIi29+LPsPHc22hCaN6spzj98nDerWynavLDZAoHv8wZEyQMtRNfGTb2TP/iMWDBD49h0Il/EvPy5VKlWw3GOFBEiABEggiwC8pt794h/5Zf6WrEabUoMgD+lQzUdaV/WWukFeUt3XQ3xdXFSvhIwMOZWQIkeikmTbmUTZcPqiHIxKsRlBlPA1+rVfZMQNbeSFB/vQmyobITY4AwEKVM6wC1wDCZAACZAACZCAQwnsPXBEnn/tA0HeKbO5a3k77ht1i4wcMkBcLv8yYL5f1ssN64XIlMkvy48zF2geVbMkVcubohuEvgefeF3ee+1Jadywrt7MKwmQAAmQwGUCEVHx8tArP8vuQ9bcfrjt4+4itzYuL0NDA6VRoFeuzDzERYI8vKSZJlwNrBeo+u2PTpLf90XLH3tj5GJqhuVZCGE79p2UqW/cJhWC/Cz3WCEBRxPIlF0dvQrOTwIkQAIkQAIkQAIOIhC2c5+M/e972cSpWjWqylRNfLlj2I0Up/LYGwh3YARWYGY2CH5gC8Y0EiABEiCBLAIntFC+kU9/l02cctMSnI9sHij/3NZQXuxYNU9xKms0awmCFp7FGBgLY5oNghjmxhpoJOBMBChQOdNucC0kQAIkQAIkQALFSmD7rv3yzKvvy8XEJMu813duI19++Ko0rF/b0s5K7gTACszAzmxgC8ZgTSMBEiABEhCB59T9L87Ilm+qdoC7zLiptrzUqZoEeV79r+oYA2NhTIxtNuS6whqwFhoJOAuBq/+ud5Y34TpIgARIgARIgARIoAAEduw+IM9qwklSUrLlqdtvHSBvjRsjvj7elnZWrkwAzMAODM0GxmAN5jQSIAESKMsEkHMKYX22ydCvr+Unv99cX1pUsP//ezAmxsYcZsMasBasiUYCzkCAApUz7ALXQAIkQAIkQAIkUKwE9uw/rAQTW8+ph+4ZJo/cP0JcylnDIYp1cSV8MrADQ7A0G1hDpAJ7GgmQAAmUVQLvTP0nW1jfgAYB8nmfEPH1KLr/92BszIG5zIZwPyRpp5GAMxCgQOUMu8A1kAAJkAAJkAAJFBuBw0dPyDOvTJaEi4mWOSGoIJcSzT4EwNJWpAJzsMce0EiABEigrBFYseGg/LrAelofvJomdK0pri6XihwH5sBctp5USJyOtdFIwNEEKFA5egc4PwmQAAmQAAmQQLEROHs+Qp55ebLExiVY5nzw7qEUpyxE7FOBSAW2ZgN77AH2gkYCJEACZYUAwujenrLI8rrIC/V+j+IRp/SJIVJhTtucVFgbQ/10Srw6igAFKkeR57wkQAIkQAIkQALFSiA+4aIKMTsfEWWZ957bb5I7hw+0tLFiPwJgC8Zmwx4g3A97QiMBEiCBskDgp7mbLHmncLLehF41ijSsLzeuCPfD3ObT/ZCPCmukkYAjCVCgciR9zk0CJEACJEACJFAsBNLT0+XVdz+XI0dPWuYbOriP3H/HLZa2slpJSkmXpVvOFcnrgzFYmw17gT3B3tBIgARIoDQTSElNk29mrrO84vCm5YskIbplkjwqSJyONZgNa8RaaSTgKAIUqBxFnvOSAAmQAAmQAAkUG4HPv/pFNmzZaZmv5/UdZMyDIy1tZbkybc5h+XbekSJDANZgbjbsCfaGRgIkQAKlmcCiVXvlfFRWaLmPu4s8ek0Vh78y1oC16IY1Yq00EnAUgazvRketgPOSAAmQAAmQAAmQQBESWLJivfw6+2/LDC2aNpRxTz3A0/ouU9kdHiu/Ljkuh07Gy/7jcRZW9qrgdD8wB3uzYW+wRzQSIAESKK0E/lq83fJqtzYuL0Gejv9VHGvAWsxmu1bzPZZJoKgJOP6/iqJ+Q45PAiRAAiRAAiRQZgkcP3lGJnz8jeX9q1auIG+PGyMeHu6W9rJaSU3LkPd+2CMZGZknSC1ad6bIUIA52GMPzIY9wl7RSIAESKC0EYiNT5R1YeGW1xoaGmipO7JiuxasFWumkYAjCFCgcgR1zkkCJEACJEACJFDkBFLT0uT1CVPkYmKSMRcEkrc0gSQoMMBoK+uFHxYeVZ5TOod/Np6R9Mtild5mzyvYYw/MAiH2CHuFPaORAAmQQGkisHHHccu/qQ2CPKRRoJfTvCLWgjXphn//sWYaCTiCAAUqR1DnnCRAAiRAAiRAAkVO4Luf/pJ9B8Mt84x9aJSENqhjaSvLlSOnEuSHheEWBJGxKbJhd6Slzd4V7AH2wmzYK+wZjQRIgARKE4Etu45ZXqdjdR9L3RkqHapZ12S7ZmdYI9dQNghQoCob+8y3JAESIAESIIEyReDA4WMy/bd5lnfu0aW9DO7f3dJWlisI6Xvnh92a11JGNgwL153O1mbvBuwF9sRs2DPsHY0ESIAESguBw8cvWF6lVRVvS90ZKq2rWtdku2ZnWCPXUDYIUKAqG/vMtyQBEiABEiCBMkMg49IlmfjJN5Kenm68c8XgQHnmsXuMOgsivy09LruPxOaIYlXYeYlPLPpwO+wJ9kY37Bn2DntIIwESIIHSQCD8ZJTlNeoGelrqzlCxXZPtmp1hjVxD2SBAgaps7DPfkgRIgARIgATKDIF5i1bInv1HLO8LISTA39fSVpYrpy4kypezD+eKIDk1Q5ZtOZfrfXvdwJ7YCofYO+whjQRIgARKA4GYOGvC8ep+zidQ2a7Jds2lYR/4DiWDAAWqkrFPXCUJkAAJkAAJkEA+CCDZ9rQfZlp6dtfCyK7r2NrSVpYrcE7CqX1JKVkeZjnxKI4wP8yLvcEemQ17aE5ub77HMgmQAAmUJAIJicmW5fq6ON+v4LZrsl2z5QVYIYEiJOB8/3UU4ctyaBIgARIgARIggdJN4Nc/F0lkdFbYmpenh4z5z+2l+6UL+HZzV5+SzfusISc5DRF2MFpOa55WxWHYI+yVbthD7CWNBEiABEiABEig7BCgQFV29ppvSgIkQAIkQAKlmkBCQqL8Mssqatx2a3+pXCm4xLw3jvfGyXr/bj4rX889LF/MPiR/rTopOw7F2CUv04XoZPls5oF88YCn1aL1Z/LV92o7YY+wV2bDXmJPaSRAAiRQkgn4eltD+hIysh9M4ej3s12T7ZodvT7OX3YIuJWdV+WbkgAJkAAJkAAJlGYCfy74V+ITLhqvWN7fTxM9Bhh1RxWQ8HvjnkjZfjBGTkckWZbx8a/7pWfbytKpWQVZuvWc/LDwaK5eSzUr+8jIviEyoFM1cXcr3N8YJ/+0r0DJzxdqAtXdN9SVcuUsyy6SCvZq1tx/JSYuXo2PvcSejhp6Y5HMx0FJgARIoDgIlPf3lmhTHqpT8ckSFGw9Na841pHXHKcSUiy3sWYaCTiCAAUqR1DnnCRAAiRAAiRAAnYlkJaWLjPnLLGMOezmvuLr49gfsvcfi5NJP+6V3eFZYYfmRa7UTsvDx821nKSl531y3YlzF2XC9L0yc+kJee/RVlI12Ms81BXLSzadVXNdsaOpA+bcdSRGmtcrb2otmiL2Cns27Yc/jAmwpyNu7i9ubq5GGwskQAIkUJII1KkRJEdPRRpLPqJ5sjZzMoHqSJT1jydYM40EHEGgcH9+c8RKOScJkAAJkAAJkAAJ5EJgxdrNcv5C1i8AXl6ecsuNvXLpXTzNCI978L2NuYpT5lVcSZwy9z10Ml4embhZzkRaf6Ew97Etx8Snyoe/7DeaIYjddH0NGXd3U6MNBbQH+WflgkJbcYX5YS7sGfZON+wp9pZGAiRAAiWVQL1aFS1LDzvrfKHL285Y12S7ZssLsEICRUiAAlURwuXQJEACJEACJEACxUNg1jyr91T/XtdJgL9v8UyewyxbtCTk47/bfUWvqBwezVfTOe2v3S/+b7ukpuUvl8nHv+2XqLgUcXEpJwM6V5MfX+8sz45qrCUmt3omdWlZSb54ob3Uq+5nrAOeV/mdx3iokAXsGfbObLZ7a77HMgmQAAk4O4E2zUIsS1x/KisU3XLDgZUNp61rsl2zA5fGqcsYAQpUZWzD+bokQAIkQAIkUNoIHD95RsJ2ZnkHldMSJg0Z2NthrxmfmCZvfrNLkPC8KG3/8TiZtfzkFadYuzNC/tlwVnq3qyLTX+2kvKaqV8wMfUxKTrc8D8GqWgUv+d9zbVVeLNyMTUiVNTsuWPoVZQV7hz3UDXuLPaaRAAmQQEkk0L5FLXHV/jig28GoFNkfnX8PWP25orpiLViTblgr1kwjAUcQoEDlCOqckwRIgARIgARIwG4EFixZLZdw5Nxla9m0odQJqa5Xi/06fVG4nNdyjBSH/brkWJ6n+11MSpf1uyLkm5c6yGv/aS4hVXwsy0pKsQpU3pc9qny93GSCludqWM/MX1IWris+gQh7hz3UDXuLPaaRAAmQQEkkEODnLZ1a1bEs/fd90Za6Iyu2a8FasWYaCTiCAAUqR1DnnCRAAiRAAiRAAnYhAPFi8bJ1lrFu7NvVUi/OSkpqhvyZD68me60Jeah2Hc45ATvm8PJ0kSdGNJL6NbJC9sxz2wpUnh5ZPxoiHHDs8Eby9O2h6hRC5LEqLrPdQ+yxWYQsrnVwHhIgARKwB4HBvVtahvljb4xEJecvRNvyoJ0rWAPWYjbbtZrvsUwCRU0g66eQop6J45MACZAACZAACZCAnQnsPXBETp89b4zq5ekh3a5rZ9SLu7D9YLQgxK84befh3P8S72IKlctpTYk2vyB5e1hzUuGZW7rVlHcebinrNE+s4jLsIfZSN+wx9ppGAiRAAiWRQL8ujaVSUFZexIvaHzM+23rW4a+CNWAtumGNWCuNBBxFgAKVo8hzXhIgARIgARIggasmsGKN9YS3Tu1biY+311WPW9gBjp61Jpot7DgFee5sZOHDCW09qGyTpuvraN8kWHq0qaxXi/yKPcRems12r833WCYBEiABZybg4e4m9w7pZFnir7tjZEeE9fQ8S4cirmBurMFsWCPWSiMBRxGgQOUo8pyXBEiABEiABEjgqgms2RBmGaPbtW0t9eKu2Ao+xTH/70uPy4K1p+WsFu5XUKtV2Uc6NqsgrRoESqMQf6kc6JnrEB7uxftjo+1e2u51rgvlDRIgARJwQgK3D2wntaoGGitL0w7SeG7JSUlIycqhaNws4gLmxNxYg25YG9ZIIwFHEqA86kj6nJsESIAESIAESKDQBM5HRMnhoyeM511dXaVj2xZG3REFD7fiFXH0d3z7u92qWLOSt7QJDZI2jYOlTaMgCQ7ICpPT+5qvg7pUF3yc0bCX2NP09MxE7thr7HmlCkHOuFyuiQRIgATyJODp4SbjRveT0a/9YvQ7GpsqTy09IZ/3CdFO+ssSi4wORVBIzyinzXlcMLfZsDaskUYCjiTgmJ+iHPnGnJsESIAESIAESKBUENgStsfyHk1D64m/X1aOD8vNYqrAI8mRduJ8ovy16pS8Nm2n3PT8Srnj9XXywc/7ZPnWcxKTYP1lxJHrzM/c2Evsqdls99x8j2USIAEScHYCXTs0kBE3tLEsc+XxeHluxQmBcFTUhjkwF+Y0G9aEtdFIwNEEKJE6egc4PwmQAAmQAAmQQKEIbNu5z/Jc21ZNLHVHVFo1DNSSe7tKUnKm148j1qDPqR1wKOGnE9Rn5rITgoTpDWr6ZXpYaV5WWKuvl3P/KIg93bH7gP5Kgj3v1/Nao84CCZAACZQ0Ai882Ed27Dspuw9lJUlfcDBW4pOPyfs9aoqvR9EIVQjrg+eUrTjVtH4VwZpoJOAMBJz7pxJnIMQ1kAAJkAAJkAAJOCWB3XsPWdbVslmope6IircmTt3QuZr8oQlCxWHIH9WsboBs2Rclu47ESmpa1mlMtvNnaIrV/uNx6vPz4mNaOEk5Ca3tL21DtXBATbBqUb+8eOVwip/tOMVZt91T2z0vzrUUZK7jWrL8xJR0aVTLvyCPsS8JkEAZIIAwuqlv3CYjn/5Ojp/JOoUVwtHQPw/JhF41pEUFb7uSQEJ05JyyDetD3imshaF9dsXNwa6CAAWqq4DHR0mABEiABEiABBxDICkpWcKPnzImh3dQ00bWcDDjZjEX7rmhrizeeFZiizikrmmdAHn34ZbiruW9um+gCBK07zgUo8SqzfsiZd/ROC1kJPecJri3WxO18PlhYbgaB2KXymGliVYoY2xHGvYUewtxDYY9x957eeWezN2R69XnRqjlc59tky6tKsm9N9alUKWD4ZUESEARqBDkJ1+NHyX3vzjDIlJBQBo1+6gMb1peHr2migR5Xt2/wVHJGfLZ1rPqtD5zQnQsAuIU1oC10LIT2LZtm+zbl+mp3bp1awkNdfwfwbKvsvS1UKAqfXvKNyIBEiABEiCBUk/gyLGTkpGR5S1UvVpl8fW171+cCwsRicmfv6OxvPTFDrmsqxR2qFyfq6ElQ3/nsjild4L3U/smweojUl8SktIk7EC0EqzgYXXwRLwh9OjPmK/wvtqm9cfn67lHlDdVc82rqi2SrmufxrUDlNeV+ZmiLmNPsbcnTmWGwmDPsfdNnESMzOv9sfcrt52XVWHny4xQdUl76QtnT0pgcCVx93C8iJgQH6u2yNcvIK+tuup7zvbe5hdKT0+TMSOv14Tdi/LoC5Olbefe5ttlspycmiGb9kTKtS0qiqZ/O8xqagLRj5Pvlode+dkS7gch6ced0fLnvli5tXF5GRqqnbIa6FWgde6PTpLf90XLH3tj5KL2vraGsD54ThWnOIV/v83/38aaXFxc1Md2fc5Q79mzp0RFRamltGjRQrZv337FZc2aNUvuvvtuCQwMlGPHjl2xPztkJ0CBKjsTtpAACZAACZAACTg5gfBjWd5TWGr9OjWdasXdrqksD9/SQP4366DdRSrkkZrwaGupUD5vAQD5pfALGD4wJEnftj/qsodVlBw9k5Dn2uCRhV/i8IFhPOStwumAbRoHqXxW8G4qasPe6gIV5sLelwSBSudS0oSqGV+8I4f379SXr64emthUPaS+1GnQVEKbtZPK1WpZ7uuVl8fcKvt2bpKmrTrJ6x/9pjc75Bq2cYW8/fydau5X3/9ZmrXuXGTrcKb3tn1JCALnz55QzbHRmf8t2/Ypa/VUTbB5/vMwCQ3x17xP6zlUqIJANGPyPfLuF//IL/O3WLYCwtL0HVHq0yDIQzpW95FWVbylbqD236Ofp/hq4g4sQdvjU/HJciQ6WcLOJsr6UxflYFSKZSxzBQnRkXOqOMP63n33XXnppZeMU1nN63F3d5cGDRpI79695cEHH5TmzZubbzusbBbTzOW8FvT5559LXFycNGvWLK9udru3e/duadeunSa0lhN4fDVs2NBuYztqIApUjiLPeUmABEiABEiABApN4Phljxp9gJBa1fWi01xH9q0t9TUxadKP++T0hcQ814XcVRCS9h+Lk+PnLubYt1pFb7mzf20Z0KlaoULvyvu6C4QzfGCRsSmyBYLV3kglWiEsLS+DR9aaHRfUB/0CtPGugVgFDyvtWqeab5F4I2Tu7WZjabZ7b9xw8kJJEarm/f6VpKYkZ6e55h/V5urmLjff/rAMuWusuLt7WPqdPnFE1SPPn7a0O6Jy5mS4XNJ+cYdFFPF6nOm9HcG6pM65T/v31hmEKghFrz42QHp0aChvT1lkCfnT2UJwwmfGrqycVfq9/F4R0jdudD+HnNa3ePHiHMUprD01NVX27NmjPlOnTpXJkyfLY489lt/Xcpp+J0+elH///Vet5847M8VxVF544QXZssUqPpoXXa1aNfnuu+/MTfkuJyQkSGJi5v+7US4NRoGqNOwi34EESIAESIAEyhiBM2fPW964RtVKlrqzVDo2rSAzXu0kf284ozyRth6IkoiYzL9s+3i5aonJMz2SBmiJ1REaiFxLuw7HypHT8XImIkncXMsp4aduNT8Jqepj1xA7zNe7XRX1Aa+zkUmZ4YCaaLV5b5Sci0rKEyNybC3fek590BHjKe8qCFaNg6WmFoZoD7PdW9u9t8ccxTlGSRGq2nTuJR279Nd+eUyRU8cPyfHw/XJg91ZJSkyQmT98LOuWz5e3Pp0lfgGBBr7b7ntG9uzYIO2udfyJYO279JWDe8PEx9dPWrXraqyxKArO9N5F8X6lfUxnEaq6dmggHVvXkZ/mbpJvZq6T81H2ERwqBfnKvUM6ye0D2xWr11Rhvm9SUlLkySeflGuvvVbatGlTmCEc9syPP/6oQhg9PDxkxIgRxjo++ugjLcQ29/+fIhyQlkWAAlUWC5ZIgARIgARIgARKCIHzEZl5IfTlVq5UQS863dXD3UUGXlddfbC4uItp6rS9QH93lQDcvGCEzOE0PXyK26oEewmEMnxg8KjaoiVb36KJVfC0gsdVXob7izedVR/0qxzkJW21UEBdtML4hTHbvbXd+8KM6QzPOLtQVbdBM+l5420WVOfPnJAP33hU9u/eIiePHZQZX74rDz39rtGnz+A7BB9nsOCKVVXOpeJYizO9d3G8b2mdwxmEKnhT3XNrJxk5qJ0sWrVX/lq8XdaFhed54EVO+4FTWju1qiODe7eUfl0ai4e7c/3av2rVKhUGFxMTo7yH3n77bYE4BUtLS5Nvv/22xAlU33//vVr/DTfcIBUqZP+ZpG/fvjJo0CDVx/ylYsXMMHxzW1kuO9d3alneCb47CZAACZAACZBAvglERWUmP9YfqBBc/IKOPndBr/4+JePHL3hA1axUQwZ3qaFyVYWfTsgMCdQSrm/VBKsrnVIID6wFa0+rDxhhPBUOqHlXtdc+5f3c84XOdm9t9z5fgzhxJ2cXqszoKlWtKW98MlOef/AGOXpojyyZ95P0uvF2adC4leq2O2ydhB/cLfUatZDGLdqbH5WoiHOycdUiOal5Y6WnpUr1WvWlU7cbBEKSrUVeOCM7Nq+SY0f2SZrmwVWjtuZZcv0AKR+U9Yvczi2r5djhvdKqfVcJrlRN/pkzQ04dOyQVK1eXG4f+R0RLj7byn1nq2q3vUPH0yvToS0yIl+V/zxQvbx/p3n+YxERdkK0blmreYdskoHyQtLuur9QPbamWhBDBDdqad21bKx6eXtKgSWtp17mPuGk5c8yW23ub11ijdkM5uGeb7NiySrGoowmA13TsIUEVMkNuzeOhfGjfdtm1dY3KH5WmhUAh71er9t0UW9u+ev3k0QOyO2y9HD28R6pWryMt2nZRucP0+7ld4R13aF+Ytqd7VZhurTqh6l1r1WmU7RHzO+XE3VvzWCvp5gxCFQSlQT2aq09sfKJs3HFctuw6JoePX5Dwk1ESE5coCYmZobi+3p5S3t9b6tQIknq1KkqbZiHSvkUtCfCzjxdrUeynn5+fSiQO76FXXnlFkpOTZfz48cZUu3btMsq2BYhbK1eulLCwMImIiJDGjRsLkpgPGTIkR2EIz587d05mz54t69atk1OnTsnFixfV/E2aNJF77rlHjWE7T0HqyP+0c2dm7j5zeJ95jPbt25fI0EXzOxRHuWT8hFQcJDgHCZAACZAACZBAiSEQF28NfSjvX/J/KXIE/D3hsYITrdLTMwQnR7UNDVZhhbZrQS70utV91WdI95oqFBGnAureVTgtEDmq8jJ4ZOHz1yprgvu8nlH30hOhNRh29HSUdBm9xKiXlkJOQpUzvpurq5vcdv+z8t6L96kcT2uXzTUEqtk//U+2rPtXEF5nFqjQ5/P3nlHhgeZ3QkL2T39cbRFpFs/9Ub7//E1JvBhv7irfffaG5hH1vlzbI9MDYc6vX6i5Bg5/UBOzVirBTH+gqZYQHYLSlx+8qJoaNW2jJXjPTFoMgeyrj17STg5zVaLPpJcflLjYKP1R+f37j+Suh1+Wrn1vlfdfG60EH+OmVkAC+OfHf62FDvobzbm9t77GoVq+Lohu/87/xXgGhYDACvLcW9MktHk7ox0JzSe/8pASqIzGy4Wfpk2QAbfeK/eOed32lsz7fZrG7S0txCjduIfEyTePfNSo2xZSkpMU17//+sH2lkq6fOOwB2TkA89bco3p75Qb96atOmYbq6Q2OINQBXYQmnp1bqQ+JZXlldYdGhpq6ZKenvV9rN9AjiWE/3355Zd6k7oivxXs9ddfV55XffpYQ4xfffVVeeedd1SuK9XR9OWvv/6SSZMmyX//+1958803TXcKVvzhh8z/hoKDg2XgwIEFe9imd3R0tCxcuFAlPYcAFxAQoAQ4hA16exdMdMQ4GzdulOPHjws8tbp27Srw5MLpiWZDYvfVq1cr0e/IkSNSt25dlay+e/fu2gnJvuauRV6mQFXkiDkBCZAACZAACZCAvQlcTLTmc/DxLlz4mL3XVdLGe+6zMImKywrdmzPxegnytya+zumdEIrYqJa/+tzWJ0SFn+w9GpuZw0rzsNpxKEZwCqBdzMXqrSIZeQthdpnTgYOYhSq/So0kMeakpKdYBVkHLk9NjRxT/uWDJS4mUnkt5bUeiD+6OAXPKoTDQUTZum6pbF67WBLiYgyBasu6JTJ10vNqOHgWwZvJ1dVVCVHnTh+Xf/6abghU+pxzNaEK5utfXuo3aqnlndombm5X/hUHa3jj6ZHKmwueUdVr1hMIaci5NX3qeJn/x9dy4exJ8Q8Ikms69ZTDmkfTCeWhtE4W/PGNDLnzcX0JV7xC9IJ5evlIG22siwmxglMGY6Mj5H8TnpUPvvtXCULoAy7wnvL28ZMO1/eXkLqhWv6ai7JswW/Km2r+zK+l7bW9pWXb69Fd2ep//5JvP80UrbAv13ToLufOHFcnKs6a8aneLdt14ssPyLYNy1R73YbN1UmHyDG2beNy9e5gm5aaLPePfSvbs4Xlnm2gEtDgLEJVCUBV6CXqIpM+QKNG2b33hg0bJgsWLNC7ZLvCMwri0L59+6ROnTrqPk7fQ/hgToKXPgDuoU+/fv2kS5cuenO+r3ge+adgw4cPF+SgKqxNnDhRXnvtNeXhZTsGhLY5c+ZIy5aZHp629811JE5HqOGyZcvMzUqoGzt2rHz44YdGO7zRRo0apUQso/FyAacrTp8+XTp2LD7h+cr/etuuknUSIAESIAESIAEScDCBlFSrSOHpWfgfCB38Kg6dHnlKzJaWfslczXcZ4zSrW1597uxfR+XY2nUEglXmCYEop6Zlnqj2f/bOA7yKogvDH+mdEkISEnqTHnrvXbqoiAIqiALygwVURMEGNhQLCip2kabSERAUBESKELr0FkgIkB5SCf+ejXuze5NA2k3uTb7zPGFnZmdnZt9Bc/nuOWfkNMI7nWpomLiUvaGKWwUkfBlHtdKacX+saZF+FaukC1SKR9Lt7LSSqFxED7Gpb32LMuV81HLPASNUwaV8hQC1fvNmquLNk+7BIOFsb3y8HGXL+6r3HpmQiuULP8n6dEGlh19AVbz20c8moUsekgTpdzIRqSZMnYNOve5Vu0pieMmxJe0iTtVXPLFkzRIeKO8wZXQvhF0+jz83/pIrgUoGr1qjHqa/v0gV9qT+9cczIGKT5PI6eWw/xMtLTDzPHnp8qirkuXt4qW3yR6+BI/HksHZIUsQqEZX0AtWKHz9V+0kI5MxPVsK3YmW1LqGJb0191MRfbfzvj4N7t5nEqe79H8LjT89Cqf+8KuLjYvDGlIfUkMTfVv+Ifoonla+y3+aWFXfzPvq6rXs+UqjS72b+ynv37kVERATk5LuffvpJDb/TjzhmzBh9FRs3bjSIU5JEff78+fDx8cEnn3yCN95IF1Elj9XLL78MzaNJvAgdlZBcEZG6dOmiJi+XkEA5OfCdd97Bb7+ln056S/lm4Oeff86TQCVjhIWFqevNLrzP8DK3qYg3l4QfSoL43r17qyGLwmrx4sW4cOECxo8fDwlxvJOJN5iIU/LuzzzzjJrvS4S7r7/+Wg1x1J4XrynhInzEU2rAgAGoUqWK6r21YcMGnDp1CuJFdeLECVSqVEl7zKJXClQWxcvBSYAESIAESIAELEFAvhXVm7m7uv4ey9kTkFMC9XZTCfMrCHN0sENQrTLqzygl2kG8qcSrap/iXeWunF44f8XtRQ3jGoxrBApmjcY5rKem/HsK7Rv54NF+1TDo8Y+sZ2FmK9FO74uNzgiPM+uiVu113kz/Ht6D1h3vNnXz8Q00lUPOnVRPC5SGgQ+MM4lTUpewQgmTy8ocHZ3UMLnscjll9YzWNkoJldPEKWlr27kfFnwwDXExURCPoqlvfWPKXeXi6o523QaqJxhKGF5uTISc6XMWq95Y2nN3DxmlClRSFzFME6gqV78L8mNuIj5JbizJd3U17JLptpyweO7UEbV+9z2jTOKUNGgC24xJ6QKc6SGlsGFlekJnEcEeHv+ySZySPtI2Yuw0zJh0n+phtmfHRlWk0j+fH+76cWyxTKEq/7v22GOPZTvI22+/DcnXpDfxLNLMxcVFFbX8/dMP9Hjttdfw+eefq3mmpI8IK5qJQCXCU8WKFREUFKQ1q1epi8ClmYgxeTFNDBNvIxHOsjMZf926dZlui2BWvXp1tX3KlClo3ry5KgrpOwYEBKihiCIoxcXFQXJ43c5+//139fajjz6Kt97KOMjipZdegoQQavb888+r4pSE/4nQpme0cOFCiOAmJxCK99ZXX32lPWbRKwUqi+Ll4CRAAiTeMY3XAABAAElEQVRAAiRAAiRgvQTMBaq8elDd6Q1dnOzRoq6SHF35ERuueFnlxjr0Tc/voT2zfX43rWiV152Hr2PK3OBcrU0vTEn4pLWbhOaJ3UkYqlknyBQO+MGrTyqC0BCIOFNF8SjSW9ilc6ZqkBKillOTxOGVlDC4vJg+T5Y8Lx5E4tGVLlDVV0Py9OOW90339kpJTlI8mRJM4pW+T1blStVqG8Qp6SNjyT+exXNDQv2ysshrV3D9aiiiIq4qHh9JSE5KULvJ/JrpxaqmbbpqzaZrrXpNTGV9IfTiGbXauGUnJWF85hwzkmtL8mzdiI/Flcvn9Y+mP5cP7pkGYwMJKAQknFdEEvHoMbejR4+amkSYkjxJ8qOZePdIInSxq1evQnIqeXqm/39UQt30JqcESh8RXqSP9BWTfE+5NXl2xYoV6mPDhw+/7eNLliyB/JibJGkXzyaxyZMnm99W63379lUFKqmcP39e9YjKsuN/jeI5JbZ//35IuJ+Wu0rEPT+/9IMpDh48qOadkn5PPfWUQZySNgn7W7BggeqJJSIfBSqhQiMBEiABEiABEiCBLAiIx5Tei0rK9KLKAtQdmuzt7Qw9UpVk6dZm+n2WtRW3fbY1YUr7+yE5jsT8AqtpTVle5VS3ya99piQcH6eemCeJwuVHkmlLsm8tebleoNJC+7Ic0KyxlFmYqtntXFfFMyg7u9297J7Jrl28whydnBXRKdHw/zLpLycHLv36PUPi9+zGuRae4U1V1js9JDK7vvr2K6EX1Kq3cgJidiZeciJQiUBmbgXN3Xx8a67XqeyJUf2qo23D8orIaM0rtd61iaeOeALpvZYkzGzevHmZBCpJji75pTQTYapdu3ZaNcurhNxpApWM++2336phcvv27VNDC0UYLggT4UZC8sTuJFCJt5avb+b/Rhs0aJBpKSKenTlzBqGhoYiMjDSIcXLvTtanTx81FFASpAvr5557Dg888IAh4bmE7WkmJyBmZRLyJ6GCMTExuHbtmppoPat+BdlGD6qCpMmxSIAESIAESIAECoWAk5Oj8u1nhhdBUlKy8g0hE6XnFr65B9XNPOagyu28uekve6s32fviYLYqTAl7CSuLjrymbkPFStXvuB3ijfPxD9vw25qFaqJzEaOOHtiFF8cNwJvzV6veVOJNpJmINgUpBmnj2sJVxCk5xU/yYIlA1LxtT8Wry18Vs/74dSn0Qp68TyndGZepqSk5fkU5xVAsVcnFk53JP+zFXFxvH06U3fPm7dbi+Rh3IxW9n9lqvrw71otKmIqJS8TeQxew/1gIrlyLQfj1OFyNiFOu6Z4/Fbw94VPOAxW8PeBb3gtN6gaiecPKyul/1vs78ZtvvkHjxo1V76N7773XlMR82bJlWLRoEYYNG2bajytXrpjKOSlIknIt/E+EKAknlPksYVp4n4T21ahR47ZTPP7446ZcWdl1jI6OVr2ZJC+XCHh5NQkVFK8zCdMTIUoYiHeW5KN64YUX1NxUp09nhNpnl1/K29vbtAQR/SQU0NJGgcrShDk+CZAACZAACZBAgRNwc3E2CFRyqh8FqtxjNheoLBXil/uVZTyR6cRGZe9t2WxZmNK4L1rwrhqaJvWGzdprzbe9iifVgKFPoL+ScHvzusVKrqeX1BPzJFH4uOdmQxKjayYhZXLiX0m0tcu+UMWpgMo1MWveKjXMTuMgp/uZC1TlfNLDdaRPaMhZZPKIysZTRBKpnz99zJT3S5tDuybciEPU9fSQKX8lh1ZJtqIQppKSU/Hbjn+x6vfD2Ln/rCLgZO/dejbkOuRHb+Id26ZJNQzo2gA92t0FZyfr/Gf/oEGDMGHCBHz4YfpJl/IOkvBcvH000VryR2nhsHJfcjbp8ypJm97kvpajSQQvvThVrlw5PPzww6hTp47qYSXhdZIwPS928eJF1btInh05cmRehjA8k5SUpJ6899dff6ntcqKg5KMSUSgkJERNCm944DYVCfGT0/fGjRuHjz76CMuXL1dzT02fPh1HjhxRvcn0p53K3JIk3dwkHFIzzSNNq1vqap1/Uy31thyXBEiABEiABEigWBDw9HBHRFSM6V2iY+PgXa6Mqc5CzgjYmyVJt8YQP9lbvcne26IVB2HqlhJKu2bZAvy9da26Ba069jGcJpeTfZE8T937PYhdf65TTpHbigglz5KYJBLXbMv6ZZkEKhFMLl04jZp3Nda6FcvrmROH1feSBOeSA0ozYR8TaRQh5J5/YIYHm+xLgyZttUfU648L3jHUtYoIYCJQHfpnh3qaoj5hvfTZvGaRIoqk/+O0Zt0g7bESdS0KYSpVOe30l43BmLdou+Itle4hlRfoImht33ta/fEt/zvGDWuPe3oGwUE5wMLaTLx95CQ+TQwRz59NmzahR48e6lIlb5IkCReRRkxyRckJd87Od/6ywvzEuzVr1qBNmzbqOPLH//73PzXkz9SQi4J4J0kIuqzj/vvvz8WTWXcVYUoTp+SEwieeeMLUUfJFSVtuTUIh5UdOAJScUsJD8mBJ0nktMbuMKSf86blo80i7mLxjdl5WWt+Culrf39CCejOOQwIkQAIkQAIkUGwJlC2bcfy6vOT1iPSEzcX2hS30Yg6ZclBln5cjOi5v3zLnd+nme2u+9/kd39LPizDVobEPvnyxJd4c1wi2kABdmERGhCM89CIkmXbw7i3qiXMvTxyC7+a9riJzdnHDI0/OuCO+lYvmYd47kw2JtmXcsyfTkx77/5fDSnJRaSfZbVz1Azau/F71sBKRRE6Rmzy6J76Z+8od57P1DpoH1J4dG0zJ0xPi4zB7xhOQUxDNTTyhJJ+XmHDbtmm5KixJEvt5707BmqWfmz+i1vve95h6lVDCmVNG4MKZf9W6hPxt2bAMPy54W63XbdQSjZp3UMsl5Q8Rpt4e3xgLprZEu0aFl2dqx74zGDjuc7zy8a/5EqfM90mELhlTxpY5rM1EfBo4cKBhWeL1o7eGDTM8KiXBueRU0gQtrZ/kahLRSBKDayZ5k/Smha1KmyQA159op++Xk7IW3icJzMuWLZuTR27bR/JjaTZ48GCtqF4lvC4/VrlyZVWU0saQnF7iRaZZVgnQJdeV5n3WrFmzQsv/SA8qbVd4JQESIAESIAESsBkCPt7GD4PhVzN7FtjMyxThQs1D/LLKQSVhfx8vO4G721ZEaY/Cz/9kvrfme1+E+G47ta17TP2+djHkJysLUk5+Gz3pDfUkuqzu69vClHA9SYou+ZNEjJIT484rYshNJV+Sk7MLuvV9wNT9oSem4o3JD6nC1BdzXsRXH01XTtYrZcqT5BdQzdS3uBbadx+EJV/NRqQSXjfmnmZK/qmKuH4tTOWV3Tvf+/BTePP5h1VuH70xEfLjoIT4iNgkCeclp5d26qI2hoiBXRX2sseXLpzCs6N6wN3DS+2bkpKsdpMTGkdPShckteeK87UoPKaEZ3xCMmbN34jlGw9ki7d0pavwCzoF71qX4OodC9eyyk+5dA+rhAhPJEQqP9c9cf1kAMKCayL6ok+msSQMcMy0RRjcszFeHNsT7q7ZHwiQ6WELN0h+Jkk4rtm6devUJOGal89rr72G9evXm0KLRcBauXKlmgBcQtXOnTuH4OBgNZfVxIkT0aRJE3UoEbYkp5VmEjooib8lObucGJhX++eff9QcT/J8QYT3yTiBgYFyUU3Es6lTp6qhjcJCn5NL65PdVfJYycmFEt43dOhQNd+U5OISXprVrFkTkl9KPNGEq8xXrVo1CDsJjxSPKwmF1E5GFP6FZRSoCos05yEBEiABEiABEigwAn6+xg/fl8IynzJVYJPZ8EAXrtxAeEQimtctl+VbmAtU5jmoImKS8fLnh5CYfBNPP5DxbWuWg1mo0XxvzffeQtPma9hAH1fVY8pWvKX0Lyu5n44f3mtqkpA8nwoBagienNjXuEVHtGzfy3RfX3BwSBcwHR2dTc29B41EyNkTOH5kr5LvKMN7Q+YZOf5lNUG61lmSqc/8dCU+fXuyEn52ND3ETMnTLUJWu64D8NDjU7WuSqhS5rlMN5WCdsy6tOkTrmfXrj3rqCRYFnPQvUPGvfT3kgTj+tMks1tLdu3aeHJfxCMRkzS7Z/gESDjjr0puLhGK5LREEfV69H8QgVVqK7m7psFe+Qe53ho2bY8X3/4OIuppjGWNzdt2xuPPvoU5r47DsYO7Mz03bsq7qFOvGZZ+OwfXwy8jPi49bFqEqsYtRIR8HV5lMpIky5x3eif9umylXFTClPC5GBaFcdOX4MxFo6eP3HNwSUbNXntRq/ceeFbM/ksYx4AkeAWkP1+100F5FLGXvXFyfQuc2tAcqYlGIUqEsAPHLmHea0NRyc86QuMlnE/EKDm5TkxC50Q0mTVrllqXXExPPfUU5syZo9blj/Pnz6s/pob/CvrcSnKy3uzZs01hfJcuKe+tnBSYX9O8p0TkETGoIKxbt27qKX+SFH7atGnqu0ruLfEYs7e3z/EUcqqgFi4oIYwiRokHlhYi+eCDD6rilAz4zjvvYNeuXepJgTLnyy+/DMnTpfc8E6FL1lZYZvy/W2HNynlIgARIgARIgARIIB8EKlU0HtV84eLlfIxWfB+t7OuGj5aewOJNFzB+SE1Ur+hheNnMIX5ppvtHzkbjpc8O4WpUEp576C5Te2EXzPfWfO8Lez05ma+Swt1W7Y25y/O89GdenQ8JR3Pz8DSNUaVGPbw+9xfExUSpYov8w9NXSYjuWTpr0bRarQZ4d8F6JCbEI+T8KVVckpMCHZ0yRC8ZPKu5TJMqBZn3u3XHVA8EEXg0y65duz/9vcXq3Pr8T9q9tl36q8KNiDT69WS3luzatfG++GWfKkKJIKSZCEsjxk7DA6MmqwnRhVelarUVQcxe9R5p320gXNwy3kd7rkHTdvjw+62IjY5QOIegSvW6JuFL3ikp8QbcPUtr3U1X8aKSnxvxsYi4GgZnF1f4+GV4cpg6/le40zuZ97fmelEKU8LlYmgkRkz53nQan8aqlF0aavTYh0bDfodLWWMOPq3Pna4iaDUdtR71Bm/HwUVdcfq3priVlpHdRwSxEZO/w/fvjkAlf6NH8p3Gzst9H5+ML5Ukp5R5wm0RYiZNmqT+aONrgopWf//999GzZ0/1JLpjx45pzaarjNu1a1eMGjXK1CZ5kyTv1JNPPmkI/RMRS7yLOnTogLFjx6r9zdckQo14I4lJWTMJLdS8sjQPJe1eVldZl4TL3SlnliRDX7VqlZp7SrzBNJGofv36av4pySElnk16kV1OLNRMG19OMHzppZfw6aefqsLc3r3pXzi4ubmpXlV6byjxMDt06JDKYPPmzUhISFDnFQG+du3akKTqufHe0taSn2spxd0r+2QD+RmZz5IACZAACZAACZCAhQgcO3EGjz+d4XIeqAhWi75Iz5lioSltdthLVxMw4rW/Id5Rfdv6Y3S/6ihfJv0f+1PnH8S24AzvszfHNkKHIB+s2nYJc5acQIqSsNfNxR4r3uqgXosCwrAxzyPkcsYx45/PmY66tasXxVIKfc56fWYa5ixXpbWhbl6JOP+3oenor9MMda3SfuxmrahenxpWx1BnhQQKgsAHi9ITLGtjbZ9feF4Y2pxZXZNS0rD3WATaNizY/FLm/71m99+frCkmLhHDnv4m0+l7pSuHo82k5ShX81JWS89zW8SpAOz8cDCiL1QwjFEt0BuL5jwCLw8XQ7u+kpv30j9nXo6JiVHzRnl5eSmeeFn7ycTHx6siiYgwpUtnFlS1MePi4tQwPRGQJK+Ur6+vKqjoxRutr1xF8hDvrNOnT6vimIQAinAkJuuSMWQ+vWekiMOSo0ra5J52qqCE3EneKbG///4brVql54BTG7L4Q07IE+GnTJmce6uFhoaq3mHi/STClZiMIScOCj+9xcbGqmvTTi7U7sn6JZRRhC4RCKtUqQK9oKX1067SXxjJHNJXBK2isKz/ZhTFSjgnCZAACZAACZAACeSQQLXKAeqHRvlAJXY5NBzx8QnKMcmuORyh5HQLUMLNhveqiq/WnMHq7Zfx254reLBHZQzrUQWuTsawgcjYZLzzwzGsUvpp1r2FX5GJU7KnsreayT8UZO9pJEACJJBXAs6Odmry87w+n9/n0hSxZPLbKzKJU1XaH0ZrRZyydyr4AylE8Or93mf4WxGpzm9vYHoFyUsla5mvhPvZSeI8C5q5sJLVVO7u7srv8cweguZ9RYwJCsr56ZIiLtWoUUP9MR8ru3XJ7xu955T23HfffacWxcPoTuKUdBTPJs27SRvjTlfxgpIfvbm6ukJ+zM3c80u7L+uXNcpPTkz6iyBW1GZX1Avg/CRAAiRAAiRAAiSQWwIuLs6oWqmi6TH5wH9U8aqiZU1geK8qkLxIYolJNxWx6iyGvvwX/jyQ4T0l995Z+K9BnJK2gR2KThCSPZW91Uz2XPaeRgIkQAK2SuCLJX9h+97ThuVX7xqMdpOXWUSc0iYS4UvmkLn0JmuRNdHuTEDyO0kYnpjkt6IVPAF6UBU8U45IAiRAAiRAAiRQCATq3VUDZ86HmGY6eOQ4WjSpb6pbSyEuIRXnw+JxLjQekrS8tLsjqvq7o5ry4+vtYvFvrYWDk+IxIEnOn/044x8mkgD9TiY5WuSnqEz2VG+y5zQSIAESsFUCR0+F4ZOFfxqWX6H+ObScoJywVipDjDd0KMiKMofMFXelDMKPVDWNLGvq0LwG6tX0M7WxkJmAJCt/+umn1VBFLXdV5l5syQ8BClT5ocdnSYAESIAESIAEioxAUIM6WLNhq2n+fw4cw2gr+kIzOi4F3/56Fj//EYKbaVn/w6NmoAeeHFILLbI5Zc/0cgVQaFXfG52bVMCW/Rkhc3catii9p2Rtsqd6kz2nkQAJkIAtEpDfA9M/XKuIGxmHUbiWjUX755bCzl45rrKQTOaSOX99ahwSItO/gJA1ydqWfDgK9naWDfUrpNe0yDQSqjdzpjE3oEUmKsGDMsSvBG8+X50ESIAESIAEbJlA08Z1Dcs/evwMYuPiDW1FVdmjJOB96NWdWLr5YrbilKztVEgcnv5wP9787hgkca+lbeL9teHqbMw7ld2ckhxd8k8Vlcleyp7qzXzP9fdYJgESIAFrJvDLxmCIB5XJFG8myTnlUiZvJ/WZxslDQeaUufVeW7I2WSONBIqSAAWqoqTPuUmABEiABEiABPJMwMe7LKpXyTgOXU7h2fXPoTyPV1AP/n1YSTqrhNJFxeY80e3avy7jxXkHbytmFcT6KpR1xqh+1XI0VFEmR5cFyl7Knmomey17TiMBEiABWyOQnHITny7cZlh2tc4H4N/klKGtMCsyt6xBb7JGWSuNBIqKAAWqoiLPeUmABEiABEiABPJNoG3LxoYxtv71j6Fe2JXwyES8+tXhPAlNu45exzdrz1p8yfd1rYzqFT3uOE9Rh/eZ76X5Xt/xBdiBBEiABKyEwJo/DuPKtVjTahxckhH08G+melEVZA2yFs1kjbJWGgkUFQEKVEVFnvOSAAmQAAmQAAnkm0DHts0MY/y95wBuJCQa2gqz8sGSE4i9kZrnKb/79RzOKQnVLWkO9qXwzLA6uN2J4kWdHF32UPZSb+Z7rb/HMgmQAAlYM4El6/YZllez115I/qmiNlmDrEVv5mvV32OZBCxNgAKVpQlzfBIgARIgARIgAYsRuKtWNfj7+pjGT0xKxtYdxg/bppsWLkg+qT+Dr+ZrFkmi+926c/kaIycPB9Uqg16t/LPtWuTeU8oeyl5qJnsse00jARIgAVsjcOFyJA4dv2xadikl91TtvrtM9aIuyFpkTZrJWmXNNBIoCgIUqIqCOuckARIgARIgARIoEAKlFDeg7p1bG8Zau/FPQ72wKht365Lf5mPSrcope4nJls8B8uSQmvB0y3ygc1EnRxd05nvYQ9lj2WsaCZAACdgagT/3GPNM+dQ7Dw9f6xGAZC2yJr2Zr1l/j2USsCQBClSWpMuxSYAESIAESIAELE6gT7d2BvHi4NGTOHch49tqiy/gvwkOn4kukKnkNL+TFy1/qlNZTyc8PrBGpjUXdXJ02TvZQ81EmOqt7DGNBEiABGyRwL6jFw3Lrtj8hKFuDRXzNZmv2RrWyDWUDAIUqErGPvMtSYAESIAESKDYEqgU4IfGDWqb3u/WrVv4ec0mU72wClGxGSFp+Z0zKq7gxrrdWgZ2DMBdVbwMXYo6vE/2TvZQM9lb2WMaCZAACdgigZPnjKHf5e8yClbW8E7mazJfszWskWsoGQQoUJWMfeZbkgAJkAAJkECxJjC4bzfD+63fvAMxsZZNNm6Y0EYrdop30uQH60CuYkWdHF32TPZOb+Z7q7/HMgmQAAlYO4HQqzGGJXoFXDPUraFivibzNVvDGrmGkkGAAlXJ2Ge+JQmQAAmQAAkUawId2zSDT/lypndMTEzC8rWbTfXCKJRRQuYKysp4FNxYd1qTeFCJJ5VYUXtPyZ7J3mkmeyp7SyMBEiABWyQgB1/cSNB5xCrJyJ29bljdq6hr0iVKlzXL2mkkUNgEKFAVNnHORwIkQAIkQAIkUOAEHBzsMaS/0Ytq2YqNiL+RUOBzZTdgg+qls7uVq3ZnRzvUquSRq2fy21lyUQX4uELyTxWVyV7JnulN9lT2lkYCJEACNklAF64s65fT8vQn5lnLO2W5LrO1W8tauY7iTYACVfHeX74dCZAACZAACZQYAoP6dIWHu5vpfaNj47D4l19NdUsXerYsGHGnU5MKcHEqXFFGTvP7+JlmkBP8ispkr2TPNJO9lD2lkQAJkAAJkAAJlAwCFKhKxj7zLUmABEiABEig2BNwd3fF0MG9DO+5+Jf1CL8aYWizVKVmoAc6Bvnka3h7u1IYeXfVfI2R14crlHXO66P5fk72SPZKb7KXsqc0EiABErBZAv/l99PWf+tWKeUQiPScf1qbNVyzXJfZ2q1hnVxD8SdAgar47zHfkARIgARIgARKDIH7B/VCuTIZp9IlJiXj4wWLCu39nxpaG+KNlFcb2acqqvq55/Vxm31O9kj2SjPZQ9lLGgmQAAnYMgH50sHNVZdTUBGnkmIyPH2t5d3UNemEM1mzrJ1GAoVNgAJVYRPnfCRAAiRAAiRAAhYj4ObqgsdGDDGMv2X7HuzYFWxos1SlQlkXzBjVIE8f7FvV88YjfatZamlWO67sjeyR3mQPZS9pJEACJGDrBPx9Mr40kXeJuVTe6l7JfE3ma7a6BXNBxZYABapiu7V8MRIgARIgARIomQT69uqIurWNQs/sud8gJja+UIC0buCN2f8LQhlPxxzP17dtRcwa1yhPwlaOJ7HCjrInsjd6k72TPaSRAAmQQHEgUKuqMfT72r+VrO61zNdkvmarWzAXVGwJUKAqtlvLFyMBEiABEiCBkknATsmbMeV/j8LePiPh97WIqExCiCXptKhbDgtntMH93SrdVnSSvFVzJjXB1JF1Iaf3lTQTcUr2RjPZM9k72UMaCZAACRQHAs3qGwWpy3trW91rma/JfM1Wt2AuqNgSKHmfhIrtVvLFSIAESIAESIAENAK1qlfG8Pv6alX1+ocSRrZq/RZDmyUrpT0cMfG+2lj7XkeMGVA901TL3miLr6a1hIhZJdFkL2RP9CZ7JntHIwESIIHiQqBD85qGV7l6tArirpQ1tBVlRdYia9Kb+Zr191gmAUsSoEBlSbocmwRIgARIgARIoMgIPDxsAOrUrGqY/8PPFuL4qXOGNktXPFwdUK9aacM0Ikr5l3ctsZ5CsgeyF3qTvZI9o5EACZBAcSJQuWJZNKxT0fRKcmLeibWtTPWiLsha9CcLylplzTQSKAoCFKiKgjrnJAESIAESIAESsDgBRwcHzHhurCHZdnJyCl6a+TEio2IsPj8nyJqAsJc9kL3QTBKiy17JntFIgARIoLgRGHp3U8MrndrQHAmRnoa2oqjIGmQtejNfq/4eyyRgaQIUqCxNmOOTAAmQAAmQAAkUGYFKAX54buKjhvnDwq9jmplAYujAisUIiCgl7GUP9CZ7JHtFIwESIIHiSKBflwbwLZ8hSKUmOiH42x5F/qqyBlmLZrJGWSuNBIqKAAWqoiLPeUmABEiABEiABAqFQLeOrXD/wJ6GuQ4dPYmZ73+BtFu3DO2sWI6AsBbmwl5vsjeyRzQSIAESKK4EnBztMf6hDobXO7ulMUL3G/NTGTpYuCJzyxr0JmuUtdJIoKgIUKAqKvKclwRIgARIgARIoNAIjB89FC2bGr8V/n3bbnz8+Y+FtoaSPpGwFuZ6kz2RvaGRAAmQQHEncE/PINSrqfMUVXJR/f3hYCRGeRT6q8ucMjeUNWgma5M10kigKAlQoCpK+pybBEiABEiABEigUAjY29vj1RfGo1qVAMN8P636DV/+sNzQxkrBExDGwlpvsheyJ7I3NBIgARIo7gTs7UrhtUl94eCQ8U9wyQG1/Z37kXaz8P4/KHPJnPocWLImWZuskUYCRUkg47+OolwF5yYBEiABEiABEiABCxPwcHfDu68+Ax9v4+lE3yxaie+XrrHw7CV3eGErjPUmeyB7IXtCIwESIIGSQkC8lJ58qKPhdcOPVMXuuQMN3kyGDgVZUTymZC6ZU2+yJoN3l/4myyRQiAQoUBUibE5FAiRAAiRgmwQuXU2wzYVz1ZkI+Pp4Y/brz8LL091w7/Nvf8IPy9Ya2ljJPwFhKmz1JuxlD2QvaCRAAiRQ0giMGdoW7ZvXMLz2md+DsGP2fbiZ7GhoL8iKjC1zyFx6k7XImmgkYA0EKFBZwy5wDSRAAiRAAlZL4MSFWIx7d6/Vro8Lyz2B6lUCMfu1Z+Hu5mp4+LNvllGkMhDJX0XEKWGqN2Eu7GUPaCRAAiRQEgnYlSqF2c8PQrVAo0h/fnsDrH/2CUScMoaiFwQjGVPGljn0JmuQtciaaCRgDQQoUFnDLnANJEACJEACVklg77EITHj/H0TEJFvl+riovBOoW7u6GmLm5upiGEQElU+/WsLT/QxUcleR0/qEobk4JawlrE/Y00iABEigJBPw8nDB/NeGooK3pwFD9IUK2DBlDHZ/OgCJkflPni5jyFgypoytN5lb1iBroZGAtRBwsJaFcB0kQAIkQAIkYE0ENu25gpnfHkVKalqullWvz8xc9dc6H/11mlbktZAINKxXSxVMnp3+HhITk0yzLvr5V4RcvoJpz4zJ5GVl6sRClgTibyRg5vtfYNvOfYb7Li7OKmthTiMBEiABEgAq+ZfF97NHYtz0JThz8ZoJya00O5za0BzntjZCzV57Uav3HnhWvG66n5NC7GVvnFzfQh0nNdEp0yPVK5XHPEWcquRXJtM9NpBAURKgQFWU9Dk3CZAACZCAVRJYuvki5v50kl40Vrk7BbuoRvVrY7bi1fPcK3NwIyHRNLgILGOeehWvPj8OtWpUMbWzkD2Bk6fPY8bb83DxUpihk3hOvfPK0xDWNBIgARIggQwCIhAt+fBRzJq/Ecs3Hsi4oZREWPp3ZVv1p3Slq/ALOgXvWpfg6h0L17LKT7lYtX9ChKd6Il/CdU9cPxmAsOCaiL7oYxhLXxncszFeHNsT7q6ZhSt9P5ZJoCgIUKAqCuqckwRIgARIwCoJKJFJmL/iFBZuOG+V6+OiLEOgcYM6+PDN5/G8IlJFRMWYJhGh5YlnX8eohwbjwSF9YGfHzAgmOLpCWloaflS8zr5auBwpKam6O0C5Ml54WxGn7qpVzdDOCgmQAAmQQDoBEYpmPt0Pd3eqh1nzNuJsSGZvKRGcbic65YSl5Jt6cVxPtGvKMOuc8GKfoiHAT1pFw52zkgAJkAAJWBmB1Ju3MOu7oxSnrGxfCms5IqB8NmdGJm8pEVwkl9JYRag6dfZiYS3HZuY5eeaCykYYmYtTtRXPs88/mEFxymZ2kwslARIoSgIiHK2c9zhe+V8f+JY35qbKz7pkLBlTxqY4lR+SfLYwCNCDqjAocw4SIAESIAGrJpCYdBMvf3EIOw9n/tYyvwsvV6V1lkNEnP87y3Y2Fh0BvwremDd7Gj78bCFWr99qWMixE2fx2KRXcE+/bopH1SB4uLsZ7pe0Slz8DcVjagV+WbMZN2/ezPT6/Xt3wqQnHoKzE0NIMsFhAwmQAAlkQ8DBwQ73390UA7s3wm87/sWq3w9j5/6zyv9nc5cP097eDm2aVMOArg3Qo91dyv+L+c/+bJCz2coI8G+qlW0Il0MCJEACJFC4BKLjUjDlk2AcPZsR2lW4K+Bs1kRABJXn/vcoWjVtiHfnfovomPQcH7JGEWKWrdyI37bsxKMPDsKA3p3h4GBvTcu3+FpSU29i1fot+PrHFYiKzmCjTVzayxNTJjyMTu2aa028kgAJkAAJ5JKACEr9ujRQf2LiErH30AXsPxaCK9diEH49Dlcj4pRr+v+D5TQ+n3IeyomAHornlRea1A1E84aVeTpfLpmzu3UQoEBlHfvAVZAACZAACRQBgdDriXj2o/24cOWGYfbSHo4Q4YpWcgmIwNJQSer94fwf8Pu23QYQIszMmfe9KlaNHj4YXTu2gl2pUoY+xa2SpiRo+/3PXfjyh+XqCYdZvV/XDi0xaexwNe9UVvfZRgIkQAIkkHsCXh4u6NqmtvqT+6f5BAnYFgEKVLa1X1wtCZAACZBAARE4fSlOEaeCcS06yTBihbIuGNQ5AJ8vP21oZ6XkEZAE36++MB69u7VXwv5+wKXQcAOEkMtX8Oo78/H9kjV4eNgAdG7fotgJVSJMbdm+B98uWoUz50MM769VAvwrKOF8w9GmRSOtiVcSIAESIAESIAESyDUBClS5RsYHSIAESIAEbJ1A8IlIvDDvIOISjCeOVfZzQ7/2AXBy5Bkitr7HBbl+EV6aNZ6JJSs24Iela3AjIdEwvAg3M976FFUC/fHgvXejR5c2cHSw7Y9YKamp+O2Pnfjxp3U4HxJqeF+t4ubqguH398PQQb3g5OSoNfNKAiRAAiRAAiRAAnkiYNufnvL0ynyIBEiABEigJBPYuj8cr351BMkpxoSjdap4oWdrP9jbFe9QrZK89/l5dxFgRihiTN+eHRVvopVqHibJx6Q3EXLe/OBLLPj+FwwZ0B39enVCaU8PfRerL0fHxmHNhq34edUmXL0emeV6Je+W5N96eNhAhvNlSYiNJEACJEACJEACeSFAgSov1PgMCZAACZCATRJYvjUEc5acQFraLcP6m9Qpi45NKqCYpxEyvDMreSMgYX9PjxuBoYN74xtFqNqoeBmZn2Inws78r5ep+ZpaNWuEbh1bwrVM1bxNWAhPJSYmYfuu/dj8527s+ucgUlKMnoXaEuzt7dGzcxs88uBAVPTz0Zp5JQESIAESIAESIIECIUCBqkAwchASIAESIAFrJ/Dl6jP4eu3ZTMts39gHzeuVy9TOBhK4HQERaF58+jGMHNofPyxbgw2//wVzjyoRerb/vU/9cXR0RCn78rjlUhFw8VeGLrq/c7eUvFJnzoVg9/7D2LPvMA4cOYHk5OwPBRCPqV5d22L4ff0QWNH3dlh4jwRIgARIgARIgATyTIACVZ7R8UESIAESIAFbICDeUrN//Bertl82LNdOCeXr3tIP9ap5GdpZIYHcEBDB5oVJozHqwcFqjqrV67ciIdGYo0rGS0lRBKCUUJRKTM/ndCDCFVNj6qJh3ZpopJwW2EC5WtIOHzuFg4oQdUi5Hj52EnIS4Z3M1cUF/Xt3UnNMVfApOkHtTuvkfRIgARKwRgL1+sy0xmVxTSRg1QQoUFn19nBxJEACJEAC+SGQpOSZemXBYWw7cNUwjKODHfq2q4iqFd0N7ayQQF4JiIDzvzHD8Ihymt/ajdvUHFUXL4VlO1xKcoLJu0o6OTs5oWwZT5Qp7aX8eKKsdlVCCqWe/uOltKeX5RkRmSKVn6jomPRyVPo1vV0rxyAyKhZJycnySI6sUoCfmmOqb88O8PTgfyM5gsZOJEACJEACJEAC+SZAgSrfCDkACZAACZCANRKIvZGK5z89gIOnogzLc3W2x8BOgfDzdjG0s0ICBUFABJ0H7umt5KjqhX9PnsPWHXux9a+9CLl85bbDi4AUFn5d/bltRwvdFE+wTm2bo1O75rirVlUlHxsPC7AQag5LAiRAAiRAAiSQDQEKVNmAYTMJkAAJkIDtEgiPTMLkj4Nx5nKc4SW83B0xuHMgyno5GdpZIYGCJiACT93a1dSfsY/eh3V/HseszzcrIX6KV1WS8pOWc4+mgl6bjOfh7oamjeqiWVA9NA+qj8qBfpaYhmOSAAmQAAmQAAmQQI4JUKDKMSp2JAESIAESsAUC58Li8exHwbgSYcwDVL6MsypOubvyV58t7GNxW6OPj3LqnUcd3FJ+gFuoH5iG3k3slXxQp7Bzz0HExsVb9JXFs6tNi0ZqrquG9WqjepUA2NnZWXRODk4CJEACJYnA0V+nlaTX5buSgEUI8FO6RbByUBIgARIggaIgcOh0tBrWFxNvPJEssIIb+ncIUPL88B/kRbEvnNOcQCm4efpgcN8myk83pKWlqQKV5I6KiolT80lFx8QiWs0vZdamtGsJziUvVWmv9B81R5WXh5qrqvR/7fo2EagoSJnvA+skQAIkQAIkQALWRIAClTXtBtdCAiRAAiSQZwI7Dl7DDCUhemLyTcMYNSt5ok8bf9jbM6eOAQwrVkNAhCNNaKpiNaviQkiABEgg7wT2Ho3I+8N8kgRIoMQSoEBVYreeL04CJEACxYfA2r8u450f/sXNtFuGl2pUqwy6NPNVEj4bmlkhARIgARIgARKwAIFbyq/hP/eHY//xSMPodnb8RWwAwgoJkECWBChQZYmFjSRAAiRAArZC4Ltfz+GLVachH4r11qZhebRq4K1vYpkESIAESIAESMBCBORLoo1/h+L4+VjDDE6Odpgxqr6hjRUSIAESyIoABaqsqLCNBEiABEjA6gmkKYrUh0tO4OctIYa1yulp3Vr4okGN0oZ2VkiABEiABEiABCxDIDklDWu2X8KFsBuGCTyUg0neGtcIQbXLGtpZIQESIIGsCFCgyooK20iABEiABKyaQEpqGl77+gj++CfcsE4HJc9Un7YVUSPQw9DOCgmQAAmQAAmQgGUI3EhMxYotlxAeaXZ6bmlnvDcxCDUC+DvZMuQ5KgkUPwIUqIrfnvKNSIAESKBYE4hXPghPnXcQ+8zyW7g42WNAxwBU9HEt1u/PlyMBEiABEiABayEQHZeCX/64CLnqrbKvG96f1AR+5Vz0zSyTAAmQwG0JUKC6LR7eJAESIAESsCYC16OTMHnuAZy8aMxv4eHmgMGdA+GtfFtLIwESIAESIAESsDwB8ZhaoYTZ30g0np5br5oX3n0yCKU9HC2/CM5AAiRQrAhQoCpW28mXIQESIIHiSyAk/Aae/igYodcSDC9ZzssJg7sEwtONH4QNYFghARIgARIgAQsRkFxTknNKck/prY1yOMnrYxrCxdle38wyCZAACeSIAAWqHGFiJxIgARIggaIkcOxcDKZ8EoyoWGMIgX95VwzsFAAJ76ORAAmQAAmQAAlYnsDx8zHKaX1hkFP79NanjT+eH14Xkg+SRgIkQAJ5IUCBKi/U+AwJkAAJkEChEdh19Dpe+uwQEpKMIQTVlaSrd7fzVz4I2xXaWjgRCZAACZAACZRkAvuV/I9/7g+HcpCuwR7qVQVjB9WEcpAujQRIgATyTIACVZ7R8UESIAESIAFLE9iwKwxvfncUqTeNn4TrVy+Nbi19YcdPwpbeAo5PAiRAAiRAAiqB7QeuYu/RCAMN+T084d5auL9bJUM7KyRAAiSQFwIUqPJCjc+QAAmQAAlYnMDi3y7gk19OZvqWtmV9b7RtVN7i83MCEiABEiABEiABIE0J5du0OwxHz8YYcDg62GHaI/XQvbmvoZ0VEiABEsgrAQpUeSXH50iABEiABCxCQMIGRJgSgUpv4izVuWkFNK5dVt/MMgmQAAmQAAmQgIUIpKSmYd2Oyzh7Od4wg5uLPWY90QjN65YztLNCAiRAAvkhQIEqP/T4LAmQAAmQQIESkFA+CemT0D692SsJV3u39ketyp76ZpZJgARIgARIgAQsREByP67cGoKw64mGGeT03NkTglCbv5MNXFghARLIPwEKVPlnyBFIgARIgAQKgIB8EJZk6JIUXW9Ojnbo3yEAlXzd9M0skwAJkAAJkAAJWIhATHwKlm8JQWRMsmGGAB9XvD+xCeRKIwESIIGCJkCBqqCJcjwSIAESIIFcE4iKTcaUTw7g2Dljfgt3VwcM6hQIn7LOuR6TD5AACZAACZAACeSewLWoJFWcik9INTwsHlPiOSUeVDQSIAESsAQBClSWoMoxSYAESIAEckwg9FoCnv4oGCHhNwzPlPF0wj2dA+Hl4WhoZ4UESIAESIAESMAyBOR38eptl5CUnGaYQHJNSc4pyT1FIwESIAFLEaBAZSmyHJcESIAESOCOBE5ejMXkucG4Hm0MIfAt54JBijjl6swPwneEyA4kQAIkQAIkUAAETim/k9fvDIXkg9Rb9xa+mPZwPcipfTQSIAESsCQBClSWpMuxSYAESIAEsiWw73gkps47iPhEYwhBFX939GtfkR+EsyXHGyRAAiRAAiRQsAQOnozCH/+E45Ycpauz+7tVwoR7a8FOjtKlkQAJkICFCVAGtzBgDk8CJGAZAqmpqahatSp8fHywZs0ay0zCUS1GQD4EP/txcCZx6q6qXhjYMYDilMXIc2ASIAESIAESMBLYeegaft97xSBOiR41bnBNTLyvNsUpIy7WSIAELEiAApUF4XJoEjAncObMGbRu3RotWrTAunXrzG9nqs+bNw/NmzdHnz59Mt0r6Q03b97E+fPnce3aNVy9erWk47Cp9/9ZORVoxoLDSEk15rdodlc59G7jDzs7fktrUxvKxZIACZAACdgkAXGW2rQ7DLsOG0/PdbAvhRdH1sNDvarY5Htx0SRAArZLgCF+trt3XLkNEihbtiz27duHlJQUfPzxx7j77rtv+xbvv/8+Tp06hTZt2ty2X0HdjIqKwvz589XhJkyYAA8Pj4IamuOQgPLNLPDFqtP47tdzmWh0aOIDEahoJEACJEACJEAClicgeaZ+/esyTofEGSZzUXI/vj6mIdo08Da0s0ICJEAChUGAHlSFQZlzkMB/BESg6t27t1rbtGkTrl83fmOlB7V//35VnJK2Bx54QH/LYuUTJ05g6tSp6o94e9FIoKAI3Ey7hbe+P5ZJnBJvKfGaojhVUKQ5DgmQAAmQAAncnkBS8k388sfFTOJUaeXU3I+eakpx6vb4eJcESMCCBChQWRAuhyaBrAg8+OCDarPkUPrll1+y6qK2LVu2TL3a29tj6NCh2fbjDRKwdgKJygfhF+cfxFrlm1q9yWlAkm9K8k7RSIAESIAESIAELE8g7kYqlm66iMtXEwyT+Xu7YN6U5qhXjb+TDWBYIQESKFQCDPErVNycjASAAQMGwN3dHfHx8Vi8eDHGjBmTJZaffvpJbe/atSt8fX0z9dm9ezf27NmDI0eOoGLFiggKCkKXLl3UsTN1VhouXbqEzZs34/Dhw0hKSkLdunVxzz33oEKFCkhMTMTSpUtx8OBB06MrV66EeHGJtW3bFrVq1TLdi42NxY4dO3DgwAGcPXsW1apVQ4MGDdC5c+cs5//9999x6NAh9OzZE4GBgfjss89w/PhxVKpUCU899RS8vG7/YejYsWP4888/1fXVqFED3bt3R506dUzrya6QW0bZjcP2vBOIiU/B858ewKHT0YZBXJUQgkGdA+FbzsXQzgoJkAAJkAAJkIBlCETEJGP5HyGIvZFimKBGgAfe+18QypdxNrSzQgIkQAKFTYACVWET53wlnoCbmxsGDRqEhQsXYuvWrbhy5UomASo4OBgnT55UWWkeVxo4EYfGjx+PH374QWsyXUW0Ec+rhg0bmtqk8MUXX+DZZ5+FPKu3Z555Bt988w2qVKmChx9+WH8L06dPN9VffPFFzJw5U61v27YNDz30EC5evGi6rxVq1qyprqtVq1Zak3p977331KTwMp+ENuqFMBG1OnbsaOivr3zwwQeYPHkyJCm6ZqWUo2VeeOEFrZrpmhdGmQZhQ74JXIlIxDMfBeN8WLxhLAkhGKyIU2U8nQztrJAACZAACZAACViGQOi1BKzcegni1ay3oNpl8NbYxvBw4z8L9VxYJgESKBoC/D9R0XDnrCWcgIhOIlCJ6CKeUk8++aSBiBbe5+Liono56W/ed9992LBhA0SkGTlypHoq4Llz5zB37lzVK6lXr144ffo0XF1d1cfWrl2Lxx9/XC37+/urHlwODg6qYCTeT5IUXeaTPFdyGp54WYn16NEDZcqUUcvimSUmXlNSlnWLF5h4g4m4JYKarEkSuovgJLmsxDvK3CTpu5jk4mrWrJnqAebo6GjezVQXD7Onn35arZcvX17N3yXvKut48803Tf3MC7llZP486/kncOZyHCZ/HIzwyCTDYD5lnTGoUyDcXfnrxwCGFRIgARIgARKwEIEzl+Kw7q9QpJqdntupSQXMGFUfTo7M+mIh9ByWBEgglwT4L4RcAmN3EigIAhLqJoLLtWvXsGTJkmwFqr59+xrC3zZu3KgKQbKGBQsWYNSoUablSLhe69atERoaqobQSeic5LkSzykxCcPbvn27Gg4odbknIo+E93l7e2PRokWQkDjN+2n27Nlo1KiRdDXZ888/r4pTsvbffvtNDSvUborgNmLECHW8GTNm4KuvvtJuGa7iZSXheiKW3cnefvtttYuEIe7cuRPVq1dX61u2bEH//v0RF2c8eUZu5paROiD/KFACB09FqWF9sUqeC71V8nVD/w4B/CCsh8IyCZAACZAACViQwJEz0di8+wrS5ChdnQ1Wvix6emhtyGElNBIgARKwFgKUy61lJ7iOEkVAPJjEy0dMRCPJD6WZ5HXSwvuGDRumNatXCccTq1+/vkGckraWLVuq3ktS/uuvv+SCo0ePql5VUp4yZYpJnJK6rOHll182he5J2+1MwvLEc0lMxC/JeaU3Cfvr1KmT2vTzzz/rb5nKzs7OWL58eY7EKclRJZ5ZYhMnTjSJU1IXLy3xDMvKcssoqzHYlncC2w5cxdMf7lfyWxjFqVqVPdWcU/yWNu9s+SQJkAAJkAAJ5IbA7iPX8duusEzi1Oj+1fHssDoUp3IDk31JgAQKhQA9qAoFMychgcwEJMxv3rx5uKV8oyUJyrVQNi28TxKHiweV3jThSp7RwuX09yVET0xC/MQk5E6z3r17a8U8XSVsT7MhQ4ZoRcNVQv7EuykmJkb1DhNPK72J55gkU8+JnT9/3tTNnIPc0Dy9TJ3+K+SWkfnzrOedwKrtlzD7x+NISzN+S9u4dll0blpBCUvN+9h8kgRIgARIgARIIGcExFlqy75wHDgRaXhAvKUmP1gHA9oHGNpZIQESIAFrIUCBylp2gusocQTatWuHypUr48KFC2qYn7lAJSF7koNKb5rwJJ5RWuie/r5WTktLU4t6gUpO+suPaXPLGFnll5J2CRXULCwsTA1j1OpytbPLudOmcNEsN2vX1plTRtocvOaPwNdrz+LL1WcyDdK2UXm0rJ/x9yJTBzaQAAmQAAmQAAkUGIGbypdE63eG4uSFWMOYzkqeqVcea4AOjX0M7ayQAAmQgDURoEBlTbvBtZQoApLkXEL4JM/Srl27IMm/xfNI81QyP71P4EiInFj79u3xyCOPqOWs/pBwPzGZQ7OEhATT81pbbq4SEqhZUlKSmiRdq2tXyWulmaenp1bM01W/9uTk5ByPkVtGOR6YHbMkIN5Sc5acwPKtIYb7dsrfvW4tfVG/emlDOyskQAIkQAIkQAKWIZCckobV2y7h4pUbhgk8lRP63h7fGI1qph9+Y7jJCgmQAAlYEYGMf3Fa0aK4FBIoKQREhNISgUuYX2xs+rddvr6+6Nq1ayYMkiRcPJNKly6N0aNHZ7pv3iCJ0TU7c+YMmjZtqlXveNW8sLSOWoJyqUt+qDZt2mi3TFdpFxORKDsvK1PnOxQCAjLczyVsLzAw0PCEhDlmZblllNUYbMsZAfkg/OpXR7B1f7jhAQcHO9zd1h/VAzwM7ayQAAmQAAmQAAlYhkB8QipWKF8WXTU7PbeCcnru7P8FoXpF/k62DHmOSgIkUJAEch5vU5CzciwSIAGVgJySJwnPxeQ0Py3/1NChQ2Fvb6+26/+oU6eOWt20aZN6Wp/+XlZlOTFPMy15uFaXqwhie/bsMTU5OTmZylouJ61Bm1vqWZ3QJ6cBanM0a9YsV+F82hz6a+3atU3Vn376yVTWCtOmTdOKhqu2zpwyMjzMSo4JxClJ0J/5eH8mccrFyR5DugRSnMoxSXYkARIgARIggfwRiIpNxtJNFzKJU1X83DFvSnOKU/nDy6dJgAQKkQAFqkKEzalIICsCWijfvn37TCfuaW3m/SdNmqQKPxJiN3jwYIhXlGbiUbR37171tD45cU9MTtpr3bq1Wp4/f76alF2elVC8lStXonHjxuqJfGoH5Q9/f3+tiMWLFyMyMlIVwsRrq169etASrS9YsACzZs1CXFyc2l/yRfXp0wfh4emeNK+99pppnLwWxBOqY8eO6uOy9oULF6rrljU99thjWSaJl865ZZTX9ZXk565FJeHJ9/5B8IkoAwZPN0fc36My/Mu7GtpZIQESIAESIAESsAyBKxGJWPLbBUTHpRgmaFijtCJONYNvOWM+U0MnVkiABEjAyghQoLKyDeFySh4ByUOlNxFmsjuhTgSliRMnqt0lb1WNGjXUUDpplxPzWrRogdmzZ+Pw4cOmId966y015C4lJQXjx4+Hh4eHmj9q0KBBOHv2rCGXlIQWiqgl9ssvv6iClYTaffbZZ2rbO++8g7Jly6pl8WCSUEMfHx9UqVJFPb1PbowbNw7dunVT++T3j+nTp6trl3DD4cOHw9HREX5+fvjyyy8hidO1tejnyQsj/fMs357ABSWvxdh39uL0pXRxUuvtXdoZQxVxqpxXhheedo9XEiABEiABEiCBgidwPjQeP22+iISkm4bB2ykHlMyZ1ARe7o6GdlZIgARIwNoJUKCy9h3i+oo9AckT1aVLF9N7Pvroo6ZyVoU5c+ao3k0iZEki8ZCQEIjHVEREhHqKnuSm6tmzp+nRTp06YefOnarwJP3Fe0qSjru6ukLm+uGHH0x9pSDeUTK2mHhbiclpg2INGzbEoUOH0K9fP/V5EY6uXbumenXddddd+PHHH/Hpp5+qffV/aKGDWgJz/b3blUXoWrduHfThfhL62L9/f9VbrEGDBurjIlzpLbeM9M+ynD2Bo2djMO7dvQhTvq3VW0UfV9zfvRI8lCSsNBIgARIgARIgAcsT+PdcDFb+eQkpqWmGyfq2rYhZYxtBQu5pJEACJGBrBPivCVvbMa63WBKQfElRUVGQk/K8vLzu+I6So0p+JMROwutESBIPquwSkzdp0gT79+9X+x87dkz1ShLRx8Uls9u35I86deqUeqrg9evXVe8o8ZLSTDyqVq9eDRGnJMRQPLPEg8rNzU3rkukqubXkhELxuMqtSbJ4Sb4uQtg55aRDydulCV7CLT4+PktPqtwyyu26Slr/nYev4+UvDiHR7FvaGoEe6KN8GHawzzgxsqSx4fuSAAmQAAmQQGES+OffCGzbfzXTlCP7VMWYATWULzAz3WIDCZAACdgEAQpUNrFNXGRxJ2BnZ4dy5crl+jUlXE9yQ+XUpL+EAd7JxNNKPLv0pwCaPyNr1idhN7+vr4vwlpf3048hApz86E2EKk2s0rfry7llpH+W5XQCv+4Mxds/HEPqTePJiQ2U/BbdWvjxgzD/opAACZAACZBAIREQYUoEqmMEwAAAQABJREFUKr3ZKZ/bJg2tjSGdjSce6/uwTAIkQAK2QIAClS3sEtdIAiRAAkVEYOGG85i/4hSUHPwGa9XAG20aGgVDQwdWSIAESIAESIAECoxAWtotbNwVBgnt05ujgx2mP1ofXZpV0DezTAIkQAI2SYAClU1uGxdNAiRAApYlkKYoUnN/OomlSvJVvYl3nXwIblSrjL6ZZRIgARIgARIgAQsRkDxTa7ZfhiRF15u7iwPeHNcITeukH2Cjv8cyCZAACdgiAQpUtrhrXDMJkAAJWJCAfBCe+e1RbNpzxTCL5Jnq3cYfNSt5GtpZIQESIAESIAESsAwBOaFvxZYQXDE7oMS7tBNmTwhCLf5Otgx4jkoCJFAkBChQFQl2TkoCJEAC1kngRuJNvPjZQew9Zsxv4exkh/4dAhBYIftk+Nb5RlwVCZAACZAACdgmgZi4FPyiiFNRscmGF5DfxXMmBsG/vKuhnRUSIAESsHUCFKhsfQe5fhIgARIoIAIRMcmYPDcYJy7EGkZ0d3XAYCXxavkyzoZ2VkiABEiABEiABCxD4Gpkkuo5FZ+YapigblUvvPtkY5TxdDK0s0ICJEACxYEABarisIt8BxIgARLIJ4FLVxPwzEf7IVe9lfVyUsUpL3dHfTPLJEACJEACJEACFiJw8coNrN52CckpaYYZWtXzxhtPNISrs72hnRUSIAESKC4EKFAVl53ke5AACZBAHgmIx5R4TokHld78vF0wsFMgPwjrobBMAiRAAiRAAhYkcFL5nbz+71DcvGk8PrdXKz9MHVkPkg+SRgIkQALFlQAFquK6s3wvEiABEsgBAck1JTmnJPeU3qpVdMfd7SpCjq+mkQAJkAAJkAAJWJ7AgROR2LIvHMpBugZ7oEdlPHlPLSgH6dJIgARIoFgToEBVrLeXL0cCJEAC2ROQU/rktD45tU9v9ap5oXtLP9jZ8ZOwngvLJEACJEACJGApAn8dvIbdR64bhhdBSoQpEahoJEACJFASCFCgKgm7zHckARIgATMCSzdfxNyfTiLN7Gva5vXKoX1jH7PerJIACZAACZAACViCgPwe3rz7Co6ciTYML6F8EtInoX00EiABEigpBBi7UVJ2mu9JArchcEv5cHTu3DkkJibeptedbx09ehRubm5wd3fHyZMn7/zAfz1GjBgBR0dHjB49OsfPsGPeCIgeNW/5KXy07IRBnJJvaTs1rUBxKm9Y+RQJkAAJkAAJ5JpA6s00JRn65UzilCRBf1s5qY/iVK6R8gESIAEbJ0CBysY3kMsngYIg0L59e1SrVg29e/fO13Dx8fFISEjAjRs3IOWcWkREBFJTU3H9utG1PafPs1/OCKQqCVdnfXcUCzecNzxgr4Ty9W7jjyZ1yhraWSEBEiABEiABErAMgcTkm/j59xCcvRRnmKCMpyM+erop5MQ+GgmQAAmUNAIM8StpO873JYEsCGjeTiEhIVncZVNxIJCYdBMvf3EIOw8bRUAnRzv0ax+Ayn5uxeE1+Q4kQAIkQAIkYPUEYm+kYPkfIZlOz/Uv74o5E4MQWIG/k61+E7lAEiABixCgQGURrByUBGyLwOuvv45t27ZhwIABtrVwrjZHBKLjUjDlk2AcPRtj6O/mYo9BnQNRoayLoZ0VEiABEiABEiAByxC4Hp2E5VtCEHcj1TBBrUqemD2hMbxLOxvaWSEBEiCBkkSAAlVJ2m2+KwlkQ+CJJ56A/NCKH4GwiEQ88+F+XLhyw/BypT0ccU+XSpArjQRIgARIgARIwPIELl9NwKo/L0HC+/TWVAmxf3NcI7i78J9mei4skwAJlDwC/L9gydtzvrEVEdi0aROOHDmCOnXqZJv/adGiRQgPD0evXr1w1113qauPiorC+vXrERwcrOZt8vLyQsOGDTF06FC4uroa3jAmJgbff/+9mrj8kUcewfbt27F06VKkpaVh4MCB6NGjB7Zu3YoDBw6gadOmkHxUesvNXPrnpCy5pRYuXAhJnu7n56eO3bFjRzUhunnf29UlN9WWLVuwf/9+yPs0atQIbdq0Qf369bN8LDQ0FCtWrMDx48eRkpKi8h0yZAgCAgKy7F9cG08reS2e/SgY15Rva/UmHlODOgfAjR+E9VhYJgESIAESIAGLETgdEodf/7oMyQepty7NKmD6o/Xh6MDUwHouLJMACZRMAhSoSua+862thMCXX36JxYsXq6KSiFAeHh6GlYnA8uCDD6ptCxYsUAWqd999F6+88oqaiNzQWanMmDEDq1evVgUc7Z6MMWHCBNjb26tizdixY1VxSu6LcCQC1TvvvIN169apgpVeoMrtXNqccl27di3mzJmTKfF59+7dsXz58kzvqn9WXxYh7uGHH1ZFOn17KeXYuRdffBGvvvqq+m7avWXLlmHUqFGIizMmHX3hhRdw+vRp+Pv7a12L9TX4RCRemHcQcQnGEALJNSU5pyT3FI0ESIAESIAESMDyBA6fjsbmPVcgpybrbYgSZj9paG3YyVG6NBIgARIgAfBfKPxLQAJFSEC8esTk5LtVq1ZlWol4OonZ2dmZPKxmz56tilPi7SQCzXvvvYdhw4ZBBJsLFy5g/PjxmcaRhps3b6r3xHOqefPmaNy4MZycnLLsqzXmdS55/qWXXlLFqW7dumHcuHGoUaOGOqx4jY0ZM0ab4rbXnTt34u6771bFqVq1auGNN97Ahx9+iE6dOqkf8mbOnIm5c+eaxhBPK02catasGT7//HPMmzcP/fv3R2JiIiIjI019i3Nh6/5wPPNxcCZxqk4VTwzsFEhxqjhvPt+NBEiABEjAqgjsUg4n2bQ7zCBOiR71+MAaePqBOhSnrGq3uBgSIIGiJkAPqqLeAc5fogn069cPpUuXRnR0tOpJpXlLaVCWLFmiFrt06WIKT5syZYoqMHXu3Fnrpl4lfE0EpR07dqjeQ+beWNIpNTUV4hU1efJkw7PZVfIzl4ODA7766iuMGDFCHT4+Ph4S3rdv3z789NNPeP/99+/ozSTrlG8bGzRooL6XhDKKiUfYoEGDVG+xWbNmQbzCnJ2dsWfPHpPn1Jo1a9SwQukv98+fP4/KlStLtVjb8q0hmLPkhOIlZ/yWtomS36JjkwqKkFmsX58vRwIkQAIkQAJWQUCcpf745woOnowyrMferhSeG34X+rataGhnhQRIgARIQHHMIAQSIIGiI+Di4oJ7771XXcCGDRsg+Z40k9xU8iM2fPhwrVkVl8zFKbnZt29fUx8RY7IyEcByKk7J89I3r3N98MEHJnFKxnJ3d8f06dOlqAplv/76q1rO7o8TJ07gr7/+Um/Lc5o4JQ3iUaa9h4RGSuiemKNjRsJvEer0VqVKFdXLTN9W3Mpfrj6D9xYdzyROtW/sg05NKU4Vt/3m+5AACZAACVgngZtKnqm1Oy5nEqdcnOwxa2wjilPWuW1cFQmQgBUQoAeVFWwCl1CyCTz00EOQXFTJyclqbqZHH31UBaKF90nS83vuuScTJAlZO3PmDCQhuISunT171tRH7mVlzz77bFbNd2zLy1zt2rXLNK7kn9IsJCREK2Z5PXnypKl9165duHjxoqkuhbCwMFNdBKp69eqhRYsWKF++PK5du4YHHngAI0eOxMSJE9VwRlPnYlgQb6nZP/6LVdsvG97OTvmWtntLP9Srlu55ZrjJCgmQAAmQAAmQQIETSEpOw+ptlxASbjw918vdEW+Pb4yGNUoX+JwckARIgASKCwEKVMVlJ/keNktA8ikFBgZCBBtJmG4uUEn+JL33kIQDPvXUU2qYnHki8DtBEM+j3Fh+5spqHvGikrxXIsZJvqjbmeYVJX0kz9btTPJqiQknCR+8//771bxVEmIoPxJaKLmrgoKCbjeMTd5LSknDKwsOY9uBq4b1y2lAfdtVRNWK7oZ2VkiABEiABEiABCxDIF45mGT5lhBcizKenutbzgXvTQxCVT/+TrYMeY5KAiRQXAhQoCouO8n3sFkCIhpJknPJDfX777+r3j+XL1/Gv//+q76TPrwvKSlJTRquhb7JiXuS8Fy8hkTgmj9/foFxsMRckk9KO8EmqxxZ+sVLTinN5JTBcuXKaVXD1c3NDX369DG1ieAn3leSIP2zzz7DqVOn8Oeff6J169YQTyxJDl9cLPZGKp7/9AAOnsoIDZV3c3W2V5Oh+3m7FJdX5XuQAAmQAAmQgFUTiIxRPOEVcSomPsWwzuoVPTD7f0GoUDbjc42hAyskQAIkQAImAhSoTChYIIGiIyAilAhUksRcPIC0cDYRnnr37m1amAhTmjglYtQTTzxhunfw4MECFagsMZeEJKakpH9wk5xQt7Pq1aubbkuSeBHicmriSSU5qp555hk1fPLJJ5+ECG4fffSRWs/pONbcLzwyCZOVk/rOXI4zLFNCCAYrx1aX9br9CY2Gh1ghARIgARIgARLIM4Gw64lYqRxSkpB00zBGo5pl1LA+Tzf+k8sAhhUSIAESyIZA7uJ9shmEzSRAAvkj0KhRI/WkOhlFwvy0/FMSqqZP/C0n4Gk2ePBgrahe9TmZDDfyWMnvXJLYXEL59Pb999+bqpIz6nZWu3Zt0+3vvvvOVM5NQbzTxowZg65du6qPiWdacbBzYfEY9+7eTOJU+TLOGNqjMsWp4rDJfAcSIAESIAGbIHDucjx+/v1iJnGqg3JAyZxJTUBxyia2kYskARKwEgIUqKxkI7gMEtBC+bZu3aqGpQkRrU2jI7mqNFuwYIEpXG7dunW47777tFsFcs3vXKtXr4bkz9q/f7/qGbZs2TLMmjVLXVvTpk0h4Ym3M/GwGjJkiNpFQvXE+0nLNSWNkhhe8ku98cYbpmEkFHD06NGmU/3khiSPDw4OVvvUqlXL1NdWC4dOR2P8u//gSoQxEX5gBTfc160y3F35La2t7i3XTQIkQAIkYFsEjp6NwSolIXpKanouTG31A9pXxMwnGsLZkf/U0pjwSgIkQAI5IcD/a+aEEvuQQCEQePDBB1GqVCnTTDVq1ECbNm1MdSl069YNvr6+atu0adNQoUIF9adv376Ij4839M1vJb9zeXt7Y+PGjRAxSk4iFG8wCe+TJOlvvvlmjpb39ttvQ8YRT6xJkyZB8laJt5kITRL+KGKU3rtKEquLaCX369Spo4YFyvXKlSvqGqS/LduOg9fw9If7M+W3qFnJUw3rc3bi/9JteX+5dhIgARIgAdshsPdoBDb+Hap8eXbLsOhH+1bDc8PrQk7SpZEACZAACeSOAP81kzte7E0CFiNQqVIlQ76pUaNGZZpLRJlVq1aZTqO7du0arl69ivr162PLli2oXLmy+ow+LFAEIc30ice1NrlqffT38zKXNq+ISjt37kSHDh3UaSS3lpiIRZIIvmfPnmpd+yOr+eWeiHRHjx7FAw88AEmGnpCQgEOHDpk8zFq0aIHXX39dGwbjx49Hu3bt1PqJEyfwzz//qKJYs2bNIF5mtpwgfe1fl/Hi/INITDbLb1GrjHpan709Pwib/iKwQAIkQAIkQAIWIqCc94Kt+8Kx3ez0XBGknh1WB6P7V7fQzByWBEiABIo/AcaCFP895hvaEIG1a9ciOjpa+dbNDpLoOytr2bKlGjYXGhqK8+fPo2bNmqo3kfSVk//ES0n/rIgyMTExqndWdifnSfid9CldurRhytzOFRQUpK5fBCcXFxf19Lxz584hNjYWAQEB2Z7El938shjxElu0aJEaznjhwgVERUWpebkkBNDd3Xhcs7zr9u3bERERoYb2SUhgtWrVTHwML2dDle9+PYcvVp1WGBgX3aZhebRq4G1sZI0ESIAESIAESMAiBG4q3lLiNXX8fKxhfCcllG/GqPro1KSCoZ0VEiABEiCB3BGgQJU7XuxNAhYlICF+ZcqUydEc/v7+kB+9SSid/Jibp6eneZOh7uDgkK14JB1zM5deHJNnq1atKpfb2p3ml4eFjYhSdzr9T/qWK1futu8jfWzB0hRF6sMlJ/Czcmy13oRFtxa+aFDDKCjq+7BMAiRAAiRAAiRQcASSU9KwZvslXAi7YRjUQ8n9+Na4RgiqXdbQzgoJkAAJkEDuCVCgyj0zPkECJEACFicgCVdf+/oI/vgn3DCXgxLK16dtRdQI9DC0s0ICJEACJEACJGAZAjcSU7FiyyWERxoPKClf2hnvTQxCjQD+TrYMeY5KAiRQ0ghQoCppO873JQESsHoC8coH4anzDmLf8UjDWl2c7DGgYwAq+mT2kjN0ZIUESIAESIAESKBACETHpeCXPy5Crnqr7OuG9yc1gV85F30zyyRAAiRAAvkgQIEqH/D4KAmQAAkUNIHr0UmYPPcATl405rfwcHNQT+rzVr6tpZEACZAACZAACViegHhMrVDC7G8kGg8oqVfNC+8+GYTSHo6WXwRnIAESIIESRIACVQnabL4qCZCAdRMICb+Bpz8KRui1BMNCy3k5YXCXQHi68YOwAQwrJEACJEACJGAhApJrSnJOSe4pvbVRDid5fUxDuDjb65tZJgESIAESKAACFKgKACKHIAESIIH8Ejh2LgZTPglGVKwxhMC/vCsGdgqAhPfRSIAESIAESIAELE/g+PkY5bS+MMipfXrr08Yfzw+vC8kHSSMBEiABEih4AhSoCp4pRyQBEiCBXBHYdfQ6XvrsEBKSjCEE1ZSkq33b+SsfhO1yNR47kwAJkAAJkAAJ5I3AfiX/45/7w6EcpGuwh3pVwdhBNZVThQ3NrJAACZAACRQgAQpUBQiTQ5EACZBAbgls2BWGN787itSbxk/C9auXRreWvrDjJ+HcImV/EiABEiABEsgTge0HrmLv0QjDs/J7eMK9tXB/t0qGdlZIgARIgAQKngAFqoJnyhFJgARIIEcEFv92AZ/8cjLTt7Qt63ujbaPyORqDnUiABEiABEiABPJHIE0J5du0OwxHz8YYBnJ0sMO0h+uhewtfQzsrJEACJEACliFAgcoyXDkqCZAACWRLQMIGRJgSgUpv4izVuWkFNK5dVt/MMgmQAAmQAAmQgIUIpKSmYd2Oyzh7Od4wg5uLPWY90QjN65YztLNCAiRAAiRgOQIUqCzHliOTAAmQQCYCEsonIX0S2qc3eyXhau/W/qhV2VPfzDIJkAAJkAAJkICFCEjux5VbQxB2PdEwg5yeO3tCEGrzd7KBCyskQAIkYGkCFKgsTZjjkwAJkMB/BOSDsCRDl6ToenNytEP/DgGo5Oumb2aZBEiABEiABEjAQgRi4lOwfEsIImOSDTME+Lji/YlNIFcaCZAACZBA4RKgQFW4vDkbCZBACSUQFZuMKZ8cwLFzxvwW7i4OGNQ5ED5lnUsoGb42CZAACZAACRQugWtRSao4FZ+QaphYPKbEc0o8qGgkQAIkQAKFT4Bnlxc+c85IArkiMGLECDg6OmL06NG5eo6drYdA6LUEjH33n0ziVBlPJwztUZnilPVsFVdCAiRAAiRQzAmEhN/Ass0XYC5OSa6puc80ozhVzPefr0cCJGDdBOhBZd37w9WRACIiIpCamorr141hYURjGwROXozF5LnBuB5tDCHwLeeiek65OtvbxotwlSRAAiRAAiRg4wROKb+T1+8MheSD1Juc0ien9cmpfTQSIAESIIGiI0CBqujYc2YSKFQCUVFRmD9/vjrnhAkT4OHhUajz6yezprXo11XQ5X3HIzF13kHEJxpDCKr4u6Nf+4r8IFzQwDkeCZAACZAACWRD4ODJKPzxTzhuyVG6Oru/WyVMuLcW7OQoXRoJkAAJkECREqBAVaT4OTkJFB6BEydOYOrUqeqEd999Nxo1alR4k5vNZE1rMVtagVXlQ/BrXx+BHF+tt7uqeqFnKz/Y2fGDsJ4LyyRAAiRAAiRgKQI7D13DrsNGT3TRo8YOqomHelWx1LQclwRIgARIIJcEKFDlEhi7kwAJkMCdCPysnAr04ZITSDP7lrbZXeXQoYnPnR7nfRIgARIgARIggQIgIL+GN+8Jw+HT0YbRHOxL4fnhddGnjb+hnRUSIAESIIGiJUCBqmj5c3YSyBcBCZVbv349goOVHEdKjiovLy80bNgQQ4cOhatr+vHIiYmJWLp0KQ4ePGiaa+XKldi/f79ab9u2LWrVqmW6J4Xdu3djz549OHLkCCpWrIigoCB06dIF7u7uhn4xMTH4/vvv1fZHHnlEzZe1ceNG7NixA35+fmjVqhW6d+9uesaSa5FJQkNDsWLFChw/fhwpKSmoU6cOhgwZgoCAANMaLFmQD8JfrDqN7349l2kaEaZEoKKRAAmQAAmQAAlYnoDkmfr1r8s4HRJnmMxFyf34+piGaNPA29DOCgmQAAmQQNEToEBV9HvAFZBAngi8++67eOWVV3Djxo1Mz8+YMQOrV69Ww/gOHDiAhx9+2NBn+vTppvqLL76ImTNnqvXY2FiMHz8eP/zwg+m+VhCxZ9myZaoAprWJECT5rOzt7VWR65577kF4eLh2W73K3F9++aXax5JrkbWNGjUKcXHGD6IvvPACTp8+DX9/y35LejPtFt754V+sVT4M601C+SSkT0L7aCRAAiRAAiRAApYnkJR8Eyv/vITLVxMMk5X2cMS7TwahXjX+TjaAYYUESIAErIQABSor2QgugwRyS2D27NmqONW0aVP07t0b3t7e2Lt3LxYvXowLFy6oQtP27dtRs2ZNPPDAA7h69So2b96sTtOjRw+UKVNGLYtnlGb33XcfNmzYgFJKYoaRI0eidevWOHfuHObOnat6JfXq1UsVezTvLO25mzdvomvXrkhOTkaDBg3QsmVLrFu3DmFhYfj222/V9ckaLLUW8R7TxKlmzZrhiSeegKxJ1rBmzRpERkZaVKBKVD4I/5+98wCPqujC8CEh1IQSQg2h19A7KB1EilJUUFGRogKigBTlFwQVsKCAFBUEERQbFhAElI4UAZHeewk9BAgJBALh329wLvdudpNN2ZBNvvM8yZ12Z+a+l5J8e86ZEdN3ybodoRqJuuI0ICRDR1J0GgmQAAmQAAmQgPsJRFy7JXNtofYXr9ywLFbAdnruuH7VpEj+bJZ2VkiABEiABFIPAQpUqeddcCckkCACgwcPlpo1a0rjxo0t9yGcDeIVwuzgTQTh6vvvv1dhewi5g6HfPkk6QvMgTsGmT5+uBB9VsX2DZxTEKoTQTZ06Vfr376+7jCtC6nBKIMQh2MmTJyU4OFjtYebMmUokc9deEI6oPacgSCG8ENarVy85fvy4FClSRNXd8S08Mlre+Gy77LTLb5HVFkLQvnFhyW/7gZhGAiRAAiRAAiTgfgJh4Tdl7soQuXot2rJYyUBfGftqVQnIldnSzgoJkAAJkEDqIuCVurbD3ZAACbhKYNCgQbHEKdzbpk0bYwqIM64aRCRYhQoVLOIU2uARpYWw9evXoymWffLJJ4Y4hc6goCBp3769GpeQfeCGhO7Fx8dHrYNvEObMVrRoUeURZm5LrvK5sCjp/dG/scQphBA8+VARilPJBZrzkAAJkAAJkEA8BM6EXpc5S0/EEqeqlsklnw6sQXEqHn7sJgESIIHUQIAeVKnhLXAPJJAEAkg8fuTIEeXdhFC2o0ePGrOhz1U7ePCgGnrHlul73LhxsW5DiCAM+ZwcWcOGDWM1FytWTLXpe2MNcNKQ0L3UqlVLAgICJDQ0VHlqITyxb9++UqVKFScrJL35yOkIGTRpm5y/ZA0hyJs7s7RvVFiyZ+U/r0mnzBlIgARIgARIIH4CR05FyKL1Z+TWrRjL4EbV8smI7hUkkw8/k7eAYYUESIAEUikB/gaVSl8Mt0UC8RG4cuWKCrX7+eefjfC2+O6Jq18LT3v27JGBAwc6HRoTY/3hz+lAW4efn5/qTsg9uCGhe8HpheDQqVMnlaR9xowZgi+IZhMmTFCnEKqNJNO3HYcuq7C+q7Y8F2YLsuW1eLRBIH8QNkNhmQRIgARIgATcSGD3kSuyfNM5icFRuibrYPuw6LUnywgOK6GRAAmQAAl4BgEKVJ7xnrhLErAQuHHjhrRu3Vp0uF39+vVVPip4EYWEhKhcUJYbXKhkznw3LwPm6tq1q9M7EO7nbkvMXho1aiTwvPriiy9UnqxDhw7JX3/9pXJnbdy4Mdm8qdZsvyBv2xKi34i2CnWli/hJy3oFxZs/CLv7jwfnJwESIAESIAFFYNPui7Le7oASdPR4tIR0a1OclEiABEiABDyMAAUqD3th3C4JgACEKS1OmROTo2/Hjh3xClSOPJpKlCihTt3LmTOn9OjRA1OliCXnXuBJhdxcAwYMkC+//FL69OkjEPMmTpyo6kl9oPlrT8nH3+2XmBjrp7RVyuSWxtXz2XJdJXUF3k8CJEACJEACJBAfAThLrdpyXrYfuGQZCm+pQZ3LStv6gZZ2VkiABEiABDyDAAOyPeM9cZckYCGwZcsWo96hQwejjMLZs2ctdV3JlCmTLipPI6PyX6Fs2bKqtGzZMpXPyr4/Oevu3ouXl5e8+OKL0rRpU7Xt06dPJ3n7Xy08KmNm74slTj1QOUCa1KA4lWTAnIAESIAESIAEXCBw2/Yh0aL1p2OJU5lteaZG96xEccoFhhxCAiRAAqmVAAWq1PpmuC8SiINA4cKFjd7p06cLEpvDFi1aJB07djT6zIWCBQsa1R9++EGQUP3MmTOGoNWvXz9bngYv5XEE0QuJ17Vh/s2bN8vgwYOVh5ZuT+w1ufcyZswY5fWlc1dhX0gWv23bNrXF0qVLJ3arSpAa+/1++XLBPR6YzMvmLvVQnQJSu0KeRM/NG0mABEiABEiABFwncNMWXj9vVYgcPHHVcpNftowyvl81aVAlr6WdFRIgARIgAc8iQIHKs94Xd0sCikCzZs0kf/78qjx06FDJly+f+mrTpo1ERkY6pITxVatWVX2//vqrQCQKDAxU+ZrQiBPvcPIdDDmbSpYsKUFBQaodua1wUt7HH38su3btUmOS8i259wJhCknRIUTBE6xmzZrqeu7cOcmaNWuiQxbxg/Bb03bJ3NUhlsfNmNFLHmlQSCqUyGlpZ4UESIAESIAESMA9BCKv35Kflp+Qk+euWRbIZzs999NBNaRyqVyWdlZIgARIgAQ8jwAFKs97Z9xxOiOgw+F04nA8PgSj+fPnG4JTaGioXLhwQSpUqCCrVq2SIkWKKEo+Pj4WWvC2Qq4pGHIzwfRYlMePHy/wrsKYDDYPISRcR06rsLAwyZMnjxJ6WrRogaHK9N5QMe/vbq+I7tdX3Y5rcu7l5ZdflgcffFBNf+DAAfn3338lOjpaatSoobzKIL4l1CJsJ/QNmLRVVm89b7k1SyZvebxJYSkR6GtpZ4UESIAESIAESMA9BC5fvSlzlp2QC5fu/uyiVylaILt8PrimlCjE/5M1E15JgARIwJMJZLCF7liz/Xry03DvJJAGCdy6dUvCw8MFycu9vb1jPSHC9I4fPy6lSpVSwhUGXL9+XQk0SBpub/grf+zYMbl48aIULVpU8uZ17A4fEREhJ06cUEIWBDF4Uzmyq1evKjHL1zf2D4dY6/Lly5ItWzaHAlZy7wVCGkL7kHi9ePHiBg9H+46rLfTyDRk4aZscPhVhGeaXzUc62MQp/xz38nlZBrCSpgh8YgvtNNvaKc3MVafl4FajLX3+Reta6roSdnyDLqrrnsVDLXVW0haBf/aGyWsTthoPVau8vwpJMhpYIAEHBFz990Tf6uq/K/V7Lde3qGv/p+/mobQ0ppLKubAoFdZ3/cZty44qlcwpH75cRXJkt34YZxnECgmQAAmQgEcR4Cl+HvW6uNn0SCBjxozi7+/v9NERqmfO6YSBCGvDlyODZxTEG3zFZRCcgoOD4xqi+vz8/JyOwVq5c+eOsz859wJOcbFyuhFTxwlb6MDAiVvlzMUoU6tInpyZpUPjwuJry3NBIwESIAESIAEScD+B42ci5fe1pyX6VoxlsQdtB5S880JFgVczjQRIgARIIO0Q4G9aaedd8klIgASSSGDP0XAZ/Ok2uRIRbZmpUN6s0q5hoGTmD8IWLqyQAAmQAAmQgLsI7DsWLks2no11em6bBwrJ68+WE2+vDO5amvOSAAmQAAncJwIUqO4TeC5LAiSQugj8veuiLSH6TomyCyEoWdhXWtl+GM7ozR+EU9cb425IgARIgATSKoF/94XJmq0XYj1el1bF5MW2JW2pBWJ1sYEESIAESCANEKBAlQZeIh+BBEggaQQW/31GPpy9V27dtqbkq2jLb9GsVgH+IJw0vLybBEiABEiABFwmAGEKApXZvGyKVL8ny8jjtlB7GgmQAAmQQNolQIEq7b5bPhkJkIALBL7987hMmXdI7I+LqFMxj9SrFODCDBxCAiRAAiRAAiSQVAIxMXdUSB9C+8zmk9FLhnerIE1q5DM3s0wCJEACJJAGCVCgSoMvlY9EAiQQP4EYmyI1+eeDMmf5SctgJHbHD8GVS+eytLNCAiRAAiRAAiTgHgJIgo5k6EiKbrbsWTLK+70rS/Wyzg9cMY9nmQRIgARIwLMJUKDy7PfH3ZMACSSCAH4QHj1rjyz755zlbuSZalmvoJQKcn4yoeUGVkiABEiABEiABJJE4Lot9+O8VSFyLsz+9NxM8vErVaU0/09OEl/eTAIkQAKeRIAClSe9Le6VBEggyQSuRd2WN6fukM17rfktMmfykkcbBErhfNmSvAYnIAESIAESIAESiJ9AuO3U3F9t4tTlqzctg/F/8fi+VaVgQFZLOyskQAIkQAJpmwAFqrT9fvl0JEACJgJh4Tdl0ORtcuDEVVOrSPasGaWDLfFqQK7MlnZWSIAESIAESIAE3EPgwqUbMm91iERev2VZoHyxHPJRnyqSyy+TpZ0VEiABEiCBtE+AAlXaf8d8QhIgARuBUxeuy4CJW9XVDCR3jkxKnMqR3cfczDIJkAAJkAAJkICbCJw8d00WrDklN6NjLCvUCc4jo3pWkqyZvS3trJAACZAACaQPAhSo0sd75lOSQLomAI8peE7Bg8psBfJkkXaNCvMHYTMUlkmABEiABEjAjQQO2v5P/mPDGbl9+45llYfrFJD/dQkW5IOkkQAJkAAJpE8CFKjS53vnU5NAuiGAXFPIOYXcU2YrXii7tH6wkOD4ahoJkAAJkAAJkID7CWw/cElWbTkvtoN0LfbUQ0Wkz2OlxXaQLo0ESIAESCAdE6BAlY5fPh+dBNI6gWWbz8nomXsEp/aZLbh4Dmleu4B4efEnYTMXlkmABEiABEjAXQTW7wiVTbsvWqaHIAVhCgIVjQRIgARIgAQoUPHPAAmQQJokMGf5SZn880GJsfuYtmawv9SvkjdNPjMfigRIgARIgARSGwH8P7x80znZfeSKZWsI5UNIH0L7aCRAAiRAAiQAAhSo+OeABEggTRGAHjVl3iH59s/jlufCp7QNq+WTamVzW9pZIQESIAESIAEScA+BW7djZNG6M3LkVIRlASRBRzJ0JEWnkQAJkAAJkIAmQIFKk+CVBEjA4wncsiVc/XD2Xln89xnLs3jbQvla1C0gZYvmsLSzQgIkQAIkQAIk4B4CUTdvy2+rT8mZ0OuWBXL5+chHfapK+WL8P9kChhUSIAESIAF6UPHPAAmQQNogEHXjtrw1baf8vcua3yKTj5c8Uj9QihTIljYelE9BAiRAAiRAAqmcwNVr0TJ3ZUis03MLBmSV8X2rSuF8/D85lb9Cbo8ESIAE7gsBelDdF+xclARIIDkJXImIlsGfbpM9R8Mt02bL4i3tGxeWfLmzWNpZIQESIAESIAEScA+Bi1duyNxVIRJx7ZZlgdJBfvLxK1UkT87MlnZWSIAESIAESEAToEClSfBKAiSQ7ASCW41O0px7Fg+N9/6zYVEyYMJWOXHummVsTl8feaxJkOBKIwESIAESIAEScD+B0xeuy/y/TgnC+8xW3Zb/8f3elSV7Fv7qYebCMgmQAAmQgJUA/5ew8mCNBEjAgwgctiVdHThxm4TaPq01Gzym2jcOlGz8QdiMhWUSIAESIAEScBuBwyERsnj9aUE+SLM1qZFPhnerID4ZvczNLJMACZAACZBALAIUqGIhYQMJkIAnENh24JIM+XyHRFy3hhAg1xRyTiH3FI0ESIAESIAESMD9BHYdviLL/zknd3CUrsket4XZ93uyjHjhKF0aCZAACZAACcRDgAJVPIDYTQIkkPoIrN56Xt6ZsVtuRsdYNle2qJ/ttL6CglP7aCRAAiRAAiRAAu4nsNF2OMnfO0MtC0GPerFtSenSqpilnRUSIAESIAESiIsABaq46LCPBEggWQn4F60b53xhxzfE2a8735q2S2JirJ/SVrPlt2hYLZ/wQ1pNiVcSIAESIAEScB8BOEut/Pec7Dh42bIIPiR6/dly0uaBQpZ2VkiABEiABEggPgIUqOIjxH4SIIFUR8BenKpfJa/UDPZPdfvkhkiABEiABEggLRK4bcsztfjvM3Lo5FXL42XJ5C3vvFBRHqwcYGlnhQRIgARIgARcIUCByhVKHEMCJJCqCVCcStWvh5sjARIgARJIQwRu3IyRBWtOSch56+m5ObL7yIcvV5FKJXOmoaflo5AACZAACaQkAQpUKUmba5EACZAACZAACZAACZCAhxKItB1MMndViIRetp6em98/i4ztW1WKFcjuoU/GbZMACZAACaQGAhSoUsNb4B5IgARIgARIgARIgARIIBUTuBR+U4lT4ZHRll2WKOQrH79aVfLlzmxpZ4UESIAESIAEEkqAAlVCiXE8CZAACZAACZAACZAACaQjAmcvRslvq0Pk+o3blqeuXCqXCuvzy8ZfKSxgWCEBEiABEkgUAf5vkihsvIkESIAESIAESMBMIPRShGzeeVK27wuRcxevyoWwCAm9FKmuGJfX31cCcmdX1/x5/KRKucJSs1KQrc3XPA3LJEACqYzAsdORsnDdaYm+FWPZWQPbASVv2xKiZ/bxsrSzQgIkQAIkQAKJJUCBKrHkeB8JkAAJkAAJpHMC4RFRsmDFTvl95W7Zsf+U4Nh5Z3b8VJjg655tkgwZRCqXDZRHmlSQR5tWkhy+We51s0QCJHDfCew5Gi7LNp0V+9Nz29YvJIM6lxMvL9tfYhoJkAAJkAAJJBMBClTJBJLTkAAJkAAJkEB6IXDl6nWZNXeTzP7tH4m4Zk2WnBAGELS27zulvibMWi3Ptqslz3eoLTn9siZkGo4lARJwA4HNe8Jk7fYLsWbu1qa49Hi0RKx2NpAACZAACZBAUglQoEoqQd5PAiRAAiRAAumEwO3bMfL97//K5Nl/CbynktMgdE35fq18t2CzvPJsQ3n6kRri7c3QoeRkzLlIwBUCEI7/2npetu6/ZBkOb6nXniwjHRoVtrSzQgIkQAIkQALJRYACVXKR5DwkQAIkQAIkkIYJHDl5UYZ89JvsOnjG6VMWvXZRGp3fL1Uvn5S8N65KwI0IdcUNFzL7SWhmX3XdlitIVucrK8ez5Yk1F4Sv96YskfnLd8oHg9tJiaDYY2LdxAYSIIFkIXA75o4s2XBG9h+/apkvky3P1IjuFaRRtXyWdlZIgARIgARIIDkJUKBKTpqciwRIgARIgATSIIGVGw/K4A/nybXrN2M9HUSoZ49vkIfO7ZbikaGx+nUDxCt8wVqe3SVD9i2Wo9kDZGn+CjK7aF0lXumxuEII69Rvhnz0RntpUqe0uYtlEiABNxC4GR0jv689JSfOXrPM7ps1o3zQu7JULZPb0s4KCZAACZAACSQ3AQpUyU2U85EACZAACZBAGiKwdN0+GfD+XEF4n9n8oqPkxaN/KXEqy+1oc5fLZQhaLx1ZLV2Or1ci1bTiDeWqz71E6RDE+o78Wcb9r4M89GA5l+flQBIggYQRuBZ1S+atOiXnL1lDdwNyZpaxfatKyUCetpkwohxNAiRAAiSQGAJM7pAYaryHBEiABEiABNIBga17QuT1Mb/FEqcePrtbFq/5RF44skYSK06Z8WEOzIU5MbfZIIxhD9gLjQRIwD0Eflx6IpY4VSR/Nvn89ZoUp9yDnLOSAAmQAAk4IECBygEUNpEACZAACZBAeidw9kK4vGrzXrpx85aBwvtOjAzdu1DGb/tB/G9GGu3JVcCcmBtrYC1t2AP2gj3RSIAEkp/AlQirF2Rw8Rzy+eCaUjDPPY/G5F+VM5IACZAACZCAlQAFKisP1kiABEiABEgg3ROA19LAD+ZK2OV7IlQGuSOjd86VZ2z5ptxtWANrYU1t2Av2ZB9qqPt5JQESSB4C9SrmkYn9q0tOX5/kmZCzkAAJkAAJkICLBJiDykVQHEYCJEACJEAC6YXAFz+ujxVS9+rBFdL29LYUQ4C1TmfNJRNLNzPWRJgf9ta7c32jjQUSSE8EwtwsELeqV1DeeLa8ZPTOkJ6w8llJgARIgARSCQF6UKWSF8FtkAAJkAAJkEBqIHA05KJM/WGtZSvNzu2VnodXW9pSooI1sbbZpny/VrBHGgmQQPISeObhovJml2CKU8mLlbORAAmQAAkkgAAFqgTA4lASIAESIAESSOsEPpq+XG5G3zYeM9+NqzJqlzXczuh0cwEhflgbe9AWfeu2YI80EiCB5CHglSGD9O1YRnp3KCW2Io0ESIAESIAE7hsBClT3DT0XJgESIAESIIHURWDn/tOyauNBy6aG7vldckZft7SlZAVrYw9mwx6xVxoJkEDSCPhk9JLh3StIp2ZBSZuId5MACZAACZBAMhBgDqpkgMgpSIAESIAESCAtEJj560bLY9QMOyYPndtjabsfFewBe9nsX8xYHnsd+78ORp0FEkiLBPYsHuq2x8qWxVve61lZapb3d9sanJgESIAESIAEEkKAHlQJocWxJEACJEACJJBGCYRHRMmy9fstT9fr8CpL/X5W7PeCvWLPNBIggYQT8M+RSSYPqEFxKuHoeAcJkAAJkIAbCVCgciNcTk0CJEACJEACnkJg9aZDgvxO2opcuyj1Lh7R1ft+xV6wJ23YK/ZMIwESSDiBzwfXlDJF/BJ+I+8gARIgARIgATcSoEDlRricmgRIgARIgAQ8hcCG7ccsW21xdo8gSXlqMewFezKb/Z7NfSyTAAk4JxCYN6vzTvaQAAmQAAmQwH0iQIHqPoHnsiRAAiRAAiSQmgjsPXTWsp3aYanHe0pvzH5P9nvW43glARIgARIgARIgARLwPAIUqDzvnXHHJEACJEACJJDsBELOXrbMWTrivKWeGir2e7Lfc2rYI/dAAiRAAiRAAiRAAiSQOAIUqBLHjXeRAAmQAAmQQJohcDvmjkRcu2E8D8LpAm5EGPXUUsCezGGH2DP2TiMBEiABEiABEiABEvB8AhSoPP8d8glIgARIgARIIGkE7lhFHi9b3ftOTNLmdMPd2BP2ZjH7uqWTFRIgARIgARIgARIgAU8hQIHKU94U90kCJEACJEACJEACJEACJEACJEACJEACaZQABao0+mL5WCRAAiRAAiTgMoEMGSxDY2z12xlS348I2BP2ZjH7uqWTFRIgARIgARIgARIgAU8hkPp++vQUctwnCZAACZAACaQRAt5eGcQ3W2bjae7YMj2FZvY16qmlgD1hb9qwZ+ydRgIkQAIkQAIkQAIk4PkEKFB5/jvkE5AACZAACZBAkgkULpDLMsdB33yWemqo2O/Jfs+pYY/cAwmQAAmQAAmQAAmQQOIIUKBKHDfeRQIkQAIkQAJpikD5UgUsz7PJv4Slnhoq9nuy33Nq2CP3QAIkQAIkQAIkQAIkkDgCFKgSx413kQAJkAAJkECaIlC3SjHL8ywpEGwJp7N03ocKQvuwJ7PZ79ncxzIJkAAJkAAJkAAJkIBnEaBA5Vnvi7slARIgARIgAbcQaFS7lGTy8TbmPpEtj/ydJ/V4UWEv2JM2n4zegj3TSIAESIAESIAESIAE0gYBClRp4z3yKUiABEiABEggSQRy+GaRZvXKWuaYUrKxpX4/K/Z7af5AWcGeaSRAAiRAAiRAAiRAAmmDAAWqtPEe+RQkQAIkQAIkkGQCXR+rY5ljs38xWZrfGlZnGZBCFewBezGb/V7NfSyTAAmQAAmQAAmQAAl4HgEKVJ73zrhjEiABEiABEnALgUplC0njOqUtc48OfkSu+GS1tKVkBWtjD2bDHrFXGgmQAAmQAAmQAAmQQNohQIEq7bxLPgkJkAAJkAAJJJnA4BeaWXJRnc/sJ8MqdrgvCdORGB1rYw/akHsKe6SRAAmQAAmQAAmQAAmkLQIUqNLW++TTkAAJkAAJkECSCBQvnEd6PlXfMsfy/OVlaslGlraUqGBNrG22Xk/XF+yRRgIkQAIkQAIkQAIkkLYIUKBKW++TT0MCJEACJEACSSbw0pMPSLXgwpZ5JpVuKvMLVbW0ubOCtbCm2bAn7I1GAiRAAiRAAiRAAiSQ9ghQoEp775RPRAIkQAIkQAJJIuDt7SVjh3QQ/1zZjXkQbje0Ugf5tmhdo81dBayBtbCmNuwFe8LeaCRAAiRAAiRAAiRAAmmPAH/KS3vvlE9EAiRAAiRAAkkmUCBvDpn01hOSOVNGY67bGbxkdPk28lrVpyQs0z3xyhiQxALmxNxYA2tpwx6wF+yJRgIkQAIkQAIkQAIkkDYJ3PvpL20+H5+KBEiABEiABEggkQQQUjfm9XaxvJb+LFBBWjXoL9NLNJAob59Ezn7vNsyBuTAn5jYbPKawB/uQQ/MYlkmABEiABEiABEiABDyfwL2PRT30WWZ8O0+++m6eh+6e2/YkAt2faS/dOrf3pC1zryRAAiSQZAIPPVhOJtq8lwZ/OE+uXb9pzHfVJ4uMK9NCvi76gDx7fIM8dG63FI8MNfpdKRzNHiBL81eQ2baQvtDMvrFuyZY1k3z0RntpUqd0rD42kAAJkAAJkAAJkAAJpC0CHi9QUZxKW38gU/PTQAylQJWa3xD3RgIk4C4CEIjmTOguQz76TXYdPGNZBsLSJ2Waq68i1y5K4/P7perlk5L3xlUJuBGhrrjhQmY/JULhui1XkKzOV1aOZ3N+Gl/F0gXlg8HtpESQ8zGWjbBCAiRAAiRAAiRAAiTg0QQ8XqDS9NcsnKmLvJJAshNo0KZrss/JCUmABEjAkwhAKPp+fFf5/vd/ZfLsvyQ8IirW9k/YBKeviz0gX8fqcb0hh28WeeXZhvL0IzVihRa6PgtHkgAJkAAJkAAJkAAJeBqBNCNQeRp47pcESIAESIAEPI0A8kE9266WPNq0osyau0lm//aPRFy7kSyP4Zsts5r7+Q61Jadf1mSZk5OQAAmQAAmQAAmQAAl4DgEKVJ7zrrhTEiABEiABEkgVBCAg9e3SSLo+VkcWrNgpv6/cLTv2n5I7dxK2vQwZRCqXDZRHmlSwiV6VBN5TNBIgARIgARIgARIggfRJgAJV+nzvfGoSIAESIAESSDIBCErPtK2lvkIvRcjmnSdl+74QOXfxqlwIi5DQS5HqioXy+vtKQO7s6po/j59UKVdYalYKsrXFTo6e5I1xAhIgARIgARIgARIgAY8jQIHK414ZN0wCJEACJEACqY8AhKaWDcurr9S3O+6IBEiABEiABEiABEggtRPwSu0b5P5SD4Hdu3fL4MGDpWjRonLy5MnUs7F4djJ37lypWbOmzJkzR2JiYuIZzW4SIAESIAESIAESIAESIAESIAESIIGUJkAPqpQmnsD1Fi5cKPPnz5elS5dKZGSklCtXTrp37y5dunSRDEje4YIdOXJERo0aJdu3b5djx45JjRo1pG3bttKzZ0/x8fGJd4YrV67IwIED5csvv1RjM2XKJAcPHpSgoKBY92Ls7NmzZe3atbJu3To1f926daVevXry9NNPS548yXtc+JYtW2TEiBGx9oGGPn36SMuWLeXw4cPy77//ypNPPimlS5dWItvzzz8veA4aCZAACZAACZAACZAACZAACZAACZDA/SdAger+vwOHO7h586b06NFDiT3mAefPn5e//vpLli1bJt988425y2EZ4haEoWvXrklAQICEhYUpsQuC186dO2Xq1KkO79ONBw4ckKZNm8qpU6dU0xNPPCEffPCBlCxZUg8xrufOnZOHH35YCWFozJUrl3h5ecl3332nvj777DO1dmBgoHFPUgsZM2aU6Oho+fPPPy1TNW/eXIlRaBw0aJBAJMN148aN8tJLLwm8qsAG99NIgARIgARIgARIgARIgARIgARIgATuLwGG+N1f/g5Xh+DSsWNHJU7ly5dPeQKNGTNGypcvb4yHl9Knn35q1B0VLl26JC+++KISp4YOHSoXLlyQ9evXS9asd4/v/uKLL+Tzzz93dKtqu3jxorRp00aJU9mzZ5fVq1fLTz/95FCcgoD14IMPKnEKog/C6SCmQbTq1auXmm/v3r1Sv359Q+xyunACOipXriwLFiwQ7E9b9erVZcmSJZZ9Yt2///5bpkyZooYtXrxYXnjhBduJUwk8ckovksavu45ckfOXotL4U/LxUpLAol9mSLe2lWREv44puSzXIgESIAESIAESIAESIAES8BACdB9JhS8K4XEQXSpUqCAQUnQoHeoQjLQNGzZMEKrm6+v4BCR4OkEkQv8bb7yhbkOo3bvvvqvC4p599llp0qSJns5yvXHjhnTo0EEOHTqk2qdNmyYNGza0jDFXxo4dq0Lp0DZgwAAlsOn+SZMmKa+vPXv2qBBDeFKNHj1adyf5Ci8teJxpQwiko/BHtCGscd++ffLJJ5/IrFmzJH/+/PLhhx/qW3n9j8C/+8Jk5qJj8siDheS5lkUlX+70ffT7HVvusvWrfpdDe7fJscN75MSRfbY/Y16SO08+KRNcXeo0aiWVazTgn584CFy6eF4iwi/L+bOek78ujsdhFwmQAAmQAAmQAAmQAAmQQDIToAdVMgNNjukgGsEDCEKVFqcwL4Qds12+fFmmT59ubrKUf/31V1Vv3Lix+Pn5GX0QkODZhPA+5LRyZBCx1qxZo7p69+6twgQdjUPb9evXZebMmaobOa2GDx+uyvobPKpeeeUVXZWvv/46WZOVI2E7vM60NWvWTBcdXuGNBqEOhvLvv//ucFx6b4y+FSNzV4fIk2/9LWO/359uParOhByVt/o+Lp+820d+/2ma7NqyTsIvX5Qrly7IsUO7Zcn8b2TkwM42z6An5Jbpz2F6//PD5ycBEiABEiABEiABEiABEiCBhBCgQJUQWik4FjmUkMNJG5KP//HHH6pq9piCZ5MjQzid9n5q0MDq2QGPI/Mc9vfDewrhfzDki4K3UVz2448/CsIJYVWqVLGE2+n7WrdurYsSEhKiclEZDUksIPm7NiRhL1u2rK46vEJEQ/4u7WU1YcIEh+PYeJdAehaqtm1aJYN6tJD9uzYrGOUq1ZJOXQfIG6NnyOCR0+TZnm9KqXJVVN+e7Rvl4oXT/GNDAiRAAiRAAiRAAiRAAiRAAiSQCALpKsQPJ9gtX75ccKrdI488YnjRaG4rV65USbSvXr0qVatWlccff1wl+db99/OK/E+3b99WW0DoXv/+/eXWrVuCsDkkO69UqZJle1rMQiNyQyXEkD8qNDRU3YKk6PGddjdv3jxj+jp16hhlc6Fo0aIqSbueF95hSKieHLZhwwZjGqyvhSej0UEBSd6ROB15qfBnAsngy5Qp42AkmzQBLVT9vu50ugj9uxF1TaaOHSI3b0SJd0Yfea7Xm9LmiRc0DuPa7unesmvreln403TJ7pvTaGeBBEiABEiABEiABEiABEiABEjAdQLpRqAaNWqUvPPOO0rUAZ733ntPJfyGAANPI+Qt0iFtGl/Xrl3lq6++0tX7el2xYoWxPsQ11HUIHzyY7AUqLdrAW6hGjRrGva4UzMnXO3XqFO8t+oQ/DHQWMog+iEJaoEKIYXIZRCZtOnRP1+O6IhE97kWidIRPxucpFtdc6akvvQhVc2aOl9Bzd0+v7NpnhMYnHNUAAEAASURBVLTs8LzT11yx2gOCL0d2NfyS7N72txw7uFuuRV6VoiXLS5kKNSSoWGxBFOGDyG9VpVZDCSxaWuW82rllrSB/U7FSFaRanSYq75WjddCGHFmH9m2Xk8cOiH9Afts9wWpfmbNks9xyPTJCVi/5RbJkzSaNW3aUfTv/kfUrF9j+LsRIrQcflso173pdRkaEC7zIEMp49colyZrdT4qWKCcPNHlUMmV2nJfsRtR1m2C3Tu0l6vo1KVuxhtq3ZQMOKtevRShPtWOH9qg8VfkKBEmREmWlQtV6ktj9O1iGTSRAAiRAAiRAAiRAAiRAAqmUQLoQqPr06aMEiLx584q/v7/s379fvQ54CuFkPITTnT17VipWrKiuWkRBP8SabNmsv9zdj3cJDyoYQtjgjdStWzdDoPrhhx8EApzZNm7cqKoIucuSxfEvkubxurxjxw7lRYZ64cKFY3mZ6XHm6+nT98KazLmuzGNQLlasmDF3cglU8CrbvPlu+BXWgFeUqwZxcuDAgUqgQg4t5KOKz1vM1bnTw7i0LFQhKfofc2eq11goqIQ81PaZRL1SiDuT33/Nlq/qrkeingRefo89+6p06jbA5qXprZtlwZwvZMuGFfJEl34SFnpWViz60ehDIUeuPPL6qOk20aempR3izvTxQ+WvpXfzzpk7CwWVlIHvTLGJPffyzZ06eVi+nDBMrX3rVrR8Me5/gmeGhRw7qASq+T9MkTkzxwkEJ3v78auxMuS9r5TYZu67eP60fPC/biqRvG5H3q7CNrENHJ3Z3h2bZMKoVwX321uBwGLSd9hEKV2+mtHlyv6NwR5QyJa7iNrlZ7/ePZTCA7bMLSaCwD97wxJxF28hARIgARIgARIggfRDIF0IVEg6jhCu7777Tp1qByEKhlxLCDO7cOGC8pSCxxQSfgcEBMi1a9fU159//qlOs3P2RwICycGDByUiIkIJXAkRg5zNad+OvezatUs1awEG+4bYFhYWpk7P+/fffw1PKYg/x48fV+OdhdzZr6HrEKi0Idl4fOFyeH6z2BSXQGXOe2W+R6+XmCv2GxkZqW5Fbq3atWu7PA0S0COsD4IlcnydOHFCSpUq5fL9HHiXQFoUqnDSHEL7YM0e6Sze3gn/p/LA7n/lvTe6KAG0YOHi0vjhjpIlW3bZ+Ndi2bN9g/zyzUTJkdNfWj/R4y5I0/efv76bFw2eQ9XrNrV5XoXL9n/+UsnZPx8zWMbPWmH5uznu7V42T6fVqq3Rw09I6eBqcuFsiCz+daactolRowY/I5O/WxfL6ykm5rYStiBOlSxbWW7bwoYz2rwuYfN/nKrEqRJlKknV2raDFnLklsP7t8u6FfOVZxkEsZGT7wli8EQcNfhZCTl+UN1frmItyVeoiPLAQptuV52mb/t2/SNv9+9kOzjhtvKUqvXgQxKQv7Dy2tr+z2o5e+qYLQF9R5k0+y/Jk6+Q6U5R92AfjvZvGZjKK1ly3H2u75bc/Xc7lW+X2yMBEiABEiABEiABEiABtxBI+G9dbtmGeyeFpwy+YFq4QXn+/Pm4KC8piFOwrFmzSvXq1dUJeqhDxHJkOixs6NChStzAGJxWBwEJbS1btnR0m2pDjqsOHTpIly5d1JfTgf91bNmyxQhN1AIVQvcQooaT+GDwotKhfJs2bfrvzoR5FOEmiDTaChYsqItOrxCadG4sDPL2vucNYn+T2RPt/Pnz9t2JqutQRtwcHBwsOXLkSNA8hQoVMjzqXBGo6vdanqD509NgR0KVpz7/yaMHjK0HFilplFFA2N+Jo3e9MM0dvn45Veiebvv681FKnCpSvKyMmjzPFh7nq7padegqY4a9IJvXL5Vfv50sD7V7Tnx8MunbjGuxksEyfNz34mcTsWBfTRohi36ZIadOHJKDe7dKmeDqqh3CFcQpWK/BH0nT1k+qMr7VadBK3ny5rQoRXLpgtsMcWrdv35Lneg+Ttk/2NO5Doe1TvZRohRA7s/kHFFDiFYSlqOuRtjDB7Kr7n3VLDBHq0Sdfki6931LtONnwk5F9lDBnnkeXv53yvhKa8JzDx36nQhl135plc2XS6H4SffOGwGvr5TfG6i7j6mz/xgAWSIAESIAESIAESIAESIAEPIJAujvFb+vWrZYX88orr8jLL79saYuKuus5gUbkTXJk77//vuBeeN5oQ9JyJP9u1aqVSkz+119/6S7jGmPzVHjmmWdUYu53333X9ovZ3bAaY4CDgjPB6emnnzZGa7ENDebxCfWgMgtUBQoUMOZ3VggPD7d0QaRzZhDVtCVXKJ05/5QW7/Qarlzz589vDDM/u9HIQoIJaKHqybf+luz+xcUrY2zxJcGTpvANEIG0IReS2eBZ9P6Q52N9De3TXglBGHvm5BHj5L8nnu9viFPoy2Dz9Hv0PzEIoX/nTsX2mkFY2/DxPxjiFO5r/Xh3XJTp3FiorPpjjmpDTiuzOIXGUuWrqhxOKO/f9S8usax+8/axxCkMgmBlL06hvXq9Zrgou/Bfji5U1ts8q2DIU/VU98GqjG/wyHptxGcqr5bR+F/h+OG9AqEL9ogtAT3ybJmtQfMOElzlbtjuBpvnmSNztn9HY9lGAiRAAiRAAiRAAiRAAiSQegk4VxNS756TtLP169cb90OA+fjjj406CvCM0jmqUHcU8rVt2zZ5++230S21atVS3lkIE/zll1/k6NGjqh3rNGrUSJ0WCCGrWrVqgjYkXV+wYIFkzpxZJWlHWFp8pgUn+xC2+vXrGyfjIYQR3mHIT4VwPxjyVZUuXTq+6S39ZpHGLN5YBpkqWMNs8A5zZghV1ObK3HpsXFezB1VCEqTrOc37MD+77uc1fRLIZhNZtF2+dEGCbF5Q2oqXrigIe9MWEX5ZJfVGXYcFnjl1998BtMHb6eL5Mygadinsngfh2dPHpHAx69/ToOJlVEidcYOtEJA/UIXw4d+o8MsXja4zIXfXuiN35Pc504x2XdBjz52OLYRhzKOdXtJDHV7hvXTu9Am5FHZOIq9ekfNnThrj0KdNi1UVq9WLFUqIEMnipSupMEU9HtczIUeMap1GrYyyuVDTFvKHJPPXbQnmr14Js4h2GBff/s1zpdbytUsn1NYG9WiaWrfIfbmBQME8Wd0wK6ckARIgARIgARIgAc8lkO4EKrPHzauvvqqEIvPrg9CjRZYSJUpI7ty5zd0q5O+5556TaFvYyptvvimjR482+lGeMmWKjBw50jit7vfffxd82RtOjINo5YppgQoJ3c0hbAina9OmjcyaNUtNs2TJEnnxxRcNgSoh+Zj0Pk6evPfLZ86cOXWz0ysEKuxDh/lpdo5uMHubueKd5WgOcxuS2SP/l7bEeFCZeZqfXc/Ja8IJ+GT0kjYPFJTnWhaTJs+MTfgEqeAOnKCn7dTxQ1Kpen1dlSatOqkv3bB2+W8yYeQruqquZ01eUQt+/MLSZ1+B4OSKQeTxyZRZiWBmz8uz/wlPSG4+67N3nU5lvsc8yJlIjhMHv5r0tmxYvVCF8pnvcVTWXl2588TveanvN3MKsMsvpccg95W2y2EXYglUzvav7/GEa1T4abXNzi2KesJ2uUcSIAESIAESIAESIAEScAuBdCVQXbx40fCOQrhZr169YkE1C1iOPHLgKYXE4RBD3nnnHcv9CFvr27evdO/eXSZOnChjx45VSczNg7Jnzy4ID3S0tnmcLkOE0V5ZjgSYdu3aGQIVErpDsNL5nRyN1/M6uyJMUVtc+aT0GPxyCJFKr2kf8qfH4WruM3sumcckpKxPKsQ9ENMg4CXUzCGJ5md3Ns/aKffCm5yNSQvtsxYdlWnz73m3uPJMZmEqv7/rJ0e6MndKjylctJSxJHI8tezQ1ai7UjDnlHqu11Dx9cvl8LZMWbJKtdpNHPa52qjXKlepljRp2cnpbQj3c9Wio2+qBO/7d21Wt2DukmWrqKTuFy+ckSXzv4k1lT5Q4datm7H6nDWYk89H37ypkqTbj9XiN9qRZJ5GAiRAAiRAAiRAAiRAAiSQNgmkK4HKLD5VqVJFnYJn/1rNIWOOBJ7ChQsL8lghT5VZ3DDPg9Pq4F0FD601a9ao8fAeqlChguBkPMzhqmnvKYx3tJ8WLVoITg7EfpYvXy47d+40pk5o/inciJPt9u3bp+aAGOeKQWzSAtWxY8ec3mKeLzk8qMzvE8+qf0F2ugEHHeY94dlpCSeQloQp/fQ5cuURJDdHMnQkM9+xeY1UrtlAd8d7zV/onidMhWoPqGTj8d6UyAFYC55F2XxzSNM2TyVyFuttEKa0OPXSgPflobbPGgOQN8qRQIXk6RCvzv4XcmjcYBRie4rlt53ypw2nDZapUENXjSvaYRDiAvIWMtpZIAESIAESIAESIAESIAESSFsE4k+AlIae15x/ypl4YxY9HHlQAQc8i+AJFZ/5+flJ69at1al+Y8aMkeeffz5B4hTmN3sJQdyyN+yjSZO7HhiXL182vKkg1jh7Rvs5zPUiRe79wogT+lwxnJ6nbd26dbpouV6/fl2Qu0tb8+bNdTHR1/jERFcmNj+j+dlduTe9j4Ew1b5hoPzwbj0Z1LmceLrXlP377N5vpNE0dewbcuzQbqMeX6FgUHFjyOo/fzbK7igUCiqhpt1pE9EuXbyX2yopax09sMu4vXYD64mkEMMcWcH/9rFv52YlmJnHIP/VysV3k7mb2wsFlTSqKxb9aJR1ATmuVv3xk6qWKFtZJZjXfbySAAmQAAmQAAmQAAmQAAmkLQLpVqBy5I0UEREhu3ff/SU0a9asAi+rlDRHuWi0qFaxYkUpXvzeL73mfT388MNG9fvvv1flMmXKSK5cjsOKjMEOCmaRRntFORhmaUJOLm07duwwcnjpNlzhCXbTFsIDgydZYvJjqZv/+4awH7N4lxgxDlOZn9H87Oa1WLYSSOvClH5anGDX4KEOqorE4EN6PSo/fPmR7Nm+USIjwlXC7m2bVslKB8JK3vyFpW6j1urepQtmy6JfZsgd04mdSDYOQeaXbybq5RJ9bfNEDyXcICzvo2EvqITmejL8m3J4/w755vNRAs8nVy1PvoLG0OULv1eHR6Bhy4YVMvbt2KHR6ENuLlhMzG0ZO6KnXAq9K3AjRHLYKx0EJxbaG5LDV63dWDVjnV9nTzLyXSGn1ejXnzPue7L7QPvbWScBEiABEiABEiABEiABEkhDBNJNiB/yC/3zzz/Gq3MkaKBf5zupWbOm0xA+Y5JkKsC7CPmqxo8fL6tWrZJKle6eEIZ2LVC1bdvW6WoI87M3RwKc/RhHdXOY25kz1pPHHI1HGwQyhPnBGwn8kLQdJxeaTQtnaOvWrZu5S5UHDx5snKjYv39/xSLWIFPDli1bBIKitqpVXc+vo+/B1fyMFKjMZGKX02IoX+yntLb0fv1jyZO3oMz/YarcvhWtBCVnolJGW147X797Bws82/NN2b31b7kafsmWbHyEfPvF+1IgsJhKcn7uzAklWBUsXFwef66vddEE1oqWDJbWj3WXhT9PVycGvtL5QbVn3xy5VMgdThmE4fTBoiVdy9NWqfqDkjN3Xps4dEG+nz5Gfv9pugqhxYmAXl7eDndY0RbKWLlGA9nx7xrZt/MfeemJmuKd0Udxww1BxcrIyWMHYt2LHF046RCiHdb6YcbHKmcXTuzT1qJdF0uiet3OKwmQAAmQAAmQAAmQAAmQQNohkG48qBBedu3aNfXm/P39pXTpe6d06ddpDu9LrMCj50rI9Y033pC33npLJVTv2rWrOiEQ90PUwZ4RKvjyyy87nRLJwe3FFUcCnNMJTB3lypUzashp5ciryxjwXwG5uMxJ34cMGSKHD9/NG4Mhy5Ytky+++EKNLlq0qAp1/O9W42LOH3Xq1Cmj3Vnhp5/uhv3ofiRqT6gdP35ccGojDHm8sDdabALpxWMq9pPfzXv0zEv/k9GfzpNa9VtIvoJBsXKdwVuq1WPd5OPpf0p2k0CF3FDjZ62QB5u2tSX/zqqEqRNH9snZU8fUUqXKVZGnegy2LJvRJujAfHwyW9p1RfdDDDNb11dGyGsjPhPkdMLfJeSCgscUxCmcgofcVFVqNTRuwSER2nSSdV3H1S+nvwx5b4YUK1VBNUMsgjgFkemdCT9JQP5A1W5Oco6GIR/MlMYtO0qWrNlVP0S9AjYOPQd+IN36vqPaMtpOIzRbkRLlZNxXy6RGveaSKXMWJdxhvQy2AxgCi5SSfm9NlhdfG22+xcYn7v1bBrNCAiRAAiRAAiRAAiRAAiTgEQSsvyl4xJYTt0ntiYS7nYk35pxGzvJPJW71uO8aOXKkEnH27t0r8AwqUaKENG3aVObMuZuzBaf+BQbe/YXQ2Uzwopo+fbrRnViBDc8NwQt7gVAEJq6wGDFihBLYJk2aJJGRkcoLDM8AgQ15qSB0lSpVSiVyDwgIMPapC5kz3/uF3OwZpfv1NTo6Wr777jvBOmYbPny49OvXTwoVKmRujrP8yy+/GP1PPfWUIKyTdo9AevSYuvf01hJOwHt91JeqMep6pJw+ecTmSeQlSAyOhOrOLGfuAOk//FP15z/0/Cm5ZgsN9Pb2kbwFAh2eWDfgnSlyPTLClvDcz+GU037dYhOwb0p2W0J0e3ugyaOCL+wv9Nxp27gb6tS9PPli/52A19XXi/YqMUuLSfbz4Zk/mv6Hymt14VyIFAwspoQrjJvw9SqbZ9QtyZrd13IbxK4+Q8ZJ79c/kpNHD6h9ajELA2ct3KOSnVtuslXAccj7XylxCt5lELYCbMIfhD1H5sr+Hd3HNhIgARIgARIgARIgARIggdRLIN0IVK6cbmcWqBIr8CTmVefMmVPmz58vvXv3ltWrV0tISIh8/fXXKqE6wv6eeOKJeKdFmB0EKghZ8MZKSv4s7KNv37thRz///LNLAhW8NiZOnKhOAfzxxx9l+/btsnDhQrVv5M5q0KCBQGhzJiAhwbs2CGTODAIevMzsDUno8ex63/b9jupmgSouDzVH96b1thrl/KVl3YJpLvF5crw3CDolytwNw3V1Pvz9gKeV5I/7DngkITTPmcHDCF9xGfaH3E7xWdZsVnHJ2fjcefIJvsym9nBPUzZ3qTLCAB2FE2bL7lh40xPAawphkK6Yq/t3ZS6OIQESIAESIAESIAESIAESuP8E0o1ANW3aNMGXM0NI2oULF1Q3Qr0KFryXJNjZPcnZDu+ipUuXypUrV1SidnhRFShQwOUlcFog8lfBOwzhakkxnDb4v//9T3lCQaD68MMPXc7HhVxS+ILn1MGDByVv3rxORSnzHuGxpa1hw3uhSLpNX/F8roQd6vHOridOnBAd0lmrVi3BF+0egYol7uVSutfKEgmQQFomgPBsfAhgNvx/gtBvfOjxwAMPOD2sw3wPyyRAAiRAAiRAAiRAAiSQGALpRqCKD44WKzAuJb2n7PcFbyr8EpBQy5YtmzRq1CihtzkcnyNHDsHJfFOmTBEIOfDIgvdTQgz7cdWLC0LWypUr1fStWrWSdu3aJWSpBI+NsZ2mhkTtWujq06dPgufgDSRAAiSQ1ghMmDBBoqKiYj3WggULVBtyf73++uvq/wRzWHasG1xsWLRokWzevFn9n+vosA8Xp+EwEiABEiABEiABEiCBNEIg3SRJj+99mcP7XMm5FN98nt7/wQcfSOXKldVjwINq8eLFbnkkeIy1b9/edjR9jJQtW9aSR8stC9omRc6vFStWqOkfe+wx6dKli7uW4rwkQAIk4HEE2rRpo/4t/vTTT1XYdLNmzcTX11cd4DF69GjBqalhYWFJfq7JkycL8hfiwxAaCZAACZAACZAACZAACVCg+u/PgFmgup8eVKnljyQ8uf744w8pVqyY8jSCiLN///5k3x4SsR89elTlqEIie2c5qpJrYeT6evfdd9V0CBecPXt2rFPZkmstzkMCJEACnkgAAlSPHj3U6bHwqsJJrMjjqP9v3LdvnwoD98Rn455JgARIgARIgARIgARSLwGG+NneDfIlIak3DGEL1apVU+X0/g15uP7880+pX7++ys8Fj6pXX31Vhg0bJrlyOU/knBBuwcHBKhcUcpwkR8iIs7WPHDmiQlN0YvSSJUuqxPQ8uc8ZMbaTAAmQwD0C+LBizZo1UqNGDdmxY4fysHrhhRdi5e9DyB5Cto8fPy43b95UOasQvof7tG3cuFEgcukPPTB21qxZqhvh4R07dtRD1fXq1avqNFj8P40PNHDwRsWKFaVx48aSPXt2y1hWSIAESIAESIAESIAEPJcABSrbu/v333/llu3IdFj16tUlU6ZMqsxvImXKlFECEsI64G00duxY9YvEkiVLkk3IczVXVWLfB/aNX6Ru3LihvKV69eolCGFEri0aCZAACZCAawQyZswoo0aNkrZt26qw7J9++skQqCAy4cRZCFT2NnToUPXhBryxYG+//bby0NXj7E9nhbCFfFcwiGLPPPOMnDx5Ug83rjhcBP++wxuWRgIkQAIkQAIkQAIk4PkEKFDZ3iFyaeCTWBjyIdGsBOBtNGPGDBUaN27cOHUaYu7cua2DUnENv1ThU3aIVPhC+AqNBEiABEgg4QQeffRRCQgIkNDQUOUFpWfQCc/9/PykQ4cOysMpMjJSZs6cqbypJk6cKLi3efPmqh/jID6dPXtWAgMDjcNB8H+LFqfWrVsnTZo0kdu3b6t/wyGM4ZTdbdu2Ke/eQ4cOqf+7Dxw4IEFBQXorvJIACZAACZAACZAACXgoAQpUtheHU+PcfXKch/75sGy7cOHCAoEKJ/q5MxzPsmgyVDp16qQ+2YdQRSMBEiABEkgaAXxoAYFKh+hhNoSCwzO1Z8+elhDw3r17C8ZDrEJeQwhUL730kvpq3bq1OoCjdu3aMmfOnFibeuONN5Q4BUFs6dKllg8Xvv32W3XaLE4dRKJ1fIhCIwESIAESIAESIAES8GwCaeY39gZtunr2m+DuSYAESIAESMADCPj7+6tdXrx40dhtpUqVBF/2lj9/fqlZs6asXr1aeVLZ9zurI88VPKhg/fv3t4hTaEPY3/Tp02XVqlWC3IIUqECFRgIkQAIkQAIkQAKeTcDjT/Hr/gxD8jz7j6Dn7L5bZ/5Z85y3xZ2SAAm4i8ClS5fU1DhIw5GdPn1aNm3apA6iQJ4qHEQCg7eTq4awPW2PP/64LlquCPmDhYeHK48uSycrJEACJEACJEACJEACHkfA4z2oIBpQOPC4P3fcMAmQAAmQgIcSOHbsmNp56dKlLU8wb948FW4H76ek2uHDh40pnOWXypMnjzEGuawQCkgjARIgARIgARIgARLwXAIeL1B5LnrunARIgARIgAQ8iwDyTp07d05tGqe8aoM4hVP8kNAcIYDwboKwlCVLFvnqq68ECc0TYuacgTiBFQdd2Js+fRftSLpOIwESIAESIAESIAES8GwCFKg8+/1x9yRAAiRAAiSQYgSGDRsmd+7cUesh4bm28ePHK3GqXLlysmHDBsmZM6fuks2bN8cpUMXExBhjdaFEiRK6qJKx16tXz6jrgk7SjkM7nHlZ6bG8kgAJkAAJkAAJkAAJpH4CHp+DKvUj5g5JgARIgARIwLMJQEQaO3as/Pzzz+pBHnvsMXUin36qLVu2qGLjxo0t4hTuO3/+vB5muWbKlEnVDx48aGlHpWzZskabowToyGc1c+ZMNaZGjRri5cUfZwxgLJAACZAACZAACZCAhxLgT3Qe+uK4bRIgARIgARJwBwHkczp69KggUfkff/whEydOlAYNGsigQYPUcgi3g8eU2QoXLqyqv/32m1y4cEGVkbwcYX/6ND7zeJR1knV4Qq1cuVIgZm3fvl0NCw4OlpYtW6oyTut77733JCIiQtVPnDghrVq1MoSvd999V7XzGwmQAAmQAAmQAAmQgGcToEDl2e+PuycBEiABEiCBZCXw5ZdfCkLs4MUEIahfv36yfv16tcbDDz8s27ZtkyJFiljW7Ny5s6qfOXNGCU/FixdXScvnzp1rGWeuYC4Y8lY1bdpUfH19Bd5QqMPGjBkjuXPnVuWhQ4cqz6y8efNK0aJFZdWqVaq9d+/e0qxZM1XmNxIgARIgARIgARIgAc8mQIHKs98fd08CJEACJEACyUKgevXqlnkQNlesWDEVygchCGITPKpKlSplGYfKm2++KYMHDxbkg4LAhJP+UO7Tp498+umnaryPj4/lvnbt2knfvn0lQ4YMqv369esql5S3t7eqV6pUSXbu3CmPPPKIZM2aVXlYhYaGqnA+5Lr67rvv5LPPPrPMyQoJkAAJkAAJkAAJkIDnEmCSdM99d9w5CZAACZAACSQbAWeheK4sAFEJHk8It8OJfQjXq1ChgqAdSdWffvrpWCftQZiaMGGCCt/DPRCwzCcDYt3AwEBZsGCBmu/IkSMSHR2tPKiyZcvmyrY4hgRIgARIgARIgARIwIMIUKDyoJfFrZIACZAACZBAaiaQJUsWqVixomWLEKJ0qJ6l478KclpVqVLFUZfRBm8uR55bxgAWSIAESIAESIAESIAEPJ4AQ/w8/hXyAUiABEiABEiABEiABEiABEiABEiABEjAswlQoPLs98fdkwAJkAAJkAAJkAAJkAAJkAAJkAAJkIDHE6BA5fGvkA9AAiRAAiRAAiRAAiRAAiRAAiRAAiRAAp5NgAKVZ78/7p4ESIAESIAESIAESIAESIAESIAESIAEPJ4ABSqPf4V8ABIgARIgARIgARIgARIgARIgARIgARLwbAIUqDz7/XH3JEACJEACJEACJEACJEACJEACJEACJODxBChQefwr5AOQAAmQAAmQAAmQAAmQAAmQAAmQAAmQgGcToEDl2e+PuycBEiABEiABEiABEiABEiABEiABEiABjydAgcrjXyEfgARIgARIgARIgARIgARIgARIgARIgAQ8mwAFKs9+f9w9CZAACZAACXgUgTt37sj58+fl6tWrHrVvbpYESIAESIAESIAESMC9BChQuZcvZycBEiABEiCBVEPg8uXLgq/7aY899pjkz59fKlasmKhtQOA6duyYREVFJep+3kQCJEACJEACJEACJJA6CVCgSp3vhbsiARIgARIggWQlsGTJEsmTJ4/6WrVqVbLO7epkP/30k8ybN08yZMggn3/+uau3WcbVr19fihcvLi1btrS0s0ICJEACJEACJEACJODZBDJ69va5exIgARIgARIgAVcIHDp0SGJiYtTQkJAQV26Jd8yiRYtk8+bNUrduXWnRokWc469cuSL9+vVTY9566y1p3bq1w/HxzXnw4EF1X3I9g8NNsJEESIAESIAESIAESCDFCVCgSnHkXJAESIAESIAEUp5Au3bt5J9//pEcOXLEKya5urvJkyfL4sWLpUOHDvHOOWHCBDlz5ozyfBoxYoTTJeKbc+TIkbJmzRpp27at0znYQQIkQAIkQAIkQAIk4HkEKFB53jvjjkmABEiABEggwQQCAwPlq6++SvB9yXXDa6+9Jp07d5ZixYqJl1fiMwz07NlT8EUjARIgARIgARIgARJIWwQoUKWt98mnIQESIAESSKcE/vjjD+UhdfLkSQkICJCGDRsqryYtBuHUvNmzZ6v8T126dJFs2bIpUuHh4fLNN99I9uzZpWvXrhIWFibIV7Vu3TopUKCA1KlTR5o3b26hunHjRtm3b5/s379ftR8/flxmzZqlypi3Y8eOlvGbNm1Se9u9e7cUKlRIqlatKk2aNFFr6oGuzrl69WrZvn27VK9eXZCPyt5OnToly5cvl127dsmNGzekfPnygsTs+fLlsx8q2A9CFHfs2KG4VKhQQWrXri240kiABEiABEiABEiABFKWAAWqlOXN1UiABEiABEggWQlcv35d5XOyT3z+/vvvq5xPn3zyiVoPgtLLL7+sysgZBZEIBpHplVdeEW9vbyldurQSc86fP6/69Lfnn39evvzySzUGbW+//bZAENO2ZcsWJW7p+s2bN8XHx0cgimFNCGP2VrZsWUHS9EqVKqkuV+ccM2aMIE8VQhbtBapp06bJwIED1brm9QYMGCAzZ86UTp06qWYwQ9uUKVPMw1QZCdzh7fXee+9J5syZY/WzgQRIgARIgARIgARIwD0EKFC5hytnJQESIAESIIEUIYCcTBCnIAhBdIH3D0QnhPOdPn3a5T3cvn1bmjZtKhCXKlasqDyJIASdPXtWeUfh1LynnnpKzYecU35+fioXFPoRPvjAAw+ovty5c6u9oAJPqj///NPw2oIwduzYMUGeKezx4YcflsOHD0vWrFlVHitX5lSLOPi2cOFCeemll1RPwYIFVY6qjBkzKjHr6NGjSozSAhU8qrTAVq1aNWncuLFERESovZ44cULGjRunvK+wTxoJkAAJkAAJkAAJkEDKEKBAlTKcuQoJkAAJkAAJuIXAihUr1LzdunWTDz74wFhj2LBhcvnyZaPuSiE6OloJOTrHE8IFg4ODlXgDDyQtUEEIwhdO4kOSdITFzZkzx7IEwgQhTsGmT58u3bt3N/ohEEGsQtL0qVOnSv/+/dV88c1pTGBXuHXrlvKcQnPx4sVl7dq1KpQQdfTBmywqKgpVWbp0qSFOYb3PP//cyIkFXhDNEJL4xRdfKE+qkiVLqvv4jQRIgARIgARIgARIwL0EEp+l1L374uwkQAIkQAIkQAIuEIDnFGzr1q2C0DVtWbJkUTmkdN2VK8IBtTiF8UFBQdK+fXt1K/JMJcQgaMHg0WUWp9AGQQteS7D169era1K+7dmzx8iHNXjwYEOcwpzwonrrrbdk9OjRagkIUrBcuXLJ2LFjDXFKtyGEEAaxbv78+arMbyRAAiRAAiRAAiRAAu4nQA8q9zPmCiRAAiRAAiTgNgKtWrVSHkP//POPyiv1+uuvK08nJD1PqCGxur3h1D3YhQsX1NXVbwcPHlRD79y5o0Lm7O/T8yHEL6l26NAhYwqEIsZlBw4cUN3wlPL19Y01tFGjRpIzZ065cuWKCj+MNYANJEACJEACJEACJEACbiFAgcotWDkpCZAACZAACaQMAXgMwYPo22+/FYgvL7zwggwaNEjloxoyZIiRDyqxu0FeKFhMTEyCptDCE/aGxOXOLKHzOprHLFDhlMC47MiRI6q7cOHCTof5+/srgQr5tWgkQAIkQAIkQAIkQAIpQ4ACVcpw5iokQAIkQAIk4BYCCPHDKXm9e/eWiRMnyty5c1XuqeHDh8vu3bvlhx9+cMu68U2qT8DDSXtdu3Z1Ohzhfkk1nLynDWGOem3dZr7itEIYksE7M+Stgmlxztk4tpMACZAACZAACZAACSQfAQpUycfSo2bCD9+lSpWSyMhIddLTI4884lH752ZJgARIgASsBB588EHBF06he+aZZ1TY348//igfffSRyiVlHZ28NUdeUCVKlFAnACJcrkePHgle0NGcziZBYnRt8JCqXr26rsa6Yl87duwwclbZD7h69apK3o52/D9JIwESIAESIAESIAESSBkCHpskHbktPvzwQ+ncubNKwIofgPGDJHJxIOmpzm2RMhg9bxUcJ46Et6GhoWTlea+POyYBEiABpwSKFCmiRCk94PTp07qY7NdMmTKpOXW+KfMCZcuWVdVly5YZgo+531k5rjmd3WMWknRydvNYiE7I0QUrV66cui5fvlz9P6gqpm/Tpk1TJ/+hKTm8u0xTs0gCJEACJEACJEACJBAHAY8TqPCJ6vjx46VKlSqC3Brff/+9yr0RHh6ukpn+8ccfKvdG0aJFVZhDHM/OLhIgARIgARLwaAJI5A2vKYT44dQ5GJKS//bbb8ZzmcUbozGZCgULFlQz7d+/X1auXKnyVG3fvl219evXT52Qd+PGDenQoYPo3E/oxB43b94syJ8FbyazxTWneZy5XLVqValbt65qmjJliuCkPqwLb2GwwM8M/fv3V/2vvfaauuKDGiRU37lzp6oj5G/WrFkydOhQVW/QoIE89NBDqsxvJEACJEACJEACJEAC7ifgUQIVxCmcujNgwAB1lDaOiH7ppZfkiy++kD///FNmzJghL774ouTIkUP1//rrr+4nyBVIgARIgARI4D4RuHbtmqxfv16ee+45yZcvn9SqVUvgQfXBBx+oHcHLOE+ePG7bHf5PhkHsadq0qToVr0aNGqoOUahv376qf+PGjVKyZEkVaoj2gIAAtdePP/5Ydu3apcbob3HNqcc4uuKZkXsKQt3LL7+s9oKTDNu3by9Hjx4VfaohhCwdcrhv3z6pXLmy5M6dW/3sgFxZUVFRApFs0qRJjpZhGwmQAAmQAAmQAAmQgJsIeJRA9dlnnwlCBWD4IXzLli0ydepUJUq1aNFCunXrpsSqkydPyptvvhlnDgo38eS0JEACJEACJJBiBCCkDBs2THDq3OXLl5VXUkhIiGTLlk2dnIdwNW06dA51cxJxZ+3295nH6b527dopEUonKUeC8qCgINGJyOHxjCTtyPuEMdgbPKbCwsKUcAahCP9/my2+OfU+zM+A+xs1aiR///23wJsKa8F7Cl5RWbNmVT8fwMtM2/Tp0wVf2CsM7OBxhQ++OnXqJPACg5BGIwESIAESIAESIAESSDkCGWxu9ndSbrnEr3TmzBmVNwKhfPiBEuEE+KEzMbZp0yaViwKnG+E4avww26RJE+PTVT3nihUrlOs/fnguX7684D7krMBecA/yXelQBH2P+Xrx4kVZtWqVbN26VbBvfEpbr149lTPLPA5933zzjVofn96uXbtW5syZo0Il8IO6OcQAIREIo0D+KPzgjcSw2B8+sXZme/fulb/++kv9UoBPsJs3by7IDZIlSxZ1CzzPIO7ZG/hgPfwygR/2K1SooPJx4GpvZlY4uhvCId4R3hXCKuDVRkt/BIJbjbY8tH/RuyE4lkZTJez4BlNNZM/iu6E2lkZbpX6v5Zam/k/fzXVjaWSFBJJI4JPv91tmWDulmaXurOLqn3tX/7w7W8fcDg/jQ4cOqbyCefPmFYS5ayHHPA65mPDvua+vr7lZnLVjEH5MgIAD0cteFNKT4MANrI8TBcuUKSMZM8Y+gyUiIkIlcIcQBA8qLQ7pOeyvzuaE8IT/N5F7Ugth9vdiLfzfh/1iP/r/O/txqCNM8tSpU+r5ihUr5mgI20iABEiABEiABEiABFKAQOyfIFNg0cQsgbwQ+IEUNmrUqESJU/gBHG7/5k9R9V4g2Pz0009SqVIl3aSSrS9atEjeeust9cMrhByz4ZeAefPmyQMPPGBuVmXkwnr++efl/Pnzlj78YgDvrnfeecf4wRpCziuvvKLqCE3o1auXEqdw4549e5RABUHqiSeeUIKRZUJbBfkyXn31VZkwYYJ9l3zyyScqJxfCL7RhD8jf5czwCTjCKJHHw95wL/J3vPfee5ZfVJCYHqxwH7zczDlFGjduLA0bNrSfinUSIAESIIFkIuDl5aWEGIgxcZmfn5/DbmftGIx/9xECF5chfC4+jyOIYsHBwXFNY+lzNifEL3iMxWVYC57WrhiELnzRSIAESIAESIAESIAE7i8BjxGo4M0Dww/RyLWRGOvYsaPKVYUftrt06aISqh47dkwmT56svH2Q9+Lw4cOxxK+RI0eq5fDDcuvWrdWnrUuWLFGn3yE8ASIS5tSGEAOMw6fOpUuXVkIV9o2cWKtXr5bRo0cLxC0kkDUbRCQIaPgkvGbNmiqPhv4EHOIPvJkwD5LNVqxYUfDpMk4rgng1ceJEefTRR5V3lJ4TYRU6GSw+rUYyWDzvunXr5P3339fDYl0fe+wxgcAGq1atmkBgwqfRyPOF48vHjRunQiHAzd7QB8MvM/DqwqlJ+ESdRgIkQAIkQAIkQAIkQAIkQAIkQAIkQALOCHicQIVPh81iEB4MYXcIGTAbxuDTUwgzMAhKEFhgyDvRvXt3VcY3CDJImorQPYSm6ZN+jAG2Aj4ZhmeQng/iEkQhJFhF8ld9ehDuGTRokBKnICJBDNLhbfCSQrLWBQsWKA8keErZh0sgdOGjjz5Sc5jXr1+/vkp627NnT5UjQ/f17t1bJZ6FWAVRCeF72j788ENVROJciGbIAQJD2CHELIhO9rZ06VJDnEICepyEhE/mYQjxgIiHUEckpof4hZBBe8OJUQgpjCv80f4e1kkgKQQ27wmTmsFxe1QkZX7eSwIkQAIkQAIkQAIkQAIkQAIk4F4Cd5UH966RLLNDCIIVs8sPgUSrEGXatGlj+YIHkz4qGvfB0wiG/ElmcQpttWvXVl5CKOM0JHuD4AIRTItT6NcnE6EMDyZtBw4cMOYYPny4IU6hH0IPxCsYQv/grWVvOHFJjzH3IfTwjTfesIhT6M+fP7/ytkLZvA+EDW7btg3Naq9anEK9cePGsnDhQhRjGQQpGBLFImxPi1O6bcyYMaofoYjz589XZfM3CG5z586lOGWGwnKyE/DyuuexiMnXbr8gq7ectwnDyb4UJyQBEiABEiABEiABEiABEiABEkgBAh7jQYX8EMiNdO7cOQsWeCfBuwhHbWtDOCA8qjBe28GDB1URYXc6DE334XrhwgVVdSQaQdSyP6YbCWjhpYX59L2YQK+DMjyrcKKg2c6ePWtUsZZ9Po6BAwca/c4Kp0+fVichYS48p352HI2tzSxWQbyztzp16tg3qToENhg8peyT6KIdpyThXSCprCNWSNgOzzEaCbiTwMgXK8o7M3bLzegYY5mt+y/Jtahb0qJuQfG2E7CMQSyQAAmQAAmQAAmQAAmQAAmQAAmkSgIeI1DhFD0IMjiVx2xIlrpmzRpzkzqhTgstukOLKcgXFZcIhPxPrhjWxalAEMHM9+h1MAc8kOIy8316nNljSbfpKxKyjxgxwpKAXPfZX5ErShtOKnTVjhw5oobiJD5nhuS0EKjMYpseG9f+9RheSSCpBBpVyyfjXvWRIZ/vkIjrt4zp9h+/Ktdv3JZH6gdKJh+PcRA19s8CCZAACZAACZAACZAACZAACaRXAh4lUK1cuVIuXrwoW7ZskerVqyfonelcT/C26tq1q9N7Ee6XFNPrYA6Ewzk7aQjHdbdq1crlpSBO4RQ/JFLHnG3btlVHdEMk++qrr9Tx3ubJzHm6bt68ae6Ks6yP7I7rHuTJgsV16lOci7CTBJKBQNUyueXTQTVk4MRtEnrlXg66E2evyc/LT0r7xoGSLYvH/BOXDEQ4BQmQAAmQAAmQAAmQAAmQAAl4LgGP+e2tSZMm8tlnnynSyP+0du3aBFFHDiZ4/CA8DSfvucvMuZ6wZ5zGlxw2fvx4JU6VK1dONmzYYDkSG6f7HTp0yLJMYGCgUUfYob1HFEITHRn2v2PHDnWqoaP+q1evqmTy6ENuLhoJ3E8CJQN9ZcobNWXAhK1y4ty9MN/zl6Lkx6Un5LEmQZLTl6dI3s93xLVJgARIgARIgARIgARIgARIwBUCHhMDA+8hJPeG4WQ8JAxHom5XrWzZsmooTuLDaX3uMpwyqO3rr7/WxSRf4TUGAwOIbNoQJoiE6/Zm3sfPP/9s321JIG/uhAAGQ1J4cx4rPWbatGmiPaiS6m2m5+Q1/RAIO75B4vpKDIkC/lnk88E1Jbh4DsvtVyKibSLVcYFYRSMBEiABEiABEiABEiABEiABEkjdBDxGoALGTz/9VHx87npDIHyubt26MmfOHOXtg1xQyDv14YcfCpKI21u/fv3UiXRIKt6hQwfRuZYwDt5E8EIaPHiwS/md7Oc215E8/fHHH1dNU6dOlYkTJ1pyVF26dElmzJgho0aNMt8Wb1l7QP32229GUvbw8HAV9gfBzt7gCdWwYUPVPGXKFPn222+VsIT1X3jhBYeJ4jH4tddeU/cglLBly5ayc+dOVUfI36xZswxhq0GDBvLQQw+pPn4jgftNAF5SE/tXl3oV81i2ci3qtgr3Q9gfjQRIgARIgARIgARIgARIgARIIPUS8CiBCifeIbQPCdNh8Cp68sknBV4/yOkEL6khQ4ZIRESE6s+fP7+64luVKlUEoYEwnK5XsmRJlcMJ7QEBAVKrVi35+OOPZdeuXWpMUr5BJMOpfxB1IIzhNLzKlStL6dKl1VoIMUyod1Xnzp3VluD9VbBgQSlevLiaa+7cuU63Onz4cEFOLHhZPfvss0rcK1CggHz55ZeCxOm5c+eOdS9EPx0CuW/fPrVvjMNpicjdhZMCsf6kSZNi3csGErifBLJk9pb3e1eRVvUKWraBk/5+Wx0i+4+HW9pZIQESIAESIAESIAESIAESIAESSD0EPEqgAjaElW3dulWdZlenTh0l/phxZsqUSZo1a6a8lN577z1zlyCP0w8//CDwLkIS8ZCQEOUxFRYWpgQlCDMtWrQw7sFcMHPic6PTVtD9+qr7IH7htMCnnnpKCWfw7oInks4TBTFs5MiRergxDxqcrfXmm28qDy/0w7vp2LFjamyfPn2UZxnu1d5lKMPAYdGiRWIO90MS9EcffVR5jFWsWFGNs79v+vTpgq+goCDVf/nyZYHnWa5cuaRTp06yfft2Jfipzv++aQbO9m8eyzIJuItARu8M8maXYHnm4aKWJW7H3JE//j4jW/dfsrSzQgIkQAIkQAIkQAIkQAIkQAIkkDoIZLCFtznOlp069hfvLrB9iDXIwwRPH4hPGTPGn/sdXlYnTpxQwgs8qLQYY14QuZYQRoecT/p0O3M/hCct3JjbzWXsD+tA5IEQhBDA7Nmzm4eoMpKPQzSDt1VcBg8mCF3wiqpQoYLaF9bA/DhVz9mzh4aGKk7w5NJiEjy8IiMjHXpS6T1cuXJFTp06pYS2YsWK6eZY1/hYxbqBDSTgZgJzbCf5Tf75oMTY/RNXM9hf6lfJ6+bVOX1aIPDJ9/stj7F2SjNL3VkluNVoS5d/0bqWuq4gH5vZ9iweaq6yTAIkQAIkQAIkQAIkQALpioDHC1Tp6m3xYUmABBJEYNnmczJ65h6JvhVjuQ8J1ZvXLmDLS5fB0s4KCZgJUKAy02CZBEiABEiABEiABEiABNxLwONC/NyLg7OTAAmkJQLNa+aXj/pUkWxZvC2PtedouCxYcyqWcGUZxAoJkAAJkAAJkAAJkAAJkAAJkECKEaBAlWKouRAJkMD9IFCzvL9MHlBD/HPczSmn93D0dKT8suKkXL9xWzfxSgIkQAIkQAIkQAIkQAIkQAIkcJ8IUKC6T+C5LAmQQMoRKFPETz4fXFMC82a1LHr2YpTMWXZCwiOjLe2skAAJkAAJkAAJkAAJkAAJkAAJpCwBClQpy5urkQAJ3CcCEKcgUkGsMtul8Jvy49ITEnr5hrmZZRIgARIgARIgARIgARIgARIggRQkQIEqBWFzKRIggftLAGF+CPdD2J/ZIq/fkp+Wn5CQ89fMzSyTAAmQAAmQAAmQAAmQAAmQAAmkEAEKVCkEmsuQAAmkDgJImI7E6c1r5bds6MbNGJm3KkQOnbxqaWeFBEiABEiABEiABEiABEiABEjA/QQoULmfMVcgARJIZQR8MnrJ8O4VpFOzIMvObt2+IwvXnZEdBy9b2lkhARIgARIgARIgARIgARIgARJwLwEKVO7ly9lJgARSKQGvDBmkb8cy0rtDKbEVDbtz546s2HxO/t4ZarSxQAIkQAIkQAIkQAIkQAIkQAIk4F4CFKjcy5ezkwAJpHICzzxcVN7sEiwZvU0qlW3PG3ddlGWbzopNr6KRAAmQAAmQAAmQAAmQAAmQAAm4mQAFKjcD5vQkQAKpn0CregXl/d5VJEtmb8tmdx2+Ir+vPSUI/aORAAmQAAmQAAmQAAmQAAmQAAm4jwAFKgdsn3vuOfHx8ZEePXo46E27TQhtOnbsmERFRaXdh+STkYATAvUq5pGJ/atLTl8fy4jDIRHy68qTcuPmbUs7KyRAAiRAAiRAAiTw//buBLyq6u73+C9zyMiYSJhnwxCGxAFFQX0tKlakXq2t1drairZ1KlD72j62arWPbW3v9aLg8/T2drK3oKhVq2hAQCZRBAIyBoIkIWHIPA/nnNyzNj2Hs0MSSE6Gk+S7nkfPWmuvvYbPtrX8u/baCCCAAAIItJ8AAaomLIuKiuRwOFRYWNjE1Z5bNWvWLI0aNUo33HBDz10kK0OgBYGJo+K0bEmaBg+ItLXKO12tlWtyVFHlsNVTQAABBBBAAAEEEEAAAQQQaB8BAlTt49iuvbz33nt6+umn9eGHH7Zrv+frLDMz02qSm5t7vqZcR6DHCgxPjLKCVGOGxNjWWFhaqxXp2Soqq7PVU0AAAQQQQAABBBBAAAEEEPBfgACV/4bt3sPSpUv1i1/8QsuXL2/3vlvq8JlnntFdd92l5557rqVmXEOgxwsM7Buhlxalatr4vra1llfVa6U7SJVfUG2rp4AAAggggAACCCCAAAIIIOCfAAEq//x61N0LFy7U3//+d91xxx09al0sBoG2CMREher3D03X7OkJtttr3GdRrVqXq6zjFbZ6CggggAACCCCAAAIIIIAAAm0XCG37rd3rzpKSEq1evVq7du2yzpaKi4vTlClT9PWvf119+vRp1WLM2VTr16/Xzp07VVZWppSUFM2cOVOTJk1qsp8LHXvbtm06cOCADh48aPVz7Ngx/eUvf7HyUVFRuv322239l5eXa/PmzcrIyNDRo0et86MmT56sOXPmKDo62tbWzPNvf/ubVX/vvfdq06ZNWrlypVwul+bPn6/rr79eGzZssPqaMWOGzHlUjVNrxmt8L2UEuqNAeFiwnvn+ZP1hxSG9ueHsq68Oh0vvbszTdZcmatLo+O64NOaMAAIIIIAAAggggAACCASUQK8IUP32t7/VL3/5S1VVVZ2Db16le+edd6wg0zkXm6gwQa5vf/vbOnXqlO1qUFCQnnjiCT311FMKCTn7qfrWjG3maPr3pB07dsgEkzyprq7O+rqgKW/cuNF6HS8nJ8dz2fs7duxYayfUZZdd5q0zQa8f/ehH1tzq6+v1wAMPWMEp02Dfvn1WgOo3v/mNzPlXJmDVOEDV2vG8A5NBoJsLBAcHadE3Jqh/XLj+zztZ3tW43F+9TN92QpXVDl06aYC3ngwCCCCAAAIIIIAAAggggEDrBXpFgOp3v/udFZwyO4PMF+oGDBig7du365///Keys7P1gx/8wNpRdD6+rVu36qabblKD+w+m48aNswJVsbGxeuONN6zdR88++6wGDRqkRx55xNtVa8ZesGCBTH8mGHTixAkNGTJEV1xxhdVXv379vMEps2vqmmuukdPptHZE3XLLLRoxYoS1O+yDDz7Q4cOHrV1Uhw4d0rBhw7xzMRlzj1mv2TmVlpYmE6wKDw+3tWlc8Ge8xn1RRqC7Cnxn3igNiA/X7/5x0P2fnwbvMrbsLlBljVNzZiTIHacmIYAAAggggAACCCCAAAIItEGgVwSolixZYgVjzKtvvskEgEwAyQRgKioqFBNj/2qXb1uTX7x4sRWcMq/RmXvMa4ImmZ1Jt956q7UTyxwwbnYnRUREWNdaM/b9998v85cJgr3//vu69NJLrdfwrI58/vb4449bgaaBAwcqPT1d06ZN81599dVXdffdd6umpsY6aP1Pf/qT95on43A4ZHZ2mfVcSPJ3vAsZgzYIdAeBW2YNUb/YcP3yj1+ott7lnXLGoWJV1Th0w8zBCnHvuCIhgAACCCCAAAIIIIAAAgi0TqBXHJJuAjGNg1OGad68eV4tc95TS8nsRtqyZYvV5Mknn/QGp0xFcHCwN9hjXv07cuSIt6v2GNvbmTuze/duKzhm6h599FFbcMrUma/wzZ4922S1atUq67fx3775zW9659v4WuNye4zXuE/KCHRngaumDtIfHpmuWPch6r4pM7tcb63PVZ1P4Mr3OnkEEEAAAQQQQAABBBBAAIHmBex/wmq+XY+4YnYVZWVlKT8/X8XFxdbB4p6FmWstpczMTO9lc5h547OIAjgvAAAinUlEQVSfzCt5nmQCVBMnTvQUrV9/xvbtyATKPOm2227zZG2/5pU/c4i7ORi9oKBAZqeVb1q0aJFvscV8e4zX4gBcRKAbCqSM7auXFqdq8f/epVPFtd4V5Jys0mtrs3Xr7KGK7tOr/uvVa0AGAQQQQAABBBBAAAEEEGiLQK/4E1Rpaam12+j111+3XuVrC5TvrqgXXnihxS7M+U6e1B5je/oyv77zaHy+lKedOWPLk0zgrHGAyuz4utDUHuNd6Fi0Q6A7CYxOitGyJWn68Yu7dOxEpXfqp90Bq5VrsrVgzlD1db8OSEIAAQQQQAABBBBAAAEEEDi/QI8PUNXW1lpnOnlezzNfpzOHg5ugTW5urpYvX35+JXcLz5lSprH52l3//v2bvC8qKko33nijda29xvYdKDT07CMz/UdHR/tetvLmjClPMoeu+5M6ezx/5sq9CHS2QGL/SHeQKlWPv5yhPUdKvcOXVtRrRbp7J5U7SGXakBBAAAEEEEAAAQQQQAABBFoWOBvtaLldt71qAlOe4JQJRi1cuNC7FnO+0oUGqEaPHu29z3xBzwS5zpf8Hdt3J5ZnLN95HDx4UDNnzvRc8v6aepNMUK25XVbexufJdPZ455kOlxEIOIG46DDrTKpfuA9O3+z+op8nVdc69fraHN08K0kjBp8bSPa04xcBBBBAAAEEEEAAAQQQQMB9vndPR9ixY4d3iQsWLPDmTcb33CjbhSYK48eP99b+9a9/9eZbyrR17PDwM68F+Z575RlnwoQJnqya+kKfOevqz3/+s9UmNTXVOsDde0MbMp09XhumyC0IdLlAZHiInnsgRfOuSLLNpd7h0r8+Pq4DX5bZ6ikggAACCCCAAAIIIIAAAgjYBXp8gGro0KHeFf/xj39UQ0ODVX7vvfd0++23e6+dLzNixAh5DiV/5ZVX9OKLL8p3h5M5dN0EjH71q195u2rr2IMHD7b6MDuh1q1bZ42TkZFh1ZnD12+44QYrb9bz3HPPec/Vys7Otl4vNF8SNOnpp5+2fv35W2eP589cuReBrhQICQ7ST+9O1j03jrRNw+Vq0Oqt+fr8QJGtngICCCCAAAIIIIAAAggggMBZgR4foLruuuuUmJhorfhnP/uZEhISrL/mzZunysqzBxufJWk+9/zzz8scQF5XV6dHHnlEMTExSklJ0bhx46wzre677z757q5q69hz5861JuF0OnXttdda45jdUKZskjkDq1+/flberCk+Pl6DBg2SCaKZr/eZ9OCDD8qM3x6ps8drjznTBwJdIRAUJN0/f4weu3OCgk3BJ23ceVrmLxICCCCAAAIIIIAAAggggMC5Aj0+QGUOQ3/77bc1bdo0a/UFBQU6ffq0Jk2aZAVzhg8fbtWHhYV5dTyv2PkejG4ujhkzRvv27dOdd94pcxh6dXW19uzZo8OHD1v3XnLJJXrmmWe8/bRlbHPz/Pnz9fDDDyvoP3/ANeOYs6RCQkKsvqdMmWKNe/PNN6tPnz7WDiuzLvN1vosvvlj/+Mc/9PLLL3vnYTKeNZl843WZOpM8bRpfb8t4Z3rk7wj0ToHb3IejP/W9yQoLtf9XrNlFZXZTmV1VJAQQQAABBBBAAAEEEEAAgbMCQe5X3nrNn5Ty8/N17NgxjR071trxZBhM8Ke+vl5xcXFeFfMVvLKyMmtnkico5L34n4xhM6/UlZSUyAS3zO6lpr6o57nvQsf2tDe/ZoeXCX6Z/s0ZWL5f1PO0M68ZZmVlWWswczCBs+ZSeXm5FfQyO7+aShey7taM19QY1CHQmwR2HCzWfy/brcqas1/WNOs3h6abw9MbB7B6k013WOv//H9nPjjhmeum5Re2K3Xijc96brF++4+43Fb2FIqOfeLJWr/73v+ZrUwBAQQQQAABBBBAAIHeJNCrAlS96cGyVgQQCAyBzJxyLV66S4WldbYJJfaP1K3unVZ9Is7sjLRdpBAQAgSoAuIxMAkEEEAAAQQQQACBXiJgf/+klyyaZSKAAAKdJTBuWKyWL0nT0AT77saTRTVakZ6tsor6zpoK4yCAAAIIIIAAAggggAACAStAgCpgHw0TQwCBniIweGAfd5AqVckjz75KbNZWUl5nBalOF9f2lKWyDgQQQAABBBBAAAEEEECgTQIEqNrExk0IIIBA6wT6xobrxcdm6LKJA2w3mvOpXlubrZyTVbZ6CggggAACCCCAAAIIIIBAbxIgQNWbnjZrRQCBLhUw5009/8OpmnvZRbZ51NW79NaGXGVml9vqKSCAAAIIIIAAAggggAACvUWAAFVvedKsEwEEAkIgNCRIP793ku68frhtPk5ng97bkqeMQ8W2egoIIIAAAggggAACCCCAQG8QIEDVG54ya0QAgYASCAqSfnTbOOsvk/ekhgZp3eentGV3gaeKXwQQQAABBBBAAAEEEECgVwgQoOoVj5lFIoBAIAqYXVRmN5XZVeWbPt1bqPRtJ+QyESsSAggggAACCCCAAAIIINALBAhQ9YKHzBIRQCBwBcx5VOZcKnM+lW/am1WqdzbmyeF0+VaTRwABBBBAAAEEEEAAAQR6pAABqh75WFkUAgh0JwHzZT/zhb++sWG2aR89XqFVH+Wqps5pq6eAAAIIIIAAAggggAACCPQ0AQJUPe2Jsh4EEOiWAskj47R8SZoGD+xjm39+QbVWpmervKreVk8BAQQQQAABBBBAAAEEEOhJAgSoetLTZC0IINCtBYYmRLmDVKkaNyzWto6isjqtcAepCktrbfUUEEAAAQQQQAABBBBAAIGeIkCAqqc8SdaBAAI9QmBAfISWLpqhGRP62dZTUeXQa2tylHe62lZPAQEEEEAAAQQQQAABBBDoCQIEqHrCU2QNCCDQowSiI0P1wkPTdE1qgm1d5iyqN9bl6Ehuha2eAgIIIIAAAggggAACCCDQ3QUIUHX3J8j8EUCgRwqEhQbrqe9N1m1zhtrW53A26N1NefriSKmtngICCCCAAAIIIIAAAggg0J0FCFB156fH3BFAoEcLBAcF6bE7J+j++WPkznpTQ0OD1nx6Qtu+KPTWkUEAAQQQQAABBBBAAAEEurMAAaru/PSYOwII9AqBe24cqZ/enayQYJ8olXvlW/cU6KPtJ+WOV5EQQAABBBBAAAEEEEAAgW4tQICqWz8+Jo8AAr1FYN4VSXrugRRFhofYlrw7s0T/3pwnp/vVPxICCCCAAAIIIIAAAggg0F0FCFB11yfHvBFAoNcJXJkyUH94ZLriosNsaz+cU6431+eqts5lq6eAAAIIIIAAAggggAACCHQXAQJU3eVJMU8EEEDALTBlTLxeXpKqxP6RNo/cU1V6bW22KqsdtnoKCCCAAAIIIIAAAggggEB3ECBA1R2eEnNEAAEEfARGXhStZUvSNDopxqdWKiip1Yr0bBWX1dnqKSCAAAIIIIAAAggggAACgS5AgCrQnxDzQwABBJoQSOgXoZcWpyplbF/b1bLKeq1ck60ThTW2egoIIIAAAggggAACCCCAQCALEKAK5KfD3BBAAIEWBGKjQq0zqa6aOsjWqrrWqVUf5ejLvEpbPQUEEEAAAQQQQAABBBBAIFAFCFAF6pNhXggggMAFCESEBevZhVN0y6wkW+t6h0tvbzyufUfLbPUUEEAAAQQQQAABBBBAAIFAFCBAFYhPhTkhgAACrRAIDg7ST76VrO/MG2W7y+Vq0Ief5Gv7viJbPQUEEEAAAQQQQAABBBBAINAECFAF2hNhPggggEAbBe776mgt+sYEmYCVb9qUcVobdpxSQ4NvLXkEEEAAAQQQQAABBBBAIHAECFAFzrNgJggggIDfAgtmD9Uz35+scPerf75p58Fird6aJ6d7VxUJAQQQQAABBBBAAAEEEAg0AfufYAJtdswHAQQQQKDVArOnJ+j3D01TTJ9Q270Hj5XrXxtyVVfvstVTQAABBBBAAAEEEEAAAQS6WoAAVVc/AcZHAAEEOkBg2vh+emlxqgbGR9h6zz5RpdfX5qiqxmGrp4AAAggggAACCCCAAAIIdKUAAaqu1GdsBBBAoAMFxgyJ0fLH0zQ8Mco2yqniGq1Iz1ZpRb2tngICCCCAAAIIIIAAAggg0FUCBKi6Sp5xEUAAgU4QuKh/pJYtSdPEUXG20UxwakX6MZlgFQkBBBBAAAEEEEAAAQQQ6GoBAlRd/QQYHwEEEOhggfiYML346AzNnDzANlJVjdN63c+89kdCAAEEEEAAAQQQQAABBLpSgABVV+ozNgIIINBJApERIfr1g1N148zBthHNgenm4PSDx8ps9RQQQAABBBBAAAEEEEAAgc4UIEDVmdqMhQACCHShQGhIkJ64Z6LumjvCNgunq0Grt+Zr58FiWz0FBBBAAAEEEEAAAQQQQKCzBAhQdZY04yCAAAIBIBAUJD24YKwevn28gk3hP6mhQdqw45Q2ZZz2VPGLAAIIIIAAAggggAACCHSaAAGqTqNmIAQQQCBwBO64bpie/O4khYXa/zWwfV+RPvwkXy73rioSAggggAACCCCAAAIIINBZAvY/mXTWqIyDAAIIINDlAv91SaJ++8OpiooMsc1l39EyvbPxuOodLls9BQQQQAABBBBAAAEEEECgowQIUHWULP0igAAC3UAgLbm/lv44Vf3jwm2zPZpXqVUf5ai61mmrp4AAAggggAACCCCAAAIIdIQAAaqOUKVPBBBAoBsJjB8eq2VL0jRkUB/brE8U1mjlmmyVVdbb6ikggAACCCCAAAIIIIAAAu0tQICqvUXpDwEEEOiGAiY4ZYJUJljlm4rL6rQiPVsFJbW+1eQRQAABBBBAAAEEEEAAgXYVIEDVrpx0hgACCHRfAfOan3ndz7z255sqqx16bW22ck9V+VaTRwABBBBAAAEEEEAAAQTaTYAAVbtR0hECCCDQ/QXMgenm4HRzgLpvqq1z6a31uTqcU+5bTR4BBBBAAAEEEEAAAQQQaBcBAlTtwkgnCCCAQM8RCAsN1pPfnaQ7rhtmW5TD2aB/b87X7swSWz0FBBBAAAEEEEAAAQQQQMBfAQJU/gpyPwIIINADBYKDgvTw7eP14IKxcme9qaGhQR9tP6mtewq8dWQQQAABBBBAAAEEEEAAAX8FCFD5K8j9CCCAQA8WuGvuCD1xz0SFhvhEqdzr3fZFodZ8ekLueBUJAQQQQAABBBBAAAEEEPBbgACV34R0gAACCPRsgRtnDtavH5yqyIgQ20K/OFKqdzcdl3n1j4QAAggggAACCCCAAAII+CNAgMofPe5FAAEEeonAzMkD9OKjMxQfE2Zb8ZHcCr2xLke1dU5bPQUEEEAAAQQQQAABBBBAoDUCBKhao0VbBBBAoBcLTBwVp2VL0nRR/0ibQt7paq1ck6OKKoetngICCCCAAAIIIIAAAgggcKECBKguVIp2CCCAAAIanhil5T9J05ghMTaNwtJarUjPVlFZna2eAgIIIIAAAggggAACCCBwIQIEqC5EiTYIIIAAAl6BgX0j9NKiVE0b39dbZzLlVfVa6Q5S5RdU2+opIIAAAggggAACCCCAAALnEyBAdT4hriOAAAIInCMQExWq3z80XbOnJ9iu1bjPolq1LldZxyts9RQQQAABBBBAAAEEEEAAgZYECFC1pMM1BBBAAIFmBcLDgvXM9ydrweyhtjYOh0vvbszT3qxSWz0FBBBAAAEEEEAAAQQQQKA5AQJUzclQjwACCCBwXoHg4CAt+sYE3ffV0ba2roYGpW87oU/3FtrqKSCAAAIIIIAAAggggAACTQkQoGpKhToEEEAAgVYJfGfeKP3kWxfLBKx805bdBVr3+Sm541UkBBBAAAEEEEAAAQQQQKBZAQJUzdJwAQEEEECgNQK3zBqiZxdOUYT71T/flHGoWO9tyZPTRZTK14U8AggggAACCCCAAAIInBWw/ynibD05BBBAAAEEWi1w1dRB+sMj0xXrPkTdN2Vml+ut9bmqq3f5VpNHAAEEEEAAAQQQQAABBCwBAlT8g4AAAggg0K4CKWP76qXFqUroF2HrN+dklV5bm63KaoetngICCCCAAAIIIIAAAgggQICKfwYQQAABBNpdYHRSjJYtSdOIi6JtfZ8urtXKNdkqKa+z1VNAAAEEEEAAAQQQQACB3i1AgKp3P39WjwACCHSYQGL/SHeQKlVTxsTbxiitqNeK9GydLKqx1QdKoaTotLZtXK3i7E9V9OWWQJkW80AAAQQQQAABBBBAoEcLEKDq0Y+XxSGAAAJdKxAXHWadSXVlykDbRKprnXp9bY6O5Vfa6ru6cOzIPuV+eUiXXXWD+g2/VCHhUV09JcZHAAEEEEAAAQQQQKBXCBCg6hWPmUUigAACXScQGR6i5x5I0bwrkmyTqHe49K+Pj+vAl2Xe+tqaan15eK+yDu1RTU2Vt74zMkcz98rldGnyjCvPDtfAlwfPYpBDAAEEEEAAAQQQQKDjBOyfWeq4cegZAQQQQKAXC4QEB+mndydrQHy4/vr+l14Jl6tB//44yx2QqlTyyHhFRkVr5NhJ1nWzmyk8oo8GDx3lbd+ajAlyVVWWq7y0SDPn3NzirbnHMiV3MGrU+MnedlVFRxUZP9RbJoMAAggggAACCCCAAAIdJ8AOqo6zpWcEEEAAAR+BoCDp/vlj9NidExRsCv9JjtoyHSocqNzKRFswasSYiaooL/E0O+fX6ai3dls1uFy2a6a845OPNDAhSZOnX6HY+P62640LRadPqLK81BacqqmuVH1tuSJiBjVuThkBBBBAAAEEEEAAAQQ6QIAAVQeg0iUCCCCAQPMCt80Zqqe+N1lhoWf+FRQRk2g1/vxAkVZvzZfZVeVJwUFN/2sqc/9OZe7fpYTBw7Vv9zZPc9XV1mjntnWamnaV4voOkKO+Xk6Hw3u9cca0zz56QBMmp3kvmbp9GdsUPzjFW0cGAQQQQAABBBBAAAEEOlag6f/l37Fj0jsCCCCAQC8XuCY1QS88NE3RkfY3zc15VOZcKnM+lUkNjc6AMmdUmd1RQ0eM08VTLlFUdKzCwiOstuWlxTqw51PNmHmdQkLDrLpdn613nyl1hZVv6m87P12vaZfO8V4yrwRmfLZBMy6/1ltHBgEEEEAAAQQQQAABBDpegABVxxszAgIIIIBAEwIzJvTT0kUzrHOpfC+bL/uZL/yZL/35ptLiAuvwdBM86hMVY12qr6+zfs1resezDysl7WrvLabOBLBCQuxBME+Dvbu2aor7FUBPKi48qUN7P9cls+Z6qvhFAAEEEEAAAQQQQACBThIgQNVJ0AyDAAIIIHCuwLhhsVq+JE1DE6JsF08W1WhFerbKKmqt+rKSQp06kaPklEtt7TL37VTffoNkgktmR5Vvyti+QROnXu5b5c2fzM9WbFw/RcXEWXXmkPRT+Tm23VTexmQQQAABBBBAAAEEEECgwwUIUHU4MQMggAACCLQkMHhgH3eQKtX9Fb8zwSJP27zsTG09HKZjuaeVl5OlccnTPZes3wr3weZVFaXWV/rGXDzVdm375g/d51DNttV5Ci6XU/k5RzV89MVW1YEvPrN+fc+h8rTlFwEEEEAAAQQQQAABBDpHgABV5zgzCgIIIIBACwJ9Y8P14mMzdNnEAd5WjppS1TrD9Oqq9YpOmOSt92Q+2/SB9YW+xsEp82qfw30wev9BF3ma2n737nS/2pd6pXW+1fYt6booaYR1ppWtEQUEEEAAAQQQQAABBBDoVAECVJ3KzWAIIIAAAs0J9IkI0fM/nKq5l10kZ321QsKiVJb/haISUvTWhlxlZpd7bzVf6ovsE2X7+p7n4pb17+ry2Td5irZf8xrfgIQk68t+n2x4z73L6mr17Z9ga0MBAQQQQAABBBBAAAEEOl+AAFXnmzMiAggggEAzAqEhQfr5vZN02cgyOeqrFJ80xWrpdDbovS15yjhUrJN5x3T4wC7NnHPzOb1sTH/TXT/vnHpPhbk32n3u1J4dm6wglucLgJ7r/CKAAAIIIIAAAggggEDXCDT9aaOumQujIoAAAggg4H49r17JCeVKvutq/X1diftVvDMo5nftp8fVcPwt/XjJE+dI5eceVXBIiAYMGnzONVNxeP8uRcfGW1/7S535X022oRIBBBBAAAEEEEAAAQS6RoAdVF3jzqgIIIAAAs0IvPHGG7r88su18I5UazeV2VXlSacOfiAN/qrWfHZKLk/kyn2xweXSlnXv6Mprb/E0tf063UGv/Xs+tc6dau7LfrYb2lgwZ185astVU35CVUVfqqr4mMpP7VdFQaYqC4/I5TjzVcI2ds9tCCCAAAIIIIAAAgj0WAF2UPXYR8vCEEAAge4nUFVVpaKiIk2deuarfOY8qr6xYfr5K3tUUpivfsMucZ9NFam9WaWqrnXqpisHKzQkWB+9v0Kzv3Jbswv+2P3q35QZV2rk2HMPW/fcVFJ0yv36YLZc7mBXcsqlnmrr1wSW3nzzTcXFxWnOnDnKycnR4cOHVVNTY7U3jYKCghQWFmYFqEIj4hQak2gqFZuQbOuLAgIIIIAAAggggAACCJwrQIDqXBNqEEAAAQS6SCDE/YrewoULbaObL/uZL/wteWmXSsrrvdeyjldo1Ue5utId/3E5nRqYOMR7zTeTl5OlRPeX+loKTpn2Gz5YJXMmVW1NtTvo5FRpXo7ZmmXtugoKCVW/ftOtrwOuX79eSUlJuvrqqxUeHu47lJWP/F+fn1NHBQIIIIAAAggggAACCLQsQICqZR+uIoAAAgh0okBERESToyWPjNPyJWl67MVdyi+o9rYx+ZeX/kWP//xJb51vxgSaNq19S3fc+2Pfam++4ORxHdq3w9r9NO3S2RoxZqL3Wvr+/t68yZidUyQEEEAAAQQQQAABBBDoGAECVB3jSq8IIIAAAu0sMDQhyh2kStXipRnKzCm3ei84vE5RQ67Qa2uPa8GcoRoQbw9wbft4tWZdd6t3JiZgdTRzr/Jzs6zD1xMuGmp9DdC8nkdCAAEEEEAAAQQQQACBrhMgQNV19oyMAAIIINBKAROAWrpohv572W59ujtbNRUnNXDsNaqocui1NTm65eohShrUx9ury+mwglEmIOWoP/N64IgxybaglbcxGQQQQAABBBBAAAEEEOgyAQJUXUbPwAgggAACbRGIjgzVCw9N01f+x//V4EnzvV3U1Dn1xroc3XhFksYMjVHOl4c0JXWW4voO8LYhgwACCCCAAAIIIIAAAoEpEByY02JWCCCAAAIINC9wNOuw7rk1VXdcP87WyOFs0Lub8vTFkVKVlRQSnLLpUEAAAQQQQAABBBBAIHAFCFAF7rNhZggggAACzQi8+uqruufuu/XYnRN0//wx7kPOzzZsaGjQ26u36nRtwtlKcggggAACCCCAAAIIIBDQAgSoAvrxMDkEEEAAgcYCGRkZmjp1qkJCQqxL99w4Uj+9O1khwWejVPU1ZdqV5dBH209ah6E37oMyAggggAACCCCAAAIIBJYAAarAeh7MBgEEEEDgPALvvvuuvva1r9lazXOfO/XcAymKDA9RZWGWogeMtq7vzizRvzfnyel+9Y+EAAIIIIAAAggggAACgStAgCpwnw0zQwABBBBoJLB7926lpaU1qj1TvDJloP7wyHRFBFcrNCLW2+ZwTrneXJ+r2jqXt44MAggggAACCCCAAAIIBJYAAarAeh7MBgEEEECgBYHNmzdr7ty5zbZwlh/VH391uxL7R9ra5J6q0mtrs1VZ7bDVU0AAAQQQQAABBBBAAIHAECBAFRjPgVkggAACCJxHoKKiQiNGjGixVVVVlSaPG6xlS9I0OinG1ragpFYr0rNVXFZnq6eAAAIIIIAAAggggAACXS9AgKrrnwEzQAABBBC4AAGze+orX/lKsy137Nih1NRU63pCvwi9tDhVKWP72tqXVdZr5ZpsnSissdVTQAABBBBAAAEEEEAAga4VIEDVtf6MjgACCCBwgQLR0dEKDQ1tsrXT6ZTL5VJERIT3emxUqHUm1VVTB3nrTKa61qlVH+Xoy7xKWz0FBBBAAAEEEEAAAQQQ6DoBAlRdZ8/ICCCAAAIXKFBaWqrExMRmW2/fvt27e8q3UURYsJ5dOEW3zEryrVa9w6W3Nx7XvqNltnoKCCCAAAIIIIAAAggg0DUCBKi6xp1REUAAAQRaIWB2Ro0bN67JO8rKyhQXF6egoKAmrwcHB+kn30rWd+aNsl13uRr04Sf52r6vyFZPAQEEEEAAAQQQQAABBDpfgABV55szIgIIIIBAKwUiI+1f5fO9ff/+/UpOTvatajJ/31dHa9E3JsgErHzTpozT2rDjlBoafGvJI4AAAggggAACCCCAQGcKEKDqTG3GQgABBBBoV4H8/HwlJdlf32tpgAWzh+qZ709WuPvVP9+082CxVm/Nl9O9q4qEAAIIIIAAAggggAACnS9g/1/onT8+IyKAAAIIINBmgaKiIg0bNqxV98+enqDfPzRNMX3sB64fPFamf23IVV29q1X90RgBBBBAAAEEEEAAAQT8FyBA5b8hPSCAAAIIdJHApEmT2jTytPH99NLiVA2MP/vVP9NR9okqvb42R1U1jjb1y00IIIAAAggggAACCCDQNgECVG1z4y4EEEAAgW4uMGZIjJb9JE3DE6NsKzlVXKMV6dm2OgoIIIAAAggggAACCCDQsQIEqDrWl94RQAABBAJYYPCASC1bkqaJo+JssyytqLeVKSCAAAIIIIAAAggggEDHChCg6lhfekcAAQQQCHCB+JgwvfjoDM2cPCDAZ8r0EEAAAQQQQAABBBDouQL2E2J77jpZGQIIIIAAAs0KREaE6NcPTtXzf9+v991f82vPVHTsk/bsjr4QQAABBBBAAAEEEOiRAuyg6pGPlUUhgAACCLRWIDQkSE/cM1F3zR3R2ltpjwACCCCAAAIIIIAAAn4KEKDyE5DbEUAAAQR6jkBQkPTggrF6+PbxCjYFEgIIIIAAAggggAACCHSKAAGqTmFmEAQQQACB7iRwx3XD9OR3JykslH9NdqfnxlwRQAABBBBAAAEEuq9AUIM7dd/pM3MEEEAAAQQ6TmD7/iI98cpuVdU4tWn5dR03ED0jgAACCCCAAAIIINDLBfi/hnv5PwAsHwEEEECgeYG05P5a+uNU9Y8Lb74RVxBAAAEEEEAAAQQQQMBvAXZQ+U1IBwgggAACCCCAAAIIIIAAAggggAAC/giwg8ofPe5FAAEEEEAAAQQQQAABBBBAAAEEEPBbgACV34R0gAACCCCAAAIIIIAAAggggAACCCDgjwABKn/0uBcBBBBAAAEEEEAAAQQQQAABBBBAwG8BAlR+E9IBAggggAACCCCAAAIIIIAAAggggIA/AgSo/NHjXgQQQAABBBBAAAEEEEAAAQQQQAABvwUIUPlNSAcIIIAAAggggAACCCCAAAIIIIAAAv4IEKDyR497EUAAAQQQQAABBBBAAAEEEEAAAQT8FiBA5TchHSCAAAIIIIAAAggggAACCCCAAAII+CNAgMofPe5FAAEEEEAAAQQQQAABBBBAAAEEEPBbgACV34R0gAACCCCAAAIIIIAAAggggAACCCDgjwABKn/0uBcBBBBAAAEEEEAAAQQQQAABBBBAwG8BAlR+E9IBAggggAACCCCAAAIIIIAAAggggIA/AgSo/NHjXgQQQAABBBBAAAEEEEAAAQQQQAABvwUIUPlNSAcIIIAAAggggAACCCCAAAIIIIAAAv4IEKDyR497EUAAAQQQQAABBBBAAAEEEEAAAQT8FiBA5TchHSCAAAIIIIAAAggggAACCCCAAAII+CNAgMofPe5FAAEEEEAAAQQQQAABBBBAAAEEEPBb4P8DNVGoKLOnJ5cAAAAASUVORK5CYII=)", "_____no_output_____" ], [ "Función Objetivo de la WGAN", "_____no_output_____" ], [ "Sean $ y \\in \\{0,1\\} $ la etiqueta del dato $ x $ correspondientes a “falso” o “real” y $ \\hat{y} \\in [0,1] $ su estimación obtenida del discriminador para el dato en cuestión. Luego, la función a optimizar mediante el entrenamiento de la WGAN esta dada por\n\n$$ L = \\dfrac{1}{n} || W (v, \\hat{v}) ||^2 $$\n\ncon\n\n$$ W(v, \\hat{v} ) = \\dfrac{1}{\\# B} \\sum_{i \\in B} v_{ij} - \\dfrac{1}{\\# \\hat{B}} \\sum_{i \\in \\hat{B}} \\hat{v}_{ij}$$\n\ndonde $B$ define al conjunto de imágenes reales y $\\hat{B} $ al conjunto de imágenes generadas.", "_____no_output_____" ] ], [ [ "import os\nimport tensorflow as tf\nimport sys\nimport numpy as np\nimport matplotlib.pyplot as plt\n\nfrom tensorflow.keras.datasets import mnist\nfrom tensorflow.keras.layers import Input, Dense, Reshape, Flatten, Dropout\nfrom tensorflow.keras.layers import BatchNormalization, ZeroPadding2D\nfrom tensorflow.keras.layers import Activation, LeakyReLU\nfrom tensorflow.keras.layers import Conv2D, Conv2DTranspose\n\nfrom tensorflow.keras.models import Sequential, Model\nfrom tensorflow.keras.optimizers import Adam", "_____no_output_____" ], [ "from tensorflow.keras.datasets import cifar10", "_____no_output_____" ], [ "def reset_weights(model):\n '''\n Reinicializa los pesos del modelo \n '''\n for ix, layer in enumerate(model.layers):\n if hasattr(model.layers[ix], 'kernel_initializer') and \\\n hasattr(model.layers[ix], 'bias_initializer'):\n weight_initializer = model.layers[ix].kernel_initializer\n bias_initializer = model.layers[ix].bias_initializer\n\n old_weights, old_biases = model.layers[ix].get_weights()\n\n model.layers[ix].set_weights([\n weight_initializer(shape=old_weights.shape),\n bias_initializer (shape =len(old_biases))])", "_____no_output_____" ] ], [ [ "# Lectura de Datos", "_____no_output_____" ] ], [ [ "'''\n# normalización\nx_train = x_train.astype('float32') / 255.\nx_test = x_test.astype('float32') / 255.\n'''", "_____no_output_____" ] ], [ [ "Lee los datos reales y pone variables globales", "_____no_output_____" ] ], [ [ "# carga datos reales \n(X_train, y_train), (X_test, y_test) = cifar10.load_data()", "_____no_output_____" ], [ "# Normalización\nX_train = X_train.astype(np.float32) / 255.\nX_test = X_test.astype(np.float32) / 255.\n\ny_train = y_train.astype(np.float32)\ny_test = y_test.astype(np.float32)", "_____no_output_____" ], [ "'''# carga datos reales \n(X_train, y_train), (X_test, y_test) = mnist.load_data()\n\n# Rescale -1 to 1\nX_train = X_train.astype(np.float32) / 127.5 - 1.\nX_train = np.expand_dims(X_train, axis=3)\n\nX_test = X_test.astype(np.float32) / 127.5 - 1.\nX_test = np.expand_dims(X_test, axis = 3)\n\ny_train = y_train.astype(np.float32)\ny_test = y_test.astype(np.float32)'''", "_____no_output_____" ], [ "# Dimensión de imágenes\nX_train.shape", "_____no_output_____" ] ], [ [ "# Variables globales", "_____no_output_____" ] ], [ [ "(buffer_size,img_rows, img_cols, channels) = X_train.shape\nimg_shape = (img_rows, img_cols, channels)\nz_dim = 100 # dimensión del espacio latente\nnum_images = buffer_size\nnum_classes = 10\nbatch_size = 64\nnum_epochs = 100\neach_save = 5\nsave_epochs = 5 * each_save\nlearning_rate = 1e-3\n\n\npath_results = 'dcgan_results/'\npath_checkpoints = 'dcgan_results/checkpoints/'\n\nbatches_per_epoch = num_images // batch_size\n\nnum_images, batches_per_epoch", "_____no_output_____" ] ], [ [ "# Generador de datos", "_____no_output_____" ] ], [ [ "train_dataset = tf.data.Dataset.from_tensor_slices((X_train, y_train)).shuffle(buffer_size).batch(batch_size)\ntrain_dataset", "_____no_output_____" ] ], [ [ "# Generador", "_____no_output_____" ], [ "En generador es una red convolucional que\n\n1. Toma como entrada un vector aleatorio de 20 entradas.\n\n2. El cuál es pasado por una capa densa de dimensiones $(7)(7)(256)$ y luego reformateado (reshape) a un tensor de forma $ 7 \\times 7 \\times 256 $.\n\n3. Luego es pasado a una convolución transpuesta que convierte el tensor $ 7 \\times 7 \\times 256 $ en uno de $ 14 \\times 14 \\times 128 $. Con posterior normalización or lotes y actrivación ReLU.\n\n3. Luego es pasado a una convolución transpuesta que convierte el tensor $ 14 \\times 14 \\times 128 $ en uno de $ 14 \\times 14 \\times 64$. Con posterior normalización or lotes y actrivación ReLU.\n\n4. Y de nuevo usando una convolución transpuesta se transforma el tensor $ 14 \\times 14 \\times 64 $ en uno del tamaño de la imagen de salida $ 28 \\times 28 \\times 1 $ a la que se la aplica una activación tanh.\n\nEsto se ilustra en la siguiente figura", "_____no_output_____" ], [ "![dcgan_generador.png](data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAABGYAAAK2CAYAAAAMp5rmAAAMR2lDQ1BJQ0MgUHJvZmlsZQAASImVVwdYU8kWnltSSWiBCEgJvYlSpEsJoUUQkCrYCEkgocSYEETsyrIKrl1EwIauiii6FkDs2Mui2PtDEZWVdbFgQ+VNCqyr33vve+f75t4/Z875T8nce2cA0KnmSaW5qC4AeZJ8WXxECGtcahqL1AEQQAR04Az0eXy5lB0XFw2gDNz/Ke9uQmso11yUXD/O/1fREwjlfACQOIgzBHJ+HsT7AcCL+VJZPgBEH6i3npYvVeIJEBvIYIIQS5U4S42LlThDjStUNonxHIh3AkCm8XiyLAC0m6CeVcDPgjzatyF2lQjEEgB0yBAH8kU8AcSREA/Ly5uixNAOOGR8w5P1D86MQU4eL2sQq2tRCTlULJfm8qb/n+3435KXqxiIYQcHTSSLjFfWDPt2O2dKlBLTIO6WZMTEQqwP8QexQGUPMUoVKSKT1PaoKV/OgT0DTIhdBbzQKIhNIQ6X5MZEa/QZmeJwLsRwhaCF4nxuosZ3oVAelqDhrJZNiY8dwJkyDlvjW8+TqeIq7U8pcpLYGv7bIiF3gP9tkSgxRZ0zRi0QJ8dArA0xU56TEKW2wWyKRJyYARuZIl6Zvw3EfkJJRIiaH5uUKQuP19jL8uQD9WILRWJujAZX5osSIzU8O/k8Vf5GEDcJJeykAR6hfFz0QC0CYWiYunbsilCSpKkXa5fmh8RrfF9Lc+M09jhVmBuh1FtBbCovSND44oH5cEGq+fEYaX5cojpPPCObNzpOnQ9eCKIBB4QCFlDAkQGmgGwgbu1u7Ia/1DPhgAdkIAsIgYtGM+CRopqRwGsCKAJ/QiQE8kG/ENWsEBRA/ZdBrfrqAjJVswUqjxzwFOI8EAVy4W+FyksyGC0ZPIEa8Q/R+TDXXDiUcz/q2FATrdEoBnhZOgOWxDBiKDGSGE50xE3wQNwfj4bXYDjccR/cdyDbv+0JTwlthMeEG4R2wp3J4vmy7+phgTGgHUYI19Sc8W3NuB1k9cRD8ADID7lxJm4CXPCRMBIbD4KxPaGWo8lcWf333P+o4Zuua+worhSUMoQSTHH43lPbSdtzkEXZ0287pM41Y7CvnMGZ7+Nzvum0AN6jvrfEFmL7sLPYCew8dhhrBCzsGNaEXcKOKPHgKnqiWkUD0eJV+eRAHvEP8XiamMpOyl3rXLtcP6vn8oWFyvcj4EyRTpeJs0T5LDZ88wtZXAl/+DCWu6ubLwDK74j6NfWGqfo+IMwLf+sWrAcgYH9/f/+hv3VRzQDsKwOAeutvnf0s+Do4AcC5Kr5CVqDW4coLAVCBDnyijIE5sAYOsB534AX8QTAIA6NBLEgEqWAS7LIIrmcZmAZmgnmgBJSBZWA1qAQbwGawHewCe0EjOAxOgDPgIrgCboB7cPV0ghegB7wDfQiCkBA6wkCMEQvEFnFG3BEfJBAJQ6KReCQVSUeyEAmiQGYiC5AyZAVSiWxCapHfkIPICeQ80obcQR4hXchr5BOKoTTUADVD7dARqA/KRqPQRHQimoVORYvQYnQJWoHWoDvRBvQEehG9gbajL9BeDGBaGBOzxFwwH4yDxWJpWCYmw2ZjpVg5VoPVY83wf76GtWPd2EeciDNwFu4CV3AknoTz8an4bHwxXolvxxvwU/g1/BHeg38l0AmmBGeCH4FLGEfIIkwjlBDKCVsJBwin4dPUSXhHJBKZRHuiN3waU4nZxBnExcR1xN3E48Q2Ygexl0QiGZOcSQGkWBKPlE8qIa0l7SQdI10ldZI+kLXIFmR3cjg5jSwhzyeXk3eQj5Kvkp+R+yi6FFuKHyWWIqBMpyylbKE0Uy5TOil9VD2qPTWAmkjNps6jVlDrqaep96lvtLS0rLR8tcZqibXmalVo7dE6p/VI6yNNn+ZE49Am0BS0JbRttOO0O7Q3dDrdjh5MT6Pn05fQa+kn6Q/pH7QZ2sO1udoC7TnaVdoN2le1X+pQdGx12DqTdIp0ynX26VzW6dal6NrpcnR5urN1q3QP6t7S7dVj6Lnpxerl6S3W26F3Xu+5PknfTj9MX6BfrL9Z/6R+BwNjWDM4DD5jAWML4zSj04BoYG/ANcg2KDPYZdBq0GOobzjSMNmw0LDK8IhhOxNj2jG5zFzmUuZe5k3mpyFmQ9hDhEMWDakfcnXIe6OhRsFGQqNSo91GN4w+GbOMw4xzjJcbNxo/MMFNnEzGmkwzWW9y2qR7qMFQ/6H8oaVD9w69a4qaOpnGm84w3Wx6ybTXzNwswkxqttbspFm3OdM82DzbfJX5UfMuC4ZFoIXYYpXFMYs/WIYsNiuXVcE6xeqxNLWMtFRYbrJsteyzsrdKsppvtdvqgTXV2sc603qVdYt1j42FzRibmTZ1NndtKbY+tiLbNbZnbd/b2dul2P1s12j33N7InmtfZF9nf9+B7hDkMNWhxuG6I9HRxzHHcZ3jFSfUydNJ5FTldNkZdfZyFjuvc24bRhjmO0wyrGbYLReaC9ulwKXO5dFw5vDo4fOHNw5/OcJmRNqI5SPOjvjq6uma67rF9Z6bvttot/luzW6v3Z3c+e5V7tc96B7hHnM8mjxejXQeKRy5fuRtT4bnGM+fPVs8v3h5e8m86r26vG28072rvW/5GPjE+Sz2OedL8A3xneN72Pejn5dfvt9ev7/8Xfxz/Hf4Px9lP0o4asuojgCrAF7ApoD2QFZgeuDGwPYgyyBeUE3Q42DrYEHw1uBnbEd2Nnsn+2WIa4gs5EDIe44fZxbneCgWGhFaGtoaph+WFFYZ9jDcKjwrvC68J8IzYkbE8UhCZFTk8shbXDMun1vL7RntPXrW6FNRtKiEqMqox9FO0bLo5jHomNFjVo65H2MbI4lpjAWx3NiVsQ/i7OOmxh0aSxwbN7Zq7NN4t/iZ8WcTGAmTE3YkvEsMSVyaeC/JIUmR1JKskzwhuTb5fUpoyoqU9nEjxs0adzHVJFWc2pRGSktO25rWOz5s/OrxnRM8J5RMuDnRfmLhxPOTTCblTjoyWWcyb/K+dEJ6SvqO9M+8WF4NrzeDm1Gd0cPn8NfwXwiCBasEXcIA4Qrhs8yAzBWZz7MCslZmdYmCROWibjFHXCl+lR2ZvSH7fU5szrac/tyU3N155Lz0vIMSfUmO5NQU8ymFU9qkztISaftUv6mrp/bIomRb5Yh8orwp3wBu2C8pHBQ/KR4VBBZUFXyYljxtX6FeoaTw0nSn6YumPysKL/p1Bj6DP6NlpuXMeTMfzWLP2jQbmZ0xu2WO9ZziOZ1zI+Zun0edlzPv9/mu81fMf7sgZUFzsVnx3OKOnyJ+qivRLpGV3PrZ/+cNC/GF4oWtizwWrV30tVRQeqHMtay87PNi/uILv7j9UvFL/5LMJa1LvZauX0ZcJll2c3nQ8u0r9FYUrehYOWZlwyrWqtJVb1dPXn2+fGT5hjXUNYo17RXRFU1rbdYuW/u5UlR5oyqkane1afWi6vfrBOuurg9eX7/BbEPZhk8bxRtvb4rY1FBjV1O+mbi5YPPTLclbzv7q82vtVpOtZVu/bJNsa98ev/1UrXdt7Q7THUvr0DpFXdfOCTuv7Ard1VTvUr9pN3N32R6wR7Hnj9/Sf7u5N2pvyz6fffX7bfdXH2AcKG1AGqY39DSKGtubUpvaDo4+2NLs33zg0PBD2w5bHq46Ynhk6VHq0eKj/ceKjvUelx7vPpF1oqNlcsu9k+NOXj819lTr6ajT586Enzl5ln322LmAc4fP+50/eMHnQuNFr4sNlzwvHfjd8/cDrV6tDZe9Lzdd8b3S3Daq7ejVoKsnroVeO3Ode/3ijZgbbTeTbt6+NeFW+23B7ed3cu+8ultwt+/e3PuE+6UPdB+UPzR9WPMvx3/tbvdqP/Io9NGlxwmP73XwO148kT/53Fn8lP60/JnFs9rn7s8Pd4V3Xflj/B+dL6Qv+rpL/tT7s/qlw8v9fwX/dalnXE/nK9mr/teL3xi/2fZ25NuW3rjeh+/y3vW9L/1g/GH7R5+PZz+lfHrWN+0z6XPFF8cvzV+jvt7vz+vvl/JkPNVWAIMDzcwE4PU2AOipADCuwP3DePU5TyWI+myqQuA/YfVZUCVeANTDm3K7zjkOwB447IIhN7wrt+qJwQD18BgcGpFneriruWjwxEP40N//xgwAEtzPfJH19/et6+//sgUmeweA41PV50ulEOHZYKOrEl212PHDufnfehmA0uU6baIAAAAJcEhZcwAAFiUAABYlAUlSJPAAAAIFaVRYdFhNTDpjb20uYWRvYmUueG1wAAAAAAA8eDp4bXBtZXRhIHhtbG5zOng9ImFkb2JlOm5zOm1ldGEvIiB4OnhtcHRrPSJYTVAgQ29yZSA1LjQuMCI+CiAgIDxyZGY6UkRGIHhtbG5zOnJkZj0iaHR0cDovL3d3dy53My5vcmcvMTk5OS8wMi8yMi1yZGYtc3ludGF4LW5zIyI+CiAgICAgIDxyZGY6RGVzY3JpcHRpb24gcmRmOmFib3V0PSIiCiAgICAgICAgICAgIHhtbG5zOmV4aWY9Imh0dHA6Ly9ucy5hZG9iZS5jb20vZXhpZi8xLjAvIgogICAgICAgICAgICB4bWxuczp0aWZmPSJodHRwOi8vbnMuYWRvYmUuY29tL3RpZmYvMS4wLyI+CiAgICAgICAgIDxleGlmOlBpeGVsWURpbWVuc2lvbj42OTQ8L2V4aWY6UGl4ZWxZRGltZW5zaW9uPgogICAgICAgICA8ZXhpZjpQaXhlbFhEaW1lbnNpb24+MTEyNjwvZXhpZjpQaXhlbFhEaW1lbnNpb24+CiAgICAgICAgIDx0aWZmOk9yaWVudGF0aW9uPjE8L3RpZmY6T3JpZW50YXRpb24+CiAgICAgIDwvcmRmOkRlc2NyaXB0aW9uPgogICA8L3JkZjpSREY+CjwveDp4bXBtZXRhPgoOOmNaAABAAElEQVR4Aey9d3yc5Z31faTRaNR7l2w127LligsuGJveQ0loAUJIJdmwm/Ysm+TZbNpn82QT8ibZQDaVFBZIQicBbIopNsa9y0299941I+k9v0seYwzJu5/38YCBcxFJ0+4y3xn/cZ+c3zlhk1zQEgEREAEREAEREAEREAEREAEREAEREAEReNsJhL/tR9QBRUAEREAEREAEREAEREAEREAEREAEREAEHAEJM/oiiIAIiIAIiIAIiIAIiIAIiIAIiIAIiMA7REDCzDsEXocVAREQAREQAREQAREQAREQAREQAREQAQkz+g6IgAiIgAiIgAiIgAiIgAiIgAiIgAiIwDtEQMLMOwRehxUBERABERABERABERABERABERABERABCTP6DoiACIiACIiACIiACIiACIiACIiACIjAO0RAwsw7BF6HFQEREAEREAEREAEREAEREAEREAEREAEJM/oOiIAIiIAIiIAIiIAIiIAIiIAIiIAIiMA7REDCzDsEXocVAREQAREQAREQAREQAREQAREQAREQAQkz+g6IgAiIgAiIgAiIgAiIgAiIgAiIgAiIwDtEQMLMOwRehxUBERABERABERABERABERABERABERCB0Aszk8cgB/+KuQiIgAiIgAiIgAiIgAiIgAiIgAiIgAiIgCMQMmFmcpJKjIkxYcdI299j4ow9pSUCIiACIiACIiACIiACIiACIiACIiAC73cCYRRQQiaTOF3G/Xq/Y9b7FwEREAEREAEREAEREAEREAEREAEREIE3EwiZY8YOZSaZwHgAzc3NGOdft5wMdIIWxJtT9054bOqV+i0CIiACIiACIiACIiACIiACIiACIiAC72kCIRVmTGoZGBjAxo0bceRouQM5SbVm0n4FF2/avdcfk0ATRKO/IiACIiACIiACIiACIiACIiACIiAC720CEaF8e2HUWMLDwzA4MIiuzg6KL5MUYcJw+MghHNhfhtlzZmNOaSk8fE1YWFCsOfbX9JngQ7zptj3+mlCetfYtAiIgAiIgAiIgAiIgAiIgAiIgAiIgAm8PgdAJMxRWzBjjjYxETm4O+noHUVtbi+zsLDS3tOJ3v78XJSVz8LGP3Qa/P4Ch4WEsWrgIsTHR3M6EmhMBUNB54wMnPqnbIiACIiACIiACIiACIiACIiACIiACIvCuJBA6YYbCimkr4+PjiKQ409TZhMcffwxebySioqKQlJSCtrY25s+0oKzsAB599HHcccfnsHr1aiQnJ8MbEYGICO+xOadjOwvWPJ3kpnlXktdJi4AIiIAIiIAIiIAIiIAIiIAIiIAIvO8JhE6YIVrre+rp7sHOHTsRnxCPQCCA2po6+KJ8WLxkCbq7ujjSdABdPV04cvgwDh4so6MmG0nJKejk6JPXG4EZM2YgMSGBYg6dNDYIdaIo426f+MD7/vMUABEQAREQAREQAREQAREQAREQAREQgXcRgZAJM8FMGP+Y3wkuWRxhio9LQGtbK/Lz8xETHYNArB9bt2/F0OAQ5s6fh/LycuTmTkNp6WxU8HZVVTXOWLSQrpkw+Hw+zJo1y/2Ehx3LLH6DSmPUXxdpgsd/F30WOlUREAEREAEREAEREAEREAEREAEREIH3GYGQCTOWCWMyyah/DANDgxik+BITF+dCgGfOmImLLrqI7plq55apqKxAKoWXfXv3Ydu2HThn7VpkZmZgdHQE7R0daGxowN79+zGL2910801IT0+nyBOH5JRUajZ2FFs2OMUfp81MhQwHn7EHLXRYSwREQAREQAREQAREQAREQAREQAREQAROJwIhE2bsTZoUMjY6ipamJlRVVmPBwvnIzMpBNMeSsnOykZqWioLCQjz80CPYsOEFZOfloK2lDUfKj6CzqxPx8fFIS0vDIIOBhyjumIBTU1OLFubSBCb9mJ6Xz/1xP8ykMUdNGNudpvSXqb+SYk6nr5rORQREQAREQAREQAREQAREQAREQARE4GQCoRNmnHOFhwsPp/NlDNXVVbwziW9+81vO5XL40CGkMEsmjs6X1LQU9Pb14nN33IEIjwcbX9mI5rZmFxw8SmFnoL8P/f39zJeZxMDAALNq/OjjY+OBCZdFk5mZhZGRYUTHxCA3J4f7TZ16n1Rmpk7jdYnGcm9U8HTy10D3RUAEREAEREAEREAEREAEREAEREAE3gkCoRNmjmkhiXS9lM6di5bWVoyNjSIrK9M5WzZt3IiEhETnmhkcGER0dDTS6Y6ZlpdHkaYPnRs6KeZUY6CvH3HxCbBcmREKPN0MCg6wXnt4ZARzZs1BRGSEE2ssj6arsxP5BQWYv2A+CvKno66unvtMR2xs7HG2bxBlguJR8O/xV+mGCIiACIiACIiACIiACIiACIiACIiACISeQOiEGTt3Ch7hdMBYaO/Bg4dQX1+HZ9c/i3kUTsYnJimydLufvoF+95qkpCQKLZHwRfqcc8Yer6qqwpw5czB79mwMDw3TNTNlgxnsH0DxzGJkpGfgaEU5XTmHnOumtaWZYkwqEigI/eK/foEhOmmuuvJKrFixEjEx0UELzRTZY+KR7fLYTTZJMY/mDepN6D8EHUEEREAEREAEREAEREAEREAEREAEROD9SSC0woypHTY7ZGIH81/8fj9e2fgK3S7DmM5mJg/CEcFK7PHAOCuxE9nUFItIr5fjTbG4+OKLWZudjI101phOkszbGQz9NRElPSMDMbExzJ0ZcvuMpdtmgKNOdbW1znkTGB/HxPgE9pcdwDNPP412unXs9QvmL+DI0wi6u7tRVFREMWZKjpn6PfUFkCjz/vyHoHctAiIgAiIgAiIgAiIgAiIgAiIgAu8EgZAJM6bHWGNSfFw8kpKSKbpEOyHEHCljY2PwRkQ4EaWgoACzSmbQTdMIj5c12Hzew+cWL1mC5XS5zKZbZqBvAL5oH/p6e8lowjlf8vJyseGFDRifnGCYcBS6unswylGpEWbSjHL/YZ4wTJ82DXk5eZigO6erqxt9HIuqYRPUH/7wB6xatQozZ85ESkoK/87gfqdapEykedNk05seeCc+Kh1TBERABERABERABERABERABERABETgvUYgZMLM1DRQGOIT4nHW2avR0dnO8aVxemT4H8ebbJk7JZKjSzHMgEllFow3wsvA3wkkJSawXnuQjUzRKJiej0ifuWji0cIxpdc2b4MnPILOmkhXtd3U0IRyjjLVM0/G54vE3HkL4KcrZmR4xDU6LVq8CIsWn0HHTSKamhvR39uPvfv2OddMamoqenp6cPMtt+Ciiy+k8NOHZIpIr0s0PEknykiZcR+YfomACIiACIiACIiACIiACIiACIiACJxSAiETZoJnOTE+SaElETOKZ6CtvQPlR4/SpZKM4uIi9FAImZgcdwNFFu5ro0WT/M8cNpVVlS781+aYfL4klztj9dmd3e3up6GhAY0NjXTDBNx4UmdXhxN5oqN82Ll7F7Zu287Gpk4XGuzluFQsx6Q2bd7sjnHFFZejtqaGYlEnNr+6Cfkcq1qyeDGeffZZlkiF89yKMY1umxw2PE3lz9BHI20m+JHqrwiIgAiIgAiIgAiIgAiIgAiIgAiIwCkiEHJhpqa2Ghs2bEAvRRhw7MhcMjbWNJ/OlpaWFrS2tjCoN5GumEiXC2OjToyjwRgrsbs7uji2FIu42DiOLJlkA0RGRGLMP+JGk46WH4Wfr/N6fbj00svoful14b39DAY+yDpuH904nggPtry2BT5vFC06PD6Fl+XLznTHamjY4gKFJ5hJM+YfQ3l5OcrKypBMISkzNwcL5s5HIbNorII7h/e1REAEREAEREAEREAEREAEREAEREAEROBUEgi5MNPPUN7HHnuMGTL1yMvNxYKFi5DIIF9vpBfRFF2sNcnyXbKzsine9KKpqQn9fUOgEsMGJnur/Gt/OAY1yayYxCSr2C5mYHAA27ZthX/Mj4hwD84//zy6XKZj165dmF1SYlugvqHeVWtb7XZFRQVu//SnMY3umCgGAVuFdgRHp4ZHRxhGzFwa/ngZPGyjVS2t7aijI6fySDniOIqVztDhOz73OYozuW6/+iUCIiACIiACIiACIiACIiACIiACIiACp4JAyIUZc60kJSahiqG7Vn+9YuVylM4pdQ6ZwcE+dHC8aVreNIokzI2ha6avrxf7Dux3LUwJyUlMpJkab7KRJgsGDqcIk8rAXhNLupkPYw6b9Iw0e4p5NokcjZpgjfYsrF5zNh68/wHYyJM1OnV1dzEQONxta/cvvuRiF/obzVDhZIo943TNmLsmlqJNc3OzE4miY2Iw3DaCap57dc1lEmZOxTdO+xABERABERABERABERABERABERABEThOIHTCzLFMliiKG/MXzOfPPFRV1aCzvROTc+BCgHv7+pybppY119bUZALO4OAwR5K6YcG8GVaLzbGnVjpYWjjyZMsCfm0cysMA4XGOH5mfJoGCzMjwMGo5NlVJZ0xmRiZDg6cjNi7OuXMWMbOmu6cLe/fsQ1RkFIpmzMAcumrMwbN06TI01De4kab4uARERcVQpAmgo6MDKakpyOK+YjlK5WXYsJYIiIAIiIAIiIAIiIAIiIAIiIAIiIAInEoCoRNm3PwREOXzuVrs9PQ0OldS8OJLLzmR5TyOHrXTLdPV2YXKykqsX78eo6y67u7uRg9/YtnCFMdxoww6Y+o4BlVbUYeczCwsW7bMNTaZkBPOdiary7bAXnPUdLMye+/efazErsPRo+V00Uw4l401OoWzPvuVVzZi+/btyMrKxK0fvRXz5y9AAp8zAcjybgoLCnjMGFQwu2aCFpx4Cjs22pRMh04MH9cSAREQAREQAREQAREQAREQAREQAREQgVNJIHTCjJ0lXTMpFDUWLljEAOAXKLbE0OkSxvGiRqx/Zr0bLxqlwJLLYN2U1GQ6WvaziamSQk0VXt38GubNLcWHrv2Qy36ZZMaMn+NGJXS61NTUuqDePDYnWeNSWkoq25kmYW8mwDBgG3HauXM7hZYiRPD5/PwCNwlVVVWFg2UHGfDbiefWP4fqqmoXGDyDDVEZmZk8Ny8boZI4UuWjDycMN998C8+nyh1rcoIij5YIiIAIiIAIiIAIiIAIiIAIiIAIiIAInEICoRVm6JoJUExJTknCocNsSYqKwm0f/ajLcGlvb0NLW6vLdpk1qwRLzliMhx96xGW7RFJMsfGlnbt3IyExGWvXrEFBQSH8fj+GhgbdqNHw8CCdLmmu3Wn5mWdi646t8LKxKTo6xuXOhHFYyvU40UmTmpaCuXPncqwpCfV1dU5omaDQ085xJRNqDh4sQwnP4Wzm0sQnJDjXjDl98umgGRsLYN26Z2FNT1oiIAIiIAIiIAIiIAIiIAIiIAIiIAIicCoJhFaY4Zk2sBlpy5YtTiwJjI65vJnp0/I5OtTEXJgRvMzRpjaKNFaP3UOni7UyzZgxE7fffjuqa+sQG20115N01KRx7KkDzz33vHv/fubLpPKxhMQEFBQV4qFHH0ZLcytKS+cwxNeLsZEx53qhLuNGnawFKoVumBnFxbj22uvwn//5EzpmqvgaIC4+DkPDQ3TdTCCRVdnTKQJ1dna68/GPDVMQGmMt9/ip5K59iYAIiIAIiIAIiIAIiIAIiIAIiIAIiICb/gkdBo4yNTY24Y8P/gmzZs901dUvvrgB6anpFE8iEMO6bBNjnnj8STQ1NjJ414eE+AQ6a3xsb1qFT336dhfCu+nVV9HZ0X48g2ZoaBjx8fH48E0fRmtLq8uoGaXIM8rqbcuxyWWor5f7P1BWRlElwHGkGmTlZCFAl4wJNTEUeyxAuKOjk8eMwpnLliOOY1ZVzLoJ0CET4YmgyyeVQlAnXTWdzJeJdS6a0IHSnkVABERABERABERABERABERABERABN6PBELqmJmkCBLg+JE/MOpEmLGRUdx11w9RWFiIvLw8JLJNqZhul0GOJ9lIkbUizZ49h61KGdi8ebNrZoqleBMRHoZK5sGYo8VGkBI5buT1eRkMnIHBgUG89NIr6Orq4shSuhNlbHzKarqtGCow5sfzzz2LPXv3YOniJejr73ePj3M/zi6DCeeYmc68mp/8+CfOvRPp9Toxxiq0AxR2oui2sTpvLREQAREQAREQAREQAREQAREQAREQARE4lQRCqjbYmJDVWHd3dcPD5qRIChyT45M4sH8/9uzajfMuuABf+NKXOc70IkebWjHBmmobJcrLy8WTf33KNSitXn0282ZaMEJRJ5JijGXwZudkw3JofvGrX+KMRWcgKSGewcABJCQlYNHCRbjvvj+guqbGjTnFREVjYGAAzWxdSmMQsQk+vX09TtSJYhX3JNuX7DxN9BkZGUJbW5trhUpJS8Mwzz2DDU65edMozHhPJXftSwREQAREQAREQAREQAREQAREQAREQARCN8pEvcONDY2PT7gAYKu0jk9OZn7LdIwzr6W9rQM+byRWnLncCSDbt25DT2+vE0gszyXaF82Q4CYXzGvNSH19/RwnikNqeipKZpewjSkSW159DYkcffrABz6A8soKeMM9rLmecK6YVzZuhLlgUjmSdO555zpxxsKH+wf6sXHjJpx37vnIpOOmu6cb9fUNrPWOgjcyyuXRWFW2CUg29pSWloqlS5e4tiZ9X0RABERABERABERABERABERABERABETgVBIImWPGRA1bGRnpWLBgoRM2Jllp7Yv0Ia84DwsXLkRsbAydLRRdnCAz7qqvzblio00rVq2An2HBiQzsPXr0qBtV6u3tcRXb1193PVuUZiFvWh4i6WSJYmaMvS6cB+3t62M+TLITZDwRHtfitIDHOvecc7B71y5Mo1gTxtEoq+c2B09PzySFmToX8FvM2myfLxI7du5ARGQEnTTh7j34KNhEcF9aIiACIiACIiACIiACIiACIiACIiACInAqCYRMmAmeZCSFmJkzizF9er4TP7p6ujCNo0FjY2McR4pEeUWlE17q6uudYBLJbJgxCjKlpTPYzjSDwk4Gnn7qKbR3tiMnM4diTiwFlUhMy5+OkpISZDPoNyYqxrlaqiqrOJM0ibiYONZoz2V6zCTG/eNuRMrjAn1TkF9YgCsuv9wJONbKZM4aG1my80lLT6NDJs1Vdg8NDdHpE6A4EwaPx+POLfie9FcEREAEREAEREAEREAEREAEREAEREAETgWBkAkzU9ktYRRHJiik+BjsG0vxI5LBvplYfdZqpGdmsFGp2bUubR0dobNlyAX5mktmguNPyzg+NOYf5TZ+mNOmvbUd+fkFdNosQFVVJcea4l3ddgSzZsz58oHLr0BjQyP66baZw8rs6ppq9HP8KSYxhvseYTsUhR/+N8qsmvAE1mbPnMlRp+koLCpy40zt7R08RhsFm143MjXJkSh/YIwNUVFsikpkAHHIUJ2Kz1H7EAEREAEREAEREAEREAEREAEREAEReBcSCJnaYHXU1orkH/W7RqMcBvqaUNLE3JjomGgUFhRgOpuZhuhWefDBPyIQCGDOnDmo41jRwOAA67AT0FDfhAcfeBB9A33MgPEh0uPlKNJ0lB04BKvQrq+bctlkZ2Vh7rx5dLykw98awDnnncPa7GT8y1e/gvSUNAo7E6zaHnMtTN3d3RR2qpzYU0RRxqq5p+dNR1FBIYWgALq7uxBOh8y2LVso/IxRyplELKu0fVGR78KPV6csAiIgAiIgAiIgAiIgAiIgAiIgAiJwOhMImTBjb9piZiyM17JfcrKzMcQcmKrKaoyODWNwcNBxMWeNn6IMY1/wT1/4PLo6O3H40GHWVrdjZHQUe/fuc6+dVpDPXJgUjhUxRJgtTE1NjWhoaKDLZoCjRz04u7kZmZmZsFwZq9HOzclFZ3snBigGWatTe0c2/GMBZtwwAJiumK3btmPza5tROqeU+0vg+WUihXXbto9z1q51ddvPrFvnGqJsDMujumz3eemXCIiACIiACIiACIiACIiACIiACIjAqSMQUmHGHDO2wtjIFM6MF2tm8ng9HCcKoK211bUnjVJ8sYBgGx3KzsxCUWEB3TLxeOCBB5jx4nehwZYrU1dXy2Ben8t7WbRokWtcshDfHob9DgwM4sCBA6goL3evDzBXxhMRhgyOS1ld9yjzY3p7+1FZUeGanyzbJm9aLmu8ezDMiuwKNjo9/0InwibDsGDhfOTm5jkBqLW1Gb09vRyhykcUx7G0REAEREAEREAEREAEREAEREAEREAEROBUEgidMGN12TzT+Lg4JCYkIuDnKBFXmHt8kqNL42xE6uFYUTVdK3MY1jsf0bHRGOgfwPjEOCu1J+iIaaS4kukalQYHhjhqNOaEF8urSUiIw/nnn48OZsP0c9QpLCwcMWx5CkwEUFtXwz6lcKxcvhIdnW0Ugkb4M+yEHcuRMdfNpZd83LU3WZ333r278dprW9DQ1IDa2loGCmdjRtFMXHzRJUhISnTnzfJs937cHf0SAREQAREQAREQAREQAREQAREQAREQgVNAIHTCDE+OBhTWTnuRkJjA8N0WZHNcaC7bkszBEhifEmrG6YqJYHZM3vQ8ulKiUNVWiZaWVpTMnoUIumts1CkxKQHz5s5140SWXXP//fezWtuLLGbL2FiTly1N1qw0Z/YczGKNtuXYdDAsODElCXEJsa6BqaOjm81NS9HW1sbnm+ngAQoLC51YM4tBwEXFxVj39Dpm3NRimGHBFjx8ySWXuHwcq/SeNasEnjBVZp+C75x2IQIiIAIiIAIiIAIiIAIiIAIiIAIicIxAeMhIUJQxx0xNdTXzYJpcrkwE818WLz4DQwz33bVrD/bv34/6+gYngoTR4ULTi3PSDDF/xoSYVStWunGknq5eZsR0YIJOGi+FHgvvfenll9DZ1YVq7r+mtprNS4OIjYlxjUwruN2CRQuRyTDg+Lh4REdHM+vGwoVns7Z7OoWXIZSVHcK2bdvc9t7ICI5QFbOeuxhnnLEYnZ0d2L1rF+pr65xLZ6Bv0LU5hYyVdiwCIiACIiACIiACIiACIiACIiACIvC+JBBaxwzHliopojz91NOwnBgLAY6Kinbiiwkt27ZsxcZNG3HBhRe4YN5cjhBZ5oyNJYWxGSmSTUixMbGuztqamqzBKYt129kMErbsFxtRMhGllu1Mcdx/MV0vfb19Tnix48UdE2UsDDglNdVl3NjxbSjJzuvokSOI4f5v+shNiImKcaNQo34/3Tcj3GcdnnjycaTzeLbfGTNm8LUx78svid60CIiACIiACIiACIiACIiACIiACIhAaAiETpjhCJKNHdl/Jp7UUeiw0N0+hvDGxcchn7XXO7btoOulEy3NLcyWGcT6Z9dhyeKlHFHKdO4XP3NoOMtEp8yEq9PuZSaNhwHCZy5f7mqumzmy1NbG9qahYfd6a2fqYKvTc88967JsYijO2OvDmSmTnJwIrycSCxbMR2tLCzooDEVTpGnr4GhTYxMDfguQlJzEGm1wTMrjWpkm6eKpqqykG2cIXT1dLpMmNB+D9ioCIiACIiACIiACIiACIiACIiACIvB+JBA6YYaijLUynbliOfbs3eMyYHpYUx0REY609HlIn5nOKupsRPmi0NXd5cSXhvp6pKSkYYwZNOVsWIpjO1OkL9KNI4WHe5zYYu1NPmbKmHNm185dOHiwDFHRUbjs8kvpxolhtkwbTLDp6e7FipUrEe5jKHBMFPpYm93c1oIYvtbGmSrZxFRUXIhzzz+H20VxvGrQBQbHMqzYGxHp6r3XnL0ajazlDvgDFGzs3WiJgAiIgAiIgAiIgAiIgAiIgAiIgAiIwKkjEDphhudoGTM5WdlYe/bZePLJpzAyOoJxZrnERNdg1pwSjipFuxDftpZ2DNIxM8HKpiYKIbFxMRgbHWPb0oAbHxoZHkU0xZVEBv2GMYB3cmIM45NsbhofZ3ZNgK6aUVxwwYU4evQo1tMt09XZTfHGh6TERCfyxEbHUIipwTe/+Q1kZWYjNycHsRR9fDz+8hUrUFtdh507dmLX3l0omF7g2qG8EV7nzPFztKmtvZ1jUXGnjrr2JAIiIAIiIAIiIAIiIAIiIAIiIAIiIAIkEDphxgwmVGYi6Ubx0RVjdpee7h7UhteymvpVTOMoU1JiEmbPns0WphYKLKPweCNcGG8/3S2l80o5ojTiWpliYqLBOBmOM/ldTkx9Qz38/nEkUngJC5tgaHAYBR57K5MM9T3gDjxv7jwsWbIENfV1OFJ2kDk1XaisqMShg4fZzrQY//iP/+RGmPbv2++O0dffhy3MvNm/dz/WnL0Gg8OD7lwyWdedmcEQYYpCWiIgAiIgAiIgAiIgAiIgAiIgAiIgAiJwKgmEtJXJTjTa53PBuynJKU4AGR8POKHmABuZAoEA7rjjDlx6yaUoKSlBWkoq3TAT6GZWjGXT9Pf3o7K8AiuWr8JitiVFM6g33GOZNZ0UakacgJKXk+cqrT10uDCOhgJQOMa5Dw9fl5eXhyJmx7zCgOGdO3e6AOL4BHO+hLlRpYSEBDz99NO4+6d3U9Apc7k3tpN41ntnclTqoYceQhuru2NjE9z52PvREgEREAEREAEREAEREAEREAEREAEREIFTRSB0jhk7QwollhdjNdk+Omcmuidca9KVV16JHrYqtba2uOeLZ82geJOC3t5eZDL4t6O9k0IIOK40gX62Mc0pnY0cjh81NjRStOlmaG8Ut/M7F0siXTeDbGtKpMiyYOFCF9ZbUVFBfcbjjm9ajTlwaqprMW36NDfeZILMUQo+Nlpl4lBFdSWyBjM5DnUBepmD09lhx590x/ZTSIqNjXZOHXtLWiIgAiIgAiIgAiIgAiIgAiIgAiIgAiJwqgiEzjFjZ0h1ZZChuhbwu5whwElJbD3if1Y//dHbPuqqsP/85z/DPzLmGpt8kVHIn56P0tI5rtEplS6badOmMQDY67ZNpXhjY1EW0BugYNLY2IiCwgKkcL9WxV1cWIirr7kGxay2DjAbpokhwAN03Uwwi4bdTlN13Qz/tbDfLVteYxZNFwroqPEw9tfGrEZ5HpZL00dxxlqkou1YdOlY2HA42520REAEREAEREAEREAEREAEREAEREAEROBUEgiZY8bGisKYKzPMKuswjhedd+65DNmtwY6dO5zoYmKHCTdVVVVYvnwZ0tIyEM7GpuzMEpQdOuQEndjYOCfMREXGsLVp0o0o2T5NWLEK7XZWZS9dtgzz58/HhpdexpnLllDE8SGS+x7mqNMLLzzvaribmppdyHAs67NtWY23OXlM3DGHjiciDOUV5dzHC5iWm0enjslHzMdhgHBklI9iEB0zrNzWEgEREAEREAEREAEREAEREAEREAEREIFTSSBkagO1D44JjWM3q7KPUvQYY7hvNEeCJgITpoy49zBJAcRaj9o7ulyAb0xMDM5eswaXX3apy5exl0V6TWQZdMG/5oLBpAczZnL0iW4aa2ZKiIvHrJJZ2M0MmT279zBIuJVjS9V0v9iY0ji6errRyTru2NgYrDxrJQJscbK2J4+H9ds8gAk8gXEKSCPDLodmw4svcpSp3Tl08vOno7+3D0NDA6eSufYlAiIgAiIgAiIgAiIgAiIgAiIgAiIgAo5AyBwzZjmxOuvdO3dh7/69LGWaRAezW7yRXpc5Y6LLlDAzht27dtFZM4RsVmubg8balkZHR934UEpaCh5++GEsXLAI6RlpLlvGmpLypuW5euyOznYn6oz5RzHAPJoh/lhrUwTDgC1QOJxuHQ/dLibwxMcnYJSv623qxbzSUnR1d7vnXGowz3eSrhw7ruXXWA6NOXHKDh10jU/nJp3H7dXMpH83IiACIiACIiACIiACIiACIiACIiACp45A6IQZc7twpCgnNwdbtm5BK8eOzIni42hQpNcLT3gEhRtWXfO/HgokNbU1zsXym9/8BhnpaYhiRbaNLEV6fdixfSe279hFsSQeudk5uPWjt1C08XC0aRw9DAweGfXT/RKOuLgEJCcnIzU1DREUgFJTU9Gb2eMybrJzcjF/3lxs27oVzcye8VKo6WRNt4k4HubHWB22n4KMl+fG9m0wdsad/0D/AMoOHsKSpcskzJy67532JAIiIAIiIAIiIAIiIAIiIAIiIAIiQAKhE2a4cxsVuummm5zYcfRoOeKYGTNqIgrHiCyzxZwppoLkMNclioKNjTUdPnwYh8omsWDRAqRTYLERpBKOKjXU16O6psaF+UYy8yWSwktbaxviY+PhSWVFNvNibB/5+fm45eab0UzRZWh4iGJPBPcR61w4s2bOQkFBoRNtpnNMaZQ5M+WV5aBRxp3biGeEolAGFi9eRpEnFv7RMb6LSRca7PMxE0dLBERABERABERABERABERABERABERABE4hgZAJMxafa24YP8UPa1Z6kdktl1x6CVJTUlFBkSYxMYGjQ35EUDiZOaMYftZWd3Z1uiwZCw62GmsfXTMx0TFYvGQxZs6cifT9++mQ6YM3IsKNQfUN9KOxpYkZMh2oZrDwtGmHUFhUyCybWBQVF7na6wiKQNewqcnaoVraWhnkGwlre8rNy6XrBuju6cK+PXvd8cIoEkVF++iOWexCiw8cPOhGoKJiotzxTiF37UoEREAEREAEREAEREAEREAEREAEREAEbGAnNCtsMsyJHTuYH2OiyPx582g+CUNeXh4OHzmKe391L7YzsNdCeNMzMpyrxhqcxgOTDP7tQ3l5OauqY9z4UsA/jjSON82kc2Ymq7Dj4uPoalmMiy44HwnMfRkcHEJLcwuefPIJfO+7/wf/fd99aKxvdBk1k3TtXHDBBcyvycL3v/cDDDD4N4UjTjFRMYhmFXYyRRprcLJztBwaL0envN4Il4XT19/rXDhZGdnMrAmZhhWaD0B7FQEREAEREAEREAEREAEREAEREAEROO0JhEyYoVnGNR719/WxIWkUn/2Hz6KLjpj1659l5kskNZpJtLd30P3ixSWXXIKzVq1CSkoKBZhUF9hrwbzmqikoLMQDD9yPZ599DoN9A4iOjnZijo0s3XbbxzF/wQLk5uY6B4zlzFRWVmHTpleZSbMdySnJGBkextYtW+m06WGeTB1GeC4ziouRnJbswohHKAZZmLCNO1m+jNfrcS6eaDp1kpOSXAhwpI+5M2av0RIBERABERABERABERABERABERABERCBU0gghDYQDjMxvMXqqE0wSUlJY77MCLZv387cmVtQMqsE+/buQ3d3j8uRycjKdO6VJRxbGh4eYUBvi8uHsZGoffv2YsOGF/Cxj33CjRn19vS54GCr106lmNOflo45c+ZQ6GnntkMci2IlN7fzery8P4y77voh668T6bRhbgxbmWxUKtITia7OTvRSOPr61/8Vu/fsQUV5Jd06vRhj1o0TZuimmZY/DXkUfiIiPKcQu3YlAiIgAiIgAiIgAiIgAiIgAiIgAiIgAiEN/2XCjCeMLUxRrgp7YiLAFiUTaiiYcFQoOSWJob+5HBWKZu5LFPJy8hjcO53V1x5WY8/HmrPXoLa+Di+9+LILDB5jEK8F/Mbw9daqVLGpHEMDQxRbEpBIZ8vQ6JAbnbKqbcu0sYyardu2OpGFuhAaGhuc+8aEoA4KOD093TwWn6BIU1RUjOVnrsDg8CBeeuklbHltM4WeeWxqSqBAE4fEhGQnEukLIwIiIAIiIAIiIAIiIAIiIAIiIAIiIAKnkkBI53M8FFmmTctDf18/1q1bj+GBQTcuZE1J7GxyATfWrmQCidVqX3rppXiZwsgzHHdqa2tx77OR40djY6PwMP8ligKOx4J/Ke709vRi1+7drj3JwoEXLToDGRk2khTjhBoTfPbQBdNIQeayyy7D/Pnz4YuIdCNJo2yDsuDgaJ8PA4P9+NnPfuZGn6waOz0tAy9v3IT9+/dgAcekhgb6MDg0YPqNlgiIgAiIgAiIgAiIgAiIgAiIgAiIgAicUgIhHWUK+AMcHQq4XJff/+H3DPONZbvSDGRmpLvsFhspskpty28JZyNSDFuYLMS3jG1IkxyBuvjii5HPeuvIyEi6ZijOcJzIQni9vN/S0sbMmHrXlmQjTQXT8zHMEOD2jnYXGOyjKycQ8Lu2p9kls1mJPYFytkHFMzh4ku6drs4uBgEPsjXKj8cefxwvv/ySCxROS013x7BOqUQ6Znp6+3mcBsyYMQO095xS+NqZCIiACIiACIiACIiACIiACIiACISSwC9/+Uv09vY6s4Llu2qdfgRCKMwAvWxX+utf/uKajfLypqGro5MjTCkoYj12RnoGwj3haGttQUVFBRLi4hgSPOYElHFmvFh7kgkyhRRmUpKTnICTkBDPsScfPBRy+vp6cfjQYWbUdLuWpTCqPOF04kRG+lw+TBYDfWeXzsUg9xMV5UMdRZxwzjSZqMMbqOOY1KZXN6GdFdrTpk9DX08PDh08zNGraiRzHMqycawZyqq5vTxPu68lAiIgAiIgAiIgAiIgAiIgAiIgAu8mAt/85jfR3NKMD994oyveeTed+/vlXEM4yhR2zKVy1I0dfePf/g3prLyurqx2GS/j4+N8fhKH6WL59re+hW9/5zt44YXn6a4Zca1LVGKcGMOGbZc74+UoUzzzZJKSkpGQGM/a7GLk5OUyt2YCE3TlTHDYKIKNSk644bY2KmVOmtklJbjowoswt7SU7poJ9Hb3crSK40kcW6qtrUNNXT3WrjmbocJLORYV49qZEuiUsXOrrKx0Y1HZObluv++XL4XepwiIgAiIgAiIgAiIgAiIgAiIwHuHAEuReX0dwsv/9w6qd+SdhMwxY5ksCXHxWLlyFQIM7U1OTkYUq66379iBP/zhPgomc1BfW89qag/HlMbwwvMvUJjZwArrdhQUFHDbBDfeNO4fN42Ga5JOGQ/HmcIQQdfL3HkL8EE+ZxVLFVWVbhzKXmaV1xH8qaCosn//fswrnYdZFGcscNhEnOqaaubKDKCwoAhZmVluv5ZNEx7mYbX2qKvrnjt3LuJ5fG+kh49T4DlW0W1noSUCIiACIiACIiACIiACIiACIiAC7yYCFtWhdfoSCJkwY2m55lj5X//ryxRjdqKqqsrluZhw0ssxpN17duFo+WEXDvzpT92Ohx76s8uWsedj42KRmZ0JTxhPzxsGG4Nqbm7mqFIY82E4icQvlYfjSEkccRqnW2bfgf0cS3oVlzCTJjsnByOs5R4dGcXRo0fR3toGPzNltm7d5h6zcxoeGUZnVwddMmfQCbOCmTJejkB5OcYUSddMHIqKi5Cbm4fJ8QkXCmyiUn5Bwen7KerMREAEREAEREAEREAEREAEREAEROCtCDhNZtLls77V03rsnScQMi/TlMsFCNDVMp0ZLhUVlRREhiicZGMB3S7Tp09HXGw8YmLjcDZHieaxNclLceTLX/4yx5Rmsg1pkONJEy6Id9nSZTj3/HMxNupHa2sr2topttCFY8uSX6xK25w2ljczxpDglqYWpKWmoGROCfcf63JkKiur0N3Thds/czuuu/Y6ZGVluTDi1LR0N6Zk1di5HI3KZDZNMselRjlSZRk4/f39aGe9trl+tERABERABERABERABERABERABETgXUXAxlm4GBYydYO/h4eHccmll7jCna6uLtxzzz246KKLnFFh9uzZ+OxnP+uiPWyDnTt34tZbb3XX8LE0MixZsgQPPPDA8X2deKOlpcVd059/3vk0O+S6mJJEZriuXbsWDz7w4IkvfcPthx56CGeffbabtLFokQsuuAC/+93vsHHjRndet9122xteH7xjwcbW7pyTnc1r/xisWLECX/rSl5w2EHxN8O9vfv0b935//OMfM9ak1mkDrr2ZObbWyPyZz3wGnZ2dwZe/rX9D6pgxoeSRRx9x40X2FYjyRTPINxxnnLGY9dceHD1yhC6VKPiio9yYk8/rw3nnncfRoRjcfc9PUVhcgFmzZrsGpzA6ZEYDYxxP2ufcM1ZrbeHA8YmWB0N5hl82U5k42cQmJj+i6YxZNXMVmlubUVFewdGkWCcE2fiSjS7FxcejID+f5+RjzkwtXTx9iIqMYihxOmYUz+AX7X66ajrday23xiq6tURABERABERABERABERABERABETg3UiAnpnjpz02Nob169Y7qeaWW27BM888Y8qNu64+cuQojvBafeOmTfjhXXfhhhtuQF9vH7ee2n7Xrl24+eabXUGObRtcTz/9NG648QYMMM/12K7cU319/XjllVfcz5atW/CTn/wkuIn7+4tf/MKJIq9vFMaYE0adbHgBhQWFqK6upiiU/4ZtTGuwc3jmmXV8/PX3tXXbVk7LbMWjjzyChx5+GMuWLTu+XdnBMjz37LM0bHTjBz/4AZqamo4f0mJQ9nMSx97Dtm3bnJHj+IZvw42QOWbsHdoIUU11jXtzNmq0YuUKp2B5IsKdIJKXl4fCwkLEREUjlyNIs0pmsQXJRoqiXGBwV2cPXxfplLxzz1mLpPhEJ7wMDgzj0JFDLi/GGpqsQtsfCKB/cBDj5mzhu7IK7shoH/cdg0E+np6RgZmzZuKvf/0r9u3b6xwxKcdGlNauWYsb+WW7+JKL3QhTBJ073V3deOKJvyCVzpviGcU8hudt+Dh0CBEQAREQAREQAREQAREQAREQAREILYEwN+LCqBAexkSZhQsX4j5mwdrtq6660gkWZQcOuBYne+09P7sHu3fvhokoUbxet/Xv//7v7m/w15133ulEmWy6V+6mA2fv3r3YwYzZb33rm4wnyXP7/PnPf+6mYILbvPbaa/inz/+Te27Z0jOdQ6aeDcq//e1vkZiQ6EQZpxhZevEJ62tf+5o7V3sbn/nsZ3Dw4EE0NDTg97/7PQuDklDLfZhwYwLUicv2sn3bdrS1tTF25Z+xlSLMsxRrPvCBD1DfCUN9fT3MhfN2r5A6ZmzOyNws0fzgRpn7Yh9QMd0ozU0NDpA1H40HxjHGemyrqLbQXXPGZGdnYeWqVXTRRFFomUAKK7aTKZD0UKWLpFCTwNdazXYXxZMJtjuFc5sxjh69+OKLjp+pYqPDI/Bz9MlUPVPr7DX2Bdq3d78bbSosKnatTwsXLXLjSxkZ6e449qULZ8hwQlIiXTVxLn/GcmcsUFhLBERABERABERABERABERABERABN4bBKbEjnROjZiBwcQTW+eccw7biRNdZIjdv//++3HZZZfZTSzi9XNdXR2+S1HGcmStbdnDQp8NGzagrOyge82vf/3r46+3B2z0ya6pLbbEhJIDFHwsQsTWF77wBXecoqIiOmpePi762OhSCUt8VlEXMEfMibqMuVt+/atfue0///nP40c/+pG7bb9s5KqgoMCNTpWXl+NnP/uZO8bxFxy78Y1vfAP/+q//evzhNWvWOD3CmplNqPk3tkq/nSukjhkTZSbYhGSBvHPnzXeWoIbGBmIN5+PjLnzIbES7aYVq4ixagEG+JuAkJSY5ZS4lJclZpmxOra+7x83BWUiv2ZnOWXsOlrLiehdVuJ6eXrY1RbBVaRjVVdXMisnDQTpq/vTnP7kvjY01UfFxXAN08QyxfWmEAcBmidpMJ882Wp3W0ca1Z/ceNPL8qM24L05+/jTuc8QdlwrT2/m56FgiIAIiIAIiIAIiIAIiIAIiIAIiEDICwcyZD37wg8dFGTuYGRqWn7ncHddyYoKiTPBEZs2axWt6TshQZLF8Glt2bf7qq5tc9szFLOU5eZ111lnHHxqhqcJWgFMv5qqxZW6boBPHPcBfK1eudPkxdn/SMkuOrXXr1jEDlgYNXribwHLyMpEln7Eltp5n+/NbLROETlwWX2Kik61+xpy83Sukjhmruh4aGmS+SzztRInY8MIGF6T7iU98nB9iALEM5jXXSyvdL2Y7slDfbbQVmVOmp7uX40dpHE8aQG1NLVW4CI4u1VBNG2FOTQrHkmbBQoRqKcSMjA7jzDPPZMPTNIouDA2m6DPQN4D165917phbP3Ir9uzdQ1tVPybo0KF5xtmUTDSyqu7+/j6n7jW1NGGQocMWNpSZnuUyaQLcl99P+5OpNVoiIAIiIAIiIAIiIAIiIAIiIAIi8C4iMDW2ZJe0r1/T2mTJpNlQ+L/CwsI3vZusrGz3mDlZTl7msLHFyZ/jy66hzd1iPwfLytjCvJstyeUuq6aM9+1nanF86pjIUsbxo1GW99iIyxlnnHF8XyfeWLpsKda5/JvXD2auFlsTPPlbP3qrew8nbmOC08BAv3uovPzo8aem3n8Yc2TTnBHj+BPHbgQdQyb6vN0rdMIM34mF9Tax5jonm2NEvhjmtaS69+djpoyZWGbOnEmHSiPHkdzXAgGKNY0m0HR0oJUuGbM8mXhjNdbxbHCyx6qZWdPd3QULGTIhpri4CJUUZ5LYpGQpzia2mPMlnlap/KIC56SxxqW4uET33ACFniR/Mj/8KTfPJF099o0aNzGndwD1DXXYtn0HPnzjjZgzZw5aeP62zYlf4rf7Q9LxREAEREAEREAEREAEREAEREAEROD/DwETQkyTCQoitg/nlqEoY8sak09eFgViK8Yu3E9ax/dzbPvg01u2bIGNFll4rlNbnK9m6llzw0w5ZV7f6PChQ1NP8qFk5r++1crKzD5hL1OvsMZlt7jdX578i7tpZzvphCc+6A4xdf7WvhRcU+c9iTQ2M7/VspEst6Y2fauXhOyxkAgzUx88P2p+Aay+2s8MmfiEWAbqXIHyoxV0zrzAtqVZbr7MhBc/BRyvx8tpoXEcOXrUiSDmbumhm+YoFa4d23YgIzODIk0MnztWYd3RjtkUTj507bV4+JGHUUNxxsacTJjp6ekh7DRccdml/IAZ3ss8Gav4sk/IBJYAz8dPN8/wMDNq+HqfLxKRbISyUGKeMnNvAu4n2hdFYakJbS3tx2fnQvZJaMciIAIiIAIiIAIiIAIiIAIiIAIiEAICdp37t5b3LfJUg+JL8O+J25qtwpYTQ47t+PDhw7iMtdXdvBY3FWhuaSlWnbUKC+YvcCNJ0dHRLsPFtgvuM4MFPcFlNdVm3Dh5tbW1uodONEokJSW4x3IYmXL33Xe/YZPgvoOvt1Kgk1fwuZMfd2KAe1NveibkD4REmAm+0bi4BPaBL8Tw4JDLj0mfns6xpXaGB/23a0ny8QuQnZ2DacxyMWAG0SxOAY4bmaLmH/e76uxWfhjVNdXuA1686Aycf/75HEma4NjRAFuUclhllY2tW7bCS2GG0b2u8cmqtKPYyGQfVhaDhUZGRp0QdN211zlb1bDNtfF4HXTnDA0PYeXylRRwCrCO40/79x+w3GJn7bKabHprQv5B6AAiIAIiIAIiIAIiIAIiIAIiIAIicKoJ/C0DyJSLxi6z/9Yr3vpMwo/lt9ooU3Bba1HqZvarlfk8+cSTuOKKK96w8eOPPz4l5PBRCwy2NXv2bPeXe8GePXtcg/OxB47/scdtBQUXu11SMrWd5dXacd5KWDKhaJgZtMGsGdvu+Ppbb9cBeeOxjm8T4htvlo9O2QEn6WCJwHyG/qaz8cjqq159ZbMLCLJgnTrmxhw+dARlh8o4vtToWpcsWCiKLhVPRJgLAR4dDWAVA38sSfnTt38amVTUTJ0r4Qc4yCwZs0jFxccjh8KMjUnZF8QEnViOMSUmJrhcmwhm04yMjZpo50amLLjI2qFeeO55l32TnELLFAW/JP69/PLLOcL0YQYMXYyRoWHm3/TQSeOFVXIHv3CnDI92JAIiIAIiIAIiIAIiIAIiIAIiIAIhJvC3zDJBF41NkZy8/t7173GRhDsOvs5qse3COisj602ijO370Ucf5e8pRcSMGLbsutyyZcKYdXPXD+9ykzbuiWO/LBj4L3+ZGlU68XGr9rY1zGv2e1jLffKy1qjly5dj8eLFuPDCC09+GpZm8lbL3oudYfA9vdVrQvVYiIQZfkKUzyK8EYjm+JGF8th40atbNzOEpw//+3//b1x/w/Ws4IrnGw9jeO8IEiiwlM4uxbz587HmrLVYvuxMF/QbHuFxKdBf/9d/w5e++EVcdNGFSGCYcFSkj/sacO6YJcsW46qrr+T9IUKe5H4TKNJ4mE0ThTBPuAsVNuHGVW73dDnnjLlw4uJjsWTxEsybN48d6fEuYNganHKyc10S8/79+zCDdqr09NctVqH6ILRfERABERABERABERABERABERABEQgVgZMFhymZhFfkb+GYCYo2x7SUN5+S2/j1IN+CggL3muaWZhc1EtxgaGjI1WTfd999fGhKIgoG89prvvOd77gJosqKSpcZu2nTJtTX1+PBB/+ICy64gHEnb1ZRPvShD2HxksXcOgz/Sm3h97//ve3KLQsGtqrtPmtW4jl+7nOfCz51/H0eM/wcfzx4wwQn5wIKPvA2/g2JMMMkFwfB3CrZWVnoaGvDkcNH2XIUS7BhWLZ0GW5guO6KVWcxETkDkRE+eCO8CI/0UCBJQOm8Ulx7/XUM4q1nlfU2VBCuiTeJSUkcN5qJlJQUzF8433Wfr1v3HHq7+5CSlOoqsM1Rs3btua4Ke8+u3Wx36sYYx5YsHDgnJxsvvbwRNQwQniD0cYo4EazZTklOdQ1ME1TqzGmTOy2X41XTnSUqkdXdmZlZxz/Et/Gz0aFEQAREQAREQAREQAREQAREQARE4P+OwDHLzHGnC/dmYoyJEEGx5OQDWEnO1NPu98lPH9vs2I75rIkhwfDc6xkfUlBQiOUrlrvM1x/9Pz9CWnqaK+axHR06dPj4/mxq5atf/aq7v3XrVifOWBjxTTd9GJ2MHUlOSnHPBfdtdywG5ec/+zmfS8IghZ/bPnab0xVKSkowY8YMvPTii3xVGK5lHq2dV3Adf/9Tbzz48Bv/8i2ZseTtXiERZoKfr3vjhGaV1LV1NRRpslmfPYQ//ulBlB04gIhwDywEKNLnRVJyEt97mBNjJsYnEB0TjRHOhJWXl2PLa5vx6uZNaGpsgDld+gf6OSYV5RKi7/3tr/CrX/8KZQcPoLKywtVzJ1NMMZXtp3f/FOufe46vi+UXjz3p/oB7vKGxnl+kSQwNDLIqu9+FD9fV1zFvppP7jUReTp6r3brpppso3Hj40nEJM2/3N1PHEwEREAEREAEREAEREAEREAER+L8mEM4pEltvFYT7tx730MAw9dybhZkTHTbBfVpD8v3334/cvDy3XW1tjTNZeHjsT3zyEzh65Ciuu+4699xjj9lY0+vru9/9Lh5++GFcdeVVLiM2ntMs5zFX9o9/+hPbmG92LzTDxIlr2ZnLsP/Aflx55ZXwcZqmvb2d9dysxuaFfzKNHN/5zrdx/3/f/4b3HDzXYOPUifuz2xZh4v56PO7v2/krJOG/7g3QfWJiiDUzmTvFlqlcE6ylfuaZdXTIRPC2jR0lusctAGiSgkxNdTVK5851WTMXXnQRqlmFZdlArc2taKdwMsjq6ta2FhcqHOllwK8v2okrTU3NaGDVtgUOhdOpU1xcSEEmGn0coYpkpo0JLl1dXazC3oZ0q8c6JoKZUGSCzRNPPOGquecwv8bDXJkZM2cgLTUNh8oOcqQq4M5fv0RABERABERABERABERABERABETg3UTArpNPXvGMErEYkL+1HnzgAdjPW61L2b503H1ywgtuuOEG2JhRNa/pzSiRR5GmuLj4uJPmAe7Pfk5c+/btY6FPrtvOtj15PfLII+6hJOoGJy/bzq7jTUs4cuQIWlpaXPW3Bf6+VSDwXXfdBfv5W+v7P/gB7OedWKETZuh+sYYkq7eeNm2aszAZnHEG/Ri4cYowgYAfCVTDoiigvPLKK2xXysIcijL2BYmgiHPDh29Ad2c3RhneawpYU2MT+vlcW0srBgr6kZqWinPWnoOBoUEXKtza2soGqEG0NDXhrNVnY+HCM9DQ1EjXzVHXymS13bt27sJXvvIV2pquY0rzep1CLwAAQABJREFUkBNz+ujAqaiowMaNGxksXMJ9nosM1m3HxMehubUFhcVFx79M78SHpGOKgAiIgAiIgAiIgAiIgAiIgAiIwOlOwKJCrPb6raqv3+rcP/CBD6Curh6f+tQn8ctf/vINL9m/fz8ef2yqzWnZsmVveO7EO2YAKWU9t/28W1cIhRkiMccMf6656mrMnTfXuVQsjMcsRE5hM4GOLwjnjFdHewcbkHz44uc/j/sfeBB//vOfcc5552LmjJnOZZOYkIguOmYsTNhyY2LjYl09dm5eLtuWxim8jHCfE4iPT3A/k0yWzmdOTKQvEq9uehXVtbWoZzqzfWh2/DlzZrsWJ9vupRdfwmxWbrWxyttHZ411pTc2NqKPY04m1uTxGK9Xeb1bP2qdtwiIgAiIgAiIgAiIgAiIgAiIgAicPgRMcDFh5rf33uvEnEsuuYSmjmS88MIL+AHdK2P+MUQx/sRiRt7LK2TCjGkuJnA0NzZjOgUSC/z1j/mdIFNVVeU6xU0UMZEknPXYJuJQoXHjRKOjo3jyySdQVVWD62+8HsvPXI7AxACFmEiG88az5joO89neZO6bA71liGGocBjHlywUOINBvdkM+T146BBq6mqRnJjMY3jcuJMTf7gPc+cUFRWyOusiunYCFGB6WKe1EkvPXIrtDBs+zNk0m6mLYCBxTU0NRZrm9/J3QO9NBERABERABERABERABERABERABN52At/73vewYcMGdNN8ceed/8KfO3kOx/QBRqJERfnwxOOPY9GiRW/7ub2dBwxJ+K/lt5jO0syRo298+5t4Zv06N07U3tnOhqQpxctCfwc5duRj/ksGu84D41M5LiboTPC2CSbDI4PMiOnla7zwRURi27atGOgf5vhTAmKiY5DABqVx5sPY3NqMoiLs37ef4b/DLrTHXDV79+5DC8WhYiYzZ7H9KYE12jFRMehlddbGjZvw2muvwexRbe2dGOJYU2JiIiu04+mcaaHrJh6LzliIlStW8nhxb+dnomOJgAiIgAiIgAiIgAiIgAiIgAiIwHuegLUo7d6929VaFxUWcIqG1/6celm5YoWr2X6R0y0XMXv2vb5C45gxVYYKi1VfD/T1MxumEfv2+vACa6tG6ZKxxqV0ZrhERUUxlGeqUnuQDUmtzS2ub9wyiGw2LcLjpYuGo07MqbGg4A7WZaUykNfGouoa6hgsHMYa7WHX7BTLEaf1z67n/VGksuHJ3DgBZspYoHBKSjLS09OZdTPd7SsnN4f2qCS2MQXQ29uDvt5eijFtiIuJw9jYGL8IUc7ZY+NTq9ecjdI5795Ztff6F1jvTwREQAREQAREQAREQAREQARE4N1LwMJ67777bvcGJhhJYj+mB7yfVkjercuPYXbMrFmz8Ovf/NrltHR0dqG1pRntbR3UbCYodszFDTfe6B7ro4PFLDb9AwPo7Oyk4wXIYa7LwkULEcuqaz9dMQG2OYECjYk0XlZYD5iQw9TlFgb+dnX1sIEpip9bOIWWXjYz9bh9mchiVdzmhElgvVaSOWwCWcyTKWENV64LIe7r7cMwm5kGGSBsPxYQnMhA4l/f+xtczbquGQwuimT9lpYIiIAIiIAIiIAIiIAIiIAIiIAIiEDoCLioE8advN9WSIQZM8wE1+rVq3H48GE89/zztCVFYsnSpRifHMfo8AiFmxKMTwSYJ/MX5FIo6aGgYsLMxEQYXS8pOGfNWubPeFBZXQkft53Gui1vJJudJmjIYdDvBG+M+8eZXTOCwckAVTUPg4PqUM6MmHAKQxN8jZfbZWdlM0/mQpSyCnt/mWXSRCOOOTXDw6Po6euluNOGLa9todDT6iq9UlLSUXbwAA4fPIQbb7zBhQkH34/+ioAIiIAIiIAIiIAIiIAIiIAIiIAIiMCpIhASYWbSJczY7zA3uhTJwF17yMOxpRUrluPqq69GVXUVkpLi6VYZ5ihRDzIzZzGEtw8//vGPsHjxEuSxkzxAYSU2IoqV1c2Ii43DZ26/Hdt37EQje9htHCqCIg0VGLdv5gI5R43tY8u27VizZjWKC1lzTYeNOXGsWjsrOxOLly5x2TYNjQ3Yw1k2V+rNXVjPeiOrtT/1yU/iE5/8ODZt2kSDTjhFnwD/nig1nSr02o8IiIAIiIAIiIAIiIAIiIAIiIAIiMD7nUBIPEJWkW35L+OssW5tb3d1V0X5BchIz3QZLh62HS1etJjjRYmop8iyZ+9etDBfxnJpmppa2KIUhtK5c9FAscRcMHk50xBBUceqslOZF3Pw4EFXbz171kzX0DQwMEQBaDSo0WCSY0+eMA+SWLNlc1Fbt27Fz+65G69t3uI+77T0DCQlJKGFo1CWc3Pm8uUo4XiTiT/eCB8bms50dVxLFi/Gzp07MDQ4/H7/nuj9i4AIiIAIiIAIiIAIiIAIiIAIiIAIhIBASBwz7jwpzljWjAXw+pjRYs1IJsQcOFCGX//ql1h25jKU2CgT82PaKd7UN9S7UaUVK85Ec1MzklNSkecqqyMY9juPo0UH8cPv34UC1lybgBPLiuycnDw6YdKwbftW19I0i+JKL8ehoqKiqQxNuVxspMnPQGBzwDRxv0Vsb9q79zk2Pg2zwWmEuo0HpaWlyM7OZm7NAEZHR3Cw7CDiGFxsTp/q6mrMntPO4+WHAL92KQIiIAIiIAIiIAIiIAIiIAIiIALvHQJm0PjRj37k9IBrrrkG1ryk9fcJhE6YMfWE/7Pa6x3bdzjHS0xMDNpYX/3II49g6ZIl+Nwdn0M8BZAZM4rdOFIYRZLp0/PxItubypjvYmNPaRxByi/I58jRJNY/9xyFlUKce+55CPd4mBUTw3DgGAwPDjkxZ+2aNazeTueo0ggqK4+y1cnjGps84RF00YDBvmP8coy7Gu29e/ZQ/ElmC1Q/m56SKRwVs7kpA2VlB/Cbe+9Fbm6ey7EZGxvlOJP/71PUsyIgAiIgAiIgAiIgAiIgAiIgAiIgAq4h+c5/vpNX8JO8vp8uYeZ/8J0IySgTFQ0emk6VgN9VUm/fuR3PP/8cShi+m8JQ3+HhIXR1d6KZLU2pqam45JJLER8b76qzrUHJtS7196OmtpZjS0c4grQVFRUVyMrKdKNNHo4neTjuZCLMeRRpljA3JjHJGpfGmVWThWW8f/RIOf7y17/g5VdewZGjh925hHsiXO32BEed/IExBKjkWQV3GR0ydvyCwnxERUcx92bQjVFt5xjTc89toJjU9j9AqZeIgAiIgAiIgAiIgAiIgAiIgAiIgAiYKKP1PycQEsfMJMeHbJCosbERmze/6s5mYmIS6RRhUlJTeD+M9daxyKRDpZ3CiIkhviify3uxDvPM7Bw6Wfpcf3lvTyfa2lsxStdKYWGhG2GyXvOmpibeL3Kul6KiYo5DdWDDhhc52pTC/UVjjCNUJroM79lNkeaIs1HFcsTJ5/Mx+2bCaUdTqhTLu60nnRk26WnprNROcq1P9fW1GB6leOMfwQjHm7REQAREQAREQAREQAREQAREQATeLQSmzALvlrN9L52nTcq8QoOERZvMYyyJ1v83gZAIM1PpLpPMcBlieO4u5r/EIi4+DhWVrL32RSKBty2/JSMjE7vYjFRZWcXA3WXubGMp2CTRNZOSmubyaXbv2YtBVmvHUGwZHh52NdcmrFRUVDk3TlFBEdIo+MTHxmFwoN+JMXHxCQwTbkR3dzeFmxS6ayIosgzBa8dOsCaoQXR1dXHUiYHC3K9ValvgsJu9csLeJF0yHRR3xtgclQQPA4i1REAEREAEREAEREAEREAEREAE3gUEnCYzdVXqrvGcbeBdcN7vkVMMo1Hj7LPPfo+8m7fnbYRmlOnYuXsY3uvzeulGCacAEoMqCjApFFGuv/EGZHIs6cjRI8xvGXWjQqvPWs2q6k/jrNVnOREmgtuGc2SptqYWT3Mk6fHHH8f27dtdrXUv67Xtn9kQs2X8dMbk5k3DzJkzWbc9gH7+WK7NIEUh+8lIT8e8+aUuS6anqxsDg4MYZ+CM1WrX1tVyFOoc3MDz6e8bwED/kMu6maSDxtQ9+8/Cgb0UcLREQAREQAREQAREQAREQAREQATeBQTCjg3SuP/TPSjQvAvO+x08xT/96U9OTEmhsSEhPh7nn38+fve732Hjxo24+OKLcdtttx0/OyvNufTSSxlJcgmOcDrlQx/6kDNQrFixAnfddZdrYra4EtvOSnhOXPvYyHzLLbdg5oyZNEh4MW3aNFx00UX4j//4D3cdf+Jrg7etkOeTn/wkFp9xhjNW5BcU4Oqrr8Zfn3oq+JI3/B1l+c/3v/99mj+Wu+gUc/CcwW1vueVm7N+37w2vPV3uhFBxsHGlaOTl5U29V6pm5krxeqL4IcxwIs19//3fWLVylaun7qK7JTra50adLNQ3mvCmcdukpGQcOXyI7pdeiir93IfHtTxdxi9CV0839uzbi1XLVyCWLpyhkSF0cKSJs0n46le/is1btiAiLNzVYA8ws2b7ju148vEnMYdZN7XVNWhubXVjUxddeAH+8If76MI5itmlc7Fo0ULU1NVhcoj/oN2/Y/1jPl2+sDoPERABERABERABERABERABEfh7BEY4cbFj506sXr36mGFGY01/j9cvf/lL3H777W94icWEbHhxA4oYH1JVVQWLHAkum2RZt36du3vVlVc5w4Xd2bptq4sn+exnP4v1fN6uoj/xiU+419mv1157Deedd54LB7b79nxDQwMa+fMci34effRRxpNscPEl9rytP//5z/jUpz5FI0Xf1LU5P8p6ZtHW0WTxxJNP4LOf+SzuueceljJPXbNbTMmHP/xhPPbYY27/Tpvjfqz8Zw9/Hn74EfzmN7/BzTff7PZ/uvwKmWPGAeAvc8okJSe7kaDYuDgG7BZQFZuO5ORUDA0MujDfJUuWYs/uPXhhw0tobGpkq9IQwvmfhfuuWrkSN9xwA846axXiYuKYD+PnWNIIbF/1NXX46X/+J+obGxDP8aU5rN+O5LiSHfuGGz+MH/7gLnzkI7dixqxZmD9/PkeSItDKVqiSktn4h3+4A1dcdjlFHi8rsv0MHPaiobHZjUV98lOfxkUXXIQ1a852qlxsXMzp8nnpPERABERABERABERABERABERABP4OgQAvzi+4wP7P9z9MXf3bJXrwCt0V1fydjd9nT23evBn/+I//yHcdhmXLljmHSx1NCr/97b1ITEhEdXWVI3IiNieCGE/+2BRMYWEhr6//AStXrHyDEGMvMaEkuL70pS87UcZeb06cTsaLWBbNrR/9KF8Shm3btrkG5+Dra2pqcNvHbnP5s/nc5tFHHkVLSwsNFzvwwWs+6I7/X//1X/j9738f3ATPPvusE2XsgTv/5V/cxE19fb1rXs5mlq25ab7+9a+7CZnjG50GN0LnmOGnMDIywtGiPvc3jkJKFIN3rcM8Jy+XLUqJqKqu4Ac1jtGRUYwRUHNzE9LqUimQNKCf7hjLhumlMhbPbc3iZMG+lgdjapkpXoPMimlvbUNPTw/mlJbills/4jJtLBvG9pmcnIScnGwM8HVXXHEFlbhGh9yChu18MrMy6MAJdxXeF1xwvqvl7qezprevFxl8bt68uVjKL6ftR0sEREAEREAEREAEREAEREAEROD0J2DCgUVTfPxjH8cOXsTf+S93Ii/XJjl4kTplrDj938TbdIZf+MIXXLaqiSUmkkRFRbkj2+hSSUkJVq1adexMTGZ587KW5S2cVMnIyDj+5CDjQ2xZJVDQyTI2NsbPYrt7/HOf+9yUm4n3LIvGRo7s2D3dPdi1axduvfVW97o777yT+bDDzIlNwCsvv+w0AXsiMzOTzpeHnfhmDpuvfe1ruPHGG925b6Erx45sbc/f/e533fW+bfOxj32MJo9xfOYzn3HPW+uzxaGcLis0jhn7vvMLHxgPuMwXa0fy+wMUY5KcoyUpIcn9wzhj0WKKK8MMBS5HMtuakumsiaTw4mcD0zhzYqwpqezAATz116c4ttSFQgb9xsbGupm1Z2iNMuVrzty5DPptofI1gtSUVGRnZWGwfxBNFHeaKPR0MVfGQoHDONKUmJiAMI5TmfKXQIdNOB8boDvHQorTKcScf+H5TqX70he/hJdfeskJNJPMo4mPiz9dPi+dhwiIgAiIgAiIgAiIgAiIgAiIwN8hQFnGXf9NTI7j7nvuxg3X3eCuHe2Cnc/8nS3fX09ZNus+y1zh9ftXvvKV46JMkMJKTq9YloxbJ2M7dv+66657gygT3HaKtX0OUy+08h8zW9j6yU9+goceeoiTMlMCjj139OhRdHV34cc//vHxXTzzzNS4lIkqwW2DT9p+P/KRj7i7zc3NOEDdwFZRcTF/T6K3t9eJMIcPH3aP2y8bqzLzSFVV5Wklyti5hcYxY+z54U6O89cx6cf+CZhwkszMmEiG/HgJ340d0drUwzDfUma7lM4pRVZ2lhNnTBBhkTWGCa66usbVYl97/XVu9MhquLs6u51AYyHCB8oOUNyp4GjUAFLS0l2LkqvLbu9AQ1ODa32q4xyajSvFU5wxQcZD0Wecx+6gFaqbWTXmsjEVdZTH27ljJ4aGByks+Z3C9MGrrzlW823ItERABERABERABERABERABERABE5fAsw35TVpfFwCCgoLsJkuiu//4Ae4hbki5s54w+Il6/tVqykrK3OTK8Zj0aJFb8ASvLN06VI888wzvLifEljscXMj2fW+PfK3XCd2/W9o3WttI66PcmTpW9/8Juob6nH99dezsdnHyJKznPhjYb42XRNcTU1NLmPW7j/99NMu5yb4nO3ThBmbnJk6i0kn7Ni5WiCxuXfa29rwq1/9yv1Mnz4dl19+uZuisYwbKxo63VbozoifUntXB1qaWqY+EoKzliWnmPG5aFqkMgnM5sYsPMhPYcQCf21syEaNJicmMTw4ihwKNZ103JiYk0iXy1yOF529ei1q62uw8ZWNaGWAbzSVt1aCb6JgszItDQsWLXCcAxyT8jM/xpS4w0yLtrGlwsJCCjLjiPZ43Dk0u3GrfgYMH2HwUCPSeU7nnHcumhhANB6YRGtLM/rpuEmho0dLBERABERABERABERABERABETg9CZg15xhvPYrKCxgyOstbA76upvI+PGPfoTCoiJ88YtfRDrbe90ydeF9usxNYuKJLWtjequVxYkUJ36w6Sq47Lralj1SUFBgN9+0nI7z+ibu+W984xuuVem7//7vLrJkdHTMhf3aONKd//zPuJq5Mb/73W/d6JI1MQVPrry8AuUV5cfvv+lg/AxracSwZaKMjVZZAPDWrVvdY5aZY1k0P+dPMt/nvffei6uuuso9d7r8Cp0www+hqaGJ6uRm14I0Z85sFwA8QbHDiJrKNcEfE9tGOW/W0d7mrEs5udnIZi6Mhfr46VgpLp6BrKxstHd2YM/ePW4UqYDiii/S5xS2Bo4z5TGzJjcnh5k04c6yFBXJuTgKO+aI6ezq5GjUGGuyUtxYU8A/TtFn3Nm0MjibtoUflrlrsum8sSAgm6mzwGH7h2pn6qWaZtYqLREQAREQAREQAREQAREQAREQgdOfwJTWMuliMDa/uhlzS+ccv/783v/5nguPvev7/4GFi85wb8auSanlvO+WZbUEl5khTnSsBB9vb2/nTTpU7OL42HKOFd62h/6m++TY650xI7gh/1puzB133OFamNatW8f2pvWckKl2+3rssUfpgul2Yk3SCTmvX/jiF7CGWTQnL9t30JFTyszZ4DIzhokz5eXlzm3zFGu1LcPGrve7GDh8DQWgF9k4tXbt2uAm7/jf0Akz/GJbnovZiwboWImj28Vg7WSYT0tbC5OVe918l/uQicGymgPMljER5doPXYtZM2dRWAmgrrbeZc9YiLApaevWPcP5sny2JS2ik2UAY4FxBghPOlUtJTkFjz/+OPYfKMMHP3g1BRWfG3cy4WUOR6VSmGFjbU7tHV1OKbPedPtgTARaumQJ4mPj+Vyrc/XEs37bvqjRUXTxMBtHSwREQAREQAREQAREQAREQARE4N1AwBwzHhdh0dzajPkLFvECfmIqCJbXqc+zmvmc83bgjw88iIsvuWhqqsO9LVMT3j8KzSy2FwfXbpbrLF+xInj3+N/du3e720EBxO7Y7b9H6uTX2jbWzmTTMpYlY41Z5lgJulb279+Pm266yUWUmIDS3d2NmTNmus/LtktjwPA111xju3nDsskbCwvOoUkjmEHTx6KgQ4cOOVFu3rx5+PznP+9+LFf2pz/9qcvSse/CE088cVoJM6EJ/zVc/KRmzZiFM89c5lwnAY4qmfOklUG9zzK49/HHHned5BUVlQgE/IjwRCDMY4oXt+W/haLCYhTkF+JV1nf98cEH0dba7kQVsyE98/RT+Pkvfu6sSZMUZurr69y4UlR0FAaHB/j4a86t09rc6sSYVLNlEb4JLaWz57Caexc2bXyVo1PRVPisLnvEfcCx8TEMAh6G/eNta+twt82CZTY4LREQAREQAREQAREQAREQAREQgdOfgI3aeGgSsBzRSV7YJzFnNJiRYi2/a9euwQDbeG+86UY0NTWjjyYApzQERRm7Jn0fLBM0Fi9e7K6/77rrh87UcOLb3rt3L/76179OUTnBUnR8lOlvaFhulOyk5yynppjBvJdeehnuu+++Ew+D+fPnw1qg7DOw5iQrDzLtwFqh7LC/+PnPMUBTxsnrK1/9ClafvZqBv0VTOTh8wfnnn8/a7hW4+uqrqDMEjm8Sw9iUf+a4lLU12bJIlNNphVRxSM1IowJ5CUN+fdi+fRfWP7veBeqOU32pb6h1diKrzBoaHKaKucCFBVtIz/g4/TPHvFIT/GDK2drUxsfPodVo/vyFrLUu4OxYJv8xDfCzm8C+vfsZ/lvJcaVU5E8vcAHDQ3TpDI4MISYuFtZX3t3d59wxw3zMjt/T2+VChq2pqbenD09R7NnP/ZhIMzgwhHaOVtkXYpBfgCk98HT62HQuIiACIiACIiACIiACIiACIiACb0mAF/OTk+G8pmvndWM6pyx4gR424ZwxXpbALFm8FCWzZ2OcMRfldHB8/OOfwKHDh17f1YlzO68/+p689e1vf9sJIpUs01mzZg02btzoGqz++Mc/OmeLXZufrFMdd8Sc/MQJhCaPhQUHR5nOOfcc52Kxg9kxX331VSfC2CYWQvz/svfdgVGW2dc3vRdIgBBa6L33JkVARMSCuiqo2EBXBQsougq6dkUBO4oIIiACoiC99947gQRIIwkkkN7znXOHQSy7a0kUft99IMnMO2955szz/nHPnHvO559/rkfTTNhpKMy4a54nBv6vVMwwfYmDLUnTpk2TD9//QOdePaK6mvvytX79+ul8jx+PllGjRimPwO0k39555x21PuFzmgFfTqNUiBn9fHAzZGdmODlHiYuPkSmTpwDEPNwILaR585ZSDzcDPyjGmFGqxL/0fNkElUzK2bPiBTaTRr1SBONgsp4wDy4Hc9/69esL+8yqRVTDzYYobnjR5EDG5I8o7VatW0nr1q1B5hxTA18lenAOOM5AAZOp8d3eIIqoiGE0GD1l8gvyYCS8XsaOe1dmzPgaRsCHpQDtT3nZucqyFhWw0cqGIWAIGAKGgCFgCBgChoAhYAgYAobA5Y4AuzDQyYRCPh51Zwt0XySilQbcDGpPN4TAbN22RVjMUwzA2vW7uXOl9zXXyIplyx1v7QKpcLm/z5KYHwmKZ599Vuv2rajFr4KaiClGNM+lUIHhPByXesk4iRmU/L86HK8zOYkEmYO98fP1kzfeeEP3Z0tTp06d1ai3UqVKwpajw2g/or/rKBgEOweTmu688w59unz5cp1Xw4YNJTQ0RAYOHKjtUVTCkKRxqngeeeQRqQVyh4zNG2+8Di/Ziqq8CQ0JVX8bnqxp06aa0KQnvkx+lQox4/yAMjNzVP3S7/p+cm3va2HqG6Y3A8G/7rq++sF0aN9BfEGoFILFzIO/TCzYsL179yHmOl5Cy4VKICLO2KLEtiP0FEkRGDuC7uPlIwEBAWDPTuGcrsrqbYX6Bt2E6gnDeK19+w7A1Ge1xJyKkXp16+EDDNUFxYSnc8g1XwgToPpobRry0EMqd+LrbJWKioqWLKQ1fTfvO5k+bbqmMl0mn5dNwxAwBAwBQ8AQMAQMAUPAEDAEDAFD4L8gQEKAdSHrS7bQMJ2HwTNO9UYa2pgYHMPhIA5c5BTq0GuuvUbGjx/Pgx2v6Q6Ox/+Xf1OZMmv2bLkeapNwdJv4w5eVsdIzv5mpBAjfe2Ag2sEuDCdJQ/7KSYg4X+NfbmONzuHEnI9p+jtx4kSJqB6BZ8XoaklBK1m81ugNQc4sX7FCI83x4sVB0mXy5MnqI8PWpIMHD6KtKVNf79Gzp6xZs0baI7zHOZgutWzpUnjO3qxzKIKlCX1t8iG84GsPPvigbEJ8+qXvx3ns3/m39Mx/AXR2diaULDnSFG1KHcM6qBnv+HHjZeeOnZpVXh0sZUSNCMmFMiX1/DnNJk9PS4dxTyWkLnmCsPHHovDThVG5cmUYAteWg5A5sb+sGOxmYSF7xqCmwQ9NfOfPmwfmq5/Ug8lwDlqS4uLiJQ4R2GeR6DRg4J26Zz78bGj6e+LESVXP+Pr7SvsO7aHeqQ+VzhmJguQp4XS8BIL0ORUTK+vXr5d7771X/Wn+zg/Krm0IGAKGgCFgCBgChoAhYAgYAoaAIfC/EXCj8S/IAYo1duzY4SAHEBjjin/0nXEFaUBSgaRMMba7oLasWDEMIoBieerJJ1VhwYQiTSm6QNL876temXuwi4ReM/3799efn7+L2bNm66bgSwJxyiBUx6mE+fn+fE4VSxE7X35l3H///eonQ0KGoggqmOgx4wexxn8a99xzj/CHCh62PfH89Kv5TxHfERERMmfOHPWhJZHDQCK2R3H75TpKiZjBHYCVnoC0o8NHj0h45XCNri6HTPG4uBiZ/8MCOQMSZMjgB+EvkyXu6POjl0zK2RTEW6fIYLBYDtKlWNzwmg/ajQIDg2AGHKHMFnsF90NV44lYbHrEiCuuhxvGxQU3IB5UqFAOfjTd0EuYr/1n58+nomXJR3z8fMUL8igqY44cOYJ2ptOSDbNfOj0zsSkYLGCnTldhW5Z6y5w6eQr0EnoRL9dPz+ZlCBgChoAhYAgYAoaAIWAIGAKGgCHwEwRc3V3lGrQmfTN7lhTlFzgIGFR2LqgbCxEeExYWRlaGog3+Z06MRFStJj169ZRXXnlF9u/fL08++ZR88vHH2tpDEuJS5cdPLnaFP+nX7wY5efKEKkk+/fTTn7wbpiV99/13uo12ISU1SMYwRcmZpPRbz8s6/vdEXJPsKcl5/9Z5/pH9SoWYwbLFwhXJgylPApiwqGPHoJLZJTTejUDaUrOmzcBU5iNBKQeqlXQJCgzGT6Dk5Oeq+W4YWp4C/APkGI5jT1hKSiq7mFQF4+XlzbPLkcgjElahorRr31GOQZrk7+uP1iZ/TXZS+hMTCIAEq2OHDlDfUCETI2vXrxNPd0+NxnYDQXMG5yXrRvNf3pLwh4IXTqz2uvnhXPUb1INqx1cCLjg3/xGA7RhDwBAwBAwBQ8AQMAQMAUPAEDAEDIG/DgHWixM++RQBM43lpRf/rXUpW5mKUVSyHaZ5sxZy+MghWJliG1kZ1IFl0OZyYP8B+Jb6K1lz6NBBJPx0lwmffiaDoNaABgAlI1KE8Ue/uP8/8u19q1Yt5eSpEzJp0iRVlfRGeA+Ti1auXCljxoyR/LwCVajQc8ZG6SFQKsSMrlGsWJr0VoSfTBHkYYy03rx5M3xl3kS7UR9ZvXqNbNqwUZgzzg/eB3Kk3Jw8VbJM+HiC9O3XV1OTGJNVLjREJU6xOEezZs3QiuQOkiYXEdtuUrlyRTkWeVRCQsvC2Kk5FDUBaKHKBtlyXsIgR4OGTclQGgvv27NXlixZKp07d4RqppzUrV1Hz5sHQoisHRfd7FmzQBIFSfNWzTU/3cfHB/PwKL1PwM5sCBgChoAhYAgYAoaAIWAIGAKGgCFQYggwcvn1N14Teqd0wBf5ffr0BqnCViYQK/j71bSp2qaEh44BbsYDX+YzBKZcuXLcS9kXkjYPPPiAMEl49IsvSgV0gFxCzZTYfP/OE9GQdxVImJRzqfLM0087DHIvkE7grcQLwTlzYY7MWttG6SFQKsSMThefYouWLWRQ7t1CB2UaCLHtyA8KlHbt2iOWLFImTPhUVTAhZUO0/ysUcddsW9q0ZbOSKpSf5eVARZOTr32A2SBj6iPJqXnzZpILI6eFaIlijDaHt7evVKpUWQJB8jCTfN2a1RIOUogkDdOfeAOyPcrVjX40aG2C8oYZ5+wbZOx2KhZimeAycjL6hOw7eED27N2DufigfaqaGg7rReyXIWAIGAKGgCFgCBgChoAhYAgYAobA5Y0AuicYv7x79y6tOffgC3p+AX8UNShHTEwMfEjPqvKlGLYW9JthEi+7OnxQJ7KVgqqbwKBgCQ4Okk8++QRWHGdl3LtjUWOGX97v/XfOjvXwzl275O2335bFMEomNq5u7lpz01T31ltvRf3e7nee1Xb/vQiUDjFD5hE3Q9LpJCQq+SgpQo+Yq6/uIcFY3CRV3Nw92DwEL5dMyc3P0zzxtu3aavpSXKzDfNcdrGU+yBS2QNHgJxgtT/v27Ze2bdpIxQphuu/JkydVTuaD5CaOFLQmZUCFczTyGBybD8t2GA23a9tWOnToqAlRbGXS/kDI2FxcXZQBPI9Mc7o5kzxq1ba1eGHO5+AQnZmN64JIys7J1nPbL0PAEDAEDAFDwBAwBAwBQ8AQMAQMgcsbAapaXGEAzG6JvkgDHv3SizLs8SeQ0LRQpk2fpgbAWfAVZXcEaBklYdJRQ3qxpkS3B31GOWrXqi133HmHPPnEE7Jn12758KMP1ZbjlttuVVLn8kbht8+uGsQIH3zwgR5AtRFFDc7kpd9+FtvzzyBQOsTMBenT9h3bNf6qVcuWsh9uyNkw+vX08oQCJkfz0Js0biKJSYni5+OrbtiVK1WRylC9MGGJzKWXhyekZ0xz8lCXZpIv83/4AeZEJ2Xo0KHChKXcPLQhgQTyhsQKf/T1HJw/DXHYnp5e6nGzd88eqV2nji6wahEwdepxNRygY9T4l0xoMdhRxnCnp2WAwOmA6Oz2shusYSLUON5Q8DhJnz8DtB1rCBgChoAhYAgYAoaAIWAIGAKGgCFQ+ghoixJqQ6YHRVSPQOjMYPkYRr639L9F5syeI7fddpvjy3dHNJN+cR8TGyMdOnaUbdu2YYIkdlzVJmPN2rWwzUB7E77UZ/Ty7XfcLsNR55KsqVChAvbFhfD//8qgxYeNvx4BR7h4KVyXohn26TH1qGXLVmrC6wtX5OjoaIlFRjwJkXLly0mrVq2kSdNGqqLJpdcLbgD6zQTjJnKDhOofMBl6+KFHsE8zJUnCYQzMuGwyeNkZ2RKDc52C3KoQzCa3MQlqw4YNEg3y5nz6eWndqrVUhcM2CRaSLyFol+rT5zqNBNuxc4csWbpEzqaekQB409D0yRXsDo2hAmBGzAjtNlDb+MFY2IYhYAgYAoaAIWAIGAKGgCFgCBgChsAVgACKUdZ1DJJhKk8+WpS+++47+eDDD6Rjp46yY9d29S5l/QcOBv+LYYB7Cv4y5dHJgWAYpsLgeAbGxMaekjp1auubZtw2E33Z9nPngIGyb/8+bYPii6x/bRgCfxSB0iFmsCpJGjZv3kIialSXqBPRMPk9r8lKUceiZN68eRKfEK+mv1S3uHt4gQwpVLfsbDzPzEgX5qnzJvAFodKseRN4y9RV5cpdAwepW/SypcvR5pQvNCRishIToOrVrYeYr8FQ2bTHneGIQmsOn5vrb+gndWrVlIKiAjUYJgtIEyPmpn///fdy/Hi03HTzTdK9WzeQOzD6pTO3jiI1iHJzN9bwAiD2xxAwBAwBQ8AQMAQMAUPAEDAEDIHLGgEXN9hWoCD19PaUDes2qHEt7Sx2oSuiV69e8FGJk8VLFqN2fEAJGNZ/TGv6/POJqC2TsQ11IPbntvy8IlXeFEMIwBhuCgwqV66M1KIV0v/m/giMOadYOEUzbAOyYQj8XgRKqZUJixGyMLpWV6oYLnO/nyuJCaeF5r3NmjYVt2I3jcWmKubYsUh4wRyEGfBRYQSXBxKQ8vLy0Q+4BH4y++RVOGk3gHmvG3oEPUGmNKpTF8ZLSfLmG2+p10ynTh3Ua8bD010VOn2u7Y3rhsLMKUW8vL0kwNdPDYd9oXrJzc2T04mntQWKyU4BAQHaJlWEPro2rdtIVsMs2bp1K66fByPgYDl2/LgEwRCYN7UNQ8AQMAQMAUPAEDAEDAFDwBAwBAyBKwABkCOu8BQlR5J6/pw0atxQchAKw/ak7du3a/z1Y489JmPeGiPn8Lo3Ql8yMzPV91TfHY4jAXMer/n5+ajqgKoaWs94ou688cYbZdWqlZKcnIxujfWyC50Yjz06FJHbqB3xz4Yh8HsRKBXFDG8AtuuxZYl9eGeSzyAn/qjGWpdFPnyTxo0lMCAI7GO+FOQXyOmEBESWfSV74AUTFxcn27dtVeMlesnEx8XL8aho8QApQ58Ypil5w6eGN44rmNB28IS5ukdP9ZPZtWuPnECqEu+cMiFlpTyIofJh5SXqeJRMmfIl1DLZuDlxN4Ht9IB/TV0obB4a8pB0Qi+hK1Q07mid4iiAGXFIuVDZuXOnrF61WtLTM3W7/TIEDAFDwBAwBAwBQ8AQMAQMAUPAELi8EaA6xgVf7OfDj5RfstOPlDUqDYGphGGnxL+e/5cMvGugxmNv3rRRaqDDgsQNa0mUi9xZIo8dl5YtWsK/lKoYF60/XdHexHYndocUX/CoefHFl+SeewfJgQMHlMS5vNGx2V2OCJQKMUOOkOa98+fPl/Xr10vPnj2lRo0aWMou2ipUPixM/Pz90LKUBYbyPKKxw6VO7Tra+5cGN+yVK1aj9ai2vD/+PShlXCQHhIoHSJgA/wC0PblLwwZN5C7cRHwuhSJBAf7wo3EFsbNb1q5bK8kw7WU7U42aNdSnhp40Cxcv1ASo8uUryJp163CDwjAYzGgdtEhd1bUrJlysUjXeZBnwrsnLzZfUs6ly6NBBEDPoM7RhCBgChoAhYAgYAoaAIWAIGAKGgCFwZSCAojQpMUkqIVwmFx4zbDEiOUMuxQ91ZKeOndR3ZtOmTVr/rV29RgbeOQD1Km0sihEwIxook3A6AZ0U0bpN1TD4nj8D9SFrRH7pr61LOPF82HX0uqaXdn4oQCR3+MM/vLANQ+C/IFAqrUwMHCuEn0sBboAU5L3fc889El65kixAotImsJFcwN5IWCoqyoOKJVdqgkC5DZFjB0GC5IHVpMqF8dgdOneS9WvXSVR0lLpj14ViplGjxtqC1LZNW9mzb4/ExMeqJ00AFDjMnU9BzLUXfGkYvU3jYbYp0UQ4FAoYtid5IalpB+RrXbt1lQfuH4Q4bH9xBxtKoqYQPjfnzqVAtRMPE+F4KYd5sIuQ/20YAoaAIWAIGAKGgCFgCBgChoAhYAhc/giwtnPFF/xnUs5KLShhqGTx55f5rgiMQW14GHVnBXxhzzqPnjLDhj0uTyBl6aMJn0g7fMGfmY2OCZwDhav6o9Jrxr9mTaF3TSFq2cKiQilEsAzrXrY48W852GmknDkr1/XpI8tXrNR61Ad1Kcdf3d6UmpqqAonIyKMgnTLVo7V58+YIt6n3Xz88Wons2LFDA3vCw8NRezeSjuguKclx7tw5WQehRGRkpHbB1KpVSz2AftPc0NESHRWlQT4NGzaUTp06leTUfnKuY8eOyaTPP9ekrr59+/7ktdJ4UirEDHkMNxc3LNgiLF7IxUB8eKN1qFxoeTCJ82X2nG9lAPxkatepJVWrVQVJ4yPVkJy0efNmycrMlmvANFaEiiYfLUX+UNZwMR06fEi+nfuthOEGuvGmm1SBEwSm8zS8a7JAwIRVqChl4SmzdesWJXsqhYWLX4CfFICYccccmiGaO6J6DTUhPhp5TNLPpUvlDlUdLCZJGexHBc9ZqGR43bNnc5DmVEXC4eTtD/LGhiFgCBgChoAhYAgYAoaAIWAIGAKGwJWBgKbtgjyZM2eO5MBrtCMDYlCXUr2SA+9Tki1gVDBcYQZ8Sp566inZjhjsKZOnyNkzKfLee++pNQf9R52DKhp+8e8LH9P0tHQcD/mMqmGKpUWzFlIWdhozZsyQWbO+kdn4GQrCp07t2qrSYWiNJteQDSrFMX78eHnppZcgOEh1TE2vBWoIl7333ntlzJgxamZ86RTOo4vl0UcfVXsRzpNEk2OuIl27dIUp8udaf196zB957JwbiaMfh2NugwYNknfeeedX50Y/oKlTpypyl2qPuqElbeLEiSUytx/nA6UUunj69eunHMQj/3xE/gpiplRamYgY/V/4BhrUry/n4VRNZpEfMaOoszIzJPlsMsyRQqBaeUDqY5+9+/ZqK1NBQZ6ylzQGzgHhUr16dalbrw5Ml/ylKL9QsmDadAIpTz4gc9p36Khx1ozX5o+Hp6ckgKhZDRna519MUiaO/YSUo7F1qm7dOtKwYSOcN0s2bN6AG2Y2jJp2qa8Nb8zs7Cwoc7aoJw2NhnOREOXv768/l35Q9tgQMAQMAUPAEDAEDAFDwBAwBAwBQ+DyR4DKh7jYGJAjyowoIUPhgCpdUJsWo2uCipayISGyYvly6QabC1cwMGvXrJMb+vZDLQktA+pbRzuSqx7v5+enQTOsbZWXwQ4BgQHqV8MwG0Zqz/1uLurh62TV6tU83HEOPVHpYUYC5fHHH5dzID5CQkLl7rvvlkceeUQaNWyg73fSpEly5513/mIC9F396quvWK5LMLpM7rjjdhVOELLVmP/111+PWjn7F8f9ng2fTfxM50ZSJjTUMbd//vOfqM8bEFz54osvZMCAAb845ZAhQ2TqV1N1e5DO7Q6JiIjQ56tWrSqRuV16URJxxOjQoUNOburSl0vtcekQMzrdYqkeUV1lSevWb5DFCxfLyZMnoGwpj1chLcM/LlASKkFBgWq0m5mZpclJR5DQdA5kDtUuvn6+UNqEwizYHzeTi5Iox45FQXJ2FoZOrlK5UiUJBHnCdiia+F6LVKYQ7M+o7QP7DyjjFhQUrEqcDMRwF6HFii1Px6CamYbF9xkYtolYJJNA5OwESUP1TVp6uvj6+Mrp04kSGxeLmy+/1D4AO7EhYAgYAoaAIWAIGAKGgCFgCBgChkAJIgBCYeCAuyQQZAlKSAd5Ap0AVTSsRSvDZoNCAj4m4ULfGdaud999j6xbvw5ql1mydu1qmQKVxmtICfbyhA0H9mP9ybF2zRrUsEFazyqbgW30Rt2ze4+qZhitzYqXCcS9evWQTz/9THfj1S8cpNfl05Ia9FUdOnSonr9Fy5ZKLEyZMkU++OAD2Y2QnYcfelgvtXjxElUROa/LCPGvv/kaT13khhtukHgE80yfPl2i0DL03vvv624HDx5S1YzzmN/7l3N7HOohjpaXzO3DDz/UAKCH/8m5uciiRYt+Mrfdu3fLzJkzgZ2LJmElXJjbcaQnv39xbgf/1Nx0Uhd+sYOHLV8LFy7QLaocunSHUnxcisSMC0iWXF3ABYX56hPDFqLbb79Dnn12pFSLqKoLm/1/RVi4xYVkGx358Xth4rsUcdmbt2xB31kO3r7DA4aUFSPLDqOtafnypTBcOq+eMVS30NOmWrVqkKANl2eeHoG+s0qqknEDw0nD4GwYCJ84eQq9bMf0HP5gOamwOX36NKLOVsNo+H2ZgJ5CRp6lnD0jMadOgRTyAzOYp/2DpfgZ2KkNAUPAEDAEDAFDwBAwBAwBQ8AQMARKCgEwIE8+9aSMf+99eMzUxllRazLrWokZQUdEgHh5MKmpSBOYXEC6hEIxsxceK2XLlpOzKany9NNPa906YsRwWbFiOb64R2w2z4FxFP4o83+Y5yB1KMTR87tIZlYmOjt8cU76zxRrhwhFAg8//JCMGz9WUtFe5BxU6ZTk2LBhA0QGWUoAsQ2LqhTncIO1x7hx4y62CdHw2DkYH65iIsx35MiRCMhx+OJQVcT2pprw1uFJt27d6jzkd/+9ODcc+WtzGzt2HJQ6Qfr5bNy48eL5t23bduHxf5hbDc5NZAt4gz8zSM7RY4ieNSShSKrpcADzZ079m48tFWJG1yamwIXHiGq2A4VgoQcFBIC1DISypbO2FHmCNAlFH56Pr4+4ujum4sK8efzk5uYoaZKZcR49gJBNwbOGKUxVqlTVxUYCJQ9x21TUVKxYEXn0O+SHHxbAG+aMfqC8AXhNpjWxZ44mQzlQ5JCtozKnNlKgBkCiNGjQPdIA8ikv+NPEx8erRCseShnm0pMt69SpHQgcj98MqO1oCBgChoAhYAgYAoaAIWAIGAKGgCHw9yIwYsQI6dS+A2rEH2Bg20m/bHdTYqZYvoH/C31lOEiPwBlVPFHzUSVSMawctsB+IztXPvnkE+nd+1okO1VSFUXNmrVQX7rxMCQBo+7Uo1H94iQFBYV6DkeqE86AYr9xg4ZyD1Q4FCIsWLBQPnz/AyHxoPVyyfIyKjBo3Lix1satW7fWOV76yxOihHDUzXxvJ0+evPgSW684H07H0a518SUVWrCW5mB7FgdbfaisYQLVTfB+ZX196aBFyMCBA+Ebe436tHB/1u6NODekM7dq1erS3fWxzi0ccwNmMTExF1/n3JzDeX3n8xzYjpxPc8zNOW/n3HjtG2+8EUIO+ABdMvLROcN2KefcKCTh4PbxIK7oOxteMUwWL1msXACBoZDkrxilQsw4Jl6MyGmkLoE8KRNcRho3aQxn6nYw+62C3ruymqzEHHn255UJLqvtTKGhIcpUBgYE6uJQqRlIGsrGaKxUrlw5efOtN+WBBx+EIqay5LPFCC8x4SkGapgvp0yWl5Ahv2L5CjCaXnrT0C07OytHW5IOHTmssWZlywQrGVQVJM+tt94qz458TgbjnDfc0E/q1qmrsrQzSJOqDYfoHj16OWK5/4pPw65hCBgChoAhYAgYAoaAIWAIGAKGgCHwpxFgS1JHpPwmJyfJ+nVrpUuXLprWyxMzcGbjhbRgbTtC8U1D4GKQD55ePhdqUdEv+s+cOQOvmOuR7HRIbrnlFlWVeHl5OuaHGlUHatLkpNNQl9RSgqYI3SBkOkKR1MTUJw94npI8OHHipFx11VUyY/oMHIZ9+MM/JTBIhuzdu1fFBpcSGs5TJyYmSiT8djhI4DjHtddeq2E9nAZbiy4lQOg7w/fPcd111+lfkihUlixbuky+/+47IQF26Xjuuedk2rTpsnTZUunevTvIKk8lavZxbmhF4vOfj8SkJLR9HdfNTIJyjj5IuHLyIr82t7O/MjemSC3l3OZ9r6on57n494UXXtA2Lb5O42AvdN44Bz2G/v3vfyMp+pD07NnTuRmf24XP+OKW0nlQasQMGaxtO7bLyRMn0LIkeNPeaiDUAOa7jZo0gWKmoXiBPNmBfZgBzxhrf78AJCdFSOdOHZWt4uJh5DWtmbhACEr58uXRl9ZCPWXiYuPw4cZrIhNBPXc+TeVn7Dnzg0rHHQzbMcjM/Px9pHy5CojFdtMbrmbtWlIFiUsFhQWYXzRuPk9I3GrJtb37yHBI1R4aPFhv3DNQ35BF+7XFUzofh53VEDAEDAFDwBAwBAwBQ8AQMAQMAUPgzyLA7/aTUfBfh0CasePHyZCHhsgtt94iEUgFvkiIwLuELUcUAVDVwe4O6iNYf1IpUaVKFXiYXgu/loMy8tlnZPPGTfLKK6/A2HeeMAobQhiKPHTsQJRzQ5AKWZmZOLoQ29gQ4yaHjx6RiBrVlaihrw1TnQbeNVCGw4JDVTe8oAoRHOcprd8kTKgo4eVISjhHhQoVlMBga9W06dM04ejxYcOkbdu28sADD+huJFguTSZighXPQQ8WqorWwHOHg2a8TFYivtf16atmv/rC//j13LPP6tw4OV7LOVj7P4P2Kn4WOrfq1fWcFHw8CGEFr8950JzYOYYPH45tXTkF+XTChJ/M7e233+YlQDL10dYl5zGs99k9Q+KG3kE/GU5m6CcbS/5JqRAz7NGjUuV04mmJhJEvQWFPng+ULe5oSaISJhxx2J4gY954623Zh0Sm6uhdCw8PUzKF/Ww06D2KY/fu3ydHjkaq2W8+ZFFsdXKweEWSB8VMOqRTNfEBXQs2rXuP7lIJ3jKFatbroooZsmihIaHy6GOP4Ka8TioiVpsL3xvXprnwYrBlUcePSWXcdG5uWI6YW926daQzmEyygLO/nfMLeVbJfwx2RkPAEDAEDAFDwBAwBAwBQ8AQMAQMgRJBAPWeG2pKKkfcUH/+C6QE24gaQBwwa8630qZNa5aoqP0cqcEkJY7hy/2rr75aa0wSLqw7mey7YcNGqVq1qmSkpcmGLZvkesQot4GB7S4Y0zZt3hRHstSHCgeCgrFj35XUlBTwPA7b2CIkPuWiJaosEocdZjbYF/NiS8+7Y8eCZBgK+49zvJgaEPM8nFdJD/q6fDFpkp52AJQ1nTt3/sklXn/9dfn3y//Wi7PNaTz237oV/i6YLhUoy5Yt044X50Gs12ksTF8Yvn+SJKehhmEKFFVHbP36AuE6v2VcOreBAwaqGufS415/DXODkoV2LyfRfsbI7W30lMHcOnRoD+/Z5T9JUXbOjQlOVClxbjQNvueeu5VHCMfcJk+efOkl9PEvxRi4gP7Hr79glAoxw4VFEuUcTJOSk8/IefR2sRfPG+QMTXjZf7dtG5QyULhw4SYkJErrlq2kU+cu4ombh6ZLnu6eSuSk4wY4fToBsq9oOXr4qCxZskgXr6enj344JHnooh0A/5q6tWpLHZAqNPWlTOsUFtXZsylqMtygfgNlMGtBLcObJrhMWfFD/nwOYr8yMrKgqGFalKt89tlnkoR+wdo1a+NGdpe4mFi8/tPetL/gc7FLGAKGgCFgCBgChoAhYAgYAoaAIWAI/BEEUCPyC/dQWGE8BONd1p/r129UcoYamS1btkq/6/uhTnTXL+1JhqSidk2AaiIJXRtkKHC4ev2mpp6VmuiuKFb9i6v8MH++XIMk4OjoaJn19UypXYd1o8MLJRFepayDeT5ePxtdJPQydXeHjwuKVjeQPVTaDBx4F5KHy8nWbTvk2zlz5Pvvv79owOugAUqOnqGihRHapIqqVYtQ811M7+Jwpjm9CEsQjnr16kn//v1FfWowjY0bN8Cv5SbUyEkXj+EDqok+/vhjnLcYATuRUr9BAyQjx2rtzRYo2pD8rzEBihbH3ESDfEjSXDo4t2FQ74wePVrnT9sRzq0VPXR0bhvl5ptv/tW5faJzE51bA8wtJiZOW9k4t0uNkS+93i8el9zH8ItT/3xD6RAzuEohFn8aWpTYqvTphE91sR05fBjGu5nqDbMYBAudrJmRXqVKZbBXhXCAdvR4LVy0UELKhUi9uvVl/4ED2m9Gd2hXKFp279ojiUhSqgcCpnpEdWUjaeyrGfK4LrFzQVRZclKybMe133tvvMaDkTkjEcOWJ7Yw+YP9DA4OlFOnTqqRE6O52D/Hn7SMNLnx5hvlxdEvqTHQpb1nPwfQnhsChoAhYAgYAoaAIWAIGAKGgCFgCFxeCJAYCQoMEtpj8Mt5qlfOppyBh2gPmThxosyYMV1eeulFJU60nQmV5JdfTtWWJkoleDzThWncGwCbjGLUq2xFojCALUGD7r1XPocKhe1LX037CrUsUpvAqpBYgVBDxyEk/NSvXx9Gwlk4exFeLAZJ46Y1aCfYd9Cqg8azN/e/WUa9OBrpwM7UJp6F448zA1SL/Otf/0Ii1MOqHGFE+MqVKy4mMznOL0g1fkre/+B9dLwUaBvSwW4mRWoAAEAASURBVIMHZfbs2VDMbJHFixdrgtX8+fO0Lnaa7DqPvf3229VMl2+aNTn/vvD889K1a1fnLr/6l+d5HnN76KGHdG4keVauXPmLubEtiWQNO2beeWeMRoA75rb1wtz8lWfo3bu3nufSi12cGzbq3IDlv/713P+c24/n+OPY/3iO3/6odIgZvAdPqE2oYqHBb1xcrLpQT/z8c/kOBkGbN25W75bcnDxp0rSZsnK7du2GHGyXEiM52Tm64MPgiLwcsqmZYCJbtm4p7dt1UGflvPw88UEaU1hYBW15mvDJBF049LUhA8nobXAzOuLjEyQe13d385DAoEDJRsJTArbxBmDX3/m0dNm5Y6e8+tqrMnXaVGGPnbubu+TDuDi8crjKvELQCmXDEDAEDAFDwBAwBAwBQ8AQMAQMAUPgSkCAihdXfGEPxcSxSGlQr74W7q5oazp37ryqMIZBRTJg4ABtnWmGmpSDMddMFVI+BMeePZOCL/OD1f+UBSa/7HdBfdsBaU+tWrWU1157TbYhRrpf336yY/s2qVOnjl6XogNe/1TMSQlBwE1CXIJe3w0KHSpXqD6hrQYJIRJARbABeeXll+Wuu+9SA19e30EWoaj9A/wAiSOqcjg/kiVUjDCyukaNGj/58JhCxJYkXoPGxk8++aTOx7kT04v0HNiwe89uuTRm27kPY6Z5/IXyG5gOdL70q385t7vuugv1949zW79+/a/ObfJkzA2Dqphfm9sbb7yhr+/atUvTrvTJJb90bpiYs92MrVK/dfBz4b+fk1G/9fjfu1/pEDN4855QpjSo10D6Xt9XDZMCA/zVyJemOps2b9Z2pAoVykswzHXOnT8nKYjW5oI9deqUqmDmwUV5x87tKgU7j5YnnmsATJKio6PAep5WRU5mRqYUYtGnpKbI4UOHJeXMWTCgPaX/Lf3lKrhuM90J3ItKlrhaPCAhY1LUDhAxZNoiI49KIyxSF9xgp6HCiTwSqSwd9+O1d+3aKUxnYo+gDUPAEDAEDAFDwBAwBAwBQ8AQMAQMgSsBAZAoqEmZjpSbk6s2Fpr0i5oQm8UXX/JPhs9IH4S/eHp4IrVpPRQhvS6yC1S8QBsjJ2NPQWHRRVJRb7rQPwPbSZiwfqWPKb1VmbTUrXs31KSpsmrlKrkbXiZK4GA/+q5O+nySJMFYmBWlEjE4hjWsdnwwvenCCAkpK+wc6YFEoKioaAcpwNdorvI7BuOtSajMgJEvR9cuXZWUoU/OzwejqfMuREZTSfRrg9HYHHz/h9EBc+mgkuXRRx/VTTpL/KLPDAmfXxvOudHIl6Nb125oMVuvHj4/35+8QG5ejm7m+/m14ZgbP1H5xdw4B50b5uREkJ/Nf5rbz89PDqD4d2L/83P8nuelQsxwwXExlkU7UudOndVwp2mzZhIQGKBO1/RuoSmQr6+vRmnv37dfe+vKQ61CF2S2Ih08cFC+mztPGjVsIE2bNZFzWOjsycvPL5DzqenaDnUqNka8QQA1bdoUN1sZBTwABFD3q7vLq6++Ks88O1KqVKqi6hx62/AudBgxFcsZ3BxkKv8B+dVTTz2psdnsHdT+QFw/KRGRXZFRchwMq8Zy/x5UbV9DwBAwBAwBQ8AQMAQMAUPAEDAEDIG/DQF++c4akLYUED8oIUPfUxbpd9x+h/qMHEDbztKlSzVaed738+S5Z/+F7SHYpwjEjguMe3OgxNgkxyOjtUhnKxNpAHZqJOCLfTQn6bk3b9msLT0zZ30jH334kdx3333qr8o3r7YbIGIgvVBDYnrQkAzicydhwP26XNVFWrZopUlSU76cLGPffVcy0zPwCotY7vG/Bz1ZmCK1evVq9WElSbJkyRJV/fza0f5s0brwQhq8XX9tOIJ36OuKAB10p1w6aMq7GaILTpHR1vzL50yu+vmgpYlzbrwo58ZWqTI0Rv6Vwe4b5+QcrUi/3EnVTdwJ1/WBn+2lg3yAzg0b+/S5Vl/a8h/mdulxFx/jnM7rX9xWig9KhZih5Ieqk4iqEWC5chGD7acLnOCWCy2PeOvymnSUBB+YVChlaIoUA6+X8PCK2oNHg6TWbdpIVUSZ+cIXJjw8XA5D3bJj5w5l/ypUrKByLy48mgXffPNN0hzkDD+QA/CkiY+LV0VN1cpVJbxSOHr1zsrXX8+Q+TBqioGZL29Mx93JBzAuAnvY5arOmtrED5SsIaVlnHs2bsbfyqrpyeyXIWAIGAKGgCFgCBgChoAhYAgYAobA34oAaz5+Ed+4URMEvsDj5QIR4oov+2Pj4rT+Y+VNkmXwkCEay/zIo4/I51C40Li2mJIbjBUrVkjimUQVrrC9hYQKjytiyxMf8Tn/Qx0z4qkR8sCDD8hVSPj9/vvvYL0RhhLVQeboziR7oOCpWq0atoOYQeuUc5BIuhoCA/W2wbmfff5fcv8D98uRI0eUZHHu99/+0pOF7Uac+YjhI7RN6ZdpQz+egSa4lStX1g20HHEQHT++zkezZs3CbyiN8B6bN29+8cUtSEZytjn1va4vrEsWyLW9HQTIq6++AoPlLRf35QO2IjlboUaM+J1zgznyr82NXTA6fmVuL6M1jINR2pyb+tDgOUmjn89Nd/z5L8fHr5/Hz18qjeelQszoWsUbIbFCZ+aRyCVPA9vnUKOwzckHyUiZEhUdhXSmrZrlXq5CmLA1iWudHi+MKrtrwJ0Xc8TTIckiEdOoUUNp2KChumUT4F179upiZ+ITsZs/b77M+fZb2YQos+gT0eolQ5lYHHxlGPO1avUqKHLcQfj4qIkQGUAex24lLjY3sKC8uehj4+3lqQxraQBv5zQEDAFDwBAwBAwBQ8AQMAQMAUPAECh5BJQsAQGTjQReBrvExUPdor4voElQ7zEZWNUuWgkWSxbUHB9++CG6KPqr18m9994rnTt21InxXEX40p5/SZqwSYfiAZSOZGVwBlaTLtKgcUN5+J8Pw1R4hpAU4CskAG7qf5O4QbTA49lVQh+b2FMx4u6J4BuchOQMz8vXTycmXFR+FOOaJEWo9li5YiWvpsNxXeezH/9uhdfNp59+qhvCKlZU75zvYQ8yb968n/zMx3Pu6xyMk+bYuHEjfGkGCtuNnGPatGkyatQofcrYbCY2cbAup09MAXxby5Ype/G6TFkKDAzULheei/tx8HpMP+aoBNFFp06dfjKnS+f4a3PbxLkNGPCTuU2fPl1GY27En+dzzo3KHF6bRE6ZMsHCOXEQG86N2+m/w/3+23B8rvx8FPH/tmuJvIZ8sJIfJFeyEUnNfrscyMf27d2LiLGaush+AHFSLaIaWopE3azjcZNwRUaANdy1ZxfMe32kTNky+IAdEjJGX9MgaOeOXejlS9MEpzT09LHFb8+ePbIJ0rId27dikbtJWfTlZedmgRlNRLb5NvxNlrS08yB63MTXx1uS8CHwpnID++nj5avnXbNmraYz6c0AZkZxV6aG5k5u4g5FDl+zYQgYAoaAIWAIGAKGgCFgCBgChoAhcPkjwPqNNR/HHKgqGBDTtm07reuYslSEBKKCPCheIFNg/UcrjaKCYtm8aYt6le7cvl1mfTNbHhs2VKZ9NVXOwzCYjUckVsDwoKOiQMqWLQtyh01PoGpQRwYHBksewm18UHce2L9fE5AGgEyYPOkLmfj5RFnww4ILwLlALLBSKlYM1/Ygkjo8N5U16WkZ4gO7DyWB8JyUAIUOveF/s3LVShAQnZWIUEXBxeMc75OtO04O4XRCAiKub7j4nHv8SC+4QD3SSxYtWqzzYXLT0mVLZcP6jTJz5kwlgxo2bIhQnrOSkBAPzAR+OiHy5dSpF+tiJjkxIhtTkLHjxuK9VNRzMV1pzJgxMmTwYDl+/JiqZEiIUFnjJDgomHD61vB4nZj+dTyhnwxbnDiehcCD4ooN8KGZ+c03MhvR4jQyTgGxFgfVE48nB/Dll19enBsNf48fO6bzHjdu/E/m9s4778BmZTBSnyOF+zmJLL3Yz35xnXD8VVxAqShm+AYywY7NWzAfqphoadykCRKagvVNxcAXZvWaNSA8XDX7nIs6LzdfY8riY5EbD1KFi7Rx40ba90dvmbz8XGU4zyLKuqCgSDaDeVy2dJnKrhi1fepkDDLgt8mWzVukWtUIoas2iZ4iSmB4U7q7o50qSBcDb5TQ8uURx11WSaCzZ88gfjtJb0Z3KGkY2U2nbsap8YNmb6ENQ8AQMAQMAUPAEDAEDAFDwBAwBAyBKweB0aP/rfVnVlYWYrJTUNs5lClUznihg4Nf9NPQlooVEhoVw8OkJ4x3WX/OmjNL3n77LXlx9GgoaT6S2khbYqHOIp3V4bq16xF77eEgFXgiHO+BmnP/oQNSHrYdrEFPnYqToUOHyYinR8jjIAE++PC9Cx4t6M7Iy5eTDL1R818cjP8kW0j8ME6bdA/L0HJoNbrhhpvUKHjatOny0UcfQfwQzZ11zk5VBz8VNebl5BzTwet84hh6+otP+T5+pAHou0PTYrYf0XOGHSX79u+T0yBl+NoddwyQ/bALqV69up6MrV2ffepQv7CFiV4xlw4qcK6GkTCvT5UMY7APHTqk7+fH2eAR3yAnxqF/HU+U/HJs1Y4bHv8y2o84N1qM7N+3T61LOLc777hTvWl/MreJn+np+va9/hdze+CBB6Rnj6v17J8hMp3v5T8Nfg4cl2L1n/Ytie0lrphRdo8gg3XMgIEQDYtq1qypi+wYmKszIFdoGuQKEoTgMrb63PlUyJyylInzhKSLLU/lyoWiFzAb/X+xumDbd+gAKVqWKmAoSWMbVNMmTaVFixYw/g2SObPmaEa9N4iXWjDxbdmypdRrUE/Kgd1TIyNX3nBF4ufnD0PhUGVMKWWi2zPPS8du3gyUpdEkaBPixKpVrabx2SUBtJ3DEDAEDAFDwBAwBAwBQ8AQMAQMAUOg9BEg0dKmbSuY4K6CMuJJ2blzJ4UuKLId7ETlKpVgrYHEH36RDyaDW2uhZmWs9rr162D6myefT/pcVSrjx78nixYukhHPjFCChEckoOVo7rdztMal2oWD6pk8+NX4+cJQFxNwcS1G60yAfAubjaioKPV7oQKEc1m5cgVIBh6JSenVRduCyoXC2wbHUiPAs1KEUBU/IahpmeREkoIqjw8/+lA6dWCrleP9cF960fzRwfr7ueeeU4XKiZMnoDg5rv44jP/+uUcNLUfYjvXfBlUul44/Ozeqeji/Eycwt+PHpUJYBalbp+6vzu1/JSov/dncLp3npY+3QfjxV44fqbISuqoudqwMfrjh4ZVgoJsHZ+lkXbT0b6HsKDExUYKQB0+Chr4yrmC72J/mDpaR/XA1a9aWrZCPJSYnYnuBZGRmSPv27ZG+1FyjzHgcjX15LT9EnXXt0l3uu/9+ufUft6k/zRlcLxVMZ3h4FXWHbtyoMVQxiXLy5ElNaOLiCgH72ARKntatWqN1qqwa/ubC9JeMqhs8bo5D6bMSErOdiNZ2yq5KCCI7jSFgCBgChoAhYAgYAoaAIWAIGAKGQCkhQP5l5DMjtbabPWu23I9akb6jTmKGChBfepRiR0et5yJBSAfaso0dGNWU76A6Y8mSpZoyzK6PObPnSOOGjdSrlKxJDmpHFctQSYP3Qc8Sbx9fPWcRjIBpMkxig341JCpGPP20igJmfj1TRkGJ44WalIOkCi+4HYRNaPlQNQd2kj2+EBXs2rVbRQeM7+Z898ImpEuXrjL1q6/0SCWBHCfBe9FNf/gXz189orowOrtRo0a/ID7+8IlL4ECdW3XH3Fjf/5wwKoFL/K2nKHFiRt8NVqYrFDPly5dT4oVMJSVHhflFqkChPMkLEWFent5q0EvDoKrwnaGXDCPJQsoGQ061EjfCMhAnrVQBQ7dsGvaSJaPSpk3b1pKcnATGcZdsg5mQM6GJkqbIqGPy3XdzZe3a1SpzyoeRL4+PjYlTZ26SL4sWLJTo49Hi4eWBeXhJLvY5i/SmqONRYN9qS0codMLCwjX1yYiZv3WN2sUNAUPAEDAEDAFDwBAwBAwBQ8AQ+F0IsH3mroF3y6zZs1Rl8ug/H9HWJZ5k29ZtsnDxIqhWoHPRFiW2EbnAXDYdtWhZJWvoO0OxQSZCa/rdcINM+HSCdL6qk7z77lipUb2GzqWYMhwMEjSRSBFu3qypCgucdEuAX6BEHo1U09kTUdEyG741L7/6sox6YZTMX/CDnt95/ViQP1OnTpOM9HQHWYRz8vzpGeloswqHtobXQuQ26t2ggEAZNGiQPPPM00rWOIQzJJ50OvbrCkSgxFuZHBiA48O6ISuZDw+ZGLQj0cSoXv16iNC6Bm7TSXiOWOq8bCVMuOArVKggRw4flo2I9zqNTHj2050BUUKZEiO8tuLmKY/YssZNG2t8Wfly5cGAZmP7FjkaeUR7/gL8A8UdpE/lapUlO9NVjhw9qkqdTCzmdCzwAmjCmLRE9pP9ZKdPx8uge+9Td2dGZJPZZNtT48ZN5LbbbpO169bp9S/tc7sCP2ObsiFgCBgChoAhYAgYAoaAIWAIGAL/3yBAxQkVFmdSzkA584z6r7z//vtQgTSQOXPnoE48go6K08pnUGTC2vV86jlN6HX39FCfUZrxhiE5uEfPHjJ58mR55+0xUrtWHfjQ9JB27dvKkCEPye5dexzdUDjH0aPHpFv3q3HNs9rWpDWkmwtabyIlIiJCVTi5qH/HgtghWTMRHifLly2Xvfv243pu2qpEcoaTUn4FbE96WroEBAVomA1JGobTUKww4K475YsvJiNx6DMZ+uhQSUhKkFYtWitB5PiQqeIxlsaBxZXxu3QUM1gEXOCFSEHSBaRYYGng5qheo5Z0aN9B/KB6SVKn5wR1nyYpkl9UoOa/q6F0uapLF+lxdQ9djNnIei/Ea+npadKta3e5rs910q5DO5VY+YGMORF1UjbCRXoZ3KQ3b9ks+XDYrhheEYa/fjoHziMH52C7VC30otGkqGPnTiCH/NHalKPXT01JFU8fL02Pok9NMKRslG8xmpstVjYMAUPAEDAEDAFDwBAwBAwBQ8AQMAQufwRIStC01cvLC/VnTZkMEuP5F14QN3RtsLPi9tvu0DehrUhkZVAVH4N3CVOFCgsKVZtCxiU0NERNftnhkYtAmujoKOmCejQ/t0A2b9ws/fpdryQMT8aa85OPP5YUiAtoaEMShUbD+QVFGimtBBDmRMLm4IFDECxcq50e11/fV4Y9MezH1hzsWIDj6T8TE3MK6pyaqsJxKGYEoTm+knYuXVo0awFlDRQ1sP3o1q073td3SsWQjjFKRj/eK+pXyRMzbGzDf8ZSs2WoS7du0r//rWp85GTtuMjcsCCzMxltnQRlTLJkwYA3KCgYDKAvosqKJDAoEGqV8qpgSYbXDFUzzCFnu1MBVDhhFSrKgw88KDfd1A9+MSHwpqmg0q1ziNLmqFSpkqpwXNzc1OuGZsLs8/PEzdiqdUsZOfIZue3WW2HEFC0pqWfRI5gDJ21P7QP0xWJPOJ0Ao+Jkh4wNhJINQ8AQMAQMAUPAEDAEDAFDwBAwBAyBKwMBlnBlgsvIraj5goIDZR+8WWbO+FqWL18hk774XN588019I2pbAW4mJi5G/UnpFUMJDQkUJvPs2LkDqUy1sa/D4yU29pT07NVTPv7kI5k06XNp1aqlVIK3KlU67Mxg1whLYl4/7Xya2nCoyS82uqEeZStSDyQD5RfkI076Kdm7Z6+89cZbMuPrGep7w0lR5kBRw7lz56RJ0ybq0crrc06uMBU+fOQw6uUA7Ikzg/zJhQihf//+QpPcJITtXErNcC42Ln8ESp6YwQIi88iI6yz040VUrSqDhwyWrl27SkREhPrOeIAcwS6O3jn8JSvJdCb2yrGnzx/u1ez7S4WcjNvy8wrhbu0Hk94QJVEYE8Ycc5r5YmnDANhP6tath5uitdSFkzbvgmK0LXl4uGtrUyWQOsw7ZwpTVNRx2YPFT4+boKAgzT+nrwzdnXfv2uXwwgFxlI02qUS0XDEuzYYhYAgYAoaAIWAIGAKGgCFgCBgChsCVgwBJDB8Y/FLl0qRxU60/aW/x2GOPyVMgRIYNGybff/e9kiHUpzC6esyYMejGiEahChIG/2A/g6TgLE1FYoy0KnGwjaExw4c/LUMfG6qKnOkzZkjLFs2VjCErU6iKGVf4zkRK125d0PmRodcnepwXrTuaNm0qZ5PPIBF4k/S/9Rbp1bOXbEMATsP6DcALoVrGdfBf3n1nDIJxtoKAcVyf8+T1qcjRHUD28MJe8E59443X5a4BAzQIx/lJ4SUbVwACJU/MYPW4YDHS/2XP7r2ItY7W2Gt//wAoXsqoy3XVatUQkZ0mvkhUIjkC6lAZQeLlClbS28sbaUg7JDY+Vtp36qBGv2xDyoLCJi83D+fdLe+OeUemfDlFVq1aJclgBelTQ4ImPCxMrx8VfQLZ64lIdMrUGyQwIEjVOAcPHkQv3gTZtHmLJCbh9YwMySvMVyUOb5wvpkxWZrMaCKUyIWXFDW1MvAlsGAKGgCFgCBgChoAhYAgYAoaAIWAIXBkIsJWJqpP4uHh0V4SBLIHBL1KCOT6e8LHcAEPfMNSOvftcq1YX3B59IlpS2YGBNiYeSwLE3d1Dv/AnS8PjSZZ06NhR04FJyCxdvFgCUIdu3rJVrunVm6e5MFwQVpMsKWdSNMCGJI/z+ulZGUqeUK1DUQMJoh6IoU5EuM2yFculPWw7nHYaCQmnJfVsCg1cEcHtIgUQNXhckuhUDJUO6+9atesoAbV06VKYCH8pCxB2o4MTvjCo6rFxeSJQ8sQMGTksYvbYnYXx0UmkKMXHx8OwN1Rbm47CrToXCUnZbF0KDNLosp7de0AZEwQCJ/cik8gboQC9e1WrVFV5WCbaog4c2K8sYihMgAOgqknCwj2FvrskMI5U2oRcIFI80UvI67OtKRXx3KexmM+kJKu0jIufSpvIY5ESFxuHdiVIvcA6enl44dqumEMO1Dt5UhGx3WQr+ZdyMxuGgCFgCBgChoAhYAgYAoaAIWAIGAJXCAKoS9laVIhWHypnLkpQ8KhOrTpIAF4id94xQM5AtTJ58hcyetQoxy6oF/mA9ejphASpUbOm+pWyUOU2/lQoX17uuP12nL9AUtFudDXipWd+/bXM/e5beQFeNvQ6dZIgX+DcqdqFgXPiH+tRF3zvT49V2tuwfOZIwLXuHngXSJWpcsst/eVj+NV4o67lIKnEfyRmmBJVp3ZttQ7hPJVqwZwYnnP/A/crocPuj+vhf/P6669LGnxanYPXdxzg3GJ/LxcESs3V1tPbWwIC/NXTJRcqF388poRr7ty5UqNGDamNxcT+PUZf165TCzHV0TJz9jfSsX17qd+gPpQxIGkoy9LF46pyMLYWHT56SB6D8zQ9ZHyRE7/MdbmmKnlAZUMClEbBoViU5ZHyRPUNF2VeXr5kggjKx9/mzZtjXgG8JSQ3D4bAIHCoAvPx84EZsLfetEehnDkeFYUbregiG3m5fGA2D0PAEDAEDAFDwBAwBAwBQ8AQMAQMgf+CANgKV5AV7Lhg7ccOCCVEQGxwdOveBWE0frILVhYrkdZ7CClNTF1q2KixDB78oLYSuaDlKAWECiOwf/hhgR6nxAoeZWdnS9SJk7qNxAiJl8EPPSSbkRj83MjntFXplVdfkQP791/svnBen2SRr583xAA0GXa0J5EtadGypapzRo4cKRXDwuWuu+/Sdqe+118nsTFxjutDXkPP1XIQKqSeO6/b6GnD4Q4VzVdfTZVAWIFwE0N4nv/X87IDnSgvv/yy1K8Pyw8OJxPkeGa/LxMESlwxc2FdSAAIl3ZIXyLJ4o1IrzSoV5islHQ6SdubSKxwQe1GW5Kbmwf8Y8pgARUJ1S49unfX106eOqEmTTm5Wbqw+HpxkauUhTqmXdt2cl3f6+SmG25EZFlPKRscLD7ePnCz9lJmMRA3YPXq1aUFzJhI0JCICSkXAplZgDRo2FCTm0jKVENbVXoGSBuYL3kgGo3O3Xt27wLjORMs6iLMN1H7AC+Tz8umYQgYAoaAIWAIGAKGgCFgCBgChoAh8N8QAPkAXkZSzqVI29Zt8YV8noOYwUZ2QySgJm2JOlFJCmwbP3as3HjjjRqnvXjRYvjGVNOXSOh8+OEn8CmNUqWJqzvao0B6kFxJZfoSCRD8kOugsGDCJxPkhptuRN0ZKlNgkXE1TH5VpXJhrvSX4bGVKlfV2pNzxOE6V6p6atSoroqcrds2S3fUxHkQEqxft0n+8Y9/6HlcYAHCMWXKFIk6dkwPLmKLE7aRiIpFRwgJJ3rm6EC1P3fOXOl97bWyYcMGxzb8vvDqxef24O9HoMSJGS4uftT0gqH0KjQkFMa9vrIFPXdFaBnqARKFbU1cZFyBCQnxmsveumVruW/QfRIMgiUDzKY3FDdsM5r7/Xcy99u5Sr7QkdqLufJY0Fx9XOTlK5RT9q9M2WBNcgpH/6A3lC9ukMFwcQb4+V9YqK7oI7xRPWNOoHcwIzMdBsX5SG4qL2eSkmXTxk1SuVJl6d27N3oJ86CYOS7bt+9Qd+6//2OyGRgChoAhYAgYAoaAIWAIGAKGgCFgCPxWBJjIy5CZg4cPIdTltIoAHK1IULzAJiM9LR11qLsSNjTS/f777+W22/4BJUy0/AN/77v3Pr1UUtJpKGfg8cI6l61HF/xnKDoognCAAgASHRHVIlQYsG3rVhkxYrjs2rlLFoHkGTbscRUhMD2J1+fYvm2b+qoW03hGBypbvHYCNiABCKzh3NevW6+EzMKFP8jEzyfKgIEDtBuFZ2Ba07HjkXplx/VhOIz3SiKJ74mX4X6hoeWkfccOcurUSRmI49evX4+2rGx9zXFd+325IFDixIzjjSHaK+28bMWipOqFiUtsMSKfMujee6QPGLs1q9eifSlKCRQa8JJMada8mRoERx49LDVq1dTHBflQ2aDvr1HjJnL7HXdI3759QeyEwD/mvEZccyk7OEIXmDJ5avKSr4+fuHq4QVJ2Dh40J3WbB5Q05UEIsTuKNwuNiXfv2i3HIo9py1U2Fij9ZCZNmiQ1wVS6IdGJ7VbqxnS5fFo2D0PAEDAEDAFDwBAwBAwBQ8AQMAQMgf+NgNpiQDEy91tEUu9xKFuwjWQIv+DPg6+ofuGPM5HccIEvxn60Ht1/3/2yadNmeXrkM/LVtGni4e6BmhDkCwpPNf8F6+EJGw0v/GCLEjucTMPGjZRIoSLnPNqMnkMb0ehRL8izz42Uj5EqXLVqNb0SCZNdO3fKho0bLrY5OSbnaJHyh7BAjYoRQnPi5Cl5HMTOsKGPIWK7h4wePVqCygSDeKHyBiwRru8IqnGRDHjPsLOErAvfI99lvbp1IHDoAXGDp+4/a9ZseWjIQ6jDj+NYvq0LxJDzr261X38HAqVEzCBoCSwMe+2iopDKBLPfB9GrVw2LMT4+AX4ziK2OjkIb0x6hke8PCxbIAvTtMTaMvjMZ6VkSimjs8mD4GLlNhnPOrG8k8miketX4+QfKcfjA8Kb56suv5HR8nLKVzoVFpQwz4kn48PrVq1eX8PBwZR0zocbhEs6C50w0kpvWrl+nLVePPfqYRERE6Gfg6emjcdq3gylt3rzp3/G52DUNAUPAEDAEDAFDwBAwBAwBQ8AQMAT+IAJVKlcmRwHiolgNeh21IkJqoCqhfQaYCyVmlNhAPHbl8Epy7z2DVFGyavUqRGo/IZ2QvrRi5Qpp1boNzgSqA3Umi8mYkyekTp3aIDZYWTrUKbTS2Lt3r/qdchu/+H/3nXflzjsHSJOmzdAS9SHq0hq8LGrXYklAXfzjIJvCsGJXtFqxRC8Wd/jh0GSYhr5fTP5SjXw1vXj7TggdesOPxu3i9Vn/nkLADa0+aNfheK/FsBcJUJKpRq1aeinyL1OQ2HR9vxtk7dq1jvejrxApXJU72PhbECgV819+nEHBQTBKag6ZWCYUMwVSEwqY8/CZmT5tuipT9DmIG0+oXA4dOCAbIKsKpaIFx5YvX0EXibuXpy5QOkm/8eabSHEKhLolUNuc6E9zAjdENH6qVqksLVq0lHI4zhfKm2IsTBWUYWHR9IiGT8FgFucv+EGqR1SXuogSq1W7liTC+Xr/gYOa5tSiRQvx9vGWJEZoo82JbVTxIITowm3DEDAEDAFDwBAwBAwBQ8AQMAQMAUPgykHgvQ/el9j4WFm7Zq0a5rI+dHN1EA8h9DflWyEfQh9TtCfVAnnhA3+W4KBg3b5kyVL9Yv+FUaNl9qxZ8hFULyhStaFi566dapFBHoPnpTqFahqmOJVHnUqCg88L0L60EsROl6uukoWLFsrddw8AIbJavpk5Ez6nGXqk8wxJSUlaexZuR0sSCR9ci8QM26wi4ItKNc+498arjce06dNA9Hwkb495R/dlQ1U8LEJ2o30qDZ0lOlhYg9w5m3RGqleLkH379zneP+Z8+NBB9bDhef5x2216LR7jbLXS4+3XX4pAqRAzXAM+Xj5gFltputGJ6GiZ8dU0XbBUxKSgxahJoyYaYZ2MhaKMJY6JjYnV5CZGhbH9qXLlSmqAFBsbrxHXyVHJavB7Hm1S3bp1gwxrMBb4YnXFZvqSJ9qPataoobIyNxc3qGsCSIM6JGogaPJy8/WmDIeXTLcu3WCM5CsxuCaVOmdh3sRWpgwYAefm5sihQ4fk3XffQYx3gLRuQ4bUhiFgCBgChoAhYAgYAoaAIWAIGAKGwJWAwB4kLs2Y/rWMHz9OPpnwqRTD75S1IcemzZulXbt26suCjdhSrPYXO3fuAEFTEwa/MPkFqbFtx3YZNOgeGT58uLzxxhsSA68W1rpM/XWa6bqoyqRY8lFrUgnjDX9VqnDIAVWByW/NmjVkxYqVMvzJ4fLm22/IW2+/rddmlDVrUMf1RdOT6qD1iB0jzut7envJwYMHpHXr1thPEEyTIB9+8IHs2L5dxo0dJ506ddL5QI2g7UtLly1zkEeYB+eZlQk/GRA8Pr5+Dn8ckDxstWqD+nYP2rtGDB8h7dt1UFgqV66i17Bffw8CJd/KpAsTywsMoRcWkoe7O1qaUtSw6JNPPpbdWAAxkFkVFheCSURvH4x2/ZDzzthsRn/Vq1dHKkFGRnaxDNjKNm3bSrfuXeXmm/tLRJVqilIuHK9J8HTt2l0TlhKhclm+YrnsxM3HVil6zfiCdKkJtUt73HDe8Jdh4hLWJG45zA1/ObeatWpL3+v66o2zYf0GzbJfuXI5evC8MDeR6KgoGCWd+ns+GbuqIWAIGAKGgCFgCBgChoAhYAgYAobAH0Jg7Lix8uKLo+WJJ56UyV9MkoqVKmpbE08WCVuMefPmXWj7gSkv/rHFKQs1psOnhW1LqGfdPfGlfS78Yp6Die8wCAeqypKlSxGr3fDinNAphYFEpLg4TRZmx4YWk6g5K1UKl6bNmiE52EMyszNlNXxW7xt0L+rYrmo23LJlS3i/5CuJQhuQr6ZOlbS0NBA8NAqGf01BkYoQKlSocOF6sOvAHL+Y/IW2LdWEyqd58xbSsWMnJVz0RBdIGdazMTGnEHYTJgXw01H+CQUxa92OaNHq3Lmz1sHxUBXdesutsnnLFse8L74ze/BXIlDyxAw+bK5ND6hXggKDlAXMgprFA4uabUKJYPmi4TZNhQsjxbKQAc/0Jv8Af12/7dq1V5PfDes2yKHDh9ULplxoqHRs315ag6QJRDsTF5434sS4+OlunYNz7d27D1Fg38qnn06Q1WvWqPN0ytkUbalyg2FTSEg5eQD9edf1uRY3CPsNITe7MFeeI+lMEljKnUJDJPrekH3kfu64iWwYAoaAIWAIGAKGgCFgCBgChoAhYAhcIQigIM3JzhUKA7p376r146SJk2CBUQUEhiOJ6RxsNkhesO0IegE5g8AZfsFPgQFfYFVbDoTI/fffj84ML/kG7UzrYL/him/5v5szFylHAy+AwepXUI/uhiigvWRC8UJuxhUdHCRkNqxfB0IHahReCy/Mmz9funS5SuvcFStWQJhQX6iMIbmThqQoEkQUFPD6qWhLYow250w1josryCIY+d41YKBsQ7LTTYj4psKGcxs+/CnxRM2NA/X6rHeZxtSuXRvU5EiVwvVZ/7p7gCTKzFAbEc6cipot27bCT6eTzJo9G1sc78f5lxjZKH0ESp6YwZyxFpRZdIdaJhG9clz0N918k5oeeYKho6ERFxvz5Kl8KYfeuV49eslddzkWdx5aiU6CvNm4aZMcPHAYsWGnsG+uHsNjmzRpLL5YoKNGPS+bN26W/IICKSjM0xizqKhoHLMfC3W79vGxNzANrU/5YAkjYLbUGf19N9zQT8IqVlBSKAtRaTSEUkdgLMJsEEX+fr7SsEEDVeOULRNa+p+CXcEQMAQMAUPAEDAEDAFDwBAwBAwBQ6BkECCxAQKF6Ub79x+QRlC4HDx0QAUAjzzyiKb/OviHYhAe2BfsQ+TRo1r/5eU5OjuY2BQQ6C/JyUlSBZ6mJDXO4Mv8uwfeLQsWLkRHyGfyNtqStJbEyc4hiWnmNzPVKgNn1aKY9A5JkYoVw/C+HMQKiaAzZ87KNT2vUd8api29/uqraHuqjGNYSbs4iB1cPz4hTho2bKDkC9uVGNXthnYk+tG0bNVCDsCr9fDhI/LB++/Ja6+9DmJlFvxYfXAltDbhH1uuvvhisuzbtxdcE+fE86PFiSKJHKhoMBcqc/j++f6Ygvz0088qqYXDdfCanLON0kWgVIgZfmwHDhySzz77TAkRX/TZ1a9fX9q0bKUu1VSiVEVbkj+2k2ih6W79BvUhM3sCHjPpsnDxEni9pMErxgueMh4gdlJBouSq74w71C/BwcG4mYLlKFKajhw9oi1Q3M5FRcYwGca9yWA846CoYZ8gjZCSk5MlKvIYFpyrMp5e3t5YcNnKItaoWUPlaJStqWQM+5Cp8QCbWLZMYOl+AnZ2Q8AQMAQMAUPAEDAEDAFDwBAwBAyBEkWAShA/kBTXXHONJCSclueff0G2wltm9OhR8uFHH0pERITjeiAdilH/0X6jChQ1KWeTQGCQ1oBXKbbv27cfUddVldigEW8COkAef+JxGTp0mDz88D+lDzoyqkZUw7lc1KA3IyNdiQx6uRTAtoPzYLQ2uQ03PCZB0xZ2GwWFBfLCCy/IqlWr5UbYdnzzzdfC1iYqZUiiuLi6o7UqC+1WhXLs2DE9p5IoIJxiYmMvKH9E6+3X3nhd+ve/BUqcLugC2S5hYWHqs8qrHUc0NpU4nB+vT8ECVUE0GCbhQgKK5sEkgOi5+s6YN+XLqVPkIAgfx4D2Rud94an9KRUESoWY4Uxzc7JkNqRQ6ekZ0hVGuw3wQbuhvYmLLABu10xlqoIFXh5qmSaNm2KxMlu9AGRJjhw5cgRkTJo0AJnTo2cv3TcH8dY0UeKiYGpTv343aq8f05mwjsQD8i56x5yH3ItGTOvgds39/RERVojFnAMVTixitXl9b2+0VSUm6bUKCwslMCBIWdTuMBTu1rUrpGaVQRb5KoNIltOGIWAIGAKGgCFgCBgChoAhYAgYAobAlYMAFTN+sMtgt0X9enVVALBp02Z0T9wg1/buA0JkpdaARYVQkmCwVegtJAEfxZf5JCzccDz/MpTGz88fe6B0RmnIejQoKFA+haFwv+v7qphg/rz5aCvqp+eh0gQlp+6bdCZZ1Tb56P5wqE4ctWXVKlWlV6+ejtCZgweVPPICibQd7UnX9umj3jckaDgmTJggW7ZsxfG8OOaJGpc/lPo4SBz+dZGlSxZJz549JTY2Tu65+x4ZjKActi1xqFoGf0ns0Ne1TFBZbtUfHsv3VAk18JDBg7HVBQKIo3LXwAHyww8/YB8tix1/9c3pQ/tVwgiUODHjWHBwoK5WVZlA9sAFQ3VSt249CQRJkoAEpDSwiPkFDvYwEDdLu/btNG1pxoyv5ST64PJycvRtMua6PnruGqGtiGQNibpKlSpBJpaqi4vGvdWrV9eFf/XVV0uXrt2Q4lRTfW22bN3KlSdXX90dEdo+6nHjzxsK21zAEJ45ewYRZenwucmRreipO4YbMBuqnIrhFaUSmELGpTXAdX1J0NgwBAwBQ8AQMAQMAUPAEDAEDAFDwBC4ghBw0S/vd+3ejS/6a4FSABGBenIDfGKaNG0i8XHxsmrlamnUuJFaWfCNUcVCqw0SHiQraHPhjU4L8BnYgqQlPsBo2bK1NIep70qoXZYtWy6bNm2UyVOmyIsvvaSvFxUX4HhXOXXipNx8yy1yPj1Nr02yCFPQWpSPnAQO26hYz06ePBnR3LPllv43a6cJT0YShiocqnh4fQoLtF0JpJFSMw7+RqpFRKiXzaB77kEozg4ZN268TJ70hXan6EVxYV6/CMenoiPFD0lNvD638QEVPtuR9qRBOPBgpQkxSayXX34Z4odz+r5cLpBD+sR+lSgCJU7McAGzR616RHV56603YSLUUT9cfNxg7NylMB8fMlQt51LOqQQrES1HWTAfImGyd89uZedyQMyw1YjxYalYBOwNzMjIVMavTp26Mh+GSd/OmS27sD8NkDzc3XBDNZYnhz8pc2bPkcfQN9ipU2cJCQ2Rykh4YktUYGAAYsFaS8b5DMlAu5QvGEmu4Tyoafaj5+7buXNl3brVEh0drSbFadinTJky4uNnxEyJrjg7mSFgCBgChoAhYAgYAoaAIWAIGAKliQDqUUfQS7Gkn0+Hga63KkxYc7IGTIiPl34gHSZO/FS6du0qY94eo4oRfRGvqwYFJEgyFC8Mp2GSMGtcfsHP48sEB0n3HleriCYDHSIPDXkI/qejZfSoUahHZ6NDJEC5EPqkjn93nJxCKrHDzMZx/UJ4pLLeVP8WnI+1Mp8/+uij8sRTT0qnzlch1OYzJZAcr2IHqmRAotCnlcIFVxApxSBrOCEW9eXKoROlaWNJTUmRRQsXycABA+AZczvat7aiNo/4yfXnfvutpMIT58frYwbAhrUwVUYov/HU0e40evSLcg/InpMnotUkme/fiRMf2igZBEqcmNFp6YJ3gT9LWfSphSPZyFPlWA1gutSjZw9tRcoHQZOXkwv362Q1SApGNHar1m2ldu3a4ufvr47SsbExsmTxYjmBRVCnTm0sJq6nQnjGJIDN2ylLFy7GNcpI9Ro1NNY6HYurbEhZlX8NRZzZvfcMUgVNDtKfSPaUDQmBWicN3jPx6KFrLFXhjs0WKCplkpOSJSn5rJ5n/YaNMmfWHBBA85TEKRmo7SyGgCFgCBgChoAhYAgYAoaAIWAIGAKljgDqUXqnMCTGDV/ia6qR8hqO8vcqEB8kVF586WVZsGCB3Ny/vyxasEh9Vjg39XgBV8F9wsLKg8RAghPoDxcqTPA6z7sdYTPuIC9UiYNt48ePl+uvvx71bB0Zi8etW7dW/iLyWKR6vLAVica9HPSqod1GEbbxfCQ76tWph5YpX/l84kQZP24cYq7Ly/q166Vnrx48RPd1wxyY7HQGAgZP1NhQMGA7JwUfGwgbZn0zC2nEoUra0FbkKnjOZMESZO3adXLTTf219uW5aFS8Yvly9ZZxXp+1dl5+LvxYPS6+fypk6tatrUlS/3zkUdm/d5/O1/GLZ1Kahg9s/EkESpyYcX407NWjzIofllPyVQHeML1795Y2bdugd++I1K1XRxpDOkZ5FuO1GRtGA+BHH3tUGjZuKCFlQzU5KSUlFWqXNtr+tHfPXumCZCVGaFNJ5Y2UJ5ob7dm7FylMi2QO2L8NGzdo3BmjtemAzePXQ7K2csVKXcCUa4WElgUrGi4+vt5qGEx5GHsFg4KC5DxSnKJORMniRUtAGsX8SYjtcEPAEDAEDAFDwBAwBAwBQ8AQMAQMgb8SAXZy0Dy3AiKv2aLEViQqRLBZvU4HDLhTPV4iIyO1PqX6ZM3qtUrSMFjGOV599TU5DvNdtvvQNJdEBvkIKlzckEJMbxpu5DE/LPhBbr/9H3Ly5CmYDI+WZ55+Rk/Detd5fW7wh+cqY7eL0DKkA8ez/n3wgcEqaDgeFQUi5SZZumyJLPhhAaKwh8NiA6of7Mfrz2b3CFq0+Gac1yfFkwrLD3aKcEocW+FNc/vtt8vc7+bKW2+/CZXQ9ertyteyQdhwR7UiwUmJka+3n+OclMxglCkbLPfdd78Eoa5OSkyUV197VcaNHYf6OYsQYDgUQPrQfv0pBEqcmOFC4ae0Y8c2mT5jOhYbjJO4gCmFwsLzQka7D9qIziA5qXGjxvL8C6PgAeMrJyDvou8L245q1awlrVu1kg4dO0i7tm0hFSujfXCFBUWatMTM+MEPDwGj2VDOofctF8obsppMYdq0Ef19kybLxM8+1cQmxnP7+flg4eVor10gyJc6YDGTkhLV4doD1yOBxBuL/YUkjkgkFcDkiexjTk72nwLYDjYEDAFDwBAwBAwBQ8AQMAQMAUPAEPgrESABg8Te7EwNoSnIK9CLs85zRccEPVNojKtEC8iNqONR0gMeL9/M+gamvhOkFwJoQsqFgHZwUaKEFhsUBZDscbQ5OUgVqnKUBUH9W6d2LSQ09UEs9yF56cVRMgteMW8gLYk1MetfDqdggTVvCLo5cHodvI6vj48mPtHThvU0fWDuHXQf0qSeV2Lm0aGPSXh4uF6fRE90dJQeSwUN1S40GOb8qMbRE+Cc4ZUqwoe1i4x4aoQ8++yz0qpVa4gYNuHaMP/V93KBmMH1EpBk3BrWH/Sg4fVJVJUJLgu1zVppBj8dtlGRxHkKJNHgBwcj7QlkFYaz/tcn9usPI1DixIzOBJ8O24VeR5Y6mTeylIcPHQRjma2SLTJ4ZPYYR10BqUwenh4SCRZy2ZJl6kR98MBBHB+Lm8VNQpG6FAp1zO6duyUH0qpQLCISPBUrVETWexWJxX7HsCjobM34bS4NysmOwEApLiZOzY7atGmL5KfGiNkuo3I0LvwkeNvQ2Klxg4Yy5KEh2v5UJqiMVKtWTRciGczKVehPgxvDhiFgCBgChoAhYAgYAoaAIWAIGAKGwBWCAAgK1KSsO2d+PVPi4CnD+hNP1SumAPVkSmoqui8YY80EJjdJhsXG8BFPySPweSGBMeatMVIbHR7OwbYe/uNZ2B7FYBn6zlygJqRxkybSGqbArG0L0T0y7aupcsstt0pb1KLL0DZET1Tn3gsXLVRD4OJCBzHCObD+TEpOVPsNntbNxU3SM9Nh4jtWVThVKlWRYUOHKdGic+LJSMpgTvx/Ft4yYRXC1HCG5+NMK+MYbvNA29OMr7+WyV/ADBik0PYd27XtitdwEksHDuxXIisTahgSP3y/rqjHExNPa7IUyR8SS6zFSTYxiGfdmjU6FV5f35zjmf3+AwiUCjHDxRoRESFNmzfTRe4D9m/V6tWybfs2Vcrkgc1zh1mRZqbjQy+GqoYKFbYQnTx1Sveb+uVUOXoEUWX44AvxkwefmCL40lQCGZOVmYNF4ioVwipIjerVlRksD/KnLAicSpXCZdgTw3AT3CIBwYG6QMhGMp6biU4rV6yQBQsXyFkodtgzSPnZbbfeJl/PmCHtO7TXuO3QcqHqUh0UWAZkTtAfgNUOMQQMAUPAEDAEDAFDwBAwBAwBQ8AQ+LsQaNuunV6a6b579+5RbxY3KFwKoAhhQ1JhfoGqZnQnEAtME3Z38wSR87WMfw8eLxAQbFi73hFlDQIHbi4XW5kOIuK6RcuW5CowwJDgeD8/P9m4ZRPCZyo7yB7UmXO+nSO33norauBkeWfMGGnRvLnyF+fh8bIaiU7kVi4OPMnPK7zQsuQwL/Z091TP1tVosXp65NMgmOKQJLVShgwZIv5IMC6G6oVECdUtJ06ckD59r5OcLEfCMRU+7FbZvHmzeufwvXOfm2++GS1ba2Tq1Kky8tmRmDt2xP/8vHwYF8/VYB7W8zRPzs/JU97Gy9NH3z/3c8f7agySiUKIntdcI6kguAqKoEjCa/p+fvKmLr47e/A/ECgVYoZMImOuX335VSVPzuHDKoRyJh3tQhnpmWg9yoYBcHn1j0mIT0BsNVKZ0Dakciqwe3SvTkxK0tamPCyQXJAyeTierUVsO8qEJI0MXs2aNdG7NwoO2C+gLaqRtkFRFnbLzbfIB+9/ID2695DMrEw5ER0NdUyGEjinTsXK+nVrIb06jp69pWr2xGuRAPKB4iYDc2kGQun2O+6QXtf0upBZ/z9QtJcNAUPAEDAEDAFDwBAwBAwBQ8AQMAQuGwSeQrrRyJEgHlCb5iD2WgdJCAzWfZ7wOOVTqkvIJdRAbTl8xHDt6qB6ZtD998mSpUtk2rRp8syzz0hoWbQeYfAMR44cli2bNmmbE4/nCUj2MHmY4gG2F/HcrE137twlg+65W31eOnXuLJ99NlHPk4twGtpxOAdrYh9ftjw5WoagcJCK4RU1EckPnjQZaekyadIkefWVV+TVV1+RD99/H0E3NXT+rI3TYfFBA2PagyibcoEgyczIkoiqVfUyTDQ+CkKFqqDn//WcDB06VEkW4sHd6TtDKxK+pyLMn+eib+vF9CccH4Cgnhtu6Cf169bV2px19fAnh0sMfHWIjeOXXs5+/Q4ESoWY4dpk+pGLm8MNOwmJR9nwgSFbeejIIRAxWbrgycrth2RqxvTpGnvdBpIxmgBfdKvGJ0s1DcmYIjhfU1KVnJwEl+rxIFfWqemQP8x6O3bqpExmJfTcxcXGyYEDhyQhMUHoJ0MyiNnya9etl3hI2M6lnYMyJwPO2FDnwNeGPXNroObZgBtr1YqVsg/R2WQw/x97Xx2f1dktu+JOhBCSYAnu7hrcvVDcXdpCnXppKdAWShUoUKC4uxaHEiRo0EBCCA5RAvHkzqyd9Jxzz+/ee879hfOV73t2S+R9935k9n7/WJNZM/wQtAMxU6xYsf8GnOZUg4BBwCBgEDAIGAQMAgYBg4BBwCBgEPhHIkBiYd36DfLBBx/Ib4t/k6BgEhgkS5Q6gFjgqfgi9Yi1pr6OxQYEBKAWPC/FQWKwsyMBrUFjx47FGFO0heinn36CaS+ipDEG691Tp0/DpiMTPATHhNFwWgrIjRwlY5QTQdtP0aJFZcTIEVp//ggi5cCBA9IJqpZ96OKoXLWKEjq5X6DquQCv1WAVJXD9LNQ9oIqJiLguNarXUDjTEZU9ddo0GTJkqNRv2FA2b9qkfqxZ3AeWEYYWpSQQNGoIjCtSSEhhIEf1uKHvDpKh0LnSCNcuWrxEO0ecnBzlPIyEQ0JC0E1i0QNWK1OOihYY1kNPV90lr4ey6CY8eSogslvRxGZ/w1jdX+khf6Km1gOvKQbWb+brfwEBC/n/won/nVP4vDMvff4v8+VPJCTFIPY6FewbZU6nT52SiOvX0NvmrP4wsbFxasRLF+shQ4dI9Wo1pAYkXk0aN1Vna/b/6ZOPJ40PWCYe/n1/7FVZ2Fmwj0/jE9Uhmqwmz01Iikc60zZZ+OtCNQE+e/aMpi7FQj72CMqYu3fuyDm8tgNsIh24n6KdiYlMD+4/kMNHD8uSJUvV9Cn84iVcR2ZVH7f/zvbNuQYBg4BBwCBgEDAIGAQMAgYBg4BBwCDwD0KABMRGtBENGjRIaoLUWLUKthX1G+auxgaJvUck+la0pKZTtYKXUfI5OjhJVOQtVc6QGIHEQP/IvwDx1U2bNEbrUk1pCEKjUf0GqFPtcJ1FhmTnoF5FK9Fl1I8lipVAvZqhihk7eMb4+PiIk4Oj+Hh76XgUKjCam90kGzdslP4D+/81/80bkXL27Hn1ismb3xYkyoMHD9AdUkm7R0gYpUNpsx3pT00gTngMRUuNmtWlf/++4oJ2K27EuhY/YU8UIpSCqoYdK2RuiIsD1lO+fDnp0L697D+0H4TQRSQcn0cb0wZ579331Y+GYyjTgxHPnDmDDpRInZ/X24AESkD6E9VAJINoU8O0qzMvNf7iAABAAElEQVSnw4BTU1mxYjmXwf+twzA0eUj8X7+/EGKGM9LDZdnK5WDPFsvq1avBDh4Wes0UKlQIBEq2snEkRvih8IHb823EUt+IuCFu7q76YXgPcjFvb2+QONcljXnqMFGyMtXhjA0FTSpYu6cgVFLxYN68GSU/oHWJcdhslfIs4IkeuXQ5cfqkrMLc1zFPdPQdZQ/5AMUhh/7Q4QOye/duqGvCoZBBshPOd3V21Zana1evy7WrV+FMfd+kMv1fHx/zpkHAIGAQMAgYBAwCBgGDgEHAIGAQ+PshkJNjKxs3bpTuPbrLPdR1s7/7VmrBF4ZajsREtP3s2gUfU3qowPMU5EEaSBp2arBVh9HWNOOlb0zFipU1aalx4ybqR7p5y2YoaN5AlDSSjchfkILA9ytIYypXsbyaCqvRMOYnkXLw0EEpgi4MG1tLscIU4YEDBgnHmT/vV2nePESvz8zKkP3wj3mOgBrOz2Rj1rQoncUV6+AaedArhmKGOCh6unbpiq6Q+zAIniMzZ86U4kEllBHJm/8hIq7ZARKNWpvzcz0kbFgz22B/Ak6Jlh8jkbL09TczZerUqSCMNoCcscdM1oRHjhyTMJAzTLni9fS0URlO7r6VncGpvgV9xQMdKwNBhrEr5eHDx1yurscayfpqvWi+/u8IvDBixgYPkhNSl1KgOrkNQ9+Tp0I1lqsnGEIfEC6X0W7EG8t/7gXcIZ86LxMmjpdgMHqBaCN6hh47O9z8+w9AjjxPgzFvQ2nTpjWiuqppn5s9Wp701uJ5SktPAeMZJZfCwyUB0WdDhw6VHq/0lOqQh1GZQylWfEKsRoCVL18eD3I18UQCE42LEhMTJRlu18+Sn6F/b5CMGjVaCuJDxg9oZnqGfjj/d9DM7wYBg4BBwCBgEDAIGAQMAgYBg4BBwCDw90SAJIZFRNioaqR7t26ycuVqadCggfz888/iDx+YrEwSDDysZqR7d++KF/7Ar3wD3wKDwSCYV1BXMiX4HuKkL168KDNmzJApH74v8+bOlRJoe2LdyCMtI11+X7JElTiqWgERQnKDrUWsL0mwsPblERv3RN6EL8voMSNhzFtZfU+pbFH2hTxP7vyxUMQw2joTRsX0W6VHDNdVt35dTR2mr8xlpB+/NnGidOzcSdavXSu1a9XWGtYWShaSKEuXLpVwrJvzU/HCgyTUM5AzVPoQK3rCfvP119K9azep36C+fPLRp5oiRXKKe6AYgm1avJ6pTT4FvXRsvkuCh68HlQyWkcOH83Q1HB4xchjUNmGcjpD+9VV/NF/+EwL5Tswok4ebwYeuWPESmo7EHPjgoGBVzJREfx8NhLYhGYkPeLNmzSC/qqku2MyHZ6R1mdJl5ez583Lk2BG5eeMmHly+7iWMvR47ZqyaOBUNLCL88LANKQNpTXgatJUpG9FkjM7mh65r1x4aA0YjYj6g/v7+cLl21V4/EkD8IHGdZFDjMHcwHqZ+/foh676SOEBZk55JhtL6oP0n5MwLBgGDgEHAIGAQMAgYBAwCBgGDgEHAIPC3QwClofqWsluiS+cuWod+8+23mtDbskVL7apo1rQpS8jces9GbkFMULdeXShWnueSELZQjjhpVwZDZxgtTTKH8dUdOnSS1q1bo149qkoRJhiRfYiJuaM+qGBRdH6G2DhArEBfF9I0ecRIU9TA/kgVXgJvlk2bN6M+9pRNW7agzamHEht5JMgT2H5UqVxFxQc0F9Z5MFJU1C0VNEDro6TN+nXrpFXrVpjBVk7BOqQyyB57O6peRO08nqkKx5qf1iCMz2bycd56uPgKqIFvRt2Utm3aKnE0f/58mARPRNeKo46jceHYZ1paqhQuHKCvaSsTrsXW1d81ArW7qxuMhLH/Y0ePaa2/ZfMW61z9ar78nxDId2KGzyTvDL87OzlKmVJllBwZN26sVIIMjEoXSsJS8MC7u7prDnv/fv2lEkyFKB2LuR0D9QpalGBUFIkevxOI91q/fr2sWbNGW4sy0LNXukwZ8QZzePDgAUi2vpYomA+VLVtW7EEL8uHiQ88n2sXFSQIDAkEQFZO+ffvI7FnfSU2QQGyncnd10/46PpCHDx2WVWtWq8v15ctX9DrGeZPwMcSMgmm+GAQMAgYBg4BBwCBgEDAIGAQMAgaBlwYB1oVO8F0JLBKoNSDJkqtXrknPXj1Ra6bK6rVrZDgUHiQRQM8o4XAd1hpRkZFQkVAdApoDX6iU8YRIgHHZaOhAzZotp06clKpVq0JEcEPjp4fD4NcT4gPWoKQpLGIFpAjaeaqhWyMjIwNT0NDXKr99vH2kJYgUihpY/w4Y0A9ExlFZvHippi6xLqaAIBvfN27aAA+YC5jdWhOL7efo9uAeOIC1epA1WHebNq1kAVKfBgzoL5999hnIH4tA4U2DOEbnZz1NsYIdukdY63L/9IhhkE63bj3gx3oWqqJfZC3UN4z4Xvr7UnGGJUkWZDzEhGvesG69RUApTgoUiChbCT3+p3i4Icab6h6cnIx1spXs008/lXiQTFwsV8wxzPEfEch3YkafDHxxdnKRouhn8yroLU4gSNzdPcTb10fJGpIoHTt11Bak+Nh4ZTDZ3kQG8k/03IWFnQHzl6E3PgPGSPyZMWMX0O504+YNuXj+gj5YriBXGFV278FdGTZ8mJQqXUaSkKgUc+c2HmJ8aPCg8QG2RTqUOx6QGrVqSrlyZZTEadioIXxr7JVhJDnD1qmdO3aq1IuKnXj07EVFRf4lN/uPsJnfDAIGAYOAQcAgYBAwCBgEDAIGAYOAQeDviIBFjIAiQcoQg2f4h3mKB3JsstHWEy7t2reD2e06ja5u07YtfGVAqoAs2Lhxg0TCIgOnoha11VqRv9DrVIkN9gfhxMCiAfLw0UOQMq/IRZjnTp/+FUQAc7T+5dxag+Lch08eSa2aNTQEh+Po5RiBXqmJ8D215mG71W0ZM2aMTJnyvryLiO93330XxsFWPPcFtCFdwj8SO1yTtk6hTYpECakWvs41FSgAg2GQIxNffw3qm00gZwbIMhjx1oAViB6YjO1d3Cdr6gB0k1gkEl5Ck4gz1EX0f3Uv4KGpVd9+O0v69e0n3bt3l5MnTkhwiSDUxty/SFR0lHrn5NBQWHeBPWMdKWh5coFihvvnYQuT5MAAf/n8889l8JDBEhv7WM8nwWOO/4hA/hMzCrIVw+UKZo2ES6B/ANygS0NOZSfpUKHQ94WtRElg0NauW4v46+/lNh5GWwc7fUAZdU2DX8q+XBDtVQ6tSA0bNQDB4wyDpgy5ixYmPoMNGzSU9pCRubm44+FJFlfkvic+TcIHapNsgZkSlTTMe8+AqRMfFD6FNrb2Om5BtFHZIgmKcdquzm7iBpLH2c1FmVIHJwfIxRJlE+LHHsAwyRwGAYOAQcAgYBAwCBgEDAIGAYOAQcAg8HIgYNEFKBghzXiegnYi1H38oz2qQd2Ai5OrjB8/XkaOHikVKpSX5SuXwzajrr6XQZ9RKkFweQqShQv6+OKP/vyDv6UCoe6lUoXKMmzIMBAucRJ6IlT69iGB0U32IFymUqVKqD8ZL22rcdqfT/1S7sTcAX9CwiaXkci2kaeoU9EfhTlRp+KbI9qLfkCkdufOncUXRNKi3xZKu3bt8LaNZKD9KAfMEFUo2fjdq4A3RT56LYkgDlAM0dzjx41DAlMqEpJPQP3SXf1pjv0Zqu1J6VTt5BIrf+zZp0od9IfkEj0kbWy1zmaQDvfP6Oz169fBnLiFJMEs+deFvyIJqgk6YNDahCl1/URUfWjQ4oRwHifsgcoZFN+6JrZrjRw1RvHftn27/DJvvuzcuUOv5xjW2rkPcwC1fD5wEwlyyVKlpGq1qjBBovLFRlxcnfFo2qjZLhnGxo0aqbkvzYFPnjohETciQKy4yRGwdFFIWeJNonSMD2iNGtXUV6Y2FC8eYPC8EDdmC4LFGURNTTCA3j7e+tDQ3Mjd3U0S4xPUcDj6drRcBxt47BhUOGdOy304VlNGpiwj1wmyhjI1exBC7IUrXNhfnJ2d5frV6zAETlbpWI4+VPmMkRnOIGAQMAgYBAwCBgGDgEHAIGAQMAgYBF4YAiRWaJrLP/aTEiFJQ9ULj/oN6qnvKNt+2LLjgm6PVatXycCBA60TcltxGAvdpWsXECMgNVjNoj61gerGEZYdGfAjZWw2a9A9e/agZq2BP/KnyUEkEnVBWpKTs0VgpKenIYgmTikZhtvw4Dea+SojwxewrrLlymp71M6dO9FC9DWSh2/K8mW/y3vvv8czsHzOTSInW89lmjHrWqu2FZAwBSU8PFw8IYxgqcv04RYgVTbDw2bL1s0wG56MfTJSm3NjzXv36M8ajY2lZKJThbUvrUcooWErFfdH4qU/WqP27t0ngxCW8/NPP2jNrBdj4VwSiTDGehODDAgsKIrg6wEQaFy6dAmijCDdP0UTneH5882338BEGObD+mreSLk3x/r1X+5r/hMzuAN4ZsSvUEGpWKGCxCE2+yIekFu3boGg43Q5UKe44yFpbrUJ4dwGyIIPRE8bfV9uRUXJRTxE7HUrWNBXfJCexLYkH5gC+4E48UQEV+nSpbX96WniU7kD9UwK/Gji4hIwto2ec/L0Kbl8OVwe4uFIhZfNfXw/ffq05r3HxNwGu5es0WN5RkzWB8xGPWrI8EVF34IZcJw4Q4FjGSz9yz0XZsMGAYOAQcAgYBAwCBgEDAIGAYOAQeClRYDEAtOEihUrahEFuTvh6/Qh7dWrlxIidxAl3alLZzl08JDaWnzx+VTUgPb4h7hqWG0c2L9Po6nVdwb1Jk1w+fPp02e06wIUhI4cjRqybds2smLZMpnywQcyZMgQCQiAxwveJlHCLxaxIuLh4aGEEYUL1tWC5ODqqpbh7/fuPpD33n1Pps+YKV9N+1JWrVql5/Nskkv8nfoB/s75ObwziKDw8ItSBHNSncO54pG2NHLUCHn//SkycsQIEEZdEP9dUdebR8hwPtbvTCnWrhL8wvqbg3qiBm/SuJGKG6ZPny5Lflui0drHjx9HsE6AYmQDFQ7nj46OkhYtmsNL9pnOz0lovhx1KxJ2IuWxSlBbwJ7+OWzVor8PPXp48Pp/Q0Jf+pf7kv/EDEHFg+AMQyFXFzeN07p37y6ImWg8TI4SVKIE2pnS5dHjJ8oS8uY0b9VC+vTpA/lXD3jPdJJSwaXQwuQsLmABC/n7aQTY3HnzlCzhDWOPn6OdA5jHeDly+IicDguD2/RDmAsxYttGsqCKiYRT9enTYUrc0NOmOBKiHOGqzZanJLQ7JYKto89NU7hxs/eNDy7XQoEbpWr0taFbNfghcxgEDAIGAYOAQcAgYBAwCBgEDAIGAYPAS4QA/VjYLVHIt5BV51nVv9Z9aVC7sHbkH+qzoA5hG87QoUNl8ltvyohRI+GHMkiCg4N1t3v3/iHXrl5VQoRCA5IsHIptTva4XhkS/O6KgJtkRFC/9/77anYbEtJcrTGYEMyDBIrV9iMSn5igvqd8XRussFY3V1e5FnEdHSLwu8ltW/oaEdY9e74Cn5qaMuOr6VCeBOvc9+7dkz/+2Ks/c37W35moYdNS09Wol6QL/WBcYC3CwB22SA0fMVI7Tfbv3y8hLZpzaj24f9bYNyIipAPq43Sog1RhhHfd3dzlAjx06terp/v/M/RPCWkWgvXZyAm0cHVFDDnrb86fDnXSt9/MQp2dZA2MMdPT0yUdayI2rKvz9l8AMeQM9+nRo4ckQzlD0obAEld8+5c88p+YIZC4OYy3ZsvRTThb86DUyt7eHpnnBbXv7RiixaIio9Tw1x0EjpenF0ibIOnfv59MfG2Csnm+MDxiPFcOpFZH4VJNgicV8rAoqG/cC7iLm4c7fk+VmOho9Mjdljt3YvRmVoeEij12JIAY50UyqFlIU2VFQ5qHaK8dJWelSpaSVq1aiZuzq7J9tWvVgXO1v35wSdJYKU+6fPPFIGAQMAgYBAwCBgGDgEHAIGAQMAgYBF4WBLTaF/l1wa+IsUadCAKCxaIqRbJyEPYSr50cJCV48OtsGN72798ftaynfPTRR/JKz5763lOoSXiCEgv8jlfZaWGLNp88HqFqlSry4UcfSsrz57Jt61Z5863J2l50GK1NnSA+YFsQVTg89qMtiLUwl8SxLCLFVi05aFTMteapazZs2CCv9OotT2KfSD/Uyq+99ppeQ69UnqfqGKzlKQgRtljR15Vj8ksJ1MFvgmyiSiU09LisWL5CTp08Jfv2/AE/mvEakGOlUjHqO0bOQNiQAFNixYpDYIwkzFMYRsGkTdjmdOjwIXkVaqO9aN9atGihFEHd7QRRBed89PiR1voYQM9//PixWobkxXwTLNbZ7Vq3lTKly0AxEynnz54TRnMzxptjcM5/xSP/iZlcINPAjvHJvwF5Elkv5sd7QwrFIx2GSuxBi42L1V42L+8CepNt7W3EGz1xhfz8pERQCbQ3FZGChXylbp36Uq9+fb3h2dmZcHOOhR9MYekJhU2fvn2lPFqmaELEuC8+mM1CQsACttC+NX7gqNRhSpQv2NLatWvDkMmK5r5y9YoypW7uLjA3spOmIU3gPN1XI9U4X9GixfTB0UWbLwYBg4BBwCBgEDAIGAQMAgYBg4BBwCDwUiBAr1IeSSAszpw5g5oURSlqVdaMdqg76fFij5ReLVbxupu7u9SqVVv27dsnv/w8Dx4vkbIG/jNTPngfo4BowPXaxoTfSIqUCC7J4bX+5PciSCS+hXYmCg6o1nn08JE0athItm3bJkuWLJGxY8fCd9WFp2K0bNkLxQu9XixiiIqXLF0KPU9zYA5MM15XqGiatwiR8+fPCdUzJ0NPyndz5sAY+Dedg+QN618e0SBW6PHCpGN6vJAMoZFvGLpLioI8ISFCUUPffn3l86mfyqeffSpffztTlToWjSII0Nmq9TFbmTg/RRGM0qayiAf3T5VQUHCwvPb6G+oDS4XP5i1bYAPiqueQfFGPGfxG0+Pq1aurckavB5HFsQKL+EvrNq3V0Jibfu3112UM8GGXjR7WlhQP64V//q/5TszkYogWo8NyDvHWpcogjQng+/oWFB/EZeMZBJP2WB7cf6BeMVTW2Ds4qf+MvZ29kjEOMOONRVw1ngVtW+KN69C+PWRovshrz9BeQT7oZAQbwp+mLSLOqiFHvk7tOmrgSzmVO+RSzHZ/CNaOkV+c+DF+vhlx02LjcA5Jozt37kibtu1B/NRTw2Cu3x1GwHTAfuvNyVZf4D//c2B2aBAwCBgEDAIGAYOAQcAgYBAwCBgE/mkQGDZ0iHZwkPiwAmDyKlXR+pMkCUpCJTaoewkG2dCvXx81tn32/KlGYH/yySfy9lvvye+//y7BQcGwwIArDIY5f/68VCxfHr6lNAXmDPBTARF0KfwSQnBKqmeLHdqeHj15JK9NeE3efvttVeKEhIRImTJllHBQ819cmLeq+/fvixvqUBJACMPWMamead6sldakDM05cOiADOjXX1q3bi37QSDVqF7jLxLkCWrsgMAA2H1A8YJxuTeuNzoqWsmRPFUOW7i+mj5TBg4aiBapWlILwgX6yPBQBY6a11BdlKPeOp4FfHRNXKgNUqRoIFwF6iC2V82bOw81dIz4oU4Pv3hR18V2L7ZHcWeM9E55lqJqHI5PIklbzDKzNTLcMhdGexiYoEWLFmmy1bGjx3CihQr38K9y5DsxY2GXA+Pdh2rmM3jgICVlNm7YCLYSciuAnJqagoSku/LKK6/Iq6/2RnyWPUx/b2nkFhlCWxs7VdTwprmAeWNkNv9R1ZKJfkB6xCwG6/jpJ5/KmbNhKs3y8vaU0iCBypcrq5FlN9BCxdYpfNKULaREKzIySubOnysXLl7Q9Cb2vGXBfbpokSJqtrR61WpZvHgxcu791KiI5sMFYDZsDoOAQcAgYBAwCBgEDAIGAYOAQcAgYBB4eRCg+iUs7IwURzuP5Q1DNxfruBUZqQbATG0iYQEmQDs39u8/gMSjcsqWsO1o6tSpMnjwQKlcpbIsWLhAvBFMQwUJ231WrV4Jf5RknGuNS6UKzYZ9fXwtUgTXu8GjhX6mvy3+DbVvT53jzz+Pqc8pI7z/WhBWcOLECUR3VwBhwTFBquB6Z2cnBOmcl5pIJyZTg+YpWblyBXxnekLNkiKr16yWksFBKHmztSZesngJjIPvYjTWwLaonbPUM6aQXyF9TReGn1zhPbNr5y4lUrjmw4ePyKBBg1Shg7f/ImKYMlWmTCntRCFdxPntQM7cgrVIQKC/rikBkeFssWKL1OrVq+WDKR9qO5XuDftYs26tXLp8SacmMUObEnIG6WnpILOc8LpF1nh6eirh1TSkmZw9czYXm7w7xlX9cx/5TsxYcMGAFw8HZVWeXgU0t3zvH/vkbkyMkiVk855AEZOGm0FfGTv0wW3dtlVWrlopDx8+lMcw8iXTx9edwcg54p+Dk5OkwJSJN5CMJvvVaFy0fccOZNLnaCRZ2TLlpACkVM/BJsbcvQPyxV1jyhjRFX7pMoihBPjU3JNoPEjeXj6QaaFXEPea+eo+BaHmwXEXD7J/YT/xACGTRQ3Xv86zYN0689UgYBAwCBgEDAIGAYOAQcAgYBAwCLzkCHz55TQVA9BbpS+CZhgEQ20LSYFTp05J2JkwbdshMcP60gZKD9aYpUuV1vqS7Rts59mCNp3u3bqi0+KuqmFenzQJI9igrryPwBgqQ6wjISlR1TZUqWgNCWLEHxYdU6d+rmoatuls27Zdw2s2wjdmMroz2D1izS/y+MljOXTokDxhSA7qW20fwRi05ihaJBDnYVZ62mADJ06cBJEyGK1W61QNM3v2d7qIxMRE1Mwp4IqwH5j/JsFk2MXF8n/hmknW0IOmHTpOqlarqhYhTC+ehfjq777/Tr755hsJCgrGvtBWBcKEbMqe3XvkEtQwvJ7/c8fP0ZmSBdULqSIe169HyOixY2TatGny2eefgswaDIGFm773HOa+DODhxaqWgcqIwgxtJcN3JWYAWOWqVaQbzIQZxLNw0W/y68KFcg9dNlY9/s9flL8gYgYRYCBFXCHFohs0+/ho5kPTIf4ei9iu0NBQ+RW58QcPH4KjdapmzEdGRSkxEhoKQyK4XxdBjFkRPIR8JrKhlMmgqzMMfW2RyESZlzMesiw8NCDtpGKlitKkaRNlH9nWRMPfwv6F1V8mCuMuW75MqNqhb4w/orkZ3QXuCA9WjgQFl5QqlatI165dpU6dOlD4+Ik9VDtU2xhmRj9P5otBwCBgEDAIGAQMAgYBg4BBwCBgEHhpEOAf5Pu82keWLVsuP/78s5r50nuUFV460poOHzr8by1OOfBTwR/3Wf+xzmQNyBqTypI2bdrAOyZGBg4coCnBX079UubPm6t1al6tiFPl/LlzUg1kx3OY/+YdTEXaCWVKMdSmPCggGDx0CDxevpB333lXFoJ8oFJEGQ8s7MTxUE17ojqF86dDsUIyxM6egTi5hsA4r2WL5lDtxMsUJECdPXcWCcevyq5du6D2Kavz8AtraHq81KpZGylQibmvW8SOK5Q8bGMiKcI2r48+/ERGjxwt7dq0lQ0b1kmZsmVQf6NNC2wQ1S5nYFHCvZLYyau/reQmrBTn0M+GrVw0Wu7YsaO2Xh08cEDTnOxgV8I98CC5RPFGfAJILIgvyA9wDQL83ZDqHIAgHgcYDHOvWzdvkld795ZTp09h5rwBOIgO9U/35YUQM8Sqeo3qSnYwF/4ZsszZj4Y7pkjmgAWj8dCNiKty/M/j8gTsoBs8YWxxzoMH91WWdRspS0Hoz2vdqrUkgOk7ceKUtkaRxGFcNh9yTzCMzFqPvh0tJ8EaMoLbzd0NxkvFpXad2jIKkWANGzRQJtTB0UEew8k6IMBfGtSrL42aNJK27VqJH9qW7MGOcn4SOz179pAatWpo29RzrJEPjzkMAgYBg4BBwCBgEDAIGAQMAgYBg4BB4OVBgPXnHdhnfISkpOFDh8mrIC/69e2NP8Kj1QjboHCAP5AEoN3Gvfv3xBu+qGxJIjFCZoO+pXXq1EW7ULASCDF3Y2To8GFaN65cuQqhM831XBIH91HH0nOG9a+SFfSjQZ15ETYa/n6FdS57/O3/Kew9fv7pR20jat6iufTt20eqQC1Cc968ZiuqSajWiUeXib9/ANZktVzxNR7FSxQXxnFz3oiICOnV+1U1Ft68abOSGWq+i9qbUdTRMCS+DO8bnIqOFH6zQXJSinaScJ8kWFLQFrV+/XoJadlC1TLTvpqGWrmdKls4HwUSquzBflLT07TbJAfSHc5vjWCj/jaTYAjMtKZff/1VIm9FyhZ0xbz++mtW0jLOtLO1iuuDB/aLh4eHii6UlwH+JGj2HzqoFibcP888hmTmpk2bypbNm7kM3QMVQzy4nn+m4wUQM7w9cLVGBDbNii6ijYgPiw3YR3SkKXaUNTmS9QP4VNOQKaPxL3VR69ZtQM9alPj4+EB6lYgPTAYMg1JxU7fI4qVL4BNzU4KCgjT6yxvtRlTmPHz4GDdri3wBmdhcsKGXL4VLRlqGxnqVBttHSVpRqG8qVawM12k37dsr4F5AGjduLJUqV5Y9e/cqsRMD4yInpDd5gbVkGxVlV1ybOQwCBgGDgEHAIGAQMAgYBAwCBgGDgEHg5UGAodFo3tE/wG/YuEk6d+yMFqVsBNSck85duqC1yTFXfcGSOAdtT/dV8ZKEliTRVCRcDSZl//59EhwcrEQAQ2bWwEelZ89XtD1n3do1Uh3dGjoQeILfly2TWxAY8OD8GemZqkhxhhokBxHdllAB3SWw3zh79qxURlpwbGw8kqD+kDFjxuSaFePi3PkT4a3KDhIrwpr1tEVG3L17T3y8YcUBAgMLUwJjQL8BavWxEIlNXbE/0EK6LFqGXEOrkc6PoUloEIeUFKvliQlQPJE2II9gKdK2bRt0r+yX6TOmyxfTvsT8RfC+dXB+kkTsgqFChrRMNrtiMATr64NoxfJCHc+AnYnjJ8hPP/4sM2fMlOUrVoirsyvLfT2iYC2yGWSL2pSwUyV3X8+eJaMeB0+gqVAYHeOyLu/Ro4fMmDED4T7PuSs9eHf/mQ6wIfl8ED38H3MnRugszRz3VN50vEZFDNm7tm1aQSXzSB2c7XBTVb6EZRDchIQ4eZbsJ8WDguTo0SMgWS4hsz1WF+lgay/sm+vcubP20kUiwoyKmfjEeO3Ju3zlily9elXqP3qisi0a95Jk8S1YEF4zTlK8eDGwewXl6JEjcgAsHYkadxA19LWht8yjR4+QFFVKalSrof43johP+2dj4vL5bpvhDAIGAYOAQcAgYBAwCBgEDAIGAYPA3w4B+rE4omuiW/fuquK4dOWSbMIf81OePdfAlwULFsjnn3+O+jRT61AqaA7Aw5QeLzQM1qIWZApJEYoGrAMkDsiZy1cuS4eOHWT2d9/JG4h6njVrtnzxxVSJgxEuOQaUvuoHk4YOjAKoR0lggCnBoV9UIBAHNcxRKEK2bNkkd+Ff8zmMhusjKfijDz8CcQI1D0dB7bxq1SoNpCnkWwhTW0IH2nskPk1UOsMSjtho/T1+wngodC4i7vsXWbrsdxkxfITWuZyZU9N3hh4ueVHi+jroEtbhZcqWlbJlSssqKIF+XTBfDh05JKNwfX10m7wx6Q1NPM6bn2RP9WrV/8P8NFhm21EgBBHcf2xcrHz22acavLMMqVYnT52UQ4cPg8SBMgiEEskbdsFQBMTXGBfOligN8IGChvu3BYnQHfdvw8YNwHiWtj81ax6CuaspPHqWdaq1lZf4q/Vk5OcGcBeIDdmu29G35c7dO5pyVCSgCAiQB1CgZEi58hVkyntTNGKLefF0tnZ2ddZYbUUYN4SEDc15b0F6xfYmKmN6IMWpIIiVJ09iwZzZSzVkojdp0lh70ai8KYvoMTKKZAAfPLwPE+HHKv168823EKddHddYLtIJ6GnbBNb0669nylG4YvM6T5A4JI0e4ZpIKHaYTU8m0BwGAYOAQcAgYBAwCBgEDAIGAYOAQcAg8BIhgIKUahnWf8+Sn0rNmrXUEDcZNd7b77wjkydPkomvTZQ9e/YiichNMhn6gitOnTqtNhmsaKkCyURdyRYj1oX0RiHZQ9IhqHgJbYma9MYbqnbpCtNapiXVr1tfy1ltJcL8jx8/kk4gcBh6Q0NeO21FAk0DYqJS5Uqog23xXppMnzlDevfqLZ06dZY9f+yVUqVKYgls57FBK1I0oqgv6fwkhXjY4Tq2XPE3bcXCdw8PdyletLjMhf/NvHnzJMA/ULZv3wZRQye8izOBCYkP1urZ2C9FDCR1KESgiMIf/i7eUOE4QrGShqSmLfB4aQ+/GIbrhB4/gfatplpn43KJQz3OIJ5/Pz/Tjtm1oj6wIH+IP/1iLl44Ly1atJRYXDNs6FAZOmyYvq4X4wtFGlwTDYJLlAjCHGglw975qgtayfg+fXJI3IRDtNGqZUskSu3gMnKH4NUv/5H/xAwwycEDTPKDsdY0OOLNKVqsqLpGz0BmekzMbdxwB5gDo6UJEij2p5UKLqWx1YzroucLfWTs4NKcDWlVVmaOMmdl0ZZUEGzl/Pnz0NJ0Q0keRpxprx3IHbYmjRkzVsaMHqW9gBlog+IYderWUQdtnAjFTZI+z15ennpzScCQES2J+TPAPPL3LLZPQenDcS0G8uW/0WYHBgGDgEHAIGAQMAgYBAwCBgGDgEHgXwIBMAYkUcCF6B/elbxQoiNH/UcXI1Y6pGkI/ohfGGqQSdIiJMTyUwEJgFJW/5EQSEl9bv3hH6QDSQhlNzBmqdKlZRBitHnOhQsXYTLcW0Nn1qC1iWlJ9F3hBU8hNIiPT8A/KGlyX+MojLG+duWaZXSLEzNA3LBlqkbNmvCVSZTpX80EodMJLU/uPF2JChamyuvgd2/UxHZMf9IDFAjmYshN3359UTc7qKJnzJjRcubsOVmyZKn67BCD3E3IwYOHpEK5CrgarUhYKkdyRjjOcQT0+PkFaH3OoS9dDpeWLVvIWuzrpx9/lA4d2qlHKy6x9ojvJKE4P0UV7iCHOAn3z//8QPaw7esSrEYGDOwv5y9ckB8xzrx588XZ0dnaF2p0XnEpPFx6QoiRgaQr3gNi6+ziBFuU6xIYGEAQ8FqOJkl1BIE1fcZXSF1O0vnwRY+Xudsl727m7SVfvlOxQmLm2rVrwpz2yMhIJWjo9XPmzBm5cvkacZUYsH90k6bahQlOXt5eCq6ru4eULVtaKlasCKavEB4uO5yPG6YflBxtPeJNyAITR5MmGgeVLFlKGT4fH2+YMIVA8tRD257uw42b5sOUbIF+02x5Mp7BwSVl3NixUg6SrWvXr4GkQWoUPiB00X4KedupUyfkClqjsrLhRm0Og4BBwCBgEDAIGAQMAgYBg4BBwCBgEHhpEGDtSJ9T1pqsPfnP0rvkSCG/Qtp207p1a4lCrbpi1Wp57713oZ5xVZKAtAOce+U56sJatWrrGPoa3uX/TPiNiLiBVhwrwYktPO3bd5C98C5dsnixvP3223qODSZdjN+jom7p/GzZ4Tj8j2E4DhAjYDjOJgVQ09Jeo1PHdvBA3S2T3nwLBM30vzxelFjRs0WvK4QQG+6J6hcqXvxgMMxam2E3PKi0GTlsOFqsvpDPP/scCVCLxAVJSJyNbUQrV67U8zg/9Sl8nSKFAh7wglGTF3SPYPx0GBqPHj0aUdpfS6NGjWXRb4ukcuXK+h7n5w+c/z6irXv16qUesSRIiH9htF+RE6hcuYrGje9BctTgwYNkCJQzoSdCdW9U8XD+ZCibvv32WwgknmJfZGbstD6nTw/HoiInr6OF9f+UKR/IMOyPvrC6DKyE572sxwshZvjwp4FdpKHQgwcP8bDfkoePHogbGDTfQgVh8pOi75G8oTEvk5bq16sHciVYCRQ+oAGBRdS06NU+/dThOQ3XJCcn87ZrG5Mt/GbYm5aWmqbsXg30uHXr2kXKly+vY0N0o3KxWPTuPUafYBy8a1Lw8POGZuMfP3R14bBNR+u9e/5QCVrys6c6Pw2MEhOfyq2oKDVsellvrlm3QcAgYBAwCBgEDAIGAYOAQcAgYBD4V0SAihm2KLmgeyNH22NQvudW8PXrNZCQkGZy8+ZNWbd2nUybNk3eQ/T0BiQTeXl75/qdkPTIRtz2MomLiweEaB/CODzop3LpYjj8TNkOZI2bkZ4qw4cPl9fgOdOzxysybvw4qUovFBxP0UnCwyKG0IqEWpYcgvIIXBN+oE1H3z594K8aJwsWLpSvYLzbt29feNBsQY1bAfPgakpTcKxfv06eshOE1XHu/E7wRyUZE4g6mkyJKobwPr1ZunXrqnVy585dsO8QjIDEKliOkABhixUJjWcQKLBrxRaWIWzbssNczs7OULz0VHJmxYpVsnjJYg3hOYj0pJYtWrDgVoy4r0ewLQlAgtRTtCSRV8FLYodWJtqSlEKdz4OtW2vWrJGQZiFalx88eAB+sL5KNBGG21h/GoggJXawdgosmK5MVQ7xJxdAroD+Pr4+BWXjhvUaE34A0dx6c3Eaz3wZjxdCzDDJKAEmvfYAjYqU7JxMVcw4ITfer3BhiYJvDN2wGWntiJvFOGxKr6rDdJfKlQSQKTRZonqFWeY+UNKQveM17J2rg+vYimSX++GgwTXdmn3BGpI5pJomGTcxAH4zsU+egHmMQc/aM3V9TkNOexpalvigUKXDsZPhh8NEpufPUvTDxjSo0jA+Koy1vrR39mV8Gs2aDQIGAYOAQcAgYBAwCBgEDAIGAYNAPiBAEiALthclS5XSdCUW9UoWgHDwRC1Zr249nSUF0dHfzZ6FNKK2UMfUknNo/wnGH+9ZX/KIQfz1uXNncS0dXyxihMRDOiKkHRydQCLoaRjTS4KCg+WH77+X19+YiLq1gKyGEqc32px4GecnWcIR7FEDe+F8lZrgBRIsZeB7akV5Q0ECa42DBw/qetxQsx45clgaNW6I2hlJUjjSoSIhYcMV5c2fhvWwc8UFZAqyk3CWjXi4uUv1GtVl89YtMmz4MLRWPYW57wo189WdgMBRXLAhBveUZe1OfxhlVkTXSHVQ9erVMHam1swkdo4d/VN2796NfU76S8VCGD7/YqrW35Aq6bpSYWvC113d3XSevP1H3IhArPersnvXHrU8mQ5ijPMQHItowhpAFmWgdi9WrBjmIEhsmbJVw+CiRYqhraoDeAarhYpjsUWKUd66L50VE+feG/z0tz9eCDFDyRZVMM5IQmrdqpVGYUeCjXyOeKvsTLQTQUkTByaQBkJOkH9N/+orKGoeIsrMBSa8nvqQ34GMi1nquyDjonu0Gx4qetYUKOAp48aNk0YNG4iNg51cRxR3GggcsowZMBbWRx0PJFm1AgU85OaNGzIBUV3Lly/Dw3ZP7CGJolyKnWxkQ2tBNUNDJrKZfn6+6knDD1lwyZKa/OTs6vK3v4lmgQYBg4BBwCBgEDAIGAQMAgYBg4BBwCCQiwDqURboJB0uXbikQTKs4lmn57CdCMRDNGKtC6gniijhwYSkpqhPz8GsdtLkN1H4d9Q/5JMtYH2IAhMtN9YYGSAAXECYoKj8q/ivUKEiFDPDtNY8EXpC5iCx6TRanFavXI2kpQ9xPYkZK/D6PAQHVatWAwmRhZexKvxPb1RaaVBcQMLBASRM1K0orClENm3aJB9A0dMIwTeF/f2xIrRo8SKQFdY3K8TGHQSI7hsDsOYNxFhdO3fVaOvrV6/JfihL5vzwo8yGimbJ0qWwDGFCErGykStIN27UqBFsPZJBwtDPxk6NgO8gMYrCCNb4NlAKxSIxeeCA/vIJEpfem/KeVICapzRqZx5sq9JEZu4Vv7PG9/LwAtwke/AC1mQDwqs51EokzN56azIUNQ9k/IQJcgT4FwCZxYPzk1jjIAFIeVIFTe78trg+GbwA1TjEn2QNO2Jef/0N+M2O1tY0YqIHF/GSHHyU8v0gWCQ93EGmkCnkA0ifmbj4eEiknmlSkpeXN9qEIFPCTTp54qTmxtdE/96XkGwFIis9BeoVJjtdwwNC+VlzxGJ5FfBSZ20XtCF5evnozaXcjDnoVNSEnTwlNyNvqgwrA3FnvCHsieNDFhYWpt+94UFTo0ZNKGsKKznkAP+aJk2awbCpj/Tu3UdK4wEh05iWkoZ2Jxd8SB3yHR8zoEHAIGAQMAgYBAwCBgGDgEHAIGAQMAi8IARAAPAP/qzPDx4+CN/T4yjuST+A0iBJgD/WMwXIzd1qReJ5tniNRrMj0I60e/cu6d+/nyxetBg+pp66SFAVkgORAWv9+yAT6tSug9/h75LLAtDj5cL5C+r1QgKCoTYDBg6UDz78QCZNmiyzZ88Rz1zi4dataKQebYHPDIUFudwOVDj0S2HYDVuMMARmtEWdmwhlyhvy3ZwfpCY6TFYjPpvR1jotFk57D57MmrhVy1ZQmaAOxu9UnLBuZkz1X61QOP8rqFP69OkrzZo2kx07d2snCsmZdBgQf/fdbCQbP9L5OQFFFekZaTAh9sQ8EDcQQ46N+v7rmV/LwH79gaGbrEZ7EkUXXBOx5D8ej9G9Ur9hfSuVCjtlmpQ9uALW4uXKltP9P8E5VL94eniqMql9+/ZcPg5rsGXLlsKE+DL2ZEchDkiYLGGtzxAge4xH/Lked/jOLIXRcZcuXeUufGb/05G3qP/0xt/jhXwnZnS/ALIger6YxOSAXrfUtDSYANHbBYQHfvYr7CfFiheVy5fDwbg9EXdEYbP3jD1sdMcOCg6S6zeu46F4ArlVsrJ1NFZyhaKGzB1bjc6dPaPsGulEGiexdemjjz9WV+0ffvxBToGkoRyLSUtk29T0CedSIlUiqAQUMUHy4YdT5LfffpMEEEZ8+H0K+qhUiu1UoSdD5fq1CJVP/T1ulVmFQcAgYBAwCBgEDAIGAYOAQcAgYBAwCPxXEOjfvz98TBFfjYMiATW0pc6EygtU/ozStkJmUMGyfoUX6vBhI9QDdfv2HTLty2nSpWtnGOqeldKlS+MU6zxyBlfxh/8SQUHoCEnBuNbrjNVm21P58uVwhkVg2KF2pX/NgIEDkORUElYetUAcdOaSMAaCZlB36oFBU+CpSu8Z2oFwSI7HVKYJ4yeCNEmTnTu2y9Lfl0q5MmXl2LFj0rxFC004zpufSpXEpEQE5TzgSvV6qklIfNCmQ1UNKlsRWb16NYx6X4GvaoI0qN9ARo4aiStyYAFyT/1a8+anB6sDUp6ovqHKiN46DlD2vIL0JBIhf+zbB+LnlNyKviWHDh7Sdil3JC9bnjA6JEJ1TmlKlBItOn+O1ujazQKRBPdJz5qu3bvK8T9DZfmKFRrkQy8ZHvT3uQolEfHn/PSYZbIz902sFGu0OjWs3xA+PdU0AWrBgoXa6qVA4Awlz3jj8l7gj3+zI9+JGe6XapMnyGy3+sNsAB7VL8810ssJ7U1PYMZLOdiD+w8lHgbAJUuVhAnQWvlt8W+akJSI1xhXzdYnkjCuIGX4UN6/90CJEpIs9Km5fv261KxdW4JKBIGZLKQ3NfziRVm0YJF8/NGH8v0P38tZ9Ag6wjCIDKaTkyPYNHwI8QHhA3AVEWVHjh5TlQ1VNZSK2eHh4Hycn7+TtTSHQcAgYBAwCBgEDAIGAYOAQcAgYBAwCLwsCFitODt27pBXQECAJtGF531njedTsCCKe4uoYb0eGBCAeGZn9RmlIe7la5cRFlNPbkGJwjanvlDQ8GCTDr1PfoQYgL6oefV+JgQBGfCG8fSEdwxzhkCKuLm7SgNYcOzcvlNGjhypnqarVq9BC8+bFkVgLUv5ApoJByAWOhNjc/4ckBie6BjhGmuj5gU7ovOx/t2/b79GWL/55pvappQD01UORdHBbXinkkQhYZGRma7XF8hVBlEVxCMIQgUSKsNHDMf5t2Xe3Hnyyy+/6Hv8wvkh1kEt/lSTk+ghy/lJaNEcmGqVtm3b6PlpUAYNHTwUCpoZSG76BmbDs6Eywrq5B/y3c9dOuXIJxAquxdXapsTN0xPHBR0qfJ2/x8Nnduy4sUiQ+ky6wKT4o48/EjdXN52DIg/Oz4PX8XUSaxbJRs7JBjW/vbRr105xJ+H10Ucfy4SJE9TkmO9bR9733F//Rt/ynZjh3tLQc3fq9Gm5eP6i2AMg9sfZk5G0c8CD4wSG70/Zum0bnxXt9/OAF8x59PL98tMvshYyqMOHDiNSLFIVMuy1s8c/Xnf9ZgQYzEeYIUdz3y35mZsqcMqXKy8fTHlfJqMfsCtcp3ldxPUIuXjpopoWTZwwXjp36aLsGh+kTD5cfAJA8mTgxiWhFeoBiB9XFxdp07qNtGrVUlU92tvGTZnDIGAQMAgYBAwCBgGDgEHAIGAQMAgYBF4CBGxk4YJfkbK0AWa3q9Ap8ZGa89KygqX5sT+PaavT82R4lVIegv9pY3Hh4nmpUbOmvmQHEuNC+AVtafoNEdFfz/hagoOCIRjw0P0/TXqKOGcqXqxin2076pGCX1Vxgu/esN8YPmy42nzcQzcH24pmIxL6s6mfy68LFkgxmNjmHZevXBL/gMJIZYr9a34beMDcgFFuCZgRg+tRgQGVJlS4fPrJp/LOW29JSEhzFTpgCxovnYbWI87PVKX4uATU2x5Q5qgjDUgLa7UNGzWUjp06SXxsvFy4cEGGDhkKFUxPjekuX6GCkipso6KrTipEFuxu4fykVki2xMbFQsiQAkKEszLR6Zl88umnaAMbASVPc/WErQtfGh4keags4pm8NhuY2cIrlr/bQyFkETY2mtA8dtwYtGzNUYKnaJGiEoYumUZYK0kYzk8FEEc5HXZaseBYuirsleMdg+jCwYmKI/wGcurnn36WXq/0khOwTuH7Cgy/49BzrB//Fl9fCDFjiweIAMcioppg+6BPjn1wlBYFgIlMQ+vRHci8KIMKDg5WwyAaDRX0LQhG665EgpSJRt+dXyF/KQrDoltQx9zGv1T4zmjPXI6tHD56GB+cC5r6FAd2LREGQP6BgeiXe1UlXisggerVu7cEFQ9SMqZsuXKQjXWRQshSfwB5F3vTKuChY99dZq7XDR8UKmkK+xeW2vC7adSwiaZA/S3ulFmEQcAgYBAwCBgEDAIGAYOAQcAgYBAwCPyXECAJM2bsGJk4caKMGDlczXj9ChXSAv3+vfsaNMM/zrNgZ4sMCYSU52lQzvjjNYvAITnx5EksSJDPZdiIYZrmFHr8OLox/MAHWGQALTHATchFdG5UQwISCRs98D67MebOnSulSpVUUoCCgw9gBDx08BBp2KCBrIRfTIkSJfR0erysXrUWdh4PMR7jqi1Pm1S07ri5emBF1jw8md0nP0PhEtKiBfxbC8uBQwdg0FtH06d0QzwJ87PubYB5EhLjuUStdblfEiz+TCDGOTyWLF2CiO+eEFI4yuZNG6VFy+Z6PsfatGmzHDp4RBU87D4hucEOFipXcnA9DX+p7qGqZQOupXcN17dj5060S/XSWhwX6Xi6J8x3Hy1TXvAAolhCFTMYll0rd+/cx15d5AwImS+mTpWrl6/o/JMnvaWCD2tRAs+gPyX0xAnEoQMTjM35SbTExlv8g9qYWO9IaGioNG7SSKPQ8/aLt6x5+cPf5HghxAylUk0aN5HyZWDoA2mTM3xdfGC6W7VqVWnbpp2qZ6KiouTR48f6AaCMrEvnTpqCdP78eSVwSOSQKKF5MFuW/oCDNFOZGKHN/kCykcnwn+GHi+qaY4ePaAa6K4gWPmglg0uq+VHbdm0hASsgO9CTRyUOY7NTwOjxxs2YOVPKQWlz6PBhbb3ig0FyjYyiulA7Q+mT29v2N7lfZhkGAYOAQcAgYBAwCBgEDAIGAYOAQcAg8P9AgBHLJAvmzZ0P1URPKYias1v37tKxY0e98nZ0tHq8qHICBANJGv6Rnt0e+hrIGrbajBkzBoSHrezagZYceJ08Qw3K+OyeGEsjnjkaaki+fjfmrtpkWCQIGncw3r2796GcsQyElYTA6WvQJdKzR3eoY55I586dlUAiYXL37h2sGUlNIByyMQi9VHUOVu0gjjgex6a6JQDECmvnVSB3zp05Bw+anTJ+7HjtFrEMjkVbrk6eOIVOkmu4Cgf2yf8YkHP92lXsM7eVC68fPnJE+iNtaeeuXfLuu+/Ka5Pe0EtYc0fHRKtFDefn2FSwWGoZ8jRYF/4rXMhPWjZvCWPk+0qCUWW0cOFCmA1Pl+LwnuXBPfLgXM5OjvCzea7X83Ua+B44eEDrf55zK+Y2CLUR8j3ix7+a8ZUs/m2xYsH907f2KgyBOTcx5fzafkVVDdU1+jpUQzAa7tAe9xv7HDx4MJQzJ+DB8zB3HXmr4Wz/+OOFEDP2MAhq0qwx/jWVe5BaFQKjWKx4CQXSwxNsH3q8SMzs3bsXLU+n5BHAyQAp4uzkrGbA7du1B7M5AUzYUTz05/DhcJDkpCQleZKTn+rDOhnO1nyI09CGxA+PSscgK6PpMMkVd093zYH3RjqUKwyI9u0/ADfr12Xi6xPl6LHjGvPFD6cHItJ2gc3bjbU8gzP3nTsxKtVKhfkSOdC8D88//laZFRgEDAIGAYOAQcAgYBAwCBgEDAIGAYPA/wsBi2yxujiKBxVHC1GYdELrzl10Z2yDpcakSZPEDea1JClITtDc9s6dOzDJLanmshyfxAMVLo+gYKmJVF8eTALuARXIhg0bZYGSDtM0ahtnQ7UhsnnLZh3HIhrslBjJgqjA3sFJ58mrLYsghfgqbDe6d+0m4eEX5acff5If4I9ayM9X5+EXO6yJHjalSpXUxCS+xuupUkmDuqYt/VTw2rPkZzIICpzZc2ajnegT+emnn9Q/B0wHL5F9B/bJjZtR/zY/L0Kd+xxKHO6fP/MrU6weP34obyB2mmt5pUcP2bp1qxQvjnYrYJFHgqSjbvdFFwqtQ3KH0hF8ChWUWrVrob720GCej9FqNX78eOnxSg9ZtXqt1KrFFjGLDGHy0wYocbKYagX8dRysISXlGRKZXZSY4uppXDxjxnRV8zRD6hMJo0oQe5Agovkvr9M9YK8ZWRka5oOX9D2+yz2VKlMKXEQQSLgsNSKmKTTjypnSZR3WmnJ/+Yd9y1dixgKaZIaII2RQTngA2bYUgKz1mjVr6IOdDfBpwkvDHoJz//4D/LsHydZjKV+xPOROvZUZpIzqdnQM+tz2Ik6rEOLI6oo3WqKUPQQZ075DO+nXr588jU+EXKqpNGvWTFnOtPRUnd/B1kFblV59tbeMgtFSMAyOvLy8tIXq7JkwCTsdhg/UBhBEt/RheA5z4szsTPTw3VRZW/jFy3CpTrKYt3/Y7TETGwQMAgYBg4BBwCBgEDAIGAQMAgYBg8B/BwESGPRY4ff2bdtJtapV1LB29+7dMN59W955522Zi1Yg1p/slmCLzlP8kd4DnR+379xWIoJGuSzZb968ibYbxGrzP5ABd6JvgyB4RyZMmCDj4GM6YGB/WHL4qKKE7UhqlIvz2PVDZUdhiBRy0O5D5oaEAqmEGmh56ta1q7bgHDhwUNOM2rfrKCtXrJKGsPjARHo+W4Jo25GMzhGaACuxg/eeJiVaxsNQh7CbJCkpQb6BB06Pbj2kfft28ja8ZypoOhSmw/k0Jub8ej3mZxoVySi2R3EqvCRlSpeVkSNGkf1R4untt94RX9Th27dv17X+NT8UM9ehwKHIgdxGFmpobsvN2VXWrFoD82Q/DicZ8J2lvUi7tm2RvuwkVO6URVdMnodrHLx0OL8mOOGC9LQMkFxOGAu4Y1HUwVBF07//ADly+BASsrpgn0myd9duGQWPHYo3OH8OMrSJdcTV65rMpK1luJ73n+nPx9H2VKFC+b/q+n379kvz5iHaUYNl4uBqcfyD+Zl8JWbybjT3RXaqBCKpY7BNLQAAQABJREFUS5cpAzOg53iYYKyEIxNMVgG8V616FWUtycBlIR/9KZQwfr5+MmTIYM2ZX7hoAQyYXNH2ZAc2zhZu0IHqlJ2AxKZjAJeAxsBxOh0PWeXKVaRrl25yE6TKuXPnNQ7tCdqkMvEBKFEiSNrgYWjQoLH2uL399ttgzcpAReOsDFw0vGsy0jOFaVE+3gW15SoJfYFPnjyWuNg464OlKzdfDAIGAYOAQcAgYBAwCBgEDAIGAYOAQeBlQCAHvqQ82FpToWIl1N/ZWtvNnj1L+vXtj9cqyIZ1GzTemj6mLOiXLVsmZxGPTcKArVBkLdJRK5IgUAID4/F11o+MdW7UsJE4ouvjHJKAW7VsBR8UxDfzMp3ZTkmV1q1bSxqSlji/VS+D4EHLDX1NLfNbeLwsXiK9X+2lhATDcDp27mgZ3mKctWvXyzm0LHFgXk8agR0oCUgvZtowKQz+x6SoI+g4qVa9ur73++8rUFsPkQKw9VDuxdZqheLqmLZUtBiUMFgH98290XiYNiS0ASHLFAaD3f4QQpwAobJ4yRJ57bXXdVecnyTLESRVcWDOz+9c2wOoizwhhqC/TzZ6sUhSRUZGwrOmpfy+bAlisLvLx0hbKozEZFzw1/zc2yOoderVq4drcC+wML7NYB6qgxo0bIj1nJGlS5diLYtlzpwfVLHEjhtlZ7CmaLSmbdu5Xb1vqKbh/PyPKicvdNEQf+vIkVSkSHcBMfblF1/ielJuOLixf+CRr8TMv9+HDdg7dxArxYsV14ixI0cPaaZ6MqRWHmCu/P0DpSvAqF+/vgLPzHI6UZOxZFTWubNnpQAcpEtA6RL7OA7ETbKSJmQZoyKj4Nb8ncydN1elYucunJV79+4ifuyZRlyzd+zY8T/l2JHDcvLkScjPHikRw1iw3jAE7gwZW6VKVSQQZsHBwcH6EJKdKwT5VX2YIzVEpNk19Nw9gYpHb9K/35j52SBgEDAIGAQMAgYBg4BBwCBgEDAIGAT+tgiQaLABEUHGgB4ptMMAV6DkATsw6GXSPKSFXI+4hgSh8TJ61CiNnU4C2RGfEC8QofBkJQkcSbZo8Z6jhANHZXtMQ9SxZ8LCoApZDtHAPlm1ZrV8OGWKuLu5Yy6qcHJAlqTLWdS1VIewdcYiZrg2kfCLlzSFmGvlcfbMGVX3bNu+Td6a/DairEdA8eEGj9VnWk9TscI18XR7pBqx1UrNdPk6DnfYerRq0QIdIndAXMyRdevXILr6W5kJX9WiRYtiL+xssdiH46F/SpHAIlqHK1a4niTIOZBSgRBEcP+2EEiQVBmBSG0mQI0bP05effVV7WJh29DNGzeUyyCxw4N+OFSp2CNOm/snEcL1t2nTBkRQInCeIAfgGzsC+1qydLFUrVL1P8wfjVjyqhBccJzsTOwSayX29N3xLYgWLyyd9/L9999H69ZAGQDS6CTqfqZkaWsTRrscHq7KqLz5nz1PVp8Z4sQl8XB2dtbuG6Y+f/zJRxIG3K/Ar0aP3HuhIFuv/I99zXdiRtuZcjfEh4+/Z4DlovrkPiLCSKq0hYKlCdqPCJDF8jHa6qhs3rxZzqLfKykpmVSkuLi56IPN+C22GrXv0FFatmgpFStVFB9Ej5EVTMAHZwUkXyRplAzCB8EFzNljKGbOIa6bPYR/7P0D9xFsGZQwD9A6xZtBJo8tUsxI//TTz8BwtsTDmgUZlC0etoJy7sIFkDvHYBRsKX3+x+6ImcggYBAwCBgEDAIGAYOAQcAgYBAwCBgE/r8RIP/AOpMH61HWeXk1N18j+ZCBNqGePXvKLrQ3ffbZ5/IdTGZLlipJLgf/WyRGKurYCuUr6LU0mCVBw3cLF/bXepYKGda5b7z+OiK5P5S3334HKUKbUKtCoZE7IcmIq1evqu8J6189sEDWmTT2tU6zkQKoT9083EASjZY538+RFiBZlq9Y+Ze/jQ1qa7Yf8fD09BJ3eLngJW3R4atBQUFokaqpHi/0nfnxhx/V8LY3rELYjlSuPNp5sH6SM3Hx8drKxZ+1fketzISjFPis0nLE0pBgVLxPD1eKIsaOGaserrt375KWUAfx4Pz0ieG6mIhUIijYImWwKdbVTK8qU7qMCiLYRUP/WCYlB8Jj58jhwxjHqsHJuqRArTNt+jSINOKVhOHaMiHayMY/rpHx5XwN/8uePXulWUgI1psq9Kft3q2bOJBAwzi6HxolY/5EEELFihbTNVFFw4PtTSVLlpJaGoueI3fv3YNgpJtaqGg/Gk/CHP/TR+6TkX/TEixuhH1kjo40BELvWi6LxjuXCWmSPx7k5s1CEJcNx2vIpMgqFizkqw8mVSrPnz3FA+GK9qViiMsuAnlZeTx47iqrKl2mtNSrX0+mfDhFhg8fhqSnauKB65nWxAe9MNypy6KHzMXFGWzhbciUUtXMl310Dx/CZGjjBjhQX5cE3HA+sK7urpoGVRdjcgzKocqXKytVKlfGA+9tPVj5B48ZySBgEDAIGAQMAgYBg4BBwCBgEDAIGAT+BxCgf4kbgmBYrOfQS4ZMAo56detJ+44dNFRm+9Zt8IsZJ43h7bIKHinNUfCTyGFdm4VUJ9avNAFme5OlDrERN0Q770J6UUF4oPK8BHi+/ALPmtatW0nlSpURFb0DxX8w6lN7nZteMyh7cVjzcy12DvZ8QV+3wZs1qleTPn366Rnr1q2T6dOnI83JCwlOq6VP3746P0ttHlE3IzVmmwlJuje87utbUA4dOiilS5VW8oIihq3YGw15mU60Y9t2CS4RrElTZIO4L1WW4GfOHw9ixcXFTWtqFSxA/UIVDb1d6NW6/8B+eLGuk4iIG7Jy+QqQUG+DFLLaswgr047GjxsLwgttW7l0k7ubh1y5ekUju7l/4k8FUfOQ5rJ33x+yZcsW+NEGiLc3UqswBlU96cCK8/NQ8gq8AnapvAH3z6Chzh07qS1JHyh4YmJiZOY3X8t3s2crB0H+wSKWmEqViajsxtqiBVJC57cH7vdAxpRBzc+D59+/f1c6dGivLVKpKSmKKd9TLieX0OHvL/LIV2JG2Sl94HJAujiJA0x+qWLJAjHDh4GMoj1+J+gEi/IuMlle3l6I0W4Dn5iueHBqQ8nip21M3ni9RHCQ9IbzdUX0BTLumtFkdKcOQspTgwYNpVWrljL5rUnSuk1rGDNF6YfHwdFOk5pOoI2JzF1h9Mv544Y7ujjBUDhaYu7EoNfwAdqVrsil8Etq6MQPHW+Kg6OjlC5ZWibDqbtnzx7qc/Mib4AZ2yBgEDAIGAQMAgYBg4BBwCBgEDAIGATyDwHWmmq0i7rTDwIAkjNag6K9hq97+3irAIB+pixfd+7YpfVoXFws0nz3S1N0d+TV4ytWrpDLaHUhsWC9hh9Qxz56/Ej/kM+Cn4cdWnhC0VrTNKSZXAi/ICNHjpIJr02E4MBF37fWpD/Kc8REBwcFWfwF2AaOGxgAFQ9an5wgMKDg4PLlK9IJBMSfocdlxfLlMnL4CKzdInNCT4bC4+UIiCVQFjo/12AjiTDH9fUrpEQGCSNbiBPuoval+oaJUbXr1pLP0C0SEBigC7FIDCwDbT43rkdI3bp1kG4EM19cS5NeCijoqdMQHi/cfxzEDcOGDUUC1CxtkVq2bMVf86elpsvPIKcePXms83OCbFyUgu4Xy1MGLwA3HkkwMx6C+OpPPv5EvXZ++P5HqVCuPN4BSrhn+F/nZ9x5yZIl1XdWiR2sie1NhQP8lHCJioqWk6En0N70nowbO169aL1wb8k/6IGBFvw6XxVLxJ9GwbzhaTAmdsczwf0TJ07KIPLJkyfJyNGjLCKHa+AgeF95Dh3wxX3JV2JGN2WtHsQIWEUAScKD3+mu7IX8djc4XUffitZ4a0Z06SZhzMObXjy4hDKV3Xp0kw6dOsLUtzIcsL3wQNnjw+OFTPWdAHaRbN+6XcIvh4NYuYY47ocS6F9EypUph/irkxJx4wbkTgJiyEEfKLJ4Bb190MNWTSpVrIgbH6ByJhoLxyckIYXpBpyaQ+XBvfsSVCoYKU2REnU7SomcoBJBYN2sD9KLuwVmZIOAQcAgYBAwCBgEDAIGAYOAQcAgYBDILwSsgtsqdffs3aftOGxlgTRAp7CBMTCDZLwKeqMeBVmDsJl7sN1gpPZsKC/o79Kvf1/UjUW1aI+IiFDFBi8G5QOrjkw1rmVXBtUcPNjp0aF9e3RnXJN30dJ08MAhmTr1c5jeroDyBZ0YnB/EAo+w02elYoWKkpaJtCRej2UxZYi+LfRozcK5tnY0rk2SQQMHy3vvvid9+vWRNjASLo/reImaFKPw1V1hXNp2kNBxxDiQwmj9TcVN334D5MnjJzLlgw/kDEyKhw0bJmtXr5NyZctJFldPEgPLYkuPP9KU2ZrFtiYyMWwPevIkzmrnwu8ktVJBwHw1bTrEE73hz1pPmjZrinarWorsJXi8PIQAgvOT2HkGTxg7dMloghLZp9z9VyxfCZ4+TvLdD3NUJFGhUgVtJWLSMvfG/asBMNAhGUTyiZ025BtsQE49S3ouzxEulMupQMmzEYKNVuh+cVLfnyqVKilRRqyjb9+RCJBO2sqE+Wlw7IAEaM5D4ohjcP/FS5TQ1Kjly5bLVtihHAchlvu44BtPwgUv8LCejHydAAvG/wTtJiRWJ0+fVgWMK5hC9vmxj+7K1cuyG9KvsFOnNULMDioapiuRqeJ1nh6eEhQcJA1gqERz4GNHjugNK4B2prSMNGS9h6sr83Iwh9OnfynzwYJF3LgOdYsLss/hCYMbqcqX0qXlMYyWrly5ClYsVTPmGevVsFEDadSoobJkfEiYFPUM1xUrWlxOY72/zpsv8YnxGiOmNypf8TGDGQQMAgYBg4BBwCBgEDAIGAQMAgYBg8CLREDrONSlt29Hy5q1q7VGRB2uwgCqQKiacIM4gEIBttjYwVCXRMY777wj8+b+Ar+WGkhp+h3kQ31dJv1T2fpjA9VFIlqX/AoXwusYkP/wf3BwsFSpUgVdI07wNklC688+EDUd0N7UGoRIGFqMSqnqhIMlPU1UHxkG4+iB65kozPYjphuzGYd1Metkqltmfj1Tpkz5UI1uV69aId3gqZJ3KLGC66OiotTWI0sVNKyrBanDPhprXbRYEUkDIXEbAom6deuChCgmR48dVa8YzgN+BsRFtkz78ktt86GEhiRMVmYGDI+zkDwFC5LcWp2bKAISau26tUg9fgX1fKq2cY0ZM0aXlEk/H6yHw9LflYE7rLd5EH920tSsVUOGDB2irUbshmnTqo1cunxJDh48iCjsUSBY0HnDAXCs37BeSZK8+bOzM7DcbPAHaSoG4WlUu1yAR2yPHj3kELiDwYOHoD1tvJI/vDkW18D7bBkIBwT6675433jw/peASGQSumZ4xMU+UaPjRQsXaLub8jLWF33/RXx5AcQMHiNiiF0/hrzrAByqW6NNidFX5xHz9QBxZenoO7t35666PPPhc3R01PQkGgDfQ39XKj4kbHFiLBn7927H3IVrdbjUqFlLWiBzvDQIF05CY6Ad23aAiZwq6xB11rRpMyQsBWpEVg6spN3cXCUJN3r3nj3qTB0eflGdr9mXxmQokkRUztDgiDKpRJyrsjGYMyUjx16fpheBuhnTIGAQMAgYBAwCBgGDgEHAIGAQMAgYBF4YAmXKlFGlBuvphIRE1I9WFU4ihrUffWPsIRzQ10HU+Pn6Sf8BA7R4P3b0uMz5bo52Ke3Zs1uTff/yY0Edyq6L/v0HaF2Lcl/34AzbjNOnT0mlCpZZMAmIIyAJGjSsr4KFTRs3SkBAAHxcrI6MeBjwZsM417paVPHiiPqUa8pBRwnrYTekFA8aOEBNd48dOwZxw2605VxDhPZaeffdd/Vazs+dxYJMaN++Hfxak/V3EhsUIZw6fRJKkHJ6LtcUGRmp5Mwe1MirYC5Mb1X6vfKiDOBCU2TOz/+egjiiCkX3jjn4mh3WFRISAp/WGkhxOof2rZOyBiTNz7/8JL9CMEE8qU7h/CnwaylbtozEQnWjFAFe4+tcB31eSKjwfNqPUIHzxRdfyPRpX8EAeay2P3Ff7IBRxRLHxH/pqeka5kP+iV0yqsLBmJWguqkJMm3CuHFo29oiQxAVvnfvHtT/rnp/8nDmvU9D5w4Tnrhntm2Rv3BxcpGVK1eKKzpmNOob92YkjJgnThyvPjZciwKro+X/l3wlZvJ6rxjdRfIlPS1DyRXsVYpBkrUf/Xo7tu+AW3OQBEIWxt46EjO4u6qI2bBhg3wFk6MT6BN7CpbRBh8QIIWbxxzybCVQgoKCpVatmkqgODk7astRseLFtR+PTtNFwcg9Q1b58+cpmi/vit4xlYVBvXMHZJA3DH35IbDBAx8MQyYygTaQrtH8iDe8AZQ0HZH+lIl1kbXM21P+Q29GNAgYBAwCBgGDgEHAIGAQMAgYBAwCBoEXgUAX+JcyWpnhLjxyhST6c2R0lKo+WKSrHypqQ190VpAgCEKNSHbm0aNHmia8EuTF8mXL5KMPPkR7ka3yOLHoyti7dy/IkFicqiU7alZbCAcStNZlDUkrDxIQ10CkDBw4EBHRv0tQcJDMnTsXCUwwu809rFYoKD4gYqAh7fPUFJ2fdSpJnKcQDLCThGoPRkkzsvrLadM0BWrhokXwZnXH5qxRVq1crdYenJ/r4XcG3vj6+ur+WV+TGKEPDBUuTJIqByLpd6yNih7lSbCubIgceB7DcYqj1qY9CQfI238qDHrr1KmtPEUmiJxZX38jgwYNls6dusjeP/5QbDk/j19+/sVSvJBWwfxcEzEndkxL5s8kXJ4/eyZTv/xCBg8dghTmCko80dtGD1zD+Xm9KnLwIr1msqhg4l6xEheILQIDi2i60x9Yw8SJr8EgOUjOnz2vRJItPIDYIMVjxeqVcvHiRf2ZpAxJMN7Y23di9N5Q5MH9k4OYh26anlDinD55SvHRi17Al3wlZrh43Su+UyqVAekTN7hzx06JjIpUN2dKpPwQU10ODx0VL0+ePEGPWLLGYD1Lfq69chER1yUWxktkrchYcVB+aGiGRCippOFhB5aTDNy7cIRuhwhuGhvx5gBFdXMm2VIc8ViMwqoDI6NiRYoiJtsTpFG63lS6QdPb5hF8ahLwIUpNSYOnjSduXDN5hgeDCVHmMAgYBAwCBgGDgEHAIGAQMAgYBAwCBoGXCAHUoGzV+eyzz2T16tWW+S8qScsoVyT02HF0UHhIUuJTJQVYx7KLg60+derUQQnLliXUlKgbx40fr//6D+gvVatVleo1quMaUWImIQH+J7mwMM2HxAHThfk+PVI8EGk9evRoTSyaNetbJWn69e0nZ0+f0TQiy3LWGoDpQu4gF0gIWUW1peag9wvFBiQVtN7Gu5998okMHjRIBQudOneWVq1b65xhZ8LU44Xz8wWmQZEcolCBY5IA4VGrek2tv3+dv0B2IUGqPhKKQ0NDQQA1UCVR3v5Joly/HiF37t7V+fVyDMG9xty+rWOqOgb1+bLff5eer/RUhYsG6eBnHg8ePkA72W2W6LotJY1QsmciMckZ9T4VMTyKQrjBgJ9t8Hd5883JSgqxo4aJVE7kBHA98eWxH8QLcSY/wIHJD9DI+Q8kPdHUlyoY3v/OwObCxQvwoFmPFKm31FqF1yfGJ2hHDgfluTyUMMKarBQva00cswa4hFOnw6QDUrzyztMf8vlLvhIzulDsi8ZJGzdvApmSoezTLTzgN6BGCUFEdkMkKVE2RgbQDmZJT6FuSXqaoIRJqzYtkczURR88EiwkXuy1v8wORE0cQL2Ifj08/HhANLIMLtQ2ACsQhEtpyLPolk32jn1sXoguc3d3U+KlNGRsNGJq266d1KldRypVroReOx/t0WNcVkJinETdipJTYafkPkyAOTc/YJmQL3EucxgEDAIGAYOAQcAgYBAwCBgEDAIGAYPAS4IAiu1Tp05JR6QaVa1aVQ4fPqzepVR88KA/CX1KE5ISLGIANSzbeNj14QPljLYSoc5kzRoIP5KFCxciyrqPONg6yJatWxEh3RWjWOoPNZXFb9EgVgr7F1bCgSUkFSOODo5KjpQrV06FBkxyomAgNu6xkETp2atHLlkAFQlq0AULFiqJkzc/X8uEQTBrXyU0lBkR2HcEyIaNm5Ai3FOSIShYsnQpEqBe49aUcGFMOAmH2Cex6FwppgQT38vbf1BwcXn9jddVGUOxRMVKlaBqCdV47YkTmSTloYQHr9mxc5u2ZOn1aD7ifrmWDLZhYY68tiV7EFtstxoM5UxY2Bn59ptvZNasWbzsr/hrEitUoWC3ygXQdFcLb+DvV7iwjB07TvcaH5coM2bOlO/mzIFi5ReZMWO61vj0wuERfumSHDp8JHcsLhXj4nWaJdNUmPunRQl9Z+hZ8w3WMg0qoy7gGqjSwemWiASTK1mF34ljnucQGmd0/8S5V6+eGmJEoUgKunL0Yi4in498J2a4vqeJiWDMlsHrJUNGjRktAUUCSUFJAc8CGgNGwoVMVO3atfUBTc+AHwxIlBroU3t/yhRp1rSJghsVdVPZrCya86IH7NatWzAQslQs9Jnx9fP9t75A3OAsPBxASk2bgoOCwGhWl0L+fuo58xBMnZuHm6Y31a1TV/PluQa2UiU/Q1Y5/mNf3hNIqtIgH2P/nwfON4dBwCBgEDAIGAQMAgYBg4BBwCBgEDAIvFwI2IOE2bFzu3Tr0g3dG1HC4Bga+lIZQ+IkIRGkDJQx/DO8HQiDRNSwThAF0PeExAFPIiHSv98gbQU6f+68MKb6wL79smnTJnn7rbfEs4CnjoW6XhUkrGfvwVPVao8RsXe0l+NQ5wTCLDfvD/7nzp2THj17wZB4jcyfP19CmjdH3Yl2JBwkFtTPBQbDHJMJQm6uBSwCgq08dnxVYCjcSmvnmzduwmt1vSxZvFhmffut/PTzz1KosB9ICrZS2UhSchJiqCtIHKw8/v2RBRbjIvxXtTUJb9y/d1fbrWbOnCEfomXr408+hKLHXy8hSXEXihkCpYoV4EIFTjZMfvkaSRCuip0nJEGokFm0+Df56KMP4dEyUj1hGXnNI0/xc/5CuNqKpMNbVokdDMS2raVLl6hZMfFPQRsVW9HGjB6vhAr5hWLFi+k4xPLGjQidn/U/56cVibZPkfzhwvCPIowe3XpotDcTt2ikHBYWBtVTTYuk4oVgaXCJ+gaxc4a+OHn3ny1g27buVI9bTYbCO7zkRRz5Tsxwoe54sFxh2uuEh75OrdpSqlRJNdklC8VUJYLEByUAypZUkCDsJ6OvC6+hK7MnIrJ5007ASOjg4UO5ztA0LyJTiAcAkwSC7CkZVFJ8EKPNNqlnz5PFGZnv/ADyffaJ0eDXCx8WPuCLFi6SjRs2ScTNCPWTYU8aP5QEPg5kTEGfgtK4YWNl8+hPQ9LGHfI2sqTmMAgYBAwCBgGDgEHAIGAQMAgYBAwCBoGXBwEmAqPslPNoZRkFguDHH3+QBg0aQP2yQGs9vsd/qkRB7fgAMc/l4Vma/NxKSlLfEZxw4cI5VduwxswGozFixDCZPHmyvDHpdfkF6U2FYJzLofgH/3Xr1snl8Es6MOfPwvlM/2XtycnyyJl7IDqmvD9FRo0cJZVhvnvo0GElf3DSXwfpBtaxJUsGgbwg1QBSAIwP6236zjQLaaJ1L2mkKVPel2FDh0lrREbT0Ld8+XJaD5NMWL5yBbxUr2NukD1a2+qmJSoyCmNxNHITtkpMMSq8Y+dO2uUy69tZUBtVwbs4H6fp/LlXeHgU0FAdvgZ5kb7PBCiG6/gVKoz2KVuhNw99frxQr2/ZugWJSd15th7h6IS5HX0bPMDzXExAjmBtN2/ehJjD8t/h2khMrYYfTOs2rbWbhnh1REtRXo3O8UiqcH300gkKClKMuFfeP+UVQGZVrlJZduzYIevQ0sSUrp34eRzUOWoerCuyUaUSCajEpKTcLVv8w+PHD6Hm8dP9/7UBvSZ/v7wQ1qEIWos++exTOCNXwtqzVU5kDxlXBhgx9unRPCgJOeuUTSkYYLzKVyyvjsl0rg4OKiHBpUuqJw2ZMro424GUqQmSxwukTSoe7nj0hfEBdQKzdhftRxyvBpgvL0Sc8QNgsWS4QzQOxu8858SJ4zAj2itX0PJ0KypaovEw4EnQ9Xh5e6F/rY/e/MuXwiUO/YLKVuYv3mY0g4BBwCBgEDAIGAQMAgYBg4BBwCBgEHjBCORpG1q0aK5CgJkzv5Zdu3ZJ+3btZR88SmrVrqUFPskSFveMz2atGonQGHINORAKgMnAH/HjsFJUl2ABWH9molNjzpzv0drUF6RJaRmIJKe28DvlceXKFYlHHckCnmEzmfBAIUmBkfR6vKo/V0adzLQgEjmrVq3SeOiz589JO1hvOLIVJwurRysVeBW5BKInRckiXIuFUo2TgY6TyMgonOusxAaJld8R7d2rVy9t05owfoJ0695D13QLlh23o6Pxs+WnQsKC66EfKw8lprBgH9+CatobejxUunbtJvcfPkTk90EZMXwEavACOBPXYy+c/+zZcyCk/PCaZTDM91ywH/rDMJqbhJAdEpMOIPynX79+Qq+Y3xYtlsaNGuvcTH6iH0yWRmvzTqE1Cuth/U2z5ixYllj428CUuDRq90isqatam6xBItVnn0/VDhfOy4P7v3LtKjx3aqFFialSFo3k4uyk7Uzl0UrG49HDR0pgLUJr2qzZs2QxlEZqs4JhGAvO9bNdibhzfqZS8V46gVjj/VciSkfK/y/5SszkMYD/i72zgLOqCrv+HmDo7maG7i6lEVAkFCkDJBRMDFRQsQMTLOzCIkS6S7oFJKUZumvonG/9nzN3HHEQeN/h+31+7g0z98695+yzz3P2ufc866y1HoZZoXx5V1FGOQQJ2VGBAgWEnmW3icZOwlTZsWOnyySvl+w5croj0pFhTITpbsaMWVzVSlXEYKlhRkQVVYUpa+ZMooqlN1+YfQJqPhM6OXHiRLdh3QbJnQ7Kffmk6c5gukBjInJQwnJkzxlbTitAzdANHte2t+/Y7qZPn2ETIJ9Mhgg4fjdMjg3rN7md23a4ExwI33wEfAR8BHwEfAR8BHwEfAR8BHwEfAR8BP5VEUCBQdqeV8VgmspbhLZOueONKil9SCqNoZIAde7c2XJQF6O0WPkgbJPff19qCTiqIYxpSdjPCShhARJ+YTSWqEMoqFOntjEtAHwefPBBU3DE5cTKRynLTKlqrWIN2Q895VL1ILxYGOMu+bPe3LipmzR+ohvw40/uhReeV4WhVOIXgA7FuIkq1/3bYo1JfxtTRB0g1oFNAxMHMCNkYLts+TKBO43NX+eZp582n5f0UpAgFwqaPGI0gvMaB2oV+rPxav9LFCvu7hHrBpXKzl073asiWnz2qfxdJG96t897qoacI9h/dfXH6pVuqoAVYsL2iTM/R6KPGBvJCA56gdfwm71PFaB6Pd/LlStXzg0Z+ovLKlwA81/235bV9qnKnD1rNoJOpCncbIa+9erXc8WKlxQx46AbO3aMe+yRx1ynDne7H374Xv2VDfZfWzqgokKDfx5seAL7ZJHWWE/pGKQTC4dtMaJdYkb1eu55SdTucq1atXK1atVykRGFjGGDPOuCjrfFWj3s2L7d2DIcf16z4x8KpXpLzJaowEywsxweEMdT8oPZZTIhJEHFSxR3VatVE4umpEsu9Cy59HvhyUQP0vP0AlB27d5pWrU8kij9vmypATdp0qeTM3OEa9W6lWskFDK9XK3PnFZFJfUP1WzuvHlu2ozpbsasGW792vVu3779QgjPuFMCXzgQgDSly5QyZ+1MAnZA8ZBQUWqM98+ckXmPnucUNQma1CTVOc+ePbtKZKW3SXUKl2fffAR8BHwEfAR8BHwEfAR8BHwEfAR8BHwE/lURCLFL9u/fJ9PeoFRykiQxbrEq7HTu1Mk8Z/p93M+1U7WlVGmoBCw5jJQZeM2QKyL7AZjBTFZZuQEYMUokYYwUkMKjSZMmKjxzxg0cNFCJ/nNicbzs+n3UzxURKQEFR5JkSaz0dZEihd0FisroH7AIDQbGhg0bzGcVvCBqa5R7vPvjrsczT7vuj3d3ffr2MUkPeS8sHqRPgA2BFClk4ivmC3ItLQQ0Qr7bumUrA2m+/ba/e+vdt13rNm3cQEmZKIDD9lV7R0uGuS1i0fAa6xrgoP3KIfPd+YsWKjeWt4xePy7lymuvvuqoInVbixbyf/le/jUCTtQAeraJyADXxfoQGePosaMuPGm4me4aWKS8O52YNnfddafl/B999JGIFeNdoYKFxFwabywjxsT2MQHeLhCkjRQsEClgJglJEXiUSjn+PletalUDViiV/fW337hmzZsZY2f2nLkuT948OAJZo6oWlaiCWOGxozEJaDKOEEOVdw/mv+11zCl3DuhDwaIx48aYx07S8HDrx8alZ8dE3KhVq7ZVfDYAC/RE/VyLlqjATDBAOzzu5PFTbsDAge6pnj1F0RoiVouqKWmBzBkzu3B5u2TNnFWgCQhXjNUaPyvEjOfp02eU/mus+1W0p+QKTLgmbXjS5Fa1KXmqFKrgFG10o0aNGqkUdh4rk3XkcLSQuCi3VvSl35cudevXrXVbJFViMqWV/i1Pnjwuox4zZchsyOBJoWYH9h2wyWP0KY0LudN+ATtZsmQVqprHjRw5UvSwjdci5r5PHwEfAR8BHwEfAR8BHwEfAR8BHwEfAR+BaxiBpCAeSkCppkOF3xhjvQQb3C2ZTq9evYy1UrNmbTOdLVq0qEl1LO8GfdH/8wIIihQtErBqWDWWHZIjRy5XXAyT1PI4vaD+31EFobvb3W2EgMGS2lx//XXanuw19DNgwAAXhYWGsIYAllHXAn0oIU3JbkMl9PuYit3gf3PDDfVd3tz53JuqRHRdjaCiMZsGGBBUxFONJ5nLJmNa2CWwPJAC5c+bX0V38qhATnYwDTda+WybNm2tSvGgwT+7Nq3aaCnbO6tqdODgAWOAxLCvegdD3y1boiwXNqxEsqVTkndBXkAJg4dOq9tuc23btg08eqw3bEOC7R+UqqWMPGnOx4JYSQXWZM6URV6umV1RxQp/HHJ2qltRTRkz5p49ehJS2/5BHaPfflskYOywhqMx8V9jIC7IjSCBwKJhe8uWLXf16tZz48SgqVWzpgyCu7qsyuOD6HCYBIzpj6PRx+R3W8gYQqHgYzKcTTGqUaOGmytgZ+nSZcr9R7hPxQ76uF8/SakkJTsHCBbIxiZOnGDjNmZPMFgGnOjtGgAzYSZRmi4mS7SMc5bJdXqYTHbQkCExwkyJCkunxKjJoskUo8BSqx0ZEW7Qu3ZJQiQToMOHJW3ShMWg9+ChA9ZntOROGAoheSovqVT16teL1lTc6slDMaOiEt4zGzduclN/nerGSiO2XyXJ8LSBEJU8RbicmcOFeJ1yK0W/Gj16rFu3dp2NE70c/XK6oLvjZIW65JuPgI+Aj4CPgI+Aj4CPgI+Aj4CPgI+Aj8C/LAJKrGlYa5yXf4mxXWKBCUx+8+bNK2PZwSqj/JqYJMncaOWOtwt0IP9EzsJ6mI1gkovagwaIQfKfWlIjvEspp6xF9LpzE5TAY1KLauRnASEVZMcBmADjY50IBGw/JCkiyQcwAgwh/2SkeVXWGu+Z+Spb/eBDD4hRs9GNGj7SPfHEE8asMcYG1BK1ZcuWads5rcS3IRjqIZ3UJovEeCmhfWNALD9//jzXWGa5s1RQh/LgReW1gmIlZFQMSUELa+0wKxfNehTUMRaN9kkFw82ceOu2reYVs1YqFSo/vfb6a2YjwjasKpNihYwJWdDpk6eC7avfVOprg3Lz66+rbvgTDJ+1a9eahOxtgVmvvfaa5EhljDxBX1MmT7F4GYtGYzp76ozJrcj1A8Nf7b9iihrnlHL6Tp07y0NmhfX35Vdfqux3SYuPHRQFlX6QckUD9mhV9hRTZvAHKi7RKEb04gsvuq733Scz5i5unlQ5adKmDvZLMViyZIkVG2LZ0PHneWK34MgmVq8aOLu7Rzv6zjtvG+2rXfv2cr+uAZhnE/zEiWOaML9Z2axZqiefXOY+GCJhtrT6jzVmyHRY9LH90ohtF2Xr6HGBOPKEOSAK2g5NckqcVVG56wwy+U2VOrUxZgpGRBrF6xehkzpOLmOmTFbqbJcMf5E3oSs7deKUyZgovY2LNYwZTIA3btrkNm/eZJWkKlep4g7KkPiU6E8csCT68c1HwEfAR8BHwEfAR8BHwEfAR8BHwEfAR+BfFAHlpbKqtQFnyJBOjItw/c2/AKipWLGCZD6t+UtAwTrX7ZFH3ITxE9znn3/hXn+jt3m3AMGQyM+aPUvAxmw9B6cJQBQ8XWB4ZFJOSv5LEppMOqGjR466W269xdgXTW9u6t6TwWyBAvltO9aBwAe2CRhAJWMDf+hBAytRvIRr1OhG3pZMaKuqCn/ler/5pspOP+8GDRws4EUGvOyA2mblsBgHn5Ydh3mi6DXy4GNiiGRVlSiIHUkl3yLPpmpxx46dXK9ne7l8AqN++PFHgS2lrC+GAyDC9rdt3SawIpvWhcIijxcBW6nTpnX169d3JUuUspLb8wQaPSC/mI7yxxk2bJgrUaJE3PYBed6RwfL2ndvjts/YIEWgSsGfJfDC0XHQ9t4WXtC6dRsDTkaOGuVq1lSVKTXkShqUlnXupPLydLIz4VjCMuInXOyZWjIRbtiwkZk6r169yj30cDd5+RRxUydPNXnSWYFeoePfv/83Alfk0aPtq0sDx5AvocwhCIBn57XPgwcN1rr1jaXTqWNH+f7cZAbGjOnsmXO2TDAjeCXxW+ICMzbbwsxV+rT0duxscxktPdvrGVej5vVCmqIM+DgusGXb9q2iRU12kQUjXcmSpdzunbsNyUSPR/UmJhtSqHFjJ7g/1qw1zxkOYAbp1Ah6PsmN8snxOU3qNG7rtm3mvgwaCZhSXDS0IkWK2mT/Y80at1esGYKoIylqVLSbPm2aW79hnfxodECEUrLNlHK0riSKFnXNkUJBicok2pVvPgI+Aj4CPgI+Aj4CPgI+Aj4CPgI+Aj4C/6IIKG9UDm+Nm/S5c+fSczJCJedKKimjzI16mC/Gatm7xz3+2GPu4W7d5PHyhECM560aMKQZAIe9eh+wwJgcSitPiDiQLDypSWwANwARMBm+reVtymXPmmEuJaIbCmj5UYa+NSRJCgEoDIrqwHVq1TF/E+ACkWkkOUojosIf8k6Rx4uACIrl9H2vj2ssgKBWrZpuuICQTKokDIEA5g0GvYAobB8QCYIBbB/yYfFYDHjJkDGTmC53yVclhZgu/dyKlSuMKTRIIETLFi2DmKiDMP2sE5PlBpXbPnHitAErjDOlrEQOq8pUlSqVDdSIiTnvhogMUalyZZniZnMzRbTIly+/eeaw/DHl+RT4sR1STMi9k0vxQuxplpPrsWzZcsrZi7sRI4YbI2XP7l0CnwYaQMbAMScOlr8g75vsRrpgR8MUKCNQSOmSISMyMMgfMbIzWSIj5rpuqYybp0yZKn/aRi6F9pmGZ03IooTt40GTMmUqF4NcSWPk+BO//GIsAfy0atlGxI0t7v333ndvCKRLYWwpjhJTQHENhmZ9J+avxAVm2DONtIiAke7dn7ASVryURZWX0HHt3LZdk3CLyZSOS650TkZG7Gh6oX/Qo07IEyafAnJj45s1qZJJlnTIrVzxuw74DLFw9tskMvaLkEEaAYVqtWDBAndAZcw4GXao2tIJIZC8h+yJ6k0ff/yxy58/v8uWOYteT6nS2TsNGUV32LRpE9VUv83GCkiDFGqf+rrhhnpG27IN+V8+Aj4CPgI+Aj4CPgI+Aj4CPgI+Aj4CPgL/ogiQcDs3fvx4U1CQ8FsKTnKtN/YIbKFqMMCGzGUs6f6uf3/XsEF9Jf3p3YNiYTRr2ixWbsRuA8zoQetSSSkysqCBApapC5nJrUpLmaTcSC8iAWDOGhEEmt3cxJQdkydNcbfe2sL8UVh+3bp1lv9i50GnNgRl5qhG8ufLbWMJE+OFN+bIB+X660VykP/pDTc0cF26dpFESGbFrKTBABYY40W5dsFCBY3sYO8o102fJq1JporJ44VlDwu8ue22Fio7vdz9PORn11N+sOwT9juweGbPnu22b4uK2z7xIm9PK+YM2wkxfKKiomQrcp2sQ8a5evXqSnr1kIro5GCzasQJ0EOsov0HzOMF0gbNSnUrfoWLFHaNb25srwHe3PfAg+av88EHH6is9le2LVFXbB+HDRuuXP+A7ZdtX+tjiEz+j9cOgBcIC3k8/jevvvqKVVt65plnXNHCRWwbdtz0jP05KYPnbNmzaZ+1njXFUAhcrlw5LM7btm52Y8eNca/1fsM9+eSTbppKhqeWUsdAGa3P/2vREheYscmhslaiF9155+1Gx0JKhOkOAEzqtGksoKd10AFhkBQBhuATU758OaOMpU+b3nXs0EFMmkJu4rgJbtXqNUZngkqFDgyX7E2bNutk2CNqUVKV1s5ofaMjy6EAM5n69u2rCbXFTH+ZBDu271RJslyOstulS5d2KeTMTOMARUZGyim6nat+XTWhjAd0cmZT9af0OqEyyqwo07WIue/TR8BHwEfAR8BHwEfAR8BHwEfAR8BHwEfgGkagnPJL7C/wUxmiMsqABYAZBhqIkYLvTKrU8lMR6wT/kpQCOyqLCTL112nu9d693VYZ4X777bfuOVVcSiH7DfxFrFqQxoxJbqvbWrqT8k+lPwCHFCIW/C5/1QoVKmojggCUv26K2qzy1Teqn69FGPjMDINz5oQRE2amwMdUNcgAFgELZ1QMR53J3yRggsDCIVeuXq26+0OWHz16PC3Vx3qHN8tXkjllpAR0LAhCH8eOHhETpawBFGwfD9UkypdXLF9uapMg1IHnzZ133OVefvkl9+yzz7j69eq5wqokpS7cRJX93rN3n8AOfFcDACTwdgGNEHSgBwCXChXKq7jPYdet20PmidNe9iUDfvpBqhUBIcRY2BfbPyPJEIDGdpEwaBwC1kdmha2Inlr8wQZefOklVUtqJ7bLTe6DDz8UQJXX1tkpUsWYMWM1Jv2p5QG9iO15xYvqVjQAl/yoacQ66i1A5eN+n6jQUGlVWxprbCUbVOxy7Nu6Nettu8FcCMaUQsQOvHHTyZMGZc333/W3ylEcL8yG69SpFxx/AnUNWuICMxwvohXbkilQZ4VmnTh53KotVapYSQBJbmnFTomeFAQ0XJOlmEyInnn6GZUtu8eVr1jefGOYhFFC6+bLfOecpE0lS5bUckWtXjs12wFRkDHVkL6sk9bjJKIeOtStjRs3ms9NterVzcCpdJmSAoJOmtFRGqF94aI1UfMdbHLmjOlypV5pZbhPyocmryZAwcgIc3/mJPbNR8BHwEfAR8BHwEfAR8BHwEfAR8BHwEfg3xWBagI0PvrgQ6uqRHUlS1OVqpKt7j+wP6iIJACDnJDfkAAogZ1J8p/dKkqDh8vzzz/nXnjhhdiEPW0ovxfr4rR76623ZM+xQ70F+S+SGiROefPm0UsBipBMuSng0KOPP+Ye00/uPLndd/2/lRSokq3FOvZEiAWGtBkyZhCrRusaOyXMmDE1atZwlUQwOHBwv1u9arWxQdq0bmsSoLySEVkfGgVdvS/GyaqVK7U3Ks+tv6kMBTiSRn427CdgCfuaShKuV195VUDI3S6z1C0DVSGpcWMxWLQr5jETu/3jx8nj0xu4gqwrGSwWtSJFill56aPHjruVK1fZmMrLC3bO3Lnyiqlp+28gjJb9VIDUnHlzDUxJAtqjvgFEYAulSR2Mi9eQYf004CfXUoAX+X83yco6depk2zMASzImADBMk5OHpzCABzNh9h+wB0nVk089ZfjDkqW/uXvv7eL27Not4sYc1659O9vn4Fg7N3L0SIsfnQPyEBPaIcm2MoqgoYMgNk5SN2nSJLNmWbJksRuksujpAc2Cw23LJ+avRAVmgt1x0uvtE8K4TbSnYwoewddB1I5lzZrF5ciWQ0a++w3hypQps1gvcrjWgciouuslS1JhqZwAlmzuOtG1bhJaVrBwIa0OWyabqya61IgRo1SLfaAm3AqbZKnTpHaNGjZwd951lyRQN7pbmt/irqtWzahUacTQKRgZ6YrKbwZ37EUqv7Vh3XptLziojHfGrNmut9ygZ4lpk0JVm7LBmNHkOy1AyQMziTnVfF8+Aj4CPgI+Aj4CPgI+Aj4CPgI+Aj4C/xcioHxv6pQp8ky5wY0ZNUZVga4z2xPyP37mCUAoJSXF6dMnDSTAt4Qb+MvFLilTrqzlnyT8n6iE8s2SIxUrXsz9qv6ySqGRVOoQEvmt27e5I2KpkNZjkXFcrA9u/qcIJ79FthNmPid33n6nlWAePHCQ8tHfXEqRCyaMH+dubdFCawbj4XGFAJX69eqrz2jbPoSDZFJ64NkSGRkJYkKnBhZUEbCzU74s8xfOd+XLldc2A+bIAQFO5NoY3VLhmAI7eKnSPz42UFSQ7ZRW9SLy7tHywUHqhR3Iz/KOeezRx7SQWCkGdiSRj+tZKVQyGVhDNeUY1aummxOqsox9CJ0ClmwWMQLfmNlzZruRI0a4FpJtweixvs6dV/XlnXHbV9dqF8xMN2VKVcDSsSKGWJ80bdLUzV8w37308itusaohffnll+6jjz6yfvDBYf8BdSgMRIXnsyrsw7qAGqnFlqESdDodRwZ5WLYozZo3c33e7eP6qY+333pbHrTpbPywpIi3xUUx4R8sGeKWXCSOwOMmzKpxFS1azN0vw+M35TeTRMee7V+LlqjADDtG27R5s+vxdE/3y5BfzMvlsMpmg7bZLmihczo45yVkQyqEEdFe+cesWrFKOrsdVoGJQOXJlcf8X9oJcCkiDdr69evcmtV/SHe314I4cdJkt2n9RkPE8J1BPpVREiQQxfsfeMA0fAWEmqXWxMfEB5OffZqkmzZvEiUsubRuRVwxATZppBfbL33aYVGxkqdIJY1cPaOfnT0d0K6CPfK/fQR8BHwEfAR8BHwEfAR8BHwEfAR8BHwE/hURULK9detWB4vjhDxFZsyY6e6QfIey2LRt27Yb+wQfUxp5alKxQVBlRBQooOQbP5XgjQkTxqnS0q1u3foNrnqVagIvHg0MbVlJibyxMJRvRslLtUjhIuZnw1sm2VGOipdqRSlHwFWi5afSUAa742TZMfSXoa5Hz6fF/kDJgRTpmIxrJ7so5dKMyLavdY4eizYQ44LYPZSvpi1fsULKkNvdgB9+dE2bNZXZ8NtB1SZ1RIWhYGVnlZRKlCwhwoHKhbO+QBR2K5dy7bvb3608OamLVq7eQiyVTyT/eeutN01ylV0AlEVF/Y0YOdzye0EwLon6YP1zYpRQxYmBI+9KIhIGkqNOHTq611/v7Vq1aSOvl1ddRGSkltZiWi60ff6GcZQieQqtnkRgiBg0eg0CBybJ+OdwHAYPGuS6iPUC22WcwKOIyEK2fdbnbzCFc7ItQYkDYANgBl6QPp1YR9og8afa83NiPXXtep9rfNPNRuKookrMChDdEGV7JF7Rh6ODKlsc09jjnzN7ToFyRQWwpXTv9nnPnZTfDv1ei5aowAwRJ6jsX5QqMK1VdaOlckieOXOWuTePGzvO7dmzxw4ekJnmhaRHYZIZQYFaIWOj2ZqMU9yv0391x1RWO0P6jFb6GpRupRCt3m+8LkZLdqGC5eyEOXP2jLugA0LNdA4EGrr58xfoQB4UKplSrJns+slmJ8eGDRsMCMK4h8mTXmXTIgtGyIcmt9HW0oh5g3EQ72fIlMGdPnvanTsbMgS6FqH3ffoI+Aj4CPgI+Aj4CPgI+Aj4CPgI+Aj4CFyLCJBA75NfSt26dd1nn38mSc0n8gmpZeWrYT2MHDnCFB48R+Jz6owKzAh4wGuGhDaUgNeqVdvkQZSJXieyQG8BD99//70xNBi3gQDKadkWvib4z7A+3iwAQdhspEsPUwMJj8pHS5lxzz33uMcee9Q9/PBDYuTcbOsBE/wqfxuAEg3Jtn8ulhFiUIByWAOB1HuZMmXMuuN5yayGqlpT+7vbuzmS7FAqm2UZE9s/LQIDHinRYuEYZcVoM5axu6FDf4mz90DO0/PpHq5Tx86ujuLVr9/Htg115TYIkJo7Z46GBwATADu8TklqWmj/IT+QZ3/00YfuuV7PqZJTFa03V16y5QlHvO0LWFq+zHxqMOIl/mw/VZpUrv+330nulcf2HynRt5J9wb6hYtXPgwc5QCa6QgY1esxo23Zo++eFC8AQSi4VjPWp488/DJmH/DLE1atfV6+fdwsXLhCJ41Yru03M7Tir081bNruWLVsZkEf8aRmFC8yYPsOVkq0KYM0pWbLY4IK3E/V34gIztvMqaS2UqoLAk0KFCsmAJ62QwcNutdCrzz7/1IICvYsAXBBr5rz2msBRr5zyWlRtmjRxslGPymhiZ5CsaO+evRbkgwcOmSM0rtQPP9JNrtQ3uLyqtoQOkBrtiqtqxn/i7hMi9pxKdENlOnH8pBk77dy5yxWIiHBVq1V1x6WFA7TJlTO3JlcyMw/mBORAohUcOmSoSZ9iYaZEDbjvzEfAR8BHwEfAR8BHwEfAR8BHwEfAR8BH4FpGAGCCstEqj6wcr8eTPVyXLl1kcVFM+ehQK/8cMEvIs7WsGAOHdHM/8HiRZEesCcgpJPblJNGpUaOGSYwgC7Rq3drVqV1HIMpU8zkNMAeWjHHvvPOODICXacdI+OWHImAFTxTepYUek4lQ0E+Vg1u1bmU2GqNGjjKAxhbiF8CAtn/61BkrdmNlsYMu6NmsOjp27CACwim3fNky1/CGhno1Ria1y+ST01SAkAAUbZ/97/dxP7d+rew86FKsINZnA7t27zapju2rtkd+PmDgAJWLbmn58uPdH3f333e/rQdYBCuFxvqwXSz/5m8YOvqfQeDTyy+/rNeTKfdf6Vq3bqPHVZaT008GMysO4IdNmzaZbOyk5FDEX52q+zD56Bww1kr8+E+fPl3Feu5yM+fMshi9LnNfGrYp1lhf2z8lkIcKzEQZEIXHFJJKAVplkYXKtm3bVHJ8hORpn6mE+Q/uVdmZ5JEfkIE4WjZa7BpkVIcPH7bjHwORRAcBEChLlix0F4xV47wWLYhMIvccLm1d4aJFXJPGTVwtoZLIiHBVzpM7r4EsyeQrw4mwW7q4vbv3uHBVbOJAQR9jN8+IkgR7JWv2rGK2ZFA98t+tXnnhYkVM+gTLJl2adPKPKegKyYPmelGeSpUsJaAlp4CbdMbUGT5ypHvpxZfctJnTLLhMVACZ8GTh9lhOyN1d7e5yZzSZd+/ZrUlw0OhLC1V6e8CAAW7y5Eli9+xO5Mj47nwEfAR8BHwEfAR8BHwEfAR8BHwEfAR8BK5tBJRSx+bPDVRimoQdDxUS8ixiUCwQa6KFPF6Sm99pAAwcOnRQzJVSYpccVQIOLiJgQ8n56j9WyVg2rwgnwCphbuSoka5J0ybyOTnghv3yi8NkOLnyWa3i9u/b746r8A2N7Z89I6NamdoCNMDsIAdmWMVLFDePmwVSe5B7Llq4yI0dO8Z1e+RhmdRSmjmQVqHiyJwpYwCCgDWwEf2yQjrqKblUIwAwS6RSada8ufnpfP/D9yobfbssO2D+BPKprQIl2D4bBwCAe2Nj0T5Z01v4wdZW7r5q1SqrWsyYPvnsU/fee33FOsEQl30KwB6UK3ly57F9oVueZMqc2U3/dborVESSI4Fie5VLt2rVWv4u/dzrr71uxsT4w7JFAKNpAlwunINFpA70/zRsFMU7GdIu/U28iX/16tXcdvn5PNbtMWOvdH+iu3xnRyj3zxW3fRg3OySlqlu7rpXBZn3ij6wJxg/HiIas7cnu3d2jYivdceedbsCPAzONA0QAAEAASURBVAwYY/u0X6dNjTv+bP9Q9GHtS5hsU4RTECstF7tosEIi/k5UYMaCqsFlyZbFRUREuAyZM7qsmbMaXatBw4buiy+/UJmvHtLYVXBZhTqBHh6TcVCaVKlFK6vjIiIjXLgm7lmVLgumvVyfhXZBcYIGVq50WQE4GaUHXOOGjhhq9dfPCPTBXyadmDX5xJ5p3Pgmd0uLW1zt2rXNnOnwwcO2DXRv77z7jvtJjtPULccMidLYJ0+eckuXLFW99G/cj3rviNBAkDtAnBOq0uSbj4CPgI+Aj4CPgI+Aj4CPgI+Aj4CPgI/AvywCscwGPEtgTZDtk3tWUjXfH+XN8sMPP7iXX3nZiACxaIEq9s5WIr9OyTflooNE/LRIA8eOntArgZSIPHLx4sWuY6dO7tvvvnPN5PHybt93zRIj1I9FSutjnlu8eHGT6qi3uLezy56jrXxYYJeQE99x5x2u17O93PPPveg+/eRjVQpWZadYwGLU6DFicxwzkCK+omOTvGhSSeJjfigy5Y2KinJP9HjSdX+8uzFL3njzzYDpoe1CemD77BPAwjn5qWZStSi8WC2H15ulS5V2NzZqLI8XsU4EiBCjbg89bPKegQMGW/XkoIKRc/NUOZn8WqiQbR+0BwBq+aoVDl8W+uR9qhy98cYbrk3bNq55s2auWdOmkiOVNHQj2C7ra0TaPqBWnlz5LP8PxT88abikUBVc7Vp1ZVuSRISNpdZPvnz53Nhx41xjycCMHaPts/8xejx+9Lj2l2OlAkQyT6baVYGIAsYI4tidFDGj/7f9ZTTcRMc+vUmt6tWvFxybC6h6ZGei/Sdee3budrkFQCl8cfHn9WvREhWYMX2WRomUqXmz5i5XjlzS6KU09A2TX/RlXe7tKsOkR1wHGQNVrFTJ7VK1pKM6QbKqTBclyqg9flgHEArRcfnMUKccfVyY6FgwcXi+e+9u982XX7uPRf8aq4kKunlEy8PMSSen5XJlytlE7/lUT1e/fn1RvYq4zKr6tFmTd9bM2a6i6q5DSVq1erUrWrSoOTrvUkm0n376UY7UO+xg7T9wwE6kaxF036ePgI+Aj4CPgI+Aj4CPgI+Aj4CPgI+Aj8C1iwBMBxrWFjvEuMCzBCCBPPM++cU8IalOp86d3eOPP24MFhJ5qgphwQFYEBO7Pn2clzeJcnV7nV7xJz2iqj8vvviiGzx4sGvdsrWbOH6CVBylLV/Vgsa0AHTJmCGTuAaUrw5YM/SXMmUq5aUzJFPKYIABQAIGvig6yqnKUp8+fVwx+anQkP3MmjNL6AojNAxDv2NM+ZEihQx01S9sDox9UaD0/66/xvWCQJnMklX9LlCjjgyGA9NjgBgIEGsFPtWuXVPrBowVes4oJtFUybMKFIjQX2oCe7759ht39913m//Lp598alYl7D+ermNHjw72CRaQtn9O7CAK6JD/6w3bf/ABWCu/Tp0qT5kKVvkIaRIVldOpgnL87e+WF23T5k2NOEE8hJFYVatDinPZ8mUt/mABk1XC+vbb27oVkm0BrlWsWDHYvqoqDRowSIa/h237xIgqy9inpBCOwFjYT1g/efLmdX+IGYQ1CvOhf//v3LO9eplHD7tuwI6OPzK4kgKSomXAzPG/li1RgZnQQHUcDERJmTql1QuHOrRKE3yI6GNr168VuyWjixRqVapUKRn+zlG5sPEuSq7ZmBNxAp2Ta/RJUZkAW87o4OZV4GDOYJTE+1DBTqgcGejXqjVrRG3aIaft6W6MJsdBVVii1HVKUbdy5Mxhpc0q6GDdeFNjC2qYJthpJo3oVzh1d76nsxCzb91TPXq6CkLjAJBAEvft2+dOnghoaKH98o8+Aj4CPgI+Aj4CPgI+Aj4CPgI+Aj4CPgL/b0fAgBXjPASgyn55lYZLIhNqqC6+/Pobd0O9+lYs5ofvf3CNGjVS8h5jvi0sRwlsWrp0Qfllnhu4otcjIiLdYwJ0YGCgvmhx220CEmLc8BHDXbt27awwjVW60To/DvhRcqi1rE0X1vBn2S2z4IxSgxjjQ68CHEyZMsk1UNWm3bt2u4cffNB17NjRZEwnlfuybVqMlgOeoVJQiNUDvyNX3tyW2wLWrFmz1j39zLNu1qxZbujwocp1nzZGD9sgLFvl67py1R+SWgWViK1HvXfgwD4VxxFDxMCWGPO3WbRooRnwzhdLprLYRq/27m3SMMgVwUgAPMLc/kP7zeSYMeIfy2uZJW/qfG9nlzlLVvOSnTBhgoEpH334ocMrBjkUbB72g/x+3LgxsjvZbfsaAiqQOF04f8FUMgauaNn1kid1loFy79dfN2Dm7bffcenEfjkl5QtgWCDbSmJGymnTp9VY8NxhtAHYA3uneKmS8hU6pCrSa1RB6hX30EMPSUXztWEUkEOsPLeWHzZsqFvzxxpbN3QM7I9E/hXa30TpNqBWxVhJr7WafMnCZKwrSlNKTY5DMvL5/PPP3Zeff6EKS8sV+JNWCouZsW/fXrd923a3/6AOpkpe4xmTJlVaK+sFCQmzHZCubdu3uxw5csiVWbo7DrWCm0Tz87yCf/jwEbdclZ327tvjjggl27Nrj9st6hF9h0vqBCOHScZ45goJxciHfqFYVRCDpn379u7lV1+2MtsNGzV0JYqXlN4sqL2eKMHxnfgI+Aj4CPgI+Aj4CPgI+Aj4CPzLI/Drr79agmSVVq5yX9bohioSBK7nx4wZc5VrX93i3KEPF0ugnAqSXG1jnNwYzp49hxutG7+J3biL37dvX/M4wYohvSwZKlaq6J588kkpAQ5d1ebuf+B+uxmNLYRvf0aAPBEDXBpMkguyyoApEbQwV0IeLwUjI90q+cc88cSTbt78eW7E8JHu2Wd6yfYiuwETQkBsceZreoEzVBwiMQ8TSYDKvygtslAFSUScJUsWu/Z3tjOPF4rRvKBqScnF0qAdF4CxJWozPINYcEVEBEmckMwkSw6TRdvR/4gCkTIVritw5IDDePfXqdPce++/7z4UiBEZERG3fXWiZfab8oOy1RecQASBQtmzaEzbduj1IgYY4fEC2PDqy6+4Z599xgE+AU6B7wA8zBBz5eRpme8ySG3/mCRA7HNyERXYz6RJkrmGAqty585tjJMePXsI8FnjunS+xw2USTDnFiE6f0GxVWijDx1xdcTOITdnI8QfNct+ER6qVKmsTYQZ0eK555539957jwFYw1WKm5yb2DKIpUt+t2pTNkh1DuECphPHL6l8XkxJowFX0vnCedNH/jeTJk20ktaTJkyy+LE7rA/hCUVNsSJFjZQR7KgIJCpEdEZMmLJlymr88toR6IMvT8vbWrnIyILmfVO4cOG4Cs1RkkhFbd4UWl192xYS/VdodiZKxwSbgB6TPwtypCThSW1iY2CUIkVq28bRY0fdHh2cEyePiSZVwNCpWjVrm2ES1CeAk+uvq+Fy5VZZL/m9nDh+3A42TJZTJ0+axwzGQslk4stBwun6mOqJX4Dapf9Wz1yMG1yVN2+NcnOF7O2UXCpTpqz2JYDxU7ROjmThUJjyyLn6d2kJZ5lh0hm5XufKlVOu2zVdj6ef0oGJTJS4+E58BHwEfAR8BHwEfAR8BHwEfAT+7RGgemqHDh0soeRO+9U0wAgqq2zXjVZY78d1jX+tGqz4DmIanJOxKNu9mhYa5w7ZHOzXDV7u4idmgw2A1cITTzzhRqhYyXaZsmKxAOsC+Qo2C2vXwq64fBs6dKj7/LPPrbgKbH/f/hqBGAxH1JDWpEqdRgSWIKOGOFG1SjXzPUGCEx19xPVSeecXXnzeZE1ff/WVy18gnyk1WH/q1GlmdUGea8CM1k+ZMrVbvGSJK5C/gAEu5MHbtm9z3Z/s7u5/4AErh92s6S0GRAImBCWdRS0QYEDSelggBgCQUBX7mzw2IrKAq1q1qgGf5Ljjx49z93Tu7OqJ1TNg4ECBEZUDJovWWPzbEpER5I0qsgOYBtvH4yVq2xaBFMVte+w/QN+HKl/d6MYbXeUqFa0EdrlyZQW6COzQWKz6EY/a/sZNG2y8F1R2Go8XWD35BFACVtytClDk38iqHnzwAS1XwA0aNNi1UeWluDxc4MnxUyfcIZEkbC+1ffJ1Cv5EFixo8Qcwwz92+PARGk8VyZnSu/r16skn9ubA28bCwdbl8SNk5ayAmbRaBi8eTI5pGPGWK1tePjpNLP/fsHGTe+rJp9w2nfcDBg8Uk+mxgLETe/znC3TD/FiImq3PSkej5dnDKPXfFDva53nz57q6deu6eXPn6vh1dl27djHGUIzmiDFwtLYdPyCPa9ASFZgJjW+dPkxGy1U6WlIkDjrmvGnTpRYwkt1Qq9XSc+0Wo4VWTwfitpYtrOR19NEjVg8dwyFoYWckbaKE9oljJx1VlDANxpMmiKGiqAOLWe+ChfNMgoSZ7zlNCAuYDuR5yZWORh/VpD0iY+A8QvwauJtuukmSpfL6kD5vNKcjYtqsEINnyM+/yCn6PdG9ZrtNGzY61HfUS/fNR8BHwEfAR8BHwEfAR8BHwEfgvx6BFcuXm8QCYMWSriCjvOKwwCBYokT2WrcVK1YY8ME4SRpIwq+mvfD883bDNsg3rmbNyy9LjoKnxezZs23humK5fPrpp27osGFm1spQD+zfb+WYLwcowda47777lLGo6VeQ8Fu3/pciQDwAYjj+xKqwKvlKnWMNcCCd5C3r122QrwoKiQAA+EDslGYqNV1eaoq+775nYAIrbNu2VRWTxmpKBB4xzP8L8pw5LdJAurTIZNhOmFVTypMntzFTWqrkdIqU4W7ixAlSZNxqTJXQ9ulz3fp1rpbIAKfPUGxGsIrWh6Dw24JFMv6VAa4MVtjO8OHDBFo0FpvlmBs0eJBMbCPs9WPyYqXClEmc7BVVNhYL57yAyLRpNKbY/U+RIrmZD8+VfUiDBo0sFx8n09wHH3rApVYBHhqgDPMIeVNFWXscEVB14VwA1ixVtae9e/aaXUgWVW0ioLPVV/u721kpbPx1ULtYP/r14/ffx8l+iD/2IdBq2Df2n/3khypP27ZuUwGgWm67js+nn4hlpHMvV46cNhiFN3avtF9it6QQScMqUQkjYAwwXZA4MW796TYKnOnYuZP7TP281/c9Az4BYnhv0W+/uenTpokZQ0QDthDQj5F07BVpabQgx/KcAGd8fqZMnuoe7tbNfSIj5mLFitpS/AoBNHEvJOKTRAdmmAR7hASD4G7Zss3oRulUwjplipRizaQw1Po3Bae/DhpmPb8v/d2CDrUMrd8ZBQPnYw7QxqiNhgIS8ZwqhV39uuvd1ClTDXxpUP8GV0jIGxTF/fsPmiyqdNkyMnbaoZ8AiT+KI7OoSSCIlMlGw1exgvxmhBgeEmhEOa9oIY1nNIHxs8EcmA/zydoGDtt7hZL75iPgI+Aj4CPgI+Aj4CPgI+Aj8F+NAB6PL730kt3d3qibl6EWunsd+vufHmfMmOHekbEpjUTqWrTQOCupuMjGjRv/R9uZOXOme/udd/7MCBloIg6YEr9/qPgIrdGNjRyysPtlQnubyjaT+N+oG8ikjiuVj+ANcqlGvnWP/DWQvCjntfEm4jAvtdl/1esk/xAEaACCywUskvPRiB8Z+z4xt3LlyhWAGMpFz0qCs2DRAssXN0t6hOltt0ceMQKBFaTRrXtDMXSU9qmCUBol8uhlYJ2Q2CN/a9++g4Ewc2fPcRMnTDQQ7uuvvrZziDLQ5t2iZckzR48abceQMXE+4R1z4MhBlz+vgBkdUPaBoZLf3tzkZvf9d9+7Sspl33ijt3LjXKwGvCDpkKRMeg5rKo3ybms2pkBK1ETVh5AjrRF5AjbZTz/+5KjYRFVizIcpm80G2f+BAn/WraUqFTFyypmPGGuMyla5c+a2/QR8QiGDTOqNN99wlcXk6fPOu8b2YiVIEQyc+PM8A9uQdCp+KysZ1G23tbAxL1q0yBhLXXUu/DTwJxXuKSoABFAlmNUjR4yUp8x6gU7BfuLJg0qGcuXhAp7I95EtnRJQ8+qrr7k77rjDZEqffPyJeeKwXVhpocZnRVpJrGK037BlgiaPHtmedLmvixQ7J8wE+bHHHncVKlZy48eNd7Vq1dKe8e/atUQGZgK0EKkRkqBz52TWq0nGAbcfOVKntrrsMe6kdnitTIlmz51t5kbZxKaxUtVCHpliuCMvWfy70fvM/FevntPJRIkzJmjzW2/RydJeLJpK8pPZbbRIqE68v3vvHpXwmisQZ4qQvNXGujkvJg0+NTTGABoGCMNJQfAPSnp1UttOJYfsGOnkdu3aaRPRVvC/fAR8BHwEfAR8BHwEfAR8BHwE/mMR4KYl19ovv/yyeUPg8RBKliy5vYJ4kJghfyL5s6o4sakN1/OJ1RgnpqSME6YJN2RDwNGVbofcg0QcWUYwTo1OWZhJPRJpoG+99ZZ1imcJSXYolqHu8ZhhoyR/cyWnuFT78IMP3KSJk4L1Q3GMTWIvtc5/8/WA9UEiDksJ7xOkS8xdM71VoK2qkR3zMMsRa8kj5ZAqDj377LNWpOZlgZL9Pu7nikhiZuWyDTBI4rZGbXHXX399wAhRcKkgTL67edNGFxkRKR/TJMpLj7puDz8iE96nrb9BAwe5lMlTGujBvFry+xLLXQ0DEQMEkkISGdagHgHAQQKF8qR2ndoqrJPcvSKD2qVap1Wr1mKm/GCyp2BdGCTOJENV5eVCQR0mEeuHJ0tuvjAlVOEJRUiM8uGPPv7I3X+fAMFWLa3sdyaRF/CoYR1MdQ9L/kSMbH7qkZLSWJOckEwJIIOqUbc2v8X8WvtKfjdp8mRXpXpVSZsGWulwZnAIgEIyVDCyoH1+MAeTAG5pO1mzZHWZVDWZ7VD4Z6AAmWqScZVUVSvYPeZJEzunT0j6hCfVOfOyoW+do8IYYNDgJWtN48orYAWZ1mBJrDDzRTL18+AhrkuXLsEy+s3x5zOCnD+JPs/oy4AdDYvy4Rs3brZHlpmpqlm1BchEbYnS85muQ6cO6kFxCp1zcb0mzpNEBWbQXzHQ666v7l584WWXW8E5r8nLDyXB0mZIZ8ZJIG41ataQ6W45+cacEnop1EqyISoqLViwQKjUBJc5UyYrQ0aXmPTqmJlDc055z1Aee/fOXS6b6r/nVOWlTWLWDNRE+P77H1WiS3QpDQKq0k6BK5QCI5hHRf/aK5fpkyeP28FgQgAAMWmQPx0TisYkLliokCiGDVz1qtVcinjO3YkTbt+Lj4CPgI+Aj4CPgI+Aj4CPgI/AvyMCv/++zK1atdISKbwhYXHEmczatfTl9+Phhx82iUSGDBkcJWlDWc3FOMKPP/7oqlevbj8AJAkBP/iysMx1113nfpCMI9QoCbxCRUAYaEREhJs5a6Yls6H3r+Sxm2QLeOhkUBL6nZjztntKRC4GTxgn269+XXXJOdr/4zirV9c4pRCgrVy50vIc9r9nz57mfXnxuJA5bZdPyclTp9xzzz138dv2N/08I9AAKQcGsXXq1rWxJhSvBDv4j7wI46NBg4auvKww4k/VmFifkS1btwgcoMCM6WsEwAWJefXq1VyByAjzNp05c5a7ve3tinEd5afjXA2pNzCK5RgeUAWinCIi7Nt7wMooM5+TyCx30+bNrmTpUgaUAA6ePnvafaECONhpFCwY6WqpRHWjho3kdxoLKJDka2WJadxOKUZy586ldSA3MPcAVsJNilSzRg3Ji867KAFCt9zSXEBNMjdBMqnbbmsZl/9i1bFFjJgNAofIoWnYrOwTaQF5E4HAKDi7cmjKgjdr2tRYJ3fceadr3+5ugUYpWEPLBAwgQCIArHz580naM9mlow91grQruWRat7a41bx6ojZtdj/99KPt/+Cff3aPaV6S/9MAs2bMmOa2aNw0k5cxEP1fsniJyniL8aJEn/3fsGGDFeWZpmrL77//gWLW2GXJjHxK/ShnN+xIMeE5FYCCGIppJHBFi4ixlE9jamEg2Zp1awXUvmIMoC8+/0ySpE9s+1Y5SM9+GfqLSw6Aq+ccAoYLyDNbn3GZJdnideIPC4n9fFMMo/f69BUukc3Go7cTvSUqMMOEZC9SpUrjykpWBGoFReiC0MmiRYuZ83Wq1GmN1oQpUY1atY1WdOjAYa0W5rLlxIPmlCbTBtPPderY0UpYQ6vCAImJkUGltjfr4Pf7pJ+tmy17TtGq8riDkjOhgUNbBzJ4vU6cGtfXcLlFTwMdpOw2jtCbo7YaHaxkyVIuomChoGyYZiwHlBJn1KSvKsSvY6dOf9ZwT/Sw+w59BHwEfAR8BHwEfAR8BHwEfAT+344AkgHsBPq829f98ccfBooEQIUu+MlmLtMGDRrkflTCRtr0cb+PlTgpN4hbh8Thz9ZCCdVBMRW4SQuY8dlnn/35pp79rISv73t97X0sCVoqIY3fcsqbok+fd40tD3gTN86/bib+KnHPQ+PkhY8/Zpx5Y8f552hDC9+msshIiBbMX+B+/OFHqzobeo9HG6cqLi3UfsAWwmuEFrU5yh7psX79+vacsr7zVKgEnw7iS8JpVWRl/5BQo8IMpZhP6sZ2KeUyvV/vbXnO30eZ0Nr/rdcM6hCwMETz5u4OHS3XJAJBCu/cvDnzTe4SffyowDWxU5IlkUlwKjsOJYqXEJwgMEBGJ5MnT1I+WtHKKk8XYFCsWDEtl8YS+d6v9RZQstmADJJ4GDn8ZBQIqYNp7BfmYZmy5Yzh1Pb2291h+aMyv596oodZfQQAio6gsnJAnTq16yhn3S9gQ68pN8WAd6/KamMNAhgHALd69RrXUH4xw4cO13wb7G5u2kQGuem0j872Fz9X276kRFR+uqCNAPCYP4rGCdOnXp36boPYMV0fuE/ypi3u088/dR+o+hOVjpj8oe1jaIw3LDFJnTa1wJckAkFXu72797kypcqoqvErAltkoHzkqPtMfkmzxDLp8+677ukePQQ+yi9GbdbsuQEoqVjEaPvEJOZ8mBE00sjrh3LZ7D9FemCu3dP5HrGUPnY3NKjvvvn2G9mXFLZ+2P/Qeb1x/SZXpFARKXRU7psBq9/0qu4MgAbgRqnr41LFUI2qU+d7Xds2bd14lerOkDGT9bVL1ZtHjRplABbHPyxpYPALOyeNjKLPIr3S9gDIwBZefulledh0VKWppDoG1kWi/0pUYCYYXYwO3i6BJEtNyxdgZWEG1HTq2MmcoA+oLDaaPsx9Ycwc0wkB3au2JiKTfe/uvcZiySoz31zSz52V4c+ePXvkmH3Yvhxw1sY/hg+zkiWLSfvXTSZN5Yz2hWSJFhFRQPXkW7jWbdu60qVLc6xskvGBxvbKlCntut7bxSZm0WJFXOEihV3WrNnsYCfVSQxli+345iPgI+Aj4CPgI+AjcPkIcLFIsoEEYNiwYZdf4aIlHlAVi5yi98fdjb/o/cT6k3FSeOB/NU4lqtd6nOzv1KlTdUcw3JWXFv9yDSZD69at7ZoHyTasZY7Ha6+95k7p7vulGtdSJLt169bVNVdOK21atUpVB3uAyjm+/bcjwF1+ZELddQccxntcU8Jyuexkm6qg4J9CztS6dSsz1CSp+7P9NbtJI/Y8bBQSSFrPp3vqzvw2e44vBka3ypRsHD8LyGCehxr+kVFRURpn97hxBgwSxhlaKuFH+ubzJxhna6scZUuS5SYA6rBdxkmCRuuhBPRv41Qyl0JJJtKO0Di379huy/MrMjLS5BWwiDBOvV0Je8mSJa0izqRJk+KWu/hJr169VFF2marZhltybzKc2B0MEtaL1/hv/z1HlhlTJk9xX3/9tXu3j7yD1Ji6tIOHDriRqoq1Y8dOAzuQ2JD4Hz0a7TLLlgPkJRTTQ7K8aNCwgcCCT1y1atXcu++8ZccQ1UXw+QrcE2agAscUrxibPOqP1xs0aCBVRqRMcf9wixYsdEjaer/xuvtpwE8GBjFXkdjgkdKv30fmpyLoIlZSJwuQUydlEnxa44ydkPQrMOS++7u6hx9+SPnsdVadK1Qa/rgqiQEeMAxybcAWQCbWRgIEM6x1m1bu9jtuF8gIuWGpmz93vuvStYsBhdmUg8cILOF8PS451h9r/nClxAI6IvCFsWbMmM5lVZnwo8eig0rMGkvoPBs2bLj7St41NWvWEmPlJXeDfGF5kxiyfQgTNOlbbB8oXW1NoEHW7Fndc5rj+MH89OMPVia8dq2aqrI8N9YrJgB1OIgwfg4fOWTGwAasqV++16l6llHHD4mSDUoP/ft/K1zgVitffkuz5na+hcbEtgM2Tpg7fuK4Pn+SBXGPPf5p06Zx93a5V/G7ICBnjNmg2OG1QSfur/ifjonUs3ZKSNMfa1Ybav2OJh6lxDDh4gM9S5bMAln2ujnS+fHBs2HjBpkF77FSXtmkNcsjY6JtO7ZJ23bYPF+QGB2NPm6T6pRkSiVLlVDpqq6u3V3t3OJFi41Fw8QoWLCQGTMdO6YSXULa1mj7lD7LoYmFORIoJq7UoF6nBAgdFVrJicNdgBIlSjg0na8ICasjFg/14HG+9s1HwEfAR8BHwEfAR+DKIvDiSy+5adOm6WIOU77gJsmVrekMyPlcd8f3CAi41mDAiy++6KZPn24XYlc7TgobcBefm0XXepxcOHfq1MnuvoYMKy8VT+4u1lZ1l19++cVxp/SULuJ37dxpx+N5VbkoXaqUySguXh92AkapT3R/ws2UOetu7RdefYt+WyR2Qz9dH5U0c9KL1/N//3ciAKjAXez4jXNceVGQ9cV/I95zEp2OYr7DGMF0NMR+CcCSeAte9JRSwc89/5ylbnhTAOzQF9Im7qSTbb4jc96yZcv+ZU3GCUgRv4WS6vivXfycvjnPKAISf5y2HEkw+5lAY5zPq8oUyTe5CuWR6at9+/ba52Ccffr2kd/Fn6AqlYGClNRZQvyVSjJbKW5tI/T6Tp23jQWGffnFF3/bKkbBfcXEoXH3vkKFCvacxN8ycPvL/4ofAXxCAN2eeKK7zJLvdaMExJD/hdoS1Bb4sWhecbhPnzytpDyZSyamSYyqIuGVyhSoW7e2eaQgQZs5Y6Zr2rSZqiUNF0gjZhadwQBQjrl9+1ZXQ0DbMeWZrAewAePl92VLzZeFZTHB5Tg2F0AAyDNj2nSpTPK6pEibtBJlnfGaQYPEsT0n6VQy9cH84sdkO1qulD7Xc8qM98svvhSrpL+pTcaMGeWaN28W7A/D0pgoS19L4AZmuWyfMRmQqD4AQ+vVrqvqTKncxs2bzIw4f/78bsWy5a5pk+amcgHfOC9WSj99J2DAi6FvbeXL+PYMGDDIHRVYE548hcugKkupU6c09clvCxdapTHsSGD0PPjgg9qyGmiRGuNCIpQ1a1aTYfGZQvyzSSa0dt06lzN7DvnOXJAEcqsrXaasqWImiO0COIv0ic8fPkuGDBmqJ/yHieNkVyLwSn0lF7gSGBoH26ogOdtMydLubtfe+vxGwNHbbwVm5KxPXFgfW5WyZcuZZAwSD+9l01i2iVEEcGrXN2ev7vqG/b3SlqjATOjDdo/8X3YLbGEPCS4mu99qwrwrWhM7DCvmmKhFaPtW6QJi/KQJCuwQW+6U0MDk4SlsQp7QhcW+A/vdlm1Rhn5Rpx19XH6VEKtSubL7+tuv5SbdXyhgBivFnVwfyKCLR/UlMFzuzV99+bU7IUNfDH6jdYIk0UQiwFx0zJw12+EADViUSvIlHLmhPdZrcIOVBAOJp1SYbz4CPgI+Aj4CPgI+ApeOAOAGJomvvfKKLaSv2atqJCJd7+vK17Mal0fXpnEx/GyvZ61iQ2gLV5K4hZZlnPfpxlCoXc26oXWu9HH5iuVWUjd0F/6f1vtQ1HOo3tC2uctLpY1xqiDBHf3atWtr1TC3cdMmd6c8BChyEL/de++9xm7mtYoCaAYP/lnrjnO8TpJx8OABW2///v3xV/PP/+MRCFNid7nznMQTIIH5RxJkDISL4napc4jPk2ryZuGinfkI84uqTnw6IHeiEsyVt3/+TAmNk/355ptv48YZ5I//vJewV/DVZKnxqqjEODEIZV3G+eADsclo7GC3bafMeDAeqrzg2TN69GjLS0jGe/V6zpJm/D26P/GklXgO7echmbF2EDhFrkV1GFg6oYaMg/M/lIeFXvePRCCI9wcyS24rFUVWeavcogIyJi+LNzXwQwFwwPSWgjRnUWDAStHnIC1bVq3XvLnlpxvl31JdgAyyp7lz57jbJFUjoebnpECe7dt2WPUj1oMYwA+EAwNWmCzqkn5Hjx0to9xmKjizS0Vtmrv7utxnJaV5n8VwemE5wKUsIi8wkgAQDd6nbHYXMTkAIKL0Gf+AQIuVq1a7kSNGuUcfe8xMg01epe8+GJWwUGz+qSP6xHuGXLi8AL5CUo5QQXmvwPmx48e5o8ePuS++/Ny8VdKq8hTsE0gNx6KPGsulsCREECGSJw83lkk+MTRLiehQuEhRF1mgoEstwIebFxMniISxYaPJGJFJZUGdArtGbfnKFcaCATgi/oBgKZXLz1ZM88rThp2FAQOgecftdzhKmT/3bC/z+8mXL7/1gQTKYqV9JP4HDh4yXCBU0pr9B7RtJNkX59sOYRS/Tv/VPfLoI+7hR7uZlCmfcIVziiHb54ZGgxsb2v4HwQ7kVfj6lJANi7Vg+MHzRP7NPiRas8mr6OzUBFsvQKZY8WIqd1VEE11oYbKkdrcGytPzQsI7derobrrxJkO8jx6OVum4P8SimWuoM6ZLTBgOvkmPNJHQcuJsnUuoewY5JicX++Ws0DsMfkH5ypQuYyW3OnToaOX8iNl+gTpUcqIs3eQpkw0VxCSIPjEQ3qRJjBYtX26hlDoZAY2QYeHgjPkQJ5FvPgI+Aj4CPgI+Aj4CCUcA+nOVKlVUuvMNuzhiqau5ZiGR4G71AfnEhdrVrB9a53KPjJOKLW/0fkOLchkXtCtNZFiuc+fOdtFna2qQ3DlL7BYq90vpUZjGlwsmYFPIIJRrIyQO3NFFfnLXXXcZM6ht2zY2TG6S4aMRatC9ueNLgzWDz0WbNq1d48aN3Zdffqlrtedt+7CDfvrpp9Bq/tFH4LIAAPMQkIFz+aGHHnTIjBJqlzr/qEKDkWi6dEHZX0AZWl7dyQfkudIW9P/n+X7xesE4e+nlMAN7bryxUdwilo9f5sOIpBGPGRJXthICj2AcIJ25uJGThD5/2DeqBDWV+SqsBTxtXnvtVfeSmId0Rs7zvhLRUINxsF2SG9bDmPivkrCg1xCIEFrnP/+oz23l2ta4AT9x4kRJWNqaUoOKWK+89EpsiPhAD0AYGExlSpY2ENt8R2KBmb1idwBIhmK8a/cud53YLt/1728yGRg0yEeZMphSb1QeSZ+8EGJmAsxwpGC80DKky2ASonvkf7Jc5wwMKz57U6cKSfSCMfE5D/kgjY49IJxtROsjudkgBotVcNK2YFY+JQUI0tQX9Pn9iOSoefLmsW0BkmKQHdo+PkZLFi22/DdVilQmH2RXITJEHz5i1ahW6AYBkt3R8mHBXuQ849YyxIB9bCAgsrpiQJwwy02q8RwX+SGtPGMK6Byoq3WTyrcH8HHU6FGuU8cObsDAAQK5AusQtjPk5yEqvnOGYVnjnD0lIoZJrwS6hOJ9TODNW2+/4+648w4DWKZNn/YX1hyrA8JskyyxabOmxngxFo46BphZJyPgOnajAs+YJCJvfOVubXaL5NM53ZjRY2WGnNW2T6z7ffiRsXlCx5/zltgCNtHsGHBsr0FLVGCG8UED4sLiiMAUJkOxosX0gZNWpknlTWuGjuy0JElFixZ1jW9u7L4QVa90mVJu+87tMu6FtuW0Xl4XrfVXCfWjhHbqVGmluTtttdXTynQI1gxAChMJZ3Kou3xYUXquhqha0Am7a2Ki2aQUNgZZlD3jwKRKncpKa7NsFk0iUDtcpXlct2Gdo/wcfXLy+eYj4CPgI+Aj4CPgI5BwBHr37u2qVquqi71luiAKSo6yZHCBpBeuoMH2QNbM+sEadslzBWte+SI2TskOgotSpzt8sXr2K+/CdO4TJ020cdpqwU5eRQ+XX5SbRQAkQbnfc8HdTu4BEpxLtCWSilN1kkWQZJPcxW9c1AbleYNXkSiFGndKQ10jOQkPD3w9Qu8/+uijwUWdFqJyhm8+AqEIhJKlUIIYep1H/DYwpz2rXIDywm+/HSsXiL/QFTwvWLCgVUGJm6Sa418pac0oADIxGuO8q91dlrMULVok4XGS1Gj+/1MOxjjjn2N8NCBRyqTqshc3wIFQo1LVxecr7z311FNm4srz5cuX8+B+UFnkQYMBVWPce++95yIjI+310K/QeXwpoCu03H/uUYEhWadRRQtGE35CVDJ6Sd4nPZ/pYUbNBl7YZ3rw/fXNt98aMI6UKZQPXlCZZhgnzH0DVvQIY+Shh7s5PisrV63iPvzoI1WAqmhSJSr9Mm8gJwDIGHinFzhGoW+5IkULmVQP71XkUczH5rfcou/EyQHogJQpNh8dMOBHY/PY+qo8TGMs28QmMU8be0XsDuW6X331pevxVA+z+YBZRQUoGtJV1sfXtbzy8mPHoyUfliz3wF6ZrMjxRePcrv5WrFhp+7B2zXozKq4q8GXevDnyZqtrgKCi4vaJRdlHrLh1GzYaqyVaZIqjAhOTS/ZIrg37BkULcxW50uLFi1UOe5BwgaKqSlXLJH8avsyONykeAqxi43JYqpfkkkUBehJ/zj9aixa3mEQK8BMwB9ktYBOEjPTp0wWyJS26RwAajCb2FZYPFiYAMUcFGGFfYp9dQZcibUwyoGfR4oWShZU2Fg2RxXIFgCq0/f0ih6SQpxNgqB1/xT/wELKhJeqva4M+cNAVCNPoieKUTD+RBSNcg0Y3uLSqyjRBiCWUPCYHgU+nAwddF4otJb7W/LHadKKUtkb3lz5DWisfBpUpaVhyTYZ9pllNKR3burXr3EvSi/fXSTTt12lCyrZLH3jSFdIHZbFixSWD2uoOymzouChZm6M2u/MqIwbIkz17dpNUAdpERx9z6zesN1YObtiUAEOnGiCbiRpv35mPgI+Aj4CPgI/A/xcRoIrIubPntC8x7pFHHjUTxdjrHft+v9xOWsnXZ57RhZdzjz/+uKtdt26Cq3AHq2HDhkF5WkmOp0+f/rflYMZa+VrRyzFZDN2hZEEbp2jpXEw+8sgjRl0OdWAXaaE/LvEYN04laYyzLuMM7Wi8dRhngxsaxI0Tv52L2+rYcTJWyuKyTqgBHLEtWkSB/LFliWvbBeulxolGv0qVyqrikteYS6G+4j9yI4rxsv+7ROMONaQlugyz97geu7gF29Tres8XQ7g4Ov/tv+Pmy9+njbESmMckn1FRUZYAFi1W1G7UclP2jjvuiAse5yPvwdJKqBmYGpqbOv+4EXs1Le68SeB8xa9i1UpVrtF7mBuXU6KK1QJjhO1/p40zmP+PSvbAe5caJ8yb4AwLRoc9Q0INyUSoFQ/JIkIvxD7i50NJZdqaNWvsEQaE3fnWX4BANkbGGfszd+48vRMmtcL6YB8U07GSVvkmeRHHXj8Xzl2wAjSh5BqforZt7zB1x4gRw13FypWCZbX4BuWEVMgyc9rYIIY5eZbo85okn7mNpAg7jLRiUXzX/zsr7ZxHwNuC+fONIcYNf3A9vh/Zfvmy5eXDEtwUCJ0/VBZOLTsNgIwkyodHjxpp61KheLgMdG+5tZmxYugl+sgxY2QZKBSbvdMPAGgKARlmdKu/W9x6q7xmIt3M2bPcO33eVSWnNI6bCl3v7Rp3U4IYAPqVEDMI7xU8cJC75hewj0wuXOoSSoiz/2eVhwOeZMyYRcDKQPfwQw8rr09m+w/4MUcSqWFDh+kcinLnVD2Z4j5FJGdi32eJEVaoUEFXidjqH3Lg/mIYHZYHE4yy3m+86dJofHZ8NHbsSJB1lS9fLpBaYQ6sfeecyJA+s8kDIWZg0NtcDCXYc3379DGvGCpSCTax7yvGc0AyXPaA+FOuHGPlJGL0gDvQiEH5ChUNxEG6u2njJtf98cfMWiVz5kx2DcP+8/3HfhYvVsrGZF0q/ozrWrTYQ5tYXRNUGQkpyLBaTguNNkMl7UV40nAxX9JYKaply343KRFIIwha2tRp5BuT3yhdIFQgdXx4Va5cxfYc7RnII2a+4cmTurHjxhoVqk6tOjrgkYZYrlq90spwLVy4wC1ctNCtVB9sn3rrlNE6fuy4KLsj3Kw5s3QwwjQ5dpgJ0rlzZ0w+Bfo2auQoQ+Jz585jyCYHzTcfAR8BHwEfAR8BH4G/R4DvyObS3C+UyR+Uey6euAy6kmZ3qyW14YYMJV+RQnFDJ6HGhRTVhubrgneBvuPvueeewDAzdmH6ukP+KaH38VJBChFqNk6ZLHJtgM9AfCPT0AVyaNmLH5E+IwmycZYq6WDfWONa+6LxMs42kg3ZOBYskE9Ll7+MM9QX7/NDv6GLRPqkP4oV9NGF5ioBOH+W++XaMOHI3nzzzRZ/bm41ahTcFQ0G+OfvuXPncm1qx4a7pKFG/xgz8sYnn3zyl7GyDOMI7WKTJk1Cq/lHHwGLQMJna2BsGgrRmbNnTGoBYLBOkgtugpJ4hRoeFLwHW+zihvyBO/9/thj3pf6mvO2VNs4bG2cCp0/czVe9h30CkpB1smHA3HS9ZCjYMoTmP+PkPZMXXrRxzGRh/3MihTaDHCWhccZnyJDsXaqFznfu8NMCAFe9a2dsfBoj4w0957OF7Z+X9MP2QTHFdNk3gBnlcgodscHXFNlP6DiNkJTz9ra3W1L+dM+n5afVLi5kQUwJeTDT06VLY0wT/gS0CBMyA1vq6Wd7mupi757dJqEhl6Qq4fPPPS+mSJZg+1ppqcx/MWVHHhNsXxIbATvz5813OWQuS4PdQ6l4jK5/GTbU+ri58c02ZnbCqjLpMUaMDUbFDQiAIZt56hebj1o1a5kX2vXyaDqr82+5THxXCCj97IvPxej5UDKg/EY8WKMqS8M1zogCEZZ/YzmCXIm/IS+Eqy+AfPJyjIh3bN9m239M4AUgU+bMGa2fc+fPWnWkQwJCUKLsFduEylOpxJjBUoSy37BpIFNY7HQcABwfVz93t7/bDfhpgFQ1FWLJEGHGSiKuB0XgsKZgARxB1kgnZgxsH44p/mcvyHy7Y6cO9r3bVjLcApEFbBXObc4ZYg2QA6s0XYb0dj5zThN/wCdiVL9efXsBls09AmgoCARoW183TgwE08Kcj0ePHjH7FGJN/K9V+/PKJVG2EEzeZMnCBc6ktVJcGO1ysVGxQiVzdGYzQY33C1ZRaeOGjQaG3NT4JrFcCslEaa5AleXmTM0HEgjgooW/uUrSsHNgqOC0X2a+CxYscr2n9nZVqlWRHu9r+cisUknJoOrSvr37JaWKdpEREa6iKGVM1AzpM9pBxauGylDQxDZtipI2rqpRCUHfdkoOlS17NmniarsiQrINxUuUuPhOfAR8BHwEfAR8BP7/igClnBMy9LySvcQXZYXu2HGnD5Na5EVxSVRwKfGXbqjGyB3g0dKpb9q0Wf4VvYzSz0IYhS7TxSeNkrN4wcRvF4/TABVdl2CWebnGdpYvX2HXEYwzqPoSDNAuhi/qwMYps1KStU260Is/zqefeTpOTgVr4OJx4sOxWXcm/w4cJRCQi7Z7qT+5QH0GVpK1GFULqRG3KKAQNPdnnn3GLpS5XuomTwKkImPGjJEfzWBbNiIy0lhIcSv6Jz4CikDABvj73AxVOr04SJwvJFSwaIaI+cL5d/PNTXTXvoRZEcRfHl8jKuhwiqaSNGPIzz+7Vq1ayQrhlAGeK1ZU+9s68deP/zwWmon/kj1nnPhx/O1TgF3Si5h9/vILzJcwjbOxVWQx9lm8nhjnvV262DIk/D9rnK3btFEl2ZNmoI2vU/x1qKJjTZ8/85TvPCaD1osbHif4XLLd66+/3t4GjN4u4+C/tNhx8tqAAQPthjO+HF0pKa7AlS5d+i+L/2f/0JyzpnhRBjuFAAPAFsKHtw9loDEFhg3SubMknZKsPNrtEREIDisx10LcpNfxohJQhgx4HmlN/UfKkkNA+swZs0QmKGC5LgDA3fJR6ao5wWd/beWTZ86c1eIq7Sz2SKZMyPBYmd/BGMg/Q15KsHsoSINCpGePnuZ9+sabb7hx4ye4ezp1lqpkc7ztOwGFm1xDMUSXx0qPCiqPRmKUTCBLSZ1X+LKGCZhYLAnrfrEr2+lmQOWKlWXtkcbAJb6D8WotGFnQ3dDwBlOcHJPPKtYfWTJnccuW/u5WC0TJmze32ChnBHacNTuS3GIGTZw8WXO8q77HYYvp7NacW7hwkVGUkHydF0s1MiJC/e8W8JTNHdVzfF7Oi7mUKmVq99mnn2ndFe6V114z7zPOJeJESWo8lM4L/MpI9SzFX1CAgSSnBWwhc/qTjRSm76qxyvUr67t1qSq/fS5VzBkZDk+wGw0AKzFJwwx3yJkzh52XfKYozAbY4OOUO09u/aX+9cb0X3918wWMPS9/LACjOXPmuAnCCzj+AG5p0qbSklqbDq5RS1xgxvY2TAc4UmhbNvnESLumQDPBKD0FnQp5UwbpLi2oCs5ZHWSAk/y58xuIAs1pwbwF7oIkR6Z304GldnuObFn1Z4CWX3f9dUIhsysuYS5v7ryueNEibvOmDaYx424VBkrLNNmoLc9FI6XIChYqJBmUStodO+H2H9rvpqqmPRM6rRDQjZs22wTNLxQRnV6qVGldLtHLYNv45iPgI+Aj4CPgI+Aj8PcIXAzKAHhwvaKv9n9s6ML79n3PlkPnH2JxcGHEFU/Qy9+7wLehTJkydl0RqrDBBeD7H7yvhWNc4cKF3eeff/63FS8ep21Hm7qiccrPgYvFl178c5xxa16iA3wwyogRw/UPHjqARdx1/PCDD+2CjptQCY2Ti+G/NbsGDC56//beFbxA4gejiePSRCajyLziN4AZEjgYMWvXrnXd5JdAAh267mzYqKHp+RMcW/yO/PP/VARC5xDnxsUNwAOZyKWaeUQM+cXWbH93e0uKL162c6dOMt+U74XaG2KpMT9feeVVAYlP2esYhlOt6cran/M5/vKM8+3LjBNghn2kBDbn8cXtHoHAyAlpsP4w8n3l5Zddj5497HXA1/iSIuRQ7MvYsePcMMlnAI2RNMZvVLAF2KHVrFnTHs3vyZ4l/Ou3334zYIZy3+/8Dz19Eu753/9q6HspTABNmMBoAGm+api5latUddmUXyI9+1aWGFujtroWLVu4MaNGu2ef62VMKuX1WjjGTZk6RZKcyiIMiIkBUiDQJqWsOpA9RUQUsMQeP5kTUnn0+7ifATEjRoxweXLnMVbGVNltwAJJpjwY0gG8mWgBOTA6jNWjzcgExBWKzK8KQg1d3/ffsxsRFSTrKSxp3ThVSnpOhr6zJR0KMTkgJeTLl8cUJRyp4ppfVBXKJCkOrEsq7QH0w/DZorLTe/fv01jquaxZsmk/b5Nx9U8y3j8gNswOY7oclVxq8ZLfrDJS6dKl5MezRQDqBZWSP2LfI83EkMXcl+DlyZXH/TxooHtXcqnv5YHETQCYNYRrqyxB8JilpPz69Rtc5kyZNc6zbpdYRSeOHleundFFFiropk2fYZWuYNzCIH3yyZ6qnPiMAKkDdnzwjkE4xvkAo+ecmCqATnZzRWOoI68aKi8vFYD0/fffm+Ew379ff/O1JE59jTUjPM3GNGXKVGPUMM4YocrE/4TsTDD2pfGZRvxXS94ICPbo44+KIfuGALlkrqVA4VMChfmxZQ0p0rPQF6W9mji/YmHExOkMRIlWuGhRBwMmXJOPScrJgL7rkDRlUIeq6URAr3w0+pAqL2lSxp4oADSwYs5IXgRidk4/aAI5AOmEAlNC67AmB+ZZ1DFfu36t27p9mzFbMumgJxVTB0OhNvrwLFO2jDkyUx0K1g503TLlyopm1sQ1a9LUpEqHBdAc00lxXIgZDfOnOkI30wsRPSHEMLQ/9qb/5SPgI+Aj4CPgI+Aj8I8RSDgF+nMVK/mqO4rciry45Cvf9YACl2rZRLMOVWRhKejeHTt2pCtdWyS3u27cMb6SllBCGX89G2eHDur7gquh5IhKR39psdc7f3kt9o/44+QCnKSuo/aZC8JwXadQGSl0hzSh9f/ymnaUfb2UzOsvy170x0svveQ++qifXTxmyJghQTCIO5N3a3xxF5i6jDP2UmxfU3UxCyOJsfvmIxCKQMiQ9NJna2jJhB//6fz79NNPlYSOtxUBJmBx0bp3f9wSPZ6P1x1x5HdX1DSn/+lz5XJ9sI+WCF60INsfa+BQmBU3CY3ziSeesDyHxSlbf/E4YQuy//hzwQKCZUOeAlMG/5hXX33VzvmSJUteUp540VD0p3YySMH+/tZ//JUQuwJZT24xXJgLYWFBbomXD+oM8k8+0qfNmGYA3H5V7h0oFlLLVi05+BZavhNmzZxpn4UxkjHBGCNRP6vjmEb+JvSrXi0njchfwAGW3SDPsQPqi/LTzF9YKHSWhO3r39YtUeYJhMzKPumVlZOvHpJ1Rzr1ydgpsc5n9JzZc9zXujnBd1HS8MAnhUpBE1SO+riqGNHvfDFWKHyTLGlSgSmHLV8+LgnQSY1z8+ZN7rhyXmQ6yLDeevMtgZ0v2/bYDiDMAeXFSLjIv5EgQVhABYNahTwbPxvIFmyL/Uda9NJLr0gi1c+lSJlCJa+DqXhIVY5XSzYG+7KQABj2QYvLwyafyyjW0KGD+wygpEoSkqO7O9wtcGWJu+/+rmJ/DXBFpVqh2XnLgRFogpkwOIABKMIWOC5ZBarVuB7wUsQLecl+oGpK3R7p5rqIbQfLjvLY1ou6gK0zY9p0wxc4/oBb2m2TuDE4A9wUF76b02i/3pD/zRdffuEqVKrofhWAWq1qNYYUNFbU/2vREheYsREGl2VpUqF5E1VILv8ZpOuaMnGSIcTRR6JFSxRzJiypnJP3in613OXOlduB8p4R/ei8jiqaMBr7zGRNomW545VWgAnADugi1QNO6UMsWhrKzJronTp2ctdfV93NklvzSb2eLk06K201Vh/ulOJOpoOYRsbDrJclS1aXPFlyQy5BuvlBo85JkkrjZhsAM8GJYkPxv3wEfAR8BHwEfAR8BP4hAnYX3b75uRRKuFnJ123bdfGaNoGSr8GVzj+BAHiq0AcyCPwVYNmytTfffNMqGiW81b+/Gsgw/v566JUHHnxAdxG3W3UU7sRd7Dl3ORlUaJyMjTuqO2NNd0m8Kuuu65W34ALwcuON3x8XusSI6k5snwv80aPH2HVW/OWohtWps0qVy+sig4owAHrBQCJJhMKNtAlg6ZNPP3H4CvjmIxCKwP82ebjUfMZ7onv37pYAwNJiTobOPW7iwmpATsj5RxUylr9cCz5Vgt+XW/bv7wfn38VJGOwytk9DwvSN7tCHxsnj1xp3IHuMUXWcp/4yTrydXn7lFVv+sJJnZDQksFSwefrpp803JG++vJZAXymAS+6qnNK3BCJgc1WxoYJdhCQ7Z06pNHNsY05h8kseKrhEx0QV6LTcHXfcLoDgJ9f9se4B0UDADS3kO6MPRgPLd8gCI1NGVd8ib9VrSbV+7py5XQfdMIBQgA8Keenrr/V2AHIff/KpK1ywsACM4GBtlqwvMjLSGClsn+PI/MGLtUiRwjqmGpMIC9t2bLfPYPqABVlXgAY3AAAM8VFl+9xE2C1Z0niBgVbWWf3MmzfPZUyXXowVgT0Clo4rP4bwwOc6DYncR6okFXi8qBv1w9w7IG+YVatWukMHDmms593mLZsNuMG3aL+kUh99+EHsKRGAVs2aNnGdO3UyyRJMIno/qLmNPIycGoYM3keZlK/jKXte24EYAWuNfUImtnTpMpEuugrIKeR+HjzY3arqVNq0fRbweQFewE0DfHVg0eglwwzwDSJugDcAR0iQKsv+pICBR2VsAABAAElEQVTiCjMVMkjy8BRagKrRxMqeGhuKOAVMGToI4p9dcWUcgEAL5y9yDz70oDtw+KDkl0PcA6r6bM2AtOBpYv/+3362/mU8wb6GmcnPQSGEHECCxIfFWiFVBHr23DnmNbNrzy4FI0wgyknHnRyQYQAS9GtcIITAEjRqYJAgdimTptBjoL7CqOvsGUmVhCyiG8uVO5fLp5rp/b/7Xojz2270mNGGwmGKhBQKWCxXrpwGyqSRfi9nnpw2QWJE7wKlBpjhYpByaJxkuDsH0+0vu+j/8BHwEfAR8BHwEfARSCgCdrHJBQ5fucFj/MXwaIEtwhUVUiQuSP/agouthO5Ox18Oqj/f99bUF3JljASvtBmAFHtxltA6P/zwgxsc66/yvujkfx9n3LVdQqvHvYacIyIigutFayRkCXlKxK2Q4BOtrP9Xej1CsYQWLVq4Tz/71HrjptZ43aAKSSLibwLgxijxenHo0KEOeQieC9zAwtsCyUlkZIRt/3P5AVBm1jcfAYtA7Pkduia/mqgYgBF7TpAYx29dZJjN9Thz/jV5TxQpUiT+25YrYPjJ6hhyd5GXxz81+uf8u3g7/7RO6L1gHa2s/8hP4jckSmyfz7LXX3/9b+PES+YFyU70tiXDVH2J357XezAXqMwEGACYSr6D5KNly5byyJhoRVDir/NPz4kpMcEuwre/RoCqRzSkpV/LPBoGCbgE3wP8kAPCUokRCoCkCDYYag/Klr+mYxsREenGjxkXAOocUDWlixbvKDFeatSsoTzzlP7mTd3gV4Ul2DIlihezV84piX/77bdMDlNTbBZKdWfLlt22DdCD3BVD7ND2KVmNdCeopqdeNabI/BE2xq++/sq1adNGz7Op1PMUA41C5yDzPImIC4vkJ3NGuTc+MHt273ZZZS0SIVPcGjVqWrEbWCh7ZGYd+uzn+5Pv5YIFI0zyhAdPmVKl7abHWYE+pno5ccqMrzGsxjcJEsUwGScfUtVjdpK5l0b2H5073+Nq1qppuTrRgDzx22+LxbLZbsbCVF1GFRPK/4kT31nYiJB/U12ZcuFr165xP0tedv8D92lZbUI/GOZP0T4H4JjmO29oy5ToTkHhAZZTrJBuIaWqKnBmwfwFrpckaT179rDt20htvJIyKc4ceys1rpXZPsc/jQA1mFGh+K9audp1at/RcV0A+w2wh+N/rVqiAjMcBHb6tCYE2lDMdNF+HVPQs2XNLqQrXNq7YzqgF1zU5iiXUTv/uJBxzHhhp0AHI1DZVdpxiPSnlB/bum2boYAx+tCCWsUFGxOPPtCHFdDFWY3ratgHWLhAm/3Sz0Vt2aKJs9wmGBdwlSpWMpYNujSkUpge1apRyzXRnbcS+vDMrO2jXeOgsv0MQq5xyAaZ881HwEfAR8BHwEfAR+DyEQjJG1gyuGj66zp8H4eu4oKSr8WkQy/qisgnrmjRYmb+z52xDRs2xJWBTchHAio1LA8aF2zcMWOdK21cRwQt9PjXNfv0UWlaNS77gnGq1K+0+5SmRTfPXUj2A8YO4+e9BMep6o/cZAqBKlzQXs04GcPVNCrH1JN/ANVsGHyBiAjzOahRo8bfuuFahwtmGonFxT4XvJ5WrCbu4NPO6PrMqjvZX/7Xfz0CeKNwjlNy92obEkbW5QcpT/w2a/asuPcoTZ9QQ1oXWn+W/Db+qZHIsSzn3tU2wMzQdgBL4jcYZfaeMvxLjlPmr3hjshzGrxe3hg0bGlvjgGQf7AdJ6j7lMJQE52b11bTQfq5cefX7eTXb+dctq8/BmqpShE1FihTJRRiIFjkgKMHMZ/hx5aThyk2RwvI332EwDFu3biMP0rRSeox2Y0aPggRikrN7OgcAG+AG3x5UHqZS11blnbBNWJ+vFzxbKFkNAIFCBBnVjOnTXAWVZ6bS0Q3167uuXe8zb1QkOMbkILgC6SAspEyZymRMfJDjcXKLwHbmWf16N7ip8mgjz82gHPoHkREwd4edZVIcdYFkBxkWAEnBgoVcCrFF8qvScaZMGSz/5bvhKwFUfL8F68QYIF+rVm1XRZ4w1pfmbA6RHsirs0sulFLsNYCKKBliU1F5r77X8Gb5WYAOczf4JpX/i8bart1dMhluJyZZKvUfSPYWLVqo78u1YuIcdDu37zTgi+3v2r3TyBOYJnOesP/Ep02btmLBvuGeEEZw1513SFIVoaXFmpFZMOdU6FojdAOIcRJ/rh/yyMy3beu2Ap/2CNx93fXv/73duPn+u/5i4xS278ZgvGH63h5rfrQcezuftY3kIoIsXrzEKkzBQtKQjP2DmXOHDh0NOMpvEikbUqL/SlRgxo6ydi6ZwA9QW8COY8eOmyyoZKmS+vKvZa+jY6P0NROdi4hcopBt2rDJ/aYDh3t0ep0MGP/MmTtXdK4VOminrcw1iBsRtWsqHQCeM4HySQOXI2d29Z1UKNs6t11IGUGGVQPIExEZob/1ik6OKTL9pV49LJzsObLbhVaZMmUd0ivGhAv3nt271LW0Z+rPNx8BHwEfAR8BHwEfgctHwC6sgguB4CLrolUAA+yqSL+hNQNsUEJ3w/oNer7O7vIBtEBVplwtpoFQ/eM3vtfbtWsXGBDG9saFLReCrHe17U+Q5s81uVgNLjQAXxgH4wyNd13sHTuVptUNI8bPe4d17RC/2Tg1Ju4GchFP4zljv5pxhkCdILbxt/DX51t0oQ4jZ9EiVcVQQ4Y0XwASVPGEGnHlriotoPEntJQz0Mne0YH5W1WYhFfxr/oI+AhcRQSQMQECVapUyWQsV7GqX/RyEVBSjfRm6LDhYsD0kGRMVX7UQp/JK1euMmkR1b5oSs9le/F/2DsPwCyK9I0PECAJoSb0ltBB6U1BqdJEUEBBrIgdsJxnb4DtvLOdiqggCIIoICAqzUpVpEjvkITeCTVAKPk/v3e/TSJ/RE7DneIO5Cv77c7Mvjs7O/PM8z5vpMCaHK56dVw5ndsm1sm1XbvYJB5AAxbliRN61qhbp2cf/9l4ARaJdjSTeEgG5G/R9XigaRswDn1ykrxJWrVuac+Nf0mk+bXX/u3KxpXVseQk4oFeAXlqat+DBwmt7bGpcC26WHIdTRo3NlDnyJFkY5fgCoVODIwXohWl6tkFGILbz3vvDbI67JW+69p18Yp4vExCxMXELslhQsAI0s+Ubs3xFNUXUESpatULLf9ogTGwYpiTFy9eXJGLiqo+ByQUf4GAizC1U8+ta83adWIATbJjqb9OVSmLq39RfRMqBtzi3E6oTsyvcbvak7TbWEW4KlapcoGJCKekHDUWEy6KOcTYLCEg6emn+7he99zrSomA8dFHI6S31MLyYn6ORg4pWcwidGEZW3AK2K9YseIWYTkiIlK/H1LY+vHSn20s+zUQ6+YHV7duHdXIOx5AaPTI0TpKqADH6wQAnGAtZc8ebjagnKzZsri4uFgJDA91nTp2VFkaI3hZ8HOmprBMzc2uSKo16EJinEDlQ+Rnm0Jl5VckpjxCIRMFynz//Q8abCUpXFaEQmHPcQViCjgaznrRZMuUKeMaNWkqBG6mSxY1LEqUMFCwFStXm482gA/+gIUVYx3mDIJFJOhLJ9WIaLyobscUiLE8v/32G4Xbymf7oEkzb/480cbeNPQUBLGIqF4IEQMi4Vu9cOEiKW9/41oJyUaFGVpvkAILBBYILBBYILBAYIEzW8BADh7JGkyeDvCAzr9xo1iw2sV2swGW9wwnZ0T/0IxBl+722++wPE4N+Yow5o8KZ0nCjYEB2aBBg9wcARL85umq2M+/+MKgzMrXuz8gzbjzrbd2Fwix2crP+DvnxPePVM/NVs+8WvW8QwO6kxZwIGMe1GWOFptI1BMQB72MOXPmmr7Es9KYOJuEyGT6gPf0R8DKIdpSQkKCGbZ9uyttIHumSEq4OBG6Fd2BM7EJAHwsU50347MgBRYILBBY4M9kgeWKTjRi+IfuBTEwLr2kkXS1uoUm1aluueaTRNKCbRgTE8MMXG442QxoKVI4RvPSjZpnhlnkol697pFW0CrXp09vudsd1XNnsMgDChWtLhowxj5oTrpvzz4TnTXwRltxM8uqfTxtsaxiY8xzC+bPdyMlTkv0JlyJunfvbs8yJIHx/ICdMu27byVKLBcdwQ0womBYsQgBOBITU5Cnl5XP/LlNm8uVf233vKIITZ823YCTLWJsfqM5cMlSJd0BBc4hUjJkhe2KjFS+bFkrh0WHQppPo+uCCytPRqRB2iqM/YzpU/Uc3GIARZ3adRT9OMEYoDBpjot1hM4S+rAIIAO6jFSEpk7XXCMPGdlRKUbuYXfddZdFTGJfwA6eJQBHOQUiwWBCqHjjpo22vx6uokxI0DcmWsyUm93Lcv+aMOELMWorKNpUBTGWRjtcAAcNes/m/DyP58uO6KTNmA6DTfnL/rkjpWsre8GcIUHaWL50qUUffPedASbIvU4LP4B0LFD4erKwYyj/4MH9YhZFiNUDWKMIXLJ/Hun0XHnlVaZxBXMUbSEAo3ORMheY8cYPnq/cSWm1CIwB+MgdGeUuqHaBVrgOuS2fbJHvXLINBmbL92uXmDH161+khhEj42WVbswx0cii1IAOmD/gddddZ2gdvmWs7uBLt0E3ynQBN5eKprVn1x63LfdW8xM7qTJJhAVTVub/CRrYvGlzoyp9PGqkfD0PSRU7vwx/UPXcJvBmn4kSlRbrxnRtDnu0tC8kREwkBhprkAILBBYILBBYILBAYIEzW8BADI0DWHX0184yHnHvvfdm/Pr/PkPl3yJfe1a8ThdulwERIWkpoWTJEg7XKIAZNFQIEYrWQ+vWrbUydvH/yzvjBurGcOWX0n33nVmvxnzmVR4RIln1PDXh6mT1VCEM3qgnwAyaEgBPFv5XrtQwXH41WUUxajqAdeoxROlgoEnqIgr4hx9+aIPRU/c79TsruN98860Jk7Lqe2X7K3+2C2Oul15+Sds8a9WVz36QAgsEFggs8GeyABPvgZrMz/phlqLrfGtA9FdffmV9JG5NL6uPg8VYpXIVb5ueKSf0PTsMFDtR7xXgpl+/fiIHrHQjPvrQdMj6PNPHffzRx2lPlKwiCezcsU3PoDuk9TJPXaf6bf3PJhcfQJDy5coaMAOwMU4aLRAVvvvuG73Pkq7QlwIEIhWxaI+FPOd5ivssgsKACzB3Diu8M3Pr42LsUNbe/XvdBrkXTdo/wV3W4jLTZRqkSEIjVKcj0q9JPnTY9t2zL8nm5niEIBECu2ajyBAbN24SIWGBhQzfvnW7qyLQh4RrULNml7mFAp549ACCMB9PkAwJLJO9+/bYtoTEBLdfoM/sH74316ExYz5xHaQRU7iwol/pOACYi8SegXHy9Vdfaf5/xLxRmIN/K5cs3LkQ2ecZg+YN5QAezZftcF9Cq3bP7l3uqqsedq+88op79dXXBLBUNUIF9dwkUGeChO2Py1UsbRFFYApEi2hdLwgb2J8ruEOuTTfccJ27Q1Gd/yEAC72Yxx9/QiBcAllZ+YBouGl1FCuGBQvcpmR+0/dhUam2ACrqjfaOZarfMjvB+sm8pArS2Bg8/SgmDD5YUJKgGeHX1bpVa4WkLqHQ2SkSlrvYtF0OyH1o/k/z3SIxVYiatEnuSosWLjQwBrZNTHSMa9qksSshKlURXay4uDjTpElR+G0MuE/Hf//DbDXs79xxXdRa0pMhZry5IekCIz7sFBYLTZqxY8aaqFgXhdNuc3kbV6NmDVdUeXLTFlM89hrVaxqyWFVMHwZcDKSCFFggsEBggcACgQUCC/y6BfyBEYMgVvP+08SgjFUzX5Qw4/EMtG686UYbQCMs857ChhIaGzeAAQMGqDSP8k1oagZ9Z5N++5iK0rzB6qnlUHZGdyW/ngymqScHMua4/vrrz6qefh19255aHi7fhLz2auRsjIRoKswh/p4RMyfjZ0AbP/Xp3VvjKA9C6yxNBerKyjGJVeY2bdq4ZUuW2ndWMEuKZRykwAKBBQIL/LksICBeYMsy9Wl16tV1U6dNddd0vsb1uPsuC/dM2HI9drxnllgXiDojk2FAgb4jBEzC3Qyg5JtvvzYw4Mc5P4pBMdS9oAUB+l/yIKHHRbj3hPh4bZeejQAUKc+Y2w3zX56N/GPCjw5KwwYNJUo82OalT0o/CTCfiT9gEUsIPBfp/9E9naq5LmwOQBn0XNauXafIT3vcrj1JxuaBCfmwgPpn9QzAJYvyeeicSDnhKkpoerbYpjBe5miOnpAQb8wcwJK1a9a5sZL5mKJ6pyWdT43q1RT5OL/ZAwYQuq5169aRx0uk7UbUZGz1g4gWSJeUL1/BfSCR3CFy+bFnl57VmKVenXruNjFHoxSGGoYpPybEr3Mv/vNFV01yIjhxsSfnS37r9FtphRzXRbDzpzDC0MNmbd68mZhPl9r1wNVrvqJo+fbHVhA7sglYImS4f/5hWbO5DgJbDoqc8Zpc0QgQ1KZ1K2sL1atVs3OxmqqdwKKJjYs1MoknqS3xb+WHS1w5AWuMP/TfzsEOzOSXTAVmqCeN5YRYLwj8EB+ehgHKBDKH+xIiSNwgHTp0dNfL17qEDF9PF5noShEROa2BzZw5Q6JFl7rWLVs783MT4JJL4a8jpKxdSErWcXFxLq/ijH8ydoypnZ9QA94nihZjuscefUw3yXPuzjvuMsNypzDsoEz84wgTVlOATIvLWriOEvO6UvW4WDQyLmRKyhFTdsbwtVWnsw1Tl8nXJMgusEBggcACgQUCC/zpLMCg6PckBp+Ww2myufe+e138ugTLnsgtLVu2TCuqbdu2AkOut2f9unXr3H333Zf225k+MGb5LYnzZDUQts6pibLjNSDnRHBhgn7uJ6vn9TfYoI59fo1B5B2HMRiunr62b77+emig6411CBvep08f16dvH9dX770FvvDd/4w7lZ8YZ/E7ObOIRX1ZAEP0F1d0Vgb5DR0+IocEKbBAYIHAAn8qC/jPFFW6dMnSbsf27e7WW7obeeDpPn3dYDFpChcp4j13tE82TeCZDwLAHBMQoemnkvpg6+9TjZHJpHyf5pMwL3o/pfDVEuUdPWa0RfTNFnoGomWyVgwZOlCAHcB44JmTWvDneLLFbQYBX1x87r//Xjflyy8lht3JfawIxgStIVF+VtUJdyGiHuWKyqXAOptcWI7sFmWoZvXqLn59ogHsAEwnpI+iU3bt5J71vvr62NKxmpdrFqzZfgFFHGZ+e0weIQ0kCJ81LLvAnBVGZJg/f64BIDsFPuAWbFGiePSojjzv4uLiNCdW5CQBJblVh9t17lX0jFit6EnYgecT9t0mRujcufMsMtRB2TH5UDKmM/sVUhCge3r1MsF8iyKmSu3UQgAgEpoxJuqrvWH6wGSKVDkZ7Y/71gcffCA9uetku1Sz0+233Wp5Y1SANEAN3JBh1DLvZ0zB+WO3gtEFXWOFGT8ujGKTAghcWK26YROIKd8mwAdCB/UiPdP3GdvHG5FQJ0XdUv7oBrHHSeX5S4sllsHvePFq8DsyyHgoQwfirccKkEHQN0x+eTBViMqEn/ICMWEQAipevIRox9uskeQVetZO9Nlbu9+qgUBVawBHZbTo/NGuVu2aNkhANRvhoYIFCkoZOlwGjhLFLNzN+eFHd0DiRRUrVzLft/g1a01QKFYN6Gqhoa1btpKL02Hz0SMcFtGW0KShkUdK7LdokaJCA6u72nXqqqFvljhSgmKmn5QuTgHp4eQxSlnG8ws+BxYILBBYILBAYIHAAqe3AHRrP2X87G/7tXeOYcDK6lTGhL/4kPeH2OCtRImS5hqU8Xc+E367SNEithnwgWN+KTH40nDTfv4t9WQMQUVPrSeRanzgo+QZ6ulNBLKYv/qZ6kkFPR09Rbv4hTC4awVEmdFUHwbHacm+p31L++XU8yX0MPWuHIoCQw6HNJgmsTgF42bevLnGTrKNwUtggcACgQX+LBYAlGASrfpWrlLZwjnvVWSmeQIPLhcjsFmz5m7hgoUurkxZ85JgP0CBA1rsB6xGtwRWIRngqcGEPpsADScXGSb8/3rpJdelSxdXNjbO3d3jbte4SRPPMtofgEGF23dAB7xAYM8IajFwhl/KlSvv/nb/3xTkJsUtWbxYrJm2rkihItbnon3jHZ7q9u7Z6yZM/ELCwa3FfKniUkUywJOkbPlyAmPkdqVnZ3xCvNyWTqieEq/VezUBD/dooaAIwJOeeYAqCOJSfhPVs1xZdGVyGPtkvmyAxgyiwyM/GmmCwgjxeg8X762YtGfMPUjPzpyyQVd5n9wq3ThcdyE4FCxc0G0T8FVfLrrhcmHC5Xe0IoyhA2NJJ0z513W9wZUsXtJkSygiSWyfk6x06Asl7t65O8QKQo8GxhE1zupatmphejOLJWI87dup7rPx492AgQPl9vWSadhyNPZF/61Fc+nxiMGDng3XHzHnbdLWISoWuQHA4UKFbsyAdwe4f/7rZVerZk1zowLpga1kQYpUPjZANiUyKtLAIwnZQnu1fO28Mvnl56Of35251/jLCpHDrwzROho3PnFccHz2oCgVL1HcVJxRmK5atapCgyW7OLk94ZZEQsioYmXFf9cVOiQ9GkSVoGflEmOGARGNMb+AlmqiWOVUo4oSisiFwD99weJFitXeQIjajVYWBiUUWlmJ1t0tESJEA9dLgBhhpfwF8rosObNrcCU08shhA5CiowtY/QoXLvL/Bl2/2zxBBoEFAgsEFggsEFjgPLUAVO/fs4r0SyAFwra/li9jBPRbziYRrtfo1Gez82n2+frrr06z1VnI6bOpJ6uRZ5sATc6UGPf83tSsWTMLOQ4lfuXKlebyjdBvaY3LGNQGKbBAYIHAAn9mCwDE4wqEi1B2uQDBVyGCHZohsGYI54yI/CLNIZk3vj94iCsVW1rBYzRP9PtAzUmZr2bXnNFAgJBBPv/8c+s3Ybo8/8Lzxmx54R/P2xyWSb6/EFBJrkSAA8buUPmUU6JECREXEk3wnnIQwq1Vp5Z7+63+xggZJoZIpMCMEyePuwlfTHApYrsQ/KZ9+3bGCAFAACzJkyevW6xQ6cyjcyg8dpQYIjBGCsiFlr9LGl5iMiMpx1MMlKhQQYCOwBsYlkcFINWqVdMYIctXLHdH5D1SSuyXuQJUYPTwLGCezLkg7JtLYbD1kepbuHtciWAFTZPo8FFFV9qbtM/cfhLXJ0gzbpsAjv3KJ68rJxDJ7Cagq0B0fon5jhSo1dXt2um5zxoDVfluUQhtIjBtxVZWkEJYa65OIJ8rxI599513bMGiV8+eCp+92D0m9y+iD6KdI8zFALUPP/pQrJ8tkicpZdcPtgsatrBwsoUBjmF/Z+7Ezz77jIX8njhpohs0eJCBNcuXLbPfAY64LgfEkLrwggvtnchX3vnbj7ZfZr5Qu0xN0LRA9rLn8DAfXJpQmIZBgxBwMakkFy5Y2Mo8dDDZ1albz9g0S6SYjBYNCs8YEDotGjAFpTFDTPdENdwjEuZl9Sg6OtrVrV/PPfzww66QkMDdu3ZLPLig3XSrpfz8/czv3X6BO1slcrRuXbyoVIcMLUPfJnvOHKY1Qzz4l4SQzZg1yxoQaKaHtmWxaFHRGuQBJgUpsEBggcACgQUCCwQWCCxwvlsAcKtBgwZaRW7mYmNj0yck5/uJB+cXWCCwwHlqAY9F4c/vVixbbuw/nzEJKNJZjBeiI/Xo0cPhJkuUIlyPEhMSDHxgEs7+TNAB3rMKpABEUM7mvnR528tN2Pc9sTeGDR0mEfrnTIA9WvNXwirD4iDe0NhxYwVa7PXsLIYIxxOZiMhQEBZIlLNPGie3ykXnEc1xr+1yrWvZpqVAgaoWAOdLuTuN//wz08HZJRbH97O+F+umnMCZaANijiQfUbjtRDFVRpvHiq+VWrZsGUc9C+QrECpf7JXIXKZ7tnrNamPVxMXFuU0bN1k0JRYupk2dqsjJ89zg9wcrb+8wXiP9aMUhcKa2WDY1atRwcXGxbn3Ces3ZD5n3yorlK8UqcnqWxLk1a9eIQKFIhzBjLGWRO9UlBoiZW5W8bWxRQ1gHOEIO2QW3M4L5ALZkkwcOYvTJAGP6zVyOZNe3337bde16rbE7r5c4McF+KAGNtL0SJYbxhP2TRfLIKkBGkrMmpGzXU2DS5dKbzSn3pEkK+T1h4gQrb/q0aa5z584C2ORKpfIhmbD/UokBr5VnjkfuOTegDKbJXMaMGmyKEClCREJNqiJq7CHFEKdhRIntgq8bLkWFBMzk0TsUJXy9kgSiQDOqWLGiu1bUKFyMAFhoyChZw5iBdpZXNLDyoptFCq0LVwgxaMu4SRGNoWmTpq6oQnSvktEQRoK2lSzgB3Bms6hnUTLwqpWr5c+2U+jdAbdSfnHzhJYSk7yeFKOJDLVOxwLiqGCPusaHIAUW+BNb4CupoJvg5JnOQb0YYCgPHcQ0WTX4rYnyBg4YaKvR5Pdr6feW5+eP4CaK7Wgi0HHuSdojEbLy1gchGPZL0USmTJni3h/yftr5+/md+k7kldOFxkO4fNB7g4yquUwIO6Axyv733X+f+SKfmk/wPbBAYIHAAoEFAgsEFggsEFjgv2EBptTeXI6J/yEt9IdrDunpmUh3RV4SiP2+8NwL7rBcOJ9+urfYE9PlMtRKwMx6Tew1HwRY0eQ8j6Q3jslNCHCBMTP/4iTdUUvuMd/PmGkAAIyLqRIX7tO7j/t8wufuxX+86JWvYxLiE0Qk2GMABPNfmBFEVwJwQD6Dwqgt5eDp8c677woo2ezGfTpWwMFkd/NNNzrG2ArEZHPszVs3iyWywcJbH1cekQJaYATNn/+TeajwGUAC4XkgH8anrRSEh3Eruja4KkGGuEw6aHs1Z/945Mc2twbMWLDwJ/1WwupTqHAh981338pbpZiD9UNemBSAIm2Yr7zat29vgAlgSZPGTVxs2Ti3XYyZrQJkcivcNBqzM2SnSxqJWStcYOzYse5iBQL6VC5JLypKkrkyy66kD4YNVV4S8NUfbkzHxBiiUBhLETnF2KF8/SNS1qyZs9wVinI4QyDVY48/ZjZ99tlnzRa4MgEOIXBcUBo1yKtQAtcfPSAkTq5WiG9AtYPCBrrd3E36t9e5t9/pb8LOa1atUV1Vkq7/l199afprBi5Z+WYJqpupKXOBGZ0kdKeIXBEW3zx6V4zbuH6jqSsPVySAq666ysXFxZkocEmhg0tEu4I2tXTRElehYgWjUwHc1KpTWz5uH0k9eo6MDnMlh4kIH9p/0O0W1XbSxCluppguZeLKuDJCARE8QnemVp06rlyFCoJzTprx8uvYAtKLAWHbk5TkVqxaoXjqC1TOEXfN1VcrvNcPQvHWCgRabReDcN7jx3/qqsvPjJvj3Jg8U69fkFlggTNaYPXq1e4T+Xhac/b6u7T9aeE/3yR6odTXfw8wQ3mIoFnGGQvI+DlUAzbl0wPj95RHVtBR27Vr57YLXbcUKov+hXB3o0d/4h54QAw5+QKfmqCgjvx4ZGhzeiW9T+nfCUd7KjBDvXuKSgkN1E+HBOzyoJysh+g999wTCGb6hgneAwsEFggsEFggsEBggcAC/20LhNADwBTmqLjdoA/D+LeevDZq166tMNPPK7LQbPf4E49LCuN699P8hYoAdIsW9rfY8JmdV69e4+rXu0gRgAUSaHjICJE567Tp010lEREANEBbRsqlaYUW/5/p3dcN1ISfkM9eJs50S/zTp3xIDLhYITzLd7xCiEZ0q4TYH/jb/YoktdTmqrgqMZZ97tnn3dBhQ2zxM+XIEZPlQGeMRc5NcoPaplDd2XSOYXL9SVRo653yWGGRtEmTJlZspObnl156iX32GCqp8hIJdxECaJijD9DCKoyX2NJlpKezUdqvVYywMHf2HDcyMV5hpl+U6HC0nc8xRZjKofm51Ttk47ZXtLWoxh/LTal58+YuSmAR7BXm4EkKALR9+1YzBYycJG0jbPmFF14oF7AXrP6IzONlk3zoiOoIGJbH9j8mDR6wgKNHjomwIZFeMWywP9cOIsZkRZP6+OOP3I9zZ7u7JFtSSXpCfZ/uE2I2eSAO18GLtuXpw1BvyBi4g3FCgDVIqTzy8CMCkGaJjfOWAhIlKYDRemNE8ft+4RCeKxog2rlJmQvMqI64/1zdsZMLD8upBrLDYo1Dyxox4kP36aefSun4WVGeqgrJOiF0LMxt37bdzfx+luKzbzEDI1IEwwYD4CuXpJXvXLqw1eX/lltaMtCXiA0P3WrD+g2Kb75b7Js8Eu/dYjHPuYjF5DIVIUZNqRIK66XQaOsTE02fhrxPHtelUIO4XOgajJ7xn443Jg4+Ywg8bdy40eKfp8OA58bwQa6BBf4bFoCSWVWdXsaO0zpjCqdX0Q/Ll6/QLXFCH1ONEvl76kV5+GFah8WDT/dxqBh7OJA3vpug9dQJVsvvSXS0UA6hPJIAaLp27WohdNGTevnll63j5b26hL4JY5sxAdSSYNTFxsXZZ+pMP5NmJ21F0T1jGjVqlLtNomecG/T/m266yR5Cq8T0A6nfr3q9+WY/cwuABRikwAKBBQILBBYILBBYILBAYIH/tgUYbUq/VK4tFSuUNy0yGxNrBMfi4GYBGp6GaVbzsmh7xRXu0ccec++JDT1xwgSXRUAHaeHChealgaeHjRM1xoWNsU8AQ3mRAtgGKySbGDaLFy52Xa7t4v4pDRcY1AA0PXv0kn5Xso19mePygfH3FW3bSNtmiZXB3DNC0YzRFisi7xHGokfFfMHVqrXYLo88+ogkQGpbdCDKI8oSch47NN9Gs3Xr5q02r96u70WllbpFbBWAEUJkV65Y2UAFb/DPq8AIVQJHK+yBlszDDz8kV6znjbFz5x13CizZ5Hbu3uVyhOewOfO3U7+Vh0oT02DF5SinoilH5crt1V2v5ElEqWLFiiuU9xrzaMG26zVfx2tlt/JiEZXoiZs3bXaF5fkyf95cadFsdVeKcfO1zruNBI5XSOuM/DnHVBmZGrIImiu3tHNwb1ICXENLp6jm/H4aM2qMG6u/ZxTpaaTAoYEDBpiuLL9PnjRF51g6BMRog04aseSD0sLlC2Vh/2QxqEaKPYTuEAu/wz8a4a67tqubKteuIwLD0L+1fcn0HKRMB2aoY968+V3FKpXckuXLzFe5bNmyanSLDSABzdu7L9bNnDldJ3jUaFYIGaUIqGGSBZqHYBETPFyZsodJrEnUsXwCXIiPnl/uTCBYJYTucUOdkCEPKSQX4b0wYD6FA0N/prYaLj6CoJmI2eXW+7XyI5wpytOiRYvUMLYZcoY7FDSvPQKPQCmvENqXDSROFydIgQX+7Ba4RhQ9/n4pES2tljrRPXt2y680xpD+X9r3bLafXXm1VN4eo1WysvB7Em5aiQJeSQ8++ODPWDFt2rSxFYCLpBBPhw7FMSMwg4/wkiWL7diOnTq5oUOH2uezeXlBDy5SNgHR0DEbN25s31HRRwD1Igmbq+c2/9cAmDHTBC+BBQILBBYILBBYILBAYIH/mgW8yTZOQ84tWLRAUXhry4tCkhVKTPZhz6xeu87mfzBesodlN+DgiScek37LLIEz79meJUuWtIX7NWKFMzskX0ImH5CnBYyNcAW2UXaa2Gd1+SRse9llLdwYsdX7KqpdK41FYY9MmDTBPfboYwpCM1s7sq9zOyV8m0VsmV07d2kDbjeCNvS3du1qFxsba4ALtac8ojLBSr9T4akhJKyNj7fw1DEF5aIj1kdxgSEVKlay89q4eZPNdTmf4ooQOPuH2XKD+trdeccdocA4XvnwTlLlk8Q5UT4Ljc8996xp5ADKEJmP6ESbtm8WAFJE4aU3ihSx3n0nkOKyli1dcYEiKUeP2XGWg4dtmNQI+yM78plcuhCSz5MvjwSCp7r4+HVGuDiguTvMGRY+U1MBkJbpfHeaJMHTvZ9yHwwdZuftRYdy5mLWokUrG88DWGF/jPijvGvCZEOzv0CcKV9OsUBAb7z+uvv3v193I8Sk+VQkjP0H9rmF8tDxduSVaIfZxIDC7izGKgKX8gQ3uODCKrr+P7iHHnrIdehwldmjX783Xb9+/dPEijnfc5G81pqpOaNVIVPhgye/rLxqPO3aXWGUrD59+rgdYsisWb1KVKkjQs52y8Ap7pFHHnYP/v3vrrQUsFGTpoXklr9dSTFeLlXM8SZNGkuFuqihWKBV2WT4UqVL2Sr11R06CsWrbKjicQExW7dvEeVotenLJO1OshsQ4CdCqB6NpLgAnQNqZFxw2DGlBMbkyye9GyFwufPmNv87IkoB/gDsBCmwwPlqAeh7ACmAMjwIPhJCHBsbe85OF4CV8pIEylAeoMzvLQ8EmxQlV8a+ffva54wvCJJdJ0EwFWcsu507d6b9HK+H2oED3gMaOuTZpsmTJxvQzP5PP/VUGijjH19fYQNRwKcjmzNnjgHL/m/Be2CBwAKBBQILBBYILBBYILDAubcAY034FqQDig40TLqi+/cdCBXshWNO0cJdAUlfMGbDlSirWC+QAkaPHm26p2HZsmtCv8i1aHGZATDkBlCSRRP6hIR4G8cek1bMSYfnhVz0c+d1xYoU0/bSNtHfIpAEJvcxzXdHaZx92623mYsN5bFAiDs9kYj88k8oFPaxlOOmf0JFqU91jWVJq+QiBWhBatu6jdyALjBtFu0kgCPSFSwU7RLWJljIbcJbAzYw56WcxPgE81wxQV3LgRdvzq6qKNmLXJsixQi/3dXTWBYPFnRxYLnD4tkndyS8WQoVLOQmfPa5RZRCIweXqbS8AGfM5KnmtdLhyivFVtqrkNO7FQ68tdzEFrgdu3Zozh8rV6xCBvbgApYs/Z8ePXu4t/u/Y8F53tXCax4BQzoFS4zZP/9svEKLH7Y5BPanTDQmc0aEe/bX9c6pa7dfIdFZFAVMufrqTm7IkCHGCPLbAhmiPxMenlNRpOWOpQrrkpoJ8ubJJ1ZQM0VuLiCx5iTXWy5Rjzz6qLv33nvdh8OHmYSKAX52juSUuekcMGayyGjyDdMZ4uOG3xwuS7ly5ZGvWF43SX5g8yQ8dEnDBsaKMTaMGDblK5ZX3PNwixO/S42gYKGNBtzgYlBEFKyDBw4q1NcP7sfZP5r/GfmScueOEthSwm47wnGBdOUQsIL71OFjRxS6K8U10mp2dHRBTUD3GiMnp4SfNmxMlO9gOXNpIm988gBysmZTwxRyhg4N/ohBCixwvlrg3//+t7HMOL9b5ZZzmRB9P30oTag333zTvpYX9fODoR9YR+j/zjuiurO0okBP1lOd6Y033pjx5//3+fXXX08r77Zbb7UVBH+n4cOHqwPtZ19xb/pAIQJ5oGZMXnnfa1OqabtQXoTu5cqVKlu0t0jRP0+XCivCGysCHLdt2zZXUKsLJM+NybvH/xNgxltB0cNXomGPK0zf6RLaNTwIYeOdeh6n2z/YFlggsEBggcACgQXOpQUu6vSSnoJMPJmEhJ6vPAJZomXyw3b9/rMnr7bZQusvVoypDkd5ybLTy0lNFG0SpM+2Iq98eAYH5Z97+8cWK+D6P9PFRef/uQv2L17C8/yHEsVLp53hLhEC0oeWqQItDto4Ug1TiRcAjgjXvXt395G0TjdIX2TixAliWxyQhslIG6e+9tprBliQD6BFlSoXaBHuRwvnTKsPCwtz8yWeW05j2Tlic4hbbYt0uDb10QLiwPcGmF4NoAnpgIAF5sx++ckKmsNiI94jFgRHXhxEKs4twAePDhjnsE1gsxC4pqAC6vykeTVqvBHhEQ62+IhRWmiNi5PbjtyNBLRAhIiNK226qnuSdrkZ02e5Dh072DGqsle0SkvVSfn2qSj3rNjYUjqHH82jhUA82XRu8QJ40FNEOHn8+M9d8eLFjVWSLIFd5vR5JfRL94ItSDVr1tY8vpj7TgLC+QvEulq1ozT/3irZgyo2X9+tufeJYydM/4fgPM89/6z0fr53X3zxhfRwGroxo8fYOPqEyB4bxeJh3M+4mnpiQhhLYcIZPHNmMUHj5q1aWhj016VZkywiSKvWLU2D5mmJMk+ZPMnqxUmvXrXGwpObmLNl4LlHwYQvrVDbADYpCgE+eNAg2WyGcIhZYh/9aOBQ6PRCeWXe2zlgzDi3TSrMYaL4VyhXzmKgY8CTqbqgalCgb0fkv8UlA3FjEhYmIR/Ur9fFKyqSELPcapAlS5RwZePKGHpJwz8g8ORQ8kFRsxYaPWvBwgUSB55tK9eJGxKBvgzJxDWJqE1jFJbsgMCZotKVQVgpf4G88g3cb35vTZs0kfbDJbadRgTKCCK5XvlAiVol16fdu3adK5tn3tULcgos8BstsEWRzJ4RxZIEUIEfbMaECBj6TT/qYTN82HD3rtThMyY0Vl599VXdL3TYSa6TXIHOlCjv2WczlPfPf/5s9w4dOtiDg/wAaU4tD3aNV95se0BcLfFuEsLGCIh//fXXP8sv45cZM2boa6qFJSTym598fRmU4GNjYy161f333+/uvvtuc0FaIvfL0yULoagfmqgf4VgS50dHTlQoVgYAbaCE0vcFKbBAYIHAAoEFAgv87y3gTZXSQBkqBGBisyhvUsoTy/68rxpb+2vMbPAS8EraN+LPalJpTzq98O5lpz1ssqkN2sd7FAblm30wkJ/Ogf3Xb9njevb+2O3aown/XzzBbCgdV9IAFT+Ag98sMc3iJYukP1jD5pdsx5UFr4mk3XtcTWmb0v7RHLlfOjEPK3x1r1693CdjPnG5Ja+BPgnBZga9N9AtW7rU7gqb7+rTQbFzIAuQsmqYSPvH4+P+++6TK9JdDokPiAowXjLcTFb+ZoEPlzRoqDqJ0a3yTY9GO5UqVdLmrSyWIpzrpzJxsabLkje/V17FyhXd/WJ3wPbwNFuPu+2SEUE+5KAAndkS8kUPkQXQY8dTvGxoh6pkWssM3eA5s4e73vJ2admqhUWP6tK5i4tPiJfHySaLZrVj53a3bNlSl6K8DysCFGK6m8QQAlRKS8ob9s71ElXGdaxSxSoGKhG0p2ChGAtLXax4URP3vVdBM3TC7muJAtepW0dRnXa4nrJ5ly6djd1Cntb7YFD97dm7x/ACXLnYjv1LyqOGqE/5JW1CmiNB4E6dOpqQ82jNXe655151Td4Jzp03xy1ZvNT28yI4CTgVqLxj63YXJeKH6dIqTxZa0cxBqHjGrBkO17ZzlTKdMcPJnpSBoIPVldo1YrwHRCnaJaAjRVQjJjIIBFsDkF1RdAbtSlgXL6DlB13YI66h3ADKlC3jIuVKBNqHeNGECRMdSOe+pH3WuInVvmb1GqNRbdu+wxpufqF3hRUye5/ixG+VkBDCSRdIiPSoGgz+dSdENePGQL+msoR/t6jxf//99wJt8rvCQhxx7cC/bcuWrYoJH+MuufRSF3GuLB/kG1jgf2iBp59+2jpoqvCcRLIAETKmXGKqEUmNe/G4hMd4ILVt29Y6I1wA8XElhUvNfdSo0QrTd3q2ip8n5YHwk36pvBEjRphYLvd8xvLQwUFlneSVN8pb4bAtZ36ZOHGi+QmzVx1FbQOk9RNCbiRcFhEs3qcYhF5fzaPJ6+Dv00MUIbSM57dZIAyPrwpaTYAxxMoKfr/eiNR7u/mWbu6N19+wh48VErwEFggsEFggsEBggT+ABZiSFC2Uz91zSxtFTZF8gBITGpus6N17ELKXty3jAkPafnZUhuPsu3cMH3mWklXGRP4Z88r4W1q+Qfm/y/4Hk4+6t4ZOcQmb97qefUe6/n2u/UszZ2hXa9euld5pU1etWnVzP586bWpa09u0ebN5WDCurVy5itpnViEpWS26UUREpE320TOBHDB40GCbM37z9TeucaPG7pPRo9wseXLskIs8c17tbMDKUTEs8mePNokNbgQf7mjVspUiCs+0RUAWP6MU0OaTMWPcs32fdSM++tDqRPl7Beqgx4o4rn+/4EkCmYG5LW5S/fv3d//4xws6RioxOkdAjqoXVrWyPLZ2tGvatJm0XdeLoTPQ8p6u6FGxsbFGaEDAOEWuVYPff981uKiBq1qtappN7C4GqFFu/Gf827Hj1QrAk8d9K9ZLTIFoxzj4gERzC2nuDPNn87bNbseinS5SjJ04zd/j4xNdiRLFTGPGPwfsU6tWLd405r5AYa6/d+uzelGh0KA5pjn42rXS+1EIctg0n8lVqut1XY251P/td9yllzQyrchjqjsAGgZfuHCR66XoqNQLFy1dPRt3E9WZcwUPyCpkDDcoFoD/qUXhfylC6/rERGEERdw26c3OnDlDNuS28845GVepsCymN8T8gOsPU6hp4yZWzi033+IaxTcynUwVl+mJc8jE5HW6UdJqOZoiNyK5A2G6rGFq5HIhQJG5ZImSFvklWmGsT4qWlAJaJ1ugD7NfF3mFFKqnTftOAIvAk8oX6KIXFA0rXGyXg0LoNrjtQue2SdEZfzy0MYhJnizxJRoi+jGAKXHlysg/L0WUqhRdvKxGQ1oqNPPfr77uflCIbIRAc4TJzUoXa4eEl5YuWeq+EFUtWTHsywnFzJEjzFA97yJlonmCrAIL/AEsgMg2IAg9Lq42v+SCVK9ePffUU09ajQ8cPGAoP+g7+xNSjvSKoh2hBH+m5JdHn+dHMDrd/nXr1lV5T6mDFL1U5cFc4b4m4lFaea+88qvl+XmvWbPGdbu5m30FEH755Vf8n+zdB2bof8gfMKpBgwYSKRMSbp20c6/L3QuNGj/hS8tDGHrjbIVWbCGBNwBiS/TpSiyGDXl/iMN++L4GKbBAYIHAAoEFAgv8USxQKDq3VtD3uTcGT3QHDh21avnjXX+l2yaoejgyqeLP+z30kAudCNvSXf71m8bb/r7MU0mhN+8zsx8/aXc73n9u2qRI2/TMJwXlMz37z+2fO1eE63FTS1eiSH63YZPAGTFndid5Wnpm2L/gCxGL2l3R3tgcn47/1N14w01iZXiGOKGFwGEfDrdFedpyNslZnNR8lGYIu4MxqA8sENmIRbi69euatsz48Z+5v913vxEQ0u4DNXx0FKtUruSOy/WHxMRePAxJaUSmaRJu3rjZdb62s7nHDB7yvkCWf3jAgs3KU6VvM0a6iBvTyoegwDy4uLxJ1q5ba0wbIh1tFUDiJa+tcMPlyMH81pl7T/ny5dzV13QyyY/DYv5cKb2XbSIfRAp0KlSkkJs3d54bOOg9y5uxMMm7S71Xv1/gW0u5B3USWz2fyAxEgyJ61MoVKxREY4l0Y36yqFXTZk43ndeDB/dbtGTG8pZCefMZzAcJknUJ62jiCuJzUsF88lnk5Z9+mqdx9WWyUxNXtoxCdotAga4j8ge33X6bmzFzpotWuG7PDSyLAS8EATku7xcSEZxgz6QIfwD48rl9dC9InTzyyCOue7dbFECohPtwxIeuicqx4/Q71xC7IYVSqFAR64v8619Ac6Vq1au5ONVpn0SECSDk28syyMQXawKZl58unU6ORpF88IipNr87cIDCW8+1hgn4UlAGbdiwoUVO2rxlo1u0YKEZHkOGZZWPmCqD8BFKyXny5FbDKWxqzhXKV1QY7GJCBXOp8Z5QJKaDRkuqJarZxQ0ustBeYRL8yaMLUbx4MTFjUtxu3RwbNm10CesTjQ2TKjrZXoXfhhkTERmh1fccHmNHDwvoVzRAQp4VK1rc3Kkyzy5BToEF/jgWQD+GDpqblZDP+Gv+UnryySddg4sv1q5ZRLuc6Jpf1lzA6TTrWEGfewip/rWUVp66h+7duxvr5ZeOeeKJJ9xFF1GeE0tugunQzBDKT/lWXo8ev3Toz7bzwGrRooUJkgGUPPDA35WvIiWFEg/qbVp5IGXPHuYGDx4soDfJGDCJWmH4dNynLiZGnb+O/Wz8eAkVj7J9Wa2g/+HBMl3+poiQ3XHXHRb6Lzk52U2dOtU0b3i0EQ0OplCQAgsEFggsEFggsMD/2gK+U9LNHRu4gjFRWujc5/oNneRgWfjJn4QyHvbAET1+7bMehnquMVGxB7ReofzbLE5v5h6l/ZgwWYST0F6oZKSn9M/CcGxekBGrYb+gfOzs29ybov2n9s8jAKDHzS0lAJvPJW5Jcj16j/zLgjP+xBwQoE2b1ua+M/j9Qa5Rk0s9vUHaLuALRqcxatB3RJIakZojAhh4UZK868BiG4t3W8RkGacx4osv/sM99sSj7oNhHyi4TDF5i8CuyKIFuWRXUvok6NIwhqRNZ5Vkxy4t6qEdw3eEgpH9uP2O293jCs3dq2cvATVdLOiNd8soPLSRC6hZFiMerFbgHOoaK3bJQRESqOPX33wldyzc7qk7J6EXFYqciCpvfxeLEXObygEswtU+ToyUAjH5jQgRUzDGlStT1i1fuswNlWsTQJVuYjKy5N+P/m0MWeLmG2+SHkx2O598kgkpJAHflSJe8FvdunUUYnyyQlWPcggiIxRMwg42cOYLn8WUiZFnCm7/uFwBVMFKOaiAHAAyjP8bN21idcHFCKmABtKm5dC5YsMQjhusgUVSCBbW/2g/FnZZPM2hKFlZskk+hWsotCVM+7Zp3cpYLiMU/nripC8UhOiwGyXWEx4A5MtZa3cJB+9z3brdbIQR//rnjspt4/wGDS620zATcdA5SF5ry8SMdylG+eRJX0ovZp0Z+6f5892XQruIsAT1v0yZ8i6PIi7hUrBTPnzQuIZ9MNwUlMuULysQppSLEOKLAjSJDh5Xp9x5oyxqU7my5V20aFRHjx21BoaoUWzpOBMO3rZtq1siX7edO3fryFRD0HaIXTNb7kq0hMsliJRdAsNcDHz2smQTM0YAUDYBQlDN1urGhZUTKc0ZWDpczCAFFjjfLDBwgEdr5L7yXYR+6Ry5Bz4YNkxskkjruKZNnW678tABzDib5Inl6sGkvHr8CrDC/YjGDA8vEkAHnWXJkiXcoEGDbNuvvfz0008CYerLzXG97Xp1p6vdCy88/7PDyH+C3JwQHJ48eYq75ZZbPBfL0F6sKrz99tvWU1P+gAHv2i8wZkj2kFHN7v/b39w7oliW0YMEgIuw2bg3AeqQ3pDw2H7RUoMUWCCwQGCBwAKBBf6XFtB6sI11w8PD3I0dLnYFxZzZtk3gzPs/B2f8OgLCAArw5yU9x0FUlBgz2zBdPyEO7AMsGlnrycgEKTT/0nc/GdCj/clPc6i0fMnrdCko/7fbP3eucNfrllaueKH8bv1mgTN9PpYeR/LpzHxeb6Nt+QBjdmnH3HfvfWJe3C5h3gpuiJgq9etqIVCJMR2gYhYxZo7IqwL9UxgUJ2j7muzTQvGywF0Ifxk8L95U6OQ2rS53lza81C1dusQVFSmAfZSR6TaiSUKoZ6/dKy/lh2cGwWvIEdAjS2pW94qY4NeIPYO+6jsD3nG4OVlS0ZTPLZeo8ezuXUmugLxNwuTVQfjtBQsXmyfKogWLbHfTpKGm3n+54njeHyrMohu92e8tzcv3KqR2RTFUCkjTNVlSIwdNauR7SYls3brFvdHvDWnbpLcTHWr3sxlAn7EnwXK6deumqE0XC6Ta6lpe1tLFxcWaFk/OnBE2Nwds2S5PmQ1yGaJ87Oe/enl6c4L4dQlmk/WbNkg/cq9FR2Z+juQJobYNCQjZn2hO7du1d5Mk3lutajX31NN95IqpuQIZ6gUQWLlKszZR1+lCY87QSWF/tGfz5Y92LVu0ZGfpLyW5K664wo36eLSN4V9+7RVbpOXapoh9A0tnp6RNjtPJqXzyRwO3CEwa72S4zOckeT1spmWdauI/44RsDVLs93nz5rlDWplHyZloR6BptWrXchdUreIOieIfVzrWJjMHRHk6ItoRbgH3SOSnSqUqCk220NCpfXv3yS3pqDVgwmSHaXU7u5AvGDHkv1hx6Y9JAwO1aR4E0L0S4uNtf1bFd4k1g24Nv11yiSf4m7g+wfQ1cGdC1TpSQBB5sspOY5jUqAAAQABJREFUeNvaUpCuWKlimrBnppknyCiwwP/YAtwz3FsklNsBFH4tlZV7Hz6ZXvJ6ooEDBxrS/WvHWnkLFmi3VKHVbdR5x/3aIVYn/EBDfZ/15e+Jaonb1a8lxNSaNGlirpPsixsSyvqo5GdM+Mxy/j3F+GnWrFnGn9I+IzBsYnGqyBK5QpKKyV3SS1kMXH72mWdC39PfqCdh9UgAzMuWLUv/MfgUWCCwQGCBwAKBBf4HFjCQRdMXFkkiI3K4mzterEWEKLd1x173xvsT05gzPsvAJo5MHll5t8QvjAH0nf/687b4MxSgn9CeoU28+cCOiZjqGMrXmyX7nVwyzHK8MshfZQfl/2b7R0WGux7dWklPKK9bvynJ3fXUR39R5oxamf4DtpSvUM4NHTLEfS43pHDpoYz7dIyFsg41R7VfuS8JoJk3d66JAEOhoL2SjosBskXgRbgIAV57PSFhWc0Z69RSxNE57sknn3BtL29rbvEs4q1LSNBRABC415ywOeWJkwqEQ34gk0ox8iJB/3DSxEnGClm5fKWYK0MVovlp+x2Yk/K3i3iQfCTZLVvOeDKLCQuvT4wXOFLPhH13ijVCaOjFi8We8bIOvXGjWlYGJNx+x23mKgTrh4R4L4SK1XLJx5ukZvWa7vMvPlckqomWD+XrLtSe3KN6s7ykvyLgpEnjSwWOPKX9P9O4Pc4RHhxGD+5TgKpLNPadMWOm++qrryjKO97yUzbKCw1aPqApSwhzjml8aWOzcUMxkxinp6aKoa4+gIToMcLCDz/yqFu9ZrVAtl4G0kRLX5ZkwJr2RRgZ4Wa0erxuRS5qAsxgCxUtVsTrU1Q++fa6p6e78647XccOnYxlX/UCAW8q7luFBk8UqOTb/6BcsrKKgQMGAfjG9Q9Vy8rOzBe/D82kPOnAhRaqtgfkg7Vr1053WC5CoG9EW1m5coVONMFO6qhEeOtJHPiBBx4wBH758uXGZIHWVEJqx8R1RyT05Zdf0gr0bBMpOi49GW4SjInWBaK+e8S6Abns1eseqTZfK0ClsssjPzLZT+DLfrda7gS0JS4wscoBiFIkHITrEqvm5cqVM+FP6lK+bDmL7BRTqKArX668WDM5M8kuQTaBBf4YFvBFwOjmUJc/27RQ4Ar3kZ+IhnQ2CQDHPw4Q5GwT+i8cZ92x4OlPRv96eQPlZ9qufXsTGeY4fElh35wKypxtHdiPcIHA4zsF8qIXgw6NAUTqlNG08pk9p+aJ7o512tpv5aqVp/4cfA8sEFggsEBggcAC/1ULeGQXTTWYVKjkCIEz3ULMGTRnfOaMNxHzqmaf/Ye4PZC97UyiSP6kyfuSVY/LDDvZxvR9/F/88u34UN7+5Ne2eU9+L0s+B+V7lvQNqG9na3/AmV63tHbFCktzZituTX89zRm/jQIc9JTLEO4pO+XuAutk8qTJprnY+6nexqowNoTa23dia/8gb4tURUXiH00QfJIAMnh8wKaBKUa73bZthyIO3eCmKJLQnQpU8fbbb7nCcu+BrWHeGaF7oqiiBucQa4fyPWDHyb2oikU1BThA2+ShRx52L/3rJfd0797u4xEfyUOkoLV/7tjckvKAcACJYO6cucYoYW4NI2SZXJFg5EwS02P1mlUWtCNDc/HuJdugqFNy8+ncubPLpXnxgf0H3MWSKwAqRds1d948DnFjIqFSH8a/nDOu/rgJkexVL7gYAXY9+OBDqltuBdKJllbNIZcgwChJsiE75bFCRCvGwLOl77pbAFBoVG/5tG7d2nRnAYQAq9CJySbgY93adS6vsIDDYhelnvDsD8uoebOmEiHuaGP8DbIBZA6iIy0QY+hS6cuCP/g1JNJswrp1tjjK9ed+QTsoLFt2SZjIzYuaaDu6l2jUXNuli0movPTKy+4muWqRDziFf/2TRBKJEQBEPnbtvIIsn8x++fkycibkHqELXllsE9A7KFu+S9C2bULlFJlp0sTJLjYuVqHCymnFPa/5mInJZWghbJUHH3pQPmAx1nC36yKC2hFFiRsI9gygSn5dMD4fsQeDd8MA6BRWnHfQunxC4RLFmgHVQl05SdGcQEZzoT8j/RhuAChtIF800CIK43WBbo6dO3e5xUsXG7KJcLDgnEywSJBFYIE/hgXo/D7+6GOrTFyZsq5Vq1ZnVbHPPvvMAbD4ib6d7+3atbM/f/up75QHW4VUVuXRCZ9NQondylNBGj8qpdp3aIftBbycLj377LOm50LdcIfCRelMblpoxcBk4eEMc4b+43Rptx5GlA+zzgdhAGRA5PfKD9UGmKGHbsbjvc7b21KkcJGMPwWfAwv8qgUu6vSS0WUZ+OFDzbuf+GQryja487YyYOJeYXDl0Xm9/dnPmqflo4GI9iErG195L9qD79rTy1g5hEqw/YLyA/v/+dtfz+sby3Wnnn8L/XXf7Xnqnb7fT0TmymnMmaHjfnBbdiS5N9+frGhNrTVpy5HWWaQ95ziejkadChNS//lMjraPdSZe/v4r/Y+3H4d5EyR7p2/in2a7MBToiLxyWC+m52KDdUKh7WRkBQXl/4f2B5zpKeZM/yFfyq1pj2nOvNW3s4vJH+VfpvP8XW1HCUDlGzEhcglEsMectnW/rbv7acFP7tFHH3VNmzd1M2fMsn35HY1SJDj8BAACqJJN79ZC1R5j5AnSUXqL7777rvtCY+Xp06a6xx9/Qh4ajZRfM5EHdqvdexwI5sTRcnMHQPHLLyDGzFSBQCUk6ss9iasgpIQVAjOe6dPXvTdooHmW+PcGniOJ8QmulACJ4sVLuEVLlijaVDX3w+wfHfPvGjVquK+/+kZuScPdM8896xfjnQL3j2cKh5Dxw1q8PCxiA3PzRo0a2W8LflpgGja5NT+f/NWXrkxcnI19IzSHZkyP54uNEZQP74QMh0mOXgxAxldTppjLT2RELrFbNrvSkj1gTr9M5IuxY8e5Rx9/1NyxdKiNq9u2u0ICyNONlY4OLZGvcmhufnfPHtJxnC6LeACwuTapDkU0z8f+uFPNlXDxZdKSfE7j/7Hjxrk5CsixeaNCdes0AZLopxjbY390NQmDTl+TVTImlE8/dGnjRhYVGoBrnIShCyryFYyqCsIxnu37jJVPX3dMoBGRng8gTaDrZ/2T3zAy+d1rLZmVqXpfDALSVaxoUVOgvkBhqStXqmSxygFBUGiGHZOQEK8GPN2QuZ1CG/ktIT7R3JBA8e6//353+623mVAw9CQaBFSiA1q1btykiblEoeKMTx0RmHzxpuy68cLlS0ZY3bJSjW6sxla7Vm01lENyc9oiQaAcdqEQEa5bp65rdGkjhcouZLSwihUquhbNLzN3rN27dDNx5YIUWOA8sQAUR1/vpHnzZtYp/dqpbRfiffvtt9tuaKggyJVD9xbpVqmkE23tlxLlESKbu6jZf1AegsQcw4Pgiy++sIcSZbCd1YJT078VOckX2SVk4IQJX5wRlOF4mDSIA+PqhKjY6RLRpFbr4UhlKlepnLbLBWLDkJL2JllflvZDhg9GJ9V3HggZRYcz7BJ8DCxwBgvo6cPDVE9U7ymkL/Z81TOWDfrjacuyhP2un3lnCMNzy7bZb3xiL29fffFAGH3nP3+AMnaUMrbjQi9B+VgrsL/flv7M7a/f8Gnug3FzuKBBClnA7na6Ad3/HnOmgSNa0zYDZyZpbJ2i55d6BvvzWgETGest9JXjGJv7id9I3it3DnvqnUloKHGMCQbru1c+eXCPhf5ZHt7+tkWZBeVjg5Bt9f5b7e+5NbV0xRWtCXCm119FEFjtK/QwNTBxs8ACc30JNUuCzrCQ16XztWJERCsa6F2u41UdaMCWMrZx5pf5Qi71TPABFUsJUCFYTSF5WgBSLFq0WN4gdd1WuR6hsdpJDA8NAy2N+3ScQwPxhFyi/PsiLAuiwLssShFXmWtNfmPEEu8iFgfyHLBS0EjFNZ5qpYg4cFmLy9xFF18kjZmVLp+8RGJjSwlwyGZgyObNmwyA+OrLL90RARKhU1HmXj3slY366/3U0wJPSovxksc0aVIkHQI7fKMYKfMlfbBX49x33nnHoiAfFvBCmGy8VEj5FZ0JoIqAGGECSgoJNGmlBVg8UvIK2GmnxVR+267Q37h2EZVqlJg4yI9w/gcVxAeQg3l4sybNzGMFEISF07fe7Oe1e+sD1O517sflDrZLc3OC99g9od/i1651Pe6+W+DMM1r8bWPhtImgZIk+Ssdx/Smf+tLf2InrjTzCs+c05g3GgbHzjMAYPHlu636rMalgWflpjOYKswXgWN+GLS0v/9fMe89cxoxOElVlQmjRXWeTPxFx10HzSunkDuzbb24AhBsDvQJU2bBhg9uwMdFVl19bczW0cmLS0LiKFS8ul6ZSFhps0oRJ7vvZ38vVCdEk0bm0en2BYqAfOnhI/n1zbYWc+PSFCscYK8YeGFaXbDa5Kyo0b9GSxSbwYyG5hJYele4MDaeSVKpPnDgm0aEDbvPWTSgtqZ2kmviTdwEzz9hBToEF/pcWABH204USxjqb1L17dwsjqB7MPf/8C+7yyy936Ko8/PDDxjYBnAGsOV2CAcfgik7sbMu7VZ0hbDsS4QPbtm3r+vbta8g+7Jbu3W8RWJNeHiGxH33sUe2dRah+pK2GoJz/aymjrsxbb71lIcBB5DMmwB5omeTd+ZrOaT89+Pe/y9VSzCN17H+T+C+svozuUrBp3nzzTdv/QvVTZ6ONk5Z58CGwgCzAPePdO84Vjcnn7unexuWS2wEDCYYV3viKzxoc8k3bbZAQerd9bOCQbk7GEEyKYHPZoCb9J/uUnq81bW/ApYPSsuFzUH5g/z9Z+5uzcK0bMX6We2vYNGu/N15V95SW/xf+ys0dSuER2V03RWt6f+wPAmf2un5DprheYllERcil37oXegi6GvU7+mj9k4fepvU/TFRt/K39PHcNO4DDrFMxdkzG/of+KpS8Ps/re9jklaOtHtps+aZvD8r/LfYnlPbdCqXdf+iXLhFBYLk19e97rYvOnyt0Fc7DN2tjXjtDAoNFfogA9mDT6ZYsXVwMjhg36/uZ7t577nUvv/qqGzNurHvowYfdG2++rvbnPXOxzJw5s+Xy0tWAESa53BEQARYKjKldq5bNHWm36+LjJeB7pXv4oUfcq6++ZtsJWrNNIZbRXfTuEe9+OnJU7jr6Fy5Wk11TfWYOHRdXRnoxi1zPHvcYe4Yw1SVKlXAfDB1mwEgORSGOlGv98y+8YDoutWpWFwGisIChRQpgE2VRjb/79jsLGb1DsiJXtb8y7Zw5F//8OYuOnXAP2i/QZJTNfTmnxYuXaK4d5SqUryR3n9luqRjmORXJmAhP23dsN7f+AgKyCopBs0uSIqnSzuGMKkpY+DFFVx303kCTIKkgwkPihvWK1JxsrJr1iV6I76lTp5knTGqWk2KyhLtKVSq5YwJ86tar6yZoQXaiNHcKCEhhJYpAJYBecnZxRwUcQbzAVuokpE9TVpGdN1ugjZICmO4WSDNf0gtXXdle7l3LNb9nvOMhY0RKjS0da23AOzyLQKKjoahWqr39T3WvabGXa/rc88+5p57s7fq/01/snLkOKZbNCt9NShVWgNtmeg9mmzPlJYTjZUJeXhtTRlk0gFRUJTX+/fJdKyexHnzG8slvLbZMrOt8dWcJHdWVgvVSa1wYOIdUnMMjwl1NUbDqCGmExoRY0BKBKVs3b1FjizZ1Z6KsbFi/2SZfJ2XsqDxRFjZ7gS7CrJkz3dRvpwqNXGB0KuhOYWEKpYWlGYwStku6N3t27ZEg0D63Qno3NFouVNLe/YZ8osC8WiraxeXahN9ZAMxkQrsIsvjDWADRX78bORugBMDCBMB0D13SsKG77z5P0BY02cAP9Uj83r9//9OeI/el3X/69WzKI58JEz3QBaFuXwPHylO/QJogkDZjeeyD1hQ9Kg8yxH/7iv75jMAjAJ0+ffqkfeY7qyWkmjVrml8tn1nBuPqaqxWRbSNfjZVD505EJRKIuV8XviPUZm5gOn/6qZtuukl91RJ+svemTZpqdWGPfb83ZDP7ErwEFjhLC3j3aaorzAr2TtwLJsktF/fajMMAD7whS/N3510zJiZNTIp49NkzLHSIbbIZlbfBB1kY3PjJf+axxQAc/we9M6kKyk+3FQb2+9PA/iEA7w/Y/urVLOeuu7KhBvVZXX+BM39l5oy5Robuae5/S/67vuQMz+5ukSAw0Zq2bt/tac4c9lbHrU/J0P/wbPfbP6ve1mfQv4T6H/to/Y1foNcf+eX67xn7HxMHDu1O/2MpVM2gfAwRMoY+/Vb7A87g1lS8SAExZ4jWdP6H0vYfXUfVlnNL6iKt7cmOlSpd4Lpe19X6h6NiVQxWoAncmh559CFHRNHYuFI0XGuKG9ZvcqNHjzY9VM/F1WvTMEQKy2Wd3dBKYaGOCD4I4/bo1VPzyXxuyldTXOmSpb1nMtcvlGd8wnojCXhXlONZZIwyAV3ccI6fSJF0R5JptNau5emyooHql19dxIeuYtYUU7hutGciNac+KKb68mXL3Y5dOyRAHK8gOQvl6vSDNFsO23lkfPHuXNyK8rhbxUoPVyCcVatWucqVKwqUKe/WrF1lJItNAiQ2btji0H/E3QpG+fVdrzPWeIzmy5ERSH8IxNJ9i1ZND2n5wO/ZJhAnQnqth48edsuWLLNQ1mADeLAwXj8owAYGEF4thLgeN9YXY04NjaPpUxTRSsCMuERGxKB8xiPgwgQn6alor/RFa9euc48+9LDOfZnp/dwhUd8caMWC7urWmfbdNDd9xnQPmKG26vuQrrVoVvqd/oe2ghzCd999qwhQ7YQJrHLPP/ece+KJx5WPgByxaiwJPUm/G71NmfWaacAMnQQN65DcjTaLesVXkEkEdak8IcoGDRrsNm7eKKQtUvHbt5pbUokSxV3TJk0EnJzURd9odCPoSitWrHCff/650MtX3I/ynStZoqRQzhwCcw5bVBkuygP3P6BwV5cbCpd8+JDbvWe3W6BJ1h6JDiHeQyz0YykCZBSCC0Vs7gMaM9Q0NGRWKSY8AkeT1TgIq41uDUwe9GhQi+biBCmwwPligS0COblLuSFhnJ0pAVg+9NBDtgvC2YTGhmJI4r54//33zV+TexvhL+7XUxMdryXt9GvAzOnK8++/tPLUadOxPPjggwaocgxUTT8BjgDE9Onbx/WWcBqfAWN69+ktsEbb9RevlQwS5/KxWC+g/aTPxn9mEZh8rSqomyQ6/c8/+8xbYbEt3gvUV0TbOH90dGAF4urF+yKtcpCoA+5XQQos8NsskMXd1OliVygmr6KmJLk3BhM1RREGM2SWPqfiCewNsdjmD/psAKqbnnfufZ6BtpKtb2nACxuV2G77h0qwzxSmPyvT9vOHcXaI9vfeuS+D8gP7W3PRyx+t/dWvWd51FTjDCmf/D6cLnElnj/ot+K/wbpM572a2+9/OOXT/c7/zFy6GDNGaCkbnkeYMzJnJFq2J+5v7PGPy73/v7g/1Db9w/b2yvL6IfNKAl1PKD/U2Xl+kvDKWEZSfOfaPiswp5kwLCQLnU7QmNGdGndehtJm003TXrFujRcW6mkdK1DXUkMM1rkTsFu8O+i3YFUQFRcy3isbJb77xphbri4f2TpWwbYIm8yfcyeMwMVINoAjLns00S30AIF/efO6xRx+zeSeu9SuWr9C4OUzhrRe41m1a21zWAyFT3Rp5kBB8BhFeEuUzDo1PSHCxcXEGSDB2Z1FxkY4vKJcpFvwsTDTIhFJ0TAHtG+uqV6tqLlTIACBbcPCAJ/VxTNqsKwW29Hmmj9snYkL6fZTheR4yCIAK4bqTFFa7oQR1EdkNU90BToiQBCPcQBHZ9EeF2EbLceSoj6Xbk0u6VBF2+zKfz6s5dK0ateTKX99R/+zZcygi2G5pMyaZexbuUOS5QtImCxcsdCeOndD3ZM37v3BDhgwxxnwRSaJw/9v1U61XrVph548rFNuwP7pBK1YuN/szyiFq0y3dbnH9xFr/5z9edG+8/roJE3PBj2reTzQnrjOnS94AadSX/LznVhYby1955ZXCEpJMv2bAuwNs3jF8+DCLDKVDjTHD+7lImQbM0KFjPoCNOXMUZgwGjFalt2zdbP5mM2bOcgMHvOteUtjdKVKNRismhy4UDYjIJ4cl0jto8CADYxLi11n4bPLC9y4xcb2rpVXq2Ng404IpoVjxhOAuV76cqy816eo1ahpSlidPXonzFLF64ArF/oeSD5iw8A7lA2uHqCooMhPrnPqmqHEcEtq5Xwjj+g3rbcUbXYlDh5PPhb2DPAML/M8s4LE4Ul3BmIImlH2mitx2+212L3OPIESGj2vGVEXaUbj6CEC2/Xwdmoz7+KyRGAl/IeB9pgSAQadMOlN5/H5EfcVtt92ue36tlU8dveS/h776b+qB/cePDy7xE6GwAXNuu/U2ExJjG6sA5JI7T25HuGzcsS4Macrwu5+I5sZv3brfYisNbD8i5g79IOJr/5AiPEBQkAIL/BYLMIEiRWq158arLjJwZtvOfcacOSRwxk/MabwBBq0WQUJ9D010/H14ZxDo3R22hz5rxUlt1U9sNQFO26AlJD/pBxabvJS+v78lKD+wP23Ha0t/7PZXv4aYM1eJOaN233/4DDfsr6o5Y33Bz+9/7ue0fkM3PJozgDOFAGe2ibFn4AyMvZ8nv6vxrz+/puWTYVefCeODO6f2P+ya1v+kdzh2c/nEmfTnfHrGQfmeLX6L/T3mTGtXvLCYM1t2K5T2+RutycReZaoEgR3xitRjEYb0FLSnpvoDXHNiY2PNmDAm+OFLLfq1kWZJooCJSxpd6nqIlUHiZ3vVB4DehIREjY8riFRw3PtFwArzXwCc6jWqWymHNTZs07qVG/3JaPeh9A2fkq4LgBAPbDRSp06bqnluone8NoaJtQLzBLAoIjyXm6noUFmyKmqTxqcLRT4AcEFnxa+N957F2NxdxWLBdQnX/LvEGCFycURkLmOCUMBQgR7pt5h3Nmz37k3v0zWK2AS4sXbNWoETe11BeaDslx4M4NFhkSC2SxcWjcVCmm+vF6ECseMnn3zangOFtS2r9GboB2QKI1U8IpBKGxSVabe7VxGUZ0yfKe2afba4ix4jIcAJRU7q9+YbrrcWV2+44Qa5bQ11dWrXlcKIDCVbz5g+Q5iAAguJjWQMXlsszirAZrUiL0d5i03atkP7IIXQoWMHkTeucM/rM8x3c+UWCEPizGkBPA8Y71NfwGKuf5S+M8aHEIKthnzwgbvxppttMXbkxx9J0qFtBntZdpn6kmnATMZaQaMC/KgoVeOvvvrafNOI4FJGrgaHpe2yQZQo0Lz8+QuYq1PJ0qWFtOVyCI0yUfpx9lwxaraZJg0+bmZBIZT4BYLK1albX404wfV/q7+hbFFqoIcOHTTtmaZNm6qHP6lwZf3l57dWyGiKIY2E6UK/pq6QyWrVa1jc9AioV3YXehcIqhQCQZs3b3M7FEUKxk+QAgucLxaYNWuWdT6nE9A99Rxnzphp+8IuQ0fldOnxxx9Xp6XuXPvMlCvhqckvD22YX0scT8fI3xnLs84zVeXNsA7XL987VsTJUB7//937raFcsjImIrkRQhwRMh4QkwUaLxf7B/YclFVcGn8pAfKiII84Goyhr6Rgz7niwvWo1O6DFFjgt1rAi8LEUCnVoqN00ySJCBpoP7yhqCkHk49aW9fP9nikHO5DY72cplB/+MVqHJMk75+3I/eKP0jhMwAPgyDvs320AZf/3c/evgflB/YPNYg/Q/urX6O8u/bKBtae+w2b7oZ9Osdvzn+Rd93X3Nt0HKGU8b7mM30EE/0IuTXd3BFQOLfG5j5jLx2c8e//9HxgEHiTXX+b/+71L/Ql6f2P/5tfvr9i7ZevDsh2UZaWr7/fz45LPw3r/4Ly/zP7Iwh8980tXPFCCqW9ZY+7++nz062JSTaCsOiNfDFxos0/aV40nxQt8jOOxEuCNpY1S5gxROprrkjI5x497narJX3xohbb3nn7HVeqdKyO0h2kY9CZSdbcs4giHG2UXqp3vDfZRy7DIwp47XfH9h3u739/0N0vKQBYJu3bX+WKyWOEOsz5cY5FRvLLP6E6EXkoZ87s0kBJkQZQPjF0jhuAcVhz2mXLlnqV52BLaR/kglRZbJ1HXclSJe08i4nIQMSkTRs2KbR0KbdT4MguzYeHvP+BESJCt5ndY15WakPKDpf9G264TvN0saoS4g3UgLywcNESC8e9ZvUac826/dbudu8NGfK+yQzsFgnCIjlp7s79CNgRrXH0I9KlBNBYI+LFhVUvVLQrRW9WPVjQLBtXVuPovRLfPa6/k278uE/dPWLuoDEzecpkV/XCKgxwpPl40H0yZpzm6Uc9Bj8Yi1g8eMCE47KkpN7LsIJSpUu5r7/+2oEJJAmcmv3DbGNBAZqRUrWqzMIyWjq4hnnbGAtlkxxLpEPbtpRsSF+ZTft9IQ+eqzpcJVwh0X0igK2oIjTZxbMjM/cl04AZGhRNg6hHRYuXMJ+xY2pAhAoD5WtzeWt3ldArEKtYATGoSUcKlUcQCcYMkUvYhmsT40KjOYluVlhiRoj9oIRcQiwYVrnzIN4rpssUTYTwYYM5U6XKBWLZpFhDO6ab78MRH7phQz9wSwX04DtHSLMIuVAVLBjj6kscFLGjOrVrKPRWEXPNIE47Psg8BBBUNJo3X4IUWCCwwHlvAbSuqooZg3ZMpUqV0h9SZ3HmsHA45rLLLpPaf/RZHBHsEljg1yyg56mePzyBeK6Ga5LUrVMDF1MAzZm96ZozoUcUT18GQbxzXBpaECqGPPiNV3LlMDuUMtif72mfQ/uFtlueer57+duu3v6sVnmHWt5B+YH9/wztD3Dmug6XaCzqJAiMW9NfCZzRfcztHUrp93z6/W/XkL5C9z/MmZsEzsQUiNIq9L6QO6WnOYMg56n3v/Uy1o/4JaS/Z+x//K2/Vj77+f0PoLM/iWR7UH7m2B/mTA80Z4oWEDiz29yadu89hInPm5RX3hJTp0+TMG0FzeOZNQLAMKt3xsDApR33GFob809IABc1bGARfWFy/SRXm2433+waSvtwxIcjXJMmTWzueVI3E7fTsGHD3UIt7HE/8PzlmCNideSUmw1tFnYHIADRmJiXXtzgYpc3b5RbIlCggbw+0KTx7w/YsrgLFdBcFSkOmCQJYtMk6g8d1ESxfsjvmLaTp9VZBVMPEu+AMbhShYllA+CEhsp+yYyskjfIboXGnvLlFJ3vEfeiXH0O/8w7hMoqAzsP2UGivN0FvJTUvDsiZ7hYPuFijVQ2Xcadu3e6tQJZiLScL18BzbELiYSx1L3w4j/dF3LfYm5NlCtjqSs/BIX//vcH3IWaqxfXPP+EtHNg4KyTrus0XZsf5/zodu3Z6VIERAForZDUyK3db3PfTf3OdezYyT2hhWAWStGvtUR/oHyPCCeIiJB4tYG+urbaHhuHnu01Zv8lS5a6fm/1c0OHfaCgHG+4N/79usuj9qCrpDaQ1a1bt9Z9++233jnruvls5b3SpIUwgv3pa1SU6eJcIyb9q6++auHGbaNXm0x9VRPMnGQdrLKi80Q899D+Q9JaWOxyK4xXXv1Fix1TMMbz4SsdGyffr386oiUReguWCvHF8eMrW7a8UL0S5uMVIWSLG+rk8VTzC6sv4ObviojS8JKG5hqRQ3QpQnHl0LGE+4LiNFFqzhuE6uVR6C8aaJJQOFg3jeUr1+jSxkIal6nMrQ5FaxSsi5coYcDM0sVLFRe+lOKxlzD0LXOsEuQSWCCwQGCBwAKBBf5TCwCe+EMthgQKHx8p9wJpzhQMgTP9JAhszBn95kEt+uAfonc++l89Jo2XDxMukg0EM8x00jQf+I3VwFBKq0no8PTt3if2DMoPGcc3m9756H8N7E8L+eO0v3rVy7qucmviXug/XMyZvxA447dJ6wHO4v6PRHMGUBjmTAgUpt/J2B3419YmobT7U9o/8wOb+fAa6n/Otvy0vJVF2qFpuemDffZqE5TvWfk/tb8xZ25EcyZagsC73N1PnV/MmZViNCdL+Pabb75x199wo0CGCGOO0haXSeOkroJL4PIDi4I2BpjAPJK5qFNUJhr7WIW6btfuChEN9prkRv26dfSc9MAdWOh7xP6g4aNZcuzoMQMiDP3VUyBM20AR0G4ppbkqbJMRIz5yU6ZMEYgxwcIz55UUh18+zOuYmGjzBInIGSnQI8ZAEEACPEtIs7+fFboHVTndb3pVYuFFn1UW2pCXq75NBCKtlY4N8h+EoabKCesSJNq7QlqTF1qQC9jt3sMqiwXPISfLTy/kdeNNN7rWl7cxbZ7Y2FhXVNovRFAC5KGt1ZJsCfbgy1ZpzD768KPuX//6l8sjrRyfNcR+lFFfJIyGFzc0sAmCRa3atU1fds1quU3t3mtlECa8kjxucsil6+mnertvvv3GIeSLGHP16tWNOAGLhefqxo2bXONGjcS0EZtPhRgGUaS4Ii0tcaXj4ux67hc55Cm5Wt1yyy3u5m63uB9m/WCRr7j+hCBHr5K6YRuuP+HCs2dHaFjuuSc8Y2QX6aSNQpYD2j355JOq83rMdE5SpgEz1A67c/EPHDgon7CTRpcivjor0YSlLibA5pjQrsNC7+LKxrnsAlSIuoQYUU7RkEAt88q3K4fiihNRCZXnFi1auC7XdnGNGjV2ucIjdaOUNCFRxERB6RA0wtcNtgsU7X1C4LaLMlambBm5PNUxOla5MmW5gtLVKCRF6UVuuJAzfPpyapUcihU3IdQuwnw1btLIxI72KbQ3LKAgBRYILBBYILBAYIH/pgV48qQ/frznkA22bAUbcCbKbcG9gGhNoagp3jHewIzPeuRZJkxsWMHzcvG2AcqwPW2bNqdPmPTllERWlM87ieMBcjjeP47PPDOtWH0OyjeDmJ0D+//x2p8xZwTOMBB/a/i0vwQ4Y/crzdK7c/UpPfn3sbfFv9O9bzBnunVoYNGajLEnIfL9B70ILz87Lr3Tsvs/i/UH6mfU//jJY9f7NfG3pvcj6Vt+/on+Jz2liwcH5Wewy++wv6c501KaMwJn5NaEIPDupPODOXNALjDXKPLm6NGfuLf7v+VeevkleXd4IZcPSHqD6Jpr1qxR82Ji72llEWkJ4AAPCswK02yD3JUub3u5GzhwoOvQoZP754v/ckWKomvqJc/8WWxOi9Yg7A6uDk2XSf5JLXg0b95MmUnPUHIed991t2k1Piw3n+HSnsmdGy0YhYbWcV8KtNm2bZsYG8cUyKac2yldF+qKd8fuXbvdJ6PHah47zStY+XFHWcanvEXlinIPPfyQAI1qEt7d57pqPk20plLyXMkmRs1qgTaffPKJ2yldGiIOwdZB9DZtAKK8cS+MKxXret1zj4E7zL+LFClmpIj8cjfauGmjgTAE7kHkF0bP22+/rfO7KyRqrHFBqI56s9/7KHpqDcmKoGMD2xztSMR9YeXkk9sZzBwY7LCM5s//yV3T8WphCMXcZwrGcfnllxuIwjmvEeMGGZLDsgv2xQaEHt+wcYM0YgSs2cWTy5o8b8Z9+qkYShcZK+farl3dFW2vMKAFk3nJEwHm2uCO5Y1xcNGUIHPuvC42VgLLAoZIO3fuDtncOzIzXzMVmMHg+HrBgkG9GVYKyB+Kx/ny5teFLCyx3SMWn/y1V15z26Ujk6KGcFx/YUKnCooKFVcmznGhMTRGwTWqhhrUDddf53KIQUMj2qrjsioyDMfg15ewLt7CbBXXRcMfDsoUSBuq1YiUtmvf3q1ascr8wkqU8BoTGja5hCg2aNDQlZYvmuc+lWLaOFwUhEsRTwpSYIHAAoEFAgsEFvhvW+BU4MQb4Tn5P7OC3VAM1LwSLZQwpyZJhw572g8MiBiY8GrjEX33hH/ZQvKHb9739K3edttFGxmI+MmbxOlIbfcGnpTg6VDY8VaQl1NQPnYI7I8V/gztzxMEvoQG7/oNP//dmqx7sPaZDrLatco4xdCGn4EdoY4gXIy9mzoACudxW+XW1G/IlHStK+txdFyo/+EQ6yv03VvgpBRSev/jb/G2/nL/4x+Tsf9hwO7XMWP+QflY1bPsb7E/zJmetwDOEEobt6aRci85yCX68yY1rWxZsolZsdE99tgjrnv3WxVeuafrIoFbhGppX3PmzjFBWVtY0EI9c9asOsYWMDCnLWxkMf2W6AIxxnAZNXqUdA7buuG4Jl3cwOwDAInbCyDAYbky4WYDC4a8kPFITj5o4AFRjsiWuSZ6qJe1uMzmonfeeZcrW66c9s0modw9bq3cbHA1ipdealFJeQCabN60RTon8bbywfx3+tRpDsZO6KqH7jCddIZbKjY2znXteq3DDQdNlZLFS0ojpYiIDSsFNpQ2PZq5c+dZXXZozn7o4CEF7tlqtuHEciJGrD6SIB4tWrQU0aKyRSeqLMLFN9KRBUDaJ5FgCo3Knds1b9ZMbkwx7kvJjQA6eW3SA2fIj7rljopy7RX5qHHjRgpJrchU9eqbFEDp2FiXEJ8gICvM7KNidW66RvPmiaDRSJGg5rixY8eZexVz+JMCgwgcdETBggwA1gEH5V4lPE1YQLjsjamy0cVLZ7aMkTOubH+luYS9O2CAe/7550z8l1KMNawdUySL0rhxkzT3Nrq1nMIVtigyVu3aYgaRrF7ex8x+TR99ZVLOdAapEuql1gAn8+bPdxMmTnDf/zBLg8idhrAlJMQrhO3TbpJEfRD8oSFnz5bdxHtyhgtU0XEoVSfL0OvWJdpxMHEAfXYJ1SO8WJjQx/LSloFVQ/vLJl4TjBuoWRUrlJcLVD5j0oAQwsZBNGjKl19JCKiZu+vuO6WWXU2MnexGV0KDBjQe/zuEiJo1bWqhvjLJJEE2gQUCCwQWCCwQWODsLcCDVA9+Hm7+FMomINrGO/psN3Wo702SdhCtSeDMoSO2v41IONQexvpAPrxpwOHxWbzvbGcXRixWhv2uDWyzgplUZSyflaNQvuZbr2Gr7Rva3y+PERE5+N+D8s0egf3/mO2vnkJpX38VmjNZQ8yZ8zeUtne/ejctQIfd66H732ufoW0Z7n9rvGzWpDPSwJn6rrDcmrbt2CNQ2HOnNDqBMrf8tR/pdPd/WmfE76HyKdf+/UL/4x9jk95Q/0OYXKu83//wPSg/U+yfS5PZnt0EzhQp4BI3JblefT5RKO0/MXOGNmMNHY2Xk27kyI+l8XKRReLEhadDh44uXFoq1p70yiScaL35jCDgubdos5phqs05u3TpbF/nSBOlS+cumr9mc5+OH+uuv+56A104nvtg2PBhCmazVCXz3EwV8wXt0lRzT4I44NeJfZdIULeGIjgREnuoRHTvuP12gQWIEWfRHDWndFljXVSUIgrruKrVLhS7Z7X6K4XJlkbrCEUJeuedt61O6U9rVcLOm3p7P+Ghcueddypacg0BQc1t/gszB5emHWLjrJamSxkFyBk86D0FwzhgpIlNm7aazg3mQ58V1g9/nGB0dAEDsACSCkTnNxCDc0EMGcZNec3Dr5bOC+AS6edjD+/sCXXdXGyZLtd4rksVK1YyKRSCB0G+2Kzyw0TwoI/A/ieFLVx3XVe5Ej0uPdpi7nXpxVSsVMnOETIH5WP/DYoUlQ/BZNkb+2MDzr9Zs+ZmZ5hP02fMsBDYPSUy/PWXX3vno/15SlHOO+++6zh/z72MsdBJ07vNJa1a6yvp51TeuUiZBsz4nfBxCSgdUaNmxY1tnMwu+aHNmzffLZaAEj52sFQAQ1JlSPYnFdBFRotmkhSzk+RGFClBKgy9f98eMWS2iGa122LBE1ud/TBIIWnEwK6B3kSEp/j4BEMqI3PlUiPO7RYtWmSRUrgI7FOquKddU0JoIWLBuUXxAhiKF1AE7Ymw2ZTZXnVs1PjSc2HvIM/AAoEFAgsEFggscEYLMOji+ckKsD+AY3zlDVD0Qb8xgL5JmjOFYhTSFuaMVrAPya3JHwDZhMcmPTwu9cAkPwY4HG7fvU9st2Tv6b9TPvUgsYc/ebI8tMH7zg8MWsibVWz29fIIypeFAvtba/ijt796hNJWtCbcCPoNnyFB4PMXnOHOtTteL1wX7l2S3cP2yb6xIe3+Z4t//0dG5jTmTKHovJ4QuUJpHwAU9u//UB/j3/9ebl451ndog9ceTle+13fwS8byqaFfPnlQZb7z3+9/gvJD/W8m2D9K17jHzS1diaL5XKI0Z+56etSfGpzx+2HkKurXv8jNnv2jGzd2rPvuu2/d4PcHu77P9LV5pbmuqB3jPpNP+ii5ZAfmsLoTrElmyxpmmqhh0jel8S1fsVzAQnM3fvxn5orUufM1mrtKiFa/4Ra0SS4+JHs2atLveWGIQSMwByyRTAvLk6RFyxbG6BmrOg0eMtS98MJzAlvekYBuQQMJatSqbd4fhw4lWz2JFhQXV8ZCcueX2w9AyC7NkWfMmM6dk5b8u4kNXnnOXdLwEldL+eEOFWoqJg68XZGIKwvkKKvQ3ytXLpeNZssV66jm3jvc4qXLbG5dXHNoNFY8kFTCuRL/JfAFoa85ZwCNoyJVsB+gS13JiRBYI2P5XgXt7tVmAKvjLjYu1rRoiB71o5g7FSqU0xw+StfhqJEnnBgvHAfxgvN97bXXBE6tdVUuqOLGjBljrlCUb9dKJ71p02a5jDW3OnNNuf7hcrHKJrDnoosutvqgTztq1EjJlzQxDR/wgtjYWIX6zq6cnCJDb1Vo9dWWJ+WjO4OeLcmzm/Zix3OQQsOu358zJ04lk5OT3dq1a13Snl3WMEEooT3jY43uTFH5411cv74rU6aMQobd7Qhfe+L4SUPhaOwfiBa2bOlSQ7bq16trAr2gXvifmcCSXIyqVqsmdO6E27Z9m4UB4wJGimkzU42SkGbzRHnKI1896rRz5y63aeNmuTO1c08+9ZSFP0sUELNF7laE0j4sVk6Cwl998cXndhOhEk3obTRughRYILBAYIHAAoEF/tsWsFUaPTn9SZOVr+cZYoMMaRh8MbmJDIc5o5C2EgTGrekN3JokzOmlEDXXjggdo2cwYwlvWKRXb4ThleM9wm3UYQNJ7adHr5XDMYxGjOqboXyb0+nFy4/Jlw0D2Fspff+0OgflyypYJvQa2N9airXz/3H7q1cTcKahJk1ZJQg84zzVnOGGTu8DTr3/7WKoTXL/k7if/T6Iff3734vWBGMP5kyS62fMGdw20u9/2rhNiqy1Ww+i3zOUzfZT2r91bKcp3+6WDOVb3dhZ/ynUu59+3v8E5f8+++PW1OPmVq5o4Xxug7k1jZbmzJ/UrUkNiKaCbEULASnc4/v27v8/9s4C0Koq6+OL7u5Q3qNDJCRFEAvBBnRMBDFGEWOMCbt7DEQdCwGxFXEMQGWkDSxQpOPR3Y+O7/9b+5z7Hoh+OjyM8ewH9557zq6z9trr7L3OWv/lFhPXXH2t3G2ulXJliJUqWcp5krxvCHcl9ZLBS4u39aJkkYLHFC8uawx42XnVZOFykV16aS9rr03+lVdcqWhOjWHRiD/FibxgUR+A2ihSuJBfCngoplDNtV0hQrAb3J9eeOEF66woRB2FrzJBBg2EfcYKhH0u4LXzMubLuqOcerTLMXAwOlgvY4YvPp/ge9+BAwdqTy2Dh2h6RF2PJqZPGG//zDPOdFxVMGFRoFSoWN5GfjTSIxF9QvhuuTTNzZjnxedlzHW3owULFwirtaIs54r6/RO4519P/iu4MYkeyApAh9etW+1GDl9++ZXmZm53adJBTBBvn4pR2uBmhCIMVy8wb6DpiP985PM6Ny5f0FiZiVwFfU877U+SK7iBrXWsmdmzZ9sHw9+3Cy84P8orQF/dP4qWDPqvvN626sctDNwrFDTIKvhgwmcThGXTwUYS/UnWUz1F46BcEzSL9AwkFXV8nwoVKkivIdwg9ZnxT9HWc+XcR44pZuIugRkD8vM8+fMtEXo0aMgrFQILjRq+YKKEM2geKWEA6G2rCEsQeI0sYVDQEMYKX7Ea1Wu4L9uJJ5xoJYsXFTNmuKYMU67iAggmtvtiTZBDDz1UAL/NXdtWTKjWKwTIk6G8latUseuuv96OF0hQxrw5rr2rWaOGQH5X2vMvDLK77r7LQ3rNnTs7DJoGadPGzbZu/TqbMvm7AH4U31TynVAgoUBCgYQCCQV+QQqwkAuLktAoix5f4LGiUGKdw2/eYIeoKcW1SVIobb3B3qBnGQu3sHDMlp/NjZdmTREfhXbCAkZLIB2gAKKU3sH6nxfReerbbXPFisUXTnSGayqlb1LSfkL/3xv/tWhay86U5Qyp7wuK1jRkgh//z3wwN5mzSr6xjG7M57XEAd/sZIICRr/5i+dztvlPsULCuupxKoDAWXInWM7sPv9Tr+oppLqCNNp7+963n9i+9+4nyJ+k/ZjiP5/+xaRE6N2jo6I1lZJb03LrdcvvVTkDJ5tvuMd/8rHDXvDsxO3mmWeetvaHt3eAXaxn6tau7fvFZYryO3zYcJ8PsXKSb7w8CuiFiCaQzw1cjaqnp9vjT/xLYZQf8j3sa1LqdO16KhwvNuW5mdufpUQKKqcgNDtkWeLPR32y5/3qq6+Ef1LLz+3QHnqUQH2byzABrJXhQ4e6sgQlAhY3bdu19WhRhM7G42Pu3Ayfr7OUFyUL+T4aOVJWLt/4PcP/zgH+ISpE87mY9tG4ZfW+rLcrJ+rWqSevk1nyJCniChMAeL+SYkUoIW7AAEjvcZ2Os2HDhrrbT3lhwmK9smVzwLebMXOWKztQIKGQ2S6LlKKqa936NTKgKOf0+myC5GnUPp3DGKNosaKa9lJF+IIHPKBc9vHHH7vxBJGSXZGsc9xXIeG8YKSRW53ib50wegilfccdd9h99z9g3bqd69Yv3Opbb73leoNdtsPpvwOFiuogxDnYtXQD8YEeYp4UOJf1vtyDAnU/91wbIpBgvHhIQLMgA/GmCThBmRp66TJoROX3R8pRxQz9LFGshFUQ1guRlbYI+ChT5j+zZs2SJu4jha+aZBvWrnfrmAIyK3IicWdidsyfAAwG7ZmbnyYfOgCFimjQ8sp86FuZUh3b8ViF5TrEpim0FR3PIyya0mXKKMx1VUtLT7cWrVo4OFCdWrXclYkoUE2EjD1K4Eh33nGnDRVDVU+rLpTodTZj2nT5Bb5lzzz9rICJi9slvXo5IjbmZ/Qb5OskJRRIKJBQIKFAQoFfmgK7tHjA9DgsCHn6s6GJVgL6yUIBMD42UIS2LhRbzuBesGSVA3NiOUO+4GdNWSUWP2GZphqzNklcUlZvj3CTusQ/qW6icnEpXaNOspDBFTiREoe+co3nuRf2OpP2E/rDub8f/muF5UxnYc6Iifs+/z8WrQldLPPT56Y2KX4UPpi6DhTOgU/w78//MPfj+Z9Lbg15rfspspwpJ0BgWc48hjul3JpcDHgbkjGST/H4h8hKcfuSDXu0n5Ih6gJpN/mDWFHF2eWPK5D2kD8eCQoxpPLkTdrfN/o7IHAPLGdKWcYCRWu66TXBTWz08fldfPBcijoK/62W10U5uczA67ukEMDlZtTo0dahw9HCGZ1m55xzjqIY/VXgtMUcmxSrUX+gwVM6xpWH/WdQrZhVlhcIoZzzFcjnEYZuuekmd5V6/fXXFDb6bwLOVX4mhP59IQuYd95+27YKy5RqsR/jKbxOmC6VK1cMz18pD/JKAbNIEYu7nNpVWDUvOE5qg4MauCsNGC/FhDdzqtym8OyYrv3wYW3b2uKli620ID1KCpsGxchjj/V1dxwa2qL9LAoNbzR8+CeU6dSxo10gTJvly5daCVkM5c2XR+5LK2zu7DkeohtlD25DrQ9trbm90a644gq7+647rYgUMPc/8IDVrVPbvWEWLFrgFi8btL+uVKmyKzbWSnFElKWFCxa42xVYOkvl6ULi/mmfsQEwOG9eURUyiVZgw44RBkx5AgZJocIFuBgPnJUyrkB5lAusG+VFuXKLXNG6n9ddlkf17Jq/XOV6AOoCq9ZlimrYIkDmworsjAtTHo17TP+G0jn0PP98KblWC1j4U7tMiiq8cr4W9AqhvL2PKI5U38CBg1xphFKOpVRYn6nyHE7wV44lBCKAO6efcYYdItOoShUrybqlhLRvhTw600wpaJYuWyotW2k7+ODGHgEJkyOYPbBnELpMhG8nfevuRffKNWm9rFjwqStTqoxVrFzBxgiwCQZq1bylTzIYB4VKSeHW1KhZQ+jOLTwU2Psyb5ozZ65VqXqAgwEBRIxLE6jSWxRnvqT6hpKGmPZXaTDTZKWDu9QyMeWOyIQpx4iTVJRQIKFAQoGEAgkFfgIFcu3UU58lARsLX76FZyRFeV7yn4+wSSGv3gYCCCzMmXJgziwVMKdCaW/wTRJP15AnWLuoTp65LBZ1QFUkvr1VrvnSM+SJc4Ra9EvX+U+B7O2HFuivLoXVrPJFfUzaF1WgW7QUj2gHIaEXie+E/tDo1+e/OJR2Hq36/6dCafsk1Yfzn+jsE9nZzz9i/kMpRfINZcShQYZkn/9hbhfCYk/RmsqXLurKmT4CIl+3QcoZlfeNi+RMLH/c/F+/fIw5/4Pth63Jbu0jr9QXingxvwed8/Nh7tCOTrsyOmk/5+iPcgbLmSoVhTmzaKVdfOPvCHMGPtN/+GKb9ptYr2DlAq/AX2lp6dZVVhdTpk6322+/XW40/7Hbb73dXnzxBYfKQGbDd/xbuXK5598m7w5pI1QD+Khl7EuFc64v4FqkOQqfs846y6MR/eWqq6zvY4+7BQrtk6ZOnSaLGZXVCcqvX5+pfqhPAiCO5wt73T//+SJbJSXEgw/+UwC931kB9blNm9YedZjHa91adezGG2+wmgLZxUuly8mneHSh+VKC7JCysk7tOh69CC8RlBng0OAR4q1GbdMHUqODG9nJp3Sx0sJsnSQokRrpNZR3rS3XfnjFyhUOAdLooEbC0znSIyA/8+xz9uXnX7il0Kuvvea4NdsFDcL8xmsGRdP2rdv0kmip0wNLl6lSepUVgPBMWdZszNxky5cqklTUPrQtLMzX/DLYAPQXBdrGjZk2WPgxWCjlEqF2bJPlisQCe/2ChRRWXHmgX1p6uitk3h7ylnvCZMpdiTEEjJnqeTHC+GPBQ/+I9rTDxy6MP/QAwsQrV/6JilrVvl17WTF9acNkrXSGwosXkIUNlW3YII8aKcJ85KmDju2HlKOKGahHLPESJUsYZk7pIthhclU688yzZG7URWjLlf0WMCOqJXAf3JLWSjO2du0amS0t9GNfyOlhuEPWM9s1EF99/bWIkWnnS6P1+huvybTqa2ngCkgLmd+VMPPnzbdx48cp4tL7NnvOHI/klF9RmApJGTRp0iShYk+yiy66SKBB9eVWtc5RrzGdwhSpeo10a4UWUAM5d+4cyycfNixllsoFa70GIEkJBRIKJBRIKJBQ4JemABYzvJLxhZoWDmH7zjJD6wfWAr7B55sz4TxfhQrmk1uTNkm4FyyXWxPKmRTmDGXDQoIvd2XQQby0YJEaFobUGOr05zFNKHGOBQllQ784GbXth1EOXaeduF7KxilpP1AFGib0D65xMZ/81vgPQOAzhDkDz4ZQ2v8bgMA+t0V033DCiEp8cp97m/+ewT+Y62H+p+a95j9lwZxBOVO2jDBnPJQ27pSy2Msq7EfZ539Otu+Vqy9Y2HBLqXaS9nejQIouotHPpT+AwJeinHHLmeV2yc2/I+WMqMATablwU3BRCW68yJ+AO3OQrFFEEQd4HS2s0qNlPdP2sHYeuadRowgvRjkI1YzlCS5QMLf+OfbJAkVTatykiStEcJEi3f/A/Xb22WdZzVo15Xp0gCxTOvn58IECB/ySXe5RUlGWIeCskChfpFgRV4YWGhQAAEAASURBVF5UVTmwUMFKPf6E4zyCUhvBdxDqOrSfy4Fwywv7BCuTqvIe2bptq/Bz1rq7FJGGCa393HPPCex3hd/f8hXL/dlD264ZCdPaIULAsKkhyI+ZCs9dRDgybQ47zMaP/9jdeMZ/Mk4eK02t96WXeuTk1bIwaSEYEaxM+qn+iy+5RJY9AvrV7WfIveqTTz9Rv5e5lQtWSVOnTHHXLBQ9C2Vd89jjfaWsmeLdcELq3sGRueDCCxSaWgDD0n1lCrOWCuFbV/qIFtTFigg6QWkwds7r2dMVOitlDXXfffdZn0f62FNPPWX36xhMHBK5h0rRQvtAoegJ7OOP1R0RqXCr4ji3LHHo1wXnX2gP/POfdsklvey6664TGHQJF3s7tjNO0I4viJfzKWcVM+rfMik1hgweYkuXLPYbJPpSw4YHCWTpWuslcKTgk7bFCYBWbJNMuvATGy+fspGjRrov20aZS2F6hR8e9442kdDbM6bPdPOmatWqWWGZcm3SBGHgwKf5Vv50oEh/Lh+2eXPmCc9muxWSf+R2Wb5Ur54uc7Mqisu+yP4p7eMHCpuN1m39ug0yoRKmjdoa/OabNk2DM0fKHZgKzV2SEgokFEgokFAgocAvTQFfQPPwU+KYBSQrAd84+UJKJ/x8lImr0YIQt6buXWQ5o03SYjBn9AYbyxlPoaJwGFnFhAssNVhshDbCm2pvUae0TKA52sdtgMUI9cT/VQFdwgfcEz9SiSWU/uJzSfvZKLP78iuhf+Ar+Ou3wn8tG9ewszu3FW+bAwL/7qM1RfMYJoyt4jhmesJ/LneiN8ycJ8XznxLuFhTN/yCjmN8hFS5S0Hogd0oBRA7WFUphNkBZ89/LRPm/177q/X77nAvyJ7SfJX/2bB9oBKRhStZ4O34mdS5pH3qE9N/Qv4ieLURrqqRQ2hkLVko589rvIloTPAR/E0WpXZu2HqTGt/U6DxDshyNG+It+KJNP7kH/0W+wXDLmzvWANPXq1xNeSlHfK7700kv+Mh85Je5UaO2trlQhihOMxr4Ww4GGDRt6vaecfLIDxg4e/IZdfvnlvp+lHehP+YUC1EUBsmr1KhXnTC7Lqz6hIKp24IFMH0UKyuWgv2effbb179/f6qo/Hu0osLeH165fv74C3SxQtONd6mth++LLL+UBstyVKESBGincmkxZoWxzSxbciZi7qiCLJZwWxxx9jJ180kkG0O+BUgw1PKihOmCKUDTHr8+YMcPqydCBtEKKEKyN7hZma+/eve38nj3l8lZJ7kLbnVYL5Y61YcN6x77BXWmKFB5TFZ578uTJvv8eNXqMY8x6ZdEHoMSPP/aYlSlbxunpMsd7mssIb46Sa4uUT2H+yxVNe/WXFTK8mOqH/jtlzfQ3gTlfJIujk2VFNGBAf0tPr64adgnMeK6N+OBD6QWEN+P3rjGQBsgjSWv95Gsb6C1ZgnXRA/ffb5dfdrnj/Xz91USFS68V1jmimefNtp7Jfg/7erz7ymBfaxNRGGUGdJFMqyBYnbp1fBIslwtTRQ0Y2i5cjwAGBvQXDJqNUrAQimrhvIUePx5QIyxdFi9Z5GZbKFd2yrds286gqYIW+RXbfL3wambMnOEDiClWCQ0M1jXLFN5rk3Bi2knjeXCjg739PCAYqXvTZEY2ceIkWcRsEMDSSBv67rseKnuDQmVvkKIGX7jBbwxW6LJ5+0qNpHxCgYQCCQUSCiQU+PkU4Fnqdr7RZkhvj1iIsIaKFTC+CGG54Ys5XeDxq0Qut5zRG+xyZYNypq8DAitak/L6Y1r5/I1pKOB1eHma1R918scvXxhps6ZD7wOLFvpBRf6n7/g3p0m+aAlH+lSNnsEvUCxc4ZUYSdc8P+e9maT9hP6/Hf5rfnANO/Okw3zePT5otA36PQMCR3OPaccbaE97zD8ma5i/IXP2+c9cjRWwMY+GvKpJExvLmR5dD7UycmtasnSN4daEciae/14m6sP32qczXNut/d3lD9d+rH2qIMV9imVY0n4WXWL5+9/Sv3jRwnJrOjYCBFYo7Rtf+c0rZ8RFToBVK1fZiI8+0t5yifhf0XngNf2tlXFA2oFpzsPhKSs4jcnfWPcePaxfv+fcUuTJZ55yVx/2qjK28D0pZVeuWukRk4Jbnkprg09Un3POOscjKq2WC9H8eRkCqL3TbrvtNkUxetLS0tM9TDQbcMZh4PPPC/d0Rqp9wk5nbtzgOC48IPMoClDb9u0c1uP222+zKy67wueK31X0bK1Qvrx7pxB+GowYXJ8qCv9mvpQ1+WXYkKm9MRYhk4XXSjkMEoiAjAGEp0Aih/HAgui8nj3cgAFDCO63dr06cpkimvFmqwVQcbQm2CiPk7vvuttO69pV1jY17fobrpcio4U8XqSc2brZaQMmDpg9YOCUKFHSvp6IJ8x6Vzx99fVXUp4M8PuhH7fdepuDGz/Xv5+UY+3c9QjjDKyIv1ZeAIoz5f7FOEF/6A14b1FFg4K3+Y3b0osvvGgnnXKSy4ur/nKlK5C48ZWrVvvtxuOPhQzKOeCZqYB6sZ7pfdmlbgFEaPUrLr/C8XbGjh1jPXp0d/iToDDyqnL8I2cVM2LyEgLSbdiooQ8GQrSEcFwIof2mUI7fVkhq0KcJWYU50JYtWzwk13xZqODStGlLsIDBFIsJtGzpCi0sy8hFaZaNEThTreo1JfgB7xH9NEj8QUg0ZUcdfbR1EIjRoXJNyiUlzAoxZonSJTXw1dx8Db81wJPat2/vJkl5tLjE727C55878jQKmWnTp/tDoZT87LYKgyZJCQUSCiQUSCiQUOCXpgCLA0eXixYe/lPPOj3xwoKIb/1nTcabYo7jhbYDB+vZ6yFtwX6QW9NiuTWFTRKAwKFg6u2xCsfHAO+TaJ8/b1HnvIzyeTssyKJEnnijxELHk6qgvtBb6sr6TY6kfYgQaOL0ElES+ge++63yX8smNTyUtiab9R0oQOAhn/nQ/V4/4LeY55jU8XGs9GVeO5PqIufi+c/9gtkQp73N/4KF87lypmyZoJzBnTJzk0A4aUiJusL/MOY53X48t+L2QpvZ5FHS/j7TPwYExq1p3qJVdvFNr/6mQ2kX0qY9TryQ3yAMFLfKEyOiQOAYDws4FH7HYuX4404UxspSe/Chf3rwmjP/dIZe5o+2BvUVoEZAtfAXHAy+S8OGBwtXaYO2o9JgaFfNRn/sx+PsIOUlgflypwBzu3fvbq1atbSXXnzRalSv7sohGv1OFiSrBZTLM5T2wTjNkzufrHcU1llaEfarVSpWFqTHBYLyKGwvv/qyvEzGO0Yq9fvM0j0AD3LNtdcIw/VguTatcZyVhXKzws0JBQz7ahRJwIHkF8DwXFkEjRs7zqM7+c2oLtpH6XBA1WrWRO5ZQH9gOdPhqGPcqKFmjdpW9cCqyuj/rGrVqlZOSqEPhetyodyQ1qxb4yG/zzjrTCso3BxSu8MPt0PbHOpQIsd0OMYBkRsrMM/8+fMdn6e8yqNMIaGo+cff/yFrpfl23733yPLlQocZ4SYJVf3B+yNkMaN1DHJJ+aHVNukT8suNCoXWrtwoiAs5zu3UKVOtU6dOBr7sy6+8bDfffLNwg2SJo+QyTzTD46ZmzZqcCOOhSktKeTR71lyNdT15/uRyy6luZ5+jCF7P2kMPPmKPPfGYEVma9vdHylnFjHqIkqR1y9aWXi3NNYoLpPCAAQoK9wU3oXFjxwspepGHt2Ywi5co4bg0W2Q1s04APFMFrLNVplYldB4mQxFDSLBR0lR1O7eb1a1bVwzwoa0Wg5FgWB7mefXNZDhQpl+Ykr3x2uv22KOPueka2DHkwXfuun9cb6cLzbpKlSqGX15+IWZjxrRJJl4LFsz3Af3T6ae7ptEbSD4SCiQUSCiQUCChwC9IARZa/OfB729/tWgg+WIiWpD4b31wnauxgiSs0kJeAIHdral0cC/o+1wIpR3noY6QfGnndflbbWUI7WuRhsIlW/vam3oKJbIUQiiIsl/gLVecQh9ZSPkdRSvJ+CrfSftOBdE6oT+899vjPzBnzlG0Jl78PaZoTQPf/B0qZ7Q2dhUFc3q3tPv8y7q0l/kfRFEqS2x5l13+oBQ+r2sbt9jDralPv/dkASDLmWj+f6/5PeZ/LA+QgN+TPz+hfTr3Y/InaT81fNHBnuMf88cP0z9WzlSuUNLmL1xpl97y+m/WcqZWrZp2wgkn6F4jDkTO+p3LlWjRIuGfVnJDgXBKLxu0l8Ta5ITjT3T+y9DesJkC2gC7MfyD4dZdYZnBQ9FFd4vJ0N72G4XCJmHdgrHrimUrBOibzzkefubZ/dZb/1adx7mxwGl/+pNdeMGFoUl9Mo9iXkdJU0ugvniLkIgANUZ7YNyjrrr6L7JIKyPrmfX2XL9+wp2ZHLXhWf0Oz+95vvUSNsqEzyd4yGoe37jvLFq42JU9i3TP3ONa4a5iXTN+3FjLyMjwCsqUKeu04UUPHi6kWvJ8IWIyECBjx4+1zZs2e1+ZilWqVLbLhDsDPVbJteneu++V+9cAV67cePONghGppFy7fG/e65JLbYr6W6dOHY/EBDgw/cD1aerUKd5HDDMy5s2zu++5WwDML3kUqONOPMH35vRlK0oZ0Rd6olDDsqiMwnY77XQ9967cVkOuS8cd28n1EVu1v8fVqVevXtbzvPNFs2cNpZBHdBO1vpSxSCnhx7hSB8GgOvNLcTVV+DwVFcDI88mqZrHgWW699VY7u9vZdo6UNM2aNaM7+yVlrZxyqHpcjgpKW1W8lG5UAztdYa9LlS5tx3Q41rVSaLrkQewWNWjajjjiCEewbt3mMAc0AokZJqosxUl6epr7dKWlpVkR1VlYWk8scoi2xCQgbFljhbPCfxDXJPzfmXb6sk0yoQKz5oMPPnDwn1atDzW0crxNrFzlAJ0rb4eobBtp8Q6seoBr27DuKVSooKWrPdCjk5RQIKFAQoGEAgkFfmkKsGzM5VahLCXDf/rgiw+9zYoXIf6wZIWphybnyOtJB3Ee3JrOjTBnFimk7aOunAnhOuPsWSXDAtLbz6qKdRX/qFTP2VC3t8XDOmqfa5zzfPrm7R/H9CP+5mrcVpxPWVLn/Ih7yXbG21Fmz5+0n9D/V+S/5o1wawqAwLg1Pf87c2sKlua7K1CZf1kSBlHC/BORsyU2QiSXKcxFzUNPLneiMvG5cEUvYwVELrmDWxNYV1jsrd8I1pXP6GhCR5njc8z7aP7HV/jO3j7N7Nl+LP+yl4nzZP8Okihp3+kUDWGgWUSTn0n/okUKCXNGobQdc2aFh9JetSYz+zD86seMf+bGTVKKvGV//dtftR8NFhN0TKwmQNrZ1kR7wSWLl4ivgnUK15Yq/DSBaijPjPlSmC2nn/4nWbu8ZH379lW0njO0zy1FVnv3vffcLYiHFGQV+IZbqKCs9N96ZpMAAl60eKmdJAyXz+Wt0efRR+zhhx62SgKwZe9K+4wECqDDtDfFbSpuH+BiIhSB29JWmDSlhGmzVtffe2+oDRs+3MvShk9DVdLh2A522mmnueJji4wdUDBkzM+QRcokhXwe714lRH3Ce2W9oDw+n/C5uyptk2KEvXlB4Ql551VnIe2177jjDqtXr65He1oj9yzSThGweLES9r722eXLlfN72CilzT333Wt/+9vf7Ij2R9g999yragJ/tWzdUt4tx2jv3VSAu9Ol6Nlqy6SceUG4PcCMELIapU+HDh30osnc8ubvf/+7cGsqyFpplLVte5i7NtE2coL258rj5kTR00NjQ2edK12ujFsUoQDiN6efeeYZhR/vIuiT8h7xuboslnTJrY6el9sTljcoZ+gpSqwdO7bJcooIToTX1ujofzFZDxHuvEXz5p6HfuyPlOOKmfWZ6xQia7FVltlVxfIVdZMAERWxQ1u3skbCewHUF/Ms8GcITY1G8OCDG9pJJ55o5/Y419q1PVzhraso9GcRKWKKSGtZSYPR1toc2taeFmFnzpwpM6h8HvYMDV2LQ9Ba7bKP5Dc4XXgza9atlkZNwl8ASIAgwej55F9XU9YyWMwMffc9xWSfrj7kd+UQiM3lK5RzDeTxJxxvR4tpqqVVk6lXof1B76TOhAIJBRIKJBRIKPCjFGCRFi+KWNK5P7NWXCwO3N2DhQJPV87pKc43Bfgk+c9wqDK5FWFBobS76OUEUVNcOTNUEQylnIkLUNIXH2FJ4O1H5UN9uh5lpn364ee14vFjXYtbD1dUnTL4f6+X6sMVFffEV9J+oAVjl9Afvvrt81+rprXszFMC5gyWM78ntyafj7GU0Hz8/vyLJqfP3Jg3IyVrlNldCHwuK6/mP2+tkT/x/M8qFaI1YTlTQVhXS5ettUf74dYkrCuSisfte1nVE49/kBwhG5+u5I0yZ5c/cftUtLf2KRtcVrLkj3edC0n7OUL/4lLOXCpA4IqynJm7SNGa5Nb0W1LO8Nwh6u61115rf7nySnv66afkotLAeRZQXCA13hz8pkcigsXgb9yIt22VwkJ4qPrlv/nCLej66/5h53Q7x1q3bm3DtJ888MADfJMOVuquXIRjzm3bNm+1sqVleRIpWuA5+tFGBgJHtG9PVW40cL4i/3Q6rqMNGjRIAMBtXOkO5gnugs8+28/myNWI9vXU96hPRISqJreiJUuX2AHyDikh3JatW7fYF1LyPCvrmTi0tD9OdC+47QAYXF6GBtwnLkmVZQn0zaRv3Y2oRKmSUgKttYlfT7TtwnDFguWmm28ygG4raP9eTIYQJNqfIguSYcOGCX5kmdMKRU8u9TOP3LrmycKlbPlyqKO0385jRbV/HzZ0mBRQJzpei9OQetQn3IyqHZgmJU89WcrssvTqOq5bzyZ+M1GKISnH9JdbVkfNm7WwusozZswYG/LmW0YY8JdffllgvL2DJZLoSfvoG74VVs4GGXVAf+RDvtx51c+Vbs1D2z6kGlzodET7w23UyNGu/Ln44ktc2bJ50yaPBB2P/+pVa4TTJ8waCnNSBxicdBWODhGxvp440aM+Q+f9kWSLlUOJDuomYOZMRVVKT0+39kcdaYsVBhtgntmzMlzThSYqtxQq+fLnsREfjpBJU22Fu6rimskihYv6IOURKjY+cOTFValQ4YJWpkwpMd4z7np0sIB+D2rQ0FGpMzM3uOvTrFmzXeMWFqjSy8gXTWouTbKgsdyFCacGckPmegceLqTQb7gzgWxNyG3Ccx8gy5lmzQ6xAwR2hPtTkvadAi1PfUCTUVMNDofJnZMZi8Ds/Pax8c0FWZiWup5qWvn4EWbIHtd03q/FtYZvino9SEOOk/YT+if8F+aKZhAHyfz7vvyBNH+96GjrcmxWeEwXIL/aR5CDPmK+slBH9IOQsKQgKcM5fofcknn6Fz8HeeZxjHuvuzUplHb/Nz5WSNvVci8Yapef10nnC/izURm9Zm8vkqvUS/I3f7pO5SksGX4FEevlUlKbk1q07dI37VMv8tv7yyXvfpDy8eU4H5e8Fer1fPpWStrXOiahvzjht8F/uDWRXnprnNyaRjuLn9u5hZ/7bX9gC5M1uZh/qQkXzf+99T/7/ItlixdU+XiaxvPfZQ6TnLrVQMFCstjrfKjkjmAMJHcecbnTUWCdBYP8IGskCOK69uzD3tun/h9qXxY/1Kn/3l/kj+6Pt/Cha7SEfNJX0v4+05+x7K1Q2o/1f9/myq3pEgECP3H76Va6ZJE9h/JX+b1j+y7r06dPZKXyqCxM3rMvvvgqDL4+pwhCwydxmAzikR3+zMFqYiczBj5RatSoiUdRGixFzj333OOWL2PHjLOrrrnaLSmkh4GrHFe1jqxL5giolucl7cPDmwWcS0SjEf/5wPnxxRdfEMbKF/aoLHCGvDnERo8eZR/J0wOmBH+FeVGxUgVX9mClMWPGbLnXVHCFEWGvy8vzY84chZTWfmrxosU2ZMhbVr9BA6svRUdY8oaOX3JJbyk4RrnbUzPVQ7jquVL6FJayoaoiFgN+vHzpcqutqEOEBAd35uqrr3aLm0KsDwS/ioJzzZp1NlTKGfbNu3rwUkauRdrzc48F8ocQ14D8HnfccTZ8hCBH5JJ14UUX2btSYDkNld9nnvqLwmiiFBzc94HacwN5MkVRs1DOop8ngnN6err6OVsuUivthhtvlDvSpR7CmnsAuoQ6ofn777/v9EXRo64ESyPVQfhu2tsl9yZo2UCRq2bOni1Q527uKnapIkXXlVXNHXfd6Tg/KNIov0UeN7gqgVWLbgBlGwYm9K21wpUPHjzYwZM11Pslqes5lbh9c8AhTMNwYyJK0sECRVoiBOyLLr7Axo4dY8ccc4yVLVNGRMwrUN/ZQr4G9GiVFDB5pVlTiKp1Cq0lJtkuX7YSCmWN9rBk8VKKPV9GCh/MunZ6vPU0adnAoXn7nXfcZWmpNIhF5OqEfxqWOBnzA0oz5kmz1Z+Vq5Y76DBEh/K5pJHDf66M6s0rDd+qlasVnSlDrlczHBjKbyb52GcKoBRBVCE42AyGh7fOITWYMnzpfDgMeVnC+2l9s7jftZNfIYUjlvj8kaiDo6iMZ4jqYZyT9kUbESWhf8J/yfz7UfmDDLn3qQ9s8PCvESy/akKMIQfFtKEfLsfooc4h16LzQfLxqYSs5C/KS0nce72IPvhdWObJHtK2dPFgOaOQtvEbbMpRk7fozegjbifVPvXRfpSPNrwUWb3HuszqNJL0KGi83tA+Wak/XA2NRU1E+cJ18vl9RBfje3IXi6T9hP6/Af5rKeXMWSe30Y7F7LFBo34n0Zrimeez3PkoTPhovmlKphJzfo/5x2zXxIxPe9bU/I8qYu0f1+kKWuUvJOVMDymFASL3UNqRO6XnC11Jzf8fbz9L/sT59t5+qBT5F2RH6LPfkheMGuUrad8pEuR0TFV973X8f5j+xQSee2mPDla5fEnLECBwr98IIDD3lUuAsDuEYzJ69Gh5ZbRVdJ9JCqV8ssB0z5fnhJRHYuuQFAFJCpSdcl8pJWgLrGe0L3fFA9fT5E0Bvgg8M2vWLDv11FNtqJQ8z8qb4065+gAgzHNwh+bAGClZiHxEUuva8wpoWMYGuA8VlDUGibxgmbAvfvW1V6xzly4e+po9MmwJb9M+CgHgN7BuAau1obxMsObJJ2VGtbQDfazo9wbtkV+RVckT/3rCXXu8EX3kzZfbPVOOOOJI4cwsskaNm1gleaPgdbJ95zarnl7D6QMu0/FSqixT1ORrrrra7hH4bp5cea1J06auEOLZTp83yXLHrdhU98pVK+RxUtr7QH/BvpmjsNR1pOTpfm53V1KFPVroTZhu4bNRo0Z2Sa+Lfc/eoGF9a968hd93elqaa1BHSklVppRgRdQm+DV33HG7/UnYPB07dXRolIMaNnQgZmr2ua726d/ceRnu/sT+n8T4F5Bb2uHt2/v4b1dI7a+//tr+/Oc/W+06tdxi6li5fu0QFAsKKHq3TWHQNzF+7D91jkhNRNiqILcqUopl/FfOfuScWYg6TsI/bL7MmjBzKqXQWLgRYYa9RH51oFf3FCgRCUuWZvLrWykQ361SwuSVYgYNGBoqfOgKCuvlALk01a5VR6ZDlVwJQ5grfrdt2861e+DAEL5rinzT0Lad1/M8W608H48f72ZPaMZJ3yk8GJOIvmyUryGTJfXWj+eM8uBnhw/fv//9b/fZu07mahVlspSkfaOAa219HHJZJQnsS7t3tGJFAlL3njUzoRgXvv2h4CMjwRg9IAKHqVTqelRDtt8IhnjpQbOpMlFWF9Lk30tK2k/on/DfH3f+jflsqr327sd239MjXDr8mpYz2k647PLFciz/JLfCIjO2nojkG6+XeI6RTwnpxpH/UpnwO3zC31jOgDkzcPDHtlghbR+VcubyHp30RkhRDaK2kL8uJvXt9Ub1eG1uvRPa8va8/Sif9zvIV3K4LCdTlLz/XnF8Ir4SrGKS9kU16JPQ/3fBf245o/F6cchY6ztopObPTlmHtMxi6t/4EdM9TimrlIj/Yj7cc866FXq2gmzmyMMfKbX+claOfwW3pm6dW9nzb37imDNEa7pMFntFZb1O+n77Okm9+p8lR1Qfm+UfbB+5GcmfqH2vPPrYvS7qz7qatB89V1LjD31+Hv0Zy97nyXJmwHCbs1DKmVtesydu/dOvajkD78Q8QXAa9qYA7xLp9/rrrreDBbFx1x132YyZ0yO+CutgQj0DshvzNZxSsEBBW7FyhTBPgyXQPCkArrzyKu05P7HnnnvOg9c8+cS//CH8+edfuJuMc6P4DKwS+oJLUT5AdSMeLqDAOEXlLXKpAHQryXukY8dOCstdRmGjb/GowRE7S9nypHXscKy8StIVJSjDdkrpULx4MfWqhO9710ppgyIij+bjoBcGCYd1st17790epjqPXLZI7GkxRsglgwfyTZ823fFUt8j1CrDjHdo0lxb4b/fzzhOW6zB7/LHH3aPkvvvuswEDn7d/3v+ArH1GSBctCxSvcZcMH+bYZb1725C3hjhoLucPkMvUKkU7wrIGpc26tet8D+/zTRmYdjFdypevYH/761/dmwVvGZQjH3/6qXB02thoGXPg2RKnokWK2htvvOHWQaOk+HpB93m7wpA/P3CgG2WAcQseDNZErQ9t46DGtIOMQgmFjsHDf+scsuCzzz6zI4480l54/gW5aQ3XPbzlFkP0b+Cg5115Vbx4cR+3rfLCQZfBGDo9o/GL+5aT31rR5WTaJUYpbpWlUMH0B0UIA6P7cE0ffmpFixZ1AhWXuVOHY461WtJWzZo5S/HOt7hgZsQWLJjnvmAffvgfmUC942ZD48aNc4sYmLiW3J8mT5nsMc27dOnqEaAgMr5f+P21EUAQuDX169UVKnw590fDFGrWzODuhJ9gprSNxEIHjNi1apGw53iRzLywuknSvlMgt1vGmGMbLF6yxvr2JypI5GOssWbc4G8mQiwA/S0vTWsycZaJgBbT8/Kba1HinFeg35R3M7goL3y3Z6KuVEraT+if8F8y/yL507ZlXet6XCuJh51Sznxob74/MSUqfukD5BTWgiR/g8Pv+E1Dts4E8YfkQ4JKZFKG/7rAcaiBz5DD5ah+xcqZsgLmXCLlzCPaJG1wYE7l1MLVUyJ/nW7QMXn+JM/fH1p/tBAgsFvOaLY9NmjM7wcQONv6h/nuaym+o/nvssTXVJIjmgPhepAz5Cc5TVJySdeQOVFe9Lek6MuPC2vjjlIYrCswZ4jWtD4zrAe/337W+i+uUy16PfHH99sPLguc31v74Z6i0v/v/SftQ29o9nPoXwxA4HM7WFVhzsxbECxnfm3MGd0Cj0SPBoTFBZgin378qTw5LhKQawt7U0qFwxXS2RUbgsCAaYcOe1+hlFe7IifmNzw4lsiapISC20CbPAKH3ayX/f3797d27doK06Ws3XDzjdbx2GPdQmOD9pnOo1LY6p97aWDlglLEn92qAwiNbt3PdRp/9smndtNNN3iQm9ffeFMRhP/k7UN/DB76D+jvIMQrV6+UEqeiFZGiopqwZmiXvS75Oh7fyWpUry4MlK/tgvMvkOHCFN9zF5MHCn1BQcG+DAycBdrrAuWBh0l9ufngRTJ82HtWWm5E11zzV2vVspV9+OGHjvMySW5HfR7tY70vu8zxaNwyVoRat2atwITlHi3DC+ovLAukb76ZJBpvVLTjTW548Y48W7JP3cjJQWucMJ/ZJqIbKCTFF54zxXSMixTPXYL8kLi3lgo13kp7/AlfTJCr1zx7U25Qj/d9zB555BF3yyLfLk18FDTPDxxghMsmOf+K/kSV4n8+KYBQLjHO9LHLqV3tlltutk5Sil0rt7QaNWtQyu/JsXRUZ6aAlgsXKRyMB+ARbih0nyZyNCF5cjDlEmBQBZllHS0LmHweOx0Qo9wyASLllstQNQElYU42+M3BcmWaafny5JNmbZWIs9GqVq5qp5xyih3X6XjFEVcYbQ3sZPmcgYI8UNq6GtVrOFPNnjXLNm7IlB9ZprVo0dyat2wmpOUysp5Z6jHJQU4uWbKUfMhqW1NhxhwkJU3t2nWsoATGNili0Fhu3bpV7a6QydYyD8cNKDEL3zxiWibQfqM4Vf+BEqRk8nXveqiVK5cFPIlyhnU/byk030ISk3PIJCIxyVG0ZB2z0dCk84VDmBEO1OWFKKADleGtFeV06PMm3rSEEuGTOpP2E/on/JfMv+zy5/CW9axrJylnJDzuefL9X085g+zjf9w5xJbEG6IuyD8kGDIsnPRs5FGZYG4cv8VGBqoeLxmOvZx+Awh83qltrLwDc8pyRtgPG5HLifyFRKJj8vxJnr8/bf3RooncmgQIzEbisUEjZRUywXnot/Yh9UrUJUmEPddf8ZW9zn/JGj8fr79CZtwz2N8gXhxHMJI/bkUT1Yf1X1bCrSm/ojVpPVi2qC1dvtb6DkApHL2so6q9th/Lwrj9UKe3j1D8gfbJtWf7cV9+3v0n7e8+/j9O/+LawPbqcawAgUsJEFiYMze99qsBAof1vxhE/w48oKrvIYnwu2PXdoHAjpJ7y+G2XC/iP/popJ2s/WfYR5iehZlyUxrmc9oLw0vae+BdUUDBa9jbwqoFxM9Y3YwZM85uuuFGuRrNthdfeMn+/o9/WFFZtETbEikMFO1HUBvlFL2IPcou+SjxXCbYDe5NKFag6rSp0629lESjR4+0l195xa684kpXAJF32/Zt9pIiGI344EO5Re3QHni99sjfar9bQiDE1TQVc8mCZZZciM6zI486UvNrmStTvvvuO7fSqaBAN24ZppYA6kUp1LhxI4cO2S6LkFKlS8n6Zo3vx7HGuezyy9wgArcpXLgefPBBKS6ucWVIXik3WFGAwfOh8GRWqRx6loIyniAM9zbRKW/+sL+vJbcm6DBu3Jgw/XTs8ArqLyn1qYOiRYp5xCnoD/hwfA0rlQbCzzmifXtvZ5v6e7OAis87v4f1PL+nPF2Gid8qRArZXB4oaN6C+d4uvLttJ5i1cjOT5ZJH2wIoWEKASFnlpNi6U5Y30LZ+/Qb26suvWHvxBX2mfYxKtmze4m5MjL8GUOOX6rgOcjYhUnM0oVEsVUoYMlLG4Ke2cOECS5NG75CmzRw8Z4u0YPi4wXADnh/kJkc10mv64OcWIPBBDQ8SkNJV9pw0kL1lHlW3Tl1NoF22bPlSN7kCRfuTjz+xZSuWuevRxEnf2tbN0oBJwUOorXnSotWpXdcaCSAYDR5YNaUVVgwXqDQBDK2TlQzRokCQzl+wkCtoNmZucgsaULWry0yscdPGwdQsRynzx6yMhYDc+4RwzcNYb0rKlnAfYw+dqDclLH5JCJ1wlEUnJgSCkDxxPs/JE9WTykjz44o0zaDoea4rlAxCk2wIK5JPMJWlSerT3ErVm7Sf0D/hP58mqY8/6vxrh3JGljPcP8qZN34VzJkgs7LklJTNPjJYbwT5x2KDxaELTpd/8RzmO95ISO4pX8zbvtHWj7AwZBEVgDl5g70IQGC3nAmbpD/q+McTILn/5Pn7c9YfjjlzijBnNFPBnPktRmvSaky9kzzQoilr/cOZIF1i3o+/kS9h/eVvK/3egvyRfNGfvyPTJcDF4/UXV7LLn3j9RZ2x/ClYMK9169zayhElbomUM9nkTtw233ujf0r+xe1Lnu29fUYia/1HfXH7WXI1SMYfvv/vrz+T9qPnz/9DfyALegtzpkqF0pYBIPDNr9rqtQKq/YUTY8vGmlRA2C6TBW1BRCP4lLRi2UrhqhxvDzz4gP1DypRjOx1raWlpfo0HJ/tPwIDJDRQG4LbwN7wpnyDtJyvZ8TImwCBgoaA1Bgx43m699Va75eab7VXtc8vI+gRQWsq//4HcgPT83izjgPiJXiBfQZsxfYbVbVBPOTBk0B5FR926dbOrrrrKel5wvtWoVV1YKLV9r4O1x2KF8iYS0tyM+XKrKqo99WK5KJW2mrVryeNjifcFBUaP7t21t10vkN1JghE5z9Zv2OjuTHlltUOPSksRM2P6LFswf4HNnDPLlgoPNr/2alzH2gdQ5IoK5e1KFN3ALbfcamedeZbVrV/X2rRpY+0OPyJaf+CqpRDTkisbtMeuWqWqh89eIXiQ/HnzOpQJCqMlAhgmLDb3F9M/+qGvMBfz5s9r7w0d6lGPDpU70kZwXqTxQb6wd8MKpoAAiZnDuQSBghLs6COPlrVPPkWUutmat2jp0CZqQvQSlcknmhNOm/GH/vymfdZGaQemi049bLvGePLkb+3qa6+WXmKFR3rudk63VPv0eOzYccLAXa/7FHaN+kJ/9kcK3JpjNe9ylyXMvyAaVimZMhkj6tLJp5zkWDFDRfDlUqoAVLRdsdXBe6lTt7ajMi9bttxDiOFilJae5iA97cVcJxx/nB12GHHbS7imK1OatEwpU5ZK8fPWW4NtscJlic7y75uic0vc9Ak/si+Edv3cgAFyhXrTtZ9o3LCWKV+xnB3erp01FQBSdZl8FSteVMxYQP51O10jxiSrIOTrJO07BRy8F5EkfkA5g49xWTdjDW9oMZ8PvM1nFpfHk5SJ4/95M+M54nyc50EbzsZlKQdTe01RdV4imkEBQ0TlsNShqFLI5rnCCT8XzibtJ/RP+E888Aebf7Fbk+7c7n9agMC/sFtTMPGldW0skHOSVyTkWywb+e3yT6Iq6zyLkCDZeAaT/DvbMZd5axYngDlxLyhXGotGuZtmc2tK5F8i/xL599PlX/PGNeyczm19YhKtaeCbn8XT7DfxjSzgHbdbiqZ6FEmPSEZwOpYxyIrs4x9LGp0N//SVJXtCyXhllZI/1BfVHTCkgvzxKHGSO2VlObNYcoeXdbHlTFb7u8+/uLUfb59+0Gb07cfhR/b21XVPXPF7iAtEv7m4p/xL2g+j/VPpT7SmXgqlXUluTRkLVtqfb3jl17GcwZRD/1CmbNyU6RYZnGJ82f/honKd8GZwZ6meVt1eefUVa6PoO56kbCAfadq0aW48sG1bAJXNrbfOxWXVMWXqd+69QSNbtcd8pM8jdtRRR8mbo6Vb4hQtVsTyyu0J44Hn+veT9YWUDWhGlbYr8hOWMGWE7eLzRPoeLDpq1axpDz30kP35ogt9T/v0U086SHDwCMjj++vly5eo/HbHZGX/SzShdWvX+rUZM2fKyKCpFS5cRPXutA/e/8COFUbNV1996ZYlWP2goKD9Vq1bSSmzTLM6l9WVYcPMWTO9zQXz58t7RUF4fL2wy0oqvPbLr7xsZ555tui4URgvA+2KK67w+wgfuxzgGKXRAoW3Xqw9OsrdMmVL27fffOsAvvMXzHMazVe0Zk+Q1skbaAx+D/SY/O23btyxSpg+buXjyuSdMq5YJ8iUkj6W0B/coDFjx9hZZ50pN68vrGePHnbnnXd6WG7dYFS5+Fb9AMakoJQ6KJHi8c8tt6QvVC6/okrBH6tWrLITTjrBHnjgAXv88ccd66c04MZK48ePM7BtsA5kJkRs4ddy8iNrdZYjtWJqtcNdmMR1HuLqG/mlrRSaMqZjmDhxY/OlnatSqbLjwWyUC9Oc2XOc8LgnDRawz+y5cxxkqagIXlouSWVKl7OaNWpa1QMOdECgY+W/B4CR48EsWCRl0Don1HCB97wvEy8iK63fIFMqmY0tldXM6DFjFPlJvoIyYVq8ZJGwb/JbcdWNX15NMX/Dgxp6TPWqUiCpIgdGwtctSftOAZjfk+YHUyTGNgCdn9CJjwpzJjPCNojfRJAfAZFKcL/qoarsE8E11n4uK69rQpUpPuPffEQFOfQuZXsIhxNsgLKmQ1YNUdmk/YT+WWyko/CQ5zt+aIdjaeIT/vufmH9Yzpx6XEt/ttwny5lfMloTb4BTSfwUNlS781qMOYNoQ175324yLNTg/Ol5wm/kH/WFEuEccvm8rnJrwnJmmeSyoqZkxu4FNJDIv0T+iVVghTglz99gRbKn/G+uaE1ny3IGN5vHB43+TVnOBPEQ5ElqKH1BtLuyFtlA4tPvL8oTnfZrext/3pBmF12eMa6D7+hEkD8BELhHZDkD5kxsORO3T3Y//lntx61EjVFHxLjxlbh9r/9H7t+v0+uk/UDMmID69cPjny2T8oVQ2sda5Yqlbd5i3Jpe/cWVM/H4sy9k/5dPVhwkzqenpSsMcy/bIdeY94cPt1dee1VGA9ttrHBNT+t6qjb0YfxhgWnTpjrAL/tJh77Ilcf3rrjxEIGYp6oq9ZcpH330H2svAwAUJDVq1FAbl1hRWdWsESYL+0x4kLRKAXDAYgV3hvMULi48mEvUJyITfzz+U7fy+eLLr6y/lDoXXXSRl6V91glffvmlY6Ji2DBnToYVl4fIp5994vKHHU1ZuUhxnwDyohTp1u1ce+rJJ73fDRXVKL/ogeIDjFZls9ZtDpVSqBYTz6ZMlgtUXoXL9vvKbc0FDdJWRhKEuR4zWnhaAwfZww8/bM8886xevBM8QA2qkqWy6CH09TJZ4GyPQI/5jevVihUrbV5Ghj0mbJhPBfIbzy015+mSS3vJuyXNy42Wq9nqVWt9b7ZduDH0D7qBhRPf/4GyfrrssssdEPnJp562QS++aNfI6uXf/37L7ym8jA+1v/3Ov53+WM+QoAt/KJJ8zx/Rn/H/+9//bhddcKEd0f4ou1oRqsC2IRHEiHGK119+Moc/snaiOVTxLg0ClbJB3qyF3VcKSUVYqjlz5grQqKRr4yBsIWnxmjRt4u5I/3z4QWcsQIl41GVKQYMJUkmZWbUQ2E/LVi3ckgVk56IyIzv4oIM8TntegQvPnDXbY5Wj9FklS53vpGWbLvOrJUukqZP5WLny5b0s/nDwDJgyTIKy5cp7aG/MtfJLg0bMdKxscGeaJLMvTNaStO8UQKvoZmhi/jBnc7mP8bn4GGPGqjcljwrB3S1nYHYmBkJAibL85Bz//Wz0m+vhLXKo1/OHzGHScCJbCjXClxzRRqifb0f597aS9hP6J/yXzL8s+XNYC9yaiLKSy+576kMpZ34pQGDJImm1Y/kXiawgujgZyS3kWZY0C/IrzOFYynlW9T5L/gYO5zd/QSbyjVvTOV1aWYUyJTyU9iO8wc6URaPaS+Rv8vxJnr+aJJoo8fz6sfUHypkzFUqb+eluTb8VyxkmepTkpONHKWvIaJ5nl/9stOL575kpEsmfYHUTr6wiuoQCUQvhy5dc0RmXV1H5uP3CRQq4m3s5YV3FSmEsZ1w+0T5/URnvcrb2475Svef5L9r/f9d/Sfv7TH+UM5fKcqZKhTLBrckxZzZEXPFLfAU+nfLdFGGK1BIgrXBCdAp+LFZSFi9Tpsitp7S74uCq0qHDMfbk0095ZJ6bb7nJ94xxL98cPNg2Zcoly6vMJbckWb+IRwoVUaQm7VvZ/xJdqL3cfL7VfvTCCy/U/na53X77HdbvmX5uVMDTBL4mzZg90+rWreveI7gC8YKYKMPvvPeu1T+ogXLJ7VD1nnHGGXbjjTfZFZdfbn0efsSKCGQXa5TcijSUmZnp5YGNaKoQ1CVLlLJ6deoLpxXw3a1SvOSxVq1aKapTRzdkIJLRaFl+lBCW65HHHOWGE8wf2n/j9TcEOVHWqgsAd/Va4cZk0xKwh2/evLnuk/DjO+y6f1ynaMg9nV7vvvee2qzjrkMU2iblxjxZ3OAVs3z5Cu3DK0ohM8+tjr786mvt+dbbDBlRrFwjjFn1kwRF6ggP9pGHHrZGTRrbNlm2YNHDOJGWi44lFGCIiFSxfABPtrDClBcRYDD6h7FSGB199DHCu61ow99/344//njH6KE8ugWAiFHyxOO/U+ssAIHRP+C2Fr+gr5aeZi/Lcqrn+ee5EgaLo9NPPz0MO5WFLnGU4ykbyfe1biQo7kuYZQUzIRaW9J3/69avFUBRScePaSrzKsJh41oEQWZOneFaw0Ly/6ogRcrwYcOceHNnz3GBUFFuRShNJk782kaOGmUzZ892zWOevHk04MvkGlPGOgu06eqrr1KIrEMdBRowYQYOLWDTJk08rBjtp6WlyRqngLSm+TyEV4Y0dyBU05eVMmHCqmalTKdg9CTtOwXYEMABsRBi8gOQh/m8A8CVwsd4lYAneUO7dTfFCWWZj27KzwEp9R0exA50Ga54WRd1ykru7G36ec6p/XiSU78/zDnnJWgvlKRKziXtJ/RP+I9p98edf+1bSTlzvJQzkhP3ya3pl8GckexBaCHLRPuY/EFM8TvIrjiCSizfXG6lMvMrkpOShvGxRxjSbx9RnXYZqIuqVf7qRE1pJcuZ4o4FhkXjxk1ZgMCpjqgw5RL5GyMFhHFyKjNm0D0bzePxSZ4/f5znL25NAAIzr/oOGvWbc2uiX7Bo9vm/t/UP4iSW/y53uCOKav7jVhEnz+N5w5kU/6MsiRJlHLBXv7O3DyBwj86HusXeErfYGxqtB2l7z/bDOepC/sQp1ce4rXj+/Wj71BHWf/Qn1BH6G34n7ecU/YNb0zHylgBzZoX1+qUAgTW+4WFqxn4vPS1NVhvLnfcZb9m8OG5Jjeo1nSfhTzw5rpAVBjinp512qrslNTvkEOcs2Anedpci5V0hpQNRjHbtRFmgpmR9UkAWHYfIuqRFixYO7Jsxb6517drVTul8iitEGjY8yNunwo2C7sBCZp7wYqiAPqH0zJAxA1GO6Tu/xah2xx132DnCnqlXv567H+Hx4Rgq3jMvrOf1JmstJcyBCrQzYsSIgLeiOuZLSVK5SmXr0qWLcu+y6cK16dChg2XMnqsoUh21V67se7P1chW69bbbbPjQYR6MB88TnxdqH+uXTxViWoRSDZo3uv8BChNO9CjwXO9VWO127Q53wwrycK5woUKWXj3Nqlat6uWAJRkzZrSweoo74PGETz6z+++7P9y7enaNwIWXCJKkz0OPeF8xuogT3i/pgh8BziSmf0G5fREuu5TcrLCOQcHyH933qRq3Lz7/3J7r18+OVkCiOLpT2Nv7IPp97dq+062GRHWnMfeKRdV5wp3BYGOW3LruvOtOBSAaoJDhA+2222+38tJTIDcYk/2RclAxo1HQDRH2es2aNbZIJkwQrqQ0bA5QCMvrOgkFCcSG2dBSVj2wsg8wKLGYQ30ghOfrb7jB7lfM9I8++kjARktcw4nCZ54YHNMtlDsXXXCRdT31VDdNK6doUF0VOvvGm27yelepD4e1PUxKG4URk3VMIblSlS5V2kOilZYiB2ZDC0kf8whsCQb7/MsJrlGrWKnSbhrS/UH4P0qdPuIx7+obHog3E4W1CejeFUDgYrZ4OSFb302Zz1PE8+rb+d+/dZYf8BFfzPzAUuFcVIZr+pfiN3/ARvWE9rkaUsyT/ELQxCkuz3fSfqBnShAl9HeCJPz3x5l/bWU50+X4Vi4l7ntKmDP7GRA4yKIg3HwDkhJMkYySUAr8F/LEcszlGwJLieNYPPoJnfbreu6FlG38dMIXWirBJqmbA7WXDNgPz4aoKcn8FwGT50941sJLcFfMYDwTlJwPIVN8zHfIGZGO5z9XQ/L80fH/4vO3RaScYa372PMAAv/GojXF46cxCPM/Ggx9xWMT6z58HYTcURmfBj7+UQXZxt/zqXz85tkLUC08o3xhP8AJpWztF9TLuh6ypAaD0LGu+mNJLcuZSFy5HPT2UUoHvkqVj9vXOh7u4j8KaE/x917bz+rA/3//ypG072P439KfUNqXdDvG3ZrmAgj8S7g1afxLFCe8dXC/euKJJxz/JLCv+FHRmbZo31qqTMnATsoP/xCKur+C0HTtcpqiC+W3j0aOtJNOPlFXuKpPAH01r1cqolLjJk193xu7zeDO/s233ygKlECGPXsux3hpKcwZXHmGyfigWfNmbqmC4uXZfs/atBnTVClMCu7MDlmk7AxBaHZqm45iRnWWVJju92SZghXIyuUr7QzhqhzSrLnugcvqt8qOV+jqV9943fLIo4RQ0ITJptKv5Aq1RPtyjBI6duzoe96ZcrM6psNR3v4hzZt6f4EQ2SQrl4dllYO3y/JlKwXvkcfvNZf6umFDsHTi/rm1fKLNxwrzfcaZZ9hIuR6ddfZZdosUO47j4hTN5W5egAWvEKguuK716tX1aEhYzHzy6XiPjLVWxhsk3MSuuvpq4fy8ahf9+c921ZXXWAnpDEhLBEXyvqxg1gvQGPca2ofmq+UeBigzx9AJ+k+fPt16X9rbbrvtdjvqyCPtFQExg0eTSqIVNFsmnJ5mzZpFgL7c0y5vD6DfQ6SMw5WNei+TpVLv3r3snHPOseekjCoq/qCd/ZFyTjEDhZS2bNnqwD74kkGjeg3q21pp4OZlzHXQHiiJcONBhTIELdYBB6Q52FHhQrKYkYVLISltlgvN2S1kNBkmTvpKwESLrbIUJmnVamjCyGRMHNi8ZXO79eZbRKDC9u+33hJ+zXh3YWLjv1EWLzDWnXffZe2PPEIayaKO5FypQiU3eQKYGICh7VL2oNQpX6G8K2O4jQqK2BRr1/ymko//mgKYq4bpqyrEw764j5lZv9kEnNslCtnqyhnelOgNrbKjOMvO9vCLV5LqTbQ50W/3N+RbvAXf+UM54rW4EprN3r4LMuqK+xOV91P6SNpP6J/wH7MhpD/6/Dscy5lOrSVOYremr2PS5Ph3UKkg37IS8iqWVTynUnKVYwk9l2cIP/2Lj+OFoteyh/yLK/d6o2a8Hh2DOdPtFLkQs0kCC0yAwBs345edvUeJ/A3jJJJrZQxlkudP8vzdc/3hobTl1kTE0selnPk1AYHBpEB2kFLzPtv6xzc24bKvf0LGcIK9Yfb5j/xJSYOI/728GnCrPBEi+3qLzLGc8priduNvnSwgd8rzACKX3Fm8dGXAnJHFHikoWlItuvyL5x87U67s3n5oL9XJn9D+j96/9zNp3wdDH2H8I3r8DPqjnHG3pkqRW9PN+z+UduXKlfTS/kbZROR2AFqCwJCYA1iEFtCe06E0xLO+zdC1E088wcNJf6dw1O++/bYNGTJEbj5D7Prrr/cX/oSJlp7CeXrEhx8KC+ZjHcuNiYmiA9xmMFRAWYI1DdGWvpYLz5lnninri4F2wgknWJ8+faQEKGVr162Vy88m51/ax5ME9xz4Wk5D2HJ4atK4mVvvgHPyljBU7rnrbgXGaWMnKKqUR4tS+0RBGqYgO3fIXalRk0YKwLPR76lIsSLC+ykqRcinHpmJCtmHL1REp79c9RfH2SkmC51bFIK6sjBgt6rvIz4cYYsWLxD+TTzOefx+8TBxjxi1B3YNFjNrVq+xu++5x61VunTuLKXWAHeHCi6Lu1wJdeftd3qkp/T06k7vNXKVKqQ+EU0Zty9wYzfJGgY58bRcye6U+1fzVofYcZ06ufURSpLJkyd7vpj+7OXpXR4pj1D60tcyMvi4+OKLpWDaZI882keRsvp7SPNRH42y9u3be14nqD6++VY4OgXyubEGAYCgP7ywfNlSB4mG/tSP69azz/R3Kx4snFCsMf77I8XjnQN100NprlattAUKkU1IbJCnCyskNdYpM2fMFBr0105UfPKCOSPa8x3OyAhYwoAB5EMc8dq1a1taWpoQs4u7j9pG+fRhhnX44W3tULkrvTf0XZvw6QQpVmT5okH8bspkG/j8QBHuaZ9kWMps0+SrU6u2AIuaefiuqorZTrz5cuXKevveY6njD5CJVeNGja1Rw0bePhpGGDZJ+04BxhXhl1oEqMr4OH5zUkgP4+4CgCsvbIOljs4/zN+UeD5mhP7z5VJPPObH/FZS1VFCgIVrnPMFgV/nR2gzjHe0aNDJlOIFaarEgz/0LVSatC86QIroK6G/c1DCfz5bwscfbf61a1nXMWeQNWDOvLlfozUx8WKZFGRaLNfiIYhlKXPTNz3izt3ln8A4dS2Vj+kcDVosf+M6OR+uMemlnCkstya3aCwhyxmF0u5H1JTNCf+g7r9vAABAAElEQVQ7dcJHREr9SJ4/yfP3h9cfLZvUcswZ3rA+/gLRmgR6+Ssk30z59A7rH+9CtP7Zc/67LAgiKMoWnn/Zux3kRcz90dpMZbLLFGRTnC9ei1GC9dfe2idqSndXzhQX5oyixMmdMsacoVz2FM+/MPt+qP0smfdT2vcFjxr53v1720n7OUH/onq2XNLtaFnOlFK0phVuObM/Q2mvlJXKrbfeZkPeGuL7Q+6BkYR/MubOtWaHNHNFCufBRyXhznT4Ee39eKsMDi5Q2OprrrnKLu19qT3z7DNWtXJVsXaIzkRUoO+mfKdno2pVnTsdqFbPU2cX+FJXdNxUFhiLFFL7xhtvsBdfeslDWI+S8UF1KSqwMAvP6tyOA9NaYLPbtT+GD/mjx0Wk1Gop3FUUBzR1/fXXqY6eMjrIL++RrlapUkWdljuPIkBNlLXLi4NeFFjtIq/7sMPa2pZtWzwoD/vvsMcNc5O2+z/X35VORFH+UIqmJoIAQamCQiJYv+2SoiNT+KxlFcFYbltYDOkGcVGqVLmiY95IA2UTPp9g99x9j7v7PPHY44oE1UF9DwrhA6sdKPyXo6xlyxYOirxJSiMC8YDfkjE3w+97gxRaWPQA8DtSODg9e/T0Fx/jxo21s2WNUwiFFTImah8jjoMaNvA+0Qp/hOumfBXdS15BnhCi+8QTT7JZs2eJ7i8qktTlun9ZAakegIBfevElty4KciuXR7VifYRFVNgnagSUt5rcw74WpMoRRx4h6JOV3g9nkBz+yDntA70WtWDmmTNnOUHnzJnjk6CcUKGLYUqmQSO0daZMoWB+QlsXL1bCXYx4G4tbESmPjkuXLSMFTX2FzG4nC5xttl4WMMVVB5qqKlUq2ZtvDrEn/vWEjZLPmajnIblLlSrjBC0oZVBaerpCYH1p307+xmOSA/ZTRANap24dO0RmZ40Obmz169VzHzcGEPem9OrpbmZVVQogNKhJygkKSHpIIsUPas0ZTy5qxDKBayLz+c4tI0DgVdanfzCfJ3/EFipH7t0Two4qw4Nfgk3HqbayZUV5F0p7Dh3LYst5NmTibMx/OsoqqQtJ+zE5EvrHlIi/E/77482/EK0Jtyaze/eTWxMbKE8uf/SRTVZln4WxrIsXb74sgSmj/GE5lE0mRtXuKX9pi7riv9C4lDN6DnY7RZgzZQUIvHytW87EUfRCmT/e+Me0Se4fnknGnyn1U9YfLQUIfBbRmjQ3Hx+kiCa/FiCwBm3P9U/g5Ug4ZC14fHCD/gSpk13yUILxD2Xi+886t3ve2BIF+UKl/1/7WFKjnAHratGSEL1zgzAI90xR8yn6Z7W/e86f0n5q/bfX+6e+3e8ptBXa+f79h/PxZ9J+4JPs4x8sZ461KhVlObOIUNov77doTSsU+aijFATs/8Yr2hI4J/4M1OdqRQ0GrxQMFviZ/SdDzV52aRREBoXqFuGnEknoNClAqh2YZjdIuXJ0pHRgnLduFaCwEntbQGvBJ/H71a3vYJOvXw3kQdJNrjDbZRky+dvJwoJp7bimI0eNtD/J6gSFC81vlTcHVj24DRHExl1mVC8KkhEffuBBcOgj1jivvvqanXzyKTZp0jduWdK06SGu0EHxslxYOoAZs0c++qgjbfbs2db44EZS1ihakrxFoAE4rFf+5Up3cSI6UY/u59rwD4fb888/bxcqKhHYr3H7BMY5WPg4KC38T3QpLJeeMWPGyqghzenH/vvrSRPt3nvvdUXIDYIlcbqoBH3OLwOMMqXLysrmdLd8geatD21tX3zxhRt1QKeVq1c57EhzuWmt1NgN+fe/7elnnrEnHn/C7r3nbrVVLezXdI8LZQRy0kkn22ZZPrkSRfdaWMDIX331lYc2p9ncUlQRfvvCiy60Po88Yv984EEpiI7WvRWiS25ZgwKKvkD/NbJgAlyZDu8QBg2uXAQsOkVYtnXr1Pbxm/TNJL+v/fGRc4qZqHeNhAhdQz5kG+QDhq9YUSElgwZNiOv69ep7rPYCAgOCAdet22Bly5e1Q5ofYp99+pnHGKcazJVYl2rrLK2XkJKlncuUO1Q9KVIQnssF0osWjDDZC+YtsHwyESUO+9+u/asTrqSQpvPoHGZM8+bNl7Zsqvv/Afw0c9YsMf12j8des2Zta6lyMBLKIhIuTIUKFRXz5Pffyce+UkBcrskb/jgMCjgXWKpaw+yTgckB5sy5nQGelOXMUkVrit/QehWarvFTOOqS/w5z3c8ABOdWL9H17F/UT0Jo8pAMf+Ec9XCdPoY2NC3Uz7g9HXq/49+hVOg3NxDXnbSf0D/hP2bR91N87n9l/qGc6dKphetw73+aaE0569YUwmUHKeXyRyQNcoplS0ixPIK2KTlKZqVsep3w26VbkGM8Q/cmf8kYt+EbjUhOFimcX9gPra1sqaKO/dBHobT9DTbyOJF/KX5P5H8i//8/+d+ycS074+RDfT3R9/lfI5S25AdrG5cYLhpS65zU/NcaKZ7/nAvrnyAbQonw6fInFkacklyhXJBA2S+wRgrns8v/uK5YjnkVrMWi9t2SGiByYRAuXRpb7GUpZ/Zsn/mX1X5ce/imT/4/2/ovzhG3Tw85jtvXD88S7j9ci8vwnbQv+mQb5v+G/gACX3Lu0YrWpFDai1bZxTe+st+UMx988IF17tzFpk6bJmXGq8IpFZivduG4r7z4wou2YP6CaHjz+H3t0sv8jXKrgXH8la/4IX+B/DZu/CfWTp4beIG88PwLipJ0mQeXoRA8gQzYoT3mAbKu4KU/iWcuViywVGl5c+B2Q8WffPapnXxKZ3v732/LdedpKRhOkgGBcFCk+HjjtdcFSjzF28+l8l5YfVkmbBl3c4L2+p9fWDKzZQnSSe4+4KiceOKJdplAi0uUDLg6tM+8eK7fc9qHd7R27dt7mOrtUjiwEiivaEnfSKlTq1YIkb10yXK78PwL7bZbbhVGSy+7/4H7FRSgoLePtc/rb7wZGqZidQvarJMio7SiKHP/KF4aSw/A/h9XrTfeeMP76VokSvqEMiNU92WyXNm8WW5lul+sZMC8sV3CnpV7F+HLa9as7gDI2+VWddVfrlII8UsE6nu6vfzyS6JTEbea2Zi5wZ566klbKW8dEvMcWm2UNU5ZRZeKE+cz5XlzjxRGnbt0loFHVRsgL5umspbCNYsU5rT2iFKcocDSGa+QeyScNhGumjRt5ucdK40x2A8pRxUz9BE3ozvvvlvMV1pKk/m2cMECadSKWAOF/UIrhsawRnp14bqUc1NphCFAwHNlTjZVMeLziclWrl7pSpdtUsgwmGykMSkihBiASNQJiBL1EdILBsSS5lD52gE8VEU+bzx4ygjst6gGLzwMdki4L3P/t5kCHEJpYwJ9wn0qr5Q4IDUzkERkWq3wXYQNS1IOUABJ5P/CUoDP+CFN7b4A8JkUjjGf7xY9jBev0BtaRWuKQ2mTN/s8gCfCLKQJHqhhwRFP/Oy9p1zcg7h9b1ZlvA9RX8JxyB2fpxwTNtSfVWvSfkL/hP/CfPgjzr/DWzcQIHBLlw335ngobVEU2eRyS7JL8ocFH3SO0/fcbXXdE9/hX8yeXi6WgbH8I6/LuFSxIAtTbWRrn1DaPboKmLM0mDNSmseYM1HZP+L4J/I/kf/xBPs5/I9y5qzOh8mU3gQIjFuTopz8YknrmCwR4vM/XufwHdY43I0yRfLEf0XyJ5v48TV51v27MNGHcns9e95QVKeuB6kWrsdt8is+DjlD+wGDsJXkTlFtSNdG7pQRPshe1n8/3P7u67+4d3GbP9Q+57PfP8vZOO1t/ift/9D4/zD9sZzpRSjtymWknIkAgdfmbFRcPDKwLvnuu8l2ycWX2H2KHnTYYa3t0T6P+nPVw01v3aznrJR7mpek3NqLsvkOLzL8kWpNGjex9oo6BPjsQ488bPf/8wG3nHlclhx16gbDgV2qgOc1UY3Wy3iAks5Dah/ojmVLl3pUYalvfC7Mn5fhkYhQOoAXc8dddyi0dAXbJMgPwHLDfAjts/9lHcCzP7aqrVW7lp166mnezh133iE3q35uqHCzQmvXUxhutyLR/WPF8sorL7sRApAesuPx9gtL6QJQbrW0arpfrENyuSHDI30flbKkkzVu2kSuO0dabSlumC8LZFlE8hfcuifwcfLly+9WKfR1PRgx+iuhPTm4sO+9NzR0PlorRF+qa5e7Gl0kK5ZSwoRBkdVKblq0j7tSCSl6Ro8a48YbyCKgUV6S+1cr7f3LlSvn1k/HdupoeTVO8+YtUD/Wh7FSXpQ0BfMXcJgUGs+tOnNpn9++/eHu5oRL27vvvOXua4MFlHzBBRe40o3xkabJ06TJ3zpujoq7rMP9a8P6DR5xiwzwE9f2R4q6kANVI0dVTaa0V4AfFcxf0KZpsPs+/rgR932izIo2KCxYfg1geTEdaNUA7hK/HJRqIjmtW7tOJkI7vY6lYt5VIi6EqlWjlityIASDScKypUZ6urVUOLIQs7yCrG4+tXWqx2klvUpZRWOqKLeksmXL+CBtFEI1VjaZ+t4isyc0pUw6UKXRZoqXFJN+q62WGdXmKK66N5Z8/NcUkPLTJykTmsQnx/HzzRUeOsfv2JeviANPYjmjTcAKNgEBnT8MrDJGKdQYyqYe9qmKw/n4Z3iTFZWI+oLwkASNq/P24x+uDU21E/U3bjB1PhxQQ9J+RJyYnJBWdIl/JvSHQxL+8xnzPzT/Dsdy5rhWGlkwZ3IyWhPcEs+ewDe+EAgiJ5AxOs6eC3lK4puj1LWfIX9pzRdoEbsGqS2gdlnOEEWvXKSc6SvlTMB+SORfam7HBNd3dvon8i+Rfyke0Vxs0aiGnSm3JjYcjw+S5cwvqJyJWRQOjeUFMoP1V5z2XP9wfs+NSEo86Fp8jPhx+bEH/4f1X8gVrwV/avsheqeUwljORErhdRt4sRr67/dAe/rtuFnRNxtH3zzq2/sf5QnHlNX6X/fsslJ5YlcGr0cb4Pg8E9nPRTM663zSfhYt9o3+RWQt/+ezjrKK5UMo7YtvyFnLGd/oahwJHCNdiwLDPGzvvPOOgeMyVq5NjRs31g3AGXIX0gacVEwBY4rizuIbz/BV9YCqsrBo6vvFrXq5/4T2t2effY4rbN4WGG+jgw9OPbfxzgBqw5U9atRdYjTf2WOyP46f5wUUibiQvEgGPT/IHnnkUStRorh9rMhKJ550ohXUPpeMrhxStwiqk5aW5vxImGxSzZq1FIK8shs46IKNGvmRndrlVO1zN9rdsg4hGhL3j/sgXiP/fPAhKWcljdQX0hYC4QhsuKAsQnb5eVnhaE9MuO0JEz6zNq3auPJl7PhxjmeDF4wn7c2pISNjntx7pACS+xbrBvqwZtUaKyNrFVy9MLYgeXfVZe+1f4T2qe+M089QYJ90q1hBEZp1euqUqcKoKeeWOCWKKSKTzkEvoiotWrTImjdvLqXXWmHKPmO3yLIHBZAa9/Z51i6cv1CKm7ruHUOLqA3y5snnXjcdZDVEJ5YtX25/+9vf7PlBg+Te1EdKur7qKxAm3jkbO3aMAgqNkz5D/VfbuJd5P3QjbvGEEmc/pUCxnKgcwqmeadOm+//GTRrbVLkQgZg8XaZj06dPkxnQNtc8zRNzFBQz5hGhtsiMCb828gLiA0MAtITSBFAeBOcRR7Z3fJlMaSkx/YKJQhK7aRAKKypTseLFrP+AAc7ImzZttvS0NNeoUV8RRXvaLNAjFC7fCtGZUNs1aqQLMKmg5RIf5dPbwII6RguZKcbfLEZxAR+1knz99xQIJvniDZ+V1AOjwClYoATNMr901scycJHGVIIaQOCBgz8OwJPaBFxxXieZrzFxyKU/JKzG3wWM6qQe6uY3x8hXxjFYTLHoiJZGnlfn9Zdqm6sq90OJK7F7AXmS9hP6J/yXzD8JDTu8VX2XX68P/dSVM8iHLsdqobcPKciyuAL/5fInknJ+wZ9REkwu7yKZll2g+XnknwThz5W/3kBoVodZ7ROt6Vy5NQ0cPN4W4V6AXO4puSx5HfqcyN/k+ZM8f5kL/9/6A8sZ1iQvDRlrjw0a5VOuW2e5R+7HRL9YyyBHotVQqrXd1z/k2j2x/smSP0EJQhnnd8kfT3xH6yg2SEHBG9Z5cW2+mf+Z7bvlzCmHyrpIkVelnLnx/ldcLLFZ8+bUbHxv8R1GN5q6X86n8qicdzV1ItU7p0tYLWatEcmWWx+0F9dLZUn7+4f+GXJrIpT2K33Oiwdmn75x2SE1aHCQFBWH2sOydpk5a7b1ffQxu/TSS+0tRfW94447bJyUNPG8mDRxomOcugIPoFlVgUEAIZ5LyZoDHQT71A+E+XJY28PstddfF67KffaaXJD+9a9/SQGz2SMUF9AeFMUKfArgbEG5Q6EpoTzdqlatmjBiTrYHH3xQrjzT7SZZuhzWrp0N1H72cYHnPvnUU2Efo/zfffed/eUvf7FvvvkmlFcV7F3BU61eowa36G3EALX9nu0nV6srdK8z3P1powwOUKC8q+A5wHhwr0R4AsA3zGOC8eRxzFUiLVF+3vwMGT1MsI9FGyyN2rZta9dee60UmUEagBuLBc7rclliTcK8wuhizpw5Nlf/2X+PHj3K2snSyOccnWT+8eUf4TeeMyRclEoKnydT1ilAnuTNpwhQuvfcmmzAogBjMnz4+9IzzLBXXnvV/vrXa92DhhDbQfxoHy+rHbx2vvj8C5ry9lC4bZN71BYp1FgT6UP4QqsNDBzoSQhsaHH3XXe5ngLZBcYP4cPZQ27asNEVe2APucWS+hN3nyZyMrkiMacqhACgLK+VVUoLabSO63S8xwHftk0YMbopFCOLFi1QZKbvbMJnExQ67CtbtXq1Su1yJp4zZ65PDM4RNosyq1Qmr/zxSpQoJkDgsvb555+LYQdaI0VRQnuJdU5kROMDgJXN15pQH38yXgqh6fa2wpyNHTfeAYQJ/0XYbdpZsmiJ14vmCy3kDqFr55HVDD5oy2U+lqX8ySnq/DHr8QdZtltnKsPMPlN2Owpn/VMDipDgYdzNAeD0poSoIKDzZ26OSgc/aJ90YTaG80w4cQKfflq/3RUuTE/OhnzRby/vZ8P56KJPRD/Odo17oU5SDF7n5ZP2RZBASRd4og6/EvqLLAn//SHmXzspZ7p2aumy4b6nR9jg9/cdc4ZZlE0qIXRcfHEe+ei/fZ4FeUnj8BvXwvzTef3O/vb758hf2t5b+yhhundtI+B8ougJmFNYYJmbAvZDMv+hfEz/MB7J80d8GDNr8vxlmqboEQCBD9OJXNZXljP7261J4iBqPWtEXFZkn+k64fKFrNlS9vUPO8s4T1D6esUub7g7kp9H/sSLpmx10XrIFXLHq0LP8gPtFy6S384REDnaEOQSrgkof9waQPsAP47lH9/86Twt5MoVIrBw/2yyuAbIK9cpp5/hPNd0HpBV8oXz/FY1ao8TKhHyJu0HmkOVHKY/45OxMGCGqPp9S2KWPBp/UlragQ4mW6Z0Ge87Xh09zz/P+vXrZ7gj3XrrLSme+e67Kfb2O2+rFM9XOBSeySM3pg1WXDim8CC8w/5zs9xsThK2y6uvvmKnKFT0ww8/ZNXTqjv/YIWCYoF6Fi6c74FqtktBAEupYuGuykpI7kxpaemcUYCcJcJ26a0ISTfYdYq6hIImn5QRVIHxwjNPP2tr1q71uskPFs567WEryhOFNqhT/zw0ds/ze9ptt92uqE817NFH+1iD+vV8Xc59uHWLMhKE53C5+KC08Ep1DuUI+KwHNTzI69qxY5vd98ADdvrpp3t0o/EKDY4xBHWABXO34EtQwqDEgi60v0zQIQAgL122TEDGGFNon795o779MLr/6IefDB8lSpY0ojCjRALyJCh3Vat4jD1/muhKG9sF1gttjj7yKFkN1XRQ5yYyCEFZRnrkoUc8ghL0z63xp29Sh7kxCNZB0B++Ba/nFY3b0UcTPcrsdinozjrzLDcSie+F+Q82bT7BqVCGNgJHeJEc/8g5i5moaxUrVrLSGlBwZcpL6bFNA7pV1ipz5s5xUyBkNPHaGaT5worJlHtTObkc/f2vf7cR/xnhWjY4h8FYunyZTZz4rR3StLk25Buk2FnjShSQpmvVqun/Cxcq4sqaTVs2ujkYvoIqLMYI4L/EbMdCp1evXoq0VMjRnNG4bVe/gjDJbQvmLHSw33JS/BBPnUng3Jvj5P4DVugPZcSskmbr7g9pneU6XK8vHvR+XcfO9PpduGBA5x8gy5klbAIUremyHp1kYojZHRmDEPIKoraYfKrNr3vL3nhcP1eiRLPK6psZVRX6J95RnV4f/ZEGOGsBAl9F9eoZnbSf0D/hv2T+SSS4vGjXoq7Lljfe+8Tue/JDly1dO/6XljN7ykUkIueQWfqOZRIth8Rc5J8+lEKXol/64b8j2RXyR7n2bIffUT0u6yL5R03Z23fLmc4tbeAbn8iikWhN71nv7h2taBGBBCby10fAxwnKO/H15fRP/fSD5PkjMuj5z+ZGBBKP/fGevy0USpv04lvj3XJGVLBumlv7IzG9fYLzCV+yVIroD3+6EjHOw3j4/FceJR8bHyPhXWVb/ziLs3FVXjavLviifN9ff3nlob7/on3kDv1HqXLrNWeEm6Aer0t1q111NOKl3eVkmH9Z7Ycyyqq6Ypnqlg1+gTr3uH+/u5CXt+8uDyFi0n6ggcY/p+h/470vhH2Bc8o+fsAS+iPlF+7INGGZVlcEXudWjTFQGrfeeqtHSXpuQH/DRenuu2rYTAHqLliwyK02vLh4nL0rY59Plhw7d27X1MktBUVxO7ZjB3vj9cHWv/8AeYlssyOlLGjRoqVddvnlDrHh7SsvCo3y5Ss4qDBzjZRXOKcZczLsIOGwcgreBkz38SeekIXMVFf2lBMcR/sjDrdxY8fa7DmzRGaYLpRfL6uO/AXyudIgLl9AioeTTznZ3nn3XbnqDHSvE0CHn322n9/rO1I4xTQBt2bu7LlSGi1w6x01r/vKpbDei+VapvDb7LH0xz5+xIgPdV/NFbFpoHURgO5pwrYBrwXMVqJIxe2vVsjr7bLM2SZlBoC5hBWHPUaM+I8d0f4IRTxSpCevFQr4AAUegh5SeH4hA4xl2v8f1+kEGzVqpOcU+cgp75opVkxeLdSJzPnmW+kImjVVuO8BbrE0fPhwWS294kosdAYYdlBwm4w9sFgimFD+/Pn8/qFjHblhlRVtsHx6QjRvK5yff+m7WfNm9o+/X+eWQU5vjf8bwqNxyx9X/uiGGAI6lcNJt5ozyQdZnawiTJcWAuepWvUAq1C+vJUUMC9aM1ySli5abMOGDfOJAIF2atBKlynlQEgdju1gZ511lh0sH72eAuLp0OFoZ7QN69c4wwwb/oHdfvvtbtFSpUoVadQyxQhbfdBnzJwhBp4iP8EmDlCEUggwIgRn/DABy6agzgEafLDQlrdt3W4z5GK1QdGeAP3dtGWTa+OI4PTxZ5/JggflUZL2lQIuPyIRED/8UnUyi7nm3xzyOxY34TwPzQAAp9CJYBuwCZDlTGbmljBZmRRqJNTNmwzKh5nCZPL/XqfqVv0+wVK/dcDM2q39UBZ+pk4WIFGuqL1QJ8VcXCXtJ/RP+C+Zfy5bcsutKWDOIDPuf/oDe/P9iS4/fu5HvPHxTQPCRimIGn1yQHt+wPSLNlSRvCOvl9jj938rf5GFpD3bJ1LDucKcKV9WIW1daT5cljNanGXrVyJ/wyg5DZ1HoCLjI5o6WfkO9A3fWdeT588f5/nbglDaitbERq/voDGynPnU+WT/fGi9FJjQ+RA+858uU+IWA1/G6x/OhvWOPv15x+9opxTJH2k4dE3szAf59c3/kMJ6Kv69L+17faoge/t+Dz/SPvMtu/yL5198njr9eiS7fHKKHvH9098fu/+kfQbkh8c/pnPW+Cu//wvjsif93UIllotc3MfkvKg6CBGNZUgxAdOG8TQFjympzXkdueK8bj17nCfvCbOhQ9+zM844Q/OR/hG9CMY2NyaoqPDSO8BT0W+ez2XKlDEMEtLS0lX/NpsktxhwT3F5GSuMmVNPPdXxi3JJicA+9KWXXtb3ZimjVYEqRXGBsgD8FBJtgS3ToEED4cX8RxYqJwjLbZO9+sqrcmO6yvPBbzF5pmkfW6NGTd8PeyfVJyINE5GoXbu2zteACHc46mj7THvb/s/1l/tWb88afcgN6zWHE/EOaG/OfdBGgQLYbqifOscXRhT/x955AGhVXH3/7LJL772z9GKjKYgNQWNib9FoYuzGGGuiiV0sscSSmKixV+wiKDaQIooUsVCl9947LCy7+/1/Z+48+2DEun5v3pdnYJ/bZuZMOefcmXNPWSSFCkyvcFvy4EMPelhsxgAKwSdstgQhBNrBDGmj9tf49UEKPGXiJHdtMk7mRYGaHBrVcplKOE+mHkyhPhg+RPKCldKOkyDM/T5hlbPV/fB4/1UOHzXIAk6TlssTTzxuvRUG+9FHH/Pxo9FO1+JDxdKwQQ4RZAOhBbAnTJ4O0jjRXkJz/0vmYxdf/Ae77LLL7K2BA1ymoQZ4/wnNjW+i1PintTvVgVI4KTXBjGsZqJHuREj+Y6pUrugqW106d7WzzzzHTlLs99Zt23hIrUJNOqpUSBaJ476/nAshrSsQglapXMV6HtLTrlTo60NkZ7dFCExMeMoQSov6O3XqrPjoc+0NOVvCJhAbPrEEqW5VdIlfXQmEGjWoL2dCedYAp0jlKghZ8GKtydBfJakuQQz9+w+wV1591ZYvXe4STAgQ/zZz5LQJR8GZ9ONHAGlr+DQT6opEEq7SXtxQmbKSnImKOEnkJ/kXWkUFqVOrsr7QrpNvgxityTOpKIUDs0qq0TWMImjP8JyaXFDnT8hVAt/h6JaXFZIEdVcxSRVSran7sT6vK4GnSqmJ2jkNeXWVgZ8Z/4gvGfzbfegPh8AnHckX7yy745HB1n/Q9xfOxK9Ukf8FdqIxFD9L4ZL4jf9P+A98KuZPuJv4X+RNXoM/j4tj2pc6T2NcX8d/dwUfjcbfniBH7RKau7kpZk1bQtSUDP/L8L8M//vu64/9OiGckaNOSQMeknDmp3EIDM/wbabzkUj/cf0T6dzNeHQBW0jxFK2L4nyST15m+PFMX7f+4R75wzNfWamucB350w+B721Srd8HfmgDTY3wuaN/bAi8C+pIONWR+3H9SZmw/iPvD+0/sAKcDHwf9zD6uxz/nd9aPnQ/7kfDD6pitkMUImiMG8xp40aN7ZRTfunOcydOnGgfyHkue0Mcy95z730JXEqbfSH3G3vttbc+DG9VcbVS5fGNMnvmLPmjOTDgiu4RaemII46wBx58UKZHj1u79u2sfkM5qPUU6uLdTFq9ao2ELTjUVY26h3sNHAL/QuGjGzVuLAHIdJs5Y7rdI1Oiv/71r66VU1OmWGh3kRYIVie59li1coVfM765Odm2TCZR7H3ZfyFA2iwFBIRN1157jV1yySVSfjhCmkMtVCbgJtYmQUMuy92R1JRCRWHSRnXJ+7rnnnvKnOlX0kZZLXcho+3CCy+U8+Nf25DB71szOQtG8EVal0RWXrZsuUyHtls97cc/+nik2rRUTnvXeb7X+73uWjYJZaiUxkP/27Rp45p7aO18KSfACJXUfB9bp0XRJk56EWaRGmhc/3zlVbZVfmVeeukle/SRR61rl042ZMgQO0SyBDSCwgcuc2fG9eX/Bt+38BFSbk5Zadi84iHPWV9hrvTMs89a71695MC4jr2oOjGRQnBDIiIXY8b88/+nSIH7lEbNYXRlcrTZ1sv+DRUm7OBQI2rTrpXCUZ1vf5dzozvuvNM6SGWrctXKISSWfMc0bNzAY8Ln62vbXnvv5UIYBg6nRghKiiSU6dq1qyMUxxqScCKJGzxosN17773SdGkqE6WmtkCSPPzTQCw49m3YqJH1kNftDnu0s6ma4JUrVvqEY5vKoPIPZF4hiRwINHHCJA8z1lmaN/FlVBpDszvX4WghqXMkfuYmjq0jdjI4vAjhM/E5DILEdYJaVkFOms9EOKPQictWrrEHJJzZIufRyhTyxLp1JFG/l3f4oQVcB3bG83iWwOXS6wjevWkBAr9UC3ipJ8+B4E/gGLrnEP2ZSulIysDPjH8G/0QbuyH9HezRmro5//ibNGde/yHCGWctKe7jfNMXGMmKAi7DteMY3CjhP35M8R8y6S/8pPJw5/vyX18UCV7k3xF+RTlkhy/XlubMMgnN0WjctEWhR/Uvg/+7J/5H/MrM//eb//06t7bTpDlDeuB5RWsaMM7PS+2HtQlrFqVgthNqdp6RrL9YiAUBBFwj8BVyuVDDeUxcN+nZd1j/BDu1AEcFUnznh8IPy6uwpvz+8JMNlZrzdfDFsNTRn7L/GfjxvfNN4+/YEtA0QZwfewCTzbVZ9lNIZhzzcgP0ZZ85ZvQYCWia+L18fZT/81/+bOedf747zn3n7bckzJGAQwnhBb5aVq2WFodQEAEP337R4qgvJQAqcKEPdWvve6Uc0l58yR/cfOnpJ5+SNYmce9MQ/TEOmC0tXLRQPln3c5+q/s7kvsp+rIhA7du1V9ZiF2TcpQhLRx55pP1cAh+c6TaQlg7CAiIZ33333TZj5mzRaGgTwgiiLbnJncJiAxCwVRXx6Y677rTfnPEb923zxGOPKST2z/2ZHnubEAqwhyfyEcF5gI8vVxL+WOADhJ9Gg+V1RVzef/8e7o/li/ETXBBC2zHTqiktmoUK7LNp8wavb6vGFfOoeRIkff7F5/43oP/rqp3EgOhP/y+RM+bjjz/e1xm0GmfFPqbKNW/ePGkn1ZN8YbsH7oFv1axRyxCoYaq1Q+ZKs6VY0UkBfvAfNFhmTZdeerGbsNF/nCQTNnyb/M3SW8YfmGukKYNbFMdNJlRpuAR0mGp9/uk4O/ecc+yGG27woEX+kMmn4VT6E6RS9TFDO4uEJAWyNWMip0vKt0gTU1U2eITI3lc2W61bt7G5c+a4Rkrz5i1skcybQCCEODVk9tSgYUNXE0PAggMgKKey1M6qV69hJ//yl3I0FFScxo0ba3mS9o2TalZVqTKBMK9pkstJ+rVF/mhQbfL46kLcalWruR1akVSZaksCVksx06vrXiWVQ90sssrNmzcqFFoXO44wZdK+yaQfPwLgbSC88JtcJPgcRz7BcWX2xT8v/1hKBB5YCnmKRDxhE/Bc/9G2RN7573/iHbtU0ZrwbRDL6iQQDAfV5YsM1e31QPwukOEKOCW05e9klUVSi2unkDMcGQnPq2fc97zcS4iYGxn4YSOWGf8M/jmR7Ob0h3AGXvLa26PtrkcHOwf5rtGaivXi9ygoX+F/QX0/0Bn8KGrQpI839OcbrXiER8GwSBRSCldpfJZ7uhl52K74767gV0g0Z557fYw+cgShebtWjX3h618GVXfCOdWEKOCmJbTVf/VcUNU+xqww6TfNZUnofdUvph6+MFQelk/+hU9HB6RMoV+69r6E2xn4YWAy4w92gGX//fhXr44ca69cbw88N0JIjs+ZUorWBNFEQbloKa5zdNfpkA0V9Ae9kdKfc49nJA6iXF9bccfNl6BZPXATDR1TdYn/8Jx8YT9DTbr+AfCpn+QHnQaNnAA/8q4AX88ExvuyE/ww/wF+4CGhxlAndYT+hxH45v5n4P9046/5CFPg8/2jfsDLpILV8ifDPnDxwkWO45jNsDmfr31qkyaNJDhYlghWFC3thRdskfL97W93WY8DDrKjZVL01sC3PVwz9cnDidfBfjNHPkuypaXCDfAP3D7hhONt4Btv2guq53e/u9A39SOGj7DTTv+VvTHgjRT8fAl1FisE9LRpX3o0IyiFrhMxKGjGqFrdKCPfp4OlmdJDUaXuu/ceO+Gkk+zcESPs+eeec3+sdJGywGdPXU7Rn5zedB0FK/t329/mzZ9rg98bZM+qXPv27e1tCZ4uueRShf5+kCq8/4SOx+0I7Qijpx7zXyY948ePt1yEL3qC8AltnsN+drj1uflmDzuND59P5NuVyEhEVsaR8Irly2XN0tQ2Tt5o69etk6+cj231mjXa21ezpUsWuy+f/fbrKmhZ7jP2ij/+UbKC1na3tIToUwBuMsUaJxOjS234sKFWqM7RvrIKxz1MQpR6MjHjRhkxB/wGYQJ2zV+utpv73OwOlRFCzZTVDeZlMAfK6k3ggh94Vfmy+J+R0AlhjdJRRx6lsXnb/vinK9XPXLvxxhtdWIXLlQ3r13n/I8/xAqX4U6qCGRAChz0zZs6Q3VcV2yFTpRXSUsE5D46B+vfv71GakH5VqlhZpk4N5dhnuA0btkjOj+Z5GOtqEsBgrsQCkChKINQee3SwptKKwbP2tsJ8qyS1ryyZQh2peOQnn3CildHEbFwnnzOSyG3asVFIXOyaMzgQ2rRxk8yWZDolBEILB/WnWpKstWzRyiWcEyeOVyjvWXL4tE2b/oqOAkj1K5bPCGZKA880FVYsfhWIi4V32CRwX8MMZTjNQVEQiic9JD93wgYD8iFzIJiwCdjfnuk3Oixe9IX2YoQzHko7lPO3sgj0q4RT8tUHBqTnaS9t4APB25ZsB0rg06LQqkSZx5saFwpchNZl4PtMMkiZ8c/gH7Tsq2gnF6f/QFO7B/0dJIfAvPTdIfCjcgis85N+3ikMxjf8ZhWJm2jhUcJ/Svgf/Cmk8DzySu6F/Mrg9KcbZNb48895bCnw313Bx9z0jBO7iy8rlLZ8zmByClMv9jizagYN9KQz+uB4Ed4HJe8BMlAo9N/zOMCkjOpyMy8qS707KKGxoGjsp5+FMhn4jGdm/AN+JDjyvwT/oGdCaZeaYMYHIdBFWH+BGDuNjLMMNIWhKRdSwDN0XcJDGEPSzrydO+ShnAuMVS/8LvAfnsKWglBMWZL13/eD73WrHqd2L7pz20vgh7Z9M3yV1frv6/sfNHu+uf8UzcD3idUPI1Fa4x9cY4B3pZA0R+w5hS6e7r//n9pX5nhbmd9CKRMQJalS5arC16XCS2G/3j95zZvZRx99KPOfU+39wUPs2aeflWnSv+zGm/o4jeCMhv3pCkVKOqLpz2273G6A3wgrwCkiC2Mu9LoUBjB/OeOMM9wfyyuvvGq33fpXtSXQV6Fa0ve5vpYrlxwHKaQ08ItxdKPnQaAirIcG9b9u3do2Qdohv/7Nb23gm2/a/f/4hx3Qo4ddfvnltkaCDugD+EQ0qicnw0vkE8VnRe2hRsJ9/+q00+yWPrfaFjnoPfuss6RRcpndJUsWfNp8IU0WtGUo8+GHH7pShbcTYYXgI9TeqghUCCroP3OeI9OfDes32JXyfzNrxky7U3Vdf/0NCv99r0dmIoLVz3/+C4WiXmJvKkoyIbevu/Z666AIUURp6vd6fwXfyXWrGHXd+j7/vIQ/E+3c8892GcEDD/xLUNR6/Ud54rm+fW3DRpldad6AjzAFbZyKcnjs/EWVMAZ8xOkjoczHo0bZawpl/vY779jZZ5/t7k/oHx91mH/M0mpI7rDZlTToerasMyrK71AbW7hwoWQYyyWjGGpPPfmkz/+w4cOlCXVlSmNHJUo9lapgBmQEIeYrCtLUL6fJye5eCmPVUpLEMjIjWmFfKkz26NGjFU7seNtbdnpMCo6BJ8op0EJFQyL8VsVKFSXRm+eaLdjH4fEaR0215MiIsNYb1m+UjdoGH9B1klq1bNnSJXOEC2slCRumVAh0ULMietNaqSjRJuBUlhMiHCwxebllcySRaynHRBU0UaPVtsnWROZQZWTmVKu27PeEBSKvTPqxI+Ai28AUnHEnDAKKCi9NZxuCIqrjGS9KGJNe6OT3PBARz/1PrF8P3KzpZAlnXhulDYCiNT1FtKafaz4VBo2CXprjzim+SMmTWiQkWSBS/wKAwCYNvlenPKENYtpIanlOJcnCgyq8rbqVgc9oxFHjvCRlxj+MSwb/dh/6w+cMs97vnbFyCDxU7/0sO/Fn3xytKV1jBlr6Kv8To3T+A/vxz2KQmM7JF2kslOGBP/I28IwiXKT4lT/RXZ59R/67K/gIZ9o0r2cr12y0tq0aWYsm9QI8YMaUsIbIY2lM4JnCCS2WGJ/wLLQ1/nov2E3Saf6UT0zcjz4ODkmLcO9bAAYo5dg5ZeD7eGTGP6wx/pvx793hnzt17ozAP+7K+YLTiNZaCeE5Seg8ff0DFOcRKXBQUlj/OP1RVoSX4ik6R54M/ZEPyuM0nf4ib+JB/EhHbi/yneCz9VeCT5F2Cd9z/QTwv2v/M/CZHkbhB80/HyVcu51afnzCYS8f94m6C04TlVdI5OjP3rEqoZ8Rfij5+0fwu3bZV3vEttJSGWQjZDo0SloeTz79lBQIutipp5zi+wf2DITKxoEwm3g+wrh5nTq9XBv6qlJQ0C2HiQXI7373O/tk3Cd2tTQ5tkij5N333pHfE5zhygmwm9fQALNtBdusqrRJqI8bTodqExGEfvObfSX0uM/bdJO0ON6Wo+J9OnaU2c3JKfi8xxHUEIGY8tAd/1BM+FzmPOWk4VqkqEbAve666+yLzz9zjZfevXvbY48/asMlfCBt0F67AmZcagavXvbV9Dm8eoNgpFmzZtICOt1ul/+b56UdhEZPr0N727/++S+F6kZbZbkUNSq5BkynTp1ssvbahx3e2y1j5s6d646QD+3ZU85+5/h+Xs20/m/0d98555x3rvzq3O4Oh2kPw0GUZRdYKR/4hdCrrAROCGPi+NeRZcyll11iN91wo73zzrv2yCOPyAfQQfIf9IGdqzr7vdZPdUkkpgowfTrvnHOMiFzwPyqtVKWSTZBwqEOHDhLMrHC/M+8Nek9Oj493k6bn+z7nsgZvEJWUcipdwYwaRyz18mXLSw1rtdvxoSrVRCpMmzZsdoncSDkAQpJIwqfMvgopxmBjO4bdF6ZF8+fNt9mzZrvtGyGr5sye47ZxuZLSFcjfDPZ8aMmMHDnSVZuaieAqVK7gqmIIXurXq+8htSZMmCAnwI20hizUswoyaZKAp05tt5tbItUxEBXnwmAdtm8rV69ypEWwdL7sC2vK5CmTfuwIMMqaYE2yfwQFiXUeUJkn4V+AAvUHJuQvexBDKbzMlU8VOIPitvK5WdPJPRSydbQtlVnTA88MsovPOkKaM4RsjQkYIjaVdfpJ4PM0aOEEGFzTkgg/qMzTTv2JI/HMTQvigoB6vEJKUlZLEQmUnJGWVKm7GfiZ8c/g3+5Ofwft1965RL93x3oobRjGN5k1pTYwcBDnM//J/+A/gY8GhhMFGjtvpsih5z8B/901fN++WfOm9eyAfds7z4aVB5YJD1eTlJKW6Sze0B0/TXiuZyzhv/FZSV1BgO+V6aHzaV9ch1FxGFSpAiVlMvB9WDPj/78C/977QJu2Utyg0um44WX942ueFAVCQ8rgwtES2orrH9rBV2b+kY8/yDWhXm3cgtA04FfqrnLoM2ey/vIyCQL+UPiq0KU/3wzfc+kHXvBj4KfzH/rPJ1t42bf1PwM/jMAPHH8KU7SUUnk5031z4Jt21tln2YQvJmgONa/6xy+BZ/4o0xn2prwruA9o9qN5eU29BeR/+plnbNHixXb//fe7I9vXXn3N5syZ68+JBlRVCgQ404W+qBlHuPn4wNR5MKdFmyXbHn34UZn6fCKNl4HS4njXzvztbxUF6l3vb4S/TcKeXr172WRZm0BzIS5UtmuwYGaEXxz2G+/I3GjfrvvaQw8/pFDUQ+2BBx5wFyC0nzDWbVq1UZ8QQCV7GElU1imENOPhfEX3ad9r/frZZ9KWQWDx19v+6mZUT0pDhOTmwvIpA70iBGksh8SLNGb+XtXdJorAjFDlxBNPtM8++0zKEGslpLnNTj3tV3bTjX1s4Ntvej0qoKhSV7g/mOrVqru501aNDxYyH3wwwjp27ujBe9C+aSYlifETxqv8jfLvcq4dfezRdvff7rar/nxVqEv9CEKUbPcd26VLZ+8XljGMfwP5oSFYEK5TsODZLKFYz0N72m0SHiEw6t6tuwINVdWYZ7tD6JHy57NBfmupM84/0ZrRMKL/uun3p0yZ4sK18y8439sTGlP6v6UrmNHAt2jR0vKa5xmx1bds2uLSRMJZNW7cUJLGTm7K9K9/PeDmReUOKmc1ataw2nXr6FjTViyTGVTOdGvRvIXsz1ZJW6ayvCF3sb7P9pVt2Ezbe599rJ4EK9jO5WQrtJXsyD4f/4XbDEIQzSW5qyTiIDwXL4mpX051dS5Ml5CybVe47DqK2EQ0KGwKF8tZMHkRHM2VYGj1ylWuMYMgCOFPJpXCCLh2iZBaiM0LDWLmRLTDRUB4PYRx6VuEHpEryednvAzDvSCJ1rmIMtQju0BJfn9zfHc5yRtjS5cr1JkcAnfes7k/jy9uqgEcyWviJ4HveWgb8NWOACmVM2SMhQRX3UnKksdf+X6Pekipx+pbBn7JeGbG39Ejg387E0mgkd2I/urXqeEmPnc9hlmTyaxpF5ozCW+E2XwT/5PIT/wn8JrIvHxZGDVqkps/Ff/9Ovjw1sgy6QYpxR917YtPeH5sN3xX/Dedg8avl95/5fMy4r+8OJL9n9/z94XqDPUDLKklqTsD34c/M/7x/fy/DP9coAFal1ZKqyulMQxZOf1xIkDQoxOOX/gz6MzXX3ocP1qRZyf6i8SmPJHG/bloMhn+QLOqlusfBN/JW22Rffz/CHz1xdeJ9P1/ov+7CXxHRJ/r0kF89nq5MpfBt8kZZ/zW3pGWCSYwOXqZ4AuF/d/ihYv97cH7E/iFRTsUEWkmI64/0YQW/+8Ped9OPOkEu/aa61zrBUsOHPziD2aDBB7sf6ET3m05ZXJk7ROUEEREqjLLDuxxoDulHTdunP3j/n9Y9+7dpdHxjv1JdTz5xJPKhkYKDoWLpCTwmYRDaPYIusoWCT4WHYOkuVFLDm+dVvXz6Wef2sknnWx/knDp5ptvkY+WcTL3ec7WyZfLxMkTrVHDRqozjCN0DpmWke9VD22t27nyj7OfFCTQCLrpppskCDnPHn74325qdPfd98onzPrQJsGfNGmiR2IaPWq0w4e+K8of6+RJk+3KP/7Jpsgqhv7jPwbTrx4H9nAtkwA9NAJrGfZ/mEqtXb3aevbuJV88A10r5eCDFIlZQpTuctBcqXJFKWXMk+nT3dYsr5md97vzrU3b1j5W8+bOT/GlVaqjx/497LVXXrHcGlKmUN2YhaHcsXfHvV0wgx+hDRs22p8V7XnkRx/72BOS/N333rUJEgCNkK8extNbqCPaTDg41qkSNC8cyMm2ww491Nt9r8bl9F+dLmFSp1AmdLDUfktXMCNuS8dQA9okwcxahcWaJ98xSKzqSZOmV69eLvFbLxMkwmaVFXIgAMH/zA5JzeDW66U6dYjUmqpK82WR7OPmzJ4lh0h13BRqmbRa8jRBLarWkHRro7Xfo737tEEyhvOk6lInW6fyi5YsdI/VPMfxEOGvNspB0wrFVkeVDcxs2aqVtxHYbVq1dh83CHKOOupIayoBD0SVSaUwAtJt5SXsP3AEkq6D0CJZjOs+eXjq9/26ZHEAoftLMJYXc6EOz68ylSuXlw32/vbs66M8KsjbqP+K8t0cwH01BJjMOxADudEI3U+apLMkKY8ak7hF0L30MrrSsyx4rDK4kMYrCdXEelNmCMoQfEWoGoeVXlcGfmb8Hb0SvIuHDP7tFvQn/nS3ojXtSjDjXMp5k3jIt/A/t0mHt0VhDHxH/3Zibc5gkzuwHl0HbgQOKr/u8fSH8N+vwg818RtboNYIgG/EaBnAEvbnWK9sfi/C50j7vThtC1/1khtaUOq5xoTqycdj8roZVKru2D/lI08Gfmb8oY8EF3Tgf0j/xfjnWAxOl3KK9ObVpvc/4TWMk24rwX8SAgs3Ejrzh4xmij+VCFoDXQb+Etq+M/3peXqfvid8nBa63yg14YfDp/3JuH5P+IHP/Nj+Z+B/6/izey+lRJTgAw860P2foN3ymBzBojDgKK2f5+W3pJx8lDRu3MjxGbwCz7Zt2+ob9CBYMasmfynTps2w3//hIlc6+KM0QJpIg+SUU091h/X+QcHJpdhDRK+T9gh7F6cN0VYThZTep+M+Hvp6u/ae5557ro0ZM8b+8pe/SDiyn2vIsAem54Pl16ZGjWpqIxsO0SHvPb3sNsr6pIJccMT3K4KRBTKj+rPq4Hj3PXd7qOijFcRm+ZJlYQQTekPLhtDc1B8Ei+bhwn92+OE2fdp01W/20L8flD/YfnbDjTfJHcnedtttt3v+bM3HOglpnu/7ohXIFCyuH3YoalK++rJYPmTw25klgQb7anTLhg0ZZksXL5VZ0aWp6aYy4KOlhDBqkkyGeh5ysC2Tg2AEZ9SLMKhD+z09KtVbb71t06ZOtasUEvs2aeK83m+A3XLbzcn7zFzj5jGFNt8qa5uqkgFIjub3sKqpWT2xeknjN4TGHjnyI2nUzLS+mndkEG+99ZYrbCCsQ4qFcKhO7XpqJ/xPgiyNWwVZAu2jsOS5OeWkqSQn0OpvJ/37KdJPIn3AcVGxbNHAJyIjDZK61Zq1a6T2Ncc67tNJsdXL+sChVbNaMco36LhJakRlJNGsJCQrq+dI9Lbmb5UToAl28cUXu2bMApk4oSXTuEkTqyFVqOOOPc4OPvAgG/z+YPeeTZQlpISEvmpYt6Ed+YujBGe7rVm52spJggZSrtI5ocYIpYWWDapKIBVtQkqG8IjITBBgJpXCCCRECBPwl7GYIIyGhTTE6YwxBUZ4w/OYV/c9j2fWWXIffs1XYsrG55UrlrU92za24aOnWV6jutaiqTx0ixhDXaFOrl1gA2JSMhSOtSTHkBcGk8h/dJ/GKnPJIQUXBh5b4llolBJVZ+Bnxj+Df9B9hv7Ag3T+M3zUZF06l4BT/EeCjfBHjlj2m/gf+aKAxAtyrfod/5I6fkr+uzN8XanxfGWKKdBBGv/1Rbf3jpbSQj8yTJ5X/Dck+iB+7UgUDqnsqTKhLDyb8rH/wPex45Ye0P8U/8/A16iUjHtm/P978S9SQqCHH/krvM/SDi/SRUltsf+swZK1mj8MPKSE/iLOpJWEPoVLrK+9XqpS+g/6Ux7+xfVfyBV/vzt8PnyhwxCT8wddfz/4JeVDPd8dforBJA3IwP8h8//N489shHV1nOUfd+T1sVaRfi+/7AppTHxo9973d59GAtJM1aZfiCtBS7CS8A8hgFMZf5/otIgKlPaR1UYL+TV95qmnZP7zmk2SI94r/3ylfSLhykUK80wiSA25N6zb4MoCnIMj4KewVHvQFe63BWHNVrnSeOTf99ukL8baHff80w6XgOTBhx6yV6T9QQPXSgsH9xw7/B1oEh5sVwSoHGn6BA+o1I2T2t69eqtP99kbb7xhjRo1sl/I2e5o+U694orLbYrMoZLmKyLyPDkk7u2+XgOZZrl/nI+0J95rrz1dgIS51TPPPivTpi8UZel+Q5CxTvv4OB8LFy3w/jmfEPzVcgFSTsoP7L3z8vKs494dbdDb73ngn+bN89x/LH0JA8ohOdfpOeec4xGVPtacHH/iSe6QGDgoR2DVgkkXSh3Ferezz8dyhuhPTz/1tCIvrfJ2qBpZ2yxzCLQpWwK1NZITtGrTxkEyRvSf8T/88J8pitQ0mzZ9hoQyzyv0dn03UXtYEan+KUue0Cd4lCnKUz0X9nCP7/s5EvQQLr1m7ZqhCwD+iVKpCmYYxPfeGySnRyusTv26PiioZIGAGxQdiVjjwRFvLY/axEKpdp06Pgh8yygrZMP5b0PFg8eGbuasGY6IG6WC1EzCmJkazFdeeVmD2dA1atrLq3PT5s3t4IMPcbW0XDk6wunw+0OHKJRZoSN/rgQ8G8tulFmUwmNLCldRpkvt5EAJNS782eAgCbWyzrJRqiS5hwAAQABJREFUO+/c8xS+q5LVrlXLnSP9RGO+21UbCRoEd6RIyCleaRmg+1yFFDcUXMU88dzr0E2vE2rjDwkKC3nVT015Esr0OmjvpErl1GOvU6VYGJQkf5DKF9qme8pD/SFn8quDNz9A9jLUCXzUD+MX3FTdFKMq8mfgZ8YfPNG/DP4Fegp0svvS3wejJok1pI9FGJH467iix4EbpUZLvER3/H8ydmn8j2cptkPZwLDCEfzza3LwV1J3Wim//0P5707wpVHIex8oHNLrjPC9BbSfbKmk1uief2H0e6G9/HptqsyzJ3UG/qsy/pUyVYmDCP3NwM+M//9m/APbA72mYfcPPkWoocVSKL9T1SUwnFaTj2dkhESDQFM0CDL5/0Cnkf6CUJQKS1Kgv1AmRf8OJoH1LfDT+02zyZ6Cz5qPBF/TedQqDDfD7w+DH/qF8Ogb4as19CkD/6cZf2aQIBulkRwPko8EW6UB0/f5F+3Dj0bKTGm6Rwo++uijbbicwhYqMhNYFj8osJ+sLVMl5pkPwSQsNqpUruIf9Nnof6DQyyMkVOjTp48NkJDntVf72Xo5E+bVjiChtSwzKAm6sheG/lavCcIW6sM0qXWjHHvo9zvsobtPtjOvelERi661hRJKPP9CX2luJC410ORQwtJk7732knbLNL9Gu7iB9ssoOVSXMgGWHnNkpYJPlmuvvc6efvppj/iUIzcepKnTp2qPW9n3vu5PRe3MkV+WlVJm2LPDHrZu/VrvP32ePHmSKyrga+bss8+y66+7XsKfe1xrhroYVwRWMxSJ6YQTTrCCbdvlG6aZax41bNLIatSuFTRNpGxRIL+y7MUZl53WwLruJfMgymEi2FrBgjA5Q4jWrl0bhSx/KfHbE8Z/7ty57pNnhvqPv5iHpfn0ezlUXiprGlXlTAIYy1Yut6OPO1bjp+hN3Nf843+mQYMGCovexAUzaPUQQepNRbdCE6d79x721tsD7UtZzpDefHOA3Kq09HNCcNPfbdvybbv66bTvT36an5JPWqVQf6GketNnTpM2yjQ7+8yzrUPbDlLFqu6qXuybHcEFB2QknDYJm7yOkkJ26dzZWrRqIQljDZdMVapQWRNVxk2Ubr65jw0bOlyetAt9cKcIYUZ8+IE98cQTNnDAG64VU0HOfevXb+DSw2rVqroDoimTpihc2GJ3wlRRgh5MmurUqu3hzw4/7HCXZlaUqRPaNJg8XXbFZQrP1ci1fFBXzKTSGAHGUaQI1cDchCMRqTmGp2GsHT90zwksDTRPg21keBmyCKBC/kEsYVHOgl05xficoYcqvRYInkpDvTxPKteJM23q5xZ5koehFTEjTJUFPvVL5i2Ano3HLgnXiR4F+FTkF5x4ysBnYP2/xiMz/gmKOS5l8M8pD4oJOLLb0F/gfc4gvubH+U8YGh+b78L/Eg4kUlPUwaKNodaEZ/lFwn85j3gXBCAB0I/lvyn4LKi1joQ9kjL8L8P/Mvw/0sP3e/8pdyCiUvqlthTnEa+FJZDqFBP9KbnQMdwPR9Y+/sdXeyWEIGGdpNr0DPqGn/j6x3Ps/BPXS6m1VexSGnx4B4uqmCeOVqwJPhV5Fsu/MCwl8GnwruFTL9UHwMnB4YV+JpzL4cctUcj7dfC/rv8Z+KU8/uyu+SuFxLw7zqiuOhK07KNowQR4eUbOfIcOHaZQyq/KZ8w1VkcuM8Bz/8gqBPv88y+sobRPmNvYFiIUzZg+Tb5KpVCg+/ioQeHgvPPOs38//LD2kJfKCe7x0tqQOY8c206ZIm0VwQ1aVQFtUQooK0FJwOcsacRUsjo1ytpN5zW2mouvtvGDbrbbb73Z7r3nPtfOAf6OgkIoxB3aEu14k1xzIFjhGZGXaENrac4USxuujEyJ5si/yuWXXW5XXXWl/eHiP7jPljpSgiiUSdfYMaNd4SHC3yBtGEx10MShhfSfcNHHHnOs+2q56Pe/t/nzFtiZEs4g6OnSqbO3hfJZGi8iN+GrdemKZbZ9h5z57thunWUZ06pFKzebIhryKy+/pLpVewmL8evwI9ciLVtY8xbNPUgPFjOrVqwOYcy15w8OeEPO5s3zPMDQHXfeId8wExT5sbW0hAbYEUcc4VNUWLxDmjFZLiRatWKVIi7N8vFn/sED+o9LlDJSAqH9+Jt96qkn7dCeh8rXbBO7Q4KaLl27KG+2LZUZ2ChFkfbEfKKtJOGS7/kkw/Dxi/0JuUrtN3KhH12h+uihtECgTYqPjs+YmnVrOWJfd/31duutt7v0EVUsXjZZOPFQKivHQ/UURQnHvjhG6qAoTrOEVKhLIZFEijlm7Bj3Wo1GTjd5U+66774eThsJ2NRpUyXFKnBv1+TP1sCB9EzA7HlzbJSQkLDZHTt2cU/URHWqIAleg0b13FypQYP6tueee8kz83IbN3acm1Ztk0TMkf5Hj8puWkHBOitaP8U7H2Y5vtjDq7NQZm4wqjDG8XWavNyFSJRJvUSTIUy9DMEzkE2JnM66wyVYpXr5Ba3jzWS5ofuOdyJO8ZLwXEdyk9ePug7MSre8YvIpiZuAW94mz8MCRRm8HcD0XKk6Uj4RvGwGvg9PZvwz+AeFZOgv8B8EyIFtfMNv4E7ORrSyiPznm/gfJXKydliPrButddZrVj5rndYRwXQBfgvPjAs67pNS9cF7+dO9H8p/Q4vFf2GPdFDznaF/jUOG/wkPAg5n3r/CB6hC/wO97Hr94dST2s1Q7scm0b9XAVUKLjSq0/bZL1g3u9XqFY/VdcIHaKLWab6pTPgV/IGoMM4nENRA50mTQl8g+fSb9ID6wj1/5AX4KYEfKgpwk+p2OqTWXwIY4CUNTwdFY5QC/KRVCf/5Nvje9qTfOwFOLnaCr3wh6ZiBn8wHQ1GK4/8VnA/4FvCD99f3/wttI/zxMccdp81+rpul9Olzk5z4XmMXX3KxPf3k04ZGDMIZcHOBtFb6v/66O/CNgpVCuenYLEuLanKlQRsQAiB43CTNjL9Kg+Paa6+3tm3a2gfSwGnbvp2PDbVJFuMJ36vVa9TQYIU+UL5ObaxL1D6RU9UKO6xdpY+t4PPf2i+6FNrLLzxr+3bb10NL04ONGzcr9POjMh9a4/Vxj9f4Vu2N69ZRPaIzF4RqLgrk+wXfKyedfJIELZVs1KhRRjjsHCkiBDoM/Vy2dLk1km+doh2B1tm/VK1WxUNFsx9nO7VU/lR+8+tfa99d116RIOuE40/w/lMP807EqlEfj5IvmOm2ZMlSF3zhrLiyzLCIeNy1637e3vXyOwvUEgmNzn2hwJG/YlfOqFq9SggcJKGOmuPt1VP5vOmo/pyssyzBWWK/u/B3grXQzZJw7AtLiuTZXxovn8scy+GZQmlroJjHzYq2VLFSeZ2zr8uSYkZ5m6WQ2V06dbHPFVXqOmkGXXfdtcGEisFVP1AmKZDsomLFyt4YxojyasZPktjBlkpydqmG4uCPCEhM3hY56EUSB1M77PBesn1r6PHIK8lcqbwkcghWYKL+qlAf8cKMadPcOXPss08/VdjjSh5Cq9u+Cm2lCUbSiP1cz0N62qm/OtV+ddppErgoqgU2s3rmDnqkGdNUobb2kL1cno44BEJlqX2HtkKqWi5x7PvccwqRtlR+Zcq7NK5GzepCvGV28y197KGH/m2fyrQJAVIm/bARKNy2xrZNuMwKJl9rbetgmwixBIoJmwEYbKjbj/wkz7kLrjvScx/KJyUFVNLP48aBavyejtwjOzjoteg8fmmG64V6Q/3xNxBWgOHMXyU9eZuArysW1Ql8oPE/NCspxw39hXuBBrygHmfgh5dpZvwz+OeklJDUbk9/vhiBb3x9Yqz8nRqYihhJyPtN/A8GxPPCIr1zswqsiX1g3e0ma5fV18oXLXW+yKaK92hYpECbLFZ01GI0nod5Ejxg/gD4zuvVeq/H66XtGfzP4H+CUo7yjIaSH8J55v0L9SYJ2oMuoWkcHHyH5ONHmW/48wH351QfoDlUnVbMWm57ZD9n3bNvs/rFn8hP5A7nE0X6kOZ8Q0d4RyynGhJYrLeoObTT11+xyb4ZAo7+dC8B6bm9DPmSvCX18kSJzAn/KXlGXUqxfp16+/1WuBngh/MolP6u8JPagbAL+GEdG9miZ0tmbef+Z+D72HzP+Y/jz+ghm2He0/+iQIZ78Ty808J7LJ6nPyNsdZEiLDFnNSQUYW/ZpEkzmueOXx997BGZ0/S2vfbe0/r1e90VCNACow2rVq8S/bGnCC0jcA1aE3wo9neo8lStXtV+e8YZ7h/1w49G2LPPPucuO96Sicxvfv0bwOidXODwcfSLg+GtW7d5v9izEFnWUciFUQFujUpyHrylv7XKv9sG/PsCq1mjakIKxe5bhT0tFiS0CmFHRYW/Zv8LLVBnrrROTpEAg4hMY8eMtdFjPvboQy+//LIiE10lDZfKTtO0beOmDda5Y2cJl9CyRbCj8urjVEVYIhqSO1jRvXHjPrFevQ61sWPHWj85B0ZBguhHmGdhMjRv7lyZUc3x/fcSmVytkkuTxo0ah6hNUya5ZcuwYcNdyLMTAXu/Aa3eCE4F9WXypEm2Xv519sTvzY5wX49kaVPJBTL4wWFOZs6cKVnAKfbwvx+yu+66y2659ZagYaPqCjRGm6Uk4skFYYVS4JBvHpXLyYkaS2atWre0k+TfZu26NXbvvffZC88/b1dcfrk9+fgTDr+wQAgg4Mw3JmZY1+DwmLH/qRK6S6WT1PBsqVShIpQvL8eLFi3SRG9RRKV69rHs+erUq+sDhjlRWU0mWitbNZnEFN9rn73kPKic1ZSApNCllVm2SQKVchXKWec2rV0Qg9oSUslC2aqBfPXrNZBgpZzlKdT151+Md0dDTBRoXVHqZsCtIwkiUkIcAUGweOFGYvnqq6/6pOKJGu/P6zeudzu7ZXIghDOmZs3yrEePHo5QkSlEQcC3DdZ3zfdt9fxvfp5FiDgRUcGacfbI6cU2clZdK1Pc0TYW13fkDgyELPonvPFIR+APyM8cisFAPBBDHH/OPW/ynPskmKeLVHUJXrDBkG4Ld0MVYqC6nUqU22mOeEZ2HREg+qZCsLwMjIL/XoZ83Fc7kubFSpFQh3aGO/SLRBNdLVLHmDLwM+Ofwb9AH04T0AaXOu5O9Ofdhp8kCb6QnrTM1LAEfsaHC64cb0JB8TktFkheRaiHe84//RMT/AeeWmSNsj6xhtnjbEXRXjZz26G20Ro7L4Pfwsw4BHODwNjgx4F5fU/+m/TB+V+sIsN/HbfDZDGsGf6X4X8ijpicnnWh49fxP9+zxI0L2dL4RPp5rO7bjo5/Dk54KJheXVI/88Lyu4KttD1y+lrz4ndtVsHPbHGh/PUV6iOq1kfkydYan/U07Q1+QCB2RK/BTH2n9Y/yp/X2P/A/rTtJ/1VRGv+hgf9BM3RS2WJKrb+SmxEide+EaxTTzfR73x++ahdPSx/7DPxk/VuK4z/kkvccDbZ+9L5PM3Pm3zIkpGRzzfiDAllyd6FfzxN+kn0EKOl51Ta9KusIsee8fqD2npvctYVlN5MAYKgde3Mt217QXbSgSDxjfm0VtWfspo33nIFHKPrRegkTwnu2Xr1B1vXpbtqnLpcCQSVt+Gtov4v2WBWV30PRhpfYRYdKK0SNqlFtoIQHy7UBqm+PXNHAbjntBAW06W/972ir+nZIEPOR9byhmTVuOM+O0f62Tq2KIKbjlW9/1R96B22V09valj5ubXNrWb+HL7DLbh5o8xYuUg7lSdYA+Hg5XlpAs2bP0r1gWkgI7Nyy5Wz//fe3MaNGu4Did/LFctZZZyv89N9sX0WAmjh+fEq49JyUFXLKlrGWMj9CG4Reb5DAhj00w+vjrxOEtKeffrp9ITOv2++43R595BEJQ26TA+UvfX62btlqs+QLlsA9rD/w/YJPnjq168g07HON6UaboIA++MUpdh6i+XN6ZxrDPEKTS2XBMmTYEPm2OUcRlEYG+Ey4/pAt1KklUzLvP/zBFEHqRps2Y7rdcfsd1nGvfWz48GEq97EXoNZIozjvxeyK6EuRF9SoWdP9ymKmli/NozffHOgOmPtJuWS4BEmMG0IZ0oABA9xBsw8QgGkQA1TKqfQEM2oYApETTzpJtljb7F3FZkf9hwEhRPXkL6fIl0xHCWfKy+5rtkvSeDbsg2EeS7xtu3Z22aWXSBhSSRouW9VV/RPRsXisW7eunAK1lpOhGR5zfq2klmvltLd58zzPz9tk05ZNckq01s2iDu3dS5GZVmswhdhCTuzxcNhz0IEHSxVsk8KQDXbvzxWltVNcXd645dCnqSSDe+yxh7R9ForYCpx4nfg1+JEJx8kJL1AmBHoKjD6d2XOfvOn5uPd1KdadXj69zvTn6fBjXTFvvI7HXd2Pz+Mxvf70e+ntSb/Pefqzr4Pjwg3Ghbz6d2CrFTq7y5YU7GOzin5mmwrrhJe8zzGZglZVJEy+yvj8u6AkMBqewXT0YKfktOHEEebC60henKip8ZwiYbZ0nhB/rCTVf+UKrJ56+Lqrf2qXn6sMxdz+1OGjaFYy/7Euv0de/Qvzn4GfGf8M/mXoL3KLdP4j/qGFY+Q/8Qgviedw0CgoZgzhKRydLzn/0XMtlBzDnC9RQnbUqhpeCc/i2hd5+q2XNdHqVZhoK3a0slnb9I4sbqE8bLS0ENNfGUUzgBdnacHkm7CkLUGAo3q+hf8qu9enlzadcPg0k6YFbpk8J2OSYl9jO7md4b+Z9w/4kHn/hnUGtBPpxHmAaPar1wk5BfrTRVh/hA1zfOZH0SUbXGiZd7Nrq+u3WJtMr1MPs+VmAJgVbY11LPuitcp532ZIoLtY6zfLylHEVGg7UHWYJ507s+FeoHGHTzu51h9H0k7rL4cfclA89okNF/n4K1l/hTUVNcU6qS+0mYr9SkfxL53+VPBpZwKMkwz8gESlPv6S/WlshYsK4sLHBvBL3xg8FWl/xjwEh7yOgQlWgBvCE+ZIZXhihX4hQxZF2alRjkI6o3Kq3GaVaoEvhJ7WNbUUr7Os/HVWRy5Qa8uCQxQUPnZkb7TqzWXGoiTPJJZVI9dhyHWuyuXq3jazmuQGbzdavaoVlFNhr+UIuH5lajfr1l6V0hxScXnly7fmdatQQuW4p980+ivW+zjCz96x2g5vvdo+f6aL/bNfE/vrI2NcOEqxVStWuhYPGitlyqg92ncDZp0iIZe8s+V0WOZP9957r/tmeV2aQQcfcJCNlpuQ9we/75ou+KpBMIOPmu3aw5eViY9/LI/0qDrRoEGocfe9d9v+Pbpbm7Zt5T/mFbv6mr/Y22+/TXNsnRQdKLdkySLLl4JFS/mOrSufPAP6v6mIymgLaX+vrv7tb/fYhX+40KopMA88yUdBJ50UTnyanBRv377DnnrySQnCKqtNzKCGR/8IfV5JfnkCb8i26iqf1zxPYbyfd8WNW265xV566UW744477aEHH/Jy4AvDi6LGKaecYm8rBLcLV3Qf1yevD+hvtRT0x8cru8i1gzp37movq54XXnhBvoiGGk6QcUqMU+fgRJ1ZUwU/QSpVwQyNbNu6jTw0H287NKh5zZvLkW+u27qhqVK+YgVroBBUTOCsWbP8vGaNmpI4bnc1rB2SJpaVIGWC4pozkAXyU0OUpFa6YFD5YdO+auVKxTsf6WHOqlat4oKfDh0ktaxSzX3StGvX1pZJYrl0+TI79vjjXJq3EbMqhbvac48O9oHs/xYtXiQJ21RJLys4EuKgaC/5mlm4cLE0frZZztw7LX/hcklnITyRh4REEJ03TMiTpRdUMS8BLWh5GYgaNKm69nvhuTzfOMHxhYH8kts5YkHM5A9+U5D+Un+yKNZ9hwHzoC7q9f5DbMojeDuUHzhqnB+LaQPPilRG5+GlFtrH/aCRAmIrj8MJ5bwO+uFUQdvoX1KXtwNP1vF5aA/Pi+g/nFN56FtJ+VBHsRxA6aH33VkluKv5a1B2gtXPGW+LCjrbzO2HWb7Vcuk3kmEcRzMGbBJUu49TfEGrtJf35uk0Lk647V8vQsu5coKnHIyBl7ajTXhCdqUSYgLHwuIvycU4OHTPqPJhgQDu8ce4kgJ8zljEhH5y5XnIK4rPwM+Mfwb/MvSXcJaEqxRZlcLZ4g2F1qXJGmlpSsV6zWip6spsFi1DOa4r5ii16xP2nifHggobmT1UAhOiRUhYIt8xfMvCxCAnW/fEcLJ1zFJ92c6rxK9U1mTG5O8N8uqfsy34klg7rt3q5szyv7WFTWxGfk8Jajo43y0WL2dTBk/V2gSGp//QMZyO63DcFf9VE5QEUfyXnAH/Aw7A00NiRMI57crwX70zGWCNm16APjapkcq8f1L45zjoiLx7vX+HXz7IUSP/o8E6CkdEw2wefH3hBKV7ybg4/iR5HJ38efI4Qa8RV3h1eqL6yBSJmmr8XHSLAFbPqBaUrJS9yjpVeM3aFA61qeIXS7Z11LqtrK/d3B0BJVk3xnapDIE7AuXrNnVRt6cEJnfgNaxPHf+TxzrEdRb3I38gS+RFPg5Jn3daf2lHrtb7P/9CrjPS18NXfWry94Xv9flGlbOvrP8y8Etn/H2QS/DPcQdEjLgtPAszDU5pvlOopb2DI0qYm4DFoR5QIb7/HCuoTpjhdescQUE4/0/8/2+CXz5nk/3xl9WsWpXDbJW0O0iFWjM8JFMeaK5FXp7uaEw0KG7qpPHwVQCdVj9ryZRryJAhdsCBB9gbEkY8J7OrmxV+GqED5khk4j2+SZotTZo0VpSmdc5roEPGgYhGR/zsCLv++uts3rz59geFCH/gwQfsrbfesksvuUz1PSM6VS3KX0+hqNl3Y/myQsKjylUry03IJ1Zf2jIjFJ57R+F2F6ZcdNFFNC00UCdnnnmWW7e88OKLHoinQMIQ5y/KQSRlgvVsVf3RF1Bt+bI9+Zcn27Rp0938qm/f56SlU9X+9cADMk/b2y6W82Mvr7pxszJw4EBXHtGuU31Tv7Te2Sz5RENFbEII6GsfjdfGjRsUwvsEu+nGm+zSSy+zBx94UO5ObvZoXrQXVGOsfWA5LcVUeoIZtQ/HunPmzJGkbr071q1RPThIKq+BRBNmmhz1YoKEt2rCe5WRHVsDTV63/brbfvt1tTUK1YXWy7aCbbZ8mZzxyhYQNaZrr7tOUrYt7uCINxMvDiRWhLpeKLWuMfKc3KhBYzd5Qs1p/BcTPGwYdnKd5IMmVxLEQe8NlmConCRqspeVORRSv/kLFnqc9M2bJNVr2cJq1qklNSeplfGO2bFJyCIntqyJPQVm4JOgGYHImRMn7TBD4ZFuI7FFNY4M/JLIF6QGTuZOLNz350le6oRB+E2OSsnBz0JJlUiD761Kg+9lHBQ1h9Kc/VD4qe57X6hJf8BL+v+18JUn/Evg6xBGT78ihGZlx1uj3Ak2f3snqdb3su1ZNV0iWuwUIUFYoGwRTQBFLbxwwwtbL0M91wz4dYkWiyDG5iU99/xqa+qFrufhRQyrom5aGTYRIU+ApEdJ0tM4tnoU4LPoiPA1rtSv3NTEL4laSBn4Go/M+GfwD3pJSGt3pj++/u2z41bnEvecGLjE9ilfiEbgG+GdEvn/ZYc6C9HdGQlbgc+QT39k+ir/TR75x0DOnbeFOsgeSjvX9FrggNXLLLRulZ6z9Tvq2aztPW2pTBaKi3NVvd6x2jT5BwUmTvPHP/ge8Ev4H1+x4a/AUQ4dgRAT+byEjpn59yFi9DQWmfcPWJJ5/0If37z+cLpxgtIKzgks0FTCMkRt2liIBj1B6E6DTo06Dfd3Rf8hl0oGFQNpzekOAJ23OFEHuk3gVyqzxrpW6i+N5xE2Pf8QaUB31lf7HHcD4F/q40cqNQ6zpqgxzpYuaaH6kKzf9TzypNiOdN4ROqSmed8oDZ9xbqLaOIbSzn905c/oaPIk5AfH6M7XwY+8kDKxxlDeK0l+vg5+4H++8g/8jxpoWwa+jxoz86PH3+c6zF+Yiv95/P+fpr/NWwvt8TeXW27TX9ovzjjF5syeK5+sdaWosEJ4jphBH3Z86CWUkVIDvloQZAhJNZA8MPvFkUe7YsN7g96zZ6Vhcljv3h79qdt+3eySyy9RVgkrcrJ9r40Qhj22cyh4g6pAcQLFCoQtwCT090knnmw397nZbr3tFuvSpbNVrlSV6ffJqyrfsLgSWb58hTVv3lzBdVa7j5b3JRzCbAg/NfRjinzQHH3sMU6Jr73+mh0n06x9tHe/6aab3HkwtEUPpkz5UoKhn9kbAwc4fPhftSrV7dWXX3VBEh/4t0vB489//rPNmDnD+vTpY621t1+uCE0kapkuAQ4JzWLmdNtWRVtCAUP99o/xnOv+scce66G0773vPqtZq6YRyfndd99zf0IjR41M4Afe4RWW4k+pCGaCZNtk/lOgGOBTNPGr3fM0WjLizr7Ay5W2yogPRjgjKyuP2B07dbRNMivC3qtZXlM7pGdPxZfvaxMkVKkmzRecG22StG3dunU2e9ZMDewKm6m482iJNJNTX+GPa6T5l0R/IRRp4vMM50x3332PtGeqekizzl27agBN5Ze5F27UuTZI1aptmzYet332nNke7gup2T57dbRKFSoJoSsIzIc+zAGdfQad7/rrxG/yo2lOmHHARC+ifCxEeaa7+gkqd6qD/zTGkYyHyk81qTpCeSghCml47Bm9HtGYw+QeT/5L4dMyukkXd9F/4sK3KPuZhDRf2Jxt+0q1vqcVZEulrQwesNEoAuGDNJN6Qn/pMYNWkvyZbiGoAyY5wgZCZ34d5sLboxrdDpqaaVtSldfJnKm0g/JaIOHwj1rVFKr2FNpATp4ncH0OuRfgU3cGPsOVGX9oHlwAWzL4Fxa0uyP9FUlT8r+R/+fYFivcLnPfwnw3U3D+K9Vh8NX5cGDASduh6ZKUzn/D3YT/6SLD/xiRDP/L8L8fyv/BHhIvD1GjDv8N6z+c6xZIy72gcJvWyWGhw5oouxiNO2lZuxY12u1hM60czjvY7PiHLHXKNVoCg/D+pecJPU7WXz4A8KEwDGEsdC4Bj7MnbqfqCSWDAUiyLkuKkae04NPWCNyblYHvIxJ+hAf6Bz7EYUmf22+bf69DBXvef7jwqdiG9b3YhQyY1jjOCe+IuFuI81U9D86oE6ENdCLcZJ7DfKs2XZNWrVppJyqSEM/22nMPW7xksZ126q9s+AfDfN+KtkheizxbpI/1l15yid1w0/WWKzyuICuPkxXV6JRTT7XB771rL7/yqgtCsJSgnkJpfcyeNceOOeZoucvIt7lz59j+3fe356S10b17N6tQoaKsO0bYJZdcak888bh16dzFLrzwApswaYJ9OelLOeUdJXOmIrvjD228nf7zlf3fpi07bOra1nbS75+yNRtxXHuXjf1sqgst7pNJEfvdSXKWG1LAc/bPmOZsca0adilUahKcdJHLkfJujrNZ++/3339fChPT7Nxzz5Wv172lFfKA3IAo4pMQ+6EHH7SmzZppzEKdXoHuz18w34P3MMZo2JJu6nOTfL1+ZrfffrsC9DT2AD7T5fOF/V++FDGOPPIXrj3zzNPP2D33/M2D7DRt3MQFNqPkA6er9ulb5J+mknzPjh091t6RK5T75Ij3aYXnvvrqvzhfAc5MyQJ69Nhf8ykzrQR+jmQL8xfMs1atWoW9PGigf4899phNlczgVglnjjvuWDtLmjjUF1qsyvyjUpaES6tkbtXSNsrlCXUyVJhKVa9WzQ455BAbP/4LdzJ8tYQ9A99+x4YNH2YXKoR4sO6gNrhA6Sa2mz86gew0jjDTKyURo4M5slNDFWi1zI5QaSKGOs538QCdK5uuXwnRsd364xVXSIumvtvI5cvx7upVqySNKy+1ooaKJ97V2in02Ib1itleqDBXmrhBEIc8SxMCG78xO6QBQ8rORsYEAmbZ7DlzpK41WBPzqDwrPy4iXOJOkFCB2iH7RBww1Zdz4KOOOspOU2SnU355ivuxKZSUsWb1GtZBBFdJBEXiZYipEbPJFKR+mUBuhJs6YSg1DnwR4MyRmfKgSExqnY+4L3njzdQRWIwkwk3OSf/X4ZeR4Wi9MjOtus1yhouklzlCcskCwNXiE0FXGFdGKKSgcqsxYrz8n+7rAhxImxi/5yjK4+QZ0xdqKpkLFgsxldyluHJSgJn0QqEkCw9SxP/kIgM/GWMfjzBAPkd+mhl/DUNYtGTwrwQjOPu/Tn/OQZwRpXMXeh7S/2/+n19YwUYt72GvzvuNTVvfTu/vAr1TFe1BH1hCdIu0CCzJeyzwutDe/+S/6pe//zhGXqm86nOG/yZznOF/GogM//uu64//pvXf2u1V7YOlPe21eafZNG0UiSzDhpmoNx6UQ+sh+sWKF3qPNA/mR74R3nnKo4zxXNk9sdGJKXJI8oT1V/IkKUT5nddf8bnW36mKY20/HD41fB187mfgp69/GREl+P+PHX/KqxKUpXCbUKi/HdLqKpC/mILCbMsvUDSl7Wab84t0zNJftm3ZJue923Tcrs8Myd/mbTm2eXtZ/ytfpaHVatDalq7ZboNHfGFTZ620VRuzdV1k85bm2+LVRbZiXRlbsLzA1uRXtNmL5I6hYlObPm+L9bnzaXvvgy+tbaef26VX32uF5Zra5NmbbMbCHTZzUbF9Nm29TZi11cZOXmdTFqi+NRXs06kb7bPpm23slPU2atIGW7iuqo2Zssnu/Pdbdsp5t9n23La232Hn2dbcNjZ+xoYwdMkrk/GD7jdtLrRnB+Xbnqd/YiNmNXWhDJGFa8pZ7esK4z1QTmqryr/KizL5+Zm0SHzYNf7uXkNENX36NDkirqYxhN8GIsMiZc7cuTL1qezBbTBFnDRxkv39739X5Kcadtedd1qvw3pLeFPRlSEmylEvqnS+NmNupREBvVOO9z9KCChX7C2ToTffeFMmRae4iePll18qocbB7uAXyOSpLznAmWf+1gYNGmz7am/P9XIF3KmhPTcBgWYpwhIJrZcF8xfY7yX8GDt6jHzF/M3at2tv1WSBA4Pp27ev+6mJ8NG8QfsSf7JwFeBVKF/OunXbz0Z++KGiNv1KZlfz3FzrPmm/kBzFVBe0jZUPQpttkj9QP1/ty8u/DtY4zZo01b0gbF6xcpW0ZnrbAxJe4aunc6fOSU1eZan+lIrGTGwRAzR96lRJDedardq13dlvRXmwxttxPQlfmjdv7o5+uEc89gYNG/gf3pSR3H2oQcRRL3Hdy+TmWoUa1SXuyLIVK5e7adSee+7poJhMqBbTKYQzxHDfsmWza8QgXAnCEA29JmvBwoWudXOiwmERWmv9hg1GPdXl2wbv2khEGzZuZFMmTZaa1BR5cx5ux1Y4zhrs8wcx3nyZMiW2/66iw9feIKXNFuX44tM352JQsvPzlxHiNlBDyOXPue/SypDfy8tngL+0/BkSZkoonwsiknrwdKVrNaKkHscm2qA2AZ9ySR1oDjlMEAvYmAWBeMm5IzEwVKZIKm+u3lmssYK45C4GyXMZ1Yn/AjLRZq8bLFdbiLTkKE/9DktHfS0JL6dQtlB2X2Ucvp6pgzTfmYxKeKUc9IgauF61rZpNWNvVFm6Vj4MyROva6j6JcsmmdmHqFuVd5I8vayce1Y+WEym85FkO0F8vrJ/wjOechdHSucpF8yP1MNQqhuMhFsnHmIXJ0dMk6R7dSr2gw4WEgT4iVO4pBd9vZODH4cuMfwb/MvQXqGBTdjstPHbYnPmrZE5r1rVjS/GWHD/nAwDcXd/E7O3hX8qXmLQ42+fJ0oAoKFCTnAyKV3FehG8w2S0VqQzhsWU4IF4s3kZe1XtQzXetXBktWAQ2cKLwC2vatiPHxi3rYF+saKM1RyV9KJHPmkKiKARheFjYBepFeRFe7mZJOsKXneHp/D/5b3iUYvwJXww9T1qhchn+G6gh8/7JvH99VfM16w/WIdDjIff/TCRbZMOfu8SFpdu3b5cwhA+SaA3IJD/5C/lFcNBcQuo688S6heek0y59ynnC73/TU1esWEL6ddMH5bdqu99BUxkaLk7WV2vzK9mYRR0kjGmhNVlZfUTNF8+gZFgrUb/+O5/I4VynLmTxe5wDP8AhL2tGP1Kef/6c/oY/hLuhbVyTI95XhaqHajk6DB311CunvK//PIPf8h/a9m3wff1HXdSRBt8HCwh+P8D3ylRpBj54UsrjrwodgzTevLvAlfgHroP3/HEv4gvnJJ8jsEP/01PE/9sVseepJ5+yN958Q+8g1c0GRWnDpo2W17RZoCnBRQmAOSaS8IEHHGiPPf6YRwLCrOZBaZE89uij8styiywvNljhNn8xK1pPf1c+4IM/0YbAyB3aO+6QFioJ/6kkLDk++eQT+9OVV9kt8u1y9dVX29ghj+qJwnLrF1zekp9tf39hpo2cU1sCiGxbJwENfSaxj+59aC/757/+JcHJcrvzrjtdY+bZZ5+VAOMO+0D7Vw2O1zVkyFDbI9kzx0EhCvF27Zsrqx0d5G91ujRKcmTBslD75MelxPDLU35pF5x/geio2K6RH5l1a9YqYI+cJidp+/ZgIrVJYahpq7K5OdJ1115rRx99tDSOFtikCRM95PhpvzpdShkN7MMRI2JxO+rIozx6VKHGeJDMqfKxjJGbEvbd+8qsCZ8wG9ZvsDatW2uNNM/+ds/ddsQRR9gVUuB44fkX7Lzzz7OlUrRIuILDXy1zq6ryX8P40Xemv67G6eijjrEvp07z8OJDhw6Tls5DdpsiSLVS3Vdp/Gk/aYvgP/jQg7Zy5UprrShZcsssHCiUBrG8xKI5LDQHh0j0/9prrrGRH420V197lQmLQ+vPS+un9AQzmiHavlk2bXPmznXhDC+Qtm3bShhTVVGTNruEDmaWl5dnE6XKdf/f77fj5Jx384ZNtlbqU7Nmzlbc+JXWq1cvl9jVlf0c5lCosCGhayqbt/26dbMVCqU1X058CA9GWC7YOKpoCxcucCnXFnl9btyokdSbFvgkEZ65SIIFpHIVpSqFdK9ho4bu62az2oWTYbR1tm7Jt88++8y1Z1pJtSk3t6K/+CJjiIPONf2AEZDipCXz7DD9SXjsRO6Dk1yn8oe5jtV6Pdj275TwsauBjbCAmw7fmREDnyTycrVTLbTzK3lifam2qExE9lgXR68vDT7X3wSfMhUKvrQG6x9ObQpoT1iE8tRs5ZbqNnrJHjZ3XZ7sGXM0zrzoA0HxnBcjZXyMvd269mPsh1oa25T0jf7SflhtzBvvhb6GUeHcbYSpCii6jvl2fqn7U9VGffzojzI6AoM1C3UF8OFhLEEef+bF4tyFPBn4mfHP4F9YqO9O9AefgPbHZ//FF4Q3vtZXjKTQ3vnFhf7+4V3JV+cdEoyzAL33vfViN8V2Ye39A/+hvPgJiXp24n+ROSXPuOxePcfK6e0O1+EP/lso5/CfLm2pTZa+dheU0zuVzcU2F7Y4X5LqNvUWltGCJL7fgKXynugEjE93doKvMrFt5OMZueLCh7pDK+CXGfrP0P/uR//fH/8DHTnVO/2IhrXu5I/1KjwCnsE5NBX/oL+YoL/IY0WUfrtIQmFlVrlt/iwlkNBOFbp1wlVOztdsqWyjFrW1ycsbqyYipGpDLFIGlq/RxAvIB6/gL35Fdz9XzhESPkChBL7Tv8rRpdC2Ej7h195M+EXynCb5Pf9xeDwVuFSfvd3kT2BwoP7Yfxcm7xI+uQKs7wofwXLJ+i8Ztwz8Uhn/BGPCPCc4CW5FPAPH4jm0EHCxBDc0DY7CAYO4CniyRhGKEIjccOMN2pi3co0HnlHX+rXr7PBzzlPknWGOOEQT5j4f5efOmythZMDv1Yr0e/4FF9hF0uSgnjGKaPTuO+8GeMKJFXK3UU+CGT5s0KJcKRdgnQEyBlGJuVIA7R4zZoyb61SuVNn+dMGx2vE/KUe7BfbW5/KHOqGs9R8w344/obMiGy1xPKZC6qyg/SuRlKpWq+oCTcI//+XPV7kLEZzc9pLQBvcc5GUMpkye7NYh7H9J0DvvH8hh5YqVfq9R/YYu+Fm6dJkH1dksoQumOt26d3dfLyhGMAf0A+uTAw48yN559+1Uu3AovExhsZs3z7OGDRs5H/jkk7HyT7Paeihcd/NmeQmNCZzgYnJEom34ewnmQuNt7JhPFIr6MNeGadVyD2vUuLF9Kj+z/fu/7iZY++7b1SapzKE9e6q/X3pfEOqukZXN+eefb2MF0z/6qHP4tvn8i89lTtbdhg8dKrjFLlRDOQNnx6/1e83mzJrtTotpHEoZQSAdcALfPDmYZOofYwU+IEc4UlY2Q94fovnp79GkETT9FKn0BDNqODZtBxx4oE/yFA0c/l6IW964UWNJ6BQWbO9sN29Cg2b9eql/STjTpl0bhaBa7eZH+3Xfz2Zr4EDsXAlR2rfvoPjvZW17vhatWsQioKmnsFvYfrGIHS3btOVyfERezJw2SBtmiwQ12NcdLilb8+Z5NmzIMMHc5HZuSOkiUWPLRjSoZdK+mTVrhsJjfSbNnBXujBj4vPRgAulfJCAoypPiOYyBxP3IJL56zzMkP+SLZdPvf915rDM+i/VzzbP05/E8HmOe2JZYB893BZ9n6Sm9Lu6nw+c6PvdyKhuicSjmlKKMILyANYileTkIctGGGjZ6UXubsapu8iLXGPuYJv0RkYWXfGKrLEeUwESCna1NhAC6JDdULYLRP6TqpNAGzkRKmhJmJS6A9ZQHnlhQwJwoG27zjNyqS/+8Ts8JQYa2J1Psz8nnJVQkwKesEmPHPcZBBbibgc9Y+ODw4ykz/hn82y3pD6kvtCA2AT/ji5QLS7TBil8B2WTFr4FoMJIw7fRUwqZS9VCf83c9S7iSg2Ajgk4kXAj+K3A2YWVLGzm/lW3apo8N7tE+1E9J13QsynET0hzX3pEGpPiy8zHaqXO+HJHgkt4J3Q/8z2+n+F7gf+ojueCJnuKRexn83y3xP2JCZv6/4/qDAQvUxiIHWmQtCi1yHvlEvE/uuK7jHgn6A9fSk6/RRI7khZR9pUJ+3cNPHx8512yr7Ou0ycubyHREX/9dsKKjnmehlSdzEm+H7gOLzZC3NdajI7DZNHkQjEj+gNG92E5vV3ymC9rCJbWRAi/RXWcmuq8HsW8BosYllofXJHDJw/OUH5vIh3TzP+HHCoD77fDJFdpFjzPwf4rxZ7p9njVvUSjDXoxzjmhwgP+cOx6mzXvEmXT84Rznrn/9619tkcxTzrvgPMvLy1P46NWJP6Qie3Pgmx6QhjU7kYTBH3ytsT/MLVNOeBPef5vkiuMf//iHTZw40W684UYpHrRzbZpZ2uS7hQTl9LGZ8vWkcbMaXyheV/DZhkPdKtoLjx07VveL5bNkoLQKptnl5/a2Cesb29gFY6TRutLLICgizDQ05K46aJuERAvmL7TqMuvh3S6gILW0XZ6QcGGy/V1tO0gmRG/LR8tQOdh1HKUBTlnmfW7ZsqWPG85xG0v4cfAhB8qcZ54tkwYO/lVRaBg/frx86/zSPvroI7v6L9cESahgzZs3xwUjBdIcKS4MdLZGUZPmyfSoXbt21kk+dKpIEWO75mf23Dkeont/CWecymhHCbnZiSedKPOkahJQvWWdOnXxeX3xhRf94xQRpWrIquXkk0+2l1582dZKXtCpc2dptjxgHyiq8o033igLDzSGixNlkHneN8bCuytcQTC09177SAYwVMMEZ1MsusGD3acOwpmzzjrLoz6NHPlR4C2sWlQfbUSTp44UQ3x8pZWMTKCs8K5hw/pydnyiPfX0U7Z1s7SMd+6SwyiNn9ITzKiFOTm51rPnodbz4INtsiRQoxUtqZxsvfAZI35sTZo2lkMndVAendF4Wbx4sWvW0GnCch0taVR1mS8hfVu0aLFMkNZIpamt1cirbs2bt3AP05MmTtGASw2cl5T+EdarikJmo0Xz4YgPFXmJsNd7uhOhNq3beJguTKSuuvJPdrjs8HYIoaJUjQGEsPHYPHrMaG/PRb+/SMTWRs6Nyjtz8BcMyK8EIyDFlwPnPI8JovaXFIxCffIXhM5hKLEMx1gmMpF4pJ705zu9wPSMemKeWLff0E8sRxldOPzItL4KP9YR4cZjrJPr9Dx+oZ/0fPFePEb4Xoe++lKDWqEZKrZxc8vZ3PxuNmdtbd+UQM6pvslMqlAEnq3QsWX0RReCzEkWEzj4KpNd1gkl1JYsVLxuIDsJchL672eCzJirDzqE+yAfF2qPt9PzhWq9BojWnwsfdAz9T19ABDheZ9Izqgg1hrGiPOW8vM6j+q7Dy8DPjL/jVwb/dlv6E1uGrXoUEdhRwoOcXwo34Ifpf+GxL1FD3qR83KQlzCepJbzHIn92XiZBi94Y9uWaPBuzZG9bvaWC+/DycNxqiP+jQUqE1g7HBD95zrOEpznLE8DgAyiUCe2PvSBr4H8caVvIldzP8D8fS0Zlt8V/8CLT/+8x/6AMAlwd+VFi/cmGlPVcXAtCp3F9FzTDWXVppNPWnPE5dSi3yDN8mHIhBfPiBczWy4fMJ8v2tmmr8/QRlPW1NOf0sMjhJ5tkqcNQt9clumbDE3iFqnGeoSdUmZwjGCK6mzrjcCKf8PLK4/zC84Y+lqy/Sniat08FAFsoPyOelD0KUvzatSuS9ZuehfWXcgiut7eU4GNa5R303zDWGfgagVIaf7CAdbYmTbgP7gRhDNon4BQfL6CDHIQfmtP4xxzw3PEpmfP4jDrYz4FHmPwgMLjs0kvtgAMOkOBluU2cPMmmyR8LdWKlUV7KAKE+CUOlAFAmJ8vK5oatcs0adUSD2+09RedBu/XWm291fy+XXHyJ9rujvN3l1FbSfAkrmjdv7ue5OaIbfVwuK1PARdr34m8VzRxgPvnKKBs5fpVd8LsL7LdnnunCkHcU+WjDhvXuooPhiPAVHUU0UKTIR+XVLoQ9ZofJ7wl7X/7OPvtc6y9tkMGDBtnFcmL8nPrLHsylm8o88sORrjzBWoOEBQn739yyuRIYVZYpj8ybNe4TJJh5+uknrdehve3Rxx52mFUlcFm6dIs9//zz7kLEBbWqGu2SAikyVK9WXeNcwSpqrNdLWLNJ2jx56v/7Q4fYqTKR0gQF0qHRSerdu7f7oH3hheftnLPPcf42aPAgy8vLs3lz58qx8iw5IG4sIUyWYC+VL5hz7QIJ1ogCRfTmurXr2BKN51sKgU2AoaZ5zT3vNrk3YcKRPUCyLEEYrKaNmvo+/+yzz/a5Of300628FEX+9Kc/ujkX0ZqFRd46BG2YNLkrD90BjzZvwm2KIl2RxI8S1hiuS/G39AQzaiFqVSuWL3WJYOeOnd1UCL8zeXnNJF3aIoKYK4Sq5I6HNM4+EM8q7jlSR6SIaNJUqVzFyueVVxiwOrLle8wWSkBz5M9/bksXL7W8Fs0kydpiY8eNldnTLDvqF0fZyI8/knBnnn/tW7J0iRwKNbCT5EVbuOU2Y2VE4GWFdF9KBeosTXz7tu0VJuxZ209qUfBqkJATvmLyZaCZvFB3ltSPdqN5AwIjBIpEnj723IvMIJ5/9Xm8jvm4judfLVNyn1wROTRQghOuuL9zimVoJ+e7SjEfz+M58EG2mOJ95dCt5KWmOsmXntLhxDIxD9fBt47iwGe3tsse3moTF5az049tqFr5+qsXPO2kT6o2NJkXHC/+IBWHKdNhb5uy8oxrRoGWJYW8SRF+qC9pJ8X9NMwPFyGfF0nO2QwlVXlm2s1YJHU4GAfu8MkS+0iLYkq6EkoBh2e66VXS3iRDBn4cE4Yn4FRm/DP4B3kEYvm/T38eelV8BPz3bqvr8DiuOcbFZmqzg/mAFqGkFD9x/sId1UBdkddhAhsr1YO5m/Ns5uJetjq/hvtiy5HgO/AvqeTqOefwWyIa5PhCN9dy9WGFNvClECGM21fTNp3TYsoAQg3m11OkZRqoFng+fX3Rmf7IT97kqMNOZTL0nwylD8z/ffzPzP/3w3+oiU0FtJfOJ+LGlNpY9yGsKVmbQG4JoQVw/puiU+c3Ca/RExSRyT5yxWE2d6MEMgqmkS1TxjLyP0gqVP5sCJU2JBtJeARaBfCJHG0Ofd1GRbQUYQ27IG8C9B8YVAq+7vOcdXcQUicNoHSKr1BXwjt06lcql+oXN5KUsBfldoC0wAvEvDvzHyopBfgRuI4Z+MwLw1qa45/MuQYXXE+fy/ieBO+5z/ymn6dNjZ/GslhI0ETeaSNHfuSRfl948SW75pq/yH9pVbvrrrtSfmCIpES94Dcbe9pQVgFrSPvss5dHELr77rsV2Ga1IiPdYEcdfYw9/Mi/3bnsJ5+Mc2UE8uIWY53MpEBJd0wr+GVyy2hfmS+fKFXUFr1nRQv0acaMGfI5c4vdcP0NCtN8nH/g7SOfNtA2+1fK04H8/C1SRKjstBjalePmOpvkJwfh1Uzti98f8r4ESkVy5HuHddt3P3vo4Yd8b0v/58vdx+PymYMQAyEodMj7v3zlqnIUXMOtTjp06GANGjRQxKTl7o+nq/bDR8l/zKuvvmL4asWyxKfb90/FrlVUvUY1ac0UKWJyE1uvqMdV5MNmpoQqTRR5qVB8g0EAQ0pIN14V25577GF/uOgPLgTDzIngQJiF4Qumn5wct2rRwtsPra1fv9bu+/t99vHHH8sn7Uc2aPB7CuzTQzKAObZNwiE2/uDiCoXnxpdsoUzJKOdrGPGibvt3c+ub1/u9bh+NHCnftoPtl3JYPKD/ANeAytfchH2gKSz4O4oYnef4BV9Bi7lQfme2aA6YsyICD/1EkpnSE8yo4Zul2jNs2HD3wNxBg71eEj+k+NjRVaxQyYYOG6qJ32hNNXk4+EXlaYG8JYP0hOqqLIkdEzJetmGNpPKFORHemVfLAdEzz/WValFtO0CTUKtmDftE9w486EDbt9u+dt11N0hlaW93OLxRIbZr1qzl6k1rJbVj4dm2TVvmS0SyRrBybb1CcK9ft8EJe8niJfJKPdElhhCjv1g0sVFCG9VHhVOePxI61/+ZQDuQHaT7+gQTYVJJ35Qvlv62PDs/L4HvL6RYSdrx2+DvXF9awV2c7pw/vHALrZ7Nz7nc5N7ZJi95Wogux8GKmlUmB0YaVOhBfpgkL3kIEfU/xgVmg+DG54Gj8jkJQ13JX+hlaFCE7/1VXn8W1hEaYK5IaLLw67OjY/jnN/w5PxLUwOj9ZmQafjvc98VGrK/keZzqUC5URpvC4iTmz8DPjH8G/3Zn+nOeoAEIH1wZCXEM8Sv4He8aUvwaDo/WA+dFubmKNAA/gfc5J/SsXjZeR67FE/LyN3HjIfrSXaRFnXiv4OzYId4o/goMZ6OqnQ0VfBf+m6PFXw5HeDE82DdcCI0C/43vLAqntwRYpNA+Fl/6p+bTpoDx/jTDfzPvnwRvArY6FoEiqZR5/6avP+L6jXGCxOLaKPKKeA2v4F6kQ4aT81g+NbzJSba+2EOl5bX5jGU4Li7YQ7xA/jC0PvP1lzZTzh+0GQEGvAOYmIQADz5RVptF5xkS6JYpEwQ1kQ+4lopgOtvSMfQNvhA5g9bAvqkpaWFsjzOQWFB997VDvPby8Jy4hg7lI1yH4P3nOflC+m7wk4UjDCzC+w/4scZwZG5IGfiR34f3D++NHzr+CAuiQJJx5X3EMR3nwZUUvvgMlPxwP8yHWpBMUDlpcpDYQ+61157akA+x9we/b1/UqqlQ1pdYu/btfGPO+xGNGeCjaNBOSgOzZszU+xHN/SyPIDx3zmyPhoTGCuGor776GhfCXH/99YqOdITNnj071Rj8vZDKiEbYQsT9jdOOaInQ8jVklnTEL37uzm2HfzDchS9okuRJSeCWW261hTItKielBRGhKzMcc8xREkyMCXSvNs2QMKZWrdoSpjS0BQsWunbL408+oVDRU10zaH+Fl16j/XJIxfb5Z5+Hj3P3PI8AAEAASURBVEAiLHyrEi4bp7ddpawwS8KURgrKk9cizxDQjBz5sX0ybpwd0quXmxKN+niUNHEu0j5/hOiSecqSBsoSmVzV84A+7LmJFNWydSur36C+LF8WegRl2tVUFjOMoSZFTdFR/5MfhdhulJqrKgrKs1GCJsYfgQoCMuDQ/5ZqV2FRgfo/Sv5hBlodmXqN+/Qzu/yyS6W51D+pI8vWSfbQokVLyROWBzRUWfgeyh8OWvXxD3co/V7rZzffeos9+9yzcuI8wCM/k4e+IK8gWlWWcLAA7SnNF5GpaBN5Yvs5K81UioIZ7F4LXOKWL+e7MyUBnDd/nqtCVpLaVg11Dj7JJC6WZsvvzr/AnQi9OXCAQmat8AlgEtZJaPJ6vwFWWVJBQmUTOmutBCr5ctI7fNhEO0wIe8Zvz3KbNNTM8uVbppyEL21kfnTSSScrTPYQ5Rsm576NrGKlEPIae7x2ckI8WZGXqlZTCKzmzW3psqXelvXyrD1WUk7Ml5DuTf1yqrRqOsircx0XDKUzAIQ0keCZhHQGkH4d7Hh91lJzRblQF+V2rd2SXqe/EEFIJcrHa/KQ0vPGa5hXOvyYN8IHkbQE8rnwSr7yE+v8Krz061gk1p1+Tb7Csi38RVxeTI5oS0WaV+xCC7U54AUZ++FfXDR3mMDBqFxAI4TnawwCG+oC+Vko8Ee3ucfLnZRqK+f8saFRoliSxa/T76kmZyi+MPQHlAuM3IsJCLCoJNpjA5Nn/nL3yrlSOeX1S38arhl/b2hooufjx/PR/gz8zPhD/wnOcMjg325Af6zKxJ7Y7LiPF/iErtM3WvC8qJ1ZJlt27coAX4R3sNB1swBdRNYC3vDFDa7ieOS8S/m0cIBX4petSOrWLgAvVN36elQoJ+s4AuRdS1v4cAEfDpussLlCGERbwlemsCAGCnV+E/8lQxCoe3Nge6mU4X8aigz/F5pm3r/Q8Xfm/0IbaI4NXVxvxWvukVhPxeNX13/+QD9Ot/AHrcdAxFzxlZBYjwkx9awwR1+EtWYrZA0mvlGgKE2x7vDBEv8bEt5qveaadjq6UJdr+IV4GrwEXvHVRH+dcyV8LD1PXMdRhnP4DpXEdWzgf/RT7dQDQigTFdXLkTFWxrjG9VeKS4aWfDf4auG3wGccY8rAF7//ScZfgpjwBcNxKuI7QpNIB4y9j7/mgyOJIzgY9xdxnjiWk7CFmWujPWUXRf9hYw8evywtkOHDP7DrFIHopZde8sg95cpq8y5cm6ww0k2aNHaTpXLSdBEA/4A/b8F8CUHqB7rMKqOgMZsVwvk5d+g7dOhQaXDsb88880xwRouDNyW0UhirJUuWWl2FiZ41a7bKC8/1/q4poQpmQA0bNvT+zZw1RxF/+tkVl13u2i0P/fvfMg+SYEjw8YuzeMky95+CkIBxWrZksSxMaktYMtR9z6B80Eh1jZRGCE5y0QbqrqA5vz3jDHvp5VfkWzWEuIdwtkpDhMhNOPxlP40bkG35O6xe7XrSfmkqq5MVEjTNsQLle+nll61nz57W9/kX7X75skHDFj8zrDFefPEFCT0qerCeihJ+NFAfx08Y731kTdOvn/pz+eWBrjVnzIVPG8JZ9SHQeKCt6tVryvGuhDEau+ryQUNmjy6sMm3btbX2EhjdI40l/NRef931duqpp8jvz9/toIMOssceeyxUrMoHvvWmC1IYfxJQiZS1eOli34/CT0CzaYpMdeopp1qfPn2kAXWTZAXtbcjQYbZW0Z5wpwF8Pk/h7w/XLJGHOm8MLCkAKMXfUhTMZMnMaLsLY+hYTUkisQ/bLIkialpszAmT3UaSOSRhMPYDDthPkrbajtBI6vCijFBntZwBL1m2xL1CF6hONG323mcfa9mypat0lVddxG3n/jipi61avcqdDG3N3+beqkeNGasXRhkPgda6deug4qTBRYq2XYObl5dnIz4codBiw6zHAQfaHh3au2OfGtLE+fTTcS69O+aYY1J+ZtIJf1djH/PEF2A8pucnTzpjj8++Wtbvq728W+KzmHdXx5gvwo3HmJ/n/H0Vfnq+eO51JfBj+XiMcOJ1PKbfB2H586+zIrBsMa+y5cpbkV7e3FczxEC12BBjISw6L3V/wXPOokAEj5ofz/0lLSDedjFd9SKCdCacDtdfyrywIUG1Hzip5Of88FTJ++c5wzm3/AE5SPpNbjhMvwyboiSbb0KSzDrEsdWZMrBfysD3gQw/Pqj8ZMbf8Ycf6BFE9BspdEtQKoN//9foDx7hvE+v+f/H3nnAWVEkf7zJOS5Zwi4gIhlRUFRABMwB9ExgzhjOHE9PPU899e4Mp54ZPXNAARXMEhQEJWeBXXLYJS5xSf/ft3r6vbcISFjO+/yZhn0z09Pd1dNdVdNdUwFeyx+CGY6eL3oHn/A0otVBLdhsU49EOTtN4AuLUtAnUUJlUjZqWhCxKCoiM6ZNUsHdomOAg7AGAXTQkPHCcAQz4snwXsEHXoL/Cj48O0CiP7YJpAMhAVwJrE4tF2XanZj+Y/5nWMKPcCfmfxoITzZGu9CKpyj9it74Z74cRFvQptGk1kGcs26y8hpH6Jp7/NlaKTpageiHfGibVELrsQDJNigSavDBkLC+m6Q5g3CXtRk+FqBl+mF8AbjK55yjCXeVZ/xE/RJYJd8PzpKJVvyT0VZI9DekcB7KRY35ejYcOCqP6EeVkkIY/yS+vmqrzV+tv6yVvYePbCh0OYYvwULE/xnZAh1/oYVhXRhstQ++p673tz0HPn/bw380QlmXl9HH+lE/jrQNfhGZ3LLHQPMCHzMPPvige+qpJ6UcsNLaIZJwv4/7y5EtYeQ91uLyAprA9QaIgGYpmjN1FW57mjRrLrzoQvf88y+aM+GH//qQ+bTJUcRhzABBSqIV9erVSxYdawTbCyHR0EGAQvhq3rGYGXF96x23u8suuVThsB8zIQHWJ+xfv/ryS3vGIipnPEJ1oMPl6jfRhtHuwZ1HP5nn4Mf19ttul0lSf/ea/M0c1ratRaRCacIpQhvvd/be1apVc4sWL5ID34NMYQFNlBVydIxly3IpSxC8Z5mEQpgAoVnzwAMPaK8/2z2h8cLHy6BBX7hKlSvbuEhWIz8689w0haomAjIuQdq2a2sUiHJE8+bNNZsMn2GNTuwy+tlqAX+QBRzSpo1r1aKVmzt/ro01xTDpGjt6TOQzSBGZVix3f3nwL+77H36Qxssb7lj52+n74UfW1vSp0+zIXhLBsjgcU2Z+ZkuVLiWw/gNVlbQ042X33X+f+Fph+bC50g2Qz5q77r7bDf7uO/sgBq0Df4KEdWhX2Qc0XYNT5Bd0KjjBjMYYydlSha4q3ryYa9GshfmLmTtvnsJfrXVrJZErok262a8ppFbOkhw5CJ6oJ1MEJmP0Rdw3klIRXgxP1OslZEHKuVmIOH/uAnf66ae7li1a2oTgEZsw2mv1t1qmS2vkv+YXEcVXshcrL3Mo/MfMmztPKlQ5rqY0Z9DemS1ELCs7OvzdoNa2TN6yx08c72rXqetuUIz0b6RlgxlVtWo19PXQq5YTXjsQfzimTkBgAql5qefhZUleatnU8x2VD8wl9f6256Gd7fWNstu2EcqFetu2l1o+nIc625blOrSTWoZz/nh2HHUV1heWrWJkSFMR1ARtIdgcyB6+viRe+hLK+AUIpk76s68nEepzgJajFC4TWbwk9N+2BfQjUQ6GzXj4dkK/uU2VQFosULhHMi0GEbNvjzxrWJJbKkTt+Fz9hhzag4X7YwzfhoaZjscfVIrxzxBif6U/raJECX7jYG908Zew2YJv8LLnXci5j8Igp3/6CGEsJ+I/doh4VOo4BkqDT1Gfvy1FtLkSH0bIUnyLQneK/4ox28aFRoFtfFf3rR9a3BXTOe3yj0UrhMs/uKmhb2BqAhhxRbpkKfA9LpL3/JKWLsf4z6gwpH7zwLmfW58fv3/i9y+4YXjBWoT1B5daIyME5V7QCIC+rZyO4Txc2w39kB/aCziHo28EF0GoAwCjbfEFHF+awEXreNZurM28FrLvB+sx+5imjU7gHb5PEszQv2jzAs+wPkXPwiN4+L5noS/qYYT/Pj91vRy0SLkD/yEZ//Gnds1pKu+xMsqwfB1T7xUU/BTwHo71xMOM4WvMC2j8GUvQh3kLKYk3IWfnx1T8Ly6cBSFoY8XKVQrhnKHXsacr3o+EUT777LOlefGkTJuucX369HG3Kgz1EjkHZnYRvoB/CC8QymDZoQ2v6KGwfNRUNI2LN958Q+4ylmsf+5WFgb7zrrvc4e3buzvvuMMGRnJPt3TFUvfe+x+45UtlWqS60BF4ikNZgtXgMBe/TdACe9u//e0R8xvzyqsvu4/6feyuvupqmSvNsQfHrRPw6R8C1SLiF0VpkzxNRAk54W1/1FFupCJArVJwnUWyTKH+YW3bSRvnjxZxjYYyMzPNn0wd7YXLla2gbi00RQo+kldSMJ4ypcrauqSRhDZYlBCpaSGOeCWE+u7bb3W8yI2S1cnSnGz3qYRct956q8tSlCfgp2kvX1PaO3OyZrv09Aw3YcIEmSmtce0Vxtom2J5EPwliFbXrudCGwZnyUe2PlKNh7SMjwRZEN18aQlXlg3ar+BQflzZIeQNnzIdL+ENo7T/ecL0bqoA/X0gegK9YGmeU4CnghPn2UY4uDR9qy21K165d3T/+8Q+ZuH1tsoX2gvvOO2+7Rx56WNpQbxkvZv5Hjhrp1inyM/PD/CtrnyT6WzBJPfRmKiwqi1gYLLRVmkrtqM4BdVwVmQZNnz5Nwo9sV1G+ZdbnrXejR4+VGpg8OuslgFSN0F1IDHEAjCkSDpu4t3pNrrVZXdI7vFlnzpplKmcfSHpH9CcmDfXQXAlqUFW6/PIr3IN/fdBdeumlrn5GhmwGSxgiLdCE4iComtS+asveDWEBCFG/fn0RFwhZ2HVSqLEmTZpKOIPKJtob/o8FM3kcwx/luc91KAehhXO0hDgPbYU2wv1t26M8edv7C3UDbI7A4phaL5SjjVT4IZ88zretF8pvC5vyqX+p98Oz4fGcMuFeKI/H7yKF6aeeSffxfo1tM1+BIVrGnyNq9OGPxUAQ0oSvtbY1EAF69cYkurKQJQXmHV1FBM9XE+r4S1soRNTPOf+SjCF/OwkIPju69G2pUnSN8CnUS9RIlKVcDD8e/xj/YvpjoZHgPzqHgyBvRlCT+r6AfwbeLKNmOEjyPjxb/iHsa7XeKfBI3rMsKoO2SzBLMmG3+Ggxa0/vAfFXNGP4Ss5HiZLhWLK0mTSgrcgfvt748k3b9jUY/hk42i7xX/FD/ed5eUrbV0bXiefXnZj/xu+f+P2bIBQoAoJJrGPsgh+frRtewMI6hzUnPIO/7a0/WZMFHsI6LKzRQr4IXHSpNiK+YWaOone7L54B78CUEV7BmsyO8Av9lZSJB3nBfAn+Y5rN2kjaGswYm3gGvIK+ctRaCy7i6d8/WYL+E9zF51sb/lS/gfMkMpLrP7vlP/5xN189wQvJ8589hR9aSR7zrz9j+LY51fDsi/G3WYzmkvb35A/6CPXQliDxER8BRglFBjbhn/AMVxdXSkuC9++jj/7NzZLpDpolfWV+06VLF9Xya33WcoSRPkRhmxGksDfnPYn2xWL5MiGcM10matJTTz/lsLo4WoIRhARVFTqbj9KkhdqHbpY5MWWhDYQHRrOmELDVrAYY2/PP72XBaPoP6OcmKHhNTSkNfKpoTYSZ9mOuRYTGJjMr06Ia15VPGjTKgJ8j7RZ8pACkvBzxzsrM1B54gRv+4wgpTTSVhsn3cmJ8mAl0aGudLFXYn2fNnuUyMjLMAgWzxjRpkxSWBGiDohnXql5TjnqbKcBODRNkjf55jPGRjz7+yF13/XUOlyWrclGWWCPrmfUmXFm8cJFpqLRs3cqNUJTmFctXKuryAh5cvnF+SeFxGpiI5BkTnitPMgFck2TNztSlhGiSjPDxirFhb8n4M6v4pu3QqYObMXOmu/DCi6RRM87CWaMBRf8RLptQV2VRFMFB8UY0iDU2wGLvP27cOF9W84Jf27PPOcciWj3x5JPuyaeeMPjMP/IItHlIAb5dFPBPwQlm1DFs20oLSRlUVCEZNTbtONwpK6nblCmTFb99uDkoylIkpWxpqDA4+KZhIi+++FJ34oknWB512bhjJ79C6kor9ZcrYkCAw8Z+laSJmXIcPF8DTeQlHAUjYUSaxYI0TaZUXbt0dcccc4zr3PlYd9SRR8ujNlo8ZUw4gLDoCEnt0jMyjGAwvaLdNAltmEyINLwEtz2GFyPExAstLK45puaFF2eoH16mofz28rkX7ofz0E7IDy9ZYJEX7nNMhc+9UIY6qX3dUX4oF9pJrcM5+fwFIUy4H8qH/FAOp798/Q2CFy+00eZAmwKEMiWK+wUACwUcZNnzSNDGgsGScAk69ef6FV5ATMaYonxe9J5RUUoFqKD/ytbeR4sCE6CERsiM2rAzf2ULxXBNdd7qlijv6/ClKVFO5wGmLTrop+76ftFADD8ef6FEjH8x/eXjP8IJ+IhWCrCM8A4IPDzweDY9xjvh7+KHZtbJOZsq3TO+a+8eTEX9Zs1vuLxghS989vWbd4RttrzwxYQ10YYrCG7YaAVTBPrBYpOjJTqp//5cv2KFO+O/nm9SQZwSthnjf4z/+fDfv0sNhwyp+El5r/rL+P0brT8YLdYUqX/QJn9h7Zd6HtZj4R7X4X7IQ2vG+In4SOArOLfkvl+Hsc7zH8uKJT64SSAjvgEvsY9nrNWM/yBAlhBGde1o/fYMw9ZF0L8xAiY2mVLnP+IWqTfteZMZ/ozNlTVluyy/6eaOb9/DTPKfqOyv4P8a/7YPP3Vr5Osk4Ft3Yvj51r+afVKBjb+159u0hqOfVDrYnXN8lpDmzMmSsCPdwj1bfdER5k2FixZ26ekZVmZJdrY795xzZT60Qv5U3nDXS+gAWrONz5EVxqRJk6W1gXABrVLRovKzVae6zIFM+KJ8NvAjZTLVSgKJadN+cacoctMll1xsGijwO96jIDPvWhzU4nwXzQ9Lqk+ZSpUqy7ypKWBM4+U+mQ9NkCnQKy+/6h78618NPiVnzZgpFxw/ufNlIrVe+1/osm/fD03BYJV8qGZk1HcD+g2wffbaNavdcLn6IGrR5Zdd5o5TxGOe3+9rCrkxY8aY8gRrCQQtE2TVUlcaJZtlzoWJVPMWzbWPbmX7ObR7PvzgAzdl4iTzY/PKyy/bvnuT9uxEnaqqUNbjxo8z5YlSJUrJPGuSXJxUlJPg1RJOLXR9Xn3VfSONGz8Y9uT2c+Yf/uBqyZExtM7aZ6W0fRCmsf9CU6lixUqm3ccQ8vxV06q4Vq1amZPeHGntfDZwoLv66qvcxRddrPPP5Ceonsp5ev5xxHBXRf1aKwEddRlcTMew3iGis/EX5eEU+TaZgF166SWuc+djNLY9ZcLVlhrqrtW0KfRnPrsgf1O5z162K0/v2mgzaTlyGDRCk4/vF2zimHQ0WtbqYfEOvXzFMpclKR9qQZ999pnUqJZIuFJEA1bZBhGTKOz42h95pBz6dJLdWxNJ7EsIcddJ4rna4WWaOPQ9epxucedBbr4qIHlcLTUp4A8dOkzemkeZahgSTez3LtREdejQUcKcuSIICX4kdSM2O1o2x3Q4RmGzzjT/Mtj8pRI9L7bU6x2dp5bjfHf+wktze3W2dy/0IZTnmvPU/JAXyuzsuD0YoXy4F9pLhcG9cD8cuR8WCOqU9O3kYyZ8fZEgxm8O/EbBnMgh7ImEMta2JxljWv4FHhECVKPT7b3ojbg9VenUi1KtXKBe1TNnW1HblOcWSd1Vrv7RNrm0E9Xz6ve6oWRmWHaERVuO/VqnokvrQgzfxi8e/yQegTAx/vG68ZQDgkBipP2R/jyvgef4P3gn54Hn+nHhHu8R3WNlqGQqvVGdUJf3X+KPZaKK4teLxRXt0TabKYTmXjvGC9e9pg0Ccf0hQFc55gdIoS9Mks+B/9nFTvhvVMa+DEYbMibZGqTNeP79YGhM9nP8j59/1/ifbQIiRpmg923pP7qG1n+rjOEfn35DG+Ir1OHjZhDOIqwxnqF1WVE5C4ZnwCPgPUlBTxKWb8tjdmL9A4anrKPAeOCkJtvfGH+Aw/h+2KOGH5W39qJK3gyeC62/eAaYk5XVKdcq/Nvwk33YPvzQpo4J+L5OgM9zxfAZ+DBW+2D8mcht8EVQ9jgFaiM6EH5UFsskKQTz4APGt19/6w6oXcvaRwgwXhoU3aWZ8uYb78jR7VP2kb9MWR/JDJ80mCxtlXYFeMsel1cbtKFOmxCBhg6W71L2u6edfqrD/cY99/7ZTKUyMtJ116MuNIGgokrlKuZol/yi9iFGJlLS5MnKmmOaGQx1ztJs989//t0c7951552uZ89zFaE4zZ5j6bIc80UDvGL6KFNPPm9KylEt0ZrWKbzzkUe1V/1lsjbJMtciE+TGA1cfhx/ezrRMUFTA0S0ydHzGfC9zpWV6RtYFudKAsbWEyICw3PzhG5YoSwgwvpBWyzvvvu3OOussabh87cpLoMO7nqhM+N5BuwiBzwEKrjN69BjTzBn2/VDjKVOmTBHcrRKCSbsnSnVqH6DQ4fdLEHKY7pGpwQUddEC5A+UNlDC2bFGESf0rL40XHB8Tapvx3yxFj/fee1/yg/Ym3Bo95mfJFQ6wfWeeNH+eeeZf2vPLyblUXuCZ7P835Skal8YLbRy/1Nqq/tZWCPQ+Eqhdprne6t599x132WWXCoY6w6/mPzq164L8YaVUIImuYtqBfRyStAEDPtGDvCsBzXDz3YLQRL6O/MJTi0akgys0GRMnTrD7JRUvHo/W+KOprclUayY1O+usM6QW9qRr1669bd6xN+OlcYLCi9133/3u4ksuNps/ELB06TIaLNnmCT6SxTfeeMM99+wzZkOHXWBZaeBwzM7OURtF3QGK1b5I0ZnGy+6NhW+DBg0UoizdVZCkEknob73kfu/7YREfjoxLOOe47XXqvT3pO/WpF47btpHaPuf2EueFrz/7eivGFb7S+q8ufjMAEdvGQm0nNhjR+5N5CUyVLGCK9qAKI0qdWZ4xWcGEUjzZkO//KANh+RcqVxE9qaDPj/J0TdP+l5vga3TPcv3igT6oOd+XqDjtUBiYOrMUw4/HHxwgxfgX018q/4FR7Jx/wveQaXuBDTzSvs5p0xT4L3yVNriG69niSYu61Lb9F3E2XQhptOEyAbjnu1Y+4lnwWfuXYFrRfO0O/42eKeZ/RvJMg/3F9B/zvz3n/6ImE0J4nAq/2/KOXb329T2PSG0ryTMiAY3xC63h9DUZgQ3rt/BnazRxi8CHuGb9xc4JnOfLNin0ycMJ/N/fMx6hH/iWrZ1UyPijL8wVGYKRyEjyJ7IAx4aN5gTP2tgJ/ACHPvnqBnw78P39beGrdAI+bdjjxvD32fgj+Ap4YRNWAD/MLKYw//738zJJGme2KKCDVxpYK9cZZQyKxUxTPvN88803OFxi1E3PcP379beIS8mu6I2p9yP+TtFuYf/L/tOIQNiCFsdJJ51okYpHKeT0FZde5o6XhsqggZ/bvS0SDphDfSHTPxVVCP8rNqAmGC3sXpYGyhHtDlOezJOEgG0OaWOClk8+/cRddeWViXNvblXIoietlVPh9bIqQbsDZYZJ0gTBYgVFhVXLV7jZiig1ffov1tdV+MlRqq99L6ZIOOpFIQKiW6ZIyESpKqrnKymzryISziLUQEhbU/vmOtKi4WN7l65d7IPPFDn6fe65f8ssqLxr2ryZohw3tvHDggYBTTlp0BzU8EBT3ujU8Rj5jxmhaMrFXEPB/vKrLyXYecf6ws8rfV6TGdZSmZU96np0P9UERmJCRn8ocfwgR79Ei7Kx0jyxRloqoRO+frDUQeDKemnmzJmuU6eOcu78iTvuuOPdndKAwWwNP7YkWAvzv1JBgcrIkibwBuYfX3vHHXecIjXXcqOkQNL3o37mEPjFF1+Sg+inrD48QEX3SUphe3vXvgmsNShIonCiu1wSPEJ7Zcnpz2xJ/ebMnePq1KsrPzIHCGG8g0N7MIHdvAnJVSFDojwJbC4TIVx08SWG5GjAIP2rdUBNbeqLWZjrD97/UIP0qVuiySEMN35n0lWmdauWZi+2atUqjdcWs3H7WRK6JYuzLUoUojAmDoko2j1lypQ1FbTnnn1Ok/29yi02yR0LVJuxvRuS/+naIOF/48/rlhR2P0+cYy94XuYmMOJlL2ksL3v/gmfIU/vkUTMfc9Z97AWZH2iCl639FyLZnEUIFQiMPJglR7tPDcoyvaof4DFR/sXOmb8Z2qBJ4ITytgChmJLVsQL+Ppo61r6VpqUYfjz+Mf7F9Pdr/gOfMMUSz0p+9WumTuIfxmPC3d/gf/Ap4z/GR8V/xGvhk0EAA5/F74PxXy1c7NoWPPBa4akOgSeysCHtCXxjsMYztWGL+a+9B+L3D2/QgJ/2lk3gWgLPKKLRsv+8pJX29/cvoaEZg4JLLGhYQzHM+heNM+1zGugfIS58w0rpButzX9YLgK0w3IW+qYNJ/KaGT+TxBdrfC/Ppr7eFb3wmURFYgf/QvB8A4NOexwnWgF4zelfg+z7QVoBPDnBCvwRjJ/CtvxH8UDCGv+/GP2BUhEp7feDDPxqiJDb1CFMCXW2UtgSBZkxIovcVkXnAuGZNm7qGBx7oXnn1FXNqC+4hrLj88isMPxHogD8IQ9h/5skdR+L9qxZypdWB2Q0CRFxyfCWtjlbao66Rz1RCWRMRGKEJ+ITAaNGiRYn9zSZd4391rRQaDP8FCxcbaIVAmz/9PNoNHjzYHdz4YPeqTILukNABTRn6yN+okSPM7Kd8+QpmutP6kNYuV9Ym8+fPU3jtBQbPfLwYdLR9ipipVTs5BkYowT6HvdYQ9ZNy7JlxT4EvGlx+ENIaixg0ZujXQXJcjO8clDEsClPbQ+Uq5AgT5KC927lzZ3fIYW2kJDFLyg91pbXUWAKbWm7K1ClyIPydaTHNmzfX+MlqaRDdcutt7oP333fXXne9u/Gmm1wpabOENGjQIMkZ9JwaExJWNBIJ2/Pb+Ou8tIRJJ5xwgoUnv/66P1ro8utv+KN7//33FEq8ptVjbUKaP2+BnP92MQEQUbfAg9JlyliE6C5du9kcM7d/lMPka6+91p177nny9fOZRYA2RLFWCvbHc78CatNLyguZh+efFMZ6jBwl5UpIUrSYEFMPnCapIjHI0+tnyKxprVu6bKkefq05QjLGaaIu5xrUr+86duxoyMOk/f3xx9z3QhBs8fLyNsgR0nj32OOPugH9B2hS1ij8VgszSSorJ0dozYDgwDnhxBPdCZJQliwlVUwJABhEnECVkk0d5lDHHttZdoFVLXQZtmto0kyWqpQ5JyqgMdkfm4FZhb/bLu9iC/4vhk5xP42XcIaNgO6D/KIfnXPQtYiX+eEexOX9wvhrXookIzrVM0ala/LtSy/wuKKxRFIb+kd7VtDyfRlfXy95e0n7CgYXQqUJwQ9ES35o1auyBvgQte7Zc+hocLhn2TQSw4/HP8a/mP62y3+MrwTGIm6xveS//KTyH4Qngf/4ZlP5n/EwMSA7Rg0GvsTi0GBGwppQhumBwcF/WedY/h7y31A39Jh3LXkJBsqJrmP+618S8fuHTbJPhjvx+9fTSur6Q6hSoIt06JE1i9Kv8E8MB2GtJUNRf40wl44ZKUtgwsrGJ+WF9Q8ZgbatvF8BefoP5SmTrP8r+BH/gRGxubQmDa7vb1h/+eatg3sAX/wn6v/24NtD/gZ804oIvdvt54/h78r4R8Nrh4L4KauAMjffdLPfTHt0NjxA4IDGRPNmzeUIV9oiuscmn1S9enXzNcMUE4XozDPPdAPleuOF5593Tz75hJn/gd+gan+Fo14gR7eeTphjZyZOmAUVKco7T5odwv1F+vjfWSGdX3r5JUdUo9tuvUWRgWtHED3NBBLs3Lmzd86r1oCDqQ1KB8WkxZZRP90EPyvkf6WU9rz33nuvtE7eMy0WyuIseJT8zqTLIfDSpdm2R14kvy5VFM1onQLvLFY/EOx8Ju0bT516SNH/YW0Pc4e2aWPPbjSom4vlrBdzpNUSgKAIMXr0z1KGUBAXOQQn7PeREjAhnMFUCkUKFDLmzp3r6mWku5O0By9XXpYsGo/iCl7w+N//7gYPHezqKToyfmg2btzg6sgZ79w589zkSVNs/JEJILB69tln3SMPP+IOV59wuNz+iCPUSpTQqNL7gmcl2hRaPDg+ZpxJaVWquEYHHSSljroKCrRS/m2mqH57+chpYX5zDmkjR836x9yi4FG+XAVFlVpmeSiZwAcR3lWQYMuTuMeJZ555xp1zztnCo3Kum4Q2iYnzYAvst2B4PqOhAUIdqI7syAjZhffr5VKHWomzHqkY1VaUpXayZ8OWDifBeQqHTRz0MePGK7T1XIuVjg8ZBDgMNu2BGPMkhftYKmSDBn0udaKP3IcffiRNlzKukuzKiGFOCNCTJBkj5jmx11FHKyZtjHIS0hzd4Sj34F/+IkQ7VAi+UZJDOfVVhKCiul9XocHayvnvwU2bCWGKiqgWunnz5vh+SPgTp70fAebxzBMOcbde3tUQ+IuhkyScmW1zyxeR8EUG5IaemHYTvtj8e/hhEU8DnqGDG5TVtf7AEQjU/gkPbQNi6OiZo7UZ3bcWARQlNhBsHiyFNmkjBT4vbwOoX+uj6tjtqIwJcCKG4MlA92P4fn4YOeY0Hv8IdyMk0oEU49/+TX8prMgjRMrvVjER+Mhu8z8juIiHgWPRtZDNiBGcI48/+G/oA0dfND//2xP4RGIBu0nWZkz/Mf0HTAsIZxiyf9P/rvB/hLOiSKOlAvlhOWP0CW1Koy2lUWiVL+X20Uy7EXiETZeO4TzwDspC3LZ5Yw1GO6pr96OjcQHaMPr3gOBpPBFpR/Bpg7ZJifWZzkN75Ntt+suuOIIf1o4WCUr3Qvn88Gk3wAf/kok2QxthyLcH348BnYjh76vxtxny05ScoL0426jgMHfddad7+823zUkveAiuIxpeow14O+0F8ySw8PPvMQTN0mGypCCSIZONAOeqq3u7G264QQKbs92T/3razICgkjHjxroR0lIBtcKmGqElkYVoxzT9dV2rRi1HX+68/Q734ksvueNPPMl9/FFfExiAmfQJvoD5IMKJcjIN8vRfSNGAxkjjprVZezSRxskwmfQQZXid+r9GgpMTZTb1ztvvuBpyQowiAm5BhqsM/m1QbMDiZJmED4TWzsrMtCjG2RKmDPzsU7McCSaTCDUuvuQS02rZKL8zwMeMCZ+tS3OWWqSohdK6qVi5krkNgf7mSQBVqEght0ztsf/G101m5iyLrHxct+PUgg2hq161uuslJ8Wt5YuWPfeKlbkam83u59GjJWCRQ2WlVRpnFCsQjH2ukNe333Gnsaz3P/zAXXThxVbG5k89A0VW565yF1x4ocuTg2IymA8CB/0iTZ92MumCXjED/+WX6e6wNodZiO8vv/hSyhntXUnJC2j8CQnaFi6cb/OPUAa5ArhgfmjMqkOKA3qI2pJv4HLl3LPPdgvlpwj4+yIFHNq7tiMmWk4PiSQKqRzOwpA2YYdWRIIYfLt06dpVToiOFOOXYKRuuoQrFU1dKEeTPW/uPDfs++FuyJDBbqwcL62RwyFGGYTG/IjITJkzM+Wz5gfFMK/mLr3sClepfEUzP1qpiUFYg9MhoitVqlzZ4K5etdq0YErKyS9esrGly5C2DqpYCICQ+JWWihtS/JwlOQa/38cfS9o3b+/GI66dbwTOOL61u/1KSReF2YMGT/DCGTAaak1gticyqxgxJ84hKs8+KW7kbUSqIkZQ/iXp8ykZ/nkVeitCI9YO7UFstBRewH6hwA3uehgUkdhIV0hUk5308AVB9z18mtZ9K2Ot2il9iOH714kf7nj8DU/AKD8gMf5pHIx2I7rb3+iPHUHgZ/CdbVOhLZ7vMEa7yv9ow5cXnmkzZAl8Uxv+H/jntyIRGirDF+MkeQrtRjx1j+B7LkzLHo5ai3hkjP+Mcvz+id+/u8b/PNkkKDMQ654fjaz1oyZp2wtRk81xG5oNWgM7X//4Z7CGoibgG/SWDbtvywPinJT40LU38NUOHMo/g44RbwFGEr6VsPcs97cPX/nGoOiZT5Tzz++3Rjt/fkDH8PfF+PvQ5GBSwSTMith/HqWP9eOkDNCsWQuTQzC/zPdfH/yrm5U524AVQqtfElGmNluaJRZpGGTT/6KK3vTkU0/KnKWnK6dIw3fcfrs7r2dPw7uV8uHCmw+hCqmU9pfeQsDjEkKXDh07yBzmOjnk3eBG/DDcBDyYBQ3X+cmnnGpKA4avgv/6669Ji2SS+rdF+9VNUmbIc1OnTbGgNWvWrndVK6e5qXKeu2p1rgW+yV250oLcdDrmGAv3DNbjWBifKvhWqVlLQiE5xsX3zSRpp5SR8OIQnX83eIie4w7kK4nEvV6KRNTh6KO9aRd9kEbRoM8HytJlmZ5RUX6lZYvCQ2GZQdWXlUvnjp3MCgarFZ4VkyT82WCOlUgaU/bf1apVd8fL78taCZVKaJwOlInWVEWvmjdvnoRM67SXL6+IUd1cqxYtFUlrtqI5j3Dvvv2ue/mVF91DDz8sjZiqbqvMmdScw+wLf7Jz5TbFtOCiD04I3BAyGVdgnnW2QAKlS+Xr56G/PmwOnW+5+SaLxoQLlvXyX2uCRpVdu3qtmbcZpxEQcIEARV26dHaHS3Nn3vyF7ruvv03wlcTzFdBJwQhm1BkeOlcDgV07mjM4HURFiZjnxeUwCOe/S+UFGoew2Ns1aXqwCXGKS4qHJgvRkXIlYJkkRHz3vXfdO++9o02xpIYSquAolsQCFYFKKfmHOVhqSqVlkpQ5c5b7z3/+Yx6mCb/MCOKwDHs4Qpi99dabFkqMmO7Yw+EQqXiJ4rKDW28Ohgh/tlS2cppLgz9RKmCLJAmLU8GOQI9uLd1tV0hzRqg8cMgEN2pCVvTSNLIxBLeXHFzSiCjlxaq7ftsAltltIYMnSkM8/2O/lKAdvxGANUXJV/X5RmgwZNsdWQEIjz/bOOgocSAt6To0wNFa921YP8lSP1Xekm6bxDuG78dI4xeGJpoin69M5ice/xj/AoJAQ/ztL/SHJ3zeZztKaMygWpwQZCS4mx+n7fE/iiQ2vAxmSoJzkTxf5CQ51lAppe3eXvDfVPgGy4PhVMD8wWAIWEz/Mf+L+f/u8n9PQ3v7a2spo3/gQ/nGDiKaFKFyLyJYz38CRIhYvIIqYWHE+idxG47m10Now5AS96Iy+fnPHsKnrbCL3CF8D3mfwKfR33z+GD7TtMfjH73/aKMgEng3auQo11YaFKPHjnbfDv5GWjJtpVmhPaMSe8/Zs7P8+1OaLbx/t8jj7kZt+sP+k3K1DqjtDlO0oK++/kq+T25006dOc3369HEPP/KI0Q/r//DQRAOuXr2aannNMvbEmNygeYM5EiSCSVF7udX44IP3tY993Z133nmKxlTB3HuwPRktTZLQYMMDGyg09ixXPyPDXIDgUHfh/AVu1qxZMksaYloeKyScQUPnxhtudPfff38kVHESJhzuGmvPzJ4YNyIHNW5k++fvpREELlevUU0OeX8wB8Sb0JIxHC9kzoq7d+9hbRp16/FGy02JCYxE6/VkeTJnzhzVr2GKD2vlg+bTgZ9KK2ij+/LLLyUUyjOzJsYupMB/GqsPd919l7RrMIHKdRdccL77VuGzkQ9UrlLZzZg+0zVTeO5atQ7QCBZyt952i7viiivdOdJW6SO/P9U0tuHD+ogRI+THdq7tv1TUwmEXkcCIcWb8mH/+HaQxyFOfnnzqCTlXflFRpFtKo2icCYbga4n1pyIuVUnzCiY2pbpZREI5BG1Npc2jkvZPP/skFZhghnkEQR5SfHVCbx3SupVXxZKmTLFiXrXpX//6l+vzWh83dsxYk0yhUVNMGjV4eCY+ujlokuCmqKRxyzVZqIIRGx7nvyALghof8lOCGqlNkVZLhWuaiIOJxaEvEtBcqUctkJOjdYrpvlB2f7NmzjCP1T/9/JM5D0YbAgdFGySBXCqVK/pNaKxWrVvbxEFAcSr4EehxXEtpznQ1afSg76Q5M262AYlQ3CO65i9sLgLOG1HZdEcbFW5ARfpLvOiVhWq+SsBnksnqkUdmJESJ7kOonun5+75Z5amsN7OiGeqITKwd4GvpYZRKljKtv1E5CSUpFsNnPKJk4xaPf4x/Mf3l4z8aDs9/AqHkPxpfM9qBxcDnuL9z/mf+HlQKXOOfT1EjlkcT/prfAD+UtjrGz+B5cNKoBZ37arsKP/ri7oGoEbVsTDHmvzaO0cD68WeQ/PiQbfOusYrfP2Fpun+/f31YZk+znhr38ldNBbrH7DskcA/6N/yTtp0do5v51j+YJUHLxhNoLIVP5Fv/REguaB6/fWP52t1D+NaS9gbW1g7hR53fa/jb4J8EQrv2/DH8aMZ1AGeiK2OA4c5O8I8iBYj2NAZ+z8qc5S6+4CL31JNPuRNPOMnddPMtssCoojsKQiOBgAfrKYQ9ZxkpBnDT+yMqpMhFzeVj5BxpjPjIRf944gl3z933uGuu6S03Gx/KlUdtYbXfB3zzzTeRdQZmMTSjj8nC26zMLCkjVFGPVFJ7XQQT18ip7K233WoOa8/vdb5t/umPvcuhOf1bpX0t5kiN5Gi3QcOGcg68Ti9aRR+aMVP72gU61b5Wygf0l2Hu1q2be/75F9yBBzY0rRmcH1922WWKhuzdi1STZcsiRTkyZQXtw8dJQIHZ0+uvvW6aMGH8mzZpasKi+hkNpBjhx2i5fNsQ6rqYFB1q1qhpToWJllRD5kfZCrazQXtvFDQwgVqjvbmlCAm8NpQf6SqyfrnuuussNDhjih8fNFMqScMH1yQjf/xR0ZMPMEEZ40WkKsqUlVDq5JNONi0ja1vPmyezrmigpQSSLa2cavbcNiDR/BPCvLfM0dCQyZKG1DXSXhonC50fR45UwKGLTIbA+JGQFyyTvxwcAjMWwEduMB+zLc2jpejgLwruN0kZe9GmX3RJzUuIgepS44Mbu4fltIfY4JsUa5zoR5skpRo6dJh78YUX5bn5PTdfKkUgTOdOnQzZMH0qWriozItKuHJy1ISX7BUrl5taGPU3K9Z4sZLF3dFHH2XaNpUrpvmNOAxafyxksCNj+JAITpk61c2bM88EPnUk1Sup9lCtwgaONEOhtCZMnCCnP+XMERGmVqhNIbE8qNFBVib+KfgR6I7mjBwC4x/ms8Hj3ahxWfaCtXe9iWk9zIBTzKfZCNvHFc8wE0ShSnyhoazRhxE+L2zo01NXdPAvcdqCQUaZJt0GnCpYllXhh2sBNArlXDAwDTAgwjUjyig/gu9v+ro0GWBEoGL4TEo8/oZnATdi/DOCE1rsf/QHx2KxtcME47DPQSrHwmyX+J9vzbik6kNxAYaRH9e06/8neBQbHA/OGJw1EnDUaHY34QPYngw4dsIUx/RvY2lj4gclpv8Ecng8sUt+wMf4/QsVEZ3NE5OR5d7/REO+br3U9m2MbbgFRze4Z8PP+IcLZRn969qTsG0A6Qhl7F8g8nCM7lkZ3yAs3pLVoWmlPYVPJ9nc/XfgC06+52f957dNMfx9N/+GiEEdwqPLXv2GfTQ+UM888w+m4dKnz6vuqCOPVJCaMa6unOSawEA47HcZW03Y0LZdO7PSIA8tmnJyl/GNoivhKwW8wDrjkUcfcT1l2pShNp544kmXnpFuWL9QznYHfDrAhAC8wCE39rHr122QQ9zyJggqIsKgDZQKXn35VXfTjTean1ZMdzofe4y1A+Hxb9LUyRqDre6tt9+U1Ugp077B+S6zwPt1nQLhjJAgY+q0qTZ8VG7WvJl7RNo8FeWrhmuc96LxU1vWLPOkvIC/V3y1Enlp0YJFbqr2zaVlxjRo0EA3d94cqlgC/jnnnevayhGvNaRnIYLyF/IBg/Pcstqj/6yQ4Ph2WSmtnXnz5puzXyI39evX342X8MMzEN+etWtj7WQqVlqaMFcoOFCaKXPQxnyZNKVJawYrHIumpQpbRXZoOKFFdIxCbqOp8+TTT7p7/nSPyQxgMbAgGwtZ6NSuVdstVdht6JT55xlQupghQZaPWLnVrHjO1XM9/fTT7vFHH3e3yzStvPzTkn4e/ZMbqAhQWOkgX9gkLRreS1jbIKtAYBPGxyoU4I/nMHvZIANBB2vKwe9JJ5+osFy5Fgd9o4QyG2W3VVeekVEBw5kPf8sUjYloTcRNv+POO+2hx6CypXYwR0IqiPnT/LkLJDmrYG0eIedMNUQMaWlVXEZ6A9ewUQOpT1VXbPdDXBOpFuXJ1o0//NFghoJH5dy1q6X2Vd+1liYMdmEgKTZkG+Tcl1Bks7OyJBw60D32t0fNRg4HSni9RpUsTvtuBHoc38rdIbMmCGnQ4Ilu5NhMpt5wCFwihSO5JojhvggMPIMhkvildAgLGxa6gVH50pQUSRqO+nrWtipGoKyhAC95z1pWXc/0rCxUr2tz+MttJYjdC2rQrvXt06lUeDH8ePxT8SGJYx6HQGTLA59AtAh/7CTGPxuT/2/0B/sIcx5hQb4DfCXSJ94N/sdmlmqeZ0XcyLebYHa6DPjF0SpQHs6W7FOyb7vDfz183xK/oQcx/TOeYTySNO6nhmEK4528x8zwF79/GBND3/3w/WtYY0Qa4cpeHtJrV7YW3vhohNwLbPStpeJftIM1/hPwT3l+/H3xxDpHl5QLgoqw/qIacxbwnVqp17Tl5zR6sN2Ezw7NTD33Cr7vA33LR3//teeP4YMDOx1/duIFlSJYtaXRwv6yrPygTJ8+3fXu3VsmTLPNPKdnr562/0QWCl5PmTLNIvSuW78WlI7wpLA288tdrdq1KKS9q3+G/gP6uzPPOkvWFzPdUVIeIEoS0oCc7BzbtxhtqCgaFyWkYIASATXZu1RSSO1rrrnGfKPiRLfvxx+5iRMmun4fD3C3S4umStU07Xn0bpVp1XL5scnbsMl9+fWXUm6Qs1rlo8Exe7YC10ggslARlBAGTZs+LSwfbM9dQf5cW7RsYX3BlUea/LB2ObaLIh0vkUBnrcx20szMifFZLWHPfAlW/v74P12mnsc/vA6i01PkB+fUU053FSpVMPrmGTAjmiztmVwJhnDc27hpY9tjj5JgI3dVrkVp+uTTT02IwzDSHvU8U+FEBjAyv0qTSxMSAhn26TnZyyTwKS3lDM+NCmtiDjqosfvDmWcZ/Q/6fJAF97nq6qvke+YVc6BsFg2RQO+td95248eOtzYNloAiDCbYUGk5VqYP4CAOjR966CGHgKZTp07uz/ff5zLS0zW2hVy2xofP/8zXFvnYwamyxDvyqyPcVAO0sS9SAWG+Fg/qXVlJE0+Sl+mJ4ya6F6VChcYM5kQIYUqXLunS9bDn97rAXS+1LZj5SuXjNAgPyCMlIWRQD5Bwp54kj0gUibvOu7hp06buvPN7mTZLw4YN3eLsxRqwbDNratG8ucVf76eITU9IrQyfMSWKlZBn6KJSPJIWjezM0Iqpo5Bk1N0sh0Hj5UEbBMxoUN/Mpg6U3VmtmjVMQorHZez+4rRvR8A0ZySc4eU9UMKZUeMyBVBXCF9sAebhg1d2HTFAY+YsFKLuhfscg4DGcyT4YhK9acMWA1ZTSKX/nqzsROf+COgg+KEEZezrncEXXB1NUGTFQx1/DAKaqKkYfjz+HoX0G+NfTH+p/MdzjAR6/OqE+/wF/ibGY2V2xv/CBinwHyub0kY++PDDCEayL3vHf5Pw1bIajflv/P4xpAXP4vfvHq8/POWHkdy747//cq5Lr13JLcpe6f7z8XCtzyPhTMQ0mCf+kon1j65T+E90mihi6yNdJenf3/oV/Yvj+JXbtjAo72HuCnwTPHsQ9rtn8FMa8K1EbW3bt331/DH8/COQf/650kzkL7JXV76tYgrZPFZKAK3kaoMctC7QoOnT5xX3tFxt/EVRfG0TLpzH4e4nn3xqoaLNF7/KI1jhwz/Rf8FVBDP4MTlSEX5mzZrlbrrpJjdNAp17pcXx1htvyT9KregdLjMmlf1+2A/yZ9LA2gE+9FBW+9Ml8mvamj6pTWD8/R//MF+rdygi0UuK3lS6XNmEhhlCAXzLEJIaf6uE0TZtGQlIcJ6L6c+Pw0e4vh/29QsIG7et5u/1iHZHqO+lDE4t7bURIOEUeIue9VqZFBEJikjKaKqgUMHzY76Eg2F7ELXVuk1r16tnL9O+2WKmTVvdAsFlL0/Z+XPnm+Nj/OHMluPeCpH/WMJo8/yB1qNuBcq3S37KSI6AgKqEnotAP4yJjb8YIeZJtQ6oJW2cYia8eU3+fQhfXVf7+ffff9/MtzQ91tU5ElbNkBsT+GeCZ0m4s0nRm3hOhGXwqIrSIqohfzVffPGFnDqfa5Y1P8mPTo8zeli5AB98XCcfOiaUkraMOQpO9LpgT5Irh71oV89nabU8Xy+XzxYELhMmT5TjpI1u2jScI70m+7UJ5kOGkFotFPKLycEGj/jvixYtlOdmhahWO5UVggstFpwGz5z5ixw2jbSwV0RPwuYMnzNIE996802XJY2XRTgzkq0djn9eFgJPnjRZA3qGO/m0U0zzhfjreG1mGYpD4M1SR1qSvVQEU1NmUUdbHPbPpa60Vl6uUU3CETCEEad9PwI9jpPmzJXSnBHleOFMlgGFgUEMJM+8EjzB50szJdy3ylZIjagORGg17ZKzZFnq8C+ltRSKVbYlv4H0X3+sJeX65QS3rQ3asf4JiPXVigA+mWL40RiFsYrGLh5/jyyGKUKYxBsjoE6Mf9D//3f6E4eLGFWY9/xH+BT+JWAp8DT+SMZ3dsD/uBfKceJ5VP6jZ1G+VGg72bqHsaf8N8A3GHJuHOYwtR/JJ7KnifE/pn+PdInfmP/l53+eYhPDs5cnFcuXcs/ef46EM5X1gTPXvdH3h0g4k+Ac4jtwp6D95gFGH6E9T1GXAk2H9U8qzwhdDGVSj/noP9+j7Sb8hHYya0VBpM/5FmDJPoZ8jr8N3+Pf7j5/DL9gxx8c8v6VAjbt5TH6SIhQhYhGfJgnoamRoz3kvX++3112yaUS0pxhigD4PeUdSjRfwx80VpQWSEulcqXK+sDvwymrASkk+IjDKAmw0R8t06heF/Sy6EcfSWGAaFCbo3f2ZJkjESKafSbt8pEX9x0ZDRqY5oaw1hWTWw98qX7ySX/3wAMPuI4dO7ounTsrklQzCRUkSBW+IwA56qij3EGyFiE0dYaEImjPYGkyW8KmElKEGDlqpILfvGXlw8qgTPmyrq2EM+x/ESxUlubMPX/6k2t3RHs3R/tp/O1AywispDthUYw4//ujj5oVStSYaeEcImuV5orqhKkP+I+QKL1euiIjp7nZMoNq0rSJLFzSJLBpYPtrFCM+7v+xmzx5io0lP5lZmRau3Ew2LXerCcYIdT1VY4XFC/1BUMGxqMYb5Qn6TdqivKFDh9kYjxs71r0pucCRRx3tikorir7iS8bYjOaF+sgkihX3Dp/D+NeWgOqCiy62KM1EgHriiX9KZtHHvfji83LJ8rDBCtHbhg0bavNv7Ro/sW4U+E+BCGZ4YNKSpdlu0BfeNq1kMQlSKlfVpG2SSthk01LJyMiQbVYRCVxmSCq1xpz14JNmtpwh4ex34eKFQgYhuqSDpUqWlgPfhe6b774zp0rZi3M0oMVN3QtnQqhxoYY1SQIfvCqDJE2aNZWnadneSajTuVNn17hRY/f444+5wWoDiSRIvUU2fmg2IDlF6rleAiFCdC+XnV3Dgxq6Dh06uvIVyhX4QMcNbn8Eundr5W6XzxnSoCETLZR28hXt60BYZucLIxOumWaC0XwPAABAAElEQVSKjvzzxKX74aWrFYQxdKNGX9/UDdWob5f7Pp8T6pE88Xri59qLYkJBb6ZkZYFvRG6F6Jjvjxqwl75VsQuasRTD16D4/xqPePxj/IsIY7+mP/GeaBi2dzD+ExXgsCv8L+JA1lzE1uwc550JnDP683wNXkrboSdc25/VSv7sLnzjv6zVoiZi/hfzv5j/B3rYvfefSicJsYDOKlcsI+HM2RLOpLlFOasj4YwcZ0b078Hkh4uQNayV2EzCJ+yJtOGEvhPrr+30MayXfB0DE4GgXqigk12EL/AGOhU+DXmhS2gveaR/dHdn8P3ajXJhS7R7zx/DL+Dxx56IvwJKhjNqa63MbQhGgwADnGW+y8pJ7a233OL6f9pfUYhOMLT6+afRMtk51UybrBdRX4ga3EIuMYisZO9K4T8aM1kK1dxYjmUVN1iymsLu3XfedWefdbZbvGiRe1vCEXyzUJ7w0n0//MAc+Qb47JHxPVO0eFGjB0bytFNPl4CiqPl+wTkulh7PPPusOR7GvCc4Kj5QQo/jjutm2jNtDjtUkYqqWn8WzFtgPlswrYKFLJMgCGGOvpYoylQR16xpc7NOYXjRuDm/Vy/XrGVzC5aDv5vy5SpoXzxUToUXuWHDhklQVNL9RyGpUYaAllhrkHqcdro7+eST5Ee2jJQe5HOl0BZ37LGdXeuWrWW14iMur5Sj4Fq1aphj4Ww5G14lkyuUIHAWzP57usyuUM7wqZBFi9ooa5uJEya5IRKEENiHwD+2DtHv6txVkgNUsufR95+oP86dJc2ZBx6433U6poN74bkX5Cu2sTVJd4tECJAprSYCE20QXM/PCimUeAU3dMgQd0AdORlWe7mCd8+f7nVXXnGV63X++a5v377y0VPJ2sKa5m1Feg78MAisfd8L7jdwob1ukQfCedJyeXxGa6aCnA11Oa6rO/Gkk2Tb1tKVK1fRqwUJdadMmSqbtElywjNDZkOLFCq7goQhFdzqVauFAMVcenqGhDNlZE+33OzfZs/JkkOhFSZZTE9PdwsWLlRYr4km1cOmbbVijjeUls0f5K0ZD9KoYZnKmUYN27dvJZj5/POv3A/DUd1ca6pjRYoadklquNmiM20SYVSpVNUdobBiNWrU3OvxiBvY9RHoLs2Z26/sZrQ+8LvxbmTkEJgW/CzpGBYDwjP/JdZvQ+y+8nxZXvR6uZp8NWRG2w7ls8iBOYo3KelH/ynNiR3JApFJAbCda/Eh8TF1rZoYkBWwssnFRWjDv9xDOzF8G4l4/D2WxfgX0x8EgQAZ3rLT5LkTRXaV/1HD+FTULue26RD9WTtcG99Sro7kek7l88mz7sHrlEIfdwe+jBo8e6Sywdcxpv+Y/iPcjN+/UJaoTP89he94/WGrh32w+vfCmbNchpk15br/yOfMunV5yfUPXUxJifWPepxYI/kH4NdS4FZhneQzeQJ4jZgBz8vBKvCj54ZH2LW/kWzbaid+8sG33Kgi1aJS+eFHuRH/+S34Vhfe5zuXgBtO8sFPlBGMGP6+Gf8CxvktkS0S5kYNGzVy6zestbkG/4rL9QU+WVo1b2VKBJNkdfHRx33da6+9Zj5MMF0JWJYrhYJvFNIZny4kIgmx31glYUHJEtLEsPcsuOnchEkT3Ok9urtXX33VtT/iSPfUU09pX1xBwWkUTUnPB64Bf+WKVW78hPGuUcNGypNwQ2jVT5olhNY+WD5VCIKDJgeaNDcritQtt94qoUptb0qj+q1bt3Fnn32O/OEsgqIs2M4CmQKtXL3K9sK49xg48AuzTsHHiu2fBJ++WIrWBkRTPvfcczQ2G+ReJM+0iSqqzEQ9B/1Cs+eVV/rImW8/VfPPyOEIaducc865rrb6hACJe0Rn6nbCcW61BGE4FkYmNHHiRAmUJITSAw4dNlhmTd+7t99+S8KqzXI0PNcCB/HwF1xwoYXzhuZwAjxVc4OwCwFrngQ2haU1U0yCNTR+GL/Spcq4nhIsIWz75z+fcO+/94HCYLdwb737tjv+hOPtebfKXxF8CefGS5ErrJfT5Gj8UdKYlZnpalavac9J/9euW2MKIe3lm7aUtI969jzPdenSxQRUNn96SuMV0TDYOBbgT4EIZmC8sN8K8mbcUJGWUFmqKIlW1bTKcs7bxnXv3t117HS0CUFwHIYwZI0cDhEmDJu4bhLg3CKEu+jiC00tDLs1vEXPnztPjohWm+Rw9Jgxpv6FZA5E+0Ex14dIykU+ocLyhExEcyolaeIaCWYIi7ZeZkyHtjnUQmCXKVvKVNbwiE0c+MKSRpLWCHG2yLzJnCtJ2odHaFSy4vTfHYEeitZ06+XHGpMfNHiCOQSGSP2bW33hnAMljKA8RZBreTpCeBCq9/avE/0PX5qVy6VvUke7mThwh3uhJbtIFmZTEcFXpv0HTmjD14oWGbofw48EV/H4x/gnfu9pJ6a/fPzHFkOep8FJtk2U5Z26O/yPsvCiwKsMnhYvNEM+17oZgbKrZPtWyN+zepTjzzO6RD3aId82XzRnf/wqJeB7+vfwQ3/i+ffjqmHyg2W//sLnxO8fwy4/Lgn80+V+/P4leiVOKPdFMuHMAzJrquPNml7/8PsUnzOMe5L+A08xijf6T/YozFoQjNj6J3RZ/MfXpS3fpK/pCxh7icomYURt7wg+t0P7Ot0+fF+Adf3uwI84me/AduHDT8NaLypmXJAueZj++WP4jM6ejj+jV5CymUJSrQDX2DuW1R5y5szMCC8QrhRyE8ZPNCe7zCJ7STRobrrhRndV76vcwE8Hmd8Tex4JKH6W/5Fcue0gITzIkxCDjbRpjgkX/NFJM6O17Ulvve02Ob79XNYYR5twBBMkUAu8pP7yFctkApThipcoZvk8+wZFFaqaVtWczFZNq6ZgOsuV67Hz+OOOlz+cp+UkV2G+DUe3aM/dwLVVtKUGDeqbw9pK2nvPkHPj6dOmm3Pen34eaT5UPx/0uZsr0yQzP1KLlsJA68j+t/fVV7vDDj3UrVEU47POPktWJltdjvzDoFmUk5MtE6tPJLxZ77I1liHVrVNbGiZXuKayWuEhWN7UkOuRJgcfbMF3FiycZ46Bu3Tp6qZMnuSmTf3FgveggbJy1QpZtGRLEWOpRUvG4e7jjz3mTj35FHN/smHtBn2Y8lq/aNY0aCAfPQrgY1or6kC1GtVsz58u37TwghkzfnEnnHSimzs7y30y4BMLj+20t2f+Can9/nvvyXrG8wXWMRvy1ts6HafMyCdMw07tNsio72ZK4+jUk0+zfr377jvuT3+623zr8NwJBQHmoIBTgQhmjCHpAYuUKOoaNmjoLjz/AqmEHS9tmGyLcoSjHuzykHDxR9x0pF4MALZaON9BMnXfffe5RnLE+/PPo+VsaLWrUq2Khe7CDhB1orfefsditlNniwYWUyZUtebMmWuen3Ms2lOuy5Oa0sY82QaKh+I4uOd5Pd1x3Y4z79WEF6M9Zmm9kJ+Y8mjffP/995JcevU0m8ECHui4ud8egTOOb+1uv6Kr5qawfM5McKPGZxnjgfmA+mEzYE5+I2KAiaGhAg5SCkEMVySKJL70JvK2ISJr2Ip7rRhOhRv6sf8ATn2xk0mbqbSIqi75IS+8oGP4GpWg6RSPP5gV4ZKd+p8Y/zydaTRYpBjl7gf0Z1zInjMFF1JOtWyA9Ri+7DL/ixiQ536qS3uC4cH4XBYT4T63PW9jCuBr0XIg9EtHg606hqZR/g75bwK+IOg/lzH9M/6MuB/1MN6MvSUbWH+6P+F/9PQeSf3Q7Ff0v6vPz56pQH1tJAD7k0oVSptZU4Z8zmQvW+1ex+eMor5YAm+h/YiuQ1VP/+GKDannG55LgOkB11VGbYR1GzW2bSu1afDfUsRnqLs9+FYGuonSzuADe3fgW893Cl/tqZ/5+r2T54/h79n4I9oPkbfCPO/NUb2w6uAfflfGjVe0Hp2DO7wTEcbg14X7+m8Rhl5/43XX8eiOroYCw6DF0UG+XhImUGrP4/JWUzKoVq2620REYJnosK8FXIOGB7rTTjvNIE+Qy43LL79ccPKkJfK2u/KqqxKPgw/UyTKRWiGrENpkP1NN7jhWrVbYaSkn1G9Q373y8it6B6vR6H1Rt05dd7IC7eC6QxTo29JaG0WEdoe1NVMnyh977LFu4GcDFYV4nasrvzlER3ruuecs4A6V1qxeYyZI0WLB2sGUiQhTd959p/vm62/MUTLKFjNnZbqjjzzahBu/SODz3DPPuknSgmG8jFb13GeecaarrChT7LE3K4oUA9GpYyd3rjRqUJwgAhUOfbOkAZSVmSlrmfJmyfLTqFHWLpGR//3v590z6uO111/rbr7pZlnTlLdx4XlmzZylKM0nmXUNgBlqfPwQQjtD42Tjr3KLlyx2p5/ewz0q3ziECz9cfnVwZpwvWf1C8jmbLflEJeMTfk4ly5BiSIdOHWS21sLaGjp0qLtJyiO9r+7t8BuEdlBwfJzKX/K1vxcX0YzuRQtUZfyVSkmVq5w8TJeU1+dqkpYh3WIgGSRim+cpdDZejXl4fM2AgERtGjVylBs/cYIJTkrKlm2h/MfgSPjCiy5xl192mTtcobKJroR9IG1tkNQLoc0B8iZN+GwiMCEJI/oT5k+ocBHvvLT6gXpZmqSHxG+fL7s7QoshHCqqOkVkt7ZJqlVEePpc8dhHKDIUMcxxMBSn32cEcAh82+WdjUgGDpmkaE1ZxuhAfoQu2yaYAvhk/7iAUk1wAtNVabJSKm1LRFaXelY0OiLRUwomSeHl6uFTJtlian0PzL+IyI/ha6ji8U/BFnAqiTvgWCr+hIUtaqOkGP94PfHyjWjK6P//C/3pOXiUHSSema/l46bM3WX+R1OGX+CY+I+Nm7UPznnsCl/zLJtceJ/9pzO+Q6Y9qitrS0f7gkR/foP/qqivA+Mz+MpQk6kYH9qkLCnGf80Lc8BYhPGP6d/GI+Z/nv957LAh2Sc/aM48c/9Zrp6EM0vkEPg/Hw4zzRlPqyn81/A0f2+MfhN8AWLfZv2jHu+Q/lXc079/rAQdaH1O2hH81B78HvB9b/1vDF9cax/MP6+whMAhdcD38Lx06dK8kiytUNShdRIAGJsViSEsYc9q+GdlvHZFS/lJ4V2I811MiT6ScsAdd9xubje80EhrfO1jN6o+Iai3SsBiOGvvP9r1AWWKFsV3zBY3dsw4i/SDxsm/n33OXX7p5Ra6m/Eb8Okn7scRP9r7EvkkQW5w0YEWDL5hWrZu6VZob7tUygeJJKltCe2XMTcyZQN75xZyx8l85zxFFyovDRcECDj0Zd/N/rsMFiXaR8+WT5yvv/za9r+YUq2X1ogBj8aI4T/66A7u0ksusf03EZWwRPnmu29Me2ao/M7gi3XQ59LAmTvH2qc+z19Ce280d4BlrzQ1duSRR7obb7zJoiRVrFjRNZPLkZIlStn+f9XKVa5ylTQzZ0KAtlZWMn/+873ujDP+YFGWMAEj4jPRspAXPP74313OMrR16Gwht3njJlPEIAQ2CQVD1iw899133e0uvfQSOSuu6l7r87oJiayQfmz+VZbQ4N0VgQlFDeaC8Ud2gFZNBwUIAi+Yg9fkEJhQ6LhNGT9xvEzNaqgVRqrgE5x/71PUt5Jy2IskD5u0X6bPMGe9eXK4u1aDCTEQ/nqoHO2inbJR+cXl7IiJmCTVpldfecWkY5g2IWxhEmpL8NJOQpk//OEPirNeVYi5wrxEF5W9WQnZk1VOSzMVrgPltBfCgJCRmOVJzal48RKutBATgisjfzVMLAij5aUhMb5siknDBwfA5SuWd6VLlrLJLSFpIdGb4vT7jUCP41pLOCOHwCIIHAIjnAkEDqohWAuJL7LQhmfkwibRKszUNiCqzz++PidT8pw6/islrVKRUmAk17oU5frFAefAIF8+agy+b8c7IrZsX51mVC+GL0qLx184EeNfTH+eV8BTIiajo3gJzGoH6ZqeHXR7q/vkm/ESzszz1aii8jvkfzQHzRnf4qsu2EdVLZispj8HpHE5NWb8j0aVVDzK96WpH/ifNbSr8I0/xvQf8z8wL+Z/e8P/WNgE2vVUum9+zazpPjkErpOmaE2r3X/6DpefhTyB91sE4w351j++H7b+oYh4g/fjohm39Y/4T9TV7T+/nkr1UvlPeLLAczz/+TV8Y1RR4T2Hry7vIfxfrf/26Plj+DsdfxP2BAwKmLHnR4Qjz0oLo35GeqIRbbeNttauWyuXG4fYphyIW9D0EPwqEhZ06dZVQWjmK5Tyl+7ee+51t9x2q5nC1JCGDOZAkARCmjflEHbxEgkLVM/IRC1v1nsQ3ynsO6EfuTmRP5U817t3b3fd9dcpElJb9+jfHnWtD2ltygBL5BjXzGNkBxS0MVbgZFh9wncre+KhQ4e4r776KnoG7rCfVchnCTuKak/siW6ra1C/oXvu+edN4aHr8d1MGyhLAhSiG7MXJ2DOLzOnu9dfe932y2vlWoQ/A8Yg0GElQmpjOpSenmFhpAmg07FDBzdDfmNwR8L5h3Jm/ORTT6o0g0GtrSZ0aSAtm/JyrIurEFLd9HoKJ36jq1cv3dWW6VOz5k3lNqSUmSUR/CdrVpbtzYsW0f5b4zhgQH/t/9uZyxOC+/TorvDV2s/PUwRmhGG+izKzWpatMS6hcZeDYM0JW8Jycl58Xs+eJgP4oO+HbvDgISZ/GDCgn7vk4kusP8w/GlPM/1i5REFAE8afpdkGCdYKSyEERQ0EcCR8EbVr29Z9JXxAcBQ9sN0ryB9Yyt4nm0iaEVJqVJDAoZmSppBWSLE2SZtl3br1Cj2WJ6/PS0w9CqdABzdtZiHFSiiCE0hBiO3snCUWJx3zpjEKO7ZgwUILd12kqB9AQm3VlpOiBgc2MMlfnhwHFStSXBo4RTV4hUxAg/8YBD/MXAmpp5UrU9a8KleXLVrrQw61EFzLhZybbBIL6X55SdequlVrcs1DNM6P4vT7jgCaM5g18XL+bPBE9+PYWcYzIP6gusoLHJtAYwZiZsYQI94ASnIf1oWQxJIO4SucLRipqkUEyS8GqJWacBTsF5f0I6TQHs16Qhbe63ZYQFGSvxh+PP4x/sX0F/gFzCjwHziEbWQCU9nmeH73tq53r46mDtzvq7FuzOTZxsvgMjvmf3x2IHFMClXgQz4/gq8L/CDA0rzwQHQaeKTV9PyLlrhvLBZhDxm7BN8KGt+N8T/G/+3hf/z+TdKVUZXRX6BSTz9Gx9BouNzHx+AQ2HzO5Kwy4cyatd4hcFgnJelf/EX/TAarJRSBPwIL4U4q/wmm5f45teUwvoMTUs9faJN/20us74xHReu0ZNldge9b3D78MP6M7u7BT/Bf1dv588fwGYE9Gv+ATH4I9/oXbZGjjjrSvfHm267zMZ2tPeQvCAbz1m+0j/LrN7LvI1MHwS+hj/vDhg5z5aSRsn7DOpkAPeu6HNvVtWnTxo2bMM7CVzsJA6iSk52jzf935lsFAUF47+F2o5g29rxryWvZSiGmmzW3CEuPSfMDE6G33njbnd+zl0IHiW54z+o4VwoKKClgOlVW5j6NDz7I/fTTT+6774bIooQIRjK/0r4aKjCgar+CypWUo1qDpGycBd97zz2uhfbZl1xyqfl4mSurEfbaBNBp1foQczWSOXO2+V9dJ60ZtEbW6VkTxGyQnLvgwvMVoaiXhBcLzdnvgY0ONGUH2ps6ZZrLkxB33LixFtEowOd568l8qGZ1aZZEz8+9dAloDtEYztHzEaSncsU0lzkrU6aUORY5efbc2Rag51D5uflp1E/SMjrDBFxPPPmEBFmPaS9f0T62m4BUY700e5mcJCsilkzJ8G3DmKBFhFbUoYLDBy6ESKeddqp75dXX3PMSWD366GPyR1vWr3tU6eOPP3bzFs73T6v5z1P0LLgCLlEwa+LzAs9zYMNGNtdXXHWFnBcP8+Ov/IJOoOBeJ5iw7Ux1BLFQB2LQkFIW1h/MFbOmH77/wVWViVPt2geYwKRalapSlzpaTowqu/SMdNnDdbF6G0VECHiwuyOU9UR5rIYYkHaWQQtGUr+qVaubI6fs7EWKS7/GPCcjOYSekSyukRrYWk3MRqkgZYto8kR0paQ6dbCcESGpI+b5p3IMtEjOh5DconFDZKbpcpiUK9WqOP3+I+CFM900p1vd50O9WRO4ZvgWdY/5Bv0sV0fOwkIQZggj8C92K2QbEvIoQ1VSqGvFfZblcepf2fpFDBslD59r3wZwkn2K4cfj73Elxj8IJqa/7fEf4x6pDCfiLamHCyScubZXB2VJc+bbCWbWZLxK1yHl43+Wr2WJZaqWR0B/TDnnNgvSZFKruu//MWM+ed6nssrArwIpWSoqQ+GI/3GPC+jfzqyb8fxvb/7j949/d4JFhlOGQ1z55HHJv2END/fb96/GgK8+/6XkhTNny6ypklss4cwbHw3XpnRTyvrHcwDjEvAF/fmc0EfRf8RBEvxHfQ/n9iFLdUxTJvFMUQsRjyI7Mf8GQxiQmH/BUZ4yfBd2Ct8DoGcFCd/3jY7pfwx/J/O/F+NfwDjP/u90hXbGJGfAJ/3dxRdfLD+lXpODTTcaL0RjMl4tiQ1aMHzoz87OlgCitL1TuTd6zGgJJo5wI0f+5F588UV3wgndzH0HT2qCEu1//ftvq2mplJBiAh9jTLApRMTXae9rrjaFhUnybTp48GBTWHjx5ZfcIw/9zQ+YBJ3ZcrKLA2BcgKyXRgdBbyZNmqjQ1CebWc1iud8g2tOIET8YHob9twWuMYRPjn05RT1uLUfEpeXYF2e77Y88wi2WYOaXqdMsZHZ2zmLtodfK58tsa5tQ21ixWDJy80LQNm0OcS++9KJpyNSsdYBr0aKFfLz+4urWq6tgPxWkVfSF++abr92ggYMg4FBdTpWr+T4qx5P4VnPAfOpJp1ikZoReNWvV1L69iJ5xksuWb9rCUsTo2rWrmVHNmzdHfmAnuBtv+KO75ppr5Ej5C5km1TBZQSE59SV0OHv6XEXGQjgD6sBfFixcoBDYdQRV/EN0Sp/++Mfr3eVXXOZ69OjuLr7wIhOMhc4ihEFugZAZ0y7cobCWItQ2pmK00Uqhts9WWO4VUj75efTPalm4ksK37KEL4Cd1dbbHzfmFIAswqWpJC2WdVID0Tc4eBGkVc4SN1kIhA75nLpXfmBv++Ed3zDHHmKCmWNFi5gMGqeFKabK80udVCWMmmtRqgzRtVq5cbU54Tpcq0ymnnCzbu9Vm04Y0MW99nhB7o0J/VTITKOzCVsrXTF7eOsHESfAmN3bcOPfjjz+axHD+/Hka/K3mufrbb791v0gli75hg4iT4s0qjzOmOP1vjECP41pKc6abUfRAzJrkEDikYNsKwdjijRu6MCmyYV1UUnmUIVEOQgRXfZbP4V74EsJ5MpcmVdKIL1TyNb1Ka3Q/ai2GH49/jH+exqAjSzH9/Yr/GH/ZhcUnmjPX9OpkKrr9vxrnxk+VWVOUfs3/1Coblih5LhXxJ12Ea97HLCboQ8jjGPgf930+Z1zwbvfv8SD0IXtH8K1lni3ROP2iBlmcxPw3fv8E9EjiYMC/JJ5wBvoJZ/bX96/IaGeadX6ECvYX4cxz95+jUNr4nEFzhlDa0kAnBZrWaf71j93VfW1CU8pEuX4OU6p7/hPdjdhMqrA4YAVN+fmPykZtU/+/AR+o24NPfgxfEgSl1HdCgc3/9pDIoO3ZD32co2hE3XucLs2Xf7vnnn1WpihnJCLs0OoCmbKAXpjesTk3rQnhZpFCMq2BDnUvQ9oeaLkg2EEwckaPM+yIQ1vq2tpPsDhHUNCwfn0zaQJfSRXkNgNFASw56BPCHHzOoLTAnvivDz4oDZ3yql/ItEYy5aN1dtZcRTKaIke/Zd2UqZPd1GlTLfT0Srn2GDJ4qAmbNAtq3f/Z0EVrC3JICBmuuOJy16TpwRZNidDPaMfgbwYNnGryaTNSe2QEO5slXJrxywwJn0bK9ArLk9BKIZkh1XH33nuvBdVBywWzI4RHY8eOtxDiCxSye4IESD+P/kn+cHKsR74H/hdW7vuJmVQthdk+xwRGWNgcfdTR0lDJU5SrjbKy2WQRrFq1aKl1hipp+Pp+9LG0ntpbmO/hw4e7gxodpD07LkycQpH/S6HOp6ppBLZosikst/bypfU8jHzAURQ7+vR5TfPWw4Qtt99+u/zxnGdtWEHmSYKYrRLOlZYwDJmAZD16DvVBbSN02ySfNryPcJtid/xD+TYK6De5iiuABpGG52my58ix0Bp5gcYvTCkJPIoV81ozICcv3/JSM2rRspXr1KmTojG1d00k7UI4gjRvvfzDjB83XipkQxUbPtel18+QdkyaRWPCYW+tmrXMye+SRYvVRgvFK28lZK8oBFttPm02a1IXSEVq2vRf3NKcZRpLr0WzUlowq1evcV98/oUbNGiQHPg0cqhjESGqqFSVSkmyyR+TwmDH6X9nBLpbKO2uRpyDvpNwZmymdY6vKBAchMdN/hk/EmNJnUM/m9FvorxnwNQ0iXZUPzDQ0J61ihaYAVEbHP2FSdRhAiFR1u7F8BnVMCzRWTz+DEgSX2P8YwHkxwR/BNG/iLb+v9NfYCc2AL/xc/7ph7nePTsYa+n/5Tg3bvJcP27b4X/wr8DD/Oj6xllYcB3wz18bx0rkhbqhLOWZDmsvNBYyaWsH8LWSUqWIV1JOf4nfmP9GWB/T//5M/1DErjy/Ec8uCHBpryATwpl/ySFwXQlnFuesdK9/qGhNMvkw5E2sf9CNSb7njU/oo6cxjZTOpCyRPK+J6iee39bcns943uXbhDv55pI8DfjcZUi81k3BwAdW4H+BV/m81H4JcAr/i+EX3Pz/avx5jyWnNgWb9vRUmCpE3Cwtmdtuu81ddvkVrnmL5vJ5cpM7vP0RiUY32wuPed5qPkmqSJHAbyp8kQYNGrrmMg1asXSZe+mll9yTTz7lTjjxBGl0jFf0pAYqisaYR5PvvvnOoh95TRo1I/h9P+zr0tPT5ddkvSxF0jw9qAKBZz7u1881bd5MDmc7uCPbt7fnR8lhSbasOlS3SeMmjohJBzVq5IZ8N9gVL1nCNVfUoOlSLsDMKaSszFmyFpE5EqQYMnWsqIhIjzz0iGmbAL+ONF0WLFjgJk+bbMd5EkytkXYOggcUJBbIymXYD9+b9o41Y40VcpXlouRAxqFlcwu2g7UJ1iq4B9ksQU5JCZ2+l6DpL/c/6PuR0od8HdL8ohBxwsknuiZNmlgI8maaE0zIGP9FCvRTSO3yFDw/fz/8MNydfNJJbvC337nTu5/urr3ueldOJlxLl+bIKmeROWC2udNeH+fCpqGnFtjjUb9Zk2am8TRu/AT3j3/+w6I5vSFtqfv+fJ8JfDahMRPxtE8HfmZ92LgJX1v+4xRyinnyG0QZcDbf86Q+516e89QFkzSQW6ShslGTSoxxPCVnSKjSonlz11QemDPqp5t0kknkkfSckrYVtahJlcpXlNrWBvfC8y+aUIaHxwEvNnIgATZpjz3+mOzC/iYE/tFVUh5RmiqnVTbHSY0VYhuP2V9+8bkhKQIdojqtlOMkHPegWQNyT5s2zbxgr5JGTQPFfSeEFxPSUxKztMppglfcJGIFMyBxKwU5Amccr2hNV8ghsPBmkMyafho/279IRUgkeAbEYiIZkEs5RjicCTd9Uo42DUg9SUaAusd9znnZR7eiu5ZrRBlUabnPC8OO1GXlYMm3Ya1aI0D392L4yTGKxz/Gv5j+dv+lfkGPdtKc6WCmmP2/Hm+aM7aRSOF/ESMyfhXOOVq5FF4UVK2NC4o0A38ybifeFfiWtWG8TnessHI46iIsVEIEp0Dh1OEexUI7oX3Liflv/P6JkMmwJH7/7nT9Ac0E0oO2/psprWJZ99wDZ7uMOtKcWUq0Jm/WZCsddQq6xldVSNA9nY2mN0n/iTWSr4OJCMnPv88L19aGaiavfXs+n3P4k47auRQMfHGpsP6jR9b5XYBfYM8fw9/e+DMPBSqPVGMI0jAJwpntm2++4Z7/9/O2fu//Ub9Ia8IwUviAtUchN12mPmeeeab5PaGPJILDjJQQpF5GuuH3lClT3GEKUT116hRZZYyQRk4P+9APhm+QT9W333rLTJIwt6ENjuxD69WrpxDOh9OkW5u7xi2SpgkRjL/5+mvTbrnxxhvdedqf0g8c7naQk9227Q6TFskA+alppf1rfVNIyJQQJnPWLFc1rYp8vUwRUWBStdE0gqZOm5LgHfTexlm0d/1110njpYl3dqv1Q7W0am7unLnSzJktJ8PLTckBk6IJUpBAG2bUqJGRdlh4o6sxPeBhhx4ms6DLTWmiapUq7uorrzTtIHy08vw1alQ3B76YZCEgwenwJskIEkmdYv9UWBuqYzsf66pJ64j9eh05BkZhooiInCjP4SFoE6WOGRqn62WShEnTVVdd6fopfDXw0WADDmOGYCk9I900fqgnTmVgcaVy/PHHq4yH/cILL7jrrrlWkZsuda++3Mc1ObixlaZO1qxMU+LAeTFtagCN9zC+RaVsYlcU3Aep4AQz6jgCFTReCE1dUvHVQUKiHtWUlguaMcQ4x1dMpUqVo0HUYOnBChXmEbe6ZSuWuWWyAVwvKWG99HTXU6ZNqJOtzl1rNnCDhwxx7777tsyZTjGCweEvSIAQZ2ZmlvtUMdr7vPaa7Ny+cZUUsYk2kf4hLVsuta8BQmpCXB16aBtp0+RII8cLaHqdf55JTw+oU0thyqqbhs8+GOu4yb0cAQulLeEMDGbg4Anup3GzhTyeMoxw7FSzDsELVtgAJihbeeBCkKIawVt9SlMnCFf8lbVJPretTUr41sgzNu7B0zCVdNDCwcrG8OPxTyCH4YbHjxj/YvoLfCbgRwI9dnqCWdO153c0hmRmTURrSuE7Hr/gP3Apz4vyQYBnwZ8iYY6u7L8V1k9+/ufLco/2fIt2QZb1wfM/DyHwysD/WAaFvFQgMf+N6T+m/4iaIB2jX3+dn/78BgNOYfFfobnfIaE58wzRmmorWtNSzJp+0Jp7o61xUuk/lQ8l1z/R9iLiUTAReIIFbAjP4tmHXfnnDzeS/CPIfqzdiLcxYgUDP9mBXYUfeFwMX2MX5jY6hrFhfhKmiaHMdud/++NPfWa44JLaUqMIRK684kprdv6C+e7Bvz7o+n78oZwCv+Huu/8+cxoLfwL+EvlxwSnsiuXyOZp4vq2yvMg1yw32+kWkPDBZApFzzjnPvSQ/MX+Ss138ubD/JKFwYPxObYIv+DhFgHCwBABAwXdJ5SqVLIz0pi2KYCwntcCmXYRCt9x6q4QO1fUxWH5OlHffffdKa+QHmVMd6PDRSts52jOvVjjsmbNm2kdj9swrpHzwkfqO3xf25YwkgikTzqitunLK211+VojmtGlznrkUqZRWSSZTk91CaZ4cULuOzJPGmc+XTRu3uNXyQcMYeLEoT6akjhK16emnn5ZPmGKm8VJX44sGC9ozCLHmzJ4jPzSZimT1vgl9cqUwge9XS+pU2XJl1YzZCrkKlSoYztRR34DVoUNHX9a+hEvjRf2vLwWPe++9x9yZjPhhhOvZq5ccIx/sfhg+Qj5qO9sYMX5ExFoojR8ETRoiPTsOmfVP7Q4cONBkBkbvun72uWfcWWef5arVrGrCutaKAEUb4B8+ibZuVj3GTjlEhFa2zRvmTr6cFS7Qn4ITzKiHaKKgQoVdVlnZw63TOcIWwmSvk4pUVcUSz0jPkCSthoXAmid/L0jRtmjgvUZC9KDixJgVNZU93JpchdgeNthddOGFRlS5q9a68pr4FlLhIkTWPEn6ciRkadSwoRBYIbVXrJTN3AIRTk2bEOzxKkmzppYEMkUVnpsITI0PbuLGKDxWP6mOjR8/Xto0Sy0+OmpiNRWiu4zUq+L0vzkCPbr5aE1QxMAhE3wobXXVGI51WVQjSrSvwipDPoRkXEQnyXJRNrckrAkpsWhU2WBD7Ov6NlNqGVFC7JYERzSeYMIxfMY6Hv8Y/2L62xH/Cawj8J5dOZ5/OtGapDmjhUf/r2XWNMWbNRlfEwMyfGPIo3P4GeeWdPD4mLwmP3Ff59aOFda5YEQlLcfKRW0JhPgfvz79Cn5K/o6eP9F2zH/DMPr3h13F75/4/Qv9QcxJOksgyn/xxDsEPss0ZxYu8ZozmDXZxib0Q1++oWe/CfUfpeAJ2/KbwCeoluA7gT+F+txUSsx/xCi0N1KKLjQkvyt89SSG7+ecWdHEF/j8q1FruiB+UNLCoWwFhVEeMfJH10jmQCRwrHfva+WQt7ftMf/1zDPau5aOcG+rwlMPlVJAribb0+A6RSlDwIFGhzkI1lMXl+kOIZQfuO8Bd8MNN9r+doR8oNSuXdtwBDimhaNoS8O/l2+Uxo3NvwtmPGaapIhC3Y7rZkIcojGtXLGcKpaOkFbN3X+601XUnpd0SOtD3dlnny03IVUkkMlx+GDN1b47c+ZMHWUlIpcdaJ3M0jVCISxMXpCTYrAVesPBbkhYo1x99dWuYqU07YGXSBFCvlZUZsniRXIR0sKdeuopLjMz01FluQQUjFXC3xXDoTGFNVWRtsrdd94ps6QyNky50rJBFjB16lS3dPkyt3zlCvVrufuo30cKc73RlC3Y95MYo6rV5Mg30lKiWSxv8CPbQmZNeQqLXVTzZnxQ98qVreiysrJc+XLl1a/CbuSIHyWQOVaWNuPce+++J+e+N8gKx/ud6d+vv/tZzpqR65iSHjIVXSxVaG40cRK8SOuP4SOGO/zujFFQoCsuv8zdfNPNmmPfjpOD4YjrSG6xUho9dUzhg/nfV6w5uSNllPYmqYeodX3x+SC3avUqH35MKkwb1m2QxHG5Yo/PM8kZ0jsfAWmj2adlZs4yk6VKsn9jsC2SkzAB5EfKhSMipIgtW0h9q0F92cLNMylf9ZrVJWQpJ2TZZJK6k2Sndp4Qq1PHDq66VKjYMLM5BllRqcIj91133S27viqmGoZj4LmKh473aSI0jR0z1hwO4+cGB8Zx+t8dgaA5I17jBklzZtS4LGOctgig28Jqz0c9M+IisQjQbX8u5Ta9TGBY2oFQS/neDtEu7BrOE8qHTQ91Aplyz5e2loDDpTJj+AxEPP6GazH+RTQHTgR6iunP8wg/JrvzS7SmaxRKGzXgT2TWNA7NGTVmCw1xIHhS4IXwtMQCJAISeGH4cp0sS104WMTUONdfKE+5UDb0N9zTDQ+HutKAhfbJs/vhmA9+PP/x+yfgSfz+hTS2t/5ghBI0FtHP73EImjMZdRStaZkPpY1wJvRN1G//jHPoJ/CJX9G/HsjXgf6jdbb4A4l8f89a8ec6NZak+75NGrDi9pO499+EnwSf7FsMv8DnP99Ep4z5npwKs1RNv9rMEwDmkENamxARszo+QGDWhMADDRL8luIYOF/yKCoHt5muWvVq8h0jwSTCKO01EK40OuhAd9TRR8kU6Ss3fPgPigq82j7+9+jew6IIUxY0n6n97rBhOOzNsZDYlSpVNDMcLDWwLkGbZMCATxKgESrUr9/ApddLVx7vUyefq1UFr5H5Spk0aYIpMZSWhQoRo/B/ghJEqZKlpPGzxK3W/neBNIMIO42QhsTeOtANx149e7oDFGVJzbtWMpNinBC21JMLkWISOuEK5BP1aaNMs9hvG90ynIxJJJkopTDdHTt1cF0l3Jg8ZbKsXFab35pVslSZPTtL+/DV8uFS0U2fMT0yjdosZYj1Zo5Us3p111A+a3DkC43DBwmj3VfuSVDqsDzmTjAR0gxWpOYqVWW6pH0618gdiLT08CMPy3nyXxReu4erLKscxnLh/AVWj/ln/L3p5VYTuviPVs5V0px3P727CbJ69+5t43/7Hbe7l1962TWWEI3k8cfJXG24RYbGJAuNmfD8VqgAfwpOMKOu41yXQTKtFUndCAVGyCoGOU8mRSskPWNgUMEiHniapH5MOnNcrnw5C90FQpST5+RicmxEQh1OU+WdAOmckFalhQQgK06XDmvXTn5oUBuTF+uy5cyJUPNmLSwyFFLRlq1amvZOJZXxQptqZn92SOs20shpJmfFG0y6t3TZUlPbGjJkiMycsg12/PO/OwIJ4YyodeBgojXNFgogWRW/4H3Pif+xo8lgdMbygQShJRcBVisiWu76MpxxagvofO0l70drCmsvwLMFRL7y+voZVYnh+4GIxz/Gv/2Z/owDGY8wLrPbP14408E+XPhoTUSU0OKDfzClaOVFDsny7MQurSxF6AL17JbqhX++lH5VnUVlaI/81G6Hdm1TpTtWX/yXd3ayZMz/Yv4P5oAXHh9j/r9r/J9xCvILP4K/36/XnJHPGQulLc2ZvsPlxDTPOuR5ud9OeJ6QyiXYEEV8yOYf7pD/IxiNBF7iF3DRc6qaf/4kP/E8zbefWH8Zv9tb+EFQFB2tU7+GH/XMDjF8PxoFM//J8UcboUCT9gZoTeCvhUhDaZWrwIxsf4kvU8IgE7n3tNNOcyulSPDe+++bk+CgBWI9E47hB6aVtElycpZa9zBxQYEAx7lpaXIUrJQrocSpp5/u3n//PffyKy+5hx96WGY9pYTWm4XLm90b//mPhEMLLBLTGu2ZccA7bfo0i1KEJk66nPISrhmrDsNy9RMBjb1PbVj82LC/vfb667XtKWKBcnD98Z3ceBAp6dC2h5k2ClYp7KHx/fqnu++ycNd5cjdSuJDXnOEeCgwIetBeQZOkuFyD4Hh3nEyZGilITlFZwEyfMUM+W3Pd4O+GiE43axy9+VHqO57zcjJNevTRx+QYWHtwmVdd1fsqa2dWZqbubpVQapj5xFmvfTcCsbfffMvdI/OvdevXyZKlsfbwZfWw0szR82LN8vmgz2WKpdFHmKTnZC2ybu0aWc1UsP0aJl7AXaW+Pf74P1y3bt1cRkZ9N0TCr7bSYsIXrbGcaP5zZYZWRWZmtAlv5fnrSgOGQEAV5L8GHoRyyblS8qA/70oLp8ux3Qw+ayA0fT7q+6HB9O8ymxgmp0BTAQpm5M9F4anMpk6ThsOf6b9MN1UotFBQlUJ7BYReo0moKElhvXrpLk2Srx9/HGkqSzVlfoRH5jKly7kyQmR2swce2MDVUozz9ZoMjZrhQWEhHvNRURK4WjVqmZnSjBm/GKzlIirUxiC+OnXruIsuusjs+rhehZqXkGyz2q0h06bDDz/CEHCyQpHVUFx0JImED0NiF6f//RFAOHPrZceKuGTWJM2ZSdN8KFmjVc+7Eg9hebpiQcAiIXFbdcNCAkLlnKO9aMA3lfSbjlCOJj0x+nqJlnxdAbK7yWwqGL7aMYYfj3+Mf/s9/cELTIDMyR4mC6Xds6M+2myVWZO0PifP8XxL7YV1bWBDxtPIF3/i40jggMbhjOf5TmzL/ygH74SnhXuhTc//onvhPvwvKu9b9L/kkWL+G79/4vdvoD4jiF1Yfxjp/E/8eM2Zc+RzRpozOdKckUNgi9ZE78RXwlpJF/n6m+Q/Af+T9wMfsSbgRVp/J9df8Cz/58vBv4wb+XFLNlMA8KP1307hJx/L+hPDTxmQvZ3/lPHXHId3WBLAnp+BQ5jxLJPPkXJswPWPD66gUlEpB5x04kkuPT3dwldPnDjZ3a7ITTfJLOnFF1+w/aPUM3gBSsEgz7373gcyp8m0zmAaA96v154WoQs4UUT7yHmyxrhD5j3XS3ByuUxjCK9NRGEEAhtlNcKeGPirJChgT7xUf/Xr15f/00Ml/MlyK5atcCMUgWiJBCS8fPXfJ8CFK+FeGykY/B973wGYVZVtvYE0eiAhlFASeu+9FwEFRGEUFAEL1tHxjYO+mbG3maeOY+8VaTpYQap0pPfeSSDUFELv7V9rnXu/JOrMe/8QHKL3QL7vlnNP2afcc9a399o3DxoIc6pwOcppBW9Oe6EMUb1aNbm0Lg1tFO5ryftKJYjhw0fYHx58UI5xUFSYdxXAeELS+KNpFbVhypUrb4dBCbJh3XrtvwluVIUjn8MwUdoAkuNly1eingBLvOdcwXjiAjV67rrzTmkdrVmz1hITE+E9qp5tBw8sQat24I5J3ZcqBQ56gfoCnqqG3DZEpMhVqlSxRg2baO/P9j8AZQ7OBZg1kN8F8c+GhYMYGJpP+sEIl6nIcf/9vxNpMvllPwLXzx643x737TiZarFebBfOQRs3brIOnToJNGJbsPmKwFJn6ndTrQacCClevjCbPnWaXdmtu23btsU+/2KM1QFI47fBoUMwbUPlfzT+fQHkwncuIhD5ILTTMmPal54ONa2dRpfWp6ktAy0XIlWpAGeIIn7y8TDwxnwPpHAnOGSOAT3cLZu729GBz547g3ip4o05BXSzYoVKcou9dNkyoJwl5bv8JJA4uiDTjhy7cvLTEIHLgAszEv6sQmf48qsvjSzTtLfjRE/Vo+0pO+DRKVoDhyZM6al77SRs9Kghww5RsBDVv2jKlI05OheEHCRx6STwmysb2UN3wFsThk1a5lF8ccpFn/BHkZ+15g1+ePaCuu+mOD8uv/nHWEzFJUKc1w1CThBKxkvT8SK5jJSrEsKRl46ftb71ID+C/CUxfQTypxjUbbzvoP/9esZfvlxYeVJzhpwzXGR9O2ON45xhJ3L/Q4sJzWHoY5qnOD/imEEbIO+Ex26O82Y5niMOz7gIEW9WthmQC74cgfkicFQznxxBp/wI5j9JSR/B/EcxeN1G3zxWf+OnbvzK379YTPxoLOUYWD//ia85k1AhxvaJENjTnFF7obT41mGOorEWrmWzZh83/2gOUj/Imn8U10tET3rzDyc6DR3G5/zjnTBllyliX8L8Oc/64deYP398Xzh3mi2aN93mzp6EH7JBjIuQ2/Jnmuf9hubJRQb1PKS3c2eKNWvcFDwj+9F1zkljIhzAw7x586xW7VrIBVS00Ah55ZVXbAAAD5rYjMGetQo4TP3ibNm8WT/gs7/RNIpmTdSaUc9EJLrc5vuyc8dONvabsdauXQftLafCzKl7924WhrhMjPlzLBB8oKbM1m1boTFS2Dp16SQtk7MAP9asXisX1klJSU4CqIhfDh5wr0HvxD2u6gntkHCrXae2HQLBLstUARowhcDVWrFiJWjsRNjQhx6Ux6MVy5bb4JsH206YGNHkqTA4dbLWARfk8YkKCtgog28m3fbB+oV1TQOYRGBl3dq12q8vgBcq5u9aP2tvRFkTzCF/Db0yk/A4Gvvuo1CuoPz5/IpVKywD+/xDAICawKyM9Rs0eBAc+Iy3gYMH2tNPP20x8KBE8IX/MNi1/ti7Z6/4fGiFI60q7P1p1kQNpFq16yjuPpR5INruPZinvfi3v9m7776L+pMTiITOqTZx/ARo+h2H+Fz7F0Aa9D4VGxeLdgPS463LDsI5UN++v7HnX3henrDefvMtS0xMdO2AT/3AoPqHLuXaQViupQTZRRcvBnKlonYQmikHYVu28ehGqw+btcpwLZUPAj575pztydwD4aTb5k2bRb7Ljt8KfuTpXz4yIgqAy1ljx09OTpaWTb/rrge4c1ZER1Tduq7+dWKaptpVTMkS0oyRhgt7A4R6HmjkMYBAKSAIol1eGNSyFkNFrRHKQfUtctiQsfnD9z/UACRJMdWnOLDovvsEOHE4+QQh70iAmjMML7w3TXPE7PlrbPb81ZpMNKjRN7ipoEosN0OcQPkrM2BYDEROjW5qYRpunLEz4UgDFN/eqZ7kse7xtlJ36fmRsn8jzyB/CAz/A/kH/S8Yf5gzfjD/cDa52EDNGYY3R86xcVP5a9YFa1CrQo5kBbjguuY3zkuc/xD4zWs8o2oxVXt5JuJfLFh0w4uhRZh7DHE0C7rndOade+mG0kQMzrDIUnOtnsuRv58SFjpYI/ibNC7IBHwjv9AMjTQ0b6s8WflnpcA8vPrg2y9qkH8g/zzb/4g9uN2PN8oujy+CM28/1d/uefwz274r00Z+tdAG9W2FX62hzY4ichrIGn+og8a8K7s/FzGGN13oBn+tzzH+2WgIbvwjNk/dJe96zjx4keldyvyzZe/yUUmyivVLzj/zQLqtXbnEWrXrKssDmgVt27zOatRuGGqW3Ko/xeo1vyfhi/uiFgv1Lk7CkmM5OEV3YH/ZsXNHlygyovedhIQEnJOTDftI9MUZM6Zbz5497I033rL+N/SDWc494qIRxwpi5kM8rmupFBAXG2cZmRnuvclU8B5t3rwFTIqa22OPPmrDhw2znr2uti+++MJeevkVe+Wll2SGzDEib8Yo15rVa7BvPiQ6DmqmUNODoM261etsytTJIqdtBSuPrODerep9SIgObjgiGjZoIM7UrVu2WgV4SaL7ae7HV0DT5fYht9s8EBNPg7emt959x+oAzKD3J1qfHAJhL+uPKsnBToP69eGZaYV17nyFNGUIAFEzh+Y827Zts0UL58tFN7WBONj5HA80XjlY8b8u0uA26uWXX8ItelpqLw9KUeCuqQMQKQKAVIuWLfk4uGh22HPPvSDqk1tvuVWmSI88/DC4c1YqLQhV1CgDagywpXBZzg0ds4wCuEQ8oUrVKrZi2VLs5QkipdkTTz5pC7D3JxBD8y3Sp6yHQsZqyJnPOTMoA15xSF6awvOHe3tE1iUfOHxaymX4//z1OevQoYO1btfGRjccZY8++phNR99g+zOdSxG4CsudwEYAOtYHpEm0/yoGkKYQkDiaENGGjypTYfgjKBIF9KpAeAGpFc2eNRts0DcJLTsChKpo0cJiuT4F7Ztv4J983LhxNmvOTBAGLbVjh48KWNkBzRfy18ycMdtSQXBUPLq4Va1WVW66CLCwFeU3Hs19FOjhxAkTbRLsxkj4lLID5EggBSLatm9vqtSraAtHMuCW6CANG9ZHecNzRyZBKj+bBGTWdEcXl59Gi4aeFjWcKIhucs/BCUC3eYyJQse4SgjFP2YMxuNLhh/+cX4fwudM419HXDch4ZISwIf/zTSVR5B/IP+g/wXjD3PED+YfnOZKIDgT0pwRIfBOLS6YYdZGiFlhcnITm7uPK36Z3KYodOZuqHS8hifdBKdjfhBIYQhtppgS/rv5EMtgToXeuQOCXHwea1b1J06VICsdJcoPF907xTyrc02uupmVvxcl9MU4zN89w2yC/Cm8QP7sInmq/3nrDJb7cgsliheytwDOJEJzhq60h38Ns6bjp0LFdOOP66osIJg3Q/LnXJQt+POIu5TzHq+F5h/dyiIPzvGc5jaXgpt//t38XRrZP0P56+LPm//Jk3AxjB8Ws8LPm39G+j5L3rzB2nXqIVCGc0ny1g1WqXL10PzvxJI1aV+M/JVKtrbMqve/d8T3DTfsDCTo3Za8TWZFBFD8H+JD7z9kzvxJCksgZ8BNN4o49y9PP2vvvPOOAACmg60sX1raax6A1yEqAbDIKjayIgluSZgQFYyKskwALLNmzbTFi5fYw3/+E8x3vgDRcEnvnQZQBd6Aye+SDqsSOqIpVLgw4i4Cd8xhWJfMgUZNlM2cNoPZgi/mjL654ZCceIb83HE+gKOFpH2j9LA/pgcnvoxTofnCdXgvgE3UyjkGpzrvvfuegBlqEEVjH00yXSbEtOLjywmQOAW+m44dO9mZk6dty5Ytau9169bB2uS0HYVmzQqAN/QQ5eevwjlRh0b4nXffBSLdGgB0kqxe3ToWHhluKSm7pD20ZAnAvlat7MYbboRMz9kCmHD1vvYakBzH2tSpU61Xr55aa8A5kur+wt9eEJBG+TNPagdxv0+Qh+cE4OiGnHQlkyZOBMDTDOZjZ+Eme6INHDgIgBf4frCYoCkT2z8lJcWqVK7iWcpQMzi/SI/r1K4tzhrmsnPnLrumV2+04wGZSBHgOncJ3WXnjsaMkwbcVEUCtSsL1uIi8IQUZ5UqJkh9aQ/UjzipcWCw1T65BwAAQABJREFUK5G49wYwYL/3/nvoLGlC4UrFxcHl1Rq5GqsHlG3r1m3qpCQF4qDanpxku0CSFA/0b97c+fAV39AKgomawE908TLWtk1badlMnjzJdqDBz8EkKhxcNLSnJ2EwGZSpJnbgAEmb8tuTjz8p1O+9997HQEILoyM1btwEqGQX2eqpcwUfeUoCNGviXxACCQQSCCTwa5MAzZoIQrwxYo59C1faDA1rV8SveHhBewslASK6wwt8cbtb7rpb6FGrx72p3X1F8iPyVclM8N9PixstXtO3Fr8ubS1Qce7y9/Njmkyd/7JKwLTcFd12eSh7vwxIR3nwVyo8iWM+z8B8ec4VscqhO0iRjwb5B/JX//d7GzsF+47rP653sd9fnv1PpfU7Ojv7ZRYc50w/++0TY6A5s99Gjl1kA691mjMal6Hxl7Pg2eXP6pH1yrUITnAhe+v4TxIX1pjmAcc/xzyC8uFDTIP58T5bNMf4Z0yGH7c/Y/84f/5m7XqH/4zy1/zHdKhZ4O5fivxPYNO8d/d2ZqQ6RUYWlLedkyeOW90GzXD138//yOFDcK+8ER5zj1uz1p0luX9V/4OZ6ZaSvMUat+ygucSJPT84Rw7iR3bH36FGy1X5u7aVAHLhg/wqFcE5yh//uSFnmzIHmqPR21BMbIz6C9vS7yHVQHxbr15d+/jjj0HUO9eefeZZ+8tfnrXRoz61Pz/8Z3kaliYOHpgIAKA2NvJMl89zT0kNmBtuuAFgRwTAmULgpzkFTY/Ftm/fPuvfr5/4ZFatWm3bkrbqGQIq5cpFa/NfDl6S4uPLW926dW3UqFECba6CuRK5ULZCW+Xuu+/mEJDGCslrued2EoNlCqxGWKczsD6h5UlSUhIUF6rBXMd5SKIZUIX4CjAlyhBQw7LXB1nv5ElT5KK6GDh4CE4QJCkAaxdamSQkkn/nrK0D2NQCLrzHjv0GJlBFrDUAlUkTJsI06AQ0SR6V0oUbyZCCximEgUCPU9cCbCEIs3DhYmvWtKksVQogbbrmpqcr5i8+GMRfsXyFwKAPwRczbuxYS8L+vxiAMpDHwLrmvB07C1MrCIDgbxrqEQvvSa7+bAHwxmCv3//6/jbmizGQ+XJbuniprV6zyt5++00pYDyM9mOD5wOAcxDuyTt26Wjjx01AnTkSTFy3pGBhuShnTCu2Y+dOaFD1tJdeetlef+MNWwUX3X6eeCRXA2efiw8qHQh1y5SxqlWrWWnYhhHhqwh2afoznwSwhKRJVFkiyzWJexpAM4WdhYjkOpAM7U/fD3WiMGnb0ESpCBDD0wBd0tMz4fbqpJHEiJo2BHc2Qs2LrrCu7XMN7M1ayA0X+pHIk6rQNKplC3vggT9YMyBl9NhEFawwpM33sz8oa9SsblcBOSSqth/cNHTrzXSpcROOQRyEQAKBBAIJBBIIJJCXJDDoWrrSbocFtBm9NZEQmKsHvqL9jUxWfXCVqxCuPBAjdB/XuLzUOY79DZM2IHxEO56sVNyRHnL3mDmeUjylr1Okx+UG82JwC1geMXvmwW9dZfrec26T5Z7hYpdHTNe7wscR3BmfUdGYv46z0mGUIP9A/qG+4nqa6z3sG/yH78ut/2lzw7FwGQeZNT3dzxLiS9q+dLrSng8Q4XTO8Y/yh+YXry4efsIbnvwJkCH4H7iefarR+NdQ59zESPhTXBzyHJFD4593ecL7esbP3zvhfR7qfvb8vWd4I5S/e8bPX/PPJcw/JRkecA5mWJUada1ytTpWuWpt/CCdaFWq17ZqNevZlo2r/+36b920zvbtSbH6jVpaFPZY/1v96Xp504bV1rg53imcfyEK1n/D2qVWrUYd9zzEmOvyZ9sxs1wMX4PvhZwhNMdhoPYIAz0WNW7YWHtL12f4/riAfWMM7p0HqFJYdZ4wcbz17n21NE8mggulf//roRgAt8lIg5otCxeCc4UnkBO/acmxFyBMEViB1IC2CN9lO3akgOR2o338yUd2CuY1bdu2sWbNmjslAgAh5G3r27ePtQOJ7569e0HaG4s9cbT2v8kAJ1iGTGi30AU13X7Tq9PSpctgOQIeGAX2ZY4PuAXfs0uaH1vgGCciIkyeoejpqTTABnK5UsPH54nLSMsAENIBHLCfY89cwmjGFAV+mguwEmC5K1asaNXhuYjgzlFQhVSpUlnKF5tApkutmaLFisiKhRoxrtXwyb7CMnnNSJLldu3aq36lwOdSumxpgT7Tpk0H9UgK7rXVvp4gCJ8kwDR48K3233/8k1VOrAxvT8/L/MlLFSAKKEeQAU2wGjeB+3OAKgycG4gzkNe2UcNGAN8M/LcAjh573G6BidTd0N6hmVpxAj0IbKvXX3ld+3+2P/MnJQr7x8mTkBOANc1dSLdgwSj7wx8eAKHzEGgAVfeLonRy84Nv6lwJbACyXhOVJDM00T+6uzoGP+RHj4KUFaF923awd6sg12JTwXp8GLZqrFkqCHnS09MtDloz9CV/pRiyK0kIp08dB/J3GqRGdYHm1QeRUG0QKHWX2lXRIsVEdkQ1pGUgNCLj8hHayWFA0xc93WW1aNHcHn/icdnnRcJEiSpOVHVixw3DIGD+VIUi4e8nn3wCF2l/RPl2sbhBCCQQSCCQQCCBQAJ5SgKD+7SQWRNXHONnrBYhMN/PbkPhrZJYI65I8P7TYjR0zis6cfGxStFmFZf4pFug4ADPcgnIwGsyVcQ3j93zeg3jXFGUlsvfO+cXIvKaisFTnHspejfxS5W3QmGa2j75eTCunzgeUh0QmZcYl4HHfhSXDxes7p7i4CPIP5D/5d3/2NH9Hu367uX4WaI4OGeeuQFmTQBnMuFK++tF2tSExihnEQjaPw+NP+8666T7ocppcIbGPy/7419RKBas4zVhYFDzWe7AOL597gjmlZXfD/Nn+vzvrjNNl783QegmzV+UG89C+TNNN1ddmvwrJlS1svEVWaAf5U8+Dc61P5U/Z+QNa5dBU2OJrVw6T3XPXv9VyxcAbIgDuFNf9T5Nl8X8x8QQflh/undesWSOtWjTJUf99+zeAbChODa2MJGhuC6B/Lm59sulwl3kB3lcZk2fYR988AG0Xv4KbQ+AUtjIq7+g/hMmTpDLZlYIxhW6TkLeFcuXGEEEeQLD9ZSUnQAW+tqrr75qjzzyKLhn+kOzJZ4diYLAHyXqjsnfshlusJs0aaJ9ZwQ4TKtWrQqNGXgkApfM/HnzFbNtu3Z2NQAflonmuWzH4vBa/Mgjj9iqFcvEHVMpobK0X6h4EAPAaMmSxdDSmWTTYJZFpQfyphAUYf7lsceOiyuttMkvc+QIvAihUPVQHu5/jx87YbGxUJ4AMMMbNAG6qsdVivf3v78E/pSZUHYoCtCoLcy5iiIGOHcAzpCL9a9w/U1nPfSgXBOkvnTHzUBFjO+gzUPQ6aVXXlb/5HV/NCkO9t80G6PFDP9Kl4oTyEN5HQMxMD0qUQYXMI5Jf0Jel6jIMPHTLIfrbnqtmgDtnK5duzI5lwcyICBDnh8qWLhA65wC0ggiOMbA/f55gGhjxoyx1q3bSPniyccfh9ZSEwBWUXJWlJEBcyyOLcQlJw3xjAvnYXkDpQ6o1kiGDRs0UhuOHDnK+Jejgi7zXPn0pp2LT4vvDiJ57ATFwTETDlu19evXq7OQ2fkwrhMBYyVJzvvll19KU4ZCTdqWDNTtCJCpAmBxrm8DbrzR/jB0KNS9+qPBTui54uCsYYesXKWyPf/8CzZgwEA1Jm3diOpykkkBikh7MaKC48ePFX8M/a23atlaHa8o0qDrbaKHJDA6jrTZEAXyg5wYDcx0Nm7YIFWyi5dIkEIggUACgQQCCQQS+PklQLOmewd2lB30+Gnw1rRxpwqhxTdfdgpZGxetcrzLbkPjlgY5FsdcgHMxi39cTGkJigOlqWu4wkW+0ufykurQ/Hb/9ATu8zazcvm4TN1zLJS7GdpQMH388+P7v3Iypp5hfkxTF/hLpcuPp8w1yN/Jg7II5O/6Z17qf9wQsYvnhUDOmTefBOdMfIylZRy0kQBnjh07qXHoRjyHqps/NJ55R3OFN5Y1iFFTXHPabRraGtOsP+Nyv6Ahj2e5YcSHImX1bz8nfrv4/Hb3GdXP37v6f8xf5fXy93P4T+VPMxOG7PkfyEizVcvmScOmXoPm8ALktGGkRQCBLVs0x6pWr2slsSFnzdP37QFIg+N/Uf+lC2dbs5adJX+//inbt4Jb5ARIZavmyN/Jl+PLl87FyT9rFldVL/qD75DHn3zC7rvvXntw6B+wwZ8grQpXf1phbLQ54C31OpfrVuhrJ7CvLAjNEffGg2tm8MDQCuPPf37EHn7kYXkcGjVyJKg02igO3z/8g1EITGdWWyI0PQiElANfS0FwrlarVtXiwJ3CMANAEQNlm5iYYEOG3CZeG54zJIC4l5yt113fT96MaOZDDZlNILrdhj0zeVW4/yV3zCGQ+5Lug3tueiy+EWTFdFFNCxG+hw9A04Wuubt16wp32ru1T2ZbUWOmRHQJeaSi2dYpWLDQ+Q5dWmeCELlDx05WFvw3fv3pdvummwYibRAMr1yJvXSElBqWLFqM+Bm2e9dueWKePWuGrV2zVnJ0HzBnKloEwEshmCIR7AiX/Al+9e3TVwTF++BRmdgByxSOdGvA2c9VUNIgmLMfaffp08c2YH8+Zcp30Hq5J9TR2IU/+vAjcPMsd4LD51mYjZHGRPMI5lCZOCIecYQFC+ZD86m30ZU33XTTRI3XNbdgvcL8zyBPWegATGMfoecmhrLxZe3mm2/BUT7nreoSzc25Bsyw0OvWb1BlExISYMZUSQRGRL5OnDymDnUAiBbVp9hBOVDYgS4ArFmzBmjf1m0YFKg8648eUxx2bg3ALp1JMyMQX/W+urf8s2+HtyVGqQy/6nR1PRKDgipqRCJLASXMD0TwFAbTDKB+Y+HHnOQ/k6dMtOWwWcuEr/iC6Ny0m1sK9uaVq1bCHXd52ZFFRkRCfQkDCuAQkeIgBBIIJBBIIJBAIIG8KoFBfZoBnGmPBco5+3bqalu9wRECC9BApfgO9gMXMFjHaBGixSXtfhG4KOH7loHXHYeDdw/XeO7iuG/F8Z6gtbYWvn4CvI70XP7UsuGi0QXGC3HhIB8CKgw58+dCy3+e992Ci2m6fHhPl/HB2gX5O7lAHAqB/NlX8lT/Q3/O1UW63xUu0bcza+pvlcqXtL0Zh2zkN4u0uQ1lx3kG/xh+PP7deOdEJEABcTCy3djGsTOl4Tmf/sH45yW/bV0E5eP6v8uPURTnX+SvxP+X/FkOF1AWbtiQn4r0M+VPuWXPfwfAkoPgzmzYrC02ypEqDzVEWChycC5ZONMaNWmD/ViRUP03rlthCVWyeDk0rzJZyg7pb96wyhJxn+lJhri1dfNabLpPwrSqZo78L0X9lYGE6mV1kV8EJ86CM2Qk+Fo6dOpopcuUhmnLozDZqSeaC2Z1AvtJvTW8/d9RWHtwX+hvylmEKlWq2CC4YuY7ldwq5J85AUc1dFJzE4hl/fcWe9xCeASaOX2Gbd6yEZozWyy+bDnsUePkCCc21mkbLYXmi+ud+SwiPBL0HNeyozMrBWptlCoVI7fTJP2tXaeWQIQ94Fo5DO/HJA3etGkLiIIXg8/1uLRGqJjwHUhzrwWQ0fWKK7AnBoUHFBTGfv2NrYIVS0M4wdmzd4/GH7tvRGSE7YImUCOYc7HPkLeFHpI6de5kn4/53JoiPgEp9RGUjfdJDNweZkkcp/S2TNCoKEyujsg6Jh+IjzfY6M8+ta1btiBNN+LPowzUrCkBahENGORVDGANZbdr5y6rXq2GW08wD4AjVJ7gyGLbME866fkNgKpnn31GXD9U3KAFjjN3hitscM34geZVRVBmFBB5qWNL8bBa9Rp2y623iLNm2CfD7LPPRtvQB4fa+G/HWymAlhpb3vjaCU4ZcuMQSAoFNM3MmTNUJzUcC3cJQu7M+exIFBzsuOiWq3Jiguzm6qHT14a6U7Wq1RxShg4Pb+SqbAm456oH0p+qNWqo45NPBm3hOGCQ1gqgcTSFKg/TKE4o9es3UGPfd+/vRLz0BXzL74H9HjVxkrdts4TERLjmri925XhoxVBVixo4h9C406dNByr2jN3z23tszGdjgOrtMpFr4fkqIHi66sru8lMfB5U1djD+BSGQQCCBQAKBBAIJ5GUJUHPmtze114JzHAiBV613mjNcT3DJ4y8B+Y3XrjatWoh7Cw5/E8vYbnmFG/yPyDQZ8H9l0oIW6wBez+HSmmkCZHHPIhM/Qy9/gjcKfppeGsyDwW1C3InKiBssp19GLYSZPm66ZQjz98qH5xk3yD+Qf17tf/ytEqNKYyGvfFBz5u2nbrAEgDNpGfDW9OUCkIOeVvE5Nv1JQOYF2SrFseqbRGrg4l5ofvCe0rjHMeNSLAJwsPlisgKFcYPzj8a94mGe0vyDE55ToN6cw7S4ZfODnvnh/OPPT4jEPDS/MY//UP70HuvAEhYG5K/rV2o/Vbla7Rz1p/MTWgCsWrHAmrboCDAA5hh8BPVfs3Kh1a7fVHX5qfoT5CG1Q2xc2ZD8161eKo6SquC9cTJw+V8q+WfrJir3xX5w/BNEKwb6C4JV1/a+xpHo/vZeeEn6s36wZ5O64GaLvXBaUxd72HP6kQBtjue5ryTVRXFs+s8hzTRQcPSG16BPP/vMPh72sf0JfCjhUA5Q4+D+V199KYBh+bJlRlOZBfMWaDSTsJdaGaXAy3oI5LOzZs1i47js0bm8I+kqMK12MHcizcfKlaulwdMdmi8kQ+F4OQDNFiocbIfFCLWpzoDc9zZogpCouEnjxrBAGSDPUMx4wrcT7E0Q15YtUxaAKYAoal8hDQIZ9MrEMUQeVypQcC9//+//SxwvJABmuxeABolkic/6UJy4bcgQKwrzoxMA7O66506YHkXJRIqmRJEAfL6bMkVKGdS0pbkSlSYo55IlSkqJgj/NbNu2FaZkE61GjeqyuGH+rBc1ZU4TFNF4g5kirtNb87PPPmu33DzYKiUm2JtvvWUdO7R34ob80EoCsmkNUxXus5kvA4EdPk/coXixYihjBDh+TtucufPwfAerlJAAjqD5cA3e2cLUfmbTpk2ztXCvfQqAHuvD55kRXasTrOOs4/qp31rKKlc+cgeYQaUZ0vbuExJ3GsLjwC4BBI2cMFSLqgQNGrrKJiEPG4eupsqhc9QCKVIMOnt6Wrpt2rwV/sdTldbunbvRWWfaPWCfvh2NT5AlKiIK9m077XP4gn8MXpVef+VVCfEsAB/avZ1BunTTTRu7xrDrI/EPfZdTk4aNvHHTJrkpoy0c86dUyVBesGBhqXOVK1/Bypcvj7xoexeEQAKBBAIJBBIIJJC3JTC4bwtpznCDMnbaSsc5w7UEX9uhNYW3yGBVsSjiAoeBizGqzLslibuG24zi4mnR5K5nxWdaLmHFyxbHB3qyNlF8CsErB/PlM/wBhxe5oPJvujIhZdx3+bvyceHGwJLykHk7EyYvXpC/2pEyCuTPjsY+xr6CHuX1O3QZhcut/7mu7RfSlTEvfJaIBucMXGlXhFlT6v4jMGtaCKDgdBbQiiqxbg7EzaqRRjvuYRumiz85/nGHMxTbjG2ohHCNT6j98J01/l1GLjUCP9yoZT37z/N3W6OfzB9J/vT8c+nz37Z5HX6sxoYTZdiycQ2IXOMFoKBSofpvhwelkjGlbeO65QBlOmBD6UvTwOlxCBvuTAAMZbz+j4SyByS8Cdoytes00VXKbdH86VaufCXsqyrj7F/Nv7lX/+wAWvbi/bvHan/0lbL40b5LlytgGrTbPvl4mH0FSo2nnnrKxvxjDDQ+HDBBHIbvny1bNlvzZs0AFhyGvPEOBKgVFl7AkpKTrXqNWlLEYP8gwHL//ffbPXfdJY2XJvA4xP0n37eUyFGYH1EDZD/kTlOcEgAlChcuateBiHgnyICXLFkKa45M5ANPQ17wAUzXb10bJSYk2kP/PRQAx3GAGmGWWDEB2hxnBWLQ09JBpEHzJpIZkzfm+eeeF0BzGnHo6bgNOGNoEULrkVkzZxs1WOhVippC9GCcD+CD3g8AayonVMZeOB5aPOH20Ucf2vz54MNBMTjc6EmJnCssYzQsW5565mloHjWwpTBnii9fTpoqOwFeFYFCRTq0WJKhfUNun+34PgkAh6AJ06IXZ/LOEuyhidbo0Z/JWRDLcQ5aOCwbKiegSFpqGLrUHqpcpYqNnzDJXnv5VfDUHrKZALUGwrwq3AONWEjWY+GChVDcIL8tZgVew1F+2JiRm4fyZ/nDcD5nzvfQLOpqS5YuR10/siu6dkPZi4qI+Ps5czCXnFO9HYp7XjhGGLSbSEbk9yu/3XLr280+uZAaK00Ebh+Ald1Qk1qzeq1s7Mg7Q5CGxEFEV9PS0ixlV4qlwyayAHhoaL9HhIzo7hrY5I3+dLSNGj1Kcejui6pf7Oi0ZateoxoQrS52RZcucMUNd1+wOwtDIx4D2nfLrbfaX/76V1sONa7jSCsMvDH05FS1WhX5SK9YsRJ4b8KB4sFkCYOMi02yOtMv+zKgmTuANpKtumTJkq7j5YJMgiQCCQQSCCQQSCCQwH9aAs5bUwftKKQ5s2mntwjTssUt6rnq4opfixi36Aht5LWs4VtetxHPgTW67D602PI3vIrHuDxgcI9qgcSLWihxk+RFYNYue+aP9ZhWQVxgucfdJ0/cIsuVk+minF4azMP/xV0bNsb1H/fScRs55sW0g/x9AQXyzwv9z+/MeePb99ZUKT7WUqE5Q84ZejlRwMB048+NUH66MYmBynvehOHmH7++3vjnqfdLuMZ/6DZX9W4+cONfSfl39e1fVx4a//9G/nzEn5g4//g54PhS509HKOTnIEBDjz3RJWJy5M8fqfftTbHUvbusUdO2Wfe8Ms6cOs5at+/uZPwT9V+9YqE1bNpG8j918oTNmzXJGjZpbSVKlmKl8cf5l4fehHqp6s9ds7SbvIJf9Bc0KQA4kFyWNBb1G9STp57pIM/t1q2bNWrU0L7/fo5cKVMzgu8f8pM+A+0MggsMVC6hSRDbIBbmOOyjhLzYHhXgtehDbOpvv+MOABNFAPh8BW6Y61wf91sBsipYqKC4STZt3ggNpAgAEYcERpSAJ6SdO1OUzyFcO8e2CdUfApfo8Tzcbj8B0toY7FMJwpBblftren/KwF57f2ambdqyCWZB1S22VCzMhMbaDf1v0L77ySeegJejwQInMqEVRWSCbUmTobLlyqC+bmwSvCgHgKVyYhWYDl0nTp1Dhw5Zys4dMGU7rdrIvEv9QEW2Bx96SFYnUXD6sxtmVgRayH+Tnp4mEl5q3NByhfQlJCumggYrdduQW+1JcP/QUzN5ZqlAIc0YcMQUI1BGuUPKWldA/qVLl7G7AYBFwrMW5XTXnXeCHPhlac78/aWX1A7+3EFi5GVoa9dnXfszHWru0FMWzdH8+YAg3D333G3PPfecyIU//cenaONYaOicCj1PeZGzpyDIgjn/KB9vGCBSroZcA2bYRgReiCCeR+GpnrUOakBzoSq0cMECACBwAYfGZSc6fvQ4CItSZVNXFOzetJ0jMTDvHThwACQ+K0Hku0O2f2R5Tk4iOfBRIGwF4bKruny4/+6/7pfaEdWpCbgQWPn666+FgE6eOBkoJF4Cp09K7a8oGoGcNeEg/S0PThkSDLEs5Kg5y7KCGGnL1q121VVXwQ3WHVB3gy1mEAIJBBIIJBBIIJDAL0QCg2DWdO+gjnhRm+Oc2bhLNeMCg+sLfeJ96oMr/pqDixf8R3D3FJkXdF03lI5T9UU6eFC/0fKWbiNlXsSDSsdL2MEmjODu87IWSojrE1y6Z7BMUToOTFFaONfzSFBJM23vl2F3jsQYgvwhhED+ebX/ObfQ6sTqznntw4Ez11vFcjG2j4TA3zjNGdbDGR24GrnZgfMD6opfyf0Nk+JxRywRYKxLCwFXsXlWHI5/l8RPj389yP7vIuVIF7OUH/55/ixL9vyxIWNiyPfnzp+aLsWjS9runckgUi1mpUqXUzmy138p+GS4/6J7a6/G+HL1X7F0rjVo3BKbYJra/Fj+6Wl7APTEwswjEhocIBOGF6fWHbqDBBXUDiH5/0z1dwVUOXPjg8VnM7IeBzMPwtUx+XWgFwMti6ngY+kGT79bt22F1sZoS0xIdJtvRD+I/Sg35xewKc8PLRHuG/njfj4oFTBw/8n95KOPPiK+k1Wg4NgCzpdUENn+47N/gCT4YUfjgTbhjwYVK1UEsHNGHpYWLViEMhWAKc82WIfMsuTtydr/vvfueyLgJcjmMmE+bGU1gtqveYsW2v8eA+8K34NUbhB4gn1zMoiBO7RvZ506dbJOHTvb1q1bbNKkyfbSyy/Z7bffbo89+pgUEIj9UCgEZJi6A0uYSz5bDH6cwkUK29q1a6wLlCE2A7g4d/Y8wJkUlZH9X4FfHA74ohenTh07YvxCiwhejQ5iX03TL5oI7YU1TUxsSV2b+/1cYQXcf8/B8cCBA+2T4cOhbROvsaV3P2S1IwXev1AG4gJUpqD8aSo1YsQIuNouo7jHoQxC8zG6wO7Rowc0n/5h8eXKqzwcpwSsWEBx6BFuYzr4RzBNwBfqHwOQjc6GWN7XXnvN3n//fZE5z503V0od5Plhgsw/KSnJGjZsaGdQJn8dggxyPbjedZHJupceGIuhoRIO2y26n6bZEicI8s5QS2YrGnbj+o1C3KgFEw12Z3YIemkigkaOGWq50PasGCYdqQ+hdEQr165dZ0sWL5EdG1mdCwF1LFsa7rpq1OSUIyKnoQ8MtSs6dwbrNezUgMYRCDpy8IiIfcuCdKlgFFSPgEBSyJEwiToAlJDaPbQ3pHoXO9p5dDy65CLrdhACCQQSCCQQSCCQwC9JAoOubSbOGS4Sx37nzJq4wcApfhDyFluosP9O54qEmyEu4hiPMXwAhg/pdyN860mu7hhwokN8MF1d0kLO5eOnzTJ4N108/yHmyQyVCo+ZP24qE7cI5YmucwWgdHiTCSiSdy3IP0vElEsgf8ojL/U/EVtqXKD58mggOPPOM/2MmjP70sE581V2syZUCl1TbcJ6agi7se1O8Kn5BzdcF1YkzlU5xz8vZwnKb2PuD/igph9F4XyBAwRn1oSDf5k/8smRP+cfb1OOhPTPT9D/RpKXIv/dO5MEyNB8pUxZkp7mzH9H0mbbnZLs3Ft7ZWEMVvDI4QPQXNjnvCn9k/qnJG8F10Y127F9C7Rudlrz1p3chvY/UH+1vcAIFD8XgtoffYYaH6x+IWjOcMNfgO80gBoJiYl2x5A77I3XXxcnKs13ihbzf6BHHPY3NDv5UKhdc+IoPPpCLtQu4R/3reQ6oZtq9qvhAA++/upr8aF++OGHVgbWGMyYz5TDfpTmQGkZ6aDniJBVCK1M1q1ZZwR2uNelpsf+/enQ6qHGhnqwyq1EvKM4uJseCCLiQkUKgQIkzTp37Aj+m92wXDkBWpDNViG+gnXs3NEqQpsHVZR7b4IgCZUS7X+gGVIDNCKsP/v/F6AHWQ5THg4O1FYWLOdgUULulfKg+KBZ0EmALcehbUItGJL1qj6ucPzUs7RsefKpJ6wLSIczsHe/CfmlpWXYAWjznD4FImrUjfvrTABeDOSHffvddwDKlLdhHw2TNQzzZ5g2He63UQ83QIn3oGxeG8SWhAtz4AtsO+IHBMGuArhGLaeBg24S8bHE5JJCihR+PjkTonYTFUGocURFkgqwviHHTrky5XS+GV6v2rZtA5O1bXAs9I3deeddUv5g/hs3bnCaTVieaF5A6fyxzjLnVkA1Lz5oYYdk4sHtUqxYUfvo44+EqkVDPYvoFAMHA+3LyBZeBozYRNbo2prMx0WLFLVCMGkKA0hTqlQpCYrI2LmzF8TaHBZRQB2FNnGFYbd29gzUzOBfnERYp9Fxo6Ba1P2qK+G+7FENrBtvuhEDr5AdByhEIIZutouC8IeAEfM8RU0aIGYkBmJ5HFgTARbthbZq1UoQDp1WmYOPQAKBBAIJBBIIJPBLksDgvs3tvkHttd75dvpqEQJzfcZ3NN/lDP43rwqI4X1tBPCNxSWD4jMuNio8Dm10dMb7Lh4zcmsEd660kZ6XlRLy88u6x3LwD/kjouJqswEwiPnzFgIXcfy1kEHXeYBTPhPkH8if3UG9JFt/yOpj7i67qa4xJjua13908B/uf+rFrnt7hc2bX1lmTXClvR9mTfDWFDJryi5/byy7zRkr7uYfN/69umOH7DQI2EJZQEkW0OvaMWv85zxXE4faWFlktf//If/QPINHf678z2JPwr3Knl07rALAEz/4+Z/AhnnC2FHWq+8gAQU55z+z7yZ8bp26weMPwk/Vn16YatdvYhvAS8Na1anXVN8/Jf+frf5EQnIp8B1FAIUEvNSQOA9Zij8IwuC/KpUrgyqjhr3497/b0uXL4BDmKps5Yza0IxrIzIaARX5u7DMPIDY0J7Yn6b3INEilkQqwgiSxBG2iCkVJI2TmnFly49yyZUtRbSTCkzDbkP2S5jD14c66OjR3juP5klBUoGfhSZMni7OGe9JzZy7YuLHfaN9KBYL9MqlCx0U5FPBFIlvyo9RAOkePHYeTnHoWWTBSZkRnYaqTnJRsCQkJeDe7d8EoeKXqD1fa25OTZRLUF16OUBzbB82StTA9YizKSflj/83hcPDQAXCugOAX6U+bNh0g3zF4gDoCs6n9kIULSh2yYGD9+vbta/fd/zu5IK9bp441gJYJNYK2b9+O8lQC+LddMjp8+IjtSN5uo1Gu1LRUcMc+bt2vvFJtc+L4KaPXKqVIaxzUm8ALcYEImIFxvcEcCZTWhSy3JiVZ9+7dQNi7VppPz/71L1Lg4LpF2rSozNatW61jxw52Ajw9fJbjpGBUQQBGi8GrVEcKJaz/+nXrQTB8m4iSn/3LMyJRLlm8hPh7pk2dQd0b1dNPA0nlasilno+Ko1gESKgiNHPGTKiLHbBWLVtJY4Wd0F88Oa2VfHK/RY9KO7bvEClROyBUROeKAtihHR5ZodnAFSpUxKCpAtWlsiDyOaxBcBZIHn9Qo5bNhvXr5cedeRSLLgZioMrWpnUbi4MK1Vrw3IyDy+w18O5EAicibHv37BPhby14i4qByzKqf5GdmQmy0yySP/bMXBVykFgggUACgQQCCQQSuFwk4Dhn2uudOHYavDXBlbYWQFjEZP8FiO91nXuLLm1cscDhdQb/Pr/9X5K1cMRKLcscyaXBZ92TuOlWcnhKB6Fvrh994AcXEbA15cte+eNppsHVoh5zz/rlzVpjuCeD/LOWd1qcBvLPk/3PH2uuV+fdT6c5A1fa8SWhOXPIRmQza/LnAfZTfzy7mnL8Y5xr/PML5/zlPJsYnHZd9vnH3fzR+Ffr80nmkS0BHWbNJf9b/l5RQgn8HPlv3rBa2v11GzTzRZEj/y8/e98G3f4HUTf4U6pff/LEtGhDbzPwrCPJ5az/KfxgTguFtauWyDyqYkLVUNpu/s0p/5+j/mwNlDJbOS7ukG3KvpOWmo59aQtZcTiwAjALtFdW4wd5uoBm2ANHMp1gfUHgYdy4b+3WW2+F6ZLzysN0hg0fBoBst+KKugOyOwWzHIZiAEpKYANP2ZOyYwvMiKiFQ+sQctk0b94CfQ/gDNIJh3YJXWgTiMgEXyo9QJGDJSISHpGwL92ybYuzQME+9xSef+vtt+QW28dllCHSiQDfSnekHV0iWryqFbFnFvcNLFAaN2kM7Rb3bmf8MJD8rgQPK4GTb8aOBUfLXfbue+8BeCmkMrJc+K/xRqCERL3R8GRUAHmQTmTv3r22aMkirBsuQFkCJL4KOcej+hjS6Hfd9da5S2eYA7WUpg35Yw4fOoL8l0PzZKM0hsKRLsmTD8OM6Btop0yfPl3mQ/TSRN4dAjEMhLP43j8CzKAkTKKAykiGHM30JtW9aze5Dj8OwGXq5Kn2u3vvs0E3DYLm0nCBKv4YJRD0+Zh/KB22P+vLbwKbAuwAAHGIUARpqP8z8Obcr18/5BkLkuGZVrduXYEymIL0LMt2KULWm/tiUvd6Siw0U2KgphQGc6ZzKD79gpPbha68eI1Ev9VrVgfxTnFNLhQKF1PFoDFTBv7Ya1avAcSxgTVq0BDM2Z1B8Fsemi0R0nipCFUnel6izdsBgD50g01umtNAII8cOSpy4QhwzVDFKg0enpg/O9LChQuhwfOxLQbzNQc6J/UTJ0/L7i86OloaO+uBFKbs2mVHkc5pqGsR1QxCIIFAAoEEAgkEEvilSoDgzH2DO3CFYeMAzqze4DhntPnBNQZ+cvHDPwa+s/1FkS5wla5IbgPFeHoyv1vw8EzPeM/yHZwtNRz6KSs1Pc383aJZKSm+H4tp+X9KB3FVVBbDj86kdM4LQf6B/F3HUL/Jc/0ve6dmx87bQa60nyY4A29NMGvyvTVlzQmcEjCmsX7PHrgRYlBfxnEIvA2Nfzf/uFjuM3u/95/NMf/kEG1ohvm/5e9PNj9D/jRdWg931XXg4jpU/2z5z5k+wa646jrsgYplr77i7s9IBbnsQc+EibVnpb26evVfu3IxvOLus1p1G4MwtvTPWn/ykpJkmBoMx44eBrdNKjhd3Ibf8SvlqNK/faJao61onjMBHKSrV6+SfATmQ5ZHkb/rUxfE4bISJkU33zzIPvnkE3sdwArNZAphH8vAshJkUFuAr0daHNwzIh0CMrTqIJ8KndNQm4ZelBbiB3+G9jB36tnzanAFFZfGCJuiapUqNvTBh2zv7j3gW7lJ5SAH6969u6FVdtI2wZvwaXyTTPfzLz6XBynmxRAOhQQe5kc5ysL1diHwATVu1BRASpidgokPvUDRdMeHTqpWr2rX9+svAIhEtx988L716tnTxo4bL0UKvi9JxMuwZ88uWLIUEiASBq2UTBALx8bCDGl/poCMr776ysjxEupPekpi0CWWkNQg9MzM/CmLNu3awIzpENx+rxA4Qr4WypPaLpQdLVY+++xTKwtFjNkAQrpD7goETBD2wuSrP8ov3h9U/ALKFR1dAto+awXAEBQi/8vHAM96X9Mb96IB9syAgojjpGH9tm5LFqjE9uOPSQdQr7CISAGXvM9ycA7q1r2rLHi+++47cO4sFEnylCmT7dZbblVfYN9lHf15RgXMpY/cAWa8cR5ZMAqaLXHwR17NCkUWUgcjmkVtFrIoR4HnpWmTJta4cSMriPtuAcZ1HtXKIIwC+VDhIkBt4yD8G+3FF1+CTWURS9qWhM5wAJ3pjEiD98M2Lyk5CW7Ijlo3uD5jZ6FvedrAZWSmi+woeWuSJVRMQAOXAeByBH/QmIEgw4GOroX3p0+GDUe8beiEYWLG3gdNmt0YCNTiETFQLgk4SCaQQCCBQAKBBAIJXI4SkOYMzZqwxBg3A+AMCIH9LYpfXi0+/F+osGDRYhbf/MdFCV7d3uLEW/ZzF+V+eFISTmXcX74hvp8BDvxFDfNgxlwTuEN++xGRnJc/H3YLJ+82tGnc4po1QCp6BN9KUEkhPi66/7gQ5O+JmMIO5M9djXoFPthHPOHoV9/LoP9xQ/NLC1lmTeCcyfBdaYNLA7L35a/GyFZx/9dtXqLCnBMLDjAvaHxzk5Z90Gd7lvfZrH7aXhMrP6/53bOK52+Jcso9e/785d6l9fPkP2v6t9YKnpQiQO/AkD3/PeCUKQW+zTJly/9k/SeN+9S69+z3T+u/HyS/K5fPt5btrgDZazHE++f1P3bsCCwhMnLkn13+Bw/uB7iwI5vkTUDLqmXzbeHcqbZo3nT9LYb77SULZtpKXN+wdoXt3rUdfCR7sZ/D3gsbZJEN88f+HKohOZL9/z5hH/A1OhcumA9Tnh3e3EfzGPCMwnRI5il6dXDU5QeB7VmZ1tx4ww1Wp3ZdADoTYW5UBXGd6S4LkQ/9j5Qa9DoUhj3uhXzkPTHwn9ayJo2awHQqXV2VxM2Kj/RJPNz32r4CH3QRLUcLjlZtWsEUqYG1hIbJyROn1BZL4TGYwMwxmBHRDTb3sgQsXvz7i1IgaFC3HjhbqKGDuVwig9egQpF26vhJmD7t196WZlq+xkht5FMRDnAEuOCZ6bBuGTLkNmnd0BMRvTCxf1FJ9SAAlDVr10p5gYmTl2XPnr0o23HbiDJRY2bo7x+QWZirBz+99yvq6cBUHpjMpkgOTG2dGjWqi+eGAA/rVLduHSl0HEH/Yv8jwXC/6/vb0qXLbCRMnB599FGVh52Ye/7ly8C/A69SWntA/gSS9u/fD/ClNIpJsIV8sedgKbPGOnfuBOLjiVYLMh869A9S8kDtVFw9j/x27dpp9aAJQwoTmemhvBHhkfBsVcN6ALRie1PT58YBA2Qi9cbrr9krr74iCyHW0Z9XsmRw8UdhF5+ES4G+0GcAmaINHv2Xk9SXZDktW7USCdBRqHUVAPBSvHgxdJzCQklZIboLWw9SYHLAkGvGraioXpZfLNe0CZu/chkmj5VWEho5sbFxcId9Qr7ho6LOAdUcDJKgGUA137CmTZvAPXY1IZhENOPKlrZopEs/8W3bdoCq01YxcB9G406Y+C0GRrw1aNDA9mfst2PobCT/lV94D53LLdkE6QQSCCQQSCCQQCCBy1ECBGcY3hgxx8ZNXakFT8Pa+JWNoAuucxnDxZoOcOJ+J+Kpt2106xzEZWz+YWXK6DrFMoj3ccIvJoO1FG7jA9fcM0yHD/A/Y+GYz/KQgYstLKAE4igOfz1jfBfX31z5acgDA1eWSifIP5A/ukIe7X8aJxwLv7DgwJnr7bePj7Htu/bDrGmxDbq2BbTqI36yptxUc/yzL7tR700QnAa8J9xcglHPTSrX8Pjmn5s3GBHzDKYmjQfvKUbhoTZXTP8nc8e85eWPmC49xeODlz7/qtXrQOOlipdvzvzLlq9k5SokqDQ/rP/c2ZOtRdsrpE3hSKR/XP9li2bbLXc85Byg/Iv6fz9zIja1ceAs2WRduvfRHo6ARAqIgg8eyBCxMMmHqdVDcIYcLgSzYkqVkSZOJD074Tyn/Cl8BArdO9Q5P7jjzeXQsWNHeUNiytSc0fuN7YcLdC19AeY5LAxdIjN/anvcBdfMzz3/PzZy5Ahx0EyZMlUemL788ks9T1CAr5rJk6dYixbNAVzsEdhQARYeZ3Fj3oK5qHN+gAw0H/qNLDuYS8mYEtauXTtXQ74o8T8ijH3/gvXufY199umnIpzt1rUreGcmGnlaaGpVFI5xaDVC2X4/53vts1u1ag0PRimg4kgGsMAk80m7Y8HihXCV3U+ehC6cdwKmu/DFS5eCNsSRRxPEWDB/oQ2++WZ78vEnjC61pb2CvTotR0h2m56eaj2u6gHlCpYPyhNIgwANtXqu7n21fTt+gjRTOkAbSBVhERBctVy+lDG9HpeHXPhsi5bNrFKlSnBR/j2AokoY94Vt3ty5kOl5q1evEbRatth9991ny5cvx97+NWnpkBeW9ZsGcycqT/jtd/zYCcgTHEKQM8cA86Kn5hZtWkvT5dbbbrNWrdvCc9M9AL7q2+NPPGnrIMPQ8zDXIqizjFY13rqB+APvM08uZ9jO+2DG9fDDj9r8+QuM2kLk5UF2lyTkGjBDE6ApIC46AXvFm2+5BVou29SwhQHCFEkoAp/tka6SqAnRRdqK0W0bzZFSUrYLhIkBARLdWfP6KahIcTHGip89fQ7EQGnyvlQJ7sa2Ie00kC3RbIpsymHhBWwyULFVUI/q178/1MhirFGTRpYKtadwCJQNVq9+Hdj4dZGw6bedalU0jaJ6Wv64fEgzSczZ9CoVhEACgQQCCQQSCCTwa5EAwRmqcr816nsjITBDg5rluSNxK1eudhC0dcExF5tcuPhX9aL2NkH+4oZrFt/kgGrp/qLJrWW8FY0iuWOmx/e9znjMxAUIYbHl5aSLjKZiec/hHo+0IcNzXLRqk4KLQf4UImUTyD8v9j+64/2BVY/a85fwIXDmmf5292P/sB27M+GtaYEN7tsqC5zhHODNP5obVGnXn90k4aTAuYFtq/bFsUhBhbhw/GunSnxGEwuTdIEXOC7cNMNjN/8wFS/8ZP5ID2lIq8BL61LnXymxuiubipUzf815P1H/ozALolvtNh2u1LM/Vf/Z0MRp1Ky1A2X+Sf2P4wfrNSsWWkbaXmmWVEyoYls2rlb96T64PACjhCo1csjflweLm6PdKOv/o/zVBqG4qvhFfbBMH374ARzPlJE7ZJIAs2zcGzJUKF8RSgW8pm6iPlAKvCLbkpLw434J2wmLjGf/8qzdc89vje6smzZrah9/PEwdiFoaU7+bAr7UZIuDYxua5mSAn6YkrEXKwoQmPj5eHoB2794FDhpcg5mOy4U5I+B5nXvfNG3qd/31NmvmLKvYrKI1qNdQbp9ZHpL8UgvnGDhRyENDs59JUybZo488BuqQGBDfrlPZ+ePJLoA1n47+VGBQ06ZNXV9HBUk2TAUH7oPJiXPNtb2NgNMDf/i9zJV+/8Dv4eXoM3v6qadt7YZ1AN32W8NGjWSZwjp06tDRZs2ZJRMpemGav2CeHT18FN6sqlg51FUBDRgaR7yAurEP1qtXT5o23P9z/849eAU4DeJ6oiBIkwsBLyAR8vXXXWfTpk4Tr842WLaMHTfWxoAb5q677gYwMl+4AaXGP2rPNIIVDhtP2mTIuSA4c5o1aQLs4LToTHbtTrFnnn7annvueWAFk6AlNAQgD7SbNDDM3n7rbcw7kWr/kAYeADrKOSqiIMrP2lwQdvH1118DWGphY8eOwzVmy7rlqK2uX8xHrgAz7Mwk/o0HqLJ2zVqrDTfWFO4puCarXDkRfDCHRKREgqlwoIJxpUtBQ6a4pexIsQULFtihw4ehHnVCJEC0lSsAtGvxsrl2DI1HP+iss9QRoW7GlqDdXTpInHam7IQN4DCrA3WuG2640VbDRImusivEl4cKU2f7eNhwq5yYIPdn9JlOLZpYdG4idWTgzoSmDH3SL12+FGprdKPmBM9OEoRAAoEEAgkEEggk8GuRwOC+LbQAeXMkNGemeeBMLffLGl/CPvDhgzKUi5YjQkncAsVdw1X857ogpGnjCfFHixiurJgIvt2v0th48IWv6/jgfy18GI/XUQ4PK/KS1A84jMP8GPytVZA/ZCFgy8mFn4H8f7CIVj+jYC7P/uf2al7HzmrGX8yR45zpB82Zz23Hngwb/jV+vSc4E4VfqjXeOa5/0GasfTaR+Np0XL8z5Bj/TCNb+GFa/pzBKG7+QRpevv53zmc4/2TXWsFzmJAY53LK/5t/fGjXDbgbVfjp+tMkKQO8MqXivI10tvqfBMfL6mULQNhK7s5i1qhpW7ng9sXovwd+eJ6b9SeclpscM+wTL738EsxZhoKAt7lMlATasctARBPGj7dq4F9hYJ9g/hGREfDOsw5mYmXUN3aBA+YhcMEsWrRIJi3t27W3efPmIqYz81kP7ZK09HQ7dOCgvBLPhkZLrdq1YQZVx6YAuOG+c8rk76Rg0AKemtQyyh9H/jcLgJAIcKYYrEvWrF4r0IEWJcugPUKrErq/Pog8+FC/ftfbkNvvgPeiZTbm8zHWAuTCmzZsVH+kZiq9HrOe/pg4ADfV1DYh4TErSu/HVGSoDkuTXbt32uhPR9s9995r7Tt2tA/hPenFF/8OQGSMuGJLwRU467Jh0wbEry6y4Q0b1lsctKLKx4fbBChHNIApVm2YcRUpXlRyVPdDOdhn+EcHPMQKIvFXpEhhadqsRxpn4Eq7ffv20lbauHGTlYB1TN/r+oILaI1kRzCGNCNffP6FPf/C8/b6a6+7ekF03NvT8mUXQCOmz90767gH9CQExShn1p/txjxGwTyK3q8+/OAD27Jli9r/JHiOHLcRy5kftCinrQAUPvjDDr1FaxzhOuvH8U/34e+98y60bx5z95BHboZcQSAofAIc115zjdisN27ZLLstcsvQZ3sTmBjRL3j5+Aq2Y8cOIH77ZTJEtaboktG2DLZku/bsFoJHl2Dnz56xfVB3Wgq/6vuBbEYD5KlTr66ViokVuss+THtAAj2p8N8eBbvEl1952Z566kkQSB2z7Sk75G0pLXUf2LZXW+vWraDOtc0mTpgk9SOWtQw6WY3qNURGXJwg0a6dth1emfi8j7LnpqCDtAIJBBIIJBBIIJDA5SyBwX2a270DqWJ9AZoza+StST+K4CXva59kLz8XsTk2JVwMCAzgZkrJaGHkP6MFjn/C29zQMBEea3XKa1wWczHH5YnblPE5lz/jMmEXsj/PuMxUcZEmi6FLfmR8B/lnyY5iyS6/QP5eP7zM+p8rVbZO/As7pObMW09fbxVBCJwGQuARX83Hr+qnvbGabfxrnvB7qROC+i93oAo/Mf5xPXuPzzH+Ed31f+9pfx4Kadm4CYTP+HOKJhQX3eWIZ7Lmxcsj/1XLF1hNEPn6ZLU/Vf+F86ZZp67XqP5HQCexDh6ZZk/71mZNG2crFn9vdeD9ifebtuwAgCLyZ6+/3h3aYmcT9kUevv/eB/JG1AAusCdPnARekfrA+dkJAGBg/zdr1iyvLalJAUuNsxf0gz2BBPYTvgdpCkTzG5L4kj/11ltvs6t79hIQwHRo/nMEgFYyNG3o6YfKA+yx7du3w2Z+kfaf9GwkwJVZ40AksmokVxavY8E8p5Q4Usg/Q0WHw9CQ2b5ju8yB4ivEAyzKB2Bhqw0eOEgWKDR7Ytlq1q5lt98+BDwyUej75NA5q/qwe5MMmCZaZ6A0wfODIIamZk6HTh2lzUIToHMgDabFC6k+3nzjDXscLqxZxjAoTQwcNNAS8XzxEsVRzAt2GoAKXW+nwaJl6ZIlMHtKtz89+jA4X8BF5Hc877sQzJV8DSXWnzw69LxEsOkIHPVQg4cemI9BuWIvTMISE53mEOVB78k0s3r55ZftlVdeseEjhyt/1oF1HDFihG0G7w0DvaxzDXHyJF1iu8xVFvyoQ46a/rCqodnWgBsGwMKmn3AAPYe02IYEBIktnMO8QvMttr3mANymYodANZh67Qe4qdmFz+VyyBVgRrVBweIBvFBbhiZF33wzFsIbYathy5UJVSOqKBUCQrYBqlFj/jFGqFVq6l5UPBKET+HqdKl790nAVBEqWqQIVIuixP3C7yqVq1pCQoJcYhUrWtSKIL18UKWjwPEl86eEhMrytb5g4XykM1LqX1T3ohYN3X4thJsvqpsRmKkENLBUXIzR1KoQ1J5OnzplqfvAhA27MZpXBSGQQCCBQAKBBAIJ/NokMLhPC7v3pvZYpJ6z8TNWA5zZ5V7xfNdCGFyk+UEaGbzOG1rYcKOi3zu1oOGi1Oekcc9krWL4DBdqbvGE67rFMyaGUyyu/I2UFmDKBL+6KX2Xjoh/GR1/usJk8BwX0dp0BfkH8nc9Q33K6yU6zgv9jwPLHw9eBX6RXzJreqqfVSofC29NR2wENGeOH3cEqKxw9vGvE08KGv/cxXD8k9QCEd34hxaLF+efzj94TvMP549sIWvO4YbMbZGy5/+j+ecyyv8M9jErlnxvzVt2+qf1Z/nLlU8A6W6SLV04y7YnbYS1Q6K179LLOl7R21q16wYAghoP/8H6C2zzWzBb41zEITlPv58zxxo1bCTnMY898ah17thZHpKYrAA2jDe6goY9nEybuGcl+Ss7GP9x3zhw4ECb9/1cu+OOu6CRkgL32R9DA+cx7XGRhTb1+0FsWxh7WP7gvxKehmjlkYS9Z5kycVBa2K3+egTmZlRKYHuQk0cX2ReRhn4MwQHNctlHC8I7EkEeUoZsWL8eZjpNoekSC+WFJdbtyq7W/4b+4Fhtbg/+YSj2sHvssceesNcBqpSENsxZvMfZfzlGUgDsVIN2DF2BnwfPztkzZ2X6VxwaOTzGT7kAAEAASURBVBXKV5ALb8alBQmVGQ4CtPv9738PPpz24m0hYNEYZk21sM+nUgTf7iViouUV6Sw4WvdAqyi+XFnwzc6EB6kvQn0QuWOfnh+WMkVkDsSygDlZ+ScBxIqEdtJWgEyrVq+WS2/ux8nzQi0kAkIsE5Um/vb3v1nvq3vbFZ272CMw3yKRMMEtEgYvW7bCJYukz5w5p7F7DmsBTgusPAGbMnAsVAIg1/889z/295f+BhLiBLnp7tSpk57lfKsHkN9iaNhQfuSdze9Z6+RHHWbNnAmKlDCUiQkjotJnHrkX2OMuPqCA7Fyp0FChIMjTwgZaAPWj5cuW2SKo/bARCwMAKV2mHDRmMmwdOlcyTJmqV69uQx94EC7Dqlp+uPg6CY6aFatWQH2srFxlkVB4CTrv0WNHLAoADV2RkVw4phSJgMkxAyLgw8ckH07I7IR0X7YPIM8RmEjt3bvHRgwfrsEWD5fcEZEF0enzi+ynfMVKUtGqW7uedb2iK2wlEzG4CqouFy+UIIVAAoEEAgkEEggkkPckMAiaM78d2AG/AF6wsdNW2or1O7z1h/OKxBpxAyMvBlibcDHDcx76wS0u3SJT17CA8bVgBNggMt/ZDG4zlP1pXdXClHG5SfAD09VdfHHhylUbF8RcODMwJv+YJq/48XkS5M8lnyQK2VF0gfwhkMu2/6l7s4C/guATAlcqD1faGdCcAThzDOCMAFavn7pRzlGNMY9/wmjRhfNp/nFC4h32cG+agN6Fmxd4V+Nf/Z6bXvZ/zRRKyz2d89P3BuePE+bP/QPD5Zj/pG8/tZ59Bv3L+m+EJ6Ta9ZpY3YYtrVmrTlYP3yVKlspZce/sP1Z/v/F+slT/3kW9C7ChJhjR/cor7WsoD1xz7TX2+qtvSGuC+2y2s7RokH9ycrK1bdMWbqfPQJ7O+x85YtathXlTuTLgRt1hEydNsT//8WF7DCDBl19+IW4TdlW+ewoVLCgX1jthJbJ+/QaAHmE25/t5ABiO2hY4tVkJa44vAV6sX7dBoMNZACUM5y6ck6kRuVjZO/nHFxxdbJctEw/Q4Sz2zskw9yuI/fIZ252yx6agHN26dZXp1I7tKaD26G89e1xlc6FBUzkhwYFNSOYs3uevvPwKgKlk5GPSqqpft4EtWbzEogpz73sOyhDHYK4DT0TYP5OoNwXmUKQfiS8fj/1yCQFYhYsWtishw3JQxti7e58VL1rcmjRpbAcOHhCFCelJloFk+BvwsagGyMuNQqd1FE5NFO8KNWcOZh60g0cOyRlQDBQmaGpVBvw8teFquwDMkuS5GRpCYQXCQTb8rV3V8yp58Xrt9TdtyK1DNP4JWkleaEhSo8TElGSD4h7nAjcLxINL6MGhD0Ib5rwtWbbcXnvtNWAFqXJcNOiWwfJO5StqLUX5abJEl+Bsf9cloXVz5KiFR5J8mJAvOw0bKHdD7gAzKBsbdA/MkSiIq6++2qrARo5qP/kh4AO0uYOvcKos9YT7qZ69eoD7JZGrKhETtW3XRo1aDqpTXaGOVQs2atRyYVizZo2NHDFctmXpaanoWGdl+0e32lQrO3HiqO0DILRu/VocH7c6dWqj8YhmucUclx3sXBQq7e9atGghDZ7Vq1bZYXDfMF6NmtXtXtjV3Tz4ZtnzUaMmCIEEAgkEEggkEEjg1yoBmjXdN7A9qn/Bxs90Zk1uS5+1EtFiBadZ153Kt1sgUXLYJHnvYi5gSMzJaz5owxh6VmnwzAVeY+DmR/9kl5T9Hl7ouqMlh9Lw7/pADLJVnCB/SJPCwP9A/nmt/6GXE3X8lQQHzvSDWVNJgTMjAc6cPHnGcSVJBpCFhj62Lvjm/JM19zAC5h9FcPOCu8JjJ0MBuXjGaeopQT3j5qAsOYfmH+WBGSg0/7jcXOaXV/5HYJZyGvwwcXHlVDGvyhz2ofrLGgB14o/n+PKCJ0H/AcbXU65+Arh/7vrncp9n+xNQK1mypN1wY39ZSHw6arQNA4Fvh07tRS7bvkMHJw9MEQRn0uB1iHXPlFkO+gD7AuRAs6ca2KPyHlw52Xvvv2dXdO0Cl9c1QaI7WYS6vmQjYAYlWSNu7Zq1td8tFRcLYtuptiN5u7wIr123BoDNHJj/ZKqdloPCYwM0RZh+BEyL9HMHjqOgVbJw0UJwsh6Sdgg5WiMiClgGypcKviCaAnEfzue+m/KdNWvW3JIBHtGLUM1aNeQimy178BD7ySntv0mGXBpaPIuXLAZgdQ4aOSDLRfxMuPcm/+vRw0dgzXJQGj98tjS8F5UGwTFGkDRPBkCW1/S5Bl6htsPr0pUws4oXwTKVIujEZ826deKYpcMeBtfFLgi0IrDEvTZBD3bGkiADpgvuigBPykLrhqZ4JAY+CzfWbD9US6ZPN8B9+YoVq+ydd96xb8d9Y2+/8zb+3tJ95sF4p2GOVRLADEEdJs5rXEfQLO2rr780aggRfzgOombiEXQ//tqrr8tUKrpYtMqDR0Q0DCRKdWH9z0G+LGs43GmrTLjmjxXmnVshd4AZlIamQgRAzgKJqgiSo7r16qBjNLX2bdsCLKmrwhPRosuvJlDDovsvTo60yaMGzDEIKAqNVLN6Ddm7TZ823YYNG2Yx0SXA3lxKDXbo0GEgf2fUOcgy3bB+Q9yLRWektk66kLoBNw205gBfwqAGRf/uEQCGiL6xs9KFN68dAeL1/nvvgaX7Q9nkkTyoaLGism9rUK8+XJLRbXcQAgkEEggkEEggkMCvVwLUnLl3YEesP8+DEHiVrd4IsyYv+NwK/qKHCx/+MsVfkbSZ5MqGAREYR4dc1XA1gN2Cu8SnXPB/ic6K5667BbC2F95D7gm+9xVXiXup4Fi/YuEzFIL8A/n73SOv9T90Y6cyH+rNv/gDgjPvPN3fEnzOma8Xga/jtKu3147cFOUc/55YOP94cbILSnMILvi33KbKi+FNFdyP+CFrVuL0haf86cRPABEvt/wnf/uZXXXNTf+y/pvWr7QatRt6m8rLuP4/1Yh+4/wb32xD/pUoEY3NeUGYsFQS98radWtF6HoC5K+zwbVy04ABkh81sKhV8dqrr4LiIl1aIvlgznIO/Kdsd3oz4vuP7zv2GwIFPXr0tM0wx6kGKxACNRGg6LhAEyXkS29EjWACdAggB60+SLa7cuVK7WVpokTtFJpHsYyHjhy2v/7lrwJrWFU6zGF/RHbI+6z2p/RqRKJb7oELhBUwusymCfHa9euVH7tyJsyV+l93vX340UdGU52/A3wgWMHAbswxQDCnJsySCFikgyeGHqU2Q5uHChFxcXHQDloLq5M0wDAcG5ThBRH2EjApQPMeXOverbs988yz9u2341EWardcABawC3WLsyOgElkLDSMSE0+d+h1FoWf0iZM2bdqAO7aUZEosjrfXgQyYaROY2bJpsxXIB3AF9/hej8H+nS6vS5YoCYuYI/bG62/YgBsH2E3Y98+cMVP8Nyon0vkI+/vMDHjaYo4uYwAs+UGnslGcPZoDcJ3gzCOPPGJ33H67kafnJvDotKY8+Rz+6Gqc/4DKQRPnuNqeuIVb/1AmqhRi5l7ImokuIk1WgJWbDKbj7/BHN130Bc8OVRbmQyRbonnR8uUroba1XkRFkRgc/PWGwikMuzN2hPFgxp4z93u458pQJ1kEFK9S5US75dZbIPibpMJFOzYCOHXBDn3LbbfIpVe37t2B0pWWIMugM8WDcDg/tGaqVa1uvaC907t3b3l8ogoYhck8jx47Lru01bBpS8/MEElxTEyMRcGUKXCZfRGdIXg0kEAggUACgQR+MRIYdG0z++1N7fHeNBs3dZWtWr9TdeOvh1yU8P3PZZP+MRK1Yriwdje0uHFLHLdA8i7jdzJ35MCcbM9nT4+pap3Ai16azAPB8ctkLYqYggrJ/L1cGc/F8D5D5cUiN8if4kEbEkwL5C8JeH2LnTf07z/Y/9R5c1l7QI1+mX8QnHkThMAVAM6kZhySK20SAqNRQiJxWi/Zxj/bjrvXUBu6SmbfN2m+8u6Hxj/mMb/tuTfwZwx3jcnhyE+Tt73jyyn/A5npoHooBM5M/Kjsl9VrY7/+jisD2ooeAHVZ158yz2paryYX98W2J7ntOmhx0JMv25eyoAYJTZbeePNNe+OtN60pFAcqwfMROxqb+ww0NvxA64/iRQHKYH/L592GPZ8czBBI+P3994tYt3fva2zwoJvRJgXVd8hRwryefOJJeA9eY+UrVBSgQtLc1Iw0KBWQT8m9S0/BtIcgy/vwGjRi1AgR9JKEtmyZsgAwXP8ujLYm30lF0HEQCGncuDGIgeH0Bl6XuKlnyWpCq4d74ydAmjsTvCi33nqbzZk1R56KeJ/lJxnuE08/CQ6ZdvCAtFpmW+lp6bZvN/hfAdbsBhXIcngtfvfddwHanEDKrlEKYf9eKg7mbywOrtIS5frrr7d6ULjYAhLeU6fP2Px5C+Slagl4cI4eARHyzl1GbaCjx0A9woCk6Dr87vvugbxRD9SZUiWXDBUnvp83V+7Hw8ILOOAG8omEF6l3oSlTDs+x/AWwz/8CJmTdunbDWT6QOk+0WjVrIO18cjI0d948lxXecZQvwSxqBXGsqP1wrQAUNliOL7/80q7o1lVx5s2da9fB4xXToRkZn2XYvHkLgKhucFB0TvgFErkk4aKBGU5YKjK+6XKKxEa7wah8FIjeYSB/RNzoN/0EVOzI9zJ/wXzZ7nFRRTfaqVAXOwYf6OwMbEgKZ/To0SLkjYOrbHbYClCPooszooRH0JHoK303zKbSM9Jh+xZnreB6rHatWugh50Ho85JUzSKhqkctmDpwV0ZBsixToT62B2XwWaopVE26BHvQCaNhP1cCGjpa8F0ScQeJBhIIJBBIIJBAIIG8JQG60r4XZk38MWXc9NXSnNG7E4sphuzrEy5iuCbgwkf3sDZwAVfwvG9qpFi459YQOEM0b/2D6Lzr/WN6WC8w8D4X7Prms1yoKnAdwjj4VCLMPcifoqF8XaCsA/nnpf7HUeR6vteEv6KvmOgi9s4z/SwBnDNp+0EI/NVC7CPI94FRjp0L+7UzzXNC0bj35gZeCY3/0BzhnvHX95ovss8/2eYOPs/7nErcn2sF5qH8cXo55T998tfWrWc/Flrl/an6b5S2TKPQ/uZyrr/eIf60xcpcZGBSBEbYX2guQ35SabxQYLhJbc2hDwy1O++4U16P6CKaWhMKrun13toBpYPOXbpIw0WyxnupAPpELH7UJ41GGvak9Lp0ClYY9RvUtwdAnEvtHPRUJRUNjZ1HH3lY5fjzn/4Mgt0kOw0gZtvWbSD33QVu1t2WvD1J2irR4G2ZMnmKvQovRPRKdMcdtyPNBlIy8POmG+5ZM2ZYRWiwbIE35DP0wIR6ssiRUZEqGwG5TZs32ZDbbgMhcWFp6vTq1VN1pkwy0/fb5EmTYbJ0RE53qlSpDD6Z8tq7U6PmMEynqFk0btxYaZtkl4nKwQsYR2VhFlUbChOdOneWqdIxeKXq2aOHuGaKgJOGfLKjPh1lv7vvPiWhgQTZF44qZHfffbe1ad1WShf+eHsJrrpnz54NUIbEyAgcq/jan3lAfDu+WRHrumDhAhs06Cb7ZuxYuwp5vvD881oHnAI/jJrPA7epGRUBkItzgNofmZUD4HXXnXfBEVEUPExth6XOx/b222/b8GGf2HPPPSelD80tyJ/ux+fB1IumUkrY6xssXm6GiwZmVGCWCIggTYBiS5USoW4kkMJTMF1aDoKdr7762v7x6WdwIZYs9SSpAaGSBFamTZsGG713BJwULV5EIM7WbVsRrwj8vieINClz/0GALMWsSNEiyOi8VKnok/ylF1+yV0HeQxu6AzBzYrrvvf8+fMvPs6o1qltldLDj0Iyh9g5tKzdDLYrqWiR/KgbUs6C0Y2DDl99N8BfAEk0G7lCdclPSQVqBBAIJBBIIJBBIII9KgGZN9w3qoAWSzJrorYmrKAT/nclvvNq1DPWvuRWMqzS1M8TzoIf0ID643OKnD664M/95pscovM9D5shr2lj4C6Ns5eA9xvM34EH+vpAgiUD+ear/CZaBKv2vNVBz5i16a4qHt6b9h23k1wskCjcPQC7euPe//fnHjX9ve+PHwUXel9aBL1D/Hs7d/OPfcHF5hr28guYVfPh5uPnHaxsvHf/ez5n/QWj8Fy8RI22Cf5b/WZjgcBNO3o+8UH/KDy2gz9z4UEpIlD/0c2NOOai9sUFn0904YIBFFowEl+nnIPSdCMWAwtpH/uY3vwG44Uh4GfHYMRD3btti3KPyJUcQhO+zI9DyICksNTgm4PnjcGJDhzXlYDFy9933wBNSdfdSQmXoobhWzZpGkIYcKzGxJaSIsHb9GliUrINiwS4pHzSDMgJ5Wjes32TTAb7MnjMH5jZD5BSHoOAFsPeeRH2WLF1qw0eMFH+O3q0oJ7ssf8xYtGixTLc47ycDdHjggQfEHTNq1Cgb+uBQ7cfBnCKnO1Se4P43DLws3LsfRZ1OY98M6El1LIL9/TPPPGW7dnvmzJCnaydmBiHijz+a0ELljrvuEnVIRWgGVYbb62h4QqJGDkEx8s9uhbkUzwWs6hcZs169e1mPq3pYseLFZUaGBI0A2dkz0KThuEPZzpwBNw5yJTHyeShUMP/8aMs+ffraboBaf/zjH23ylCk29KGHbNLkSRZXKs6tFVB/RiZoRh5bas2o/VHm2NiStmjxIqtdt7bWDKzz7yGn++77nVxrf/zRJ1YUZXJ1NJszZ7bwBD9/FCHXw0UDM6ES4ZezCGip0HSpKgRPLZaqVasB4T5pmWBq3ggyI3bo9m3h/x2NRH/nBEvoRmzhoiUiNKqJztqhY0eYIFUR4zKRsv0gNjp8FOpjAGZojxYbU1oNQF6YJCCL344dZy+88Lw9/eSTtmL5CsRpC3W+QnKnnZCQKO2Zhx99BKhlRWvUBGgxFb3QYUk8RFfZJBBOTcuAbeBm2ORlqgNk/cIUql1wEEggkEAggUACgQR+1RIYdC29NUFzBu/7cdNh1rTBmTXxnek2BW61poUwJKV3qU64hCJoo62mZOjHwYorJNMQaIO4PoeDW/45UmGm4QeXoneu/GnJwPxdXHeI8yB/iCyQf17sfxgxKDbb7tcbHCHw9dKcSU0/6gQBkeQAUqAVw2Huxr8DZTnXuPnHmyPwpfmHm0gEzlfegfvGZ9b8g2NPG9CfcrAvRPCewdHlkv/8OVOsY5er/2X9N61baTXreNwyeaT+aAEKPJcC3gngHU0HF2nlypUF0KhvMHW8fw6BOJnkteSVOQbHMy2xfx0OpzMjho+wZ//yNHhQCwvYIvIwGqTBe8Gnqq0kpM509lFLBiDJOYAF6DhGTY2VK1ch3cPy1NS4UUPXdX4wlO+44w7wvMB78L69ds0119i06dNhMnRMwEMGTJzolahcfBm5r74PmiZPP/0X7YNvufkWKwIKEHRY9eNx0BSZM3sOjgGsoKOyl/L9dwycKOEgDSa6GF0y2jq272ivQAOH3ohuvvlmATVl4HSH2ifsFmeh3ND76l4AVQDOYO/etcsV0LwJFycszaug4yAuF05L9KAcmpp+UK8qkPGL0HhZvXYNPIDVAY9MjBQxihUpJlBk/vz5tnDhQuzdX0BSbkzxszZMoX7Ttw94aJvAsuUMWwf3yedCr0j57OABB3aJkgSZs6Z0f52YkACwC1YzON8AcKt923bin502Yzo8OPXA86hfgXy2b88+AWPk9GFgmvnh6YkWODVhAsVqAB5QGT/5ZJhd3etqxI8CdUpt69WrFwrDUlK2iIn5IFR/Xc29j7DcSuosSHmPAk08jcaKQEdITEhUR120cLFAFrq+oocmTnZRULHah45IXpqqAGFoJ7cbKGFkRKR8rFcAyvbqq6/IdImCIjpGMiCSCVGgRLcKFQZjM/yu01MTARUSCzPdJx5/HKpIH1nStmQBNARg6I+977XXwjVaH1swbz4QwTC5LNsJhI0ETTS7IvcNwR42eCxMqJBlEAIJBBIIJBBIIJBAIIFsEqC3Joa3Rs6x8TBr4mKoQc14LWr0KzI0ULmc4spQmx/vm2sZBl7jwgZLARzjgm7wl0vufLxIXjwkjf+Mz2j+Pbc4QhQ9r3s8wYHyV/q6oLx0PXQtyD+Qf97pf+zxWZpf7NO/zuBrztzz+BjbsWe/5oLHXhjFmQH/NIl4gvGOZe+ISxIgxjy+uc3XzKEo7jk+7131ns/55eSPZPSwFzdbdv/p/KE2YWkpq23BZpaN4cf1P3/hjB0/lGqFp2zOY/VXa7lqXfQnx7zZUYAe9Cq0CsS7YVAkoElLGDQuUveliojXvWPgEhmUGTRrIkfJM08+JY7U8XDTPHXqNIlYZjRIj68s7k+joG1D85YLaI+k5CTsP5O0pzxx7IQ1a94MNBnFnQluqB6uL5Eb9bHHHrPXX39dfbVPnz7SEuE+NCEhUR6XqHXCcwIVr732CsyKVtsjDz9ibVq1sUUw4Zk2fZq0U1J2psBqBd6GIDa9M/GCLYACynQP2VE75+jxo9a6TWsQ9Y4T8e+LL/4NgM5sGzJkCLRrFml/XgTek1u1aqX7qemp2qfHlytv27CPjoK20cFDB2Th8jUsYVq1amnVa5DThfVx+aoiGH8EwAaD75WyojkWy3b65GlbumwpFCOipI1Db1E7tm+Xtg+tWFj2MPAA0YSJnpK/grttAlWsP9PNhHbY7XfeJvfg1BgiGXGRIkVs46YN1hjuutdvgDcrtOnceXPthv797SFozowcMULAFhM/i7KMhjkVzbsIwmFZoHvUJqOihuqgdYjJModExD0A7GzZvNU+++xTAGPP2DvvviPqFfLVuHrnZj9VEYj55E4gse5e+AMnsdKqlatFGkRtmHMAbEjmQ7Kj4uB8mTh5IlSz1gO0OS2S3diYWOdmq2IFLaK4aIiMCLOTAFroR17zJoqZCVu3zZs3O9MkdFCSAJVDoxeLjoZHpxPwKb8T6OQReV1iB+AALAUi4Hbt24o1+gzIes4gzdhSsepM9/72PqBhveQRimTCRP/S0tNAGJQB1akssqfckU6QSiCBQAKBBAIJBBL4ZUiA4Aw5Z85hx+K8Ne12i0G8sLmJcZslf8EC7Re3buNdCYALR77rtRLjagzXs/givMiMiUOp3ispP72s+z5WI0BG6eARXswRP8g/kD87E7uT6zt5pf+xnAQwgwB3ujBrevtpcM7AlbY3wLHZJTTCwe7/UVK45k8V/EaTMw5jKmjSUA/QVUXFhw/8Mh7dJTP48mc+2oXjsktHT/3H8z+Ysd1KlHOaAiyZCznrfzRzjxUpgf1VHqy/V6GL/mJrhsN7ETm2hn38sTwFURzklnGaIOekSMCMJCds7mnR8dEHH1nnzl1kyjJ23Djr1bOXJ2XEQj+ieQ27UwSUCrgnpaZJXKnStgeKB+RP3bJlkyg21q0HYMAM2UR8IFsHjYELb5oV0dNT1apVrRT4b07BZOcYuFlLxcYhD5gS4VkCSTElY2z6tKkwsbnBknYk2SOPPmZ3gB/l5EmSB9NLErJgnZAPOVkT4TyHoAb7/xmY7xAsOQy+msTEyiKypeZPIRDhfgrQ4Xe/+51Mm/gCD8Me/Morr4SFSimUKxqekjbAsiRdFi5UZNgKThy66R42/BNowuxnrt6fShCqXlTBKFml9IDcDsBLFF19N23aBO61U6wUaE8KFyxiS5csAVgy2o5Du4dhxfJlNvv72fbGG2/Y8889bwSFzknWBdRGkyZOgMVNhrLUuETe9ApdsmSswxA8WINA0v3/db/98b//CC7a0gBohltBlIckxqdBTuy3P+lNaFaWffznK5BfJmMtW7QUcTE9Vf33H/9kd991N7w9fWANGjQUiOXqrWLn6keu6YXQ7IiqT+wMK1Ysl50cG4WejkhcxM4YBvu7ZeCcoS0fma1ZOfK7ENBh5yCQw8VaeHiUNW3cxMIgYKpihWFAnYA2DtFKdqyWLVuDZHgX/MynySTqwv9j7z3Aq6rTte8HQggdAqEmhITee0eQjgXsYkUBy9hFHftYUUcdu44NewOsCCpFERSUIr2X0ENLCAECCSmQ93c/O3E851zfd97rInPm5cx/aXZ2dvZe5V4rZD937nJsJ/VcOy3Dq7EKuXiru1qmgB8sqV9kgdoIsCmwXo0JIBKBkwnrV6tWHatAGFIcPjRdtIt+W0SAU6vf30SWKNJhZQGBgEBAICAQEPhfgoAyZ3iLaa+hnJk8c5m/sWnfKpHf4f/17Yoe05MRDjsBo4HHFz7pjaN+7/tTeJ/gCgG+1htmPVj8PdkM/kHm6C2FvtZa/Ma/1l++9F7DdyyyhchmfOVh+wH/k+z640KOXPN/uJj/je+KnJnw4uh/YwT+46HPnj3b+vbt+x8f/MNXCm9VbESTJk3+8Oi/313983/jDTfaddff4E1BImTaUjUtwkP2Hc2mkd8nIkL1ryTVzMyONfhDvsQG41BJtGrd2r746gtXTag1yJ8PA6I/5C9YMB/CoYs/1qJFc1duKMM0L6+A2XSXTaFxeCvRG2dCUIhM8F9H+uUJAViA8kJqDalU9PurceNmHgC8jUyYGjVroAhBycHvQsV5yO40gxIbqUzeHve2XXDBBfbXxx+j5KY5ddjP+ayt359a/+o1q02Wp4ULF9I0dMwFEj9SKX0076jnwGZlkSGDyGFjykYCbuvYfffd59dSKWXqoEYpi3KlWbMm3oAktchuZuwmTZsQV0IwMGHAB2ioGth/oK1eucIfr1cv3v+t8sBdjkNZPsVLe0KLqyOieO2NNxzXDu07Mu/n2aYtm7wpS5Eiqg/v2bMns/gx+/jDj23j2g02+urRND51RdG0O/J+gBUu+m2J49QJhYxqzbPhBvQ+wI+76NhFkJ1JwPEPcAbj3hoH8fVna9ywsc2d+4uNuX2MzcH25W8TeJ0Ip8suv9zzZ7S/ytTRsSdwPFWrVmG/lnI9FHpT1Px58+zvr75q06dPs/GfjI+8zxDYJbyUkGKGN1OloVG4uCUjErulFqQsAnnVhZ6U3NCZMqlYZDeSvEqWpkYNG3guzc8/zfZWpaacdDU7HUcOFh8fb737nupql/iEeIvmJCuYSLXbLfF7nXLKKTZg0EDr1fsUTnQNQpcUWJTv7GQ1iBn1tc+dMwdJ2SYUOGWdjctCxlW/foKtX7eO0OAXSaL+xr2C6qaXHE0MmydbB8VMCV9mYXUBgYBAQCAg8L8NASlnlDmjNzlTflzpmTOMkryN0bvOPyz+pW6QD+thv9EzuVv0xkaf9aFn+ev9G5F1Fb9p/uNaI01NkRf73wT9+dwrWo828/sStg8UAf+T7vrjYv4vP0u/X9Thzr8zAso0Ub7G/98il8G/OylTjM+o0aNp/f3cVSf++4LsEhE0+g2ikN5oYi0iv3kiErWGTRrbhRdc6NEcCtm98YYb3EI0Zsyt9hZV1s1pAtbvGv2run//ASI2foNkKbCtqEEkSshA1ZEC6bGJGTQzcz+NwLttEUG9kd9Y+l2n33+yUhWN4f4NojQgY6S2qQ4pJPWp3ByRX4yFqExqed7LkCFDXK2qDJXnnn+OQNvB9s6771l3rEURFRkzOK3HEydMQHmT7fuplmJlvdaIre62n7zcbC/CUSDuNqq2tZ89e/YgU6asq1w8c5wDVJbN3ffcY/379cPSs9GbjhUf0qRZU0tHNbNt+zZEGfm2bLlszSh1UPp8+tmnnrvjbwz0IMeWmNTA7r33HqtcsZL/cQa9hYO3d88eVxpt2brFH8giNFn8wFwanO+5+14aouLttNPP8O2rPVmLvi/chf9+sI2PT8CmxHsL+Aftt85lPI6a3vAEUhBlZGbYFSOvpFVqjX028VNXKMFUFCmmjnnezRq+5zvE6RAZTrKN7c/IjFwXfC1yax7ETN9TT/XPY24b49vX/pT0UkLEDCcPxqsnF0U8feDyb/2eWAyZEk2mSx4Xtuq49AMgW1MlTo5OoABIR+lSn954KWjSqe5KIzdGkq7q1WORUVW2alVivearb/9+KHIOkYr8s7NxlcmZaQOL2axZc1fJ5MEEbkZiJalWGYiizVu2+A+JSJ1NXFAzf/gRxmwupNFhW792vU2Hefz2229dlqUO+uSkZJqfqkAeRU5+SYMd1hcQCAgEBAICAYH/TQhIOXPjZZAzvDGa/P0yS9m6V+/DIm+c/DbyvtLfnfEdz5LRuyo9xz/r2XqzpS8i991THrmrV/j39UljavFS9Kh/WXzfCRwe0efi//QE3mbo1j/C9nnbF/DXBfH//vWnIcGvW9/dcBMQ+B2Bragm2rRp8/vX//mOWmoSEhL+88P/tl/feedd1pR2JClOBg0YJImFz5/6p1CuDrk2eJB/EyLEjCq158z52RLi6/uvjlz+YK/g3LOGnYV6ppV98tGHNA4lF+FZaAcpuVF1dmUyT+rWrusNxR0IsVUor3439iQT5vPPP7eHH3m06N/fyG8t2aGUoSr1h7avTBgRM4cJIa5cWQG/7CePZR48gOpmi4sIRMhJxRGNoODnn+fYo6xTVq0PP/jAaiOGkOJFx7WH9eSz3xr08/OPIUyozyxd2ZYtWcb6DnqejXJbJ2PTkrpqE9ms2jk5U5TpGiXSiO2L1Ln6mmvsoosvcmVL586drX58Pa/33r17j/9DumL5Mn+tnDPbt263iTQqrVoD2fGH3zXKkRFB1h61UjaZrpr9q+OqUb5NPuSO5n/lz6qNqiE2rIMc8/33/sXJlyee/Ks9/PDDVheOwRdcOkJMYg2tV6HNvIngWFVhHuXKqHLM9dqBUry32Lgxxa6//noqsJ+yZ55+xu4lp0eLdk+8wKqVq7gnpIgbhoTRZ5GfFcpXdPx1eiQ00XIpLV5fffHFP+1fZu1FiSylOYEJnPTEBkkE9Ga51UhJx/Kj5ZEnI/+4By0hEaqMNEss3Lp166067J3kS2LA9JGNqkUSqXbt29PO1NiZvK3btmBbSrdy0TFulXr1tVeBTJVZZeln3+CMlsvG8vO9VmwhzGUWJ0knVw1Nuujlh5s9e5Y98de/ophZa7Xr1ob4qeH7pqAnBQj9GbnTOUjFpJ4JS0AgIBAQCAgEBAIC/z0CImeKlTP7MtWcghycNzL8UdBJERElkbdA3BaRJkVsye+P/xfLRtEbusjWeXPqX+utmBaFB0eeECFdIo9GbvUcbT/yGn0/bD/gf1Jef/6z8sdrO9wPCEQQaEboqls3/z8AkYVGfwgPixAoJFB2tl0w/EJ3UXz48Yd2/vkX+myo786cOZOYjcX+PJ6q/92hIYKkVq2aTuCWZjLXjKrymQ7tO9j69RsJhj3TxowZ40GxWo9+JQ2E9OncpbMTGO3atqVVqR6ukWRXa+iPAirIkSpE24jc8JnfUVEQK/pdxW8tD6ZdtXKlr+PYcYJqsTsV0vikwhzNqrL+yCUi0UGnzp0gYPZA+Dxis2bN9hDjp//2tNWpW0db0KpdXXMA4qh37z6IGGJt245tnolzDFJDvyer4zKRCmbWrJn+Cv26roWVKykpiWybKN++9vW8887DjnUmhFFlJ1GimJUVOpyauoNmo13M54dxoeyxDPJklJHz9aRJ9hN2Ox2ojk3HrP1pROnPMGq19TtdWT7KmxFH8NtvC3z+V6BwCwQX3bp1A4sC5wr+9uTTHlI8ARWQXDMimooAs48+/ohQ4cU8t2gLbESkSgEuGm1P69b2D7NPL7zwvA07e5i1htS8nirzDh068IzIfmgPpaLKR1yi1+k1Hp7MV6Xg6yQYESmj2JUp33zrx+MvLuGbkmEgwEK94Fkwb2LnjnMBq35sDxIlNS2JzTruCco60CLg9BqxUVxgAwf29wrsbdt2+ImVYiUaAqdKlarIv3YR1LSSiyaDdqWtzuBVJYxIicgig76fMcOGnnEmbGRPZ72ysUtlQbQUEJ7UqWMna9GyhcVWi7XE+g1cqrVrZyoXVENry0lRq1Orlq1gJ3fj6dviypt69er+B29cCeMdVhcQCAgEBAICAYH/dQjI1qQ3Mq98+DNvYwrtux8X23ezlhQRKnrLE3lz9g/9b9F7Ab294XWFvG9wRczv94tf4w+AF5+LFn+lvuQdpD7JIBX5rm+F+5G/GGqHIoRO8br0muL1hO1HMA346+L5f/P603vmoos+fAoI/F8iIAtT8+bN/y+f/e/wNIh6/lKwijyUUaNG2+233WaPPPqIrSGU99nnnrEcIjRW08wb+S2hfxUjgbL6dyEmJlI3zQpcMSFnx/QZ023kqFEoTiZ6E3DXrl1t7NixFNZsZv4l55QoDQ39KyFXmjRu4qSHZtiqVStDWOTYCgpympNFo9zUCy88n7k2MorrjxNlyEfRr0T94WEv7pHMzEwbMHAAv7b8HylUJIcsGUXLVjJoNEPXJSu1MZarGezTu+++Y4MGDyHI9ybcJG3soUcesnm/zud3IFk4RHb88MMMjqc8GSplUYKUt1gImR2QKtpOjd1xWHTmOwmlGm+hIMFDCyxbR8hxERZaRPbVxSa0jkgQ5bi2xbVyAPWNKsN3Ml8fgTySdeswJI3IITVbKZ+nYqUKhO9KwcLC8VVH9XPb7bfbL7/Msc9Rn9xyy61kt0xFSZNjZSgTUnNTJmTSZRAhqruegrvl088/M+Xk/PbbInvqqae85lq/36X6WbkiYqXSr3fxEMoNyj2e6/+u63e+4KtLC1ZCQqJ9M+Ubuxd71l0EA0+a9JU9/sQTZMi8xXMjFintnwqFqlajTUuiEkHPCqSGSiOTpmxMTKTVKXI0JX574ooZDkCLgNixYwdBR7udEWzatJknHeuEibnN56JQurTSlFURpqwZhS6J8Y2ChFH/+4ED+yMXMOoWnVgdfFxcbdg6euIheWZ8P53nZFoX2MiJEyci4fqJawWwCOg5+5yznM1bv36DM4LFtWG6EFSPduqpfezKkaPs4osvcXtUfkGek0LqS5cHb/rU6TCmi3z7kb+uRY4r3AYEAgIBgYBAQCAg8N8jMOIctTX15on6zax3M0VvEPROx7/mMV8gToru+jP4a5S+9Pu60ev8AW78QR7T10Usi/+Bx9ctCoYtFbMvul+0rbB9YVcMnj7ra4GoJeBfDIUjdDJcf5ETF24DAv8tAiphkaohLH9AACWECBARNI+OfdSH8aSkJJQ0P1vP7j39cX82g7j+3dxK7krNmnV8/tTjvNztPW3atPYg3qNHs23hgoXYey6mgrqH24F69uzlJIL+rZXd6cyhQy05OclSNqW4dWcLZMom7qtyeilhtwt/W4haZ5ZnnWobWrpRry3hgNQohYWlIY1yI9/g32xl2Gj/FxI2HPkdaLZk6VI744zTIi1KqFzyeb4IFtV0f/XVJGbei5yU0QFoPpcdKi19L9k1uYgRmrkAYunipZaTk+3zuNqL33j9DciaNLav3xn2e5W0f8HeiOQYjvqoRfMWNBlrVo92dcuiRYu9VapBA5wzxI4oTiSX9alxaSFkShrr9IXf1/p3V3Elw7CGXYh6aQltTD0QWFQiokTRKMrw2bMnnZn/ENtjg7xGmT0dO3awqVO/Jaz4Hrv6qqvdMqZfbcWLfv+rdUnZszoPep0UM3qfEIcyZyihwOIfUsjU+fLLL+1NCJlXX3nVxo17049fJIyOc+q0ac4/ZGcf9cd1/iUG0X7KiaNqdD+I4g2X4OcTV8wUASKWUF6vQ9RhVYRVOuuss6njWm8rUbvkcqFIrlQO9YzkTQ0bNjL9QEz9bqqH/YqcKR0FnKxDH3msR8nOyn05bchAQnm3wJwddOWNasnkMfuAmq4KJFknkk2TkZHh7NZharmlmKlMenUsFqmPP/7YgUxIiHcSJjmxgYcBTZs23au4VZtVFnayItYqSda2b9/uF6pkVVFRkeqxEsQ6rCogEBAICAQEAgL/qxG44txupo+wBAQCAgGBgMD/LAKaY6RyCMt/RACjkIclq2paypKfZs3ybNLEBonkoUy0p59+2l577XVCZCPKzi2QCiMuI0tk0tfM30ouKeW12L+gAGnUqCGERIZnj8zAtdG2bTvPMJ058wfyUA5AJESyaqS2acCs279/f/t++gzIjxyIkGq2iqYnxXi05DylbNoIkfKrXXLJJda0aVMrjVDh1ltvtm+nTrNpVENn5xREtg85oVZj2XvKo3Ypw/NU666g4Q0bNkJi7MWqdIpFE94rVc0cym860lw0fvx4Dx2eBElzFJJEZIXm7FjmdDUoH0ONEgtBokgPWXV270mzfggZpAYaTutT7Tp12b7/mcNfG0E1QvSffsbpHF8D+5X979K5K4TIUQ85Fh7KaZEaRzP/9OnTLRGHihqiWrZoZQ0bJbPGf6xp6LChLr5QO1RcrdpWAGkk1VfFCjE0N6VYTfJuRIIo/FgYXnXVVfb88y/QpnQZ+bT17fY7brcNWMsiTIksZ4WemSPiR4ssaBRAe47QO++8C+FSUw96k/TjtFqtow787bff9qDstSiBtG9yAL337lv+HJ1/bR/Y3MYkYkf3fz8IbaQElxJTzJRBDROL7EfeK4UIQSxxAlrivYuFXazmnrMCLoA4gn7ksZNPTn40XQyyO0VCdSBllEfD80TJCZyqvL4SzG8MKctt8eu1J3tGCcyVK1d1Jq58ufIu9VoG4yeZmpQ1yoipSoaNlDqbIXWkxlFatQKGcyBuFJKkkyv28scff3RPoFhFnURJvoR3WAICAYGAQEAgIBAQCAgEBAICAYGAwMmAgP5YrT9qh+U/IlCKoVTV06pXFnGlwTo1NdWuvGIkf8T/yF588QWUIq/7H+UlNJBNSKSMiC5NhcUZaAcQCdQm3FeaD291YlA9wizbjxbhl156ycmfKpUreQGND+9s5xQIk5GjRyIqqMOMeZz4joG2GHVJDebhrEOagaNt2rSpLmIQIaBg3AshRW66+RYnYZDh+PblSpkOwZOc3MjVM5qSFRCcsnGjBw/r+5p/y5ILo1laRM7KFSup8u5sjz421ho3acw8nO/AiNg5lYYhuUtUspMEwVIbUkStyCrIkXBBTVNaNIsrIFmLH5MjwoHxv/JeLsWJcoigXoUGK9ulbp1avq6YmHKof9p5dbc0K8dQ/Kxdu9qzc6Rq0aJ1a0Vy0mh99bBJ7UYokUEwcBxEzB4PF44cvwQej5Clo5DeyVMm2w00ZSljZ94vv1qP7kV/DGKFOl8KWt62bTvr1FesGCmM1EaKM1Gmjo4jck5L2RdffmFDhgymqjvfrVMXQ5Lpe/sy9vve6XVaFNki3NRCLRWOVvvPWE6cmInsr+9btCtisr166iPUKmnp6WS71LcqVSMkjHJhWiIDW0P47py5c7wGLEpg8QNTlTwZ7pEnswr/3XJfhwgeAaX0aS2SDylVWp3onbt0Qv3S21q2asnJPu7Ej/rIMySrIhCpDSTOtX/6k9WpXZuL7AAnooYHDG1AvpRAZXbHjh39s7xrtWDjLr3kUryATa1iZRKYfWvhJiAQEAgIBAQCAgGBgEBAICAQEAgIBAROVgQQlzBrRtlGrERJyUk+rGv+3LcvnbyRe230qKvs3PPOtTNOP42G4HY+lyoQeD/RGlKYyNkh0qEMf9gvC5Gi9WlYlN0nOTmZ9ZWiDvo+Fwfo8QrMkjHR5NPwRJEAzZu1sIcffdRq16tjmfszrQOWnJSNG9zmpDxVFdzsJu/0RZqfpNZJhUBQyPCjjzzss6uIAn2omfjNcW848aKvK1So5I4UKX1yIWMUEKzjTNmc4qqPTXw+AGnSh+DfZ555xvqR6apoEe1jPYgi5eM0J2j3QOZBlDyrrEqlyjZv/jxvTDqE8iaTbBwRTtupxZbwgZf54i1SRV+ICOyBjatGjTgPH1Yb1MEsCCxajBo3aWiyNmmW37M3jccP+bEtL2pxikz3rEi2Kf6XlWk1iqK65L2KFyjGX8cqckhBzeecczaKl0LPuRHxoxry6TO+t5tuusmFHlqnztkcGpyV2isFjc6XQn2PQajESPHCf+JbdL9f376QUAvtiiuvcPWM6tCfe+45dkf/6Xk8kUUByS3IBlIg8R+bI/2bJXhzwsSMDp79tmgkW1WqwbzBvh3lwlCQ0q7dO/GAXcWFfobXeuUhoyobVdZ2EKozHZnWIXyQXEEeIKREZ3n2dAGp5u2dd99FVvaa/coFcoCLQ2csKrqMewSVXl0ZD5p61E8/4wzLRf0yb/58O0yjk5Q3ZWEMa9euQy3XKGuGNGwJrKHSodPT06waShtdRK1g+c4+5xy7/sabbC0Xweo1a7joC339ugDCEhAICAQEAgIBgYBAQCAgEBAICAQEAgInJwIa1CUAOI57Ih/VQyViMKSKKVRwCEttLDTvv/+enU0rr6a/Tz/7DCXNlZ7z8semH9mX4uPjmVlxdvBEVV1rcO+IauMsck4LCK09hKAgJWWTr0fhuXJ1+ETJTfmy5ezhhx7yENnRzMb7IGjKMTtrft3NvKxGYRErP2KzUn3zJsKEa6JiGTBgIGHDI71JSfubBUEhB4q2L3JBgcHbdmy12jVrU5V9jAwVrE48tnL5CicQpJ7RoF6HoOCn/vokJNQof61C8+UgGThoEIqVWu5gUUW1CnSU57oDcmg5ipvSzOkSWyguRIRRLmHJ4lGEq98wM2su19IA8cTateu8nTmK/Uusn+hBvsqwScFytWH9eg/mVcjvPyxSvNCZLs4TByQCZQX7vmUbAcdgKppBIMoipvl8O48XQLAoh2cfc/2FFw23p/72lD3y8CM2iKDkpKQkf37khewitiVVjh9GnVSVqBNtqvj8S7zRB+VQNbJqdxNgLDvayJGjCFC+2aZ/PwPhBionbZfX67h69OhheTiDpACKHL/fsoMlt5wwMeMkBoBJ0hULu9W+Q3s/yS1btQJHefqqw5g1sWguklQIl40pG9zuJClQlH4oQFz2J3W9703b51Io9Zqrh30jbKJOYhqJ0WIomzdt7uSNCJ08qrKi0NIoI0b98ut53mEIHHneqpH2nIfkSeBXrFjJu9ynTJliGzZutN59ensw0F7YyRxCf5vRa7+OwOC/3P8Xm/T1FLJm9jgzWnIQhzUFBAICAYGAQEAgIBAQCAgEBAICAYGAwP8kAoyomuq9fUnZpN7AxnwYUUKUIs5iIBXOTTyXZQZNSdO+m4Yw4FVXr2j29GYe5s20tD0uBjhwEFEBM6per7Fc828h2agK55XIQPPnRubNIxAZKrGJ4cOXyI6QE9PdxQr1E+JR4JT1aI3S1F8vXbIEIqC2ZWZkesvSDLJZ1PSk7T/y0CP24ksveLSHmo9V3yziYwXNT8pZbU10SGVan6JR6VStVs169zrFj1ltQvuwBalsJ0IllLIbb7jJ/nz7HW6hcm0FipIEmovPP/98S0Tdkn042xomJVsCJJRKcdRWHFUKFc6mDR6Kq1k9F7uV5ncRJlpKOXtiZN6U98aipsz9InLWMpvL4iSyR+SG8JA7ZtmSpXYQ8scXgVi0IsWXSI2zgxru3WDp0SbsuR6LIos2PiGBUOFtfvzF+MuN8zSEkxRPiiwRsdW5U+cIccJ6JSDRJlQrfu21f4JHyHN+QK+vWrWahzh36drdd6U0AdGfQcz169cXoUll++67b4lkqQqTIyvTUXvl73+ndjvHz3/xdeUvLMGbEyZm/MSwQzoZnifDBaEwpY4dOjpDOX78J7aGYJ24uDhnE3UBHSHnpSrqmqTkBg6+Qn5UJSamrg1p1PKMtWnXxmrB/uXCbipUWDKy+IR6Lvf66MMPbQFKGlV8qa5L5JBA14dIHYX56gdCoUaJDerz0cA9Zce5MOW/VEr0Gnxuanl67/13UdCU95RmdbHrAg5LQCAgEBAICAQEAgIBgYBAQCAgEBAICJzcCIg8UfV0m7ZtmAFRPJClwujos6VKYzSIa9F8euPNN9qYW2+zq6+5xr799jvPgpGOQLPoq3//u23GDiWSQXoZzZ8SC2SilNEMqpgObUs5prt27fR2I19x0Y1TBD7RF9oVV1zBTFrfiZVRo0b6/omfyD+Wb9u2bke98aOvR+U0Kta5kFpsuT9aYqfR8yQ+ULjuL7/+6mSMGoajaUiKiY5xMYT2WXmua3CEpO5IdTVO8fZP6d3LbUex1ZS3AuvA3ksNdNXVV1m/Af1sBUoZuUpE9ChnRzP7hPETyZ7JpgpbOTjZ9tKLL3mWq45bYEY+R4iqdCxiEjpso91KyiHN3xVxukiVInWQQo/HjRvnx6djKV4UQDz4tME8VNqVM2rSkm1IxyLlijiHTHgEHYeIlXLUal93/fWuEPoZ69KHH31oO1N32izImdGjRnmGjogZCUV0/ndyTjJpgC4+/zp/Ugc1btwosgtF52bOnLmmrJmff5pDZk5de+DBv/D9UlirMooCgTlaP+DiPS+5zxzlCS4clHZO/d6S+PTv188aJTdCLqb+8EIyY1YTPrTca6pvIWm6a7euFg0zqNajuBo17bdFiz3MR9VcutD1mmowWA2TYevIglEAk7xpGZkZ9LZvw1+XZZtSNtk6Gp+WUhO2mQDf1q1aEwzcxk+6TpSYtaJLxMmdgQP6e0tUw0YEJiGRktxM60nZuMnGvfW297KfhYStSdMmfpJPEJHw8oBAQCAgEBAICAQEAgIBgYBAQCAgEBD4VyLAjKqZXn+U1zyprBAN/hrXZZ2RmiPrcFbELcFzo7HIvP7G63bWsGE+j15KO9MZp5/pwbpyZ6RBOviwzxyr+VeuDuXPaKbHXOQBu4tRv6jgRhkt6en8wZ/ZVotns0Tu0SBc1W677Ta3UC1CmdKzZ09XZmRDtrRo2cL3beOGFJswcYIH+e6leUkFOBM/+9Tn7SgcI9q+YkJU/bx82TLcJhxnbg5OkHXYdqrR/lSNbJi9tge1jzJct2za4vsukkHHL4FEHYgHr5RmH6MJt70Swkh5Kz+S56JK7GxiRuLrJVjz5s19Rt68NbKOzZs321133uX2Hz82bnT8WrOakJo2aUZ47y7akCp6VbUapI6itFEV+dEc6rpZtxqbVMLzx2XUyJGQYlf7eRApQ0oM62S9zPfCObcgL3IMnFTl0Ii06t27t68iOyvbhg09015F8fTMM8/y8bRbwLQWVmIff/ixtzBHQoFRUUEw6cxUINtGi9gDLcnJyXAOW+yOO+6w9P377O6777FvvvkGoocaetEeOp+Rp/rzS/LmhImZYvZNO6kk6qZNm1nbdm1NlVPHYBelfpGdqCz5MN2QCg0eNNjVKwoH2imvGqxi6q5UGEe6yiFU9BqdCNVRxVD7JRZOJzh1e6p98PGHVGenID2qYjUhdXT6D5OwLULlqtFX2QBIIXnhFGwkZrM4nEfJ0LVqxnlwsNhFpWDrB1HrSagX7/k46kyXVUrKnrAEBAICAYGAQEAgIBAQCAgEBAICAYGAwEmMgAZo5ktN4OPGvWUrIShkkYnSfIlaRHkjuRAFImR8bhRlwXSsKuxLKIaR2uTxJ56wxx4b67EayqrxaZ5VRmN1qk8VtKw8mtXLkhmza9du27QxxaTg0GMSKvxxiuchTfa/z/Vjxowh97SVky4Z5Nho5ak7d1C9jcKDCJAalNd8/+MPtAxvYl92WI3YONqDLrMB5KlIjaJ91iy9CzXKEWZiKXg2Uh/dANJFcSIK1BUvsRb3ipMgHLcahop5hTqEAEtQoeIcPy5uTicEuX2HDpT3VPVsltga1QkZzkNJs9x27tjp2xSZpVaoV15+xV4mIFhFPHKsqElZ9qLS0dA0zOQq7mmCtWkdvECfPqdCGq2PNDOTlbPwt4X20Ucf0eqkLFk/dL+Rgumuu++mmryx71Mh50OtTtshdYR/8XwvvHfv3G0tIY2EcSnsTgdY74MPPGhXjLwC0usc644gRDXdOjZv3OLYdf4chXJWAABAAElEQVQl5JDzRvEnWqfwEDETRR5Rn1N724XDh/sxZdLOdObQoYQYJ3JOf4JrGODb13l0gobPJbmcMDFTzC5thln6/IsvkSjtIvMl1yVD+jkYduYwT5aeOm2aS7yUqqwQpiMQI0qeTqNJKT8336VHrpgBOR1sNZi++gkNYNwaW424Gh4etBNyRhd8LQKMBg4eaM2QcwlYMWgVkEjVqVdPPI4dyYkEFK0no2btmrUuE/PMGZhRdavr4pIcS+yf+s93UKmlurGDsG6qGgtLQCAgEBAICAQEAgIBgYBAQCAgEBAICJzcCMTxx3ktmv/2QGBoUXiuCIQYERIaHvlQqKvm2tq16tp5ZJbIcvPlF19g23nRVROTvppstQjK5Sk+f8qFoflX2Sia/AvxFynLpADpisgPz1CdM8fVOpGN+ksj2/MHIjcDBgzwGu9KVSpR6Z2NoiTPkiACNqZstOSkJIuvU89moix59dXXbW/6Hl9fW6I/HnvsMQ8vVpaMtl9AY5Dm7BzsWvvT99uWzVtoLd7vFi2ve0YkkatcV0ijdJqTi5cqCBWUBaPjityU8u136dKZ1qZmtgR3S526dZnFK9KsvKZIAFHgr1Gm7Dfffmu3336bELA/0Ygcj+NFZJCvjdlbdqyy2KKUmdMB1Y8qug9SDR4LbiJtFixc4DXa22h/Kt5+K1RDY267FS4g3kpLuMG5WrCACm+YC21H87/wz4ewKQ8HIPWT8NeiQqJvpnxjHSCX9MiUb6fYBVSQK7qkeFFEykHqz9u1besYOdHC8ZfBPlW1Siz7Rls02yqDMuqnWbPtgvMvpDV6lX0y4RM759xzHCqpeEp6OWFiJrJDhU60jJ8w3l6nB3761OmWhnTqeP5xK1O2jJ+ApUuXOCs2+espLvPSBVAXIkVBQEpgXrd2jbclDUM6po75Y6hkxITVpdNcJ60RqpjkhsmeEVOjZg1nKLUOnRglNG/csMFVMtXIuNHFcBgJ2eaUzfbGm28gtbrTJrBvKwlJytx/gB/EAt9tbbsB+TPbYSDfe/cdZxKVgROWgEBAICAQEAgIBAQCAgGBgEBAICAQEDi5EVD9saw1GqOlcvBWnsgMb3UJ3FVtsoZsNfjoYTkxWhCoq4yWQgb+Dz78wC1HiUmJNn3adDJp+nn4rvJZ1pHhouwSX5g/ZddRZIecILLDaP595ZVXmH3FVBRtVE/mbuSryK1Kcm699VZcJ+1wh2wilHiQz8W/Lphv5clx3bRpE01O+2wrZEtqaqpvrhntxE+i5pGVRxEhmn9VntOpQyc7ePigtxGLjJBaZs/uveS07nFCRsII1UJLlVO8F3o90PgS+SSrV2kbceWVPn8vX7bc6iXUc3XP7NmzHUsFHovAuPTSSz0/ZtnSZTaN1uVLLrqECu0eXtAjjCSKOP20M9xZI0yVH6tlHzN3WRqddiHq0PG99977rqwp3n4NcmObUOd965hbfV9kq5JbRk4azf+yMZVnltc51b6WRloTDbEy/PzhnMOmHiAslcvnn31u4958yx4b+6hn6UTOf2Qr06Z+5yHH2h9dH/qeMmolBFGGTaSBq5Tbw66/6Xp7fOzjdtlll+np/5SlhIgZ+EVoJTUkKXR3BxfMzj27LD0jHUVMGhdLGfxltRzI7anbOWCz7t272Z0QJlddNdpO6UX/Od8Xm6asmN69T7Hde3bjC4uGsUtyYAR8HSrNTjmll5137rk2oP8AvwjTOKkHD5E7wwlVhZn8cLqwjhUWuDJHjODPsJVffPmVPf7Y43b/A/fT774FMFWNzUWse5zggxA5kpApBCgsAYGAQEAgIBAQCAgEBAICAYGAQEAgIHDyIqDxW2G8i5YssmFnnQ3JUa5okI+oHVZBWlSgYlpWmeIlNraq/fTTT5acnOw2JZE2kydPtouGX2TrseS8iIKmNQSKwnFFMijfRUobqW4qVq7oRI+Cbnfu3IWCJs2VG1989QV2pK3Fm3AWIEIURfZDc6kIBtmIRpLxsmDBAhcPZDDH1qUZqXr1WA+i1X4pY3Uq+SxaiZqcbsMONeLyEU70FJY+7harlhBLOvbuZNck0Gak/cpAnCDVigKBW6F0mYUKJ0cNS74Uel6N7ionxl9cdF+KnmHDhkLu7LEmjRtb69atIaxKW1JSMpk9mU5SxdWIIxQ4x2646QZ78KEHrWvXrnYeTU9RZdRsVehCjcaQY2p5Ejmj5ioN7D/N/cWboISrwpCVO6OA4D9u/+KLL7YvvvgcS1JL31MRM9p+dnaOq3O8Ejuy11ioolxR1I3t63wcpS784Ycf9v26fMSV9umnEx1XnVMtqgRfv2499yLkjtZdCuWNFEYVsGFBXvjz4uProa6KsZdffsnGjh3ru6fjKunlxIkZ3yekXxxfRbxmjZs24oS1ssT4+rBaSqbO9ryY+++714ONlGgsAqUMbKIuikGDBjs5M3rUKCtdpjQX2zIYvP1YjY76czoho6pVq5Y1btTEukHmKHinVq2aBCM1d2JlwsefQMqkuFzsCNuqxIUn6VHe0XzPoYnDBlUNjxyEJwFIe2zer7/CYG5GNVNIxkxVqwkhVBOJm6RskkDJDxeWgEBAICAQEAgIBAQCAgGBgEBAICAQEDh5EdD4/Ze//MXee/tde+edt+3RR8b6/OfMCN/7beFCm/3z7Eg2qcZtz6Mp5VYbzZ/6WranaGxAa1avslGjR9mbb75prRACTCSYV21DeWSbKqRW1p8asTU8F/WMM07HqbHCrThC75c5c+255579nXDQY/qiiB/gPnvK/4VINPojPug3oL/PuZpXSyMkUEbqJqxN23ds93l2yW+LiOtY42uRA+Sss4fRIFXdYsqUc+WMrD3Kd6mJ6kTKGiltOnbsQCHPKm9kkupGliKtd/GSxb79TMQJX0+ajB3qkMPgK+f4pSJp2bIldp7zaGEuh5ulrpVh2N6DiKIRRM2BQwcsJxc1C9vM3Lffnnv2OTvt9CFWB9JIRI6H5h6PHGtZWqM6derkjhg1N1UjwDeuVpw4GreBtW7dxsZ/MsE+ptW5ePubCBoeOHAgzVCf2PkXnO8Pl2J9IkYkzEhFsSRBhuZ/VYlLJKK2Z517YSySZcL48dRod+R8JLqDR/k1OqdaZAET+IpUkWWqIJ8MGjCXKkf46ySJ3Lrh+utc9LF48WIIOz0c2YKvpIRuTpyY8X3STSlPc65YsbKH9QjAevXjAXmvB+3GElZUs2ZNv5hS6SbfvWenq2l0TJV4TWPsSgrfkeXoo48+tvUbCA9G+uSECRd7Ah6z0aNG87m++/mUdi2g09PTkBvtwb+WZou4SOfPX+DkitizKGxUsdWr0+5UnwsgwQmeKlWqWCU+YspFuxKnLqFHQwafZu3at7Oj2WqSijBjJYRvWE1AICAQEAgIBAQCAgGBgEBAICAQEAgI/AsQUO7KvfffazfdfLNddvkl9vTTT6NAIfqCRaRDcVvRcYZ9Tdyeh8qAGs0cqTlVU27NmrVt9FVXkWuaZy+88ILnonTEMvTjj7OsT+8+DPGlPWw3XjEd5WOc3GhDfolyUWUdqp+YYDUhejZvSTFZgdQO/Dv7wWbFTJSRuoSZV4RD40aNrHefU1yMMPOHHyBd4iLV3JkHPFu1dp3a9tPPc+xR1BuyT0X20uzUvqda7Xp1nEzSsZUqVYY66Ca+LgXvDh4y0Odxr/TeuccOy+mybYdvX6TEylUrmMUn2A5qskVq+MGzHglHauNcqUdpjmxDCj7ORkShIN6p302zxPoN2FaEVmjavJnt2JFqHSkAStu7z1ukGjZKdrJDzIlsW8qkGXnlKM+l7X1KH6/krkors4qB1H6VjuNG29+wYT2lPYdsyZKlvv23337HRo4cCVZRolJsGtaymT/MjBAoIpHY0WNgqGxZwYoryRdtczecRMdO7T265FxyYu695z5fpz+RGw6TRdvcYAnx8TRA5frXOv8VKlRyjqFy5Sp6UjHckfsleHvixEzRzijFuApqFvXBC5Q4VC0V8OylAYI60Ldu2+wyrMTERD/Hhw9ne7+5esXVKa+wo2MwVgcI4tmUkuLJy7lcaEqcnjVztm1A8iSZmexHaXvS3OulfnRl0VSoSCo1hMphLvIVsJNSyBRA6nzKhZWFzakSbGZs9apYoWo746Wgn7Zt23tA8WZYOLU/1YyrZTlHj5BKHVOC8IZVBQQCAgGBgEBAICAQEAgIBAQCAgGBgMC/AgHNqFKGSDXRu/epbqW5/vob7NRT+7ioQGGyWjSA60OtSg2SGtgxslJFbkgJoibfGMVq8Ad9ERM7yCc9/8LzXTWi9d50801OqJSBsLji8su95UjOjBxyURQ6rMySAsiCtWvXe2X3Qw895N/TdovkFx4JIleIlBrakdIWZdcSptuCshttrzYKlD7ss9QciurIhzhQmc10CnaO5kYsScrL6Y6NpznZLJ4bA00hZci5555npw0ZgpAh3fe/HI3FK1ctZ+4me5Ww4hxCgfMRPBxgLk9H7PDRxx/Z0iXLfPekxPFwXalU/D9Kesh0zS/Iw8YENggeykh9ImaDrxNRlyhsNwUcp38/ldDiv7too1PHjuBCbguzu8p2+vY71f5y/wPYxCZ5rXV1GpJUxpOICEMV1qtXrbH33v/AFhIOrCBl5dykIchQps1D2JPqN6jvcSR7ccT4iWNvj8EHlJMLRiQbC4kxTq5JHKIq8IyMTA9NliXs5ltusvfJtWkOkRTZdTek2bxffvX1ZWfl+PnXWoThsuXLqDmv7NxDZN2RbfiGSuimRIgZHYzkPDHIm5RwXACb2KdPH05WNABv8uwWPV6dE3vB+RfY5VywPcmVUWaMVDEiVw7xfTGUAl7rKwOosjYpHVlpzdO46NasXWe78OvppChcOJuLUZVXcXE1LSkpiZyaOPrSK3jwj4J7vp48xbZu3UricjvStc+3dvgBJb+SNK01adbbtm6zF55/wSZ++inE0BEPFi7OnSkhfMNqAgIBgYBAQCAgEBAICAQEAgIBgYBAQOBfgIBmVAW5doQYUIvSddddhy1pjX351Vd29113MUfW8L3SLC/fxC5yUpVZKmuPLyhBooi6mD9/vjVMahh5COJk0aLFNmrUKHv3/fd8nmyqemeWihUrecZK86bN3VajYhk1QK2iqlsChFWU0bRt09YmT5liCxcs5BWsrGipWQsRAeoar6/mYZEYp51+hldk6zgqsW7Zk1RVffBglgsO1uIy2ctc/NGHH/paNEerOKdnj55OwuhBzdPdunXDIdLes1PyIYlUwrMbF4vKcZRbcwzbTjr7mnUkC8Inx7797luyYFKooD5k9SjjkY1JMSXaWzlalkGUbERM0a1Hd8KNIUC0IW7KlClrc+fO9edoPc8++6xdASki21d/8moqEGYM1+FPbtO6tV1JwLDsTdpOOexXhw4fshXLV9jSZUtd1DEPN4wEHGpPVrHPdtqURVZ9QRv00KFDnQzS+kRqaP6vhKpF59v3h8eFm4gkkUHVwU5k1lIUOH0JcZZYYyF2th49ejj/oN3KgeR6+623vX1K+KM78uNTC1Q5iCVfrz8auacjKanlxIkZzj7Hy8FotyOfjyMDU85MFMSK6qp++OF7mzdvHlKxrUiL8iw5Odn6nNLbBg8c7OxjJsFBh6m4Lsin9iovn7UcJz27CQFNFZ0wETMpEkYsnjrH1UH/znvvctFGEyLcAz9fdT/ZsYAumVU1ApIaNWpobdgHXShVSIyWL+7hhx/xxGuxjNu2bHU52oLfFtjECRPdSzh69GhrkNigpLAN6wkIBAQCAgGBgEBAICAQEAgIBAQCAgGBfxUCWFyUczJ48GBrUD/RMlFNTP9+ut1z9z12w4032scff+zzXyHODM20zOJOcnhALfssBYuUIlLdVKlWhZmXViCIGlEUsiTdf9/9NmLECM9abdKkMcQHTg5YnspVK3vtdgXmWTUP1YMI0TxbtUo1y8jMsEMQEV9PmUzg7bQioiJC0XjER8NGKHsq+nwtxqAF9dG3EvJ74OABQnZxmUD0ZGbsdwHCkcOHfc7OIKP1/ffec4eJoFZNuAiVYsuUZvXKEDtnDzub0OBaThrFoSRR+9GC+QvZ5wIIozbk7Rz3RimJJ7ZTYf3gAw/YosWLXLGjSm+P/QCTI8SKLFmyxOfz3FxsP6hltK8Kz91DzIiIGD1XmS2zsW8p5mQZhMgpvXqzS6JAIrsmJVLHDuS/ELArd0wd6sFrI6IQWSRbk7Jq5Ij5hZxY5cHsB7v3UdLkHc21DyGj7rn77si6xGqw2kW/LbTkpGR/TAIOdsH5hi+oPo8HD3bRSkWVQpGz2tq0a+v14bJEdevWBYeNqs8jxyG8tOj869yrHUuv0zojz/Fvl+jNiRMz7KxQqAHbKB+dkq31UHSZmIhvb/tWkrAXQ878QIL0NAdLzJUu8DwkWCJdFPa7DWWLJFpSsuiCLQ/TJdmQLrKW1GfL0+VSJV4nqdW6NWudkbyPUOGBgwdYK0iYmJjylo9aR3KrRAiWq6+5xiVNKYQDC8BK/FA2apjsOTcTCGw6cuSQM6Ly6ekYunTqYg0hdMISEAgIBAQCAgGBgEBAICAQEAgIBAQCAic3AqWwMpVlplRttAgO5ZxICPDOu+/YaacNMSlb1vC9WqhVCiE8NBPuhEg5hJND86P+y83L9WwU1TFLm6FZt5D1dCbwt0GDJPsEcmcGZM/cOb9aYoMGFotLRBYllcxce92f7JKLL8JGkwE5NAgRQU0XI0hsEBtbzTNSFcGxL30fK42QAdqA1iv1DbvDw4UezXExzVAdIDE2U6mdfzzPCYsDBw75XJzJzFwe58jXNEjNp9WpeIGX8mBcfa35u1yFGLcDDeHYRbwM6NefrFdCfFHzJLHvffr0dpHDcfZJuTErV6y0Sy+71D7//HOrjhiicuWqKHCikFGQ5gLr8fPPc93+5bkvpaLcCVMAeVEWW5UOpxyRI6NHjfTslssuvwx70juOqfZHLILfcufMM4fa+eeeR9TJVqxkSVRxV7N4bFFSK61TaxOZsQcQc+j5GRn7aFq+z6bPmG5PPPFXe//d95w70PekAJrG4zpuhxMyRXecWMOSVoh1y88qfIFIrUsvuZR1PIHKqZu9/uprZPI08teKkPHcIY5h997dNFMNw94WCQoWkkU770dQUjcnTMz4AXPktchoufrqq6giO4uLqLKnIediSwIKK4BIUcKzTv669RthnkrbYZinxbBmakoqyDvmEiuduA4d2hMW1NG9bjt37rR9+zKwSJX3AKYUkperVo21Du07WHNamUTqxFaPs6TEZGvZvKXLutasXWOTJn1tWzZvgRzih4cd3LldoUalYNU+sBdefBnip6orZApgM9tABPXt089Zyy++/tLtVCUFblhPQCAgEBAICAQEAgIBgYBAQCAgEBAICPxrEBCJgtTBMtIz/I/0KtrRDHkMZYhUEz1697Rvv/3GLhh+gd1y661ue3EyhJcpf0V/8M+gbUh1zUfJRY0oRgifLSxt9VB3qIGJVSIgSMOWc4W9/tprVgslihqKNSerHelcIjVuufkWm4PFh0JmV3Bs3bxNoSh2EOtOauoOG/v4WLcGFaOkdSpHVcoXBQOLTJBVRzalK0eOtqT6yZaXk2cXXTTchQxqPCrPtn6c+YPn6RRK2cFKdPi6jQgjdEz+gHXr0s1OP+0020E9dZs2re0I7hVl68jyowyasogkdPync3zlED/c+ec77dFHH/HG5QuHX0wWjHJ4IvgczDrESiM2JylpRKpIeuStRoDQtl0HJ3VU7z328Sc8gDc/D5KDfRFZ5As72pLGZpEkUso0w04mkkQWJok2DhB8vHnrZjJ78lEOZSEGqWhvvPaGPfXUUzZoyGBsTcOs76mn+qqUVVOK7Qp/HX8O6pqy1JtHcWyRoiVUVBWrEEA80rKyD3tj1teTvyYkuY+re7p07sLzCD0WicMKtm3dbvEEHwujYntT8W5Hdr5kbk+YmIlc7OYBvvLoKXVZIcDqeJ8zZw6qmAK/mGQ70gWlg4GXsTwkT1K9KMD3aG6OJSUnccJquKcumirtAqRMyof5ceZMWwfZogYmsZ0KF0pskAhrdbb30n/7zWQPHObni3Uf85YmSZXE6imvphqtTOUgdgqO5dvmTZvtk08+wV/X1h564EGrigpn3759VrFSBbdYSRa1evXqkkE2rCUgEBAICAQEAgIBgYBAQCAgEBAICAQE/iUISGkiQYCGeSlc/I/2MBMiKUQJtMXKIhHA1VdfbfN/nW933X2XjRs3zuqjztBEr+fpqbIQNWnaxA4dOuzHgQjHv68A3I0bNxC0ywOQAIrLuP322+0WbEcK6VW7jxYRBD179aAJeLDFVa+BZecYapo4d5Xs2JFKXssyMlKjbe6vc1HdzPHmJflLNDuXQ3mijFRlvGiI1jF17NDOrr/heuvaoysV2tssOTmJIpuaEE2rvIq6Jdkt2xEmyCFylJlb29cOSgUSoU/4kuOqoNyXfgNQ5zSA6EmybRT2qLFJObBJzNua22uxn6OvIu4D0mbcW2/b3/72N+b1KBv72Fjr2bOHhxsXr1v7tp4mpXPOO485nO1CQylnRpkz1WKr2hmnn2455MVIAXPb7WNsJ1XXviOR3eP5hW6Bevjhhwln7u0qo+wjhy2fOV7iiz2QX5rv9+7dhcumtRf7yAr24IMPuLJGc/4tt9wi5s2P2W1HHPv+/fvhD+pDFEG08LXOa1xN8Q7R1rBhI+xbBbZzx05rB65bNm924mj0qJFO4PmJZb8egZTKzDzo+Ee8Pw6q9rzEFl1WJbKsJkTppZdeth2pOwnLKQCw3V55dRQvnSqrh5AELQ+bWEedANVZ64dEnjQFIomFqlu7DoxcjHv06tap688vR0vS/gyqwTi5tXnsAEFMuwgqkqdPmTQK79mHNOwIlV27d++xDCROR9jm/v373KdWg4Rn+edEACn3Rt4/+fLEDFaqVMl++eUXlzxVZx+VAi3CKCwBgYBAQCAgEBAICAQEAgIBgYBAQCAgcPIiECFWSnkDkspiNH+KPNCi77VCoXHa4CFeyazojYuwCnXD0rJo4W/WollLzyp1MoTXKcR2E+4NZnReGxmhtapMlBxRKFrUCBQL+SChwltvvmnn00y0AdImsi2/9arqAQP6k4XayMNlVTrToV17Ml7mI+op7eTKp599Zs88/Te2o5k58joJEBT5odwWbV+ztDJPnnryKe6VxqLVyirRGJSL2GHv3j22AqJHmSypqan2/HPPOrmk1/yHJQKD25KERUyFcrYrdVdkzsbp0q4dLUoE+S5avITZOd8tQ02bNvVg3kceedR+Zf2PPf64Db/wIpQ8Dor4EN/+Qo5HTU/FS/q+NGbwpk6MiZRSOK+UOQ8++KCLMPQ8Ha6IKC3R4FmLivIB/QcAfxSBweX8+PdBfO1h3s8+kuOqox49e7qTZmPKJkitX1C7/OjikDdff93VQ1qXji3zwH4bOXKUB/+KbNKifVgLQdS71ym+VeGfTmvVUCxLL7/4Esf7lGcPJZBLpEXXjdxAkcYnEXb/CU9/1ondlAgxo/Mqdk0naCPSo0NUb2XiU2vapBmetSbuyVOHutg4HYyCjw4hc5IKRgwV9J2zilLDRPFRsUJFgEi0Xr17Wf+BA6gG621VqOLSiVRXu+q0f5o1236d9yvd8FvdFqWwIHn3tP0sPIF5HiRMmDDrl1xNvjl58m695VZStBeRxD2JbZb2ZGrl2MTFxdl1119vyQ0bnhii4dUBgYBAQCAgEBAICAQEAgIBgYBAQCAg8C9HQAP0kcNHrDvNO3nYZ6TK0GyoQfwwGSPxqFoKeY5m1F9QdvRk/ly2fLnNmz/PunTp4o2/OggN5WoFjnAH0kxohMUSxXzqAzUDv+w3gwcNsU6dO7vq4tFHx9LGhBujiATRemSnOeecc2w4FqS0tDQCgs+z+PoJzM1NfcY9Qrjvth3bPYPVA4j9tRESQGqbKuSkaH3Fq7zkkktsEATG3t27mW1xl6C82VnUtnQEdUrHjp1tJ8KJFOZnX4r4hN95Hx7UutTcpKKd5EYNaUbKIhNmnddDq1JbLcvlUe6cd865PJe8FvC8hNrqBwgGvuiSi+2eO++y8ogrpPERWTV9+nQXXxRxYKhUSnl8yPYdqRAtpVyckUTuq9qQpFB6/Y3XI7DyAj8u7SMfrVH+XHbpZSiVslzt04umqS1btvCtUt6wtJtjVj5tfSmTwH/sY4/bE1ilzjn3XGq4J/tzdMx5ZNC+8847lkq7c9EWnEgTr6DzL6pG51LL4ewj9sjYR7GljfDQ6C+/+BIOoUNk//i+v74YfH9Fyd2cMDEj9kjY6WLNwl92AFJEDJmqqCTZeglL0wXnnw9BkueePZ0MkTGqvN6KVUlKlt2ApAshL5dGJj+DBBPBYiUlJlmP7t1t+PALTQydwn/rIOWqhgpGYUPbyKxRSvMyUpvXYEFS6FBlCBx1zkuqdBjp01ZO3sYNG1DykDPDksCFry74jz760AmgC6nvbpTU0E+GJGRSzoQlIBAQCAgEBAICAYGAQEAgIBAQCAgEBE5uBETMSCXx2aef4qzIiCgeGKxxNrn7Yj8tTXJsKDtGCpZULECXXHqJPf/8s3Y61ps77/yzV0s76SDagMFX1iUN0fpc3i1GUAWQPZqH61AqcypZJSqv2YN65RtqscePH/8PEDXU89G1c1fm5Jfsqy8neeiutq0clW5dujqpsWjRb1RCf8HH53p6ZGHb0dHRvg+av31uZu4tw2MjqKSOw52ylfDci4YPR+1TYFVpkVIosKqnRUJpTtfKNjAbezFS0b74uviGRAuKGtGsvAubkSI/ZH9qiMJHddZql3IhBTN3IbP2EyhmrrxihMXVqonVqym2qH5FO8pmwDyS0SJsShFYvNmtTyLA9LFpQwocwQWE715m38/43kkqkVyRffGD83UpIPk8uAQ1YZXBetSlW1c/frli6mK5Wr9+HSVEcY5RFOfgs88/8xp05dz07t3bLr74Yle3qK48Hf7Bn8hxHyanRliKyBKOHKZvu3PHTu6qmT5jGk3QKyHzcu1n7GXDL7rIFUtOzPxjJ38/3pK4c8LEjB8Fe3IExkkky3dTp3rP+x4YQF2gdbAntUOilcP3BwwYQOLymV6tlU34b87RHFe5LFu+xD4Z/4l9O5W+9A0b3QsnNlEysbL8oMTVqOVZMbIsNW3a2Nm1RBQ1ChyOgcDRNbV7b5onUZ8NA3n3PfdYJxhOXVzqPd9G3/msH390prBq1ap0n7dwBU4MzJ9qslq0bOaZNmvXr3FFTkkAG9YREAgIBAQCAgGBgEBAICAQEAgIBAQCAv86BBo3auwbV7XzEuxKGqxFFBTSKiTlTG6+gmGj/Q/7mimjCYmV8ODJvz5tEyZMdKXNTz/NsmbNmvp6CkmPLeIzyHdJtfbt20e+5jXrIDykNvnu2++sJfYikRMiXL777jt78YUX3Ybj078Ge5gRtURdOeoK64kSZNPmTZZDpIYUNlK6pGzcjMMky50h22k5VqtQZAci3ILfdzaBY2F9igy57oYbrG+/vggTthIjcprnzcjOJGWOrE/Lli5x9kGxIG+9M87bhn5XzojA4hiUDdOyZUtaovYLKW+ZSkiox35k+fxcrqyybkq58qdh4yY2ldlf83c27pnPv/ySVuSreVVEUcTq/L4yXGfMmOGzeHUaq4RgedqhNqZstBWrlnszlJqpZnz/PTkyqJK0FO2YsC5fPsYGE/ArAUUlnDXiB7bCO8SiIGrXrh2Cj3Q/BwXgHZ8Qz3EuszFjbvNsmSf/+qS98MIL7DLhyWxD4g52iiiUNM/NOZqX4+STtiPuogUhz6NHj+b0REHK5NuZQ4fae++9Y69jj3ocIqp6Ne3/P2c5YWKmmDCqyMlW2rGsRZKBScGiZqS9MFNZKFeOFeSThF3J2rZui43oKwKJtqOEGe7VZGomU2XYStQvU775BnJmvV8QqanbkV6lQuCQgAzrJzArVqhCFXYiYUoJSK2SrQk/bKr2qk2OjNhO1WoPPeMMO5cLJIHMmFvG3GqDqQNTuI+IoKpUfN1F37nA7dWrlzNi5SvQyc5/aojKgk0MS0AgIBAQCAgEBAICAYGAQEAgIBAQCAic3AjcefedNuKyEX4Qhw6pPQjbCgM4MzzNvzGuetFADovi30tOTrI/3/Fn3B4U1axb68HA6Wn7vdb6POqc/TWIB7QsWbzYUiFnlF+qmVjqEk39FZl5VRetqI3yNCs1atiIrJbF5LG+9HvRjEgPLbWo1K4OqXIm7UdynOzcmYodahCukGXWMDnZ2rZtY3N/nmdPPvWkK3z0sk0bU4jpOPAHhsZX5WtUbEdrXqPK6hxsOQXHC2xzymaIn82/kx4iKZYsWQbxNMEWL17kL66GRaps2bLcL3QlSVSZCL0i21VBfqGtWbUGhUpdVCtlUAtxjOTdXE6NtnJid+J+WbdmDeTTCyiN+HjxBa+6Ps7xS1q0Y9sOjzL5HuJF50DEV1RUWVu9ZrULOxokJ1nlShVd1TNv3i+2auUqYIzQH9HEnGg1UrfEVosl9iTKc2m107shmNatXWf12IcojkkFRIPJDOrUpTMk11GyaBdDQL1lo0eOdo5BbczFSy5OnbIx4i62ONECV+fXhbYnDkPHL2LtICqjO/98j/3pT3/yam2FABftWvGqSuzzCRMzxXvSoWN7u/mWm61e3TquPlE/vIDfs2u3W5CkXNlEK5IIGCUk5+fn2SnIi6SAqUwIr75fuWJlZxVz8YEpDXsWOTIjrrzSZV4CSK8rU4YfJH5wopBQxVALVpNEZT03KbmB1apdy3+oNm+BYURuJTBPhy286647PTxIpNGcuXNgHA/jt+tgHTt1sob42wYNHOi2pkykbJ7eXHxQ4XNAICAQEAgIBAQCAgGBgEBAICAQEAgInJQI/PbbIvvgww8Iqn3MSZhI+GvR0E/IbA1UF3oMrsCXevXq2RoIGX0WNZHG7HrOOWfRSDTO3iDU9/577ydyQwSGqqeP4xb5zgUIrsRgHaJbZIeqVBTUq9DhXqf0svY0/qSm0pQ0fiL5MRn+PF9J0XZ79uxlt912mxMQVbHvDOg3wHNfVqxYQdPRWixJmTb3l7nM1XuZZbMgVBbbFuqjRaREKB5WVLSuoWecab0Ixq0eW93zYaQi0Wtk+8lC0aOmp/04UbZt2WbfTP7Gj0GV3COwJakcR8oZt4BBOL3wwvOocWLJjW3kz5NiSLyUSJytW7cwT3dk+5I4FNpjYx+zq6+6ygYNGGgff/SR1UckkU/uq8Qvecz+IrpU/CP+SjYwzesirSoTXJxNe5PURUso9lm5epXNhgfQ8ShnpkqVSg6VHyj8We3aNV0NUzYmmmboXM+/0feaN2tOtXYVF26I0FJuzBdkxNx9792WRKvUp59OtNOoCC8G6rNPP6MFa66vW04dEUa6FtL2pXu+jCxX2td8VFWySHWAP6hWlSrwfxIzU2LEjC7mtm3be51XBSRG8pqp2qt8hfJ+kArjXb1qpS1fsdyrwpSmvIgflGjSnlVdppNVB1KnU6dOsHG1XVamQJ41tD19/fVkm0qIUDZflyldhsChLPJsULYAFNSWRdPcVL58BQieSEDw/PkL7MfZs/yCVRr2YZ5bBjXP3j277eVXXrYRl18OWXO3M2xS4cgLqP3Uvsl/F5aAQEAgIBAQCAgEBAICAYGAQEAgIBAQOLkReP+9d+2GG663q0aPtokTPrXk5IaMkPrP7IeZP7iQQIU0Ti7woHJOU1CkNGnc1ENh1Qq0L30fQbf3s54b7Kyzz6Kmuju5KxGL1DFZm5zVIdRW0g6WqtWqutKCuR5ywDzbZhiWmIGIAVQhnUbts6w9cnOI0IgshVhrklCcPO+Wp0oQDPUSEnjtfogf6rVpe1q2bLnnq04gL0dzcUrKJuZdtSYdYxUcgQ5KC59FsJx19tnWvm1b24oVqiHHrTKeTyd+5o3InTp2JIrkiEnQsJz1pjAzZ9J+LBtP8+bNWQViCFa1Ny3d7rv3PjJiNtFi1dpndimMRGSkg0v58jRFaZN8Db0F+fEpmTDnuULmPDJkLrn4Is9p0a4p2kQ7FwWmUhdF42hJ4Bh1fNGILiqj2lHhz1pybvYQU5KCykfP79ypM46ZBMZ+sm34rzTcwSo4Ah2/zqcWbb9t+3aei6MsoQaJDSCYIIDYsAqKbof02rJ5qz1N41Wf3n18+3qebF3+etarZi0taoyOwbJ1TOeV9YpgOqVXT1cGXfuna3lGMdD+9BK7OXFipuhamr9gvn0Ok6QL0yuvIUqkZBHLJZ+aGpliq8b6gTTlZO8niOgWpFaVYMASGyTBdJX32ixdRNWq1aBZiUAeQE9OTrJ+/ftZOU6cAn0lX9pBlde3337jKdl1a9Prru8BpDOg7I+ChDfQzrQEH90PP/zgHwsWLnByLIrE6U1Ilt59913Smqf4xfzDzB+RoUUkVqrwDktAICAQEAgIBAQCAgGBgEBAICAQEAgInNwISDXx2utv2MW0F9VEOPC3vz1jjZo0Zgw3S2P4V/6LiJnjxyFrePAYYgIFvopcKcV8KQJBuaSViMP47LNPbSgESxT/LWfOHEh7sOIyfOG5eahB9If+PoT/quymgCiPgzQDy+p0GHeIslHLlavgbM36DRtsCrOoMmAir49sX3ah6667lhm6Ffag1VaRmVqukO07drriRtEbeUdRn/C9fNavVuKff/7ZV/Ff6AKO56abbrZzzj7X1jIbK+u1IQqVzIMHrGmzZta/b3+vnVbmjrid8845zyZ9+ZWde+451rd/X4uG1dA6D/D8115/0+bN+9XzWBV9LJuSnCZSvmgRTvqoiCVJteLDhg0jpmSlPQPeTzzxV0iXqkWKFKgVVrqeXFm1JQn7ddRWR0dTiY2tS81LCmJWqU8qRT++8Pzkho0hhlq6ishzftj+BuJP5lPNnUUWj5wysahZlCerD3EKUv2IzFFV9z4Irvvuv88++WS8tevQ3iZM/NTJqsgGdBslCZQdQdBRsWIFJ4B0/tkZi6sZZ/0HDIKEKucqH794/vHCErsXQfJEVqezxd5l7s/Eu/WtE0iNqMiuVYdgXi7UGGROCiNq3qK5XXL5pZykodRplfML/OCBDCdXWhF8VLtWbZdW7UM6VHAs38FQdVki3eFjHxtrF5GorKTlVOq+5L/zECNA631qH3x51X09+2H5so9mO7snoiYXVi6DNOk1eN7knxO717hxI1PVmBQy2UdokcKftxQp2IIFC203ihpZrMISEAgIBAQCAgGBgEBAICAQEAgIBAQCAic5ApAOjJT20+yf7EKafpctXWRnnH6G3X/ffX5gh2gVlrBAShYte8hH1awaCQiOaCNkd7rl5ptdzbGdP+Yv+G2hF9dMZLi/7/6/0ApUgywXraPQVqA+UWNwDmG4YjWU9aLta35VFEf9hLo8LNWIFCfpuEKm2WIamJwB4XkiLaTWUJOwFC+qhK5ChqoyVY/S+qR59yjEkUQOFVCrpKFaEUGTzsyryA4JGUQm+MK6tN4eKHz69+1HmPAhgoa7uSKnZlwc9d7NnIBSNkst7EEK0r2RaJJbb76VBqWe1rVbd89tkaVHLU+yVSGFcGIpm1rvKlWq2jG+p2145TjH3wwS5GIajI7z+Pczv7cxt99mZ501zN59+x23AhXkH4XAkY3pAPP3AisH2SFrlZUu9NyaGGZ05daKDFn4228unlDMiQ5FnEI7VDHVqseyXYJ82YYCmNesXePbf/7F512BI5tWXVnROPeqFxf/0L1rN8KCY1FNfezhzE0aN/bWq7PPOZs1R+xZwmwRhFtrlEFHyOeRpcvJJmxp4z/5hAiVmhx9hFjSc0t6OXFixveILnE8bPWorCoO3nElCwzf0SKioyzqF11kTZBQleNxVYRdOWoU3eEJfM0JIHwnlxToXbv2YCfKwpZUGcAjTGP5mApWG7uR+ubncsGJKezXr5+zmFGcWKVnR+OLW0ra9tyf56C+4YeJq1rpzvoR00k7im9NV03t2rUhcuIgZ+JIot7rPe8KXNL2JKuSHC0sAYGAQEAgIBAQCAgEBAICAYGAQEAgIHByIyBLjvJQBg4abBnpGfbI2LGusrjzrrvsgw8+sOSkZA4QG5DUESyy7ChL5Gh2rkZwJyGkmPnll1/4A39jnz8LcFjcQUDwXX++y66lheiziRPJHoGkyD/G/Hrcli5f7vYZkS+af9ejVilANVOtSjUIh5paq5NBsVQ+5xNCO2fur1RUb3fiRQSEWAjNsklJSXb9jTfQWJzjIcMjRoywI2S4qt1YmSxqk1JUh2ZfkTEzv59pH0MgZKnBqWikdY6G+yIqEhMSXbWj+3uYgxXK26F9B2I96mHpOugOl05kxkyZMtmGDDrNG45fefXv1hMbjwKTpUApVURWyJYki0+hiKeix6SkqcYxlQFvkS2FWJQ+/ugTL/wpjWtlIvk6519woc/oYquUR/sZ1qqdO3c7USWCSWIKVWTHVq3uVq+dVHSrOnwv+1ucD3TTjTfaIM6nCBxZ0I5xPkSgSEnzI03M4h9EzqjSu0YNCoLAR6qZXj16OfG1MSWF+uvhbln74P0P7f77H9AZEfJ+/pdCzhw5jNiDa0JEmbJvZHmqDIEF86XT809ZSoiYMevUuRNBvVdwAklqhuBw5hBmLQ+yRcepznPJvz76+GP3sEUT1tO2bTvCf5tYcnIysqqGLmWSdMx9XQCYj7Rp586dNnnK154iHa18GS40hQbpAsnjQv513jykYYfd73aUx/djY1LejBDbhvxJZI3CfvUDVRrWTMxXFJ/LVShHiNIBDyeWj+7Uvqda9+7dqeGS3SosAYGAQEAgIBAQCAgEBAICAYGAQEAgIHDyIhAJb5UIoANKizbt2vgf7BcvWkz2ycWepTIRUuW00wb7/KnR/MjhI26N2b5jmz+XNBQnJDZv3mxxqEwKNalDAmj+fPf9d20AuTFJyQ3tIlQi3Xv0YOxlm6znsOw1zJ1HyC9VNkt1yIZffv0F10ZFrRFxQrzHfVSuXInBv5Tn2Lz3/vtUXW+JwM1KeDn2nto2cuQo69O3j1uD6taNx5JVy3NrlJGi3SkTFc06ypA7k+1OkRfIqVHhjoZwdlWjuO+TbnQ/vi7EzO49tmPnDp+tW7Zs7pmsq2hVVhPUE088YbOpCJfta+O6Dfbaa6/ZNddc4wQHvJO7UHScM2fO9KybyBGzcv4XiTELckQqHKlLtEVVlQ8ZPMjDg98hTkTqlTKINoSPyBjVmOdANKWl7fPZXIRVDAHL1WhhKmDeV9Cx8oBEzBwrwF7GNu659x67nuygymTSSLWjjBkpkpYtX2a1CFyWTUpqpeTkJNtIZpBsVzpnOn/Cf+PGja6gepmmrEceeZgmrCGOq3iIWbNm8Tz2nt0vxfEKV61bzVA6fh3VP2MpGWKGvRMT1RQLk05+NCzZKaRPr1+33n5DgrSLZiaBqCTod999G3byQ/thxg+2Gf+Zusp1MdagXUm96dvoJN+RmuoXp85uAf4xVZHtgGSRN23t2rW2HXZPz5VkKjNjH8zhYb+o6hE23KhRQ0ihOPelZcN0KWdGQT/tCCbeA6OYBRsokF1Kwx2papTe3AofX+/ep3IBkLQcloBAQCAgEBAICAQEAgIBgYBAQCAgEBA4iRFQxgghswzVqwmLbVC/AVM1Qzzqj6nTprlqYv36DSZypht/oC8gX0bLJ+M/ISclhedKMaH8lWPkxRxztYmGXSlhNNyLQEhhwJdVaMvWbRTWfOU2ILUMaYAXa3K88BjtRZvJgtlPgG9NWqCq+3rq1avrNqIYbFPlIY6kNNm+bbtNJDx3/vx5vh+lFHrDbK3vX3bpZViLunneTBmICX0rBeXHXgiWFSuX2y4sT+nYsHQMIiBefull24USRaIHzee+RFaHoCEfC9QetlnNA3Q1qwsTtSSr+lqBwDWxEu3etdPuwfL15JNP2n18fuHFF62Ct02BIetaumwpJMxMVh3BVJ9lezoEKVWhAk1KwoDHlENTCXfKtddea2NuvdXzbe65525yZhM9f0fPykWQsQ9rl2Zz2bX2QzqloiI6yLqk1tmH2kkijKnTpgoS8C/teT/PP/ec1SI7SEolbUuqqO9/+N5VSpw+J2QS6tfjc6FHnyhmRYDoupCQ5C8PPGCXX36Z5w9N+moSSpxBrEWLjuWYn//9nLu6qIok8ND3iuH0p5XgzYkTM5E9t+kzptkXn3/uBy1JUS0yY8RYPYpcbMLE8X6C4wj0qVixErkumbZ12xa6y9fY1KlTufgWWs7hHFeriAk7yPdF2tSBaImPT7Tt21P9QtFFJZnXylUruHC3wVpW95OciqrmSDaqmJgYErIbQcK0tg4dOlpLAoJq11Z9d77XcO3kopszZ66TNuVRzOgC1InmDhdXKeRJFZ1UKkF8w6oCAgGBgEBAICAQEAgIBAQCAgGBgEBA4F+BAHOehvIs8lXKUCKj4bp42b51u40Ycbk99dTTxGT0tTffeMPnzwIEBUcJ2JWiQwN8juepVI4M6hAtms61FlVgDx48GJtNmv2IouOhhx6m3OYWe/PNcVafFiG1DmvOXLtmnduDkhsmmQppFExbBVtTbk4uKowyEAQ1aRGqzwzbDnVGDs/9zr6eNDmyEd/ZyD63bdPGBqHQyc/Ld4VMtdhIoK5Ip8McX/PmLSAojiF02O410stXLLN7qffeR/6ML7ApWpMsSaI2ZGXajVVoE2qgimS76NH9hOT27t2XAN2IiySbrJVnnnnGTj/9dPJputNq9KQ1ppFKqhW9INJoDNHBfalWdMxSzUSVAXcRWzypOhk9F9POpM2+/Mor9hFV2nrxQw88RJZNDw9KdksSbptmOFky9mUy8++kcrwSKqEVCDS2u9VqM2TZFkijhx982F8jgJKSkuydd961eIguxZton45AMIk0OobAY/bsnyHSykA4pZrw1/yvRfk4wy8Y7par8eMn2AyyfvalpbsV7Lbb7uDcs0dS4vB0YXT99ddBepGDW3T+fSUlfHPixAwHLwDErn03dZpfvKmwa59//oVfzHv37nFJldKWu3fpakOGDIFAKed1WdlIxXSh7+JgVRfWDRZQ3eKVSVQ+dOiAS5xiq1UGWLJjfp3nzJeUMrIy6XWSN0kd8/4H79ukSV874yfPnWrNapAj06JFC+vfv5/7wSShyiIVW0FMumhEEFWEfdQPgyg/hQO5d66EAQ6rCwgEBAICAYGAQEAgIBAQCAgEBAICAYH/eQTQlrjiRVEbmgGVh1KcJ9OlaxdIg+reGvQlaomePbvbhAkTrB+2IQ390DK+wweYITXDqkFIg2/poiYixWfUxRYkRYjUHK+++prXabehlVhtQU1wk9C57DNxRkamTfrqa9tFkU1m5n6sSIT2MujXRYhQsUIFO3Aoi4iPRu7eiKtRjeDbBfYTgoLV2It82PY9KSSXNYZSnIvc7bGb3NVOnTo74SHCpR42JzVCqSlK6xYHNZkZ+UJyXTZv2uLr0egeOTZzG5H2MT2NwGNm4/j4uoQO17daNat7oxSuHl9EtMjV0r9/f3e73DrmVrc2SRShfXPbjz8TSxYkULt2bR1zCS7I9HXbljJpEsikEf7K8RnCzD/7p9nUkD9g51OrfZz1SCEki1PrVliryJzt2r2Hh/nqmJUnsx2CZsuWrdSI17NvaGhehKuGkwTRVZFcoI524403RXJgeFCtWFq0PdVlx1ar7oSVvhZBVBWXjGxUCjzWdZFBxMkFF55vH+Lsee65v9kbb7xOLXe0kzNZWYdpc56M6mmfY8qu/lOWEydmAENLZU5mOQJ8dcEeZue/nvSl+7AGDx5iXQj6rRdfj/qsSm41qkD6sgJ09JGclGxVKlX1NqQqyL6aN2uOxKsGVWMr3WMnkkUXiFhAhf4oJEhSsP37D7isqxwkTxqyLYX+zp/3q3vv5Bk7dDDT2TrJpspGx9hivGsbYdnEeopRk/RMF6sYQ7FoysBRBRaXROSAwm1AICAQEAgIBAQCAgGBgEBAICAQEAgInLQIqDzmKAIAldRIASP1jBYN42rrVWOPpnvFZVx8yaVYYdJt/ISJdu2f/kQUB6GyKCwkClizZi2qmyM8UyoaCB6NjKxrKTOmrEZa9Hw1DfXs1ct+mfOLjXtrHKqa9pFKbZQsmpPVMKQlF0XO4SNZfH0Um0xdCIDjCBVyTaqYtu3auyImDYGDiAuFFGtudRKE7VdDxNC2bVtLgIg5cGC/twipHKd8+bKE3mKZwqWi+VYZq127dWXm3kkV9jlYrb72LJ2kpCQvvokQVGoewioEoVILp4lIJYktRKpIxaKlCrEfp512OkKM3Z7tMoe5+0VsTa/9/VXWU9Gfo5m6NMSK5vBLsV3lZB/V7srN5fXTa1av8kyd4vlbWT6yEN1LVsxQarzV0FyJ9ikdqE5RJTJq4mhe7gjhsnXrVrc/qSRIddzaLfEIH0GiLESBVLyMGnWl75eOXzEqUuxIfOFhvmAt0ie/yK4moic1dYf1IrdW2xQG+ZA5d9xxu1199bVUoQ+09uQSSTiiRXk6Ite0ziJYijdbYp9PnJjRRcmiCzIWQiUlZZMlJiZaPxg1BRPVT0jAs9eFPvTzAGWpNyE1pZpLF7VqqFpjN6qPdEsd73E1anqNlYgbsXyZBPkKtGuuGg1AV1svart0IakTXoG+YjgVAjx61FV+0qJgtcTG6eTJZxcF+PXjEziRjd1aFU0okh5ryA9hm7ZtvDJsHfVaGzZugA+NEDSl8SCGJSAQEAgIBAQCAgGBgEBAICAQEAgIBARObgTgCjxTJaEBag2GdM2gutV9WYJUOS1FCEyEyeFx/oUX2DtvvW1vvP6aXXDBBU4qCIGvv55ELfNqGS18kdVF/x3GNlO2LA1EDPf6Vgy5q5nYgUZcebm9xXq6dOlio5ljYyEZfKjnOWyaWI+t5LBUYAY+ZPtQYqjmWnXYVclUPXTwECoampZY9jDT3n7bbTQH3efbUPyG2ox1FFLbqFGpQ4cOVq9epGmpNgSU8lPrN4hnZj6GLam3XTB8uEWVLWP33HOPPfvcsxA7VVx1E18vnu2y1+zPEbJcPiXfRsoYtSsrX8XJGQgX7dNG5uUuXTpbFvv2xZdf0LR0ETawEeS5zETp08zJKh2/yKd72U7W4YNOgmnGVm5MLhYnZe8UAqCOX+xGRYKQn37qb3b1NVc7qaV91XEogLc034+0KTXhGE6xFTRdqTGqUVJDqsHTnbhSNuxnuHR2k7NTrJBpiwLmUzKDOnfpKL7FFS/Kr/l57hzLJC4lV5k7rFtqmWPgIyWNjr+YhKpEGPPbb79Fps8lTti89dZb9idIOhFPikfx8+/7r2Mo2eWEiRldhBybpy5H4dvLQoZVPz7eXnrxZZc4ndq7D4RLHKFLZbyrPYO05UceftiuuHwErFxtS+ViK8uFksOFqGCeNrCEzZo19UovVVtHc+FVxgPWHSnTcGRbd9xxh7cn5UDqHOZDPeb9+/a1W8eMsWuvu9bOPutsa92yJeA6wlywdfDiDbK77rzTk5ebNWlGcnWUxUD4SHL22+Il9tyzz3pIUVJSYuQHs2QxDmsLCAQEAgIBgYBAQCAgEBAICAQEAgIBgf9BBEQUyHakrJDXaRY6hBBAGTNum+F7Cozdj6WoogJttTB8a6697y/32whmVbUpPcTc2q1bd/+2v14jphQzPFFqEBgEV1uI2NHSsmULuwLCQu2/b40b5/ksakF66KEHLQHBgMiSQtQZR4nlmD9/AUTGUR/6Rcrso+I5C4JEldr1ExuwjdI2cvRoRA9JNo7cmssvu5QmpVS+l2jxdPh8LQAAQABJREFUUgCxvQqVKmD7OWItyWbR55WrV1oO0R6dO3Z2YuGbb6ZQk13flSkKH/4Ky9YnKIKOEu9xNU1LLVo0c1JDh74XS9PHNCinZaSDm1qT2ACPq/HpwIEDllg/wdh7nCbRNhmiqucpvSwfsmLqd1MtDsGECCo9X9ao7Oxs375/vTcdZ02MRXHc2mkdv1RHvbFdJWBL+v77720ZmTAz+CxyplmL5pAgkVYkqVyuuuoaGq8Uzpxnx2BGJLTYjhBDgclp+9I8NPlRzlMqNjGdEhFJf3/l79YKTqBCxQpeo53H3L9g3nw/5zr/Wbhw1NYl+5T4jEjujnnOj4Kg5y9YSE33YnJ0Z9jrr79uL7/8spVhn3X8giUCju6U3AI6J7b4QYCA5D/ykGlX1cpUB7tQ82bNPOG6SdOmAHnMGqFUqQ5bKHtRs+YtrA6SJf2QZOzbT0XVYBjHHA/wrVWrjjVv0tybneJq1KDZaaEnY6vnvQZf62LMwMq0cN4CW4B/T8nNYvSS6if6SRvFBZzcqJGlbt/hPjSlacs/d+bQoVSaDfCLRvKmSrB06n/XRawE6ybYqCpjtwpLQCAgEBAICAQEAgIBgYBAQCAgEBAICJy8CGhIF4mgRTkhs2b9BNWhP98XehuQ5tc8ZtSyzK4ibKRGqUAIrnJKP4KgeAflxHFmxokU2Vxx5ZW8itfCv5RmttRwLotTXYQEPqqL2WBRMK5CaOUmkYpj+Yrl1gsCY8XylfYQobWdO3fyfRDpofkznTl4D5YlkTExxG/IXlSnTm0KbLDjoFqRmGDMbbegAOliM2g1fvbpZ2z1qlVUf7f1um/N30exQ0nkoODgrp27OKlyAOIhKSnJc2D3k2mjQpy//vVJBAplsDvtYz+62JTJU2imusT6InKQ80QtSjqMO//8Z6xN6b4veiw3N8ePN7osKh4UNmqlEjkSVTrKLr3oYvtq0lcuqhCJJRWMsNHxFauIUiGT2hJsrFBloS/FjDBShMmws85x3PLA4rJLLrP/w957QFlVZdvfqwhFzjlXFTmDkjMoSBJQEEVABQFzQjEnzGLOCQUVJYoZRMk5J8mhiiQ551BQ32+uW5e2ff36fa8p+9njv7ddVNW95+y9zzzn9hhr1lxzPvvsM2BY2k2B3QuWeZRspVjy+ITSliMLtTqTZ0d9s2LFStKZj3j7lNqQpk2fbtOmTRMh4col3cf7H3gAbPO7eucM91rkkO6eiLMyZcqAmVKr2BBfIsLi4kpZW4yONYm6dO7rf6/ddded+M9cZd//8KOVTSjDcQydk8bjgokZR51NSTZVFKWMbpLvFkDEhAlQsVFnzyX7xVeuXBkGLNmdlSUH0gWrbak7DODcOXOcpdtPr1y2nNnIP8/l53711Vh/fSLRV9u2bbOMKF7kHK1M8URap2QElJiUhGTqKH4ze2wHsdhZMmfx/PbvvvvWRo8azU2ajmfNJu8tUw65ZFQFCxXwD4d6y+S2vI/eOlcApTHIYbqAQEAgIBAQCAgEBAICAYGAQEAgIBAQ+Pci0BrTXnmLapxJPu1FfnQHmejMyIwPiwpzkSuqD0sSqX1Nt25eM26l7nzq6YGY9n5jnw4das8/96wrcFTaayzGpuNSOjPUquPkDK/JEFdJQmXKlPW6UkSI6tcru1xpr73xqrVt1wYSo7C3UDmDwJm7SXVS/ZonXx5q2KzYbKz3JKMC+fPblKmTaXE6Zf3732OdOnXkuJ02lvadhx56yMUKtevUcUNg3z/KIJkFb9myGYPdRPdhUTtQ8eLFXciwZu1qVCY5/XqlzLmWdp2HH37IquArU46OlcpVK3trk2r4wxAeEj6od2f//gPeHqX2Jl26rl/zyktH9f+A++/3/d/U7yb7iXSjYrQjycPHlSgcq/YmiTN27NzuOAlvrRG1J9Hvai1Lhi94nKQmiSxEpIk48d4hkTzsJS/3sVXrVlYGgmYPNb/akpo0bWobkxKdXygM+fTjuB/tI5RK0aE2L8WhV0CFo7nESWg9XZuUPauxNdHwNjfeFzkzfcZ0J460/3TpY+xNose74z8kAccLL74QuW3RBdLw+4UTM1wYCOOALGPf0vRp5ab37Ags1L7IA+rvKyqLC+ZQya30i+RQUq3I3Ei9XGL41He3fPlyQNpj5VC4rF23FuOg4zwIuWzn7p22bNlymLEVLk9q1qK5M3LHYQdlciTPGSUurV61ygbcO8CmTpnsD4HkYorYVirT0E+H2AqMh2qwpvrXRPwUwmypUaNGNmPmTHvt9dc8IiwN8Q1TBQQCAgGBgEBAICAQEAgIBAQCAgGBgMC/GwFqzg4dOlADDnWywJdHFaPyWwXpfvxM5XV6FmJF9amUMAUKFrDFpP2UpvhXbaquj7vxeLkZn5HrbrieaOZPqGeze3uUUnpkzKt61dUmzKpzRCoUSFVpiAzIkg1zYEiTgQMH2oMPPGTpUfEoAUitOFHfGfmtTkftkRlvmZx0lyxdstTy5s9r2zHcle/qKmrcPv36QhhJYBCDN8sgfF6uctPeBvXrW2bULOkhN37D6Dc/LUubEC2cxkOnJN46CxcttNyE7ciTRtejcB1dsFQm8pzpigeNRA0TfvqJtKTLzuOBfS6YpPe2pIsvvthOI6jwAakSIZy2WlxcnJMm+/bsw8S3rRsFT548xU2VRc5I7SK433/3fSeNpFIS/ufkYQPxIXJFZJb0KrorMvcd+9VX1qljJ9uwcQOvxECSRFRPah3LgSlwF677MrptZN7bCVPjolxXuTLlILh2urnyJtqcNJQaLexEzH3y8Sep15lVt4KRYqNGj7aZs2bqF7//Wl+bkMdPDu6BFEsiarS/SZMn+bpLly71Y3RoWo8LJmb8uriAzZs2ez+bosgE8MaNiXb2zDnv2RPLJ0MhvUfDExdIbBkslS62EAbBWbkBGzZutPi4eCtfrhx+NBk5Ngb1yy6XZfXv3x9Pmuusao1qTu5kJAKsetUaTtAoaltAJbJeZmRo6oGTXGopBkEnjuPATd+aSBi5OouIEdCSUrVt3w42rpeVhYk7eOigmwavWL4iEDNp/YSF+QICAYGAQEAgIBAQCAgEBAICAYGAwL8bAWrU11573f1Gp02bYVKXxPAa//N/5G2ydfNWJ1bkQ6PqPCPep1KllC9X3mvWFKplnfPBhx9aN9p+JBiQ+qILJsEcbuvXrXMCRJNqhsOY+WbMlNEVN6KARGwUxaD38cefcBNidYBMIPVIqpju3btbfKl47cZPPkE3yPfff+eCBQkO1q1dZzly5/T2quW0RKXQRiTCQOoYJQqPpavkalpsxjNfrty5iMpuig8K3SqQOzI0lsKlw+UdIXpyovxIj/pmqhNBkVasFBdVZEWhM58kqZkzZmJ+vBrFyQ/u3ZqT69R65yCWNKZNm2rr1qxzQipqoqwk5rN0vjiYXP+8+fPtyiuupKVomg1FYXRJi+YRZRBTHIYLEBGmVi2tL/Nd3yNryFrEiRnmqFa9Kq1fjWxjYiKWJ+97K5jOURdOerAUyNrRZW1ao2B63sbRXiQiKjPhQZvpoimF5clJfHuUsjVx4s82edJkV9copUrKHqUuKZ5c90tKnr343/Kjry+FjIgk8RSZIap8Le5//vz5rFu3a2z3zt20oz3h6/8ZXTYXTMzoQrSx9Rs2YKS7CKlUMe/j0mspEdMZNzFay0MrM9/sMIzqg5PztPrTJGPS/RbTqJSlihgmCdh9e/dZxQqVLTtmTGLOateqZZc2v8QSEmAvId5OnDruzs5nAHThwoWwXbNc/VKlSmUrXrK4z5WZdqcc2XKQyFQQT5sK9tjjj3n01R5MgnSD5YFTqlQp1DVTnQgqjFGwCKEwAgIBgYBAQCAgEBAICAQEAgIBgYBAQOA/G4H1G9Z7MtKmzUl4qnxH+8617imj6lp2Gl+NHetCAJEdIgdOnqDdCduN7DkjXibpkGlkyZLVGjdqQrLPdP/D/lH8W0aMGGGPPvaYF+mq4KUu0Vi2YrmVLV3Wi34RCKr6pUZRe4zSitVGcxT7jTatW9u777xrzZo1c8Pb9BAPmiwZYcPcubMhMfa7EkX1rzpAziajLjlzytdRDS2VTw3adJbSbdK7dy9bt2GdtbikhV3bszu2HxAdrKMWK7UWyTg3c5ZMNgMPloJESes9+cPUgajSe9qniJL27drbc8++YE9iVPzOu28T443BMB0uGtOnz7SFixd6y4+IlUibE9QTP2vjZyE0NO8BBA9333U3PjUDnGAZ9sUwV9H4xcEP6Bh9KXlJxIi8ad3fhxpc/+XIlsvq16vryiDxBVMmTfF0q7179nsbmSxJotW6THo7opipglXK2vVr3bNnFeTSpsRNrs4pR62/BFNhqWZU/8vwt1nz5p4mFVcqzveha9eXyCbNe/DAYctHuJBaniAM3FOoIL64JUvEWU5IsshzwvF+3ZyQhuOCiRntRftSz5VYL7kgF8V5Wpf20Qcf2divx6JC2Qfwp/y9vHny2W/btjpAAkHGP/rOk+cTSfEiYOQXo9x39cRNhd3bsXuX3wzFiqn/TOlPl9EzeOkllyBPyu1snbLa1YtWv259q1u7DuZHdazmRTVxwC7mrVZxcXF28UW1bBn9gHNnz3WmUsRQLPIk8TGFefi0nzACAgGBgEBAICAQEAgIBAQCAgGBgEBA4D8bAdV6K1euImWpu32I6uUBopwvaXGJpwOrEFeNqnEOxYxqUnm9ZMuW3QtvOceoLlX7TLvL23mNu58o7CTahO5DedGfVGClGMnOQ7yM5tu+bTtigUIec62SXwW8ukYW006kViInAXhNHikPPni/3XzLLRAgCs2pgA9LPu8o0V5EWpxjwmKIHkqiArmIqOgfvv9BW3XFiVqBGkBMtGrVykUQv2Iu3Ov6GywXHSK33HqLJcTHc2QKpsPL7Y3X36S7ZIMLFNxrFcJB9a88W2WO7IIKNiZrkYFPDUQdcq01xD/mAfxn6rFGZKTQAXPUjxUxc/LkCdZCvcNcGlLXqC+mRLESLoYYOnSIDRnyicWzj1kzZ1Ozt/CWJL9+R8p4fYYLLWSiHFHhMA+KjYkTJ7mHTRR/cQGvvfGax5lH9uIr+ixFCBwqW66s1atTz44TXX5g3wG8bxqgAJrBZCnOP8yh7h89aqR3xhw6dAj/nMM2f8ECfyZk9uzXr2vga/GSxXjuVHXyLNn5CdqrEIqMGTsG5RP3L/V6I0f/bTdp8VOaEDN6DMvgCaP+NjlRywD4NOZK8+bPw1n5WR7YL8lq3wLIE0lT2kerUzK9asf8wooC5kFYLL2fTB+clCyOClcXC3MYC1GSmIh5EaZJGmVIdlJvmkdcYTpUrmJFK0Eak3r0Vvz6q21Yv8FZt1x5c1vValWtV69e9Lh1QBkT7140e+gFPMxDpZ7ATfSf6Wb05piy5cqT0pQ9KGYc5fBPQCAgEBAICAQEAgIBgYBAQCAgEBD4z0XASQBIBFXwu/FAeeqpp+wmvGISSie4n0t56kgpRyIjQjBs27bFlSQyx4WtSSVWYpwUkeWGD8QIb2AIK5Pg0gkJNpL46bj4eEgFQotQ28h3Jok6E1qG2hKTW0QFJ06epsAnupnCXq9pJEPAfAmxU69+XdQ7p2mzuRdCJc6P0XExfA3+aDBx0JuI7K5ru2ixypsnL10ppCIhZjhKLHZuRBFeO3OdI8eMsis7X4kv6zJUIddBGJUlnSgTbUEb7eVXXmPFFFP7UXoixLNQRx/F9uMEZEZkgsj1p8fs9ie8ZqpXr24n6WgZAsHSq9cNrhriwMjg+s+QopQvf368ZqRgUdMWs3BZ8o5dAOmRC9/ZtevW21W0WinC+4svR3gbkdKqONzHrxBm0+hc0bVrDg1BI/JLvi4pKGZEilStWsVDfF584QXURHN9rfOT+Flm7dtfbj1v6OGdNxJbVKhQ0VOURIxNmzGNrpuy3qolwutnIrCH4BX0xltvQVq9Qepz3uiWbNeuHTZhwgT8a485/pHpU+ArDrry5zyJE4ErdfW0+RZ5Ki5kLt9UirskV6ANqVJlPeDpbNvmbe4noxt9DPXLISKtBcAzTz/tPWqxxIEp61xfR44ctjkkMklidIzo6k08fDI6OovEKRk2UC7M4mskzVFbUkYkTCtXrbSf6KfbnJgESDmcVUvivMlTJ9kMpGLJRGufPE7LFLKqIrQoFS9RzM+ZiYRrK/1nRw4fddaO22/FihV1Ni9P/jxBMXMhz0I4NyAQEAgIBAQCAgGBgEBAICAQEAgI/AUQUKmvwjwT7TK9IRcU5aw45VEk9h6i5WYSooGe1/f04t/1MjArZxAQrFm7Fn+XtefrT81xAHGB7DY05NahGlcF/lVXd8UHZal1xGT43rv78zo+qaT9nJYKhGN0rqw3MmeK5fdI6S1TXI0WtNWULV/WNiNQGImi47Ohn7tapWmzZl7bqv49xbmvQR58CkEi1YyIhej65WnVka9KChneEjeo7Wr+vPl23XU97e233yL2u4I98cQTTsIcgOwYDjmiaO50+K2o/m7OOkqCgv9hp7qoGKtUsbI1btIEX9ZD9tCDD0OofEkq0Zv2+muvWnxCvLcwxUAoaYwfPx5yB7FF6vqKElcb1ZTJk50wygABtBEf2VtvucUTrZ5/7gUbNmxYxL+F89WqtHL1Sl8/HceK2VI4kGbPgNBDyiGlO0uRVJj4cPnqDB8+3PaQYiXF0N+GcDarWL6SPffMc4QMHcBjtggdPdlQzOS0UydOMuc5S6LFSS1YifAHTz5J+tMNvaxbj242c+YsJ5R0/cJcRsu/5x9E0kgwoj1GCLPIc/W39dPmp8hTcSFzObkV46Y7eth1Qepn263Wo4zpXclSGNlW0aIlPE7rAEAp2joHcdiVK1fyizwtEgXfGX0QZHQ0deo0e/e9dzxBSQ/FGWK19bjorunB0fxyu15OFJmAq1ChEnNVcZOmAvkLuhuz+u5279tjt99+h+fQb/9tJ/15ybaX17agvtm+YxvsWD4njdasXutsouLJdFPDCAgEBAICAYGAQEAgIBAQCAgEBAICAYH/bAREjMhu4xR+Mo2bNPTCWnVq5yu72Ogxo+1N2nxeemkQXRTZ/A/6utpxP/5oiZsSvS6M+LnIFBbDXQJqvDCnvUmmvoVQZuzcsdNuvLGPzUHJ8fQzz9hHHw+m5i3q5MI5epGkjlHbUhNIkOSzpyPnp9abaiNq2ripr3MUccLtd95mAwYMsGpVq1q3a6/Bm5XamnWVrvTNt98S+zzKSR9v+4FA2LJlq9XBvkNkgmpkDXmsyhLkQeK0R4wY5f41n342FB/XSnS1pKPePsN6KZ7O9CpkSyniwVVnR85PcU/YRg0auoGx2pVefuklv752l7e30aPHWM0aNX2/UriIdJFXqwgVkSjCmqwlO4bSJitKl2R8cfSa2ocGI9Bo0/Yy30/fvn2s1sW1vI0qUuJzAcwnvuo45xYvWszvhWA6S/2+mlpdBE29evVdyaIOnMWLl9JudIwjNP5WwJeMK4kq6mbCfqrx/glXFuXDvHcK+9y1ewctWDIGFvFicA5T3QLlNN49HTp0tI4dOzghphnVUiVcNPNOTH/LopbSNer+/245HZpm48KJmdStyPRXXi6JSYnOMKplSPKjzJkyk4iU3ypXqYjz9LUeTS3J00Ecq/Uh8bYnwBA5U6FiedQrxZxh3LF9u23amOQJSwf2HoSAWc3N3+DtR8lIp8TqifGSUVH2HNncdLhhgwbWnoemc+fOVoU2ppz43sjNejKs3bLlS+yiiy7yrPnKPJhiCpW/fujwIZjAYSZCpzJqHyl4wggIBAQCAgGBgEBAICAQEAgIBAQCAgGB/2wE1Lajsf233+iMoD2GwlpkgVJ8pQjp0/dGu/nmW6zPjTdafFyc19xKDlZtKjJGxbkICv1BX69r6HwV7pUIl+nZo4f7o6jF5hqSe2qjalFrU4sWLTj+DDVnhJjIiR/LgQOH/PxoK9PpU6dRvGxywkcEwFHapz764ENrTeIQzUp2OSqcMqhFRE6wDfe2Eemh9WU2LB/Xbdu2OlHgCg8mKUEt3bdPX6+zZ8+Zaddffx1kxnF77713qZG7UHvTeiQGhP1vStpMR8mKCCkjXNhDNoyOp9L6kz9vAb/+ZGrt0cRK161TF3Ikk30+7HOUOxcT+R1pAfOocJ3I+fAzntKsdiXfY0oy383JHsWNT5063X1u5PHyzbdfE0F+M/W4am8RXZFmpnVr1lrrtq2xOImkI8WSxtyy5aWuoNkLv6DELBFdBw7uswXYpuzbu5fzUwdraStS28TFx1u79m1s7do12J6UcPuSXTv32NatW50DyItSSITaTjyFLm/fgXaxzTZ0yBAUN89aHsx/ZfCjdCYBLwsUefmc4DlwYsZf1bORtiNNiBk9s5JVZQA49V8tICpL0q4TxGTHxsZ6j5jUNBV4eKtipqMs8uX4wayiHeksLtMyOsrKDZQhUjX62eR8LFZz2/ZtfChOu1+NeuO+w/DowQce9ExzGRtp6GbrFojkEUNYuFBhqwdzeAmmwBdddLE/CIrIPrD/gLc8iZ2TG/eVXTr73vLkzuNGR5KXVa5cOdU1WnOGERAICAQEAgIBgYBAQCAgEBAICAQEAgL/qQg4iUKxqgQmFdpeOqqm1g+8PmbMGE9tkhpmIn/Ml0dLVsgJjYg6Jr37ssgMNwWbDQ3Vn6p/s3gozVn3RFV70XdEXXclUnvHjh1OzlyHz4u6OCQm+Pjjj913Ric6BcEc2oICbUQk6FVtKx2qFhn2NmzU0OYRY/0yqpYbIY3cm0UHcL7IHnnMyIclE+1VivrWXPqnILWwiIuCREjLP0eEUQ/Io29R3AzEY6d///74wmg91Cm06CxfvlxnQsLwxXcRTnt377GctADpGiVjEQ5qz6pXp77PI7HFc88+74SHyBitL2JISp6dtBm1JnEqSiZpblmRCBOlJP3KemNQ3rz/wQf26iuvuCGzfGREFmn93Xv32MeDP3YSROdmItFKkdc5c6ol6ZQHDv1K1wwIINBIJpVqmW0gHdoviON9zzqRvTeo38CuvPJK/ebX0PySFrZt6zZPyZKi59DBg9a0aVOUR1vwxVlot91xh/W9ua9N+GmCyVRY907Iiud44403bAdWKxFiJvUe+sxp90+aEDN6OA9yYUKiLCbARZBvHTt21KPGRMgULFjAWaqdO3bRJ3fKL2g/cdhi6dSmVK9uPWfGlpCWtClpEw9ZRjdOWrBgETKrM3bFFVeQy97Ef96AamYbjGcxPGPUW5YjR053SI48rBEzJbVQyVwpN3nuyprfRS/dvHlzyR7fY9tR4uzatcsK5MWsCNVM5aqVbdCLg7y9SSbD+vCGERAICAQEAgIBgYBAQCAgEBAICAQEAgL/4Qh4m805t60Q+6DC3dtU+N6oSWOrRCfFokWLiGT+EAJgEt8/sacGDrS8efM6oaLKUEqZn3/5xQ4fPcL5og9QeOiLNqV5EB9KJVYLj8iAdSg0ru56tb3//nv4mDxpN1zfy+eSXYeIAGdPVIH7PkSI4A0DKQH94W9VrVrdvV9Us7788st4ywx109yXX30llSyQCgU9DV9qxckMOaNxlnm016wQNcuWLsNftbjX3JKxSCVy++2329133mWdOnSytu3bYlpc2hUuIlxErGh9ET6nUAqpu0WiB12rWrkktLi0ZUs7ceoE1/SEff31N3Zdz57uF1O7Vi1f351hAHfnju3EV1fxliSRGIJLXSqqsytXrMSxJDqdPonv7DMulJA58Phx4y07tXtE3RTj6hW1HAmbk8KNDpd9eORUpLtFwCnlSi1Zv8EJiACSTcqRY0dc2XS+lE8t6cuWKWdXdLqCqPISzj5JpKFob2G+k5Y2+deIKEvhtREjhsM5NLMCBfL7M+GJzRmIMWeI7Nqzb8/5+6/bldYjDYgZ7SqFpKNDAHPOAb+CPPHu1/aw8ihkciMFkpJGMdTrNqy3NfR0qXesUeNG+M4QTw07qUjs+iQ6KQLsY/rPihYpZq1bXeagpENu1YAWpccff8IG8nDnxTVZUqdYztMHKwfOz3Xq1aZVqjL9dMfsHHvIwDkyDhYrqof0NBKsffv224HDB23p0sV84Abb+J/G0/921GO1KmCMJHmb1lcPXBgBgYBAQCAgEBAICAQEAgIBgYBAQCAg8B+OAClDyaT+lCTF171ZKMBViMvItRBKjtaXteYCY2iJ2W933XmnkyB333OPffvNNxAq+fxY/rElixfbAox1FRwjwkKGt6qBjxM2k9mtMFQTx7i3Sj4K+8eeeNz639PfqlavhnpmuNezHOAEgCdBMadMiSUikARHda2K/UqE6VQmhUiEiDxUviTd+KquV1ktOkHGjxtHdHap8+vLyHfUqDHanr+mvUlFchIhRBaIDld3cP2KxS5erLi9B1l0bfdrWDeT/fDDD9YF+w/V4lpf/ikiruS/U6JUSU2o//meciF2iI+Lo3WrovvdSPDQrEVTK0wtP336TLu83eUuqtD68nR97PHHIDGwNdEErH+G/ej1rL6nCBGlTc+aOdOaNG6C1+wpK1euvF17TTcEGrFOGPk1AbHIrEaoh/T7Lvx8pGyScfOadats5qyZkDGn7Sw2J0p2WrhwkRM42nZ0iDDKmTO7Y6jo8ew5snrYkOxQhLFsVxx/3U6ufyPqmxZ03oikEwdxz933kDAV6dTR+ymSCOlG6VsajwsmZiTkOcEDuXjhErxcJtniZUucFElISHDjIkVob4YhmzljlvfAHeNYtSfd0LuX9b9vgBUmMWn3rj12/OhxO3byOHKqZTwMJUwfiDat21hcqTgnWGRWlACjlS1rdpdfFShUANfqbhafUIYPQxZYtuzel3eOD0nePPlsEmCqVywHPjMZAF3MmljAo8dOeKT2PNqt9sK86cF69+33IGa226lk2qaQKv1vhlhGRa89g9mT+uX+2ZAi6EnIJcmzwggIBAQCAgGBgEBAICAQEAgIBAQCAgGBPw8B/UFfHUgy6c1GXXjWq3YqWGU78/Ouvbu9swNOwQmCTxAJNGrUyOKoZQfcf5/VvKgmxTsRzwz9sV8nnVNxjgeJfGgyktYTA8mDHsBr9UpEcPfANkPesd9+9517uygqW74zvXr39ohsX5g5VuOhenGdWhT7nMz66txQF8j6destP6SR0oWhAkiSmmFqpVq9ZrW9gnKmarUalikjLUxczGlIjyixIK5g/759bhGSnrnEi+j6Cxcu5nWzzIYnYYL7C+qfObPn2of42Tz59EAXNbiKiGPXkkbVpk0b1DGnfH2pWBSQo66WOvXqCAYIrgyILdZarZq1bNz4H20UMd3xcQl4tub39/XPcTfm9aYtulVojaLLRdDrS9cpD52mTZpaDDW+UqQOHDpgbxJf/QrtTeq40XG6foX+zJ0739vLDiEE0blZs+Ww/XsP2NHDR10ls//AfhK2JhMidNKWUm+LA4hgLL+Z9H6/pAoqXaYMrWIZrVTJknbsyHH3wRXCOjYGAikGzqAqxss1qte0W1EYKWjo1ttupfXqK7dL0W13Uk53Wqel8bhgYkbdcMnkta9Zu9rjq56FoJD0Ss7JezHjOYo65jAJSuN/Gmeff/a5S6FaXNrczkDOlIC5001avSbCeDXh5shBOTZDLP1xBa0M4BUtWtROYbSjubSWPhzKR5d506WXXgp508rZT73v/WbgJBZSkihlsLdp1w6znstgF0u4Osb7/PSQ8+k5h7/NbnroRo4a4cArQls9gP+bcRTXa8WQPfbYY379/+zcgcji9DVi5Ih/dliavKf89x9xFBcZFEZAICAQEAgIBAQCAgGBgEBAICAQEPh/CQEpNlydQg35w48/2K94t1ClA4GqSmkp6Po4dIQOjBxOcoicEJExnz/gt2jezENn7ru3PwTNAMuTB0NYnSe2g1pStfn+A/vwNK1J/SgVjmr1FK9d9x885O1LTGVbSU7q0qWL15uf4DPzEH6p8ofR0WvXrbVx1KwuDKAE1fkiHlRf5qdLxAkMTczXHrxXrqJFatSoUdby0hZO0BQqXDBVqcKZ7Etf8ku5uHZtFCoyKtb1pCMNOQftPhvdSFgtWEcwGe7br489OfBJu3/AAz6n6m8ZCh9BaCARwW7UOJpP16CWItmR5M2T319Lh6GO9paMJUl3SKiHH37IihQuYq+//rpVxy/WT2J1j7QGrl0IGWpDQClCPDoyYD0SFxdndTAVlv/M1s1bIa562W233WYTJkywEiWL+6HoKvDmSTL5zZ4mwVl7z5Y1M0nLe92MWMa8si0R2bQCD9tTrLF+40Z+XuG45c6V21upRPIwlRNmEnscopMmoXQC9z8irIiBddH9L1yoEGFERVHlHCC9aTWGyVdY9RrVCRJaZhfTBqXr1tOj/9J6XDAxoysW2XECMx49RL+uWOkSopwkJc2bO8emTJqEs3FeN/Y9C4ETi3+M2pyUDS/iYNu2iMGvQK5YvoJdxYMrE6O1a9a6ye9hHuzjx0+55CgGwqXzlfSI4ays8+bMnoMb9QHHRB43r7zyqt1yy2323gfve4/acXLVS9Im1RZyRiSObpicq/Pkyeu9c3rYtHd9GNT3px61/y0xI2WQ0p6EwygizP67oeubAFGkJ0K9dH/2+Prrr+1yEqquv/76P3upMH9AICAQEAgIBAQCAgGBgEBAICAQEPhLIaAiuittQGXKlvF9rVjxqxML+kU1n6cjUQ/GoiQRTaPiPFfu3NaRlJ716zeS0jPURmFU+8gjD7ufigQDUl54Qce/y5evIHDmUpQ2pyEvOJ/X1IqzYf060oblh4JiBWbhDF0Z99Ep0rdPH2KZOxD7XA9VRg1/f+liuk1Ug6rQ53+qSUWexEoRo7VErGTP5ia2stzQnkTmtG3TlrCdX6w89bNW9l1BLBzGjyUORcj+A5EaWYuIBFG0tlqRNISLauDXXnsN/5VOdKiUsga0C7Vt3ZYWr/S2csUKam8SiFLXP0FNLYNitYJpIUWFy4OmHS1MsdiVvPnW27YMdUnZsqXpWplIe9hlvh9fjH/OnkuGfNlpmzdt8pec3GAPaify9iTmkuro22++dWVQVgx/J/4yxSO1Vf+nF4b4zm7ifLVhqXbXvShYsLCtXLkK0mqfK2t++22rbd22hTl3YEK8y8UiCxbMd99Z+dJy8/3+61pE4IgbUBK07pvUMHL60fXPmzPXsmTCkJj/FMvdsEF9m0uXjQijHt17cDTvsP+0HhdOzHBzJDE6ePCAX0hCXBytZHo4YrgBu0wtQyV5ONq1a8+DCuGC7GkV7NNB2CmxX+vWrbPDmPVk4QZkwhgoI5FZ6bnZv23daj9P/MV+GPcj/WLziRc74IoXgZobg6UdtB698eYbMHyjbT29YPKtkaPzMhQiY3HXLoD8qzoyryRuoD50yqbfypxyYq6CRCkzbtuTp0521i07LKJukB6w/y0xoxvSvXt3/k2xqdOmukzuH92k75CyqX9OD6/ivP8dg2cvjIBAQCAgEBAICAQEAgIBgYBAQCAg8P8kAmXKlLXhXwz3P9JHCIzUAom6WsV5boiYZBQvziRQPBXDNLdilUqe5qv6Ul4sLS9tiRdKY5sxbbrVqVvHCQvNojai+++/3w4jEIhR0c9rZ1B1qOaLKmxk1KvWnCpVqtjgjz+x2++43evOEaNG2jXXXOP3ROeJBNBYg4omP8E5IjMkIpDiJReqD7XfuMcLXiob1m9wIuUYqUqqtTu0b+/ri0eRV8zzL7zg0dqqBVUHJ7Mn1bg5uVYt46QC3wuhDlFaUzdSlo7S4fIpRsMPon5Ry5MriphAxx6ibi9MarKnUul8/ktP7ZwrVw7rALFzDmLpELX6pS1bQWQsse/B7B5sSWIhUXQNAubTTz+1pSiWoqSMfGD1nkiWTHTa6DqFwGrMkztf1QWLlInWr28/SJGGziHoPXW7KMq6SuWqzjsoylvtSWtpCduGUigWrmHLpq22CD8gjdMYBD/55ED7BNx1TXnwDFK9HwHBbCrmydOmTYMUE9kELcJ+RJAdp1snV+6c58m7zZu3WL8+fe3ll17CbuXuyLOiBdJ4SEd1YQOU5Nws8yIxgu4DA0Ei8OIT4skYP+DvFSte1NuYZNJz9nlyySFzZMJUsEBBzJb2WRYIl6z07ykTfUHiBpdRyUhJcixlxyclJdHblmSJEC1lS5fxLPnV61Z7+1JxjHy0h87EYR1GmSIPmyKFC/sN2AYZI9VM7jy5mOek7yELBJD65b4Z+w057BdZp8s7+oOhmG4RNP/bcfXVV9uAAQP85o35aoy7Xv9xDuW/a1wGg5iPePA/DknmZs6YSZb8SmcAa9asieLlco8W/+Ox+l0eOnPmzLEk/HtEQmne5s2bu8/NsGHDMGKa7qepVevtt9/2nxVV1rVrV/85+s+vSL5mzZrlUWmKMZczeYsWLTxdK3qMvouhHD58OBKui92o+S16ANesWeOSuEcefTT1If/9GeHngEBAICAQEAgIBAQCAgGBgEBAICDwf4fAhx9+aI0hVcaPH09nxS2uynAGBbJgPma+1113vdeO2qFIgxz4lqouU4uSiAMZxM6YMcPqonJ59dVXSe4Z4Ua2eRAK7MXPRV0RPjhX9a9abLIRo63aVFSN5hRJoy4G+ZIuWrjIFsxf4F4uw78cbhXKV6Sl6AnOFG2ECodEJSl25s+f5zHTnO417br164mbLmc7du6AFElP689mVDNtCMh53D4ZMsRTg+Txspna+QxkRzJfohm0/uEjhz1sx/1RIG4iSiFSqRo2cluP+Qvm+XpKmHrm6afdaFiR2CJztCcZ9+ZlbqUj+TVR80uIIa9V1Y9A6V/H6J7p2LGjk1WPPfaoq3SmT5/mliQcAlaHmY//OPcsZJHazGSvk5H6W9ep18sklHbVzD0QIKrL+9x4I61k99vrqHvUnaMNFSpcCAFGNfv5pwmWlbo+oUxpVEf40aCqOUQrVuXKlf1nKXLEObw46EX351GblPxrZfrsG2ZPqpUja0dILJE/2kcs4UURUikFHiGPR5YPGjQI0UeijRg+UpeT5uN/z0L8YQu6EfKTaYSjsox8xVSdAYCzPMiSjeXJncfZMrF3+lneJ4thsQSSosXU2zV/wVweos2WUDbB8uXOixxpr8uKGvKwZMu+lCz1vW5AdPLYcZs9eyYPUUNr2qypVUSOtR4SR6a7UsDUqFmDyKuyzJvTjiMDW7RooS1ZtMT2cH52GL3jnC9yRn2EYudikZrpQatQsRwPQGN3XM6Wmlv/h8v8p78WK1bMmjVrZlOmTLaRI0f+F2LG25gm/OxzRNQ1f5tOTO199913njzRwy9MNerVq2+jR4/yD1LkFZypkbD17t3bxo4dG33Jv7/44iC79tpuduutt3pvXvTN3ci47iCTXUORX1FiRnIw+eK8BPPnH9LURbV+bKZM9txzz3nOvZ/IP3P5oD788MMQQK344O/1exh97zbMkf4R2RR9P3wPCAQEAgIBgYBAQCAgEBAICAQEAgL/bgRke9GkSRNMeN+zd955x//IrPhmFeObNm+yb777xgUB8lFxgoF/RACUK1teTIEbwqZLHzHFve666+zOO+60hx99xN5lvrvuutt27txOLYWag390/sqVKzwie+XKlU5qiElQ3LRqN/mnqhZW/Pbtd9zmZMHjpDeVL1+WJCURKWplOmlDPx1KLbyLWrqcr8/svie17+zYviOyT449dvyEPYTCZeas2bRafW4jEQJc2+0a/nC/iWP4jwlFhMj/9YorrrDNMsXlukU86PqVklS9ZnWvxZVm/DResetQ4wyCyJg6ZbqNwIdVyUk6Z/jwL0lMLqCt+J50vcmYD6fYGd+P7msKk0pF9Pxzz7vg4Kuvx0JijHDvmMmTJ+sQnyvyXW1X1PBZUdWwtm+I89UdcwosDHPmqahZ1PHSo2d392n97LPP3IRYe1WnTSuIqUMH91v+Avm5h0cd21OnTnL9SV5Q58ZORYol4f4NKVvyjHkUQUHLVi2pm6+xL7+U76sIMQfF97Br125XEimlS/dD+yqMl48wlMhh/twF/jpnpPm44FYmWtl0KZj45rDmzZp7LFgShjtTp0zxdKQ8ireG1UvhAWx3eTsbcO+A87ItkSKSj237bYf9QtvSOCLA5tC/FQs7lhWmMSPqmUIoajJnibVNKEM2ioQ5dMzZK7FdV3a50qpVqeokweKFC92XRjdKHjTly5Xj4SloK1CgjPpqtE2ZPAn5Vk7vHTzCh02ESDmOEcu3kYdX7GZevGdESvwrQwlRuktSn+gB+v2QBE6qn2z0B0oF8/shAkSKlsz46rz44os2g4dfD12RIkUgQ+bY7TB7vx8iU/TBFtMpidiXX37pD1heZF1ql5J58kuDXvI+RJ2XL38+PlyDfG65XEfH07ChWk9M6LXXdrfPP//c9yGzKGFz7733unQvenz0+4QJvzje8tV58MEHXamj/3MLIyAQEAgIBAQCAgGBgEBAICAQEAgI/JUQULJvMnHKN910k//xXLVY27ZtrT4tMipilQh8mj9Yu3SDjR87dtTrLNWfIgxEYGSKzWp9+/blj/17bNDLg2wj5IU8Q6dDHFSpXIW2mgy4k0TGQfxR5cWSiGWH/vgtUkc2HVKyqO7VEDEi0uTtd972tKRidHd06tjB/WdUUMrq4/SZU6nrKwH5tKtH1G4jg9r0arthtOM6pDoZiSGw4r3zUcsux/S2UZPGrrLRNaXj+k+dPOUGvBup0TWi65+mLUrhOdqLH0t9OZKQmi6du3D8Dnv0kUfY3zWskRECaqenSOlCVf+rhlTysbc3AaRjxTT58xfwuO1J1N7PPP2U7cMgedLESZgN97PspGJpRNefMm2Kkx0nEE4AmQ91tYz/cZzlzJXd5xTB1b5dR/eDuf22261N2zZ+/VovC3V7ITpeykKiyaz4xInjHnGeSJfNBsyC91Pfnzx1gpSsHqiNyjsx0wefHyVPffb5F/YU+8sLVxHxBxII5nioc+XwkYO+vl7LmiWHzaRGL1G8hOPPCXo5zUf0GfrXJ+bOOLjQZ5JVyfDo6PGj+MIspJeukrVq2cri4+I8fUmmuxfXvthjp/LkyW0rAHrJksXu0HwSdjBxY6JLx7LSCiVmS27S+2mFOkN/mB66PXv28LULtm8TztKJ7nKtD9uBAwdNZk7vY/o7l/aeAzCRih/by/GHeG/x4kVkkU+2hPgEJ0C2bt3ibJ4+HGq/SuQhff+99+377753Jc2/AsaVtFFlhA2V5CnathSdR+7ZetY6dezkhFP09V2oWaJkyeCPBrvsS1FoPXv2NLUKaShmTR9OjU20E4nt1WQya5KcToSQSBa1JL388suuBrpvwH3+fzg6p0iRot5mpf7Ha3HN1pDsTOdqPPDAA25m1aNHD1faTKPXrlatWv6eSKD/OlJcDjh79mx7/vnnPflK3j5hBAQCAgGBgEBAICAQEAgIBAQCAgGBvxYCFE4QISrAP/54sAsEMmbMYD/yh/OePXr6H9a1X5ET+sP3b7/9RjpwISdzZH6rOjc2NgN0SQrJQrUp1lPwSj3g4S+7qUt//uVnFDLXQYREGlFUC777zrter0qJIcXGGQJwPAQntT1G9adG4UKFnbSQLYaUGuq8UG2mIUJIa+lYJQSVLFWSeVCnaE+pFbxIn46dOjqpIvuQVq1a2Rj8V7/9+hvqygEeMa3rEo0wYsRwCIk15+fWD+fIET+B6kZDx2mIPFmwcIHdcMMNBNdMQBV0lw0Z8jEKlUwoZJJdMZOeddV9kSdvbieh5C4rFZCuV6qau1AVZcmaxefuSP2ruvPll18itek1ultKs0pksS14t/w0/qfU+juCid46AjkmkUb0+tetW2O33X6bPfDgA5A07S0uvpSLK5S0pLo4J36x/UnPUneO6tza1LJbIcIUkb4dhZGUO7VqXezdJbqfU6dOsZsh6nT/h3w8xFVUun69p06XMV995SbMwkf34Rx+PydR4khk4fiLmfoTxoUTM6CREUPbQjxYPM5+MWe5CG1Xhj/5+BCo1ackCpc6tWt569DOHTtg/TJCoBywaVOnoYI55C1NNWrUcAPfjBA8efPn9WNl1isX5wp4n1SpWs2KFCriAO3m4V0E4aJf1PMn5+ssWbJB0ux3F+bHkIWJxBATqPVz5MjlN03tS/KqOUmKVM2aF7mXSnaIBUWAyYxYrU7/ysiLVKoN7tga6j2MDknhFNstkZSraqJv8H3evHm0Vx1zpU6UNIm+rRSpiN9NDJK4lf6yyC6pjKROEmv7+yFpm5jgv43Uhzv1wf/b6+YmRzKLUgta9MMffV8fAm994vRly5b5Qxl5jxfAWiTMF198AQH3rymLouuE7wGBgEBAICAQEAgIBAQCAgGBgEBA4M9EQByIyI0ehLUUJtJ55apVnqT7Mwk7n372qT377LORFh3UJZJtKHCmfLnyeLnsUtcOxAOJQfy3FDGB1CAiXlTn/ooypSvx1cOGfeEx0S0RI+RXqw9DRbyEBar/tL7aa3LkkFqEX3w2zWhO9FxyySUeFz3hp/HeIqVWmyFDhlrxYiXcGkTrqz7NTA179EikcyQGAknzyJT35MnTTijp9117dtsdd97h7U1Sl/Tp2w+VRzGOpe0I1dBxJ2G0J9KHqOsyQDjJXiTipUL9zqbKIo6Q786+ffvtHWrpp5951hOgJk+a6LWz1EBqWRKRIXWQauzI0OvmteJHgz/y+Gxd/0lULGq36gXRc9llren8+Boj41J+/fpnw4YNvr7IHZ0vXxgJMpT8FMVf81/VpbN9+OFHdmXnKz14SIqnvJj5in/QZsRFXNamtX8/eOigE2fJcBIwKRArZ1PTl9JZAh42uoGKLr/6mqvdykTdKHEQX1pfY1PSJs7RfngBrPaDRSzEm2LOhZX29WeMCyJmIjeR/bLnOjCI6imTOZIMj0QeiFUT8yT2UcZJefGPOcNDsXDRIm/VyYa5klyPM0Ds5M+X31q2bGlSbpQpUx5pVAZMi2TYk2xHYc0qV6zg8c+dr+rMQ1/Ib5pMfoVMebxs2rdvRyvOW272I8Mjfa1bv87ZvGu7XWs39L7BH8icOSFoUMrooYqLi7dKmAMVQ5ZUtkxZ7ye7EJDV06excMECNyvWz99//71/mPLTUiQW8/fDH0Re2A+ZlBAfz37irBRxZfHsSySVotP05Ii51YgqZ/RAZYFU+eeDJ4778o+GcNEoUxoPIMyM/jhk8KsHVg+5TJc1YjQXv8unRq1iYQQEAgIBgYBAQCAgEBAICAQEAgIBgb8yAp62Q1G0Zds2DzgR/3Ka4v9aYo/vu/c+N3Udgnlu8eKlqCspgKh3xowZbUuXLqFlJ2J+KyrmBASI10NcbAreKiJctm//zS0l5D0jw9nJhLPExcd5/RvBRObB6b27o03bdv4HdhX8Opl/XbGitKZIUlB6e+2VV607dWu1alVoTxpptQmpOQdxojHsi+G04qzynxXrrRlEOHgysoga9iqy5cjRIzZ48GCvO5W69AuR2jVq1PQ4ap0cXV/bOHPqzHmf0HPnmJPXitJtIWGF/hifkbYiKYs6XdEJsukMpMpXhO109jpd60+ePMkSaRnSkG2JykXV2Kpx5SUrFY3wVzrSWFqtLka1Il7gamrmLld19bRiP5l/dJlaX90ulSpVZJ/cKM7X/ZBwQBxLw4YNbOLEifb2m2/bT+N+dIPmhIR4n9Nbm7DXaNnyUouLi4OASfCQmqz4x27HNiUmPYbJ27b6vZCERO1LMn9Wfa5kposvrm333tPfsmdVJ4juDiN1/Z20SVWpWh1eAR8gztP9/zPGBREzIl0izBJ8mEBH0vQlxkAiOqpVr4YvTCRJSSZHMvERCOq1E6EiT5WcObK6xKhz5yvclEk3T8a8kkXJJGkz4KmXq0rVKhgDZbFcOXNaVVQzBWArZd576MghOwYLJ7WLGCypckqULMFrJ+xmXLcH8GFr3rSZPxhly5XFQPik3+gbe/d2OdguzJq0FzGY8nQpWqyoy6L+VaAvJ5c+J3sUO6j2JY1IW1OMdenS5b/MrdYsDT2gOi9X7ly+1xzIsfR7Ffxi5OVSu3ZtP05Gvhp6738aujf/ncpK/ZH65Mgk+R+NqBpGj5xw1tBDqA+bk3H+SvgnIBAQCAgEBAICAYGAQEAgIBAQCAj8hREQCUL9eRh1h0gL8oCcLDjHH/9ff/1164wSo0SJ4nbTzf28XlP9Iy8WpR/pZ5Ed7qdCYe+1L695Tcu7pfkjd17EBd98+w2GvZ+p0rVZM2ZaV/xnslLgy1CYCdz75PCBQ7RAHeRXESiRaGihtgyPm8xZ6ERILdy+++F7zr/aOzmewq6iS+fOOgx/m93ujeKUQCqxkw7SR6qQdBAo4gq0voJsKlaogM3HSnv+xedtKqlI8n69/dbbvNaMXj9TuIWICAzhE10/M4bASgpWqI+8W2PxY5W/ynVYbcjP9ImBT1pbBBEZacs6evQ4xsELva50roI5JawQSeJR2doRr8nq5CLIIdW+9erWdsuRIUM+cb9TtWhFdq62oRiCfTZasybNqNtPcHYEf9Wmm7H0KE4XjGK6d0CU9Onbxx555FGLZ/81laCl+8R1qONE1y+j3iLU9vkK5IcnoNbOTn1N7btzxzaegRhPjVZalzDr1esGN2JW18gHqH0qV6osyH191b+KTVdcudqctKeoskbHpOW4IGJGG9FmxYItI9pLgB1jw3qtWJFirowZRxuP+tM2oNI4ifGQrsYvSHcPQkIPeoniJa0ZBIpSk3Zs30lv2C4nAASiGLvWrdo44VIcpUaJEkRj0+qk804j61Ir0DHOU79XPgxw1SalB1emwVeTDf/0c8+6RGvbpi0wgMWc8KlAmlOpEqXsM9yrlf0uaZfYzAyQO3og/tUhFYsiwnSFamfSzVMbk37/YxqT1ijJDfbvPJDLli93DJfzXV9qI1rO1yLURXXr1vXjiroUzVzu5i/8k3+czftv3nfFC/DvAKt/NNSb5wMoonvU73/WQxhZLPwbEAgIBAQCAgGBgEBAICAQEAgIBATSEIHUAuYsteJpVC9QEJFalO9ShcyaOYs/gtelQ2GbB7A8QuKSxAAqfFS0i1hRMIq6GvwP1SIx+IO1quDixSB0+vV1E9wdqGc6tO9gP/z4o73/4Ye0SD1DbYqxrDMWZl8MH27r16739aMEj67yyOEjkARZIuQIv8ubJSkp0bpjWDsa5c5NN99MmtTbtFHlP68iEVmkkYs2oix4v8RoP/ojOnsuFRdH4M7lEENZ7DDqk4cefNieYS8iMT76gBYjyI1onajAGoXvpOhadV0+UqhhD4NB7vPXr5dl/iu/machZmpUr078+E/e9iSlzu/XP3DwoBVC8ODzyd6EadVx0Yo2pgL5CuBFe9x+oo1M0dXy1lFkeK1adbg21DGMo5BnHw3+0FOWosSW9iuxgJv0MqGiyNVqpcCby9u397jt/PnyOR6aQ1cijDMj9IjNmMnatmsLkQROqIJOqyOF97IjBqlfv76TMGe5R+vhKm6g3ap06QRPRW7Xtn3EU4f7L95BQTmJiYmoZTjf779WSttxQcRM9KZK6SJWUT17GxOTuMHjbSmkgtilI8RgiWBQ2pKATgeQWSEwcrsqBoD0H+CI3dJFHz9xzFU0YsN69+pF71heWpmOAHhGwM2MTKyKt99I8hXL7wJW64sNykWL0lIIIvWuyUNFHyK1NEkCtmP3TuRrzXFYPoLMbKLt3b+XiK4jtgevGvWyyXtGbNyFEDO6NdcqnYnNLF261FnYE6h3RG7I1PePQwbHGjI+EjP7P434uHg/RPKw9evX/5fDdb3RIRWOMJEXzx9HfDzz8J4eLpkG/3F8NahRE2gAAEAASURBVOYrf0ku18I/Ov47BU70/fA9IBAQCAgEBAICAYGAQEAgIBAQCAj8VRDQH/pV0MujRXYkauWJ1jRNGjfxGk12HO+9+549R7BJf1JvP/74Y9qJqvLHfdQokBYyus1H4a/6SiSG5lMtJRJl8pQp1KB5qCHTkby02e697167GwLjztvvsA8/+gjlRsR3Jpk2mMNHD3tNH2mvAiFNQy0rDxz9pyHSo2vXrv7m4MEf26AXXrTmzVu4MXCjRo3YP+ek1nl79u0xCRfU1+BcAVOIwJmPj6lsMkQsHT9+zF5/4w3rjDdLvfr1bNKkye5vqrVUy2/BJNc3Im6Hn45SK6tzRQE7wkrXrxq5M8odtRQpYnoEJsXZ6Tj5XuoeyBWlUnmLFOvvQmDRncCZZDATscLloVTJZatWrrBadRQww17BVUSHYsyVICwzXqUjO65sQtHU2rffK65Xdb1q9cjgHvCDzJZlfLx4yRJr1qy5EzVSzpSke0YWKhJtRNffREz4CaxRzpxVOxpqJSbQNUqkULVqVd+j8J+OJUub1q3d3/WHH78ncauqpSSL5sFPB7WT/IecxPJX0v6fCyJm9CDpETqXfNZNYvVzMiTHGiKoxo//EZXMBr9gZyjpS9NDrwjsXLlyW7ky5VDKFLccMJIZ5XTtnxA9VJGeLz2w5ctXQB6WD9JC82x32ZJulN9QANW58rKRZOsQBsInRATB8OkGicyRUc/adWshfTJ6D90uZE9n2MOcOfNs1YpVgF3ZyiN1kipHN1ftUBdKzLSkTy36AXz88ce1VXf//kfz6sNVokRJHrpz1u+mfp69/vtbrKz5YcOGnX9JHjzq19MH7xZatZT9riFDqGfInZdkKzr8Z47bhVu4FDgaUeJGvXQiXTRuvPFG27t3r/+sfyRRe/+D9/gpxvr373/+df0QaWb6u5fCLwGBgEBAICAQEAgIBAQCAgGBgEBA4C+JgIgBKkwMfgtabAaKexVSqWRIgUIFUE00cMIlmT/kv/LyKyT3dLdKVSrbV2PG2MWpLTKqT0ePHkN70SY/V3OodBVRs5703Lx0bTipAALHMOiVqXCDho08tOauu+4kLbeNZc6MPyjn+LkiVhgSBxRDdaOht7TTuLh4Jysyo4SBr7HptBFJWXLkyFEnRDqoOyN1fbUYSQCh+juyfgrqm0wumCiE0bFUPyJ91LY1d+48J6FWESrTslVLDH57Orni6DCfzldln4TxbQW8XSVskG8ME7iNRj5atmphryGRxaakJLv2mm5OZAyGfHqODhVnN7gKXb8Ceg7QuqX1gRqiJIY694B3rGjvmkP1r1KWa6OW+frrsVihFLfHH3vc7UWEhc5zuQ3rixgrSbeLztMbIl4yc51KeGrRvDkJzIttKj4xb731ppUjOls2KKr/uXqfZ/q0qTZr1mw6bUigYn0Hlr3KxzY3Hrg+mFf8gwyVr8ED5/HHH8NDt5a98trrvo6uRSNCHjFH5Ib5a2n1zwURM74JNiU2UUoXqV7Ua1eVh1mu1zL/lUJFUVZKS1qwcBEEy06/0SJjpMYQ+5idni+8of2hUi+bbp4kWpGblsLNOOttTorE/vzzz2zBgvkQQccAMqe3Hwn4zVs2ecJQjhw56V2rAwEU6w+7IrLEsMlg9wNkZVo/N14uYgEVy52QkOC+NVVhxPShklLnQoaUPMq1Pz/A54+JS9H31Pr00kuD/EOtOG8paFq0aGGK3q6Gv4z21rt37+jh7Dt3aopSCmznJE+bUs665GGP8fCoJSs65E0jJlAPjcx8q/KAViLZSkMxaHIg1/+ZLKAvUK1NderUsXLlynkrlmLTatasYXfeeWd0usj3KL3896+G3wICAYGAQEAgIBAQCAgEBAICAYGAwF8PAYpt1eKqI3PmypFKYGibFPgk/2zevMn/8C0SQN4oP//yi9Wv15DWph12z933WOvWl1FTZaRgP2gr6DRIoa6VGoMp6dpIdoJBdaeTCLxWAIWMzHbnzZ/rBX4iRMfnnw+zh0kmihI40T/YL1+63Gs0EQ8ij1S4ZaR+W716HQSF0pTEi8R4h0MXvHC+wIbj0UcftsaNGnsdq/ThOXPmuLIksn7E/FfXJl8X7VJ0i8JvWrRobtu2bbXe/FF+PeKJl1562V5/9XXWKezr+PkAJXVJ6dLluF78VPhdBInixSWSqEk4ja5fBMbGxETrRxuXUqTuu2+A19+e0szxsjH5bfs2bcPXP3r0mF9HbKaMkbVS/40vFc/ezyIU6GO76L7pw96GDB0KfhexTkSw4cUsx4t8irRwoWABKqmGxD9of8LzNF6yg2htUmqT0rPqN6jnwgtdv+6/4sQXLlqgU32cwmdWYUNM4wOYmTcFz9li1NbFqJWf9yjzGjWq2XA8dC+99JLIgXIY4v7r2tJ6XDgxw6bUSiT/GMmNspPvfTHZ4YoMK1u2rBUtWsTJF218Cw/mWtQ0Yhs3075zmtab9MjL/NFmHvWOSfaUDSVMBogTyYXg+PxGwq058aAbqx43SZrq1lVMdgU38pHBUGJikhv59r/3XnoF69DXl8+NfdQGpSiwiRN+cWlU2bLlOC4nrs8HXGGjSDNJ0ZRvn54H7UKHnLllxqQh4qQ6krT/bogBnTx5shMxUv1MmTIV1vBr/+CrBerJJ5/8u1Mfeugh++CDDzBAzk9r2FFvmdJ5IlZ+mvDT+WNlZvzWW295epMitlesWPl3ihopZdRTWB5TZClpFpAkpfYoZc/fjYRv9uzZMLu0iqWOjBg/6UmWlC2MgEBAICAQEAgIBAQCAgGBgEBAICDwV0ZAxILCrjWWLFmKncWUiB8KRY0rH3hL9ZSIC1nBpE8fOfb0qeOoSi61cRAMauH54IMP/Q/YkWsVhRIp7+UPk4cWHBVJkVdivP5th6eJBAvqfvhq1GjUF4/bgw88RIE/AhWNuhYiR+/EauPdd991v1TlP2l17fnkiaOpJA5CBZgDfckP54EHHqQ2HGgJpRO8tUl/iI8yBBHFTIzJN6YAhreaR18ZqG3V3lStWnVXDR3GA0bpTj1RzFx22WU2depU9lzOkqnjRXKoPn/zjVedxDnniheqcd4TeZQNY2CmRDwRqZcVSPPGG29ahw6XU0tn9ejwdu3aIXRIH/GEgehKx/HyLxVpIzGHrtHJFK6pbr26LmgQSfIb4okrrugIKZRgUyb/Ym2IvuZwJ104xb744gtPyoquL8VNMl07p08ng3V6R1QR20ptUnqT/FpVSzdv0tTvhXDSOtq59nCE+y4STXWyXkkBJ5FgUuH06tWT60TRs2kzyqLrXNihurlPnxsjJJgm+BPGhbMQgC0Z1gHahvaSnS5fGSUrZckc69+LFy9hTRs3NcmuGjRq6G1Dp4jI3rdvn0uFVPz/DGGSmLgxwoIxXw4UNNBgtgbPGn1XO5NYPz3gejC9hw1YMyHxkuKmEC08JVlH+eMiKUTc5MqZ3bJBEqndyZUol7SwRk0aO+BZSYQSwbCbdKJNkESrV60hJmsqTN3O8x+rC8FaJIn8dEQszcdc+H8aTZs29YdH/8ewbNlSNwLejWt1UlIS7OrD/+X0fv36md7ftm2bEyq7SGuahzxNztK/HzpOOK9EsiZCTHK33w99GNV2JpNimQyrTUwfnFdfffXvSBmdo/9TUuvUEtrEwggIBAQCAgGBgEBAICAQEAgIBAQCAn9lBKSkyE1HhIgAKTN+w6A3Ukeya2pOvaaiXq0xOkZ1pwx7O11xpf9R/8svhtknH3/sNdaIESMhMlpD6EQUEzo+KTGJ+OgrvLgXCSLKJid//F+8eAkeL6WZDjIAVcs777yDYqWZJ+1+/fU33jUiUYGG0oulBtHybMmOUQ9mQIGjP4prPl2DSJkuXa5yO5Dvvv/OfkHVo7p4FHvq0f1aTeMkTDq6G1QHt2nTFnIH71FNyh6k+Fm8aKF3T4jY0F5/+XmiNWrcEOuPfbaAelXmxplS/yh/jO4JWWVIRaP1Vd9HUpaYUP/jNYkrLmnewkoTV/0Dkdr6o/66tavt008/daPhjLSNRc17ZTWijppoqpGfj4hBRI/ugbM97GnJ0mW0j12MN+08t9cQtrr26FiKwkgUmM5XdHUs4osUxBfCQiOGxCt5xmTLlt2u7nqNzZo9y+68+y4bMOA+PG4z+3VH7/+Z02fcIPkECc9aP4p/dtabO3eBz6m9qT5WHfwKUeavv/6mDXxqIMfq6LQfF07MsC/umd+8AweJAeMrPj7eTX6kjFFqUvqM6SwuLs46XN7BmjRrZnlQpsgsWH1daltSTNiUyVOccczDh+cI5IoULkd1LjetdHwchkB5XH1TFCXIWVqkBGK0dSeGY3RDZLyk1qX333/fbujVi9irNRg9ZULqlYV2nlp8sD6xxpgMHWFdGRapxUmyp6RNSfY1hsFz8Z7Rumkx1Kak1qPog/L/Z06doxamajxQYjb/p1EMqVUt1Em/95b54zmaUy1MalNSm9U/GnrgxbhKZfPPhgi3oJj5ZwiF9wICAYGAQEAgIBAQCAgEBAICAYG/CgI9e/SwqzDTlYrD62lqVxEeVLAeOFOwQCHv/pAhrQr+QoULWR5qymLURVJlLMK/pDk+Jqo9x48fZzf2RjUhcgY1yXbMY7ds3eb+KbpezSoiZC9/QC9XvozzDSripTCZPXcOvjMNbB3pPxXp+LgT75k81IrnhxgTxuo1qyF1EvBHhVhhaE8y3c1B7XrRxTWRJjA//qBXXdXVux8+x4904MCBXsNKNaNa9qfx46lvNzoJ5SQCcxyCHNG1CQO2yF5T7LdtJEl16IjdxwdWD2GBrC4KFop4lnKIExlaX+IBtTxFCCRm5DVm8bqwMYoUTao/4N9733026KWXbCAdH6PHjHRfHTlh6Mq+oB1oIeSQhvD3/1Di7N61h7jwzG514mtRp3e+ojPX9ITHl990Uz/CfyLx1TL20fVryOpEdXZ6CCxB53tiXxJlyEB5229bbBRqpaefehoFTG/7ZOgQ9+5xQ2DfUYoN/mgwZNJ634tm1Rz62gyHkQfrE7W26Xepah555GEium+0FpBRigr/M8YFEzMCWuDqgctfML97vuTlIcuSOaurP5bwMG/jgU0kdcjNd4kpU8RV0sZEV4Zs5r3atS623BAvsfjOZMP/ZOvWLTCa250hywkosZkywxrG8jAVscqQFmLbVq9aaVNQuchIiKdOSNL0xH98qCTRkkHTjBnT+VAVte++/Q4n6JVO0uTJlcdNgvXA34O5bTV6ALPADp6AFTx8+KCrf/4MoMOcAYGAQEAgIBAQCAgEBAICAYGAQEAgIPDvQUAFu6waPvrwIxv04ouusJCfaaShxmw2/izFihVxiwzRB/ItyUzduXrtGquDZYZzJVTsW6lXmyEu+AgC4+VXX8FKo6KVIv1HHMHnwz6zg8RSez3KZUnJobmyo9pQsa/2mGxZslknukfWrVkLIXMXde4OCIOBxEIPtqKpHi+Usj7Wr1tvZcuXI8Blj6+vGlsBNVtIfCrIH+5xuPG69yhEi9Qgt95+G5HdN1lHIrLVkqRFZ0yf7p0VmkDrKyZc/iyZqKdVu6ttR8uVK1+e9p4T3mo1D5WKCKwff/iRVqnSfpzUJVL96LuIlwOoa85KkaHN8l1iBwXNOPkDrmfwO33xxRdQq3RFaFDNHn7wIUQRjd3LZy3Xvn7tOj9X64v0YFpIj9PeqeFTsqfCtDzlypPbXnhhEOE2z1p8QoInPl966aVuuKzr91Qq1hfZJf5B9b8iu6HAnKyZSlKWyCwZGH+Pwqgp5JFSoUfSSnYJ80BcONYHEYPMxwtIw/HnH5Fxp1EwZUf5FL3/8rOtQGDQqBEjrMd1tDnpBH9TP6TduGBiRjfVNwcU7ZFNPfDgA66WkWzrEOoZme3upD1m7ty59hquxp9jWrRq1WpuQFb/cEhRc+WVXawOnjCjx4y1ZcRMp+fiz5DipPalhLh4i4+P80Sn4/jLiAQ6TgT1SuZ4+423nPwROyhlSB76A6M9b/rQnaXvTJTgEub88ccf3Uy5UeNG+LHUhQXLaxXKlbdKfLDUblWCr+xkwfuDlXb4hpkCAgGBgEBAICAQEAgIBAQCAgGBgEBA4N+MgIrtSXh53nzzzaTN3ms/076TT54wIheo+mSB8eFHH3s7kUgCXEYozM/aKf5gXxh1iXxG9Fp6vGcUGnPrbXfYnXwVxJPz088+84JfhbACbyKVvbkKJV9eyALqU7UM6XV5mSogJy4uwQ4iKkhKTCSpqT1kTSeSjWZ5+41oBe1J6oznn3uOLpDf+J1X3QSXBGT8XNXiJDMckSyqv/Pny2/vvfOuXX3N1b4/qWeupA1L78l/xQfHyvi2cKEirjKRgoXymG2low6uRDLwLZ6unJiU5PYjEiwo6emSS1pwOgc7D5NiwyElpk2f6fHZav3SGjLBPcO+pH7x6+W1DJBIo0aPti6E4Sh8523auO7pf4+3Zp2GuBElpPWd9OH4GDfgTYdqKXKt8qd9BWNi3YfldNU8hDfP1m1b3JT55ltutayZsjhSWn/unLk2Z+YskwmycAYqf0/JSpkg2IS/gn2mI9boCNZTpkyz1+Ej6tSu5WoepuDcKE7sjAmkjFLHDdP5HvRa/vx5rW+/frR6ZbWN6zb48+EHaII0HBdMzOiuqEUmM8SIYsaqoGjJkTOHR1ULJPXtqf3lICoWuUZ/8803NnPWTKt5UQ1r176tm/NImqU+tZUrV8Bqfe8SqvIVysM0Mg8PUtas2XiQzri78+rVazyiC6z8Q3CAh1vmw5ItVccpWjnpGZAXKUJLd0Y3VUZDOl8PkVp6+vbri7NyCzt2/Kj3tqm/TV8yLpZcKYyAQEAgIBAQCAgEBAICAYGAQEAgIBAQ+M9GQJXd8OHDXbmRO28u+2XSz/zRP959S/XeTuw1pJIQuQBrYIdJI5K/y3k7CrXPoLC4hsCWbBAsw/CdWUSqrdJvv8db5frrr48AlFpCbti4gdanZufbm2TLobnG0QZVKq4kBMZZr+mnoOq4iDhuETATJkygjq3u/qiaTK1DIi7ct4Y9nThxkrWzwoSwR9aJ6F3MGjZqRFtURbxSp9m3336LV8wCGwGB8oh7lEJUiJhgffmeNr+kOaqXQ6JQ/Nq1jspeRV+r/pUyZmPSRgiZS+1L8FIbkPxKlWSlZc9CGG2gDUs4ScgQJT2ihIz2JQhzo3a5olMn99l5++23bST7eYn2JkWIS3ThrUSp9XZSUiKmwAXdKDiKf1bURe++/x4+PPl8DSUvtWrZyl4aNMief/5Z90IVIRUdajXT0P71n3gBTa97pj0qQUrCjVqQMXfceTsGygNIfaphY8aMxpM2J2ey6dTzdf83b9pColMDDzbS+ZLpyLj4O/AtW0ZKIv7j3vwZ48KJGS5cqUdit2QM+zns4cpfV2JcdMR73ASKTIEUib1v/15/4E4cO27Fihazy1pdhiP09ShWitspEplOYAp8AjWMlC4FcUlWStLYr75y+Zn67U4hw5KMSqa32SF/evfuTUz2Flu0cBHEy2k3UspBK1Tm2My+pljQTUmb3dVa6+vB8ZsG8yXWU/OpPUpx0mXKlnXfmUDM/BmPWZgzIBAQCAgEBAICAYGAQEAgIBAQCAj8+xBQXe0lNPXq7JmzrUvnLjaL75e1bkPM8yOupnBCgS2p2JYSRT6oCbTPiAxR01P6VEWHAlKaQrioWD8McdKsaTNvk/nkk08gDJ63LKlKDokBfqAdSPHczqJoB6x//OhxFyuo3SZS16fYrytWuNplKBHRmu/NN96ABFBqU2SkiBQiaUg2HgqzURuRkwWIDTQkbGjbtq3/fPTIUbvzjtvtoYcfsvvvfwAT3qH4neZhdbUMnbMhnwwxkUZan4KYc/CjgZDakMhrXHf0+ncT5nMvdh8DiMBW9PSTTwy0uLg4X0OKFA21EmkfeSBhIibFiCGShVWM1+MXYeCr94Th8y887wlQl7S4BMXKFO+Sia4/D9NhkSaq/7W+ppcX0Epwkc/PWUgs4S/zYcVyXwcJptCc62+4wRo1bBSB13fEPyh7tL5SnqNJWU5u8UYuhBuF8BKSP+uHH3yEsfD37us6Y+ZMa9K0SYSEi9wmuIXNcBsZIOgOR0gk8BIJJxPgIsWK+nX7MxVdNw2/XzAxo41lINtc/XhHIU2++mqsvfnWmzaH1qWi+LtkypIJ0oQYK0CXuU/XLl0x8ulqu/fugajZ71IwqV0E4qEDhzhml00lMnorhIvytY5xI+YgU/p5ws+2fdcOHJ2ruLSqeNHiFh8XT9b6JnvwoQdh40ZaRpQ76gHTfvLky2f79+63+Qvnu5O2JEtZ2CNmzT5E0ughlYpGhkNFihalb6+gfyDTEN8wVUAgIBAQCAgEBAICAYGAQEAgIBAQCAj8mxGIodKXYkTcSJ26dSEmNtp9994LSTPTDXO/o1MjlvQgEQJOTFDYH6eVpVDhwni6bPViP8Wr5RTsObZ7q4vmUg2pzpDrERj0v/tuVDM30LqDNQdr6P0Zs2bgrbpLk0J6pPPAmnTUqSI01M0hPxQtqk6O3Tt3m8xy5e3Su3cvmzptipWGhHECRXUrbVQy9D1AW5CsPpzAYJ8ap4l/3pSY6F6sfhFc8Csvv2LdundzJc1dd95hDRtAYLDcihUrLTEpCWIBxYyAiUxx3n9G8+pAKWSUzjR06KfWr28/K4z4YhXerrXwhM0ASfL79Tdu2Jga/403C/Nqhmx0unw27DOPqtb1K51p2OfDiMK+0oN9Ro0aaTUIu1G9fgxLk2+//tbVOPKJ0fpnzyWD7RnnFtRGdE69V4zckCs/fPcD3TYNXRUzY+YM69mjJ/U9XTK6GL+eGFtDd81lbS5zzKCbmBKPH/Y0f8ECqwyPIMWLYsxbE8etdOKRw0fa3fj+qAVLsEhx8zn7PU4ylU7X+sePH/P7JvGHrt//0z1M43HBxIwwUKpP1erV+Kru0dV79u3xXj3dSJntSl4lIqQyMVky9y1XvqxtAZCXkTU98MADNmPa9FQJGdItbpCAXowsKUM6IreZO3eeXJYJpuocPWDFS5Swpwc+aw8/+rBnuJflwc0uw2BUNBk4JjdmwVWqVbHLMUDKkj2rHTtyDDVPUatXv75lRn7mjwybVka7DIFO8V0A6wEVQ5eROcIICAQEAgIBgYBAQCAgEBAICAQEAgIBgf9sBNQtoVGpUkVr366dqzOWLV/uSUvlIUYWLFxAWlJDSBOoEClJKORl8rtw0QJ+pkbkd9XgZ5MhDVBwqCjXC95lQf34Fh4qnTp19K4QtRNd3e0aL+ilKFHprnmlCClQID81ZyTlR3Wx3oyPi7d7iXIWEbFm3Vq7qOZFxFyftukE3FwLuSK/l6ioYOiQz2zFqhW+fuSKRBCl2AE6RBSHrb354M1xP46zrhjw7oQceuHFF+xuyCMteAYhhIbvnfXTQRSJIOGF1HljrGKFynbXXXe4qGX9unV2z113u8Lkl58n2R28LssSB4R5ZJ687Fcpg7hWCAyfl+vdt2cfKVI5XJET9X+dTXT1dT172tixY4npbmxvv/W2K4gOIewQThFuBayOn8LOJKvFQEgJL+Ev25Ru3bpbsRLFEHbs4/58aO+99569y9egl18m6rukr685FIk+YfwEJ32kCIref5ErSsESdrruVStX2c033WSvvvaavUibVA3aykTg6P6fpItHBJEfy+9Kfc6ZI7ev8Xf3X/tOw3HBxExkLylWOj6BbPfLrP3l7a1A3gLImXhwMSfSzVZkmByhH6bfLQe+MbuQiGmcoVftl18m2sjRo2DFYl350hHHahn5HkEiphspIic+Pt4a4uicLkM6Zx8zZc3k8qbCsJkd2rdn3oesbOnS7rotw6S8GPt2hpUrTX78kqVLyGn/2fsHtWYy5M4JetUkCTtMu5Wcr1etXo0nzl5/eHRMGAGBgEBAICAQEAgIBAQCAgGBgEBAICDwn42AkwWU2HtpRVJLi5MiXNLUqVOtA3XrqtWrCKAZbS1btYy0NkET7Nu3ny8SiPhZhb1IlvQZUd6kdl5EiIkUiy9VyqOc59OS88H7H7gZ7Ygvh3vKkUgAV6dAAhyjjalFs2YuRND64mU0MjKnlDjyY5VIQIE1HTp2wJP1W3v33XftFhKXciE60Dh1+qRtoG71vTCnRga+yx9HxIXUJfKekUer2psU+fzOO2+j/vicRKqX2N973iaVurSTMTtZWyoSDd8X3xWXvXzZrxBJBSyFa5eBr1Kfnhz4pD1Ai5T8YqRe8cHayzAK1lDdL3HLiWMnHLNYfHp0/cJfvjGduC6piPqjWJqGKOPa7t3piPnJo8N1jEgU7U3ESmtazdQSFsVf3MAOOmeUriRSRQTT3UqkwtRZ7VYjR470zhcX14DD/AXzXYSh9TWEv9qcdK6TUNyAWDBvfVkre/3111zQkQ9Pmx/HjbMaNWs4njovWefDzsgrt3HTxtigIOjgXBE2IqPSekTu6gXMqi0dJS1pw4YNDlLJEiUtJw9iMqzdMZgpESVZMCuSi3SRIkUsY6aMsHd7vCeuJ6xZnz43Wr68ef05v6R5C7vvvnv95lerVpX5TvlN0c0oHR/vyha5M0sFcxh2UD1wJ7hpuXLlJiqroE346SdaoLZ6+9NOortE7OyBFPqBRKbPkCTJA+fniT9D4CSy5yMwciecuPkWQ+KJOHbv27P3ApAIpwYEAgIBgYBAQCAgEBAICAQEAgIBgYDAXwEB1akRYibFTp4+Y3ux0vDCGiJAQym/t958q0dpX9XlKnsLO45YalUNPy6VQUkmPrlUyVIuNNB7EW1KjBWmtu0BwZCJ0Jkz/PG/S9cu9gheKHfccYd9jPdMxUqVdTikx1n79vsfzgsFzhf1tFmtWbXGcqJCiexT0dxb8Ii532655RaEBlfYQw895IoQnwfyQiMllZhRjR1RgaQSG1T2peJKWc2aNSF0cnPx6WzIkCHe2tSwURMIjBFWny4SCSaEwOKlyzwlWfuJIGIIJDJRK69zD1hYHycpDmEa/I6UQQgoqlSuYtdwzZUrV3ZqwlVAnO3EDuvvRexQsmRJJ1rUHaXrUqR1+QqVLI696XcF/nTs0AGyKyOmut+hhrnm/Pq7d+903xr5ymp9DXnpyPLEDZAjL2CDcsa++PJLvGYaOuHWHB7hklZEanNtGtqPd2fx88HDB61YcVQ1qUSN9qC5irNPkT3jIGTmz5/nHT/fffsN/EQfJ8ui6+tZWPHrCtuBz4x+1v1PfTS0VJqNCyZmxEzt4wYMxWDokccetQm//ITCpZRlI04q+XQy13OWKK98XHwWN+4pgsolFw+fjIxkYnT77Xd4z5rAXvbrUpi03ZaVFqSihYtDnJzEQ2azbaQfcMP6jQCdzl2qlXeu2C/1kI3jIZ848ReYMBQ2tCKdTj5Fe9JJW7p4qWVlD3K7jo8jbpvXdu3aaSJhpLB57533bPbs2cSkHXaSJmljErHeO9MM2DBRQCAgEBAICAQEAgIBgYBAQCAgEBAICPzfIKASPUp4qF5X50Rk6B2DKIiDsDmJauINe/3NNyAtGro5cDXsOTwSWoKJVBIEhYDbXqgwh3Xg7BS38xj7zddWsHABjkMJgwHuC8+/YNcQX12hfHn7ePBHtFBV8rVWQUashoTR+eeLen7Q+op2jjATMfivFnRrjeGQDn1u7OvdIzPxxBGRIV9Unc9SPtav32AlITuSz53x/ajtSTYiM4iH1vq6Zl3/V2PG2NVdr7J9+K8OxWi4VCkED1zPcSxERo0Z5QnGkRadGIQRyahzzqC8yaZL1gRO5BQhPnwe5IUUKrt37LTZs2bZjTfe6IE8Kd5vFWn7kmluq5YtXbGC1shVKiJB1qBMql+vge9b4UAyAu7RvYdNmjIJddB7BPGUsdiMmfweKfZbHTfnUOFofXnsxEA6nIEgkzpI18/LRJoXdjuT2nVqYdq83d564y179ulnPZTI7zAXoOsXeRQPTkcRd0SG2qPkf7vb26P02uEjh+0WSDp567yOCbMSpTKmV4oUpBWTTZk8CU+aNXAbgML9jz5XqROmybcLI2b0UOiqkQzJsGcDD8eM6TO9p00SKqlmlLaklqWYmAzeX1eR/r6LatWCYSzoCUvKCs8CgXLi1AmbO3se+fITULYsBRz1y8VixJPe9uzdbbOI2C4FqyWX5wm0JolQEZO2bNlSnK/H2S7kac2bNXfCZ9/efXZELBu9aRWIEGtHu1PWrDlcNqW4MB27fMUy1vrZFtJXmC1rdqtbv64TOWmCapgkIBAQCAgEBAICAYGAQEAgIBAQCAgEBP5PEfD2FXaQBf+Q9Bko0p0ViTAbdTHrva7Hdb6/xYsWW4+e3ens2Gk/YAp8Y98bnZBQi1EKta6isVehsNEQieGNQyhStv+2ne4P+ceogKcm5r1JdGLICHjLlt8Iveli/Uk5isUvJvmsCBSNyPoiGjzNiXrat8XrVWgt6tevr8prT3Z6/rnnaLuabkMgVB577HHUNTldFSLNy1Lq4EkTJ+FLg3dqqjwkBsLhIJ0lUvPAZzgxIgxWrlplV3XtbMMgPcqVK4vX6ysY/ebyDpIIWSS2I8XTj70Ni72IvNH1Z6EV6XKIoZKohnbu3OmiiJdfftWee/Z5e+vNtwj8KeZcgK5MpsjjfhoPUbLT148SGEqyKl4ikmokkkVDAoybb7oZVdCDHkk9avRIK4GfrFqVTjGPCBFRDcdoj1KisoyUNZ8UPzIVbkXCc7269ZxwmTFjpgmr3ty3zz773DmDZHyBomPk6NFg8Ov5+y/S6STCjYzpIwopzb133x5vQ1M0eq/eN1i3a6+x0gkJrJneFUBS6fjzo039CePCiJnUTWWIzeCmQ3qi1D4kV+N8GBydg5DZS3vQV0Re/woRItMdkS1SzRQsKLfrLTZ9+nR6ybbZETLjjxw97JnveyBODh066Eqbe3iQWzS7xNlKZZ/Hx8fb008/bZ8M+cTZs5OY9x48dAB1zmk3ddL6q/GMGU/P2pmTynzP5hFYMelTbAH9Zqt4KHfu2MHNL+JO2GK9JFlr07qtVate9U+AOEwZEAgIBAQCAgGBgEBAICAQEAgIBAQCAv9uBFTEa+h7ASKTI8RM5Pe8+IooTjozpIkUMuvWrLU+vW60wR8PJrr6Lbvl1luoI/GW4ShFNq9bv87nciIDckEdGZo9I+a7Pi9rFCHuWtHXRyEiuvfo5i0yjz/+uA3+6COrUKGiHxfdw47tO63mRTWoaVFgMI++8hITvQl/GClfpNZRZ8rteM3Iq/W+Afe5Wa2Cd2S2qyAbRUsrSUh1uMicI7QdibRRjLWabnTdInO6deuGSiSjPfXUU7aWNKIOHS+3EcNHWJ06tf2aONjXX0tHSmO8XZ3sYU6RI1lQ6ihBuX6D+k6MSHgx6OUXrXv3bnYpMdjTUegoZVlx3trHr8t/9eQjORdrfXm7KoI6A2SKhkgQjUoVK8EJFHDFjK6jENYkP40bbx0xU06fSoZofRFEBXlPOPsQOYP3bFY6csqVLesblypmBF4zrS691EpiBjxl6hT8Yqp5947OUf3/668rI+fz7+kzp5zcERklskfY6ysO8mktRswN6jV0tcwXXwxzBZR8gDT83gtrAZbG48KIGd9MiuXNk49I6iZW8+KLMAPK7Q+JCA9xidr89t9+w0RouT9kC+cvdD8aOTAn4w+jB287QIl9kyGQFC26WQd5Xw9cPdjGrkivFIN9mHx29fBVRAUjY6FkYrjlFXPw4CEYrn12CAmSxlFu/jK8aCTDksRJ5JDixQSogJeZkGLQevfqbdWRqh3A0GcNZI4e4H/HEHGkvsFBOECHERAICAQEAgIBgYBAQCAgEBAICAQEAgJ/HgKTUbF4F4eWUE2t4p6vbST7utGtXoMI2YWH6VMDn7KrUU00adLM7iSVKIGQGy/exXwwRBRQ4nr9WbR4RC1C2epKllJx8dagQQNvBVKtOhWT4RuI05Y6R8RBi+bNI+oSplL9qVpZsdGiBXxbdHxs3rTZvWJ8MeriQyhg3n//ffv/2HsPsCqu9Xv4RWkiitgVC1iw915i71Fj79HEFntirFETS2KNPYlRY429JfYaey/Yu2IDbCAiKCKIfGu9wzHo9eZGwfzud/97J8BhzszeM+/M8XnexSqVK1YWJknVgFFxpUqVtbfVU+KBPCHMeRN+q/R8JWjDCXneND0msOFF9gd2vYOUIZrv8hga57Zp0wa1iVbP11AQJRiOQxNeBZCwPiPHeQ5pU6fVbQRbaIRLZlDZD8pJgF+AfFSvvtSFSoXMJB08n9j1aaRMJgwBHa5PGIReMPnz5UOMdkPd/f79QPm4XTvxOeEjc+fMVbNhe4A5xBM4Dh06JBGwObFkTdyCcwBBIwgpTZyPwAzBktPwgikGhc2unbsQg/2F+sXw/tqGzobzCkdSVlpsV2kSZ8N1xqBWHrifjRAixAChFWDZbNm6TY2dx44di3vqatVEz8o6L9u8CfEzAYAZO+jBHLTYfND40B0/dULRPRegaw4AVHhTcZ2qUwsOCVZGCxHJzDDhyZY9m960CBj9UpoU8jBEHap9fI7JnDnz1GSHB/MDwIhrRnRNmDhBGjdtrDHaWbN6ijPAGqKSAX7+akhUE7nkZWFsRBmUMzR7YQBsiPKlwcPEh4Lx2e4AkIg2JndLLoF4EH6c9pOchBP2PzHoXs0P14wZM/6J5f5P1+jQvgOi34ogfWvb/+l5mMVNBUwFTAVMBUwFTAVMBUwFTAVMBf7fqkDG9B56wfzj/3KAEGzoLbbJC5Ud8Q/2VFjEwIKDrBlH9JVenp7C6OsBA/tr0vC2P7ZJ8xYtlKVhA2USYR4SDGpCTkOA4AWQGQI9LmDfHD1yVM1xCYqQRLJm7RqpWLEiFCKhsnz5CvjCZMHRABbQ/86c+QtCbULR7hLFgH0JfFUIMri6WiAAz4m+KtmyZ5cDhw5Iq1YtlZSwEvNQImVLSCIsgYtALx0sJUqUUCBFm2wgNzz++PHjCjDpIkAAKN1qAvIDr3MBpD9lSpdUdhBlW78j0vo2jG4JWvCaKLmi1wpZKpyTHjE8/+w4p3sgV9B3hv4r3347QsaOGS3p06XXmiuwg/kYF+4Cn5n7d++9rD+Pd4Aahn04X7NQvr5XAaR0VknS4MGDVSWTHLHbrCP7542bNyoDRkGUWJ8XEj1IvrAoIfCdQaoU52vevBmAleXSulUbmTljphQoYFPGWPc/GkDUbSRFE+jheQInUrzCOYmL/IG+ledLps+I4cOlE6Rln3/+uWzctFFSpwbLCqVmuRN6JAAwE6OGQ8xOd3R0luSuyRRMuYvCZ/TwkLz580mlKpVBgXKBMW+UXrTFpokR12QuyErPo/KnaMieiMYlwoMD/FJuXb+BfPJpchjyIz8/P9zQCPFF8hMNfDN5ZJL8efOBDZNBjXwnT50kxUsUV3qWC9ahAdCwocMlT768khgyK8/sXvpAurgmBUUqjXjiw+aVzQusmTTi5emlufNPQU9j2tM/OfiQ/ZMjODhYNiChiulU/8QgWEbJ2QnoH/+A/tEMUwFTAVMBUwFTAVMBUwFTAVMBUwFTgX+qAvUbfiSf9+op7u7usUuSg8Eu3FJZJHFOopIhAijszNh4M10pGYCRK5evwq/kV/kNQMVM/EF94vgJIBA4oTEHDACZDg1j16/fIHfgM8NOnV4kBAgeABzJliO7Nvxsftn0sw+rXKUyZFK/SM4cOdHn/ixeXl66piYI0egW/1O6xJ6V58Iz4n8uSV3U/DdXTm+5fv2mHD58SMZ+P0aGDvtG5s9fgNAc+MmglwblR9echmjs8xfph4NZwMABFKEkBkqakFekPTffopdL7y96A3joDElQLlkGaVOBAvlxbcRJUBDsxP8eP3kMQoOLglRcgydKcKsIQnaqV68JU+EglWwNGjRYGjZqpB4z2XPmUJWKHcAcXgvTofbu26vHJgI7RTcC4bgPCxOmS9Ejh9hUdPRzyKS+l6bw5qGnTb++/aRK1SoAg5C2DLNers8WOio6EnYklEbBRgUYAu8JbqQCWJ27fAZT5WeyDiFBffr3lUKFC8hmpDfXqFHjlfWXLV0KQ2arTgRbFBvD3H4B/mqpoueEE12KNKvy5ctrAtb27TsABGFdXlQCj3gDM3y0o6Pg6hwWrs7ONByimS4LngSIYUYYD7VDLHaGDBnVT+Y2tHSkPvEBBAqDm44bC6YVzYMTIcudF8pUJsZXkXIVGhqG/PSHiMB+rKa/P/44TTZu2KhZ5gQ2PEDVatSgsTT8qKHq365dv4aceuTOA/2im7QjGDvNEX/WonlLmCnlAwiTXFk0rkDfUkCCRVlUNs/s+oAmc+fD+g8MXDqHoojWy3/k+++//w6KWT1p17bdP7IekV7GjdWqWUs+Qna9GaYCpgKmAqYCpgKmAqYCpgKmAqYCpgL/VAXYP0+eMlWm/zxdm3arDcN3oA8+x3zEO08uTfQl0EBPF8p+/G75g1wAgAIgRhhsMwYMGChUAfT6vJdsQrSyA8gIdjEAQdDHUuURjLRgzkd2yTN4nFK1QeYMu3fKfhydkkjDBg11/oFfDYJ65DyYNrVk8aLFUrFKJeyH84ntys+dOyclS5SUMMibuD4lPzS6fRj8UPIVYPw204leyPfjxkvN6rVU1rQHgEemTJmV8YMd1E/lrkY7A5fBNUXA0oMqFnqx0s1GyQEvEJKTK7dky5ldZiE9atv2beKeMqUsBVjB6HDWKfoFAnywfiSOp4zpOX4nAGKXyB7Xj+0RkahXcmtf9PLbIPup8MEHwIISySkQAYoXK6G1IF7w5PETNfulXIjrs//ngQwLoocN+UtEXNj/lyheQlYBDBs0+CuN3160cLECNPS1tQAYAGAAskJDQjVq/Dk8dhRawPFk2NyCqXASsmhwf05BEVO4cDG1UtkEcKZrt+4A1+z1fpEscuTYUT0ROwBYXF8ZNDiOZs2MOcepa/3podugUQNhQpYTwDy9aByZkCPewIwicEDRniHqKwLJSjRIItiRAYCMmuTgYSAljAW4fu2aXAHrxT/AT1OTblxDRDUYMOpQraiTfgPqZY80pdygarWSLHjIMsIPxgPsm2AgiDTyWQ5ako/PCaVFPcS2pxFP8aDhTPCg0DiY6VDUwZGl8xhMmOtg36RJm1rRScqY3FIkx1wPFHWj2VNyN8R340F9HwV+482yLvONb73/jfgE/IPr/wKjq02bNqnW8v1fm1nBVMBUwFTAVMBUwFTAVMBUwFTAVMBUAK0du3UM/mjWvCl8R3YCLCiO5ptskBi5AFZJUOADCQLjIwaNPptzAhl3EQftDcYHAQSyXV6g8V+xcgVMcStK0mRJpXKlClLnwzpQZFitNNexrUXPmsxZswLsidQ1yDihV4pzEif1VHkBCY2/f4DUrFVD+88N6zdKO3irvADYwh6N7JPr168DXLiB44HXsG/DVzBYOERvuE4iUjswdu/ZLcWLF5MbN67DiLe1/kFcWTE4b+vSCYnYQfoUIjm9c1isGoAOCsxgjqyIkG7atJn+ftv/tjRp2kSBB4IzQ4YMVomWro9aUcYTiFq9XB/nFPU8Etfip+dCkgXn5blXq1lDFixYCLPiPtKzR09JlYomyzhnnBRPXYEZADGUkREn4GD9CY6lh3lyvXr1AH44ypUrV+SXmTPkJ1iOjBk7RhYtWqJGzbbr37J1s6RM6Y55Iq3jMTdrfQyyrRQwUSY2QLJIKAyRa8PqZAxkViO//U6qVq1mmQbrytY3Ak0ExygrszGLeK3EOlIj1Kh0mdKoY6j07t0bTKNn1vXEOT4hXgIuevehJ4sbwKIx6zsLWC6ukDI5xYIddEpmWtJZIH/JgF6lg+YrAiBKBBgzRBMPHjwo54EYUsNGdgXZLUQatS4oIilLRL1SwDE7f4ECeJiikap0Tj1i6EcTBXPfe5BMMTqbH4I7MBF+DuTrydNwBYpSuLvhQb0nkydOVBPgHEope4Eb5iCP8IAyEaoEnKjpCJ0P5kNEwuIzqH0jEkcAKgLX5+XlJQ0aNABKV/hfptXPGK71TePIkSPIn98LatV5vdYiRYuqmdKbzIkpF1qP+LajR4/hgxwoaQFi5QUwRudtBcawAGu1YMECnRNPl1LGeL/4oaAZUrNmzV6eBs2n6AfDa7kJ3SS1gwUKFpQPP/xQ5WAvd8QLfvD4wS2K8ytXrhzcy6fIxcuXAYDlUOdw6hkXL14ME6qb0rJFS/H09Ix7uBpd2da6gbVy4DhGxNE4iseaYSpgKmAqYCpgKmAqYCpgKmAqYCpgKvAuFVAAAgdaLZedeo0ePXpU/9DP3pQeIvMgsWFHRgkSuk+k9SASGf2kSxJXhNKE4GALfCmIfohsiRbNWsDQd6fMnjNbfvqxGDxIhkFWA8YG5kDrikCbO1IcvZHNX5NrO4Dpcfz4CfirZpJrvtcU3LiMnonymFmzZslsfJ1GKlEGSJLYz66E6SwhAU5ql4jsEDJVuAI2KbBinVP27DlgIHwRMd9t1ah2+IhhSCIqJH369NVgHVuvHv70CSRKBWXP7t06B+vCL9qIHNi3H+E6jtp/k5Xz2WefCZOZBiOauwSYO5QAcTDZ+B4IEElBfFApkjUTQBFIqDAXE5nok+OI5KRI9MF9IUGi2W/r1q0kZ86c8umnn8oN9I5xB3GAD8CwOXPylNaf4AhjsZcvXYZ0K3gD4ZJpSvwtkqRoaTLu++9VOXPl6hVL2gSgaNWq38DsIbeI3jc4BOcRAWKGA67JZuzLUjrBd3b40GGyd/deWb1mtaxdu1a6dO2Ce7lL7z8uSllA7Ft7AUyaN/9XS9mDd9NA3rZ//34kQOeV/Xv3oYcNhzTOApTiXk98X8cLidCHHQVLDECDSUM09nEGbegS0Md8aLBTp0wjD0NDJACpTPny5gey+CEisHBzceF0bH6C+GyyXS5duSze0NqRFUOJUzKAO/xAsMAK0qDIRL+yZ8sOmlYmOCU3BGPmuEaDOTg4qqsygZlLly7jQ+QMd+Zg2b59h5QqUUqY/nTx4kWAEvsAcmSQJo2bSGYASDQFdnImxemF5M6XR4YNG6YuzO9aUEp11q5dpzfW+thYMzHa+3s8RDRnsg1+SBR/i/2A2bbTOKpPnz7y00/TsAmz4CHiD46SJUtq7Div3zboLt4E+juaWdmG7ZDvvvtO3atTp06NaLAziFnriV2syZh+1bNnL/09fYb0L4EZes80b9Fc9ZRvWn/JkiWSLVs221LQEh5RAIZ6vcGDB+k9sa3frVs3RUdtkWz0BfL09Hx5LNdqARMt/qMU9zq5Qynk0S9evOiVtV4eaF6YCpgKmAqYCpgKmAqYCpgKmAqYCpgKvEUFbCAF+47KlSvL+AkTZNTIkRrrrB1SbOv18BHYJQASKGPRY7RREZjjllZA4uy5s3IMEqj1GzeAzTFTSpYqCUuNj0irQU+TSIGKafCPCYV6g4QFGuWSXMA+jx40sdO9/NmhQwc5eOigTJ40RebPmydDRwyXg/sP4MrQL6JXJOOFpAaa/BIQYBdpO7hw4UIad70UPdq2LVswxySAOivhvVpS6oN1wlaT/TrBHPaGlDtxfd3Gpg2DBInECPLRaXkA+vSJkyfJOYAmTCIa0H+A3Lp1C4DIBmXzJEUfrea7OJakAc5lx/PEF8+Wlia90feOwHVshbSJ/erECRPlDICnamCqHIN0SDEErE+wh3+Up8KF6/OUEkF9c/f+PQXR2Kfz+jmW4I/9BLW+G/kdiA8N5bMuXbT+7Gt1D60/jIaR3ERQRskeLBUGU50JOE2ePEU2b9ksCxcuBIBVWHaCQdW+fXslGnAdzhMINQ6xiUfAMGzPDO8h76e3dy6cJv5DPd/HsOC2+MyMK8iVy1s+RswWGTJExs6hyE/A5KgBilYhoIv0mnnw8AEMc9xxkY/0wsheYYw12RwpkdNO5M0Jej03pC6R7hQJ+tctRIVZxr94CHmzgYDhTkKn5yTpMiCHHv9lxzw5s2eXVNDEOQKNpDaQDJrfVq2SNevWqK6MN4b6OLJB+JP58gQYMuPhvHbtmoQAHeT6zIt/l0F2DEEZLy8vPCwjZQtu+OrVqxUB5ANFV2k+NLZBJhCfPusxs20VOFCPBijzE6huzvpB2A8Ec/78+cqaIQjSvXv3P3fGq61bt2otyXghM2X79u0yavRodZG+5usrI3EuHHSh/v77cS/jyKjdGws6GD9sE2BixUFDqqpVq2o8GgEygknMtidQ5IiHm+sTSIk7eE84tuE8CJSROTNg4ECpVauW0sysfa2H3HptfSeQVKVKFQVldK1xWAvMm759++oH/PDhw/+yVtzjzWtTAVMBUwFTAVMBUwFTAVMBUwFTAVOBv1sBBQO4M9sXfHXu2EnZF0UKF42dgoyLRBIc9ADyoOISDgKB1Zhbb/uC7UGrDfafL9CT/goQpSr6Gc+sWQCsHILKIAe8SCzOgx+AjDD0uNyX6z6HMS9BDBoGs6nXxh7nQNJBhvQZNY551aqVMB5OLWtXr1E7D1ujiMNViZEF8ij2sZzT1kfGQD6UFsE27H+5dQvYLTTkJYiwE4BIhYofwNoDfigYVFlcAeDAHRVMQVvN63seZaVRsc/m9pTo15nqtGXrFgUz/Pxuacz3l+gJdeB8eB7sZpPCtDdlyhS6tqY28fhUKRFZfRoJUF7Y/kJuQBVRrnw5WQOv043w5/my95faW/Ig9skrV6zSaG/b+lHAAHhe9J21ek1emSjp4xBMjzt07ADSxQX56Ycf1JBZ98P73Is9fxCAlQIAe5TkEXtNlCbxPMqWLcOpJAKhQhXKV5AfMMfUKVNlIhKf6XOjzwgmWrRo0UtgBicBk+cwxSDsHRjfjR1eb+J11vh/ixdjhsuzCHS5bg7pDP1ijgJBZJFpwsN306ZJK7fhVL0ZPiPUerF5Z2F4I2mmRMbMh9DoPQgKlmdRz8BkcQTK5aDHP7j/UPYfOADJ0nMpW76sOKFgzwDs8CbpjcIDZY+HPHlyN5g55VAGTjgKTd8YIm2kkgU/CFbpknuKFKAhpVHpDiOycbckMRA5ooRPI55ARuOtDBxe09sOrkd6Ex/iuHKjChUqwPw4o36YTp8+Ddfq6jo168NB8yPboCP1+Anj9T7Pgi8LdYIcZcuW1VjvJk2bKuWKDBNvb299r0ePHtKpUyeVG+kGfCPgQZCFwAqpehx8GAl6zJ49W2gAnBaSsv79++t7tm+jAegQMEkF2diRI4cVDOJ7zZs3V7YOI+I4HyPVPiIqjKEPL37yGeC1Eiiyyaf4PgevkO/HHQSMqHVMqWsdeXUt1LAZ1uRapJjVr29Mg+PWzrw2FTAVMBUwFTAVMBUwFTAVMBUwFXjHCrxsv2IkLxJ8fU74SNPGja1+Bb0Z/Wd84FHyADKZFAAdmBjMhkYTfNHzAWtQhQfBkB07d6ltxVeDBklLeKNev3kDf2gfCYYMlB/oV7kU+6VIEBdICqCxrLWR2wVhN5nUT6X351/IHcROf/LJJ7Js2XIFBnJ555Zde3Yp6MJj6NmpygkCKphXvzAJe0PKfzgSARg6CbPbRg0b6R/Xu3TrqmnG/AM7e03b+gSg7BIBlIFEytnFWRKF2On5cs68efNJ4yaN5cTxkxKAPvlHkAayZQf5AIyb3Llzy8/TftaT5768/o8+qsel9Xh2uM5QpJyEbIvnihKoXOourq0r1BRt27aVIYOH4NxjYIGSCUwcf/WapVky68U5Q0IeqQFzDMyJbYbM9g6JtcclWeDK5SuyHaDT1KlTZRIYMDm9c0KS1M3CB1CoqwCf6terr5HlegKYlP05FTz0wSWDh8DS04hw7Y/37d8nk8DoSQV7j2IA5DZu3ISzIGBMR68CAABAAElEQVSFq+FNwtfdu7dhyFzKwh8wH8/1fYx4M2Zo4EMEjugW5UakWhGhovPy/fuBSE+6K6SEEXU8f+Gc1KlVRypWqiSX4UtDJgc9TlKmSq0P/mMANbzQVDDx4QPMY57DR+ZxOPxkoiIlpXtKeMM8VqbHpQuX5CnMhrm+PppAH+kR4459akJaM3/uPNDUqmhUmWc2L3WMTpk6pWT2yCyZkOR0Ag8tbwwRTH5GrPHuRSaAEheU4XwErDJltqRHBIBswwZo8MNqG4eAtFLa5YZjaHocd1SrVk0drbmNlC/b4ANPD5jXR+FClqdNwO2A19/C42r99/obG0DH44e+A1KUaJAcdzQFKJQ9VsLE87QNC8UEWop/DEgJex2U4X5ElF8fREs5Onfu9C9rEYCi1wwHwS4zTAVMBUwFTAVMBUwFTAVMBUwFTAVMBRK2AlYnxu8robRo+3Eb7UvZV65ftw4eqafxkr0hvuitgUFJklJFYqUsLmCjXIaVRvfuPcFQ2Q6/meGybPkKJRnoATiWIEA0WCmFChVSkMZGMODULiAq/LH9D/VixW7od6NgMdFUBiIFqmu3z2T4sGEAMDLr+scBFh08cNACK2wtKyYJh7yJ/ik8TZ4l+8w7ICcMHPQVmDg/q3fnlk1bhIwbDq7PfV7gmm6BDVO5ciUeijesno1x4Vs3c38AK+jpmSo1etRY7U+ZFjV95nRJnzYdj1ASxBowfLiyypnwisBWOIgXzlDSxICoQUNlevcQDJo+fbqeT0ooXU6cPA0WTD6tT9z1b6N/JbASGRWhTBpemJOji9xHVHaZ0mW1/g7o+eeBsaQSMhy8GueQC8QFAjxPQdKYiZQpJmXRtJnr01iZPT/JFJZyhbNCdoVY8hXw8ykFmZp94kSyAXKtpjBAJqbBPV7EggThTyJUCfM0/KkW2cIfsEsCj/gBM6wizuzYsWMybPhQ8YV8hjeAzTgf3HDoxe7B1fosPE5IGXKBr0zUc8Z12cPV+b7UqFlTGjVqKMfAjkieLLmkw01mOlIyvCbYcxNRV5lBD6NLM1FLFig6OkoWLlqoeeKnTpySYz5H1fiXcV7E2Ujw4geAVDCiYtWqVMVDOV26fNZFPD29JBTgjwNoSJcuXVRaFm9YVjzwZNPoQx2PApPlQdkSgRUyV2iI6+fnrzcw7rQ2YIYRXrZxFYZGOHUJAWvFyysbztVTv7y8vNQ8mA8TB7104g7qFSljom6ODtY0kRo2Ypju8ib9m94yO+sfF9s8ZPBQs8f1S5YsYdv88ifPt0iRIvo773HcwZnSQRqWOTP+0XjDoAYRN0X/AeDbXIumwRzFihXTn3G/xV3Ltl/c981rUwFTAVMBUwFTAVMBUwFTAVMBUwFTgQSpAJsjjNq168CDZLISB/g7e1H2UlYjHwN/VEiC0NOoFIm0GYwsWTPL559/Lo/DQjUhiUwVetEQQCmGtCT9czh6IfqmTp8xHaoRqwdSew4cT/Ndv5t+kgJ/mNfJ8Z3Aybjx30vzZs3FI6OHdO3aVftlrh0YFPjK+myz2DsnRu+r0c64FlpWUC4UieSgfXv3Sc9evfTc94MZwr6b4ARbQfyPoJkLEvYoTAkRBIo4CErcBrDDUB9eq9W3xsjvq39XL9J7d+6hz62k58X976pdB3p0giD4j6oIJ3jAEtCxelFEWKdwk46dOqLXdkTvfgyqiDWyFyE8tO3o3KmzAlO2vpiGzDdgwMukLPamXB+YDtKfI9DjWzWLZWbIrt27pHWb1jAo3qRyKSpEyKq5j3MiqYPH87oe4f7QW1avm9/wBq1W2n7cVmvM/ppBOsuWLVMbkZGjRqG3zQRghvIuXmWMbN22VXyvoQ/mnLH3n+8k5MDlxWPoiZIK9AzshgPIB7+mFCnGg6WDqSy9YgiSUCLzHBeWFA80WRwsVAZIfIjCucDZ+Sa8ZDwyeaiHipeXF6hTPshuf6xmwnSnjoiIlC1ICnoAWRILRGfq0NAwlULRz2bX7j04hwgwbVKruzbX5weADzZjtNMjqahmrdpSDrqyefPmqjES0TR92PAApQUglBngzJsYH3+3OgMGDACoUUpG4UbSJJeMHPrXME5Lb6DtaY8zIWtjG4GBgfqSH/7k8Nmh3IoSJIJUdA1nMhJlYJRL2QbBHHrstG7dRmbOnKnu39zGiLd/P/A0vTbo8/OM54nhBmOpNw36+vDJJJMp7rAA5D+vI+571mt9evVe8HfGmNOThyMF5GVvGso8wpSvr/Wmfc02UwFTAVMBUwFTAVMBUwFTAVMBUwFTgXeqwMs2JkZ9TJhMW716DYAd9lb7TTQAwwvR0s7Oll+L1U2xx8wAIsJT7WloELsL8dX02vQP8FO/0bbtPlEmBpvBQChJGFSjx+Ibl6W0iH981x6UjBVspHIhZ46cSI7aJW0g/QmE582qlatgwttPSQ56MooWoAdHP5s1q5f+4ZutJvuyDBkyyIPgB/C+8UQfnhhpRlekdp3aMAfeivSnOQhuGSyOAE50fYApS5cttYyJ2TtjI4kV7JGdnV30+hOBnUIv1oIF8sPb5ZImGV1Gv0lfVMqJqFbhsK1PSw0aIpOkgWkUNHJDX/sYgEuxYsVxTokgC3sqnyAifOy4sTIcLCPOlYShPDwHTLQO3q13AQBFc1KsTe9ZJj49x5wEjigv4x//Ka3y8wtAJHcPhN7Ao7RVS1kP1ounJ64d/3F9kjaePg1X9QclZuxn+Z+jg4MSRpgKxUF8od+A/gq0UXL16/wF4gHiAefgOAIP1CtXQKTg0PpbLxPye7yAGRaOT1eaVCkRN50XD1IGNRiiEVEWXEjValXVD4WNNo2JKEXiA8QiU4YTGvpY7GHYWx4FIWJIcyJSm1at/F0uAsGD+g2mskeRugRZFJA7InAEM1yTuUE+46LJSkmTJJVnKDZROoIZvld95cI5xHlBy4bDAQI9R+3wIvZcg2H0uwJ0Nc5Vrlx5fQAuIBKM9CgbteltC0xJ1Lhx43CY5XhNIOohHkrGnTGdisNCG/UlbrCFsrJ2tkFUkoPME/rRnD6FL/48fSr291OohQ8Si0rZDtHYsQD/ABgc5VNzXqZiUcPH2Lc3DVQBgw9p7PqxOxEgcQWbiSOu5Cr2bf0RROAINcySJevLzbb7zwf+Pw3b9VPeRekTB+vzpkFTaE7JZ8gMUwFTAVMBUwFTAVMBUwFTAVMBUwFTgfdVAe1p0CNRosKmfSuMbwcPGSwp8Mdyq38SOXDwMFohq4eimoP9Dw1j9+7bq+wS9kk0n6XxbYcOHWX6z9M1SrtmjZpQazjG9qVRAADALsE87J7IyuEf4jliYtCWYyOTij/5pJ32pZGQBE0DaMEkogEDByFa+xe1sbCtf/ToEQAoTvpHbxIeeK4MkaH1BaOdqSYhqHL9+g35/Mve0g/gTvfuPSRXntyS0SOTrkdgiIqGF/B04fXfxL6p0NurGS/Ok3OSLVS1Wg2oOCw5FlUrbeCHWrNWTbBMlkppsIQo++H1Ux1D5QqZL+w5ub69PTxe0PcRSFEWDfr5EPyxniE0LVu0VJlQ1SrVJbd3Ll0vGiAJ57GLXZ8EAmf0/GzptV/HbSBLqEyZ0vD4sbxPTwNQ64GgHJ47bTMaNWrEqmI+XoGd3AcDh364rL8CQNgWgjh0W4+KGdVCZc6s2UJT54weGaXrZ52lRjXLI5ZQAllIHKz/+xjxAmZ4IUTW6FCcL19+qVS5CvxjKoo3bjYLyEIwLYkSIXqHdOrcSXVyyV2TiRsoTS+iI4FWOUlDFJRIIbVzQaBoUQumbAlQs5YvW4F4rTPwasmsSBvRN+r5qMmjZ0xxSG/yA8F7BtZOGHxtIiIjQFcK08LbK0p4VWbMmCGn8CGhjEm1ZqglaV/Jk7noOS7HA/UJzJZuwJT4XcbevXtxmB20g4VVyhSXCaImS3jP+sBbs+sDgrrFBTTUKwbbmEL1ulzoTefE+hyBdIpjBCK5yaSxPVjR+HBx8DGMOyyGjoXYxt3O1xqDjQNo7vv6oFnUbuTec768efO8fJuIJ8d/fjRfvX7bWkyuen3oWkCbOfLmy/f62+Z3UwFTAVMBUwFTAVMBUwFTAVMBUwFTgQSrgK2HYrOjagA0N0OHDZWBX30Fq420us45xD0zXpl9j9XYk90h8AgFY8Y9BcAWvIHjKeNhKtLIkaOkPaKws2XPLnvQ23jn9FaCQSIkM7EXZF9FdUUJ9LIquyEqguEEEGQP1CAMa1EkAosMg9dMO7A4aAPBKO70IENw7yCE5yyFWXA4SArs/gjERIOU8AzqBIas8HcgIwBGYK6LHnrW7FlSu1Z1jY/etXOHBtNQ4qMjdv3LvlehMimHJKJQACsW2ER5EBOXCxeGjyn6fzv00YuXLJYWzVtCYhQJX51lkjd/ATUTxkFgm8xXdon2nlhfgRbs5wJiBftf2/UnBvNl997dsOIoB7VMmCYMt4AlSNJYwgDXZ504yHAhiUO3YX3OTZZSUuAK3InED8aOtwLQcxieqAy9aY3UaAfIp1gZ+ubQ6Beoiq4PhEgZS7YFgB9pL50YIUQ0GC5arKhcunwZvjg/yzdff6MSMT2RWPxDXyfwt3gBMzwXFosooDOirh1w01NDvpQPCN39u/dlFZgpwWCmZICsKbmrK6LHSiJaOxey1UtIzeo1lU2SDCihO8AbDmak371zH5SoFIrcRQAlJCr2FIY7acC4oZEwH7waMPctCGmPGyQ+aYDI5cSDTsaLD4AKgj6M7bbDp4oPaAgeqp/xAI8EeLFxwyb9UGTJ6gk2Di49FlggkHEZhX/dv0VP6m98o1aPN/zy5UuQW9ELx/JSYdS0ZdaLh4oQX+zQ1/icWB8/ayNpVJmAXFJfR78Y2zy2Y+i3Q28d2yCNywkPKEdck9yz+EeD63LEnZ+/U7LFcQ+1IhuHgx41HB/jw84D1qxZK3PmzNFt/Ea0sx1oeATa0uAfiDZ4wG3jzyuybfl3P1+9/nagrllrrXllLUqqdC2cE02h467172Y2200FTAVMBUwFTAVMBUwFTAVMBUwFTAUSogLa3+Abe0l6bHbr0V19Qzm3qjfQ0Csigy6YjAsnNPKJ2edhe2K7xLDHyIJjeqDNtJNtSKxdteo3ZcZs2rJF2rRqA1DjzwQi/kGePp8PHz7C7FbnxuPuI13YM4sXtmCbAhvRsn79evUSfYg46NJQUHRDylFi9LIhD4OBNQAqQj/L/vYBVBtOADB4SghewlnSJiO59OjZCwygFHIICU0XL12C14uP2m/079dfk5NJXuAgSEQg6cb1m8p44frsXSlxUrsJ7EdmDGc+ceK4NEWC04rly6Va1SoysP9Aof1FGEAc+smisdX12eO5JHVVpgmBL+2FMQeBnvJQsPj735a9e/fIL0gmnjplioyHx076dBlers/zokKFvXo0zs+2/nOAUHcRsOOAmvKMeP0EUzp27qwSqcZgzYweNUYZPM8QKES7D67PSkeg36SliuUjA6An9rxottwE1xT+5KksWLhAxn4/Dj49PRXoUQNnAEDW/edZJeyIFzBDFggvjPHWSZImUd0ZtVcRT58pgELzVkp6PqpXXw4jgnkjkn9IW2JUNnPgK1SsJMWLFlMPFerpiILRxDcJ9HtEw5j/Tp+awKD7OrcrKEx8IJgt3wLxzdVAqbpw8aL61hA4CMSDmhweKWkQ0U0E7hFAIaKGfFBv37mt8qY8AI1aIo6Z8Wd8SKix45y8FjWqfYf61q5dW/VulBIxUYhx0mTATJw0UWfjgxKXMaOvY2++bTle+/jxiMvG9h3bt+s8VatWVRoWQSgvLy/p2L6DbXcFrGphXQ7GbJMxw/Qm1kYNh7E9LhjE/ehRYwORigEFZCoWDZI5GL3t7Z1LEc2OQHczZcoMelgZGCJlVEMlJl5NnjRZXAGw2YYt9vv1dWzv60/9B+HV6+8Ompk3os3IJuqAtSjf0rUATNG8yQF0N5pvxV3rlTnNL6YCpgKmAqYCpgKmAqYCpgKmAqYCpgLvqwLoydiDE6DZt2+ffNnnS7XciMHv7FOJBPAP50XQU0WBjaIbsI22HGGPQvEHdw/dRDVIW/xRevnSJfLzjJ9lzJixgA+wIwZ7IcZgh4Y+whZsQyMYBVUEe8VkyQFkYJlEWI99bT4oCWiX0aJFK7mBdccBMKChcHb0ntgNg8fDlgI9bwFYaVBdQSkRe0uCJTdvXJeChQvqfpFQmrAfGzlypAwZMkR+/XWeJAO5QXs7TMbrvUfZDwgDXJ9pUbZ+j9evqgnMy3NyAJumb99+YLvskLr16qqBrgIYWMnqrwFoof/PkT2bPIuyvFf1WvE+/XA4B/ZU9ctYWIN06NAegTb1ZdsfW9VCI4a+MBh+/pZvz3P+jotSaw6cQyQkTw6OTta9AohECRKJG5MnTULE97foaT0gTduG1ObqkhgkEhaL5CBeUzJI1UgCUdkWisc6poTtBj1X6fkaCTBnzi+z9bryQDWyFUBbNUib9P7rWSXst3gBM7xBqAeoUc/1rPjgXQFKxQz3QwcPgkZ0WJkfTmCUBD94CFRuMSLFLupFsgiBkMjwZjIliRfLeGsCAVGYjylJyZO74UFygGFsCOY6KAUQM0aPFTe3FPqAUQ51A+DPsSNHldlBwIeFfY6HnADPQ/jJ0Gg2a1ZPyZM7L2hmmA8gUkqAPfYwdArFhyYForlLlS4pZWAMTJrWuwwCJwsWLNB0IiKpa2FYFASQ6Atk0veCEzZJWGS42Iaijbjzr6/XHIARKXIFCxXU6yEy+Pvvv8tZSLmyIuLsm6FDbVPoTzKBGjRooA8Xk7G2A9Dx9vaWP/7Yph9gXSfOEXz4pwCFpHEVzZPIrmGiEgcBm2PHjkr79u3h4eOqWe+MxqapMsEeyrVatmwZZ7Y/z5/Usn83HEEf4/XHPRdrLR8FZVzxjwDRz0OHD+m94lp7gJi+Hhn+7+Y3200FTAVMBUwFTAVMBUwFTAVMBUwFTAUStAJARQhqaLOLH/RDmYGwFQIh9FNhc38frBBP+G8Gw2yXzTrZLvRNPX/xPNQd8GMBBJMYspu7AEsGDRoCOVI76dLlM5ATKqo3K+U47MkIiBAUoB9LEMx+U8InhfIfsnDIsKHlR9Wq1aQI/sjOPvfkiRPSuWMnbKsqS5DOWxVsFfbPRB2YhPwcx/r7+eNXyn14TnaQTV0Xdzd3vRwCKyGIk540eZICDZUqVZJm6EPLQsLEno3Xwp6a62MCDeRhX06VCY8lY4XXz77zo/r1sa5o8E3v3l+IK5gxTDdqhlSpqMjnuj5Tm0+fPiMBMOrl4BysLa1FuJ19Mq+fJr/so8t/UE4BqhWQSBUCa0mJBbgWgk1WnQCigCVEwIoR2bxGnjNvV5ZMWaSdevTYIx3rEOr+FWxCrqBXXygD+g2welIsTtBoN5hB6v+D4/T+of6OAHnWrlurXjW8flRAMY2iRYqp4mTT5s3iBgbS+xh/ogXvNDtKgyLR2yXoQaA8efJEL4quxrdu+Sk4QQ0aPUUqVKwgPvCQIfKkOjqUdRsABBrw0DQ4m1c2SZ0mNeK0UmiSUZcu3SQ15Et79+yTU6dOQnLzVA2EM2RMDyfnCEmCB/Q5EDLmwisQgzX50KLGoI1tURYO102MG+0O8IUmRHxNdOzx4zAFbvhEkDFDqlTVSlUAUlgawncpBRk8TZo0UeSUVC8aLvFhIRI6YsSIV9KOGjdurOlEZMm8Pirig3ry5Cn9AFy5ckXfJqBCac/rg2wiAjcEg27AH4esI0akcTD9iOu/PjqD2kWnafrYEBhSb5vYnZj+RD3erFmz9DrIKKJJ0+sAkm1O23VYLBzb1ld/EjAik+j1tCeuxXVIWaOEjedL9k5cAOfVmcxvpgKmAqYCpgKmAqYCpgKmAqYCpgKmAv9ABYgcYFAfYr2ykybo4crCcPbM6bMKhDB8ZuzYsar6IN+C6gtuIyDBvpbHWkeDyYF+lzYfFaEYoTntjp275KsBAxUEYL+ox2O9y+j/unfroSoCngK/Xsqb8If6O0g4pvxmydKlkCUdkkULF2O+TbJu3Tr5ZeYMPec5s+eo8qBmzZpEQdCP0k/luRIeCGxY7BcE6ri4gh2zF8SH0rIdffnvv/0mo8eOkTmz5yqLh5Mx4ZgGwTyegMwLkB9IzeA8Uc8j0aMTGAHXAyAGe9imzZrCrHikpj3Nh9cMPXHYn1MVQWNevSBOzKqCaMS+nCoJgjesM9k9d+8GaqLylCmTYHPRFkyhBzJhwkTIowKxB+VRFhDD9Cv2vwS3FETC+bm5JQd2cBp+tsm1+tdgZvzJJ5/Kl19+KSNHfSeFixZR8gbP4BTO15aMrMfzRDFHWFg4sIO02scrAoV9iTvUBwjVHwbKw4YO5+EJPuLFmNGCooIEOx5CY8c4a1fXpJIPZrxlypRVGRJBGN7MevXqSZ8veusFHvc5oYZAoSGPIFG6IJu3bJaNmzZCz7ZHHsAx2R5IVVbkwhPo6Na1Mwx1B0lLaPLOnTunzBs+vETYSK8iJYroGcEZGgDxwQ2Cz8uGDRuVdVKkSGGpWKGiymVSIHqaIyzssX7Z4yFKmy6Noor82PwVwPB3Kk+0j0AHJUI2UIQ/XwclOBe1fn8FQhC0IROHX28CZeKeD82GqdGzgTJ8j8lH/+56uJ20sZw5c1ofpLiT4TU/rJ6enrr2vwNlbIfwOv5qH773puu3Hc+1yAbidf5VPWz7m5+mAqYCpgKmAqYCpgKmAqYCpgKmAqYC/0QFLH0IIQMOEgo8NI2If+zmH5rZy2vENHYhMEN/TvbD7HHIyrDHNgIXDRo0VBCBnqgHDxyQbdu2ytx5c1RORNYGzYPZx/L47X9sQa/MaG0yS7i2tZ2xz7b+m3MzbakC+twFC36VRgCNxoPVw+RjIhxMfOI+dF+hp2hSerwQd8A3roOXai1CsIEJxevQO8+ANIryppkzpiEKPB16ZOzFHTFWrlypzCBOzj6c50XAh2QMmgtzN14/5Vbdu3eTnj17Cv+IPwYmyGTW8DhiBQQfeEnc3+KjUIUBwgX6em4n8eDjNi2xbxDsOibJcvjXfNymrcaPZ8vmpWtagTYC7GCvHsu0JGXzYH0CO+fPn4uVnVnXT/LIt99+B/VHKykD9Q09ZEoDYON9uR1wW8/NVn/10rGPNUymbAv/sa+mcoS4Bk2dH0F69j5GPIEZ62apfo3VBVLG5pqu00OGfKWGQgUKFJSzZ86BSfNCGTBE23bv3qkeIsePn1CWC12jw2DSewPatzu37+iDEICfAZC44B4p46YSGDcrVqxQH5b16zfAbOiUMnL4oHFu0pns8ZAQBEifPj1MiN2VFpUMYAz9S6pDV1YaciX6mbjF6skSwzfFAx+ux2FP5Oy5szC4jXgfNTZzmgqYCpgKmAqYCpgKmAqYCpgKmAqYCpgK/P+6AgQ0rP532rRpMgX+m16eWa0rwuZE0PcEBAQgPam49pfo6dVslnKm69d81UyY4AiBgj5f9lXLiwED+ssmMGicAShYiEWMHD3mo/uwt7YAELBJADpY/jYU11i75syeA96tSTQ4hmyc+vU/lAW/LoBZcXnuYf2PtclKKQZJEMkSCs4AAeF/ZLDkyJEd/p72UKSEy3CoPBrAK7U5ko12gzCRP19eXZ/rBYE8sX79Ruv6Y9fnHNFgkhDgoBwLPBbd39PTS3788SfYcHwj6RACtBT+OvXhP8NhW5+yrYcwLk4JZUu0yrAssIYpWGjrEQbkjnN6IkdgVlyufDkkPaeWM8AUPqjwgZ47QRzKthYClCJZgwMVUsIIGTRMftZ7BRCK+/I+LQXLqGHDRqrmWLF8JTxWuwO7oIDIOp7f6Y/rCSUPiSDKosE2kh+4Rs6cubSGBG/ex4gnMGPdcPvEVnyVHRgrfGj4ANkDoCmFTHNvsDKmA3Wj6Q7pQr7XrukDSg+WW363oENzUbNcOlIfO+ojj588Fkd7B0ULmWgUBenTxQvnVXpDmtMNSHaOwEh44aIF+uCRZeN79Zo+DIkAtBBxo2wpJ5A5Jj4R3Dl77ozq9WhqlAlOy7nhZ1OtejXJ5Z1LnkFmQx0e0T26L5thKmAqYCpgKmAqYCpgKmAqYCpgKmAqYCpgKvB6BQhG8H+OTz/9VDZt2aoBKwQ9iADQCoJAw7XrVxWESIzeOBEUFEwXZmCOJW6KkUiwLwiiVK5URdIDvPgYybdVqlRRNgrnjl1C2S1UpxDAIIBAbxcyYQjS0AaidZuPlUXClN7mzVuIKwgI6yFratOmFUAEADrYn/4z/ncCdE1ODKxCz4PhLuvWrlcTXIblcNBqpHChwsL0p98gbaLNhi2QJRJKFc5HIIbr098lsb2j9v9qHIzrTwLP11Zgl7i6uqBnPypDv/5GTp46JUvhF9MfIBSv37b+aUiO6tT+UPv/GAIhGIzU3ou4cM+sWfQ6uRb7/0qVK4K1swImysuR8FwcwE0q3T8SQT8ctvVpNOwKhosglpyMHiwnzlCifAhgiHU+5nNMtkBWNW/+PHivTpVZMPelHQo9c4lhBEAq1rVLV5gRP9X1ObcLronpy3nz5Xl5X7g9oUc8gRm9Vo1t9vTykjp160jGDBkRb/0EF2I9ULzJZLXcu3sXKNdpWf3baikC2U33rt00NjsJaFWpgdbdwft/wH35+vXr4unpKV7Z4A+Dmx2OotxGtNWBAwcVIatcuYrSwBhLdsvfT04h9nn3rp0SBXDFDea+/LC4QKqTIoW7Fvn+3XuI8Fohc+fM0Zt59iRAmsBASYfkppy5vDFHgJQpXQrgUAPQlFwSur5mPlMBUwFTAVMBUwFTAVMBUwFTAVMBUwFTgf+RChAYsZpdO02a9QHDpXjRomjgCQiI/PjDD5AHXeBLBVqew3OG+2usNbZRNsP3koJ0sGfPLg1zuYte+FeEyXw18Cv1RcUu1hIAJgiwMNqZ7JhohTZ4fIw4OjtJBBQf9EwlsEEj3XLlysof27Yp6NP244/hteKmDTv74ZNQq+i8BIvQS/P1o0cPY20pCJmImhefgZKkcZOm6J1XaRLyhAkTlDVCUOYFTlzNe7E3Q2Jy5sqBc8F/rAl+poVHrM+xYwjesUCMYLBi2n/aXoYPGybfAKSZjZ7cCdfC9RntvXDhr4oV8CQ5QwyYKY+fPgFY5KYsFUu2ZKckil49e8g3w4dKftiGLEKUdVaylXBStvVJEPEHPlAVABd9eFhALKNeOncQq125YiWsAKYNiB9fI42KwFq9eh8Cg/gDsjRXvS4aEE9AsvL9wCCtPyZXgId4hhula/idRsHWmzpdgn2LHzCDc8K5wS06kfqZ5MyWHdntmdW0aCOivw4imSk8/KlVVFSFlCCig2nTpgOK9zFyxUdL3jy5VZbkBKaLHZKSiAimAuOFSF0amCb9BlrW6dOn1IDI2dFZcuXKKQXgR+KWIqVkyuAhtWBqlAUeJbwhyQHy8ITIlilarJi4w3uFvjOBcLc+c+YMvGy2aB75r/MXAAC6CYZMOJKZQkDjSiPN4RzNiGgzTAVMBUwFTAVMBUwFTAVMBUwFTAVMBUwFTAXeVAEyK7Tjx5sKSODX1mC8DBg4EDHZmRRgYA/MQbCB6gyCNiSlENywJDJ2Gn7TuEkzuXr1qmwB82Y4TGUZ8UwTXiYJa4wzkAA7HLRi1UoAAlgIqA6+67zsf69f9ZXkbjS6xcB5EVigl0qfvn2lRo0a0gNeLyVKlOC7SnaAtkUZK9jZAi/sLPNeqoEYDU6T4vYALCjJGgop0tw5c6Vz505g0vwBIMTrlfV3794taVOl1utTsALrEyS6efOWpIO1CAEssoU4xo4dByZNaxA0iggtSpgyxXHq9Fndj+sDLlLZFZk4/NJkKNYM118ZYEsUorLnzpojGzdsgD9OHjl39pwUgm0KwRuuz/IEgHTBBOJQePUogIZzYkz2I6QxZwB2QAmVBUnZybx58+SjBo1UnlQI59UM94J+ORcBqjG1ylb/kJBgVQMRq2CN8IbWWi8gAb/FD5iJfSpI/3EDCOKO3HYWP+JZhFCa9Nvvv2kGOs2O+ADT2Mi6QVgWxyZNmkQfJEqTvLw8YV70gVKyGBFGqldiXPx+mCPt2L5TgRUmP3EORlQ5oGjcbzQMeIjiUTrljwfoIRKKmEleIH8ByZY9G2K4c0vJkiXVHJdI47379+QcDIECbvtj/1uyfcd25MdvhFSKdDN9pBOwvGYqUwFTAVMBUwFTAVMBUwFTAVMBUwFTAVOB/9UK2DrIT5H+w8TcQiARsNllb8n4bI6ixYqiX6VJLpCG2JEpkwdIB97q0RoVHSWz586WmjVqSaHCheTw4SMajsJUJLaoVHxQgsMeGiiEzsDvzxCC4wS1CPkmBGrs4ZlC1srkyVNk9OjRkiZ1Gj2nJk2bYG/u9UKlSJyAHq80+eXgOzEAQLJny4EeO5UG1VAedMznqBQvXgLXkUj27d0j9erXA2hir8cwgOfn6dNVKsXj+f8zGPEyTpz2IlbaFIAR4APeuM6169YgRruJ+ucsXrxIegE00oPirH874I54esHkF2uzfsBRdC5PEDEafFRfzzMYpAt6yBLM2g3ZU9du3RQ7sE4hRn766UekLZHxYtWE50lPWidnR2AVLKAlx8qdK5fsxTW1AajG6PPvRn0nU6f8oAwim28NkTTOxfWZPM36Y9N7GfEDZnhKODneHHf3FCoxcgRA8vDBQ0mHB4KJPTTYzZ3bG8BMImXADBs+DGZIj+XOnTt6cDQeLH1oYTrkgYeTiUYEVWhARKOipNB0UbMX9iRMwkIfq8atFNyU8+bPpyydZ2DYJIfBL3AfPMCHEZW9Vfxu+mGNULg5P8D+yfBBKCz9+vWX3l9+ISVLlUQsdjo1/aGnDCO0tm/fLsvg+HznHs/JDFMBUwFTAVMBUwFTAVMBUwFTAVMBUwFTAVOB/1ABNKHa6qNZZz9avHhxqEFOSu3aNVW5oUAMtt9Az8mel7QOC1dBIjC8T3bv2a29KUEFgh9bkdb0YZ0P4aHqKxMhqSlXtrwG2vAswuDFyjliuBDGC0RMM8mXx8WASQPiiM7VvXsP9OCJkGh8XgZ+NVDOnD0jK5atkEGDBqG3RjqTnnGMXLh4URo0bKAACiVEnDWpa1JNjcqdK7cCNQ7wkPHx8ZG6devKRgBD48ePB/EhL6xIUisOEAE5E61LbOuz/3aBxwtZLgQ3CKyQ0dOyeUs9N3rD7t23X36A3GvS5MnyI0CUlCB3UHrE9UOgZqkJpk9YaJimNoHfo8AVDZPTg81DsocdgB5Kvzp0bI8I7JHyzZCvpXbN2pIBPT5HJPABJihZ4AqkUahbUjCWWLtEYAjxSmn6WwPKG6YzkR1E9s2woUOlfYf2IG1skhzZc4CcxAhvxo1Ha0ryA8iyOMef0Joul2Df4g3MEDBiTjvZKNSKEYBhhJQXPGIqflBRcuTMAc1Wcrgmh2txalSvobFUn3/xBSKtN8g1PHRhAGA4D6lBLDYpUIy9Cn0SqjQlT08vSQ0nZpoZEVVLgZubNXNWGPu4Qqc2W9asXaMPezRuQiQ0Y4FBgeLn5y8zp8+Uw2DjRIP2lAQavPz5CkixosWkwgcV5Kqvr1yDO3YyaPt4s8i2CUc6kxmmAqYCpgKmAqYCpgKmAqYCpgKmAqYCpgKmAn+7AkAVLOAFPS0Qko0bN0n/fn3hfQrfGTS669atlyOIylZ5DgEaohDoe0NDQtV+Q5thfHNO4gR1x3nIhz6Dj8oiqV6zBiQ3c1XaxF6ZAIqyU3A404zzgPVBDxrb1pTuKeXUmVOSL19+LBGDPjha6tWtJ8OHj5Av+/SRn6b9pH6sXD8YHi/r1qyTJ48f6zny2wuAEWH4PX16gCBcLxYtuHXrpvTp3QfeLF9LRo+M8vvq3yVL5ix6GbwYMlJ4PSRglPugHMAMAi2EQBIpTkA/2ZLwdeUalCmNGj1KWrZqJVUrV5Xly5Yps4e1iQbYRB+aGzduAnnC3lifANRzpD9FhFt+OrGLgsjxVKYC3GncpIme0xzUid42tvcpGSM2QXlX1ixZldnDuVh6Rn7TqDlPLssLh3Uli6de/briCi+Z9evXwzeouNjZWwVYsmSxnPA5gSN5BfiP9zCBR7yBGV4Y46w8kJm+c9cu1WVlzppZ2SxeObJJzhw5NY+caUyn4BXjC0DkEdySg4GmUb91AFKlA/sPSPly5VRu5AwqFqtF7DEG7kKOQLPcUiQHAyet5EWhQ0IeAbU7KYlBBXN0ckT02DWd4yhMhoie1Qe9Kn3G9HpT7wXdlwvnLyg7hg+DL4CYW7du4aZEwVvmEcyB/FU6RYQuBKlMz/HTDFMBUwFTAVMBUwFTAVMBUwFTAVMBUwFTAVOBt6kACQaUA6Fv1/FF795SqUolyQuGCTfeBhHg5cA+EUgHpg0ILTmAgaj/iZdndunerasEP3wgs2fNQgrRMqnz4YeQ3OyXPEwFIi0mFhTwOX5cnsKUliqTF7E8DnvYgQT4BQCYyaunQQCC6pMR3w6XNq1bqzqFfXud2nX0VI6fOCFPoFQB50X779BHjwGcgKXiCD8V7JEIJ2YPZUu9uh8hbCdKliD6+gTWzZw5s+zZu1dq1a6Nvh+pRrHr86CwR2GqXKHvC8+XUqYgECfSQlb1cuD6Ccg0b9Fc/WCq1aiOZKqPcW3wzUF//yA4SK8T4UowB45EbeBFi/Vpd0IyCAcToNPCy2bfvr0y79d5aqVy8PAhadumrapsKIfSemHfEydP4H3fPwEVrE9g5xHAGd4uXj9lZ9u2bocXTgvZt3+f0Dy5f59+Km0KCgqWACRb6eB1xZ6DtSFhvscbmIk9O40Fo3ER46T64QIOAmxZtmSZerqweLwZpxCVdWD/QVCFzihCxThrolPheCi9waxpDO1blixZ4CdzFhSnfTAoum09d6hWDG4qgRjmhu+CezUZOjly5ACNKhXSlxwRsW2vD80HH3wg+fPk1wfbxdlFQZgo6MqegrFz8+ZNuFRvlz27d2mOPGlYTG8iuvcw+AFuekTCVNXMYipgKmAqYCpgKmAqYCpgKmAqYCpgKmAq8P9UBUguUERDrzpGZTo+x47LB+XLWTHT2E7vFoIBtwP81cqDUhk2vUw7ohKFEyhZAQDJJcQ0V6taFT6qwbJpw0Z4tDR7CUxEQka0dOlSADOQONEUGJNGPYtSgkIKpBopOwezUY6UFQE9WxCEU6VKFXmOiOkVK1fKF5/3UkkP2R/cl+d05eoVyHhyaiAPeTgENugnm8M7h9SpVVuv4RE8XatWrSKHEfSzcOFC+eqrr4QhPeq9AmBl6bKl6PfPYjZIh3BNkTinGDBVCPBwDR3ov92RGHUaCcu8piuXL8vkKVNk3PfjFE/gOT3HOXF91scF8dsO9g56qKXSEShoMiNZ+SPdRuYRk5amTp4q4yeMl0kTJ2rEuMZoo56HDh6AH83ul941lodNNOa24rZVjoWZkiD5yhdeuZ06dZLf1qyWjp06ypo1a2BmDJkUcAsdODdeW0KPeAMzPCUa4lwBKBMCOtRdRFs/R/F8b1yT70Z+BwfmsXII3i90Ok4Kvdl9sFj8QbsKg1Ny+nTpJQW8afggPkZCEoGnxKALMeaKurPRo0bKpUuXlJqUGtozDtKPWAdSoNKkSSPpMmYECplPmuKG5gPQEhkRqQ819Ww0P2K01SXc6DPQjRHU4c19AnQsb9480rZtWzyclaU1aFTF4FadDLSl9zEWIPqsL5yxj4K+ZoapgKmAqYCpgKmAqYCpgKmAqYCpgKmAqcD/bgUU7IhFaJwgT9qzdx/6wT4K1PA9cj4uX/WVgvBWJePFxsBwgCrkyJEjSkBg4hPZJscRc92u7SfCnnLcuHFSDME2DmTZYA7KkCx5E1pk9L8PQDYgUYIMHOIIBEacXZyl3Sft1Q/2IVQi1677yg9Tp4JF853MnDFTcoIgYcMZAgPvae997949AC04TzTo9KuhRIn+sTx3bFDwokvXrjJi+HCN+G4OD5m0GejxEgNrkUgwfh7q+jw+9HEo5EGMv4aMCL/znHj9eSC3atG8hYJATHPu1auXNANRYxmYNEURP44DsJc1rl+7Dj8dZ2X/xOB4glcEV8iuYdISaxEW/lg9Z9q1a6dMnh9+/EmywrSX50RhDKVbHMQcosEicobvDM+Dv+s9wYv06dNq385z3Lp5s3z22WdgB2WSrVu2Sp06dfR+vNADeGTCDstSOR5zEhmjH0wYaEBBMMTZuHmD+JzwkTv+t7VIj2Dg4+/nJ/1atYSsKQfy0FdKxgzpNUabwAmjw4jQsVB3790F4OKgiJRb8hRyy+8m0pNuI/KrJdytC8ldPCgEV55FPFWARQuIJ470qYJwv6YL9KkzJ4E0pgFAEwVFWyJ9kHhuvkD/aARcu3YtsGOCJYlTEqUlJQea2BQPAE2KSSN7H2P2rNn4MO5WIMkWV/Y+1uGcO3fuVMOibNmySZcuXd64DJHDrTBJ3gXpGeVcZB3lzZdPWgGgcoXnzuuDSOYff2zHZgs1VZlZ7AeV96BBgwbi5eX1+mGv/M6ErlMnT2nNiYrmzJnzlfdf/+U3xMSRYeXo6Ij701S8vb1f38X8bipgKmAqYCpgKmAqYCpgKmAqYCpgKvBfVwEb0MITUyYJ+vixY8ZJ2bLlAKpYScXhMLRdvHSxBN4PxE78HwADAITH2J7Jw0N9WYkaEAcIQmrQkK+/lvMXL8iiBYtkx05KblrDJ9VfQQX205T93L93X5o0bqTmueAj4FgY3YKpcuzYUSlcqDCO26F4x+DBgyAB2iejx4yWpYWWwX/1KraTOSOycNEieUaZFROluDgGGS+0A+FIRBQEfSHtSab8MFV7/507d8n8ufOlOYCWI0cPSRR8X9VjBscz1If9/z2cmwI7eIfTuiPVmcbAWiv8vnjREiUyDB8xQmVOTJZSORLe2w+50hOEAZF0kZjFwvqcK/BBkDg4EdIAOwdoFD11tmzZDIZSeTUrnjlzJixLQmQbTJWp0rGGZejLWZwA9vCaMRu+2QF8chM/9Me0UeH9YEz4gQP7ZcjgITJv/jxNu7Ij4sWDE3jEG5jhSd1EChK9X5IgMzwI7JlzgefAnnkIz5dEYMS4o+Bu8igkTNE7pjFVrFQJkdeucujQIRT4MY5zAVL4REIehKARd1Czol6f91QH6NsAZhhrxdQnom13795XmZOnl5ciYOHhT8TNEXIknghqxMx45pRTluSASKxs2bNJBpgXURdG+pSXlycy2D1Bh0oinPshwCQ+rNkBGhENfB+DNzqGsOV7HEQ/+8DQae7cubpKHmS7vwmYOXfunDRq1EiuXLmsD+HLU8Kn47vvvpO1a9eq18/L7XhB5tPK5StYXgxeh/XK9ht/9oaG8t8NpmU1bdJMUVLuQzocGUT/buj+AGMs6tl/3v/fzWO2mwqYCpgKmAqYCpgKmAqYCpgKmAqYCvyfVsBqUwFAxKj0hoBCmdJlhH4oTGviIDhAxQeBB/am6vGC45i4lCgRPF7q1VUgZfOmzbJjx3ZNdNq+fQeSlr7SOGyyZjhonnv9xg0Fe1KlSmOBHtgeGBQEcISMFq4VjRSkGFm5apUcPnRYFi9ZhD+EN5OdO3bKdMRf3wdbhi1fSlh+UDb0DFIkeuFoyhJ+Z89M0IJmvvSN3Qf/mx9++FFKIVZ71erf5PtxY+WXmb+A3WIxbni920EeSJsmla5PUIYgCJUsXJ9MGDULxvUTH2gKM9+Ro76TH378QXbgGqdOmSpPIds6ePCQKnCsdhSwDy6ZXrGODvC4wfUT4MmYyQOR3156LV/27gvGSyeZPXuWzJ0/X8Z/PxbgEECwWD8c+uOWxX3g0OOxPkkBW8CUyZwlE+Z/oSqds2fPy6ft28sXCC8aNWqUsm30oAT+Fn9gBid0995tSJhuI/EIlCMU+iZclJNDMxYZFKVgTQqgYcd8kMVeoICyH0gNIv2KTAzKm0qXKSWr8GAULVpYSqM4aYFQ1UMkV/GixeTE6ZPid+OWyqF400iBegq06xAeZEqkLkPqlDtPLrA+UmpcGG8wB1VyhFkol6pcuTJMhUL1iadGjQbAwPpUj0egiF/3kQ1fDPnyjBxL6GEhg3+CGQk9/4qVK6Rnj54vP0RxcJN/WYo6xMuQdnkAhSUTJT0Mk6g3JMrpB2YT5V1kqsRFealD5NlXqlQJ0ee5dU6+b12XSNVqVf9lHdsG1puRbRZ1zfpHx/bem35y/x49egCVBeqLHd4znvWmUzDbTAVMBUwFTAVMBUwFTAVMBUwFTAVMBRKsAlaHGvsdfdSBg/ulW7cesgSgCAkORBkIglBpUhZ+NPfu3NP+C5vQlwEkga1HpcqVFKggyNJvQD/58oveMmf2bClZqqTGa2vfhHm2bN6ihITKVSoD1EFiEiVBRDG0O459CZCICcf37t9Fr1xFCRFTkHCUHxYh4yZ8L7du3nq5PtUxlEexn9b+DydEk2N3txRSGrIq+tIE3r8vX4GFQ7/Xb8F4KVWylMyYOQNrWj3wcURucw72kPrFRh0n/ATyIyY4267fASoWkhp69uiF1Km78vXXQ+Cjs0SGDRsmFxAcxEEsges/j4rUhGWL7YJ1MLdHhgyawnzwwEHIqYJkzpzZqPESmQjPmQo4NypECMJw8BqDwbjhYO/J+scA6HkKdQ7TtMjW4fljswI2o0ePBrnhiqxYsUK3K/DBgxNoxBuYoWFOUGCwRADdY7Z42jRp5Qk0aGSgBD0IBspnp+BMxNMIoGuJpTceoGvXrymSx4gqDiJ769auB0CwScqX/wA3tAIMjkIkZ/bsCqwc9zmuHjb5CxaQpEmSqqwlu1c2mAPfUcbNuXMXxDOrl2TA+vagXL2IfK76M9ZSAQbUntFhpCR5eGQCPclPNmzcqMlMRBcZ73X7dgDkPHneCzCj55BAN+z1aZgV3wwIJ0ebNm0U1KLZkaIar++M30vDvfr7779XDR8RQY4BAwZIwwYNETu+GsbLZ+T69etCKZRt2ACYTp07gTLXyrb5b/0kg+e4zzEwlbKJm1syOQk5018N7n/smA/291IdI0Ei2/p/dZx5z1TAVMBUwFTAVMBUwFTAVMBUwFTAVOC/vgLEBdDtT/vpJylXrgzAh28UN4nBNobfZM6UGf3TcSUykEYDKAMqD3i2ADSw+qIY8YNipX//AcJk4oULFsq1q76yFXKdhwB5oqKiAFpAdqRYhZ16vrhBwaIpSVia/S+NhmkzUrBQAZk//1fZAGPhKfCd+QZASKEihWXw4MH6x3quzzXPwsyXkdN8TaBCcR78PI60I0dIs7iB0qZRo8doYvIfW7dJmTJlAIhMUAyAf3ynSbEa97JRxSQM6KGihYSNaPbkODf3VO7K3pkxY4YqQbLAIyZPrtxqcjx82HDZuGkjr0D/P4Nz+qjBR0hQ2m+dEKZ1c3OXXWDCeMCHlruxFuvWLVeJ1Pr1G/SayAxiqBBjtJ9ANsaBq9L1HwMocnZyBkMokQJa3J4U1iu04li8aLHastzBPcqQAfMn8CBWFa/B3PQzYL7cAGUqD9gUNNBNZE9j3tRSrmwZpWllyZpFU5luXL+pN4z7pAJQkj9vfqldpzboSk2lUpXKmshEXdyhQwf0QZ0EXdkeuCcz4ovIFSPFkiRNIoUKFZLaH9bRaOzq1WtojDblMdFgdtjjwXjyLBwMnjuI6QpEKUG0QmFZbn2A8IoMnijciOu+1+XqtaswTXrC+6ZIWXyKwfNcvXq1OlN3gxnSIujzyO7BM/KXg9KdCRMmSMeOHfDB/FrZQ3yI/s7gddONes+ePWoIlREPoaKleLjfND5E3BplRDZQxrZPS3gA2epDYCbuUGAJFwH86q1GCMC1QV8N0mMmTZoIL6C/ZiPxHxy6epPmN2nSJAtVxdFvC2zZJFBvdbJmZ1MBUwFTAVMBUwFTAVMBUwFTAVMBU4H3XAFleGAN9mytEWG9DSBGxozpld3y/PkLNea9du0amiBLysT9o9BnPnmKnlUbMna4Vgz1ksWLpWLFiuKSLKls3rRJaiJ2mrHP1iCskEgiIAPKlSuXKnjY70VDysTmN0VK2oGIAg+UN82bNxcGt3VBqnAGmLFO2SW2BvEPeK2cOsM/sBOYgY8rvjiCoDqh4oTXooANCBe7d+2WIkWLICDovgwbOlw6dOgQJ2QHO+JQnj8VK0xZph8N4B89Pm3a9PIAjKAMYL4QBPJFwBAjtY8gRGfJsiXSr19fPSUez5CgfYjsZt8cjbpwfXsECT0IegCAxk3Pz2qMRX1Vq0BFs3TxUuAItaV/v/6SEj6rtv4XqJeCRg8ePJTy5cop/sAJMKUmZdGfplCRQjpnRMQzrZ/+koDf4g3MkB6VJKmz6uHy5yuoSBqLgmvj/YYZD6KsnRw0uWnbjj9k2rQfLbACxz0OD1MKFf1QPoF7cr++/aQ8KEauSZOptu7u3Tt4AE6rjouxXbm8vTUznYXO7JFJ2TUtcKNatGou+fLnA2PGA1q4FFpUMnhI86KkyR0PHYseAHTrqT7QfLyxAQ5JzC+nlo+20S9vzDsUmKBCzVq1pFHDhjJm7Bj5efoMZbAQKfTlB+sNg47VdJ8uXaa0giVzZs+Bz8tIaQJdHWlgNOb9T4NmvQSDuD+HhaLi0oCCvs1gDfjg8fEjVS7u0O34tPG+vs0YgsiywKBArYstyuyvpiAoFYQPUs2atV7qL/VD/haLVq1SBcleiSGf6v4WR5ldTQVMBUwFTAVMBUwFTAVMBUwFTAVMBd5/BWx/dLb1XtlzZEdKcBtp0qgxbD5cJALSIaYGKQaDlpXqDiYSW/0q7STQ4eIP2fkL5IcVSHEFJyZPnqSqgwWLFsqgIYMlsQMkQWzG4dVKmQ6Nfp9H8w//+Es7t2NxRzBVNm3chF45pXq8cP5z586qlGrjhg3y00/TpMFHDSRturTKHjl6+AjmtHrCGPTR2lHjd1WssPkHvMIetCIkV6fPnoEH6ZfwltkhExFfPRVsnHTp0um5E1Th/xdhZFypYiXImcJfrs/rP3HypHgjKIbnz30ZENSpY0cZMmiwsoQoScrqlUXnOHDgEKfCsECoMLByAAbo9esf6/kaX+VhBhwMb9neX34uB/YfkK4gUSyDxYeXpyeOpdwLi2HBYKRaVahUEWFDEbo+a8Xo8ls3b0q+PHl1XwXHrEXxe8KNeEuZyFBp2aIV4qS2wgT4Bm4Gb1ZiLTofqOOgGEXDwZmuzgRBroKl8gwoE6lUG2FexIIxgckNDshRGaMkCxKTXgCxY045461IbQpFtHbSZC5q2psbBbkD017KoahHS+qaXJIjfkspWy+eY61oiYDezCWJC7YnA0AE9Av1YrE3gfo09JuhUu/DuuKEAttBw6aD3T+GDdTQX97y28CBA6H5245rt5MuiNWiFwsBE1KprNnx/bUbOHLkSI0FJ5NkxIjhoLKVk6tABTkXWTQEF5ib/i7jba9lExBWlsENMWh5ETsed/Be8WFndNnFi5eAijqr1KlYsWKKcsbd1/aaEiRS0MjM4QdRB+bnR4YP+OuDfkPTZ0zH/g4yBRn2Lwf2tdZ/ueUvX/gcP67vcz4zTAVMBUwFTAVMBUwFTAVMBUwFTAVMBf7rK4AWacbMmVKocGEZM2aMen9S3EPlBwGaVO6poCAJ1x5bnVTRI3lm9hTv3N6I0/aB12igkhzqQm3CHqwMfGd+nj5TCQvsv06eOIk+3BczEthhg0XAJkYjpjOi/2OLxr/Cs3djoE8b+o6CIPENZFbz5/0K5stQOXz0MHbCbDAj5vEvwO5JnjyZ3L9/T3ttypRSgEBB39lrV6/oH9wXLlwAH5w5CJhZAy+cUjAxrmfNgWzv8KePZcTIb8HoeapAE9en5InkBWf40RAT0PNCk8q+/nsArvO1qQAAQABJREFUPCQ8zIP0asmipTIcUd2bkcDE1pJmxmTy3L7trzImpjgTRFISiYsrlD15NO3qKKLICT516thZNm/eJGcvnJMK5Svo+gSVuP6wocMkCdYvWLiQYhtR9FtFrRh7jlPEekAXuChrloDj7WgVb1gYt1S8vLxw4tCn/TpPbw5ZCywoWRj0mTly5JhGWtepTflRfWwLVE3Zbf8AvHcUqNhxzOEpub1zKxiTMkVKACdO+ppR2ZREkc5EgIfzJsMDcMvPXzbCJ2bJkoXig4eRyU0OQAaJsiUDi4SmwgR2wqAbu3DhItKc7kAKRTfngzJn7my4UmfQ3HgCQzyO/jcEgd5l3L17F9q8+XroN998Iz///LN8/vnnGl1N92Zr4O7xBsaOQFC3KF/imDXrF9Cy+iE+raya7/7www96o9chIYlGvW894qzzd46lkRJjzDlYb8qj4g6VguHB4/l8++0IuH8PkhbYj+ATzYLJFnp90MCX0i6mNb2Mxtbzwrc3PMQEofhB4P62aGwipBw2VPn1Nd70O42ekwOwYy3NMBUwFTAVMBUwFTAVMBUwFTAVMBUwFfivrwBaJDJQunXrFpuSW0gJDGR7kKHhD8Ahc+asehl2zMXGcEriKOfPn9P+m2ALey/GQ1erUUPy4A/tU6dOgY9MTu2lOHcY/F1BtLGAGbyIePYU/a+j9T6ADbJwMmfKpP0dpU0EeObNnaNMl9VrVmsUNnvnRJwEbdpV3ytSvXo1nCcAlNh5qWw5CIJBgQIFcYY0NE4EWdRaDdmhp8ze3XtgX5JVXGBPwhH6KNSSDfGSgExQxkSGir1akSj5RdeqCBZL1qyeAFM2Izp7j8Z7L4W06YsvPtd5uD4HQ35af/wxMIMorRu3s7elyTEjtK0BGdTli1KufDlZv26DUKaVE+ocV6RGs/ukFcnDhw+0LsQ6HoBFk8QlCX6PlYgRMLItGDtjQvyIHzCjBbBD7NQ9FD2x6sOcwVQph4t+9OihoncZ0qcTeswkBxKXFT8/+aQdqFT75QSYDbUg/aEUhx4v9J2hpKkmHqSmzZtJtarVVBtG5M0lias220/AuLmPot6Anw19UMjACfC/rWZEicF+uR1wR3bv2SshMB6ivswNueiP8QDSFOnAgQMKvFTCTfXAA+eMG5QCTtJ8gKmLc03mquDMuxT1KDRvz+BZQ+MjSpPiDnqlFCtePO4mfU2AiGZDlFnRHTruqFatGh5GOlILEL1zcd/6z68J473F4PV/BoYPGUc0bmZk9uuje7fuMmL4CJkKwGjatGkKzOSGTpFAysIFC/T4uMcshtZxP+hymTJ5qGdO3PfeBMrQi4f0ukwemV7Zn/+A8GreBpj5/fff1cdo7NixryxrfjEVMBUwFTAVMBUwFTAVMBUwFTAVMBX4r6wAmh7bH6ULFyoM89+TUrVKVWy02nWm6NJThcwNMmg4yBChDYR7ylSqMKCTBVvBHQAaPoRChIm73bp1VfCCJru2LpEgDvure/B2LVwQrBDMRWGSHRgmGZHcG47ePG3q9LrWzVt+UqpUCcikjmm6Ud8+fdFnW2QGXyhhDiHu+inUMEwwok8oGSrsv5MDoLHW4YmKBsB8BIIG1SAFCxZUVkqaNGkU3+D66hGD9QOgjGEyc1QkI71xobhGJygqsmTJolYXbJCfPAmXdrBBmThxMiKwx4MZNB0gFdQwWP/ZswiZO3u2+AcEKLDF82EJ2auTzfOn/w78cQACde7USb5Fn5snX14ZM2qMZMqcWesRQ0YQzZbxW+ijEPXoYb/M9fU+vWXPjUv8jyN+wAyqyIfjDhgj1L41hjdKBmjH7HCzeRFBgQ8UBOnS5TPp3PkzgDNumqlOpO1B8ENp+0lbBWPOXTgv58FqIYvEEUwZ75zeUqd2bc0hP+bjozFavBLKlIh+MS/9AMAd32u+iMlOzfpoYR8D6DgEAMb3yhU1AU6TKi2kTgB+sP3u3XtKzWKCUT08FHyASUdibBe1d/qwvmOBrfjtGEitsv6LPwvPm+wdfSL5S+y4etVXXz1EHZiA5OnpKZ5ensryKYQPIwETlBA+MwG2Q/7eT1LT3mL07NlT9u7bqx9OoqKpYIL0+qgCBhT9X8iCoR6PEqzz58+rWRVXY2QYo8M4wkLD1C+H27/HB8W6dn3L+oY3+HTYRhhkav3799dfmRb1+v7cU+l2tgPMT1MBUwFTAVMBUwFTAVMBUwFTAVMBU4H/wQpon4QekGDCT0htqoC0YipHsEn2IuxFX8S2UvwjOfskJ6o+0F/zoCTOLlKjZk3t1SZNmozU460KXixYuFAcHJ20D2MvDgxErkEWRPVDBJgs2p5xLgAcFy5d1LQmlpeMF0p56KP6zZCvpUvXLlBDeEsBgCs8aNu2bfIMDBPb+o/DnuA0YA4MNQrbUoJHJC/Qi9UPfW2//v2UVPEZ5lm/fr0CLno5VtMnkbAkyV+gAPxgHlg9ILaTgBEJJQ6DfuglynPi7rQCaQmCQwWQQlpDzVEQIBPXPAK5EhUdfM2C2dKhAVvo8boVjBz6zrAWU3/6EbXdK/nz5ZNjUOKQbaQdKyYgCENLlDAwe+4E3I6tn07MaRJ0xA+Ywanw4XlBEAEnXgTRWqlTp4GxULQmNBVAvDWNc5iRnitXTgVBfl3wq4IOzyFLCg4KVmAnGAAOzYTWrlknK5Yvl/379ymYQpbUbaBmT57AxAcjPPyJ3AQ9KQjADI9/FvlMZU1c/wf4mNwO8NcHhf4xifFApEmXBjHaWSUvmDhFixVRxgwZKjmyZ1eGDM/Ly9NLYqCPY1yW3mFd6e2+EW3jeB1UsM1iAQuv3kC6VHMwh51sITKK6LND+hdfF8QDWbRIUSlZsoRtmgT/SSnVdCCMvO7x48dLgwYN/vYaRFlp7svBfygYocYx85eZcgeJWGQksS4//vjjy68AIJcc+/bu022M5ibdjveYFDNGf79p/71w2+Z27m+GqYCpgKmAqYCpgKmAqYCpgKmAqYCpwP9iBWysGVtjyv56KXw+8wI04HgJ3OD1IyTgpkqVUtky7DSpNEkHtUrJEiUkTerU8FR5rDKdevXrSdWqVbTHLpC/AKdhBo5ahDA1ifIhzsv+jjHbtCPhvLofGz0MqmJGjR4lHTt1RM+bRFbhD/NkwHDQD5Trc1fKfrg2DYdtc1IhUwIKklLwvSFhwg8snKqVqyjBYt/e/VK37ofKtOH6nGg4AJfLl8EOIjCCTWTjcA1ejzOZMWDBEMQgA4YEgWbNm8Ov5qkMGzYUZJDOeAfT6DFk8djpmkxlJgij9cUaDCEnDtC0STO95nswGG7ZoqWcOHZcrU8+/fQTxQ5UpoX5Fi1ZIgcPHbQukhf6Hkb8zH9xUSzUc3zFALkiyGAH9IlFTJs+PQx7XeXk8ZMo2EqY9d7ATUolZEjwfXrFrFy1UtKmTSuZs3ogvclBoqIj5TLMZY8e89FLfQRkijIpmvs6OCEmDA/K1auXcTP9JT2QQ4Ipj0CVosnwvfuBEorXDRs3kmOQFvn5+/NpU7pSurTpJB+AjkOQD9GQl94olC+R5UO/mqioZ9DYReBaqBd7+8Fr4F2ygS1/ZwbSsTj4MyGNavWB/hsnMBsUr36xTBX64Hz55Zd/46hXdyGAxKEfRH5qMCz2kKCez4RsHD781kdV39Zva+FVw69evT7Hs8APDOllz5SRw1k4nzWsOdeuXSdrEdnWC/O9Ygxs2838NBUwFTAVMBUwFTAVMBUwFTAVMBUwFfifqYDVB7FRqvBBeWGwCu0ufNAn24CQU6fPSKdOHeC3esLahn0Z0nICv5cuXUb7WzJW6Mtao3oNGTVqpPz2+29aIXsHi5/h539LAR32a+zBHgaHwGokKXposHAwH9kv9g6JFTzZuXOn7NqxS86cPYf9gmU1ZEl9+/aRH5HeZLFwLPPeYpAi0XbEugL6wDqorMrLy0tBD/qXnjjpA2ZPDRk6bJj8CjNfX+zvBvCEPX04UpqehkeImzsir3EO7NWpbiF+4JDYOi9uL1a8BKw1ohC2s1t+hb1G/vz54YszXY1+6ftKcEZ7UWAP9K15/DhM+3+eGC4LXrRhOJ77WODSPRAnWiG+nCqRsaPHyUWAQwzmueJ7VUkhVPewSG8TTPM2j2P8gBlcxTMY6t5AISkz8s6RHSdqIVS8QCfQpZ7CEfksfFJ8r12Vxk2bqHfMDLAqGPlloXGpIAHKBgmQuzJpbjrf1IfoeTTAlnt39KamSOGGnPN7KnWSGHjZAKBJBSlSyIsQuXLpsiJnTnCQpi+Ls1MSjf4i2kfGBtemmWzrVi1B33KQm9dvKphDACMEKCMfmrSQX3llg9sybtq7jMzQovEu3Qa9iRKfuKlG1PZRD/j6yA7WDgeBDNLIKGdKiKHIIib6K4CG6UpEE/mwfvrpp9DnTXynpXchcco2CgD44mjWrJkypIhk8lz4RcCOP+fAPOrenXtC8yaa89atW1c/xAR4+D4Hf+DW6CB4RGCOvkC2/a13zHdTAVMBUwFTAVMBUwFTAVMBUwFTAVOB//0KMF2I0qCdO3bK0OFDIV1y1IsOgFpk29ZtiHh+qFYiVIwQiHkUFiqFChWSS4ijZkAO6BMK7Hz6aXtN/aVFxeQpPwDU6ah9NYEGWpFwXEWaUutWrVVKxMaMPaV9YicQLNJIvbr1ZNXvqyQGYAa9YulBysTjAgULi/+tm1YjhznWbVivFiLs6RLD95XeNWTK2AJmOCeTjWhL0rNnD8RXt0R4zjQE4sxS71V/ECyYSKXACuYgd4LX5uzkrAbFZLtwpHR3Fy8vTwAze/T8586eI5uRNDwUCVLeuXOpcfFTqHcIztDbJkeOHOqRymNpX0MpmJJGuEHJAnaQQAXL2LFjACDtl+XLV6ivTrVqNcDguWj1q1yagMd7GPEDZnhO0BsFAzG7cf0aAI8QIF3uiLpOLNSXuQFQoQcNb8rjsMcak50LprFkllDqUgUGv7zhjLRODPNgsmbIZKG/Skr3lBKVOUq1aAR4/JDCtG79WmnfvgMYL97qacIbHIHYMGrRXF1dNMmJSFhgUKCCNIz7evQwBIU9IKVA6fLwyAR9W025A8CHRkCU0Bw5cli9b1LDW+V5ZUiy3mFUrVpV3IHwPQTQQxbIxo0b9GEkSNOwYSMgf/diH58/J/8A+euZYELMB48gyXJIuFIiQ942bt68KZTwtGnTxrbpb//kbbEBHa8fROMoagmJahJE4Qfgr0Cc3Yhc27Fju3oEecAMyja4nclTfDYZ62YDlhj5za83jR07dshdADMEZPr06fNyl7/an8AM6W19YDT1d0anTp1lx87t0rFDR/n/2PsO8CrK9Ps3PSGVQEJIAiT03qSD0otgFwuCvXd0XbuuvaOi2LGgiHQEBKT33nuHkEAgJCG9t/85783FwAIiCf+fu/t+bsq9d+abmTOTfZ7vcMpzzz13PrvYNoaAIWAIGAKGgCFgCBgChoAhYAj8LRHgWloHfr4K6xEFC1xT8x//lyNflYPODxIQVJswg7USWoSo7sBiXdzQQHz1NVfJJLTwsujlFggWBt18E3Jd62lmaAyKdUp5GQeBAtVN7KZY3Z/HBiciWTlZup7XtT1ec73+/AvPyeo1q+Ttd96VwL69NdN1Awp+tiPigg3L9erU1ePnIn6E2TClJ4TTAj+AVyx+YZQFozW6du0qfVACxPXvHRAOrML6XcN/sR33W7NurXTp3EUdM+CeSt8tlmXLl4G0clUlDK//xx9/lJ07dqod6mNk7IyfOEGbl47BqnQCNiuugfU8SDrhP4pLlCgCVvzJiJGqCCVmU9MLz78g9953D4KP18idd90ps2bOUnzJf1yMUXpZFzg1EHWDrCgYN9+nki9uqIsc2L9f5s9fIF9ARsS2HSpfKoOscUN6M0HjQ8PwnsKCIuS/RGnbEqVQK1evlO2o+yJYJHY8vT1VeTIQgcI9QeC0hi+teniEMmXNmjXRxGRaiHwBHo9dCV+urm7ag04ShzYqEhCtWrfSY2SDLeO8XSAF69O7Nx6EGlBhdJY2l7RRNi4tPU3tNxeCBMmkJ9QK5IJrn4ecnarSvkMHTZymP47qHsfj98fs3Ie5LiSIFsyfL3WgNurRo4dcd911ul907WiQUHf9scNZfiPT16BBQ5BOEUo8sc+dg3+ojvciUFfmULPw/ddQd00ZGMd4ePI8oSKi0oh/LGRi+eUkWbjNeoQvv/ba61IDliv+HwBDkurUqS3dunVXRRJxHjt27DnJHc7zx/irDzKVN3/s/We/TcQf34H9B2QWqtRsGAKGgCFgCBgChoAhYAgYAoaAIfBfgwDW36GI6VixcpV07NgRl0WKgwM/sa6kYyS0WijW2oUq7MBbsAG5YT0aDEtUV7IPchxtTH379EWMSIHMmTtH/7GeGhamwnDQWrR9+w6oU0BdYH1NMqYYa3qW8HC4Yr3I43l7+8iUKVOle7duWIPnyScjPpUHHnhAt8nDawpLeHwKJ0JBdhSjkpvHdwOzwjwcNjfzH+vprog9FCMDrrxCIz4mIgZl6NAnxR8ZMjgBnW/Txg0QCyxUxw3tVVQGcY2YnpahShxdL+JgtE2tBYlDAcKChYtkKIQEb7zxOta7bo7z5zXhP6pxWIHNY+uAYoaByMx7vfOOO+EA8Zdjx4/LQw8/Il99+ZV8DfLorbfe1hgWB0qO3Srye7kVM5RGhSI7prB+PQkKCJLKlYMAUBoqrvLhd1svrMFiqGxlKGCqwK5EHxdvsoM148120YqtPbAkEUiSAoGBwZKRmiGBvgEgOapIVO0oZLHUkDA8ZLT+pKSmSyU8CO5g/7w8vKB48YcVKVpvFD+njckDhNEll1wi/fv3V08a1StuOG7tOiUa/OvlnS/tGjWWy7peiuySx7XquxLIkgsdlIThXkIWNlxbo9agv50WKjYY0RNIdpLnVHbchKCiMGTxsGKbzCJ9exx8WKKio+Tu8yBmqIxh6DEbqZyD50GCK/5ovBKCTiKGn5Mk0xPlE4Uv3o/TB5lD57j66qtl4aKFqF2bDwnXHv3iZ2yzImnG66Py53yGJ+4vD877dj7DgVeJqo/OZ3tuQ4naetS51apZ63x3se0MAUPAEDAEDAFDwBAwBAwBQ8AQ+NsjwLUf14rBWHOvWLFc3Rqj0bikDhSQDBkgQUh4HDt6DEs9GIaU2HCRPfv2QUwRDEICS0AsFpnpOuSWIeqA+GbkN7Jj20555933JL8oT9JQD80RhDUbCZQClO64uXtifVmkREYJCBYSOQwmLsS6cdWqVTIMgoOuIGgYkcE67Pfeew/ESbHmvXKbLITzaooN1p88PkkbH6zn1yIbloIFyD3kwL79cg/cMXfdc7e88sorUMh0QnPzTj2XQpT1bN26RbNiaI3i8bmYZWasB+u7iQvmZfgxlTWM7mCIcA7Kg9Rx07iRvALb1Tbk45CA4t5H4486WqGwL+uxaZHy9/VDFupUxZDdT6lod6I1auXqVfIp1vmXQqTgCSwuxji/FfLZjoyLz8eN0oRkZLtQvVI9ojrW3i6yGcAxlJdpycwQ6dmrpwYJsR66ccNG8K2dgHImWyVKDmBF82gIcCUkPTPRuVr1MPF395HVIDlqRNbQhqV4AHgMrT9NmjYBi5UgOQjt9Rc/pEP7STrsUmTI8nFcKmAOHjigbBhJhH3wsKXjvX379+vinQ+su7uLqlnYja711Ge7zvN8/0WQMy+gqYgEBmVQ4eHhuidVMGQDycCdPijbInGTg4eV4UL8Y2Mlmva6n77xGV4zCCkj09FadYaP/+0t1pL9lcEsnOkI32WuD/N40kC68fyoxjmXBepMx5gLSRiv80w4nGn7+bA+UY53vttzjpUIeKYk7nzxO9Nx7T1DwBAwBAwBQ8AQMAQMAUPAEDAE/m4I/LH+IrXgIsPRTNyxU0dZtmy5Eih02UyeOAmCB0+JrBmpxTxcX+dDwVLoWwiiBbQEWBE3iCOYhfrMs8+gbWiFTJwwSauvr8W69UQSqqr1P8e/53OdTMUJCRDm0TiIjRJdfzdv0VLWrV8neVDovPf+e1g3ThM6OJhxw/bdIjJBmOvXyVOkbr26SqiQLOLxqVbZsWsH1noBmNdhKTqenCTvgSBiA9Lc2XMhorhM3Rls59X1OtkX7Mvzy8zMgjooTGLjYnEchUNq1agp1SF8CEDbMQUgs2b/LqPh4nnttdfQbjVW/oX1utqxsPkK2MC4DUcRjs/hCjKK7osGjRrA6UNiyRWNT9kyedIkhCpvkCm/ToFoJFq3rehvpdqdC522RK1DO3ftks2bNqlygzYlP19Yi9ChHgAioiPyRqKjo2Xvrt0qNaKSI6x6uAQHBcumLZu1KzweRIuqNHDfXF3JFZUo8ITJA+DMnjVHG5y2werk4+OlTBlrt3zhf8sCKcFucdqiKgcF6E2mNCk7M1tmIvxn05YtSiqE0E6E+dlpHnMoVgryCnAsNwWcPjJWZ9MWVN7BPxZafpykDOfjQ8eKrj/+kP79KGQKaTkiw/h3JBWoZqICqC2yeqiQOde1/PvVOd5h5s9fIVl4zL+yPY9CoqoawpxPytLOdjL2viFgCBgChoAhYAgYAoaAIWAIGAL/kQhgpQwyguvlQTffLG+8/rq0aY2IDrw+iMyYvfv26meUp3D9TUsSGBV1rlAIwO24ZmLOy6SJU+QblPP4IaJiGTJemO/phqgLTuCCKBKONWvWYG2PRl6oZtRDhA/dMOfCBfMRG+KPV5gT7AjbkfogGJhrsq+/+kq6de2u+yeB7Fm9crWqZbiO5PF5XoX5hXBi+Km6hu/TWVEXER9LFi+VTlDMsIGZrVL333+/ij14fKp4aGXailaqDu3bgXTKxTkVi2uJq/IPi3ENdNu4gCOgDmbJkiUy8IYbZfnS5TIOgb7t2rfX41MAwPBkXqijKdgFdqc8vIZjw91L1TgkgRj1QaKJGbC9kS2bXcapohdXQd8unJghmrjRtMIch/9q67atsC5tQMPQQSgWTuiFcHHMUNwA/wAN7NkCIob+s8iIcHjG/EG+FKtkaRNIHcqu+IjUBLPn7+cPfPiY8b5DQpSWovPGxcRqloy2IOFjf9Rok/wBllIFlqcIqGqCg6tC1ZGuKh6qPDIzUIsFwqUerFYtWjSXqKhotTTxeJRVkVWkKoQZLyQCbBgChoAhYAgYAoaAIWAIGAKGgCFgCBgCf2sEsAbmkpyL4cpoKPp16q9y++23w6Xix0W0Eh/MMy2A4qU6LD7gKFTQQMUIB/+xndtTCHEQa/iBA2+QRYsWyagfR8G+87KSMxDY6NgCN8yyZcswLYgezM31N7+SkRXLrFcKJfia6/n9cIHQ8jMb+TV33XUnbE3vw3ERBuoA2/DQ3BRrfRIjdLaQMeI58aNqyJBlXIY/SKKtiPr4ZfQYLauhNWrEiBHIeME82I7HOnrsKIiW8bi+Il6uql84L2NVyEEw25bXz5EOe9b9D96vubCXtGoto0aNKpOrip05aem+WsiDcyXP4YrrpVCgW7duKrxIBMEUd/iwzlnR3y6cmOE14os5IIEgXtRSA0aLjUtJsJIUwuJEmRKlP7SvMAAoK4vqlkINq+3YvqM0btRIOuOmhYRUVeA4IduKLkcuDMkZ1lvrLQLSxJTZNDyOP45HZxnJmHZQcARgWz5EJGEYbpSN4yWhmSkpMUlbj4rA3gUi/4ZZNBGlVqsxY8bIDFR57cWDw753f3zGuW0YAoaAIWAIGAKGgCFgCBgChoAhYAgYAn97BJSlcPAK/P7D9z/IsGHDIHaIUrcIls/4WYT1biXxQfQIl/ClHIQqYFjaw+IakhKHYg7Jkyi0eeqpp4T5qRMnjBd3T0eeCot7GNehh8PanBwLrUW0+pD8cM7pDvdL7z69seYXeeG5F7RuesiQW9CW9JN07MCwYhwK+/As4g7HSe/efZQn0P25vkepEKNJGqLumrNmgryhDekGKF6GDLlVGI0RiJgUBghTwKHRJfhJ2xFP6gQEIuQQXFzcsD9Dgh3hxywTIr/wKaxfv6OluF69uiB1xiGPdoCeE/cnv5CamiY3DboJpAzammm3wbkyluXIkXgoeNg8DGKKLhsFQnetsG8XTszoKeCM8L9AMFJ+/r4gZRI1UOcYQmfJmrlC+sTabObJKEEDcoZpzr6+lbAPgn1ByNwM6dXgwUNUxpSNcB5i6ofPGcTLBwj6KTkMIJhtwsAhgutAokRJmNCQUNRxecvPY0ZreFBOTi5ubg7IoFyt0T4B61IRyKB+l/eV1pe00ZvNJqJjYNjmoQ1pGnxw9es3lNq1aqtMqcKQtYkMAUPAEDAEDAFDwBAwBAwBQ8AQMAQMgYuEgGNtTE7CQW2AN0DF833y24zpmumiMb34aAJIFtZscysSGiQYSKLs3LkD6+MqeBOLcLyfA1vQd9/9oK3ILFO5//775LLLLj2pPHGqW7goz8rO1NxRzkXyhz8DgwIRY1JbGjVuqJk3CxYulKuuuloLfaZC0XPzoEGKBK1IR+KO6Ppe81JLj+8BIujIkWNSo0YtfAayBNdDF8zMmTOkNdqWMzPTpVPHTjoP+QfnIIlDIuZgzEGs7RtoZg0VM2CBtD0qOioKApAuujktSTfddLPs3L4TIcFjlYzi+ZNmyEUsy6aNW8E9wM0DPHh8NkunpqbA9ROBLYCdAz7noSvsZzmJGfBKYMhoAWJeS15ePnq+1yqTRcsSrUw+lbzBKsE/BkatQImZfDQRVQfYNaUqcl/YsR4KfxuBI/mShos+FHcY1qV9ACAVXrUCJVH27juglilKpTy9PHX7k5IpgHb8eKJ+FRTkI/U5G+1NcXjQdkKiFaFyLh6zXu060ggqnaZNm8G7Vg+soRcAL9aw4Rq1avyl9p8KuwM2kSFgCBgChoAhYAgYAoaAIWAIGAKGgCFwoQg4GRrsT4KhadOmyI35Rlo1b6H5rBQtLFq0mG4mJUPILZDwIPnhjlBfhudyCtqgWrdqgWrq+XIDGoQTEhJkPHJZqKKhM4WzO0xHIruQIduv3+Way6KKE5Ag3H/Xrp2am4qN0bXkAn5gnbRv3wFhuyvlF7hWqFYhgcN5GBCclZUthRB18PhZyG9hQLG3NzNeeCWOQdvV5s2bVDlzLOGYDEdD0jvvvAvnTahu4Dw+68JJwKTDyUNihccnV5EKnqEhCog4I4/D4OPbbr8Nc7yDGuy35LJu3UFQBSs5NeO3abJ/P0p5gBb5jOzsHOzjAq6ASpw/rr/01CrsR7mIGYLJk2UDUWR4hEQhZMcZXFsMQoWDPesNGjTUjJdKfvSfFUFG5Kuerjp16ir4mWhTop2IINHLxVqqZctXaKrzTtzwLrA3hUJdw4eHIbm0JaWnZUoCWpkY5ks5Eeut1O6EG8De9oz0VIlB8FGrlq2kaZOmMuv3WbJh00bJyEhH5o2fesSCUB3NKuksBAWrXKnCYLWJDAFDwBAwBAwBQ8AQMAQMAUPAEDAEDIH/vwioIobMC8bgW4fIG2+8LuHIeOHAUlkHbTspqSeQTYOCGsdbWNmL1KxRA7EiV4gXBAz79u5F9McMefPNN5XAGDdugobzcnOu25ORt/LdtyNVTKHvISuGRA+JDBYB6cRYoHNbtjAxO+a5Z5/TbJsRn36mChjuR2EFz4cbHgcRFIg1uubYQEBBIoThu9279ZBevfrASXNYdu7YKQ89/JA89MCDMhtr/FpoW3aoW4qxpi+C9el12LJicT24IhyXQowChPqyVYoklJ6rkki+em2Dbh6k2TbDUPXdvXt3xYEEESkYDvIN5A3o7NFrKsVQP6zAb+UiZtghzuGK1GayYxFQwdC/1bp1a9iICtTrVYifgUH+0hIESXRUlFY7L0FSMkH1BhlD21Eqwniqwk8WFFhZVq9ZLbugdElHgC+tUQnHEuRRAD/4lsEANF+OosEpDxXZrNTiw7Bjxw74044oE+iFLnQqd0gWES9+sa67akgVmTd3vlaHjUWn+TwwgPS08fiULWXBQpUNps75oPKabBgChoAhYAgYAoaAIWAIGAKGgCFgCBgC/0kIkMwggaD8Ab794x9PycTJk6Rps6Z4D/9h/cufR7CGvmHgjepEKQGhwf2CECK8ZfNmadIY6hK8Zm7sp59+iiyW/qjl7oA19bzSwF4HjbB33351zZCQ4fHSkfnKDFqkkejampmzniB5+qPpyQdOGVZqb0OQcDQqp6egQvu6a69TaIuZ6YKRAvdMzx490aCM9T7m5DkwB7YqsmWbt2jmyJbBkSaMnwgVTnvxRfYsq8LZbuzp6aVzMPw3LR1xJhB8kPChPckDjhvQRuIG1Y8SLsCHmTf169XXCuzf0eZMtc04cAUPgvDBx3p88gq5sHc1adxY26iZraOEjx6pYr+Vi5ghL8PwHh94wfwRihOApqUwdImHIU2Ztqb09DTUb01EcNBEsFYxeuZ5OXkyZszP8svPY+TgwRhZsWq1xB6KAfvmgWwZH5mDyqrp06dLHIgb2p+yQdww5Ll6eLgU4IZt2rhZtm/fDoVOuOSC+cpCIFAuvrxxoytXqSwt0aV+w/XXS5s2bVUCxTotpjwXYS7KnrZu3qo1YJs2bMQDUVv6Xt5fGTJXMHHY1IYhYAgYAoaAIWAIGAKGgCFgCBgChoAh8J+NAFUHXN/iR8cOHbAO3iLde/Rw2IbwNgt7mNnC+BBdB+Mb187HUaLTsGFjJVYcQb0u8vvvv8vlffvK4fgjyI9pLPffe69DPeNgKshiYA5XdadQdcP1Nw/Neb09vSU6Kkp6gXAh2XICAbt9eveWffv3y6TJE+Wf//ynki84Td1v7fp1+OwA98YX53CVQvAAx48dEw9YnDgpnTLrNqyXvv36yIL5C+S2IUPklsGDtRyIe6ngAt94fPIPbGlSUgFvOOgagXgjBNXg/fUYKcjEfQrnQWvVZ198Lh99/DEsUAg+1lNwkfHI6NmI45F+4nleDEVHuWqIeB9480iKhIKMITsWjlCcNBAyhWDdpLAEipadsgcyKJIsT/7jSamBSmtvKFtyc/ORJ5MqUyZNQhXWdvFFVRZTmP1gc4qHKiYlmSxXoaSnpKEi6xtYn+qIv6+/sl+ZkBZVRj5NDixI69auhU+tnjRsAFYPDwD71++480658+670Us+R/vNA4LSVb2jIALcAlidUlPTJbRqqPTs3kMfSh8ECCu7qLfGvhkChoAhYAgYAoaAIWAIGAKGgCFgCBgC/5kIONe2pBLUKoTLmINGIrYvcd3NHJbPv/gKtiWoSbiux3+5EEW4wQ3j68caa+wHJQZ0GFK7Tj3ZsnW7PPLgw3LJqtYy7OMPpU27tnLvPfdoHTfX2fxitizjSg7FHsIr8BrYmSqVxMQkKFqw3uZ/2JB5tH379JVX0bj0/PPPI/D3CAQa8+QYeACqV5gxQwsSty9CRAoFIbl5KBGCRUqFF5g7MiJS4nAtjz72KMqAvORGZOIwyPhWtDfRHaOWLnAVvH62Lbm6uCO/lnOSsHFVwmbHzt16ns7rf/HFFzWnlnXhl112mXz80ceyYf16deloMZFeKL6R8angceGKmVIGTuVJUMzwHKmaaYKgoQaot8otbUViRTappmQwcrQnEay2uIm8Izt37VDpVB7kRfFg33bv2aU3jsDiuQBZ4yf5uBFzIZnahh7zWlE1QbRsk7Vr1khcbKxakEjuFKG7fOANA0HapMtM3MjYuFityqpbt67s27dXxv4yVh8yZtmwL70YN5nHx/3A8RhS5KtflCrZMAQMAUPAEDAEDAFDwBAwBAwBQ8AQMAT+GxBwEjRYkmOU6JqawcCNGjbU11yL8zMXNxfEiByXYKzXVXGCT7leZlRIV5AUVwzor23LMQcOyk033iTXD7xelixbivaneo71NQkBzPPVV1/Kxo2blBhREgWT5ZZGkShJhDl5TuQ2Xn75Jbn99tslskakjPxupPSHm4WDxI0OrNkpqvCieqXUisRsWY6GKPV55tlnlWBasnSJPPvMs8pDLMM59erVW7NpeHyOCRMnIDJlFeJWHPMqJpib+bMcqoPhteKCf/75Z21torDkvfffldtuu123yUHBkA5e50UYF85ElF4kLUDBQcHIisnV/Jd8gOjt4Y2rK9Zg3ZSUVDQg1ZF7771PSRledBWoXchgbdywCWROIHrWa4CV81NShh4yBvN4gKlr0qSxNG3cRLzgFyuAv41Ezf4D+2U2mL5pv02XvVDi5KF+mw8TyRUG+a5auVJmz5otP4/+SSZPmYymJ4IN0REATE1L0UBgVoNFR0WLD5Q75A+dnriLgK9NaQgYAoaAIWAIGAKGgCFgCBgChoAhYAj8DRBQvQhyYXuAqFgjnTp30nOi9YgkRvzReGnfrh3KeNCODOLCFWExTvFCSDVHAxIlNNOnTdegXIoeZs/+Xbp06oI1PFubBLXciVrDTZaGHEYe4kc8PaiA4XxkZUhBlKjTpl79ehpj8sXnX2gB0AzUYj/88MMIC8Zc4FR4fI7Dhw8rYUSyxIURJPiPqhpWX5M/oLJnz65d0rVrV7RArZexY8dgnkfQvuwIIS6A2mY9lC+U3nD9TyqDah4KNkjSOK+fOTh9+vQBsbRRXn7lZVm0cLH88MMPGn7MZmiOi8TLUJdSjsGzImDo9mYTUjwAO4qw3qjaUXLTzTdrG1JoWKhUQx3245AYMVyHShrKjry8vNHIlImb5KF1WUHwfYVUCVHmrjqqrd093MQDlVRh4dXF19dXcrJyJA2MVnR0lJIwcYfiNLCIN5ohQfsRPFQXmTG1akVJNtisg8i0mT9vvsQejgXYuEyAXi00TIIrB+N83VTZE465ec4LFi2URIT92DAEDAFDwBAwBAwBQ8AQMAQMAUPAEDAE/isRADmC5bsONisvX7pcnhg6FGtykg7IZUUw8MJFi1BNvUXJjhJlC0q0fCfhaIISGEpmIORl3dp1cjtsQ6zTvufee2ToE0+owIIHUAIF85GIyUFLE2NP2LTEfUkAuWD/ELz3MKxRjENJg9XoBjhgPh3xibz73nvyxRdfSO3o2lIMcohj/vx5UphfqHOpuoXECkiW9Yg1ocXIcU6usmf3Hrnvgfvk7bffkSefGCoPP/QIGqnClUxhY5MqdjAfj8+MW093T8WDxAyv38/fVxuhyRvkI3rlw+EfQWByjwwZMljGjx0vzZs1d4QlU/VRwaN8xAwBwQm5e9CH5ieuALW4uBC/+8rdyHh59fXX5N2330V4UEME7yagkSkDsqRcrbQKDAzQqmxvZLsEBQZp5bYPcmaqhYXBw1ZbIsIjYDvKlZQTyaqWIauVi5vK1qcB/a9EArOvtip5wUZFxcyReLB7HTvKbagEIxnE91LQ9pSLOcjO8UHo0bMHHpihcvVVV0pUVC2HogdWqDGjf0bV1u/6urz4vvTSS2D/wmXUqFEXPNV+BCHVrFlTeiAg6VyDDVV16tQ52RN/rm3P9hmPVQMBTT169DjbJvq+81hMvP47D70eKLD+7Hr+ztdg52YIGAKGgCFgCBgChoAhYAgYAoZAhSPARXXZgZcffvSRDBp8i9RFNgwVJYtBzByMOaDrfFe8LiZpAYYmvxC11iA0ShDgy8FynngE8j7/3HMy4vMRcj0KeObNn48ioDBVoXD9TQEH1+JxiBrxRjsTj642IkzJ1yO//VaboDgvhR5PDP2H3AcipFfvXogj+UXqRNXWYyUlnZCp06eq+satNH6EQbw5WPNTOcPGJbVH4XhpyKgdNmwYok5ugBijiSxZsgQNTu0Q5uuhx+eEVNwcwDq4EdqneJ5FShgh4qRSAMqLJqAkKEq3hV4Iebffwro1SPxRdDR23Fi0SVdxXIOeWcV9K1f4r+OiCC7zYCppRdUENDDVhySpTu06EhEZiUV/pATvrSwzfvtNli5bhtam6tKvX28kQCPvJSdHmTKSOpVQXZ2Jymq+D2xAuvgo8HFxh4WL7RYtW0gNLLjj4uLEy8dLfHK81RoVVauWbncMxAxlTLWhmmFNVw7mij/M99yg2AnVjvbKqP+qV7e+tG7VUirhmJMmTpLg4GA9Tx6fLFp5RhqquT5GgjOvgWSUc/z0009av8XXKpXiBZ42eO733XefdsaTlMkBNgsXLpBVq1ZJB6Ron2l8/vnncvDAAaRR9zv5Me1dvwFrDuex+AfgPKLzvbZt20qXLl1AktXCsXIh01p47mMhnfoAjnX55Zfr3GW/8Z5MmTJFQ5EoZ6tfvz4ItJ7SEUTZ6YO2se+++w5+vozTP9LXZFMHI1H7TINzT5s2TXagTj3m4EGtaiNRNGjQIL2P3EevBwQer2f16tVao3amuew9Q8AQMAQMAUPAEDAEDAFDwBAwBAwBUafLbBTnXH3NNZrnSsVLMRaRJFa4Rnb3BG2ABaWSHwpYidSJjoYo4lZ5DwoXKm92bdspd4FUWbRkkTz79DOya/cu3R57ycwZvyH/pZGuSUnU4AMoZ0QJG2bQUqnD91j+8wvyYVevXiM//TxaBkOpMn/BfPn+++8lOTEJIb8MENZNpRjnSLJFrVaMkMX7UTWjZPAtt8i7774jS5E7Q9FE9249YL2aKm++9bbO4zw+W5a8QA5xkGzS9TJkK8nJJ6RRo2ogogr0+vn58hXLUC1+g7z++uvy0EMPcQ988YgVNy6YmOEF8cYU44vYOk4NbBhCeelNizkUA8lSNe38DkAocCqCebds2SIT905CQO9MECWByr7RwlS1alUJQc12CNi4w4fjoJ4JkOrVqyMDxkvy4QfjgtwfREpTBAuvWLEcAT+Z6mHzAXnji7nzYEPasGmjpEMCVVxUAmtTqrjh4cnKzlKpFEme1q3bSCSIIjJd+ajyDgjwlwSoePbt2ydPQHZFlQtvbHkGQ5SyMrMkKirqFEKAhAuVQn8M3kS99X+8hd969eqFh6CRnsd1110nX3/ztZIYZyJmSNx8+eWXIDVdEE5008l5Jk+eLM8++xyek9KbcvKTsr+4yI033qDEDAkhsptff/3VWY/Fc//qy690grLH4hvPPPOMfDjsQ60wczya9O2JvIQgp4cfelhGjBih+zm/7YL3j3lDZ7p+Mpf+uC9nImZINvGPgCQQx0kEcf07duyQzz77TN8/eT3A7lswsOy3t2EIGAKGgCFgCBgChoAhYAgYAoaAIXAGBErX9Vx/r1m9Stdq48ePx7raIVrgup8lPlSKkKRxvOuC/BZfJVb8/fyhpHGVLdu3Yo15I+qvn5Lvkcsydtw4xI3swwFLIMDIgT1qsx6cRArXcgVoSGKLkype8JPDA2v4po2byabNm6TrpZdqiDDDhJvBQvThsA8c8SOIJeFqkE3OLVu0kMWLl6jqRRBeXLNWTdmL8h+qebLAIbDtachtQ3Rd+tZbb0o75OfwGnj8Qhxz5YqVPKwOchu5udn4DE1SOEYxGqr4HoUeXbpcJstA9DD/5lqsnauD59BJnDtXwM8LJmZ4khweCOaphJCcgoICdKIXQUSECwW9VgQP2NEjRxHicwIBQu3lumuv1UCfPBAHR9HAdAItTWxfYjhv06aNkR1TG7esRDzyPXWumjVqSSAsTgSUKhioi2A1cpA0iccTVEZFVQaDgn+bMV22bd+Odu4iAJwvsajnqlu7rtx0w01Ijk7S0ODQ0KoSFlYNHebBsnz5Mq0KO3HiBID2ADFURUKqhpQbztFg9RiSdDp58RyIksNHWNnlOASxcxJbK1askO04d74XXCX45DncjIwe9qiPHTtWVTiVoEgqO+aA0UwCIUWWjySOc/RmJ7z+ATjf4TODe4XzogyNZAWpEzKTznHzzTcpMTMOfzzDhw+HT8/H+ZH+ZNhyIkKcPGEbuwYsqnPwoSZDygd3wIAB0qFjB9zPNPn6q68lA0TcZ599Lt26dZOBAwc6dznpN6RK6lYwrA4c8OcKbIgB1UKnj+XLl+Mar8dzkY9A6CaqkGnVqpWkpKTAbzhfGjdufMouxO4bxW7cGbE7ZWN7YQgYAoaAIWAIGAKGgCFgCBgChsD/KgKl63osEbkggzjgG+ncqRPWee8735KlSxZLv8tLXRpKomDdDlHD2nXrpHKVIKzlioW2H1Zx/+PJJ2XdmrVKzhyEy4F5MdlZWboOJMTMf+UhM+GgCK4a7LBKcbmK/dVd0/9yECqHNUR4KtwSsRB8vPzSy9KyZXN57bXXwTXwREtk9+7dGl6ck8e2JIhGwEH4Im9m584dyKeJlq3btuo69TjEGG8gYmU53Duz58yG6CMHVqcP0Pq83aHU0b0xJ2xbaanpsDc11jUq+QfKeigkuaR1K6yHE7Bu3yHHjsRLdWToVvS4YGLGeSLMl6mDXvPpv80Aw7ZGLThVQ0JUUlQEgOnZInRR8IdxUU9r09x582HB2Qc7SxaAwRcYNObLEMhLYa/ZCpDyEerLNGZmyDRs1EBfb9++TUODycgxQyYMeTQtmjdXwFMBYjxACkbjExUx/r5+0rd/P2Xa5s6dC1vQEolA6DADhZOxoB/zy89Q5YQr6XPwYIxUwTl7o/3pQgftVps3OVhAWmvKjpfRg3624bT70I5Ujcxb6WCiNK/vGHx7EyZM0Box52f8OXHiRH3ZD/tRkeQcrVu3Fip3zjRIsJCY4f24/fbbTm6ixwJDyt54sqOsLCs7Jk6aiD8UURtTYGDgyY9IyFDFQkVQC7CVzjHo5kFCqxRJF1qPyhIzSkphQzKytGL92eAcd99zNzyH+XIpWFMqZ8pe75nUNbweZhUx14g43XbbH9f6Z8ezzw0BQ8AQMAQMAUPAEDAEDAFDwBD4n0OACz4M/rjjzjvUXbEB7UQlICxYlMOQX7oTlAQBiVIEgiYPjcws6qHyRQd25j/A/4C81fyCIvn00+GIy1gpjz0+VJYsXoh96LrhFC4QLhyRu+64U6ZO+1XnopAgAI3NqxHl0RwkzPy587FdCQQEI9XaxHUsW5dTT6TgHF218GfkNyNB7MCGBLKG68wccAgFsDiRW6Dgh+9x+GG9PHfeHF2zLkKGzqyZv8uLcHiMHzsOap5MPT74IuUy+kDosGo1lTSO8GIfHyiDUBjUvn0HJWYc7IZOW6HfcPjyDfrBSqBUIRPFAN318GpVguKiBXJcggAAQ3wIiItricqAmJocGREhISHV1MLUv39/KC3aKynBJqb69RtIA2SUcL4igMrO87SUdOTI5KMOK0ZDgithjsSkRPjNFsjGTZskNz8HD0QlVeFs3LhBjh8/LnsgYToA5Qj9cVzIr1m7SlLR3kRioSlUFy1atNSHiA8XFTZJx8vXyjRz5kx9iOvWrXsKSXEudKmWYYYMx9DHHz9lU0q8qDLipL/++uspn/HFrFmz9L0bEGp0vuPDD4fpprRG8cFyjpPHwhtnOtbMGTOVzDnTsWghKkvKcE6SQw0aNNA/JGbBlB0kWv7KYIDU7l279RkaOXLkKaTM2ebR66GKCMeaAoWWDUPAEDAEDAFDwBAwBAwBQ8AQMAQMgXMjcHKtBuKkbr16Gn/BtiS6NyiM4OcqvACJkgOBBdf9ugjGtNhFA30HXj9Q8z9XrV4BUcBI2bBhgwb5Pv30s0rskEghCcJK7i1bN6uDhvQJv9zwwQk4biLDa+j6k7wK1Ti0QdGdsXDhIomGm+alf72k1qoCcASFRYV6XgwQZuOyHwQaDjEAnDzYn66PK+DuqF27jkar8B/6J0wcL8M/Ho7g42HaGMWMGR6f9q1fp/4qiQlJJ4+PM9BolcrBlbGFc/y1Na1zr3P9LJ9iBueTAjvQUoT9MHS1UZPGqlrxQSgP81I2b94qP/zwHeRHsVptxVwUsmm8obwUVmNFR0VL5y6dpUG9BnIgZr8sWbpUGatcgMx05cyMTFm3fp0G/NLHhmdAbUyZyJn5ZcwYWYgwoBYtW+nDkoQGJxI4+chEOQpiZ+++/Roc3KxpMw004kPAA9P68jQCiXJRq7179x5MyBTognPh9KefrUVVFx/G020159rxww8/1I8bNmqofemnb9sYBBLPl4G+ZQclYScQSsRxvsfbunWrUDnE8fhpJBDf4zx8cE+3QTHwl5Yv5zb6y3l8Y405/7wql7FncbeTf+znMQc3mQvLFke37t01VFhfnMc3By4usu/A/vPY2jYxBAwBQ8AQMAQMAUPAEDAEDAFD4H8bAafChKt1xmHw+/hxIDE+GS7DPhgmSckgLNTj4yJH4ezocmkXWbZkmZRgjc99AxBFwjV/xw4dZfee3bqOfP31N9St8cmnIzT/0xGzwWiPYpk3bx6cNm74HaQIoM/KyZJAzyDx9PLUG8Hjc0RFRUG4cVzDhjtg3jvvvEvX96++9qrWdusE2O4grFQP3H+f7Nyx86SKhtEfrAa/9LLLtIkpGyVBr7/xBuxZy0DQTJB2ECwwbJjHx2nIdrh36AzRnFcQRdyer91d3dUV5Lx+bFqho3zEDFgSSpiOgJniQvjdd94BC7UVN+koAlm3A58iSU5KhuRnm3q9FsxbiItFPzjkTT5e3iBQsHgHW8acmEqo2A5HSM/bb70jqWkp8Kx5IgtmpQT4+2mDD1UwR9G8RFaOgAUiPDi/ADXZsC95enhJ8xbNZQuYtKCgypAu+YClqwwfWJJs3LARapxQuQXpzLQ6LQXxU7deXT2HcNhp/FHbvW3rNsmDVaY8Y/369XonWT19PoPkilOd8ugjj+mDfPp+nIuPYkxMzCkf8VjMsiEQEVAfnc/4CDVofOC5PcN+Tx8MRubnPK+yQ6+Lb+BEzvdYe/bswUN/QKfhH2XZ4fxjZ/Dya6+9poomWszq1q2jcjlnMrZzn01QRHG0bNlSSH4tWrRQ4uOPIjy6sgYl9+3b94wqGse5lkjMgRjd374ZAoaAIWAIGAKGgCFgCBgChoAhYAj8OQLONZvSIvjGf9jv1LkTQnQf0vU/43v5D/jXXXetZGRmoBAIC0ksTj0Rc3Ls6DE0MDVQYobKBbbxjhjxmezZvVfLZkKqhUgrCCuOHUvQNiTmlpKWoUggISFRGjZo6MijwWtXvElSpHuPHmg+zpCpcJJw/f/C8y/IMLhB6I544oknte6a++fDbcOg4nhk2urA8emmYJkQLVckgcg/0ILFuI62bdtASPKj3H7rELT+HpAdyJChEojnVAziiNXc6chOrYKyooKiAj1HXRjzcnG8ihwXTszoybjAGhQgDRrVh2LFF2SHP26Gp8yaM0ueB1iRIBaykWbsA5tRHnJftuzYKl5QyVxz7XUatrt3725lskgyEEh6w+gJy4Aaxg3VXFMhIyKRkwcFDAkgZsdQisT7Hl0rSvr17YcA3ERJhlKGQcGZmdlyIumENGneBInL7cUdgURHkZvihSpsDp7L5s0bUcXtLQyP3bZ1CwJ/MBlBJStRjhF3mG1BLudNXgwf/olKpUgwlM17KXsKJBeowmFYEomMaqUhQ4cOHVIMPMH+sV76zwaJsjHoged1MhPmTO1TJGYIQxYYwbLHOgzSje97enqf17F4LiRceDBP5PmQECs7nH/kfMBf+de/yC2V4i/C6vOPUDdeNmCY58Ljf/jhR/iCwogvnLcML6KjozQXp02bNvjgj1GzRk19Qc/gsYRjCJIK++ND+80QMAQMAUPAEDAEDAFDwBAwBAwBQ+C8EKDroW2btojhWIPQ3C3iAn6jMK9Q3nnnPfHE+p4qEob65pcWAjGHlus+F6zpWYsdEVlD5sydI5f37yvLlsHi9N238uYbb6L590us/0GCMHsGa7wsrN1aXdJa9tLVwoUw3nNDK1KV4GA9jgsOTN5gxarl0gakyrSp02UCsm8mTp4g0379Ta/lh1E/6Hrf3Q1UB5qaigrhjME8vAZPtD5RsUPbEzNPKeK44oor5Lfp00H2PK9RK2+/87ZWc+OSdD8SMn4gdW4YXqsAAEAASURBVLLg5FEDDj/gmrSCx4UTM0QOYHl5+0jzpi20SYkkyHZUF7u7ekgS1CpszaFsiB3njRs1kUBkvRSgrSkAddjNoHCpFRUFlcsmyc7MUdUMW5VUGgQWi4iz6SkLpEQBGCuqYxLQxrT/wAGpA19ZJb9KSlQ0R1aMG7alxIiWKYblBgYE4bMwuQQ3lcdnb/rcufPgV0vGg+GGFqjm2oEeExunrVFVq4Tq8S8UW+bYZKRn6O4kOP5spKLOe+S33ygpcffddyt7d6Z9VH1DmPEhw4WdxAz353N6PsfivF9+8aWygiSk7r///jMdStuQOCcJKrKfzmMloz2Lx49Anfj5DAYM/zzmZ33gX3zxxX9rWWL2DP8IGcLMvB9eC1VWEyEji0HWDxutmLtD4oyDBA6PTxRoj7vqqqtxbqFCJc3HIHEY3EwMNyKYimyoc0QiZNo5YrCNETNONOynIWAIGAKGgCFgCBgChoAhYAgYAuePgOMf1xmG6yLNmreQl154GSqYEfoP4PkgPpRYwQqQzby+EGW4kMHgihDrSw/8A//l/S6Xlaji3rplizYSL5i/UP/RvV2HdnLPXfdAL0Oyw9Fc/PPon7EPrUwga8jqYJHK/8gFcG5lhdgCjXDhK64YIC+99JL88+mnpWZkLeTPbgBxBDdMTq64QjjC883NK9DWYWbjuJCs4bwlrtK6VWvx9PGUKZOmlLY1Zctzzz2ndqubUeZTUJgvrm6+enyGDDdr0VS5CqDAKXh5FTr+WMn+5WlxJvhfVma6BvSyczwvP1fWrFmtNpPnwTixRacEYHii7agKqrDC0DrEBXlScqJkgsjIgaSIShqSLmTCuB1vuoebh7i6k11z3By627joJttF9QcDfpo1ayEkDeJRQ02ZUWDlAK3DZnAw/WDRUVGaO+MC6RPnXbFiufw65VeZMWOG7IFSh8d2wf3PyyXg6CcHq3ehgxXRzhGCdqc/G2xNysrMUlnWI488ctbNq0IypQM4MwnbORzHc0GA8p8fix4/R/tRiQy+ZfApNdnO+fizbH02bWPOQWKE9znkPJQ5u3btElZV80Ft266tPPvss85pTv6k4un5F55Hv/0/tc3pafwRjR79k5Ix3sgmykdG0AcffHByezf+8eD4/GNjRs6jjz6CEKob5a233nK0T4Eg3II/cGbolB1/YIc/YqhubBgChoAhYAgYAoaAIWAIGAKGgCFgCFwoAliU6f9K5IUXX8A/xo9GnEkjnUwzZEF85MLpUq9+A1IfeN/RakQrEuUwl6AghoPii8+/+Aw2qOvkyiuulCVLloAnCFOLEfditMbBg8gJLT0W1/78vQBEDEOIXTEXCZvIGhFwdITIq6++Ip+jkIaZt2zkveKqq/Q4PD4JnUJwCCwZoi2JWTicjOHD5CXCwqor/0Br07Tp07QhmvMsW7pEuQwPcBScJxvZtOvXb3TMCzGJLngdryrsezmIGcf57NixCx3n7yH7Y7FmuuSRkQJgN9w4UBmsQiy0d2LBzp7xXCQ5u+PG5OLCaUGiJWcb8mecQbYU4VCx4Y2MmJAqIRIEm09lWJS8vXgDACTvCEgYkjdcxDMYeO+evfI72qBWrlwt0dG1JQgypwAoMcIjquuN8oKdiZYmEgK8kccSjuK4scrmcQ4GC/EB+quhtGXvADNz9AbjexLIonMNqms+/fRTXoladnjjzzYYusvtyBISN+egP444JZYhUJyfnf7zxx9/BNYOUuexxx47/eOTrzXgF/jzeI4aNMdHlG3x+EmJiSe3PdMvJD/4h5UKkorNVEy7PpNl6kz78j1myFxz9TV6fLKczuFHVhTHZ1L26eMq/NHpM4MP4uLiTvlYsePFAKfykG6nTGovDAFDwBAwBAwBQ8AQMAQMAUPAEPgfRcCxZtYVqvTo0RNkxQYU+XTRAh6HeaFE9qO45hjWhtyW63aKINIhykiFm4bD+T7Xi7169sJnafLQIw/LjQNvxLawKmGbgnyQH+QGwAGwKYmZLxRykBOgZYqjVq1oR6kNXrOp+eWXX0Q7c4JMnzpVHnn00dLjcA6RBQsWovgnx5Fdw7dwXiwG2rBuvRI2qCfCettdVsK50aNbd1m1co289uprMnjIrRCdBPFwJ8uCcFmcQN+ryG/lI2ZwPsyFWYcw2skI3uHPNrQPYTGfAdIkwD9Q8TyCjvIN8G/RjkPWiuE5efkFqrQ5iiDXQ3GHVJpEUgNrcJ2TjU0BQYFSo1ZNyKWaKqlSgH2ILD/jgpz5M4XwpJ3ATd6/f5+yWjehProYrFoO5Eu0TbngCSFLV7lysNY6t4D0ig9DMvvPqZTBXKl4GMozGFjr4+Ol13r4NILg9HnJ4pFEIMl0LqKE+8XGxeqcvFLafpwjqPT3eFjHzjV4nbT7cHTv3kOaN29+1s2dxAafMwYoOweTtXl8Zs2cbdCy1rdvH9mHexCCe8sq7/NR85w+nx+Cnnl8/esp/bB6OMONz3x8Ej+O7U8lk7irYlz6YeUy11M6rf0wBAwBQ8AQMAQMAUPAEDAEDAFDwBD4CwiQ0Cg7KHRYhnKdW+Ca8EXeLEmLadOmqQqmGDyBEiBQqXBdyiwZDm1KhuKFognGUdx2++2yBq1Ib775hnwy4lMVVXD9p6P0x/p16yQc6pYCqF/IJ/BtH8R0zJw5S3xZ2Q2ShXmzPXv0QgX2R3BXvClfIM4jFI4dbrx9+3aZO2++7s95yTlw0F0Ca47yDzx5ZuIciDkoDz70gLozBt08SMaM+UUiI8oU7pSek2OGivtePmIG50EZkTeICZIFVKdQDkSly+aNmzQfpBFSlflZBixPTEN2xcXWRChwRkY6iJMCJCx3B5Hjrx6xrIwsZaoIEFUszVFzTUsKA4apeGEScyHYsiCQLAzrCcD7Xt6wPxEPAMmw2SpVquoDwRrvxUsWy8e4MZlIgq4E5YU/HpaGDRvqDZgOqdLuXbtVTZKRmq4sXHlg1ZuOCZwEx9nmGjZsmH7UsmULuQyVXecaRw6XpknjAsPCIO8qHaGlQbY5CDMuaztyfu78OXPmTKG9iPg8/vjZ1TLc3kG8OJ5QZrg4R1gYHma8zd76xDOoZpiy3b9/f4QqbwGhEyRzUG9NxcxfHVQSrVi+Qndr3rzZyd1bIIuIx589Z7banE5+gF8WLlyoL3nWzK4pO8oSSRdCEpWdy343BAwBQ8AQMAQMAUPAEDAEDAFDwBAoiwAYCjIx+N+dd90lL77wgtSrV083YJsy13CMI2FUhQ+blel8wX8UdriCGakG0uTZZ57TFieKF4Y+PlTuv+8+XeNFRkYoqePkgVauWi1ZaEtiAxT5B2bMshlq7949IITgXuFrfJG4eRqRGffecy/qsS/VfFXmlHLEIs9UB84X2htV3zB/Vi1OuA6eqztamxi5wZboUaNGyTPPPC0dOnaUJUuXSefOndX+RHHIxRjlI2Yc90HZru6QIbGyehtqskmeuDFzBtkxnTp1QthuM9RcV9KmJBIqJFm40OfNuvPOO6VHr26OOioPV008ZmYMbU9DIB0aOHAglDdBGtZbCUocd2TPUE7EFOaB1w6UPn36qhIkEKoIgkkZEkcRwD2RcgJJzdPUYjV48GBV1lC9QxIg8Xii+tdYBz1/wVwNoS0PwAyr5YPD+c82WNW9DmwfByvH/mw45/Lz9ZM6deqc3LxVq5aO33Gpzm1OfljmF1Zk8+GPRljylVdeWeaTf//VQWSUKHlV9li0GCmliIs7nXSipIwNSgzrZcgylTK6/b9Pf/Kdf/zjH5rz46hFc7zNzJwHH3wQFes7+EicUuet1d74w01JSRWGCTtHbGyscC6OjnjGoqOjnR/pTycxQzLuQoiiUyazF4aAIWAIGAKGgCFgCBgChoAhYAgYAmUQwMqNC2Au4DA6YE22dNlS6diho2NFjvU416Jce3OtSJcJSQ3+5+IKRwjcMckpSdq4RLJlOqxNXTpfqkKOBfMXyXXXX6/WJs6fhxrsKVOmINSXHIJDMVOI+Tm3F1Q7VOEw29YLMSW1ataScePGS5/evWE/KtTYk3vvvddxkvzOU8bx8/LzpGpoFb5ULoIsApueWSbUtHFzWKfcZQtyTDt16gCxw049/iMPPeI4J85TwaN8xAyuIgQKldpY+AehcSkIF7IaFVrZmZnSoX0HtAbVQIgvWCgE6xKka6+/Rnr37K2hvZnYhuwZ657DwiIUDMqZunbrJp07ddRWoOLiIjQsBULt4i2Vocbgov/22+7Sz1idHVkzUrpjey7qr0cFN5uR2GuehbmZSUL5VH5BPqg6F7TyhIJly5TVyKLZBilTOhQ7DLzlTUxExXZqWmq5oHUQGCI7d+486zzsWufDyWsehKTnPxvOuTh3WdlY48aNldziU0RFzJnG5s2bZf78+frgPwqPHUmrc43de3brQ9q8RYszHMtD2dA9e1BbVmZQsbJgwQJ9h8qZLvAX0jbGL1rE1AeI45JtdA4mWrOSjEqo9u3bC2uu2T418tuRevy77rnnlLrstm3b4vXVevz3339fnzWGO1MhsxN/IMz3+eSTT5zTn/zpPFeShWWxO7mB/WIIGAKGgCFgCBgChoAhYAgYAoaAIVAuBFQWgW8kOKiCWbFyhdyHJmBajVhgxPXvwkWLVMzBAymJgg8YvjthwgSNOSHVwjXkqlUrhfahWb/PkG+xPuzevRscNZ56flxvcj4qZTjy4bDx9vLGWs/xDhUzUVHR8jDKYtwx98GDB2TYR8Pk+2+/l0/RIMU1I9edtFVBHyPrN6xXN00RSR3wDlwzesONQyFFRGR1zbZxB5exc8duueOOO+ST4Z/Im7BINahXX49f0d/OvVr/s6PhBoRWC9GKY1/UVzMxOTS0qhThwnxRT8Va6yAQK3kALRWKhyz4vti2RFLmxIlkNCwdg0piuwbL5mTnAB5XTVNu1qy5qhx++OEHmTt/nlZhF8ObVgmhwP0v7yctYAPaum2rjPh0BHrQl2ildq3ompKSmiJLFi+SediH1dmesFYxX4QWJ0qmijBHAc6R/ejZkEI1btJYhg59HMqLJ6VqldIGpD+75rN83huMHB9KEgJUfpw+aM+a8dsM3YZqGTKG5xp8YH799VfdpE+fPqdsSsVRt27dVDbm3OaUDfBi3Lhx+hatXVQlnWvwWFMmT9H5+vTpfcqmPFb37t31PeYIlR0kRfgH5BxUwfAPjV/6e6mfsKw65tXXXtV7y5rsNWvWIDBqvT4PDeo3kF/g3/vm66+d0538+dNPP8kDDzygYcJM6V6N5i8OhkWRgLrkkktObstfeD3Ocz0du1M2tBeGgCFgCBgChoAhYAgYAoaAIWAIGAIXjIDSJPymDI3j51dffilvvP6G5rzy7W1Yuy9atFCPoYoZbob1OYUVXFNSUOEcschjpZ3ooQcfktZocpo9exaiTIIcH2MzZv+SRIlFTm3Xrl2hislX/oEbVKtWTRYgS6Z+/XqwKLnp+v+Bhx6URxAuzH/sHz16tISHV1fb0/GE4zL116mlkSZ6FbqOpXOHhA/5B4e4gfm0yfLW22/JgAEDEKfi7zzVCv1ZPmIG589wXT9Ik1xx4QziJUGxa9duGTt+HHxcsdqIQwnRMVz4iuXLNCCW/rJ85MswY4Y5Kvv27dOgW5IlwWhVYi02lTYkONatXYemoxOyB9skwH5EGVQlsG9ZWdky5pcxSHB+RCajBptcnL9fgC7e4xEovB8MWWVk3Qx9YqhE1YqCNSofip4AZcEIO5k1d3cPqVunLlQ4t58zGPd8EO8I71m4BtX+QYqU3Y++OlZPp4I8YpX4n43Vq1crfrwuVomdPtTigzd/m/GbZvec/vkbb7wB+0+KHD0af0pw8Onb8bUeC0HDHM559UXpN8fxXWTGzBmnHIvKFcrHlHUk83iWr7vgOXSOoUOHwgu4V7NxGPZEYoVEnbNq+0zqFv6xfvHFF8glytBnYvu27boP67Op1jp90FrFZ4/jTNdz+vb22hAwBAwBQ8AQMAQMAUPAEDAEDAFDoBwIOLgNLl91XHPttXLH7bfhH9FRkw1Chc4WLMLpG1JihFYiEh90WrC8h4G8XOszHwZ0gYbujp8wUZ0W5BH69u2n25N/4EHi4g5DdeOl61MsQ9Ulw3KYeKx/aWeikQrGGRAvhfLtyJHIRR0AoYcvIj6ukhsQl8LzZFSHKnggLOHGjNjw9PDEutYRjeICjoMkzW23DhFfxIvQMbK1TIOw40or5juv6sIHTri4ECoUsErx8UdUucIq52TUM38A28mYMT9r9gfVInwvU2uy0hQkLpy3bt0mO2BHIRjqP4PiIioqSmJi4mTj+g3Ss3sP2I1A1ADMdBAaJ5KSJRivq1QNhSctQEOEUpJTJCWZddAlUiuqlobydAJJUgTCgOE/7dq103PIzsqShx96WB599DF4yUKVRCA5wwdAb+SFo6B7klAYdPNN+jvDi840SM6UbVc60zbO9yjrwqlJs2ZNpUmTJs63T/4kWULVTTYIqhkzZpx83/kLH3KG8Z5PZfX48eP1j6VZs2ZnPRbJMMexZjoPUa6fDOSlRYtNUbSwnc/g9TZq1Eho5Sqr1Dl93wm8HoyzXc/p29trQ8AQMAQMAUPAEDAEDAFDwBAwBAyBCkQAXAGFE7QGDRp8C4gNX11zFoF1oUDmBMQXjLXgP/RLCb0zLhqTQhFHw4YN8KpE4sAZcN3L8N/RP/+EsOBn1WXDz7jf2LG/qPiBC3run52TrdyCY31JAshB4oSiSIeij44dO0nMoRhYmz6Tt996WyNGeMXFxY4w4RysrRs1aqikjQvOg+txNgdX8vUHz9BJwclIg6XqIozyETM4U4bt5iKMhx3mGzdtFk9kyXiBgKCtKTMjU7NhMmEdWrt+rWRAEePn54sb5Kf7JB4/KosXLVbwuNj2C/CTAkiRSjArq7Ufg9plyOAhqroJQLMTgeENqIT5gwKC1CNWu24d2btvL+RR2xHE464hQi1atVZb04F9+5FR41DR0MJ0CQKDX3n1X9IR+Tdkw/z8fMQLLBuJu4oYj8GiRCJkx46deuMvdE4qTyZMnAAc5GTA7elzMaPldnjdOJRY0d/++jcey0kkPfXUU2ecgMcacuutiv94KKH+zkOvZ9IkPUVnOPDf+Xzt3AwBQ8AQMAQMAUPAEDAEDAFDwBD4r0OAa+zSdfaY0T/Lu++9i3bmCLUwUcnCxt8+/fqgMRlkCv7j8IQYgM3JLVCsw7Uwc2OUF4C44qUXX5J/vfIv4fqf+TScPBtxKOQdirgx5jwcG6eN0BR+6LFxIMaaXA2VTEsU6BQV5su8ufPklX+9rLkxP4z6AUKOtiCKSqu8cdQqwVWE7cfFJHtwfK7vKTKJggiExAGFHRdjlI+YwYknHj8OWdJRVVOkogXJA1KkKrAjRYZHqs0pHjkyGsqDsy/IL1SrUqPGTaRWjVpSMypKIiIcVVjVUbMdHBSMkJ5DqJG+VHr27Kkhvv5gqDzdvSSyRiRulJew5prWG95IdwBdBXalgwdjNJ9kM6wxBI032gWI5SDbJhCfUxKVjO0nTpgka9eshZKmCLk3eSBxguARC6wwXGvWrClsfyIVOG/evAuelzk1cXioqAC65ZZbzjrPU2glYgaMM4D3rBue4wM9FiRctWrVkpvRP3+2QZLDA6RbeY51trkr8v2y13M+AcsVeWybyxAwBAwBQ8AQMAQMAUPAEDAEDAFDoAwCIE3Imzz84MPy67SpEgDhBEkPFgEN/2g41vXJELw4iBlH5Eke1vhVThIrnInbfoncmssv7wthBdekC4UuGVcs/F04Ofan5ib+6FFpA8cMCRvyJ5yWxIo3GqLbtG6jpArDfr/97jtV4oRXj0Cu6zS4OJqpo4Y7jR79k+zeuefk8ZmXSnEIY1Dgt8GkPGDFD/dyTQnyIyk5WebMnQ1CJFmJDma/0CLk6+erF7du7VqVKAVXDRYG/HrAs+UF1iqwciCsJs3lgfvu14YkNuzQnkLGjEGwjRs3lUmTJoiPt49UDw+TI+xCBxW2ZMlSqG3mqtIlHj41ZtF4g1ljNTbDhGl9Ila9evfCeYCcyQYDB7tVBlQ7VHt89/13ejN5rBMnUqQI+TeCwOiKGt9//718/vnnarO60DnZOJSN86Z1xxE4dOaZ2BPPfJYz5bKceY9/f5fHotWMBM+5jlW/fn0QZVmOB/bfp/nbvHO+1/O3OWE7EUPAEDAEDAFDwBAwBAwBQ8AQMAT+WxGgaILXhm8tW7ZWIoVr9Z+hojlx4oTjqsl34LdUZKSGodmJ9dncwYXCF6zpWfbC/NT58xZKTMwh2bt/r0ycPFHee+99FAJ96uBKsAtLiDagXIY5tnRSuLlBh4L98+HwKSwuxGt3tTzxoCtXrlS1zOTJk1BmdDPyTLMQtbJF0tPS5VBsjDRo2FBcwXdkIRLFG5XcnM/FxU3Pk/tX9CgfMYOLLwaDlJeTqydMsmP37l1QpBQjSyVAg3MY6hMGT1cH9Jkfij2obFMKGpp8vHw0KTkIrUkpqSdk+YplIGF8QRC4K9lQE4E9xxMSQfBUkubIPtm8eRN8af7SNKKJJCBIOAFKneNQwRDsm268UfvEk/Ca+/tU8nMkMbu4Q81zRKZNnaZWJ4YLH084BtLEV1U9GzdskMiISGnRopnuz5ybihjMkinvON85zidD5s/OhQ/a+QwGM/0njPO9nv+Ea7FzNAQMAUPAEDAEDAFDwBAwBAwBQ+C/AQFdbeMbw3jr1q0rn34yQstqqKDhWpzCifoN6oMzyMcrUDXITaV4oDWiStw83OXLL76EeiZPnnz8CbkBHMAHH7wvHdq3k4mT0DCsahZBNfdCOGs8QKLQHISQFFid+HtRQZEKNEjU4H/qyElNSUNb07Xy+++/q3skKChQ3njjddmyZYse30WFIzkIIA7RKBbdkSoQskgVQx2cvK3ltjLREsTzYi4MB5uZqBkKDAySqiFVEUBbWS+iqKhAvKB+oQQoKSlRQlCr7YPQ1917dsuBAzEI8E0FIZOBryyZiaahD4cNk5atW4HUiQDZ4i25sCWx5rpOndp4rxpIH3e8lw2fWBHkUIHKdm3ctBF11Xv1+EePHoNEqljtTGz+WbR4oeShBYrnw0Bd1muzCouKkwKoZmh9smEIGAKGgCFgCBgChoAhYAgYAoaAIWAIGAIXEwEXee6Z5+QXtCwza5atzFyNF8AyJOImh+PiHSoYsAwkUdyx9j8KB42Hh5uSImloO/5g2AdyFbJjevbqJe+gyprCDnWSgJwoBPfAWm52a+ehndnb2wu/IzMGRAsHiZrGDRvLjTfcoMKReXPnyKuvvCJNmjaVCRMmY95rwBEU6D7Mq9m5c4fGrBRDdcNpL8YoJzHjop3eR+HlqoPaaZIeBI7nyoulwoIg0CqzBxXJ27dsVfYp4dhx/dwDwLDTfMbM37T2OR5gsxo5H2RLcnKihvs2qF9H/BEK3AqesKioKAT/UlXjsN2kQWbETvH169ep1ckLVVYkb3hDV65cJZu3bIY6xltq1qqp5+KOc2JCc/PmLeSaq6+V1mgFqoxOdLXw8EmwYQgYAoaAIWAIGAKGgCFgCBgChoAhYAgYAhcFAVqCdGD93bVrV9mEAqFGTZpqEzPZmJHffCWrVq/EJlS2QDeDzblPWmqqij+cShW6JGbN+l369O4tqSj2eeLJJ+Xuu+8WN0SdOIcbNiaxkowGKHIFJXD7kK3gGTBAOD0jXaNWClETtW7dWrh8OoAAipNfp06Ru+6+Vx0+nGv27NmyHJXdei7c+yJwB+UiZnhJaalpcgz2oCvQC969W1dJBWBs8akSHMRrwMBZQ42Sj7alZOTAkEQZPfpHWbNmjQLEIJ1M+Ln27t0jsbGHkNRcU5qiHroK5EKxhw6B7QJ4uCF1akdr9XVEZASInlyAgjqt0BBlsvZg31XLV0rLFi00GCgmJgYyJXfNtGFXee+evWXQLYMksmYNBAH74aZ4SLVq1SQysoaGEZPoOfmAOE7avhsChoAhYAgYAoaAIWAIGAKGgCFgCBgChkAFIuDMR3Wsv0u0POiF55+XRx97THkEkibHkB3rkHuwyho90HiTFdteEH0UF6HGGhRDTazlr7/uWtm4cZN8/PHHsmzpYuTNjJAvYHfycPdU94y2AmGmGb9N1/U+1TMcjhrtEjQpb0dRkZd+5gLLVDoInl69+8h7778vjz/+mDz66KPCDFMOunx08AQvwigXMUP7D8Hx9PCCWiZIomvXgU0pWeVIV19zjQRXqSIHDxxAMG+SglkCNJmCnAZm6njicQQHJxFpBZuesuysXCVhQkDsVK9eHQ1M02XVytWSClUMVTlMaWamDH1lvFGvvPyK3HPPPWp1YiV3ZRwvF9aqWbNmSVukMXfq0ll8oZDx8HKH+qam1IK8KSq6tubTHIk/otXetF2pp41314YhYAgYAoaAIWAIGAKGgCFgCBgChoAhYAhcVAQcahiswbkMB/Py9NNPC4t0WATkXJlzia5qmbRU5M1WUnbGoXdxldDQUGHbM7NZKQ6ZPHmK3HjzTXLbbbcixmSBREVFO4J+MQcFIhMnTdRqbc2u0WOi7SmvQG1SJQgN1oH3AxGT8sJzz8ozOJ8GyLsZN26s9O3bF+fJHByyEBeHmSkXMQNeRnNighD0O27CeFkGeQ9TjP1AhjRs2AiqlyrIi5kh48eNlzWouPZC3bKHGzrHSykvXlIx0C5Bbzh/X7dutbz51ptgvTYqgcNmotjDcWhbSkD99Hy0P82VXeg1p93JB4RQDViU+l8+QAJRh30Uqp2o6FoSER6BGyAa0NMMHjEGD5OQ8fP1BfGTJfkgbnJhrWLHOeuydyGTZtXKFZIHRY8NQ8AQMAQMAUPAEDAEDAFDwBAwBAwBQ8AQuPgIOIgXHIecAH5cccUVshxr8w6owlbGhhuADlm6dClcNY0lKzfHIarA1p7gA0b9NEqCK6OVGS4bZtlOQx13t8u6whZVJHPmzEYcSitxQ5MSJ09FsDAH9DZQzMDiBE6CmTVKEFGOg/fc0dp05913aTbuzJmz5NXXXkMmbbaGAz94//1at839lM/Q2SruW7lqdtgZ7olecA9YgXbt3CUrly2XmNhYMFSLJTwyUnJRj80GJdZYZ4IUoeerRo0akpaehv28AR5dX+CcyPBgruycPJAkKyUOczRBXfY111wNpcwxrdDOxU3YtWOXhISEwDLVQ2rVqiU7t+9QpQ6PX4zGpYiw6lDKtJUs/L527Tqwar4OUgbEDbNl9u3fp+HCLVq2gPLGAzfJFZVcJ2Tp1u0aFhSJc7ZhCBgChoAhYAgYAoaAIWAIGAKGgCFgCBgCFx8B5V5wGFIwJAWqBAfLihXL5eGHHxFPWJJcQDowqmTJ0uWSkZoOwwusTbo9WpwSkyQ8IkJJFb7HD1avXiODBw+RJ5E58xjsUfFw3jz77LOOwh/uB2KFebgpIGrq1W+IWJbjjvnASbBZeveuXdK2bRsc85AkJyZLjx7d5OPhn8iHH30sTZo0U+sTSZyKHhdOzJApwgmxZ/zSSy/VLBiqYg7jwmMOxcjcOXORIZOkDUgHDsZoFowbfFuVgytrNVWjRo00V2bPnj0a9EOShqG9lCqxTakIKpr69RpIANqdjmHO4cOHyyTIj2hVog8sMjxcJuB1gJ+/dOveDXYpMGA4H098FoRA302bNuE8DkodWJd8QcykaEe6C6q5U7QlysubvrMiPW/l4Soe24q+VzafIWAIGAKGgCFgCBgChoAhYAgYAoaAIfBfioBzUe4in434TA4c3C+0JB08eFCWLVsKhYsHiBkSKy7asAwCQMN7yR8wGoU8wJUDrpCfRv8sTz31lIRXD1eCpvUlrWXg9QNRx53MXXT/LSgKeuzxx2XhwgXKP9B2w4iW2Lg45NvWRqbMQbQ4ISsXzc6PPvKIrFm1Sl559VUJDQm9KNhfuJUJYHC4onkpPCxceqGm6hGwWo+DlQqtGioFyIFh4G91gFG1arCkpKTKgoULYVEqkuCgYFW++PsFyAaE9TDQJyAwUDNf2ORUyacSEpBd0W5FeAmQpwy4YoDcOuQ2SUc11oaNG+TQ4VgN+s2AEqdZ02ZyC8J9M1C3ze2DAoKE9ipWbB86FCsrYLHat3c/1DRtpB16ziMjImGr8sK2rsqcKRlE+ZINQ8AQMAQMAUPAEDAEDAFDwBAwBAwBQ8AQ+L9DAEvzEihlaiPDdv78+dKmTRslVHhCumrHt3S4cKqFVQOXgHyYYpRsgzugO4f8Q0hIVZAvrrJ+wzrpjNzZIIg9SOwwBoUV2qQyshFv8s3Ib7RBWskezF2A2u5i2KDcXN1UKOIKGxRJoMqVK8M29aP069cPu184hXIuQMs3KwChbehE6glJA/FC2VGrVq0lrHoY7ErpEhsTC5brAGxD7khGdpM9u/coCcNg3hyQJnn5uciccdOqqkAQM5W8K2lYb4/uPaQ1/GC5eTmlF+7we9HxVFRUKPEI7t2wbgOalxA8jNyaQrzn4+0jJajZ5h0Lh5rmEty8Fs1aQNoUjnBgLxAweVIZhFBrnN9ll14m0WDBjh6NZ4aPuKPBieSMDUPAEDAEDAFDwBAwBAwBQ8AQMAQMAUPAEPg/RADrfhYNcYUeHR2FmJI1chtEGhRwUB1DNUYsBBi3DRkibFpyvOciAf4BsgbbNmvWDJuUgGBxh0hjhQwceIMsRk7NqB9/kHZt2yr/wKtLRnFREV07qsIRcBipyJfxxifIoCmEZQpsiTeybG6/7TYVn+yCzWnlqpUXhTsoHzEDQPJUlXJIpkz9VeKOHAYpU01yQbhQrZKWmS7bt23XrBcSJW3g1WKgDpuVCgsKAEKR1ERWzDVXX6NV2LWiammY8CXYrmePnnIiORVSoaogUaIkATk1xaXES0kxgMJ/rpjHC/kyOWC7MjIzIV0Kliy0MxW7FKMW21+i6kRLI4QQR8B3RkaMtdpJSUl6I/wRIMymJ1cwa3XAxFFBY8MQMAQMAUPAEDAEDAFDwBAwBAwBQ8AQMAT+jxFQcobngF9A0oz6cRRqrN/Tem2KKtJQbT0DbcypiCohN8DNmD6TmZGpBUAUddCBw3HgwH554vGhaFp6Rlq0aCFfffkVxBueOjfJH6cVqiC/SBo3bIgdHQ1MbnDY+IAMykN7U/ee3TmVZtxQRVPR48IzZkrPJAAECFmkGPi+kpOT5coGV8o3X38jO0DIlOCEWSflDlVMBJQrXbpciryYBEkAIcLwXnrFqHRp1749MmOqSyZamNauWQsVDU4L++bl5SHct7LUbxAqs+fMEX9fP4Tv9JD0jAzJAhGTn5+vicrEhZIjVncfP35cdu/eLc2bNQdJVB37B+IY3pgrF/Kl5SByvKU6JE8toZzxxvsFIIj8AwJwY9AWZcMQMAQMAUPAEDAEDAFDwBAwBAwBQ8AQMAT+7xEo5T+UZMHZ3Hfffbrub968mSxZvETWrFmjViPlZbBROsKBWavNgiLIWvQzxpd06thetmzdJt9+951cd/310rhJEzRCr0HA8KOyfPly5SzIP9AolZOTi+KiTGFvNIkdN3AZaRlpcPf4KB6OUyKdU3py+m75v124YqbU+lO3QQOE/14GtUplJUkoAyJxsmnrZlWoMKCHtqZEyISYD8NQ34ysbDBVzeWqq66C5egS2bhhA97L0myYK6+8Qu1JVLaEh4chIHivHI47jHrrWJUM3XvPPXLzTTdDUeNoZvJH5TWtUk5oePxUVHanpKUguCdWDiHBmawYE5RJwiSnJIMxi5EMnEtDsGE+aGuieqeigS3/rbEZDAFDwBAwBAwBQ8AQMAQMAUPAEDAEDIH/bQSUAuGCH8MFRMnChYvk6muuUQEGiQBWWLMVm+v/6OhoR44siBMqWxhb0vqStjJgQH/d//DhOLkfBM9WCEkmTJiAcOAnwGOU0iI40PTfpssG8BNu2Jf7F2ruDBQ5GCRpikngkCmq4HHhihmeEIYX+sKrVQuVYOTLFMHvdfRYvNqDvFBtRV+WC7JlskG67NyxHdXUKajIdsf21aRe3boI0QmSyMgaMnLk1zJjxgzpcmkX6d6tuxIwJUWoygqrKt9/P0o2b9qoahtvZNA0bdoE+9aBPClU3EH65KGlicenKicHqhjKjjSPh9jhKw8hxFTn3HtvfQkNDRH6wugjKwaB4we1jwfOn14yAm7DEDAEDAFDwBAwBAwBQ8AQMAQMAUPAEDAE/mYIlHIhzIilWGXKlCny/gfvy8cff4Q3QMwgECYjPQP8QoRsgzqGLiYlZsAT8HM6Z/jaBQxOTMxBeejBh+TOu+6UDz/8UDZv3iwvPPciN9OYlFzEoDBEGBtLNlw9vpV8tNiI+6PNSI9f0ehcODFTeia/zfhNVq1cJccTE6Ux/svMzMInjuTidAQD82KiwFoVgWiJj4+XI0eOaEX19u1b5ZZBgyUCFqaqVUJkE+qqaGui5Kh9h45K3lSCmoUWpAMxB6R+/fqyb/8+Wbx4sdxzz90a6ktpUZUqwWh9ClXbEzNj6tStx9si/pV8JQv7Mo2Zyclurp5St15dqY7jrVi+EgqeBBA77ug1P6K1WNzHhiFgCBgChoAhYAgYAoaAIWAIGAKGgCFgCPxNEdBlOyJTQKL886l/wqbUSb6DRakI0SZc0v8wapSKL4qRS0sihXk0+XDI5ORmlebKslrbFXalLPn4o49l186dMnHSZPnk0+ESP/CIbNmyxREmDFbBFfszX4Yum4J0hARjLv53MUa5iZlDhw7JbzN/k0ywU3lglliHTZTcQHpQ/kNr0a233iohUKusXLFCVq9eo6qV7dt3yiHYkx588EFUZQfIkiXLwFRtlAyE9XB/P19fqRpcRQKCAiTQPwhVWGGqfmGT0vr160EEJSBgOB9NS/gMgcNwSKk9qV6D+mheaqW/L122TA4fhg0KX7XRwkQKjODyZ0FBkQYRL5i/UFLQKnVZ18sEvVoXA2Ob0xAwBAwBQ8AQMAQMAUPAEDAEDAFDwBAwBCoAARIkDlmFi3Tu3Fm/Phvxmbpo2LTE4QbnDpf+hXDLuEGoQaEIA38dwg2Rhg0awprkIr//PluGffCBXI4q7OnTpskbb70pI78eqfQLi4LIHSxatEgaNGygDdM6uX7qOAPH6/J/Lx8xA7Kofv160gAXFQeSha1My5YuE2/UU9erVw9WpSoI4z0GRUxV6dC+g0TVipKrrrwawTtb5PfZs0G++Ck50qRpU9iT6svX33wls2fPQRZNhgy4YoBmzbiCzaqEAJ+m2KZX796SgfwYzaopKAYRlCspJScQPByrPeZsXWL7UhfcnBo1a2J/b4T6rJMjh49Ix04dYX+qJgX5hQ7UcJfIoLE9Kh5KmywwZjYMAUPAEDAEDAFDwBAwBAwBQ8AQMAQMAUPg74uAWopKT89B0rjIw488LMcSjsvXX38piSgEwmpf1/tU0nB7V1ialEoBh0FHTRjiVdq2ayPbt++Q2Ng4eQ/kTEOQL8OHD5f27drLK6+8AvrFQb5QRRN/JN6hmGFdd+n7FYlQ+YgZGLwaN24id995F6RBOQjhWYDaKiQhB/hrgnH//v2kc6fOkpBwDBe8TdxdPTTvpSoqsKtADZOWmiar165WoiUkJBTZMK6oyD4h27Ztlb379kjvXn20/opV1vwsOipKq682btoEYI5AfpQBhU4lyUd7E+uyNkBJM33adFm+Yrn848knoaYJhE3JUzZu3KA5MiSFWLFNi1QOAogT0BDFaiw3V/rObBgChoAhYAgYAoaAIWAIGAKGgCFgCBgChsB/CgJ/kDQl8vrrr0Kk0UmeeGKo7EJTMy1LJG62bt0qzZqxDAh2JHAYJG2YVztv3jwVizCcNhMFRm+8/qZs27JVRv88WguEJk+erKQO50hMTiyFBGQN3UwOzqbCYCofMcMLgkSodt1oqV2nDtQnebIAF9e+fVu1Fnn7+IiHF/rBcdKbNm2W/QcOwN4kEhVdG7kwVZGQ7KaNS0w+puWpdlQ0JEINNS+Gypi9UMCw0alaaCgImjSECGdrgxIblWh3opWJSpeE4wkSFBgI0iZcjsQd0Z7yxONJ0hTg33HXnTJs2Adae5UN8sgT50tihnk3o378QRKRjeODMJ8/bmiFYWsTGQKGgCFgCBgChoAhYAgYAoaAIWAIGAKGwEVFwMGU8Huffn1lVedVcsWAK6WgqFD5kyVLloifnx94hhy8xn+wKNGmlHAsQXzxfhFLi/T8SmQq7Ew9uveQj4d/omKPgwcPytixY8FXFKuCpoSKmQomZXjochAzuGzUROXmF6AZKV/c8B8UQtocVbcO7E31G8gGKFWYK5OdnSk52XlgqPJl3aYtsmrVGqkLC1QKWpp4TempGSBYMiUM9diXopmpOn7m5+VLIMgWWo2yJQf5MrkgV6ByOZ4ov4z5RUmdoMDKIFV8JRMkjhvSlbt06eywKyHch1VWBK9GZKT4KRFzFEE+myUY9iq2MmVkZspusGgkZni+pprRJ9G+GQKGgCFgCBgChoAhYAgYAoaAIWAIGAL/QQg4mBInXxLgjwzbxUtk3oJ58tmIYG2HnjlzZun1QDGDBqeiwiIpRFAtW5ohrAHZ4ti7Q4cOsnLVKnnggfu1semrr7/WWJZ//PNJjUUpZlX230sxgxPC/4qLCiQ/Px8ETS4alw7rVyZIj2KQIlSh7EA9dUpSkkSAIKlbp64cjj0iJ1KSQObkynbUWG2FbakhVDI+UNdQccM67aCAIPEP9Jfq1cKURDmeeFytTBkZ6VDgeEGWtAsNTfslOioKWTZ1pWfPXgjzKRRvby+JiAgHuFDpYKxcuRL13QkI+fESD4QRp4PASUpM1tybQtyInOwcDQOickdvhu5l3wwBQ8AQMAQMAUPAEDAEDAFDwBAwBAwBQ+A/DQEHZ+LgKnqBJ/jiiy/kxRefl3379uulwHgDxYxILhqcAwODJBut0sVQzJDb8PH2lo4gZpKSElE2dFwmThgvn33+uUyeNElatm4lNwwcCFIGrUMXgZkhN3RBgxfM81ElDwgYWpHY8X0IwTmr16wBy7RCyRKqVTyQfszQHZIm4RHVJTS0moSiAalxo8ZSG7YmyorIWGVnZGn2S68+vbBNqBRA2cK+cT/fAFW0rN+wEeHCSxAqXFn8/f21p5w5NWS1unS5FOSLp3ihcju/sAC5NakgYBJkzepVUgW2qX5IWe7Vq6cevxjnmpaaCuYMX2mpUoAKLJ6/DUPAEDAEDAFDwBAwBAwBQ8AQMAQMAUPAEPjPRAD8yinjpptuRKvSYgT6ttP3NSwY/AUtSjdcPxBCkyIkzDhyZ/zAMfy/9t47uM77PvN9gIMDHBz03nsnWMEmVlEiVS1bkkss2/G6ZGdiW5lNHN+bcifJyrl37+7cZNfOrB1bO7uJvbKzltwkSyJNUlSnSIoixYpG9N57OwfnAPf5/kBqvBnvzAYSofa8FAjg4Jz3fc+H+EfPPOUKzSObN9U5rSPSE4Vfsmdm79698MX6eJ4XmdDJectd8z9d6G1+s+ooEztzXHFud3fPW/0sYbpkTOmx2ez+/gHk5+WjbutWV85rTcezriOGYguVqKTkFOzem4ua2ho3hd3LSWubsrJOmsryKoolQfziqSeRwUWnrOxM1wszxHZlA1hdVUXBJgYNjQ1uXruDk92169ZxEisKsXTUtNNN89zx48jNyXXQzLmTn1fAmNR+RpsK6ZqhY4fnN+HHRZ540+bw0SECIiACIiACIiACIiACIiACIiACIvABIUDdIi8vD6fPnMGXv/zllf//p3rT3d2NhoYGTM1MOxHGCTrsnZl2FSs5rjTY+mjssHqWT3/yd/CN//MbePirD98UMKsWZuweQ2w17uzscHNUaakpiKQAEh+f4Bwy9rU9x8+IUl5ODmbnZnH29bO4wnWm3NxcxCfEIy7WjxwqTjalbRGmgcFB10cT5fU4F42JP162JZeUFLMVeTOstMe6YezccfFxThDqH+jHX/3FX+Dhhx/Gtu3bGE+aYx9NEKcZY7IZ7nxGqIZ43vbOdoyOjrJ9OQpxcXH88Ludc4tQZWdkOQfOTSGsk4qACIiACIiACIiACIiACIiACIiACKw9gRVtxV33H/7bP3Bg6BojSj/D5YuX8LOf/dyVAC8thV11ygKrTlZGgVbSNE7S4Dr0TuoVp0+fwjf+j2/gwK0HUFtb+46/j1VHmdyd8H7NJWNiydj4pBNXNm/Z4pwp6yiKmPgRyQCX5bjsY5b5LSvbtQ6aLZs3o7ikmC6XSExMswSYc00lZcXuTVr0qaCwEAcO3OZmsIeHRtkfE4sMxp+SkxOdepWYmOhEHdsjt7WnCwRbUVHh9sktUkWyvLewi0BNTU3j0sWLeObpZ3Dy5EnXiWNCkDfaCysGys7JdPf+jtPVCUVABERABERABERABERABERABERABN41AhZfcgdFmgqOEH3/0b9HXR3jSoww2WDQEn9uz5ienkECdQawHDiCeoJ9xMbGYP++fdQmDjAZFEB7R4c71Tv91+odM7xzm5mK5yqSOVAi+Gbi+HVKaiqjSOXIK8hfmbimU8YcLIvBALNbjAtZrw7foL3Gx0jT2NgYvv/9/8LinURUlFdg+/btyMrKdItM27bW4aknn2R8qY3iSZYrFg4GwohhXGkpvMQCns3YsryFMadYPuZzPTXGPBgIMOoUzYLhAOY5kZ3Ac/uZCRsdH0VoIIQZTmzbtU2cmeL09jIDWCvK2DuNV+cTAREQAREQAREQAREQAREQAREQARF4twj88//XLywqwblz5/Cp3/kUnn3mWQoxK8NFJtRUcLBoanrSCTaR1C2iPDEu/WPVKJYIonJwU97G6oUZ3o+JI14uGll57xLCqKyqZFTIi/MXzjsnTXpGOm87jjPX8xhg50xraxvmKJSYVch9UEWxz8Psjmmor8cTjz/hlpu+893vuGiRuXGsiKe3p9eV8HT3dNEI48Hd99yN8+fOo43nKy4uxG23H0RHRzsbk7/rFphiTajxUrzh6212u7y8DMWFxYhPTEBfby+8Pi/6+/qcGBOkYGRCzlsq2k3BrJOKgAiIgAiIgAiIgAiIgAiIgAiIgAi8WwRWin9NX1kRV376+E/x7b/7Npqbm51bxu7ruedPIC+XBb9mROHTItivYmaPxVDQvY4Kxk25/dULM7wdm5jq6OxiNGmO3TDzdLpks8A3jO9+9+/dPPbnPvtZ9snksNtljKU6MyzdHaJDZRLNTU345S+fdOtKiWw+djknE3oopHSyyDfI8854Ilkg3I+6LXVOAJrjvHaIQlAEp6x27thOsaYPTzzxBCa4vlRZvc4JQX19AyzziUQVy4H//f/3H/AaY0sDAwNum9ycPOtr12HTxo2MLyXg2LHjOHf+vGthjmIcSsLMTfn90klFQAREQAREQAREQAREQAREQARE4F0n8M+dMxZg+qM/+iO0t7XiP377W+7+Ghsa0dHWjsTkZCZ9PEz+LLKeZUWysOEgp9bchHfytoQZm7i+1nQNly5dRE93J+688y5XmmMFvL8+fMRFkmqqa/jzS66HxpaUrCtmgdGm8xRFOju76bCJdItJ9jMr4jX1KgIeRqPglpkKCvL4tpddR8xW5sBsdntpOYKlwCwLpuhiBcRXeP28/DyUlZdijstPCXHxnOKucSJQgG6YpsZGvm4JycyLVVZVo5RRq7108cwz6rSlbiOFoEUXj7oJfHVKERABERABERABERABERABERABERCB9xiBG0JNSUkZvvN330FL4zU8/+ILzoCSTA3CfDVmRon1x7P7dtolbpbDFCpuwvG2hJllqicWNZqkC6a9vQP1VJcsWlRJMcYen6dIMs4lpPPn3+T3S8hIT+caU4KLKZmQMj42ija+LodOG3v83o/c67ph3HnpngnzHKAIY0rWIgWYspJSrGMDchsVLeuZWb9hAy1H0y7zZT8vLilx0ajExCTXHzM2MoqJ8XHe3xQLglvouFnEwNAgMjLT3T37/TGoqaphf02Oc+vcBL46pQiIgAiIgAiIgAiIgAiIgAiIgAiIwHuVgCkwPI4cO4r/+5t/je9+77sukWNr0JbqGaOmYSYSVxLsnmvizPUXuVe+/b/eljBjhb+xLNH1eqM5U00HCot+w6EQbDp7+9btaGpu4kx1J10yMYhm94zf50dmegY2cpFpW91WbOByk4k25oIJ8XW5OXkouS6uzLKg1yJOExPjWA7zbdM/FMEFJ0+Ux10rmuW+n/rEx9HH7hpzxUxNTtEtM+uWoCyq5KHt6FpLC86cPes6ZOw+AwtBN5k9xbblxvpGHD16DLdy+cmEGSv20SECIiACIiACIiACIiACIiACIiACIvDhI2BSy1/927/C7t178Mhf/1uEWQpsxzPPPMPHdmOJmoWrYXmHRRm7xtsSZsK08VhBr4klZvHpZ7SooKAAtevWY2h4EBcuXcDg4CCqKivpblmPnJxcCin9bDZmYXBCPA4eugO7d+2h06aei0u9dj/sq5liF80IYrnaZIJNB+eozPFCjQb97JWpZhTpxhEVFY10TmgPDPSjp7cHbSwANlFmx46dSEtLwwJ7aazk18p9s7Oz6KLxuEjT3OwMy3sWMcjSYXP16BABERABERABERABERABERABERABEfhwElgpBl5xwRy64yD27NuDH/7DP7r40sTkBA4fPuzAmGvmZhyrFmbsfuymTPRITU6Bn0KKl/PTtrK0cfMmRpKisDC/gJOnTlJsmeX8dQrdMMVuFntsbMLNU+2/dT/3piK4pORlDKmYK0vtuHq1HqdPn6GQksO1pzgXObLeGNvZfoVlvpPTU9i58xYWAw/w+lSweA/WR2MizsjwMFpbrjkXzdDgAMpKy7Fr9y60XmujgyZqpacnYsmVEUdGRCI7MwvdPT0oryhjvCrRQeeFdIiACIiACIiACIiACIiACIiACIiACHxICNzom7nxdmOZ+vnK176KFtao/OAHP6SGMMIfMcVjTzBtZkXDse/ekWNVwswNNWmZGaMAC3SLi0tgqlJ3dw+s38UXE+16ZErLStnt0opXXnmF3TDr3EdSUrITRrq7u1xvjEWWfvzjf0JRYRESOGc9w/Wml1i4Mze7gGwuOllZcCqjUalcVVqgCNRLZ0xcnJ8rUDM4deo1FOQXIic3F+OT49fjSBEUixYwQKfOvn378M07/hoBCkT/9E8/Zr/MEPx+ij1UvKwdJz0zgyLQad5zAvbs2UOBKPodgaqTiIAIiIAIiIAIiIAIiIAIiIAIiIAIvE8JXBdf/vZv/8ZpBX/6p3+Ka9euYXnJfvDOH6sSZlbUpGUncmRQ3AhyZam2dj3nr9PoOqGDhW4Uc9TMcZ3JHDTmnJliHMnL3pfKdeswNDTsVpmGKZ546LI5feY056uP4cCBAy4KtX7jJnbAXMUIVSkPlaiEhDhs2LgZ8RRk7DwWj7JeGTtPMt06D33mIf58AxobmnDixHEEXfaLahbLepK4xFTILptTp0/h2PeOoa2l1bltTOwJ8b5NpOnu6eL5tkmYeed/v3RGERABERABERABERABERABERABEXh/EaCWsHJE4MEHH2THzC7cf/8DK6mdGz96B9/RqoSZGzfop1Dyuc99Ds+/8DxefukliiTJ7HxJdytKS1xAGqJDZZjLSJG2/81Ol7mFeQo2S4w8eSiiNOKxx36E4pJiZGflULyZp1tm2pX2fuqTn6RwU4qLFy9ibGyMNqEIVzK8c+dO56j5yU8ed7eQmZUJ2xK3SFNVRSUy0tIxPz+LV0++xqjU68jKzkQtRZklTmXbc4K8B+uj6Whv50y3l2XDufj8v/o8+2gy3kGkOpUIiIAIiIAIiIAIiIAIiIAIiIAIiMCHR1yQAAA0TElEQVT7nwA3orkUncUlaTN7XLl85aa8pVUJM29FmWjj8XHtKD4uAb8+cph9MNOoqCjHRz/2Mdhs9jyFmFlGkzxREXSmTFGMacA0xZcAnSrmtBkbH2M3zCK7Z0oYScpDRXkFkulksQWldE5r21pSQrzFm6ZWYksF+S7GNM2eGXPDVFKMsYLfs2fPIZEdNykUZiwqZcfFS5fQyinuN998k8rWxzBiAhGDYEumfLHXZpYLUlM8T011FTZtqXPv46YQ1klFQAREQAREQAREQAREQAREQAREQATehwTMHrOSa4qgQLNhw4ab8h48j/D4l575RpTJHCitra24dOECmq41Yz5A18vUDIt/o+lMGcSly5fcKpL1xmRTYfLFxOD0qTMoKCzAoUOH0MLXRnFqOy1lpTy4urqSfTF5TpgpKCyk4BOHOTppLLLU1dXtJrN90T7k5eUx2xXmfLaHMapFdNIB8/xLL3AVqh87t+9AQly865jp7+/DMF0709PT6GHJbzAYoCZjy0whTnAvISEpCXfdfQ9KioucMLPyvv6lNPR8ERABERABERABERABERABERABERCBDyKBGzoBdRl32PjQe6r8125oydaQRkfRcq0Ft992m+t3sRnqRT4+NjrmnCrT0zO4fPky+2aW3FS2fb7twO14/sQLFFuW8ZGPfNQJOMkpyS7KZL0zZcWl2La1Dnv37mWs6RxOMp5kYsr8whw20+GSl5eP0ZERNDQ2uGiSh3Gp/p5e5NNVc/DQQdSyc+bP/+zPeN5+N7dtkScvBSM7t8173/+x22hHyoQvOuadp7ry76W/RUAEREAEREAEREAEREAEREAEREAEPgAErusy77goY2hWFWVyihHbfWPogEnmClN2djbiExKQSAeKiUevvPIy/LFx7JtJc+JHcHGlb8YWnLbUbcGVK5dx/LljmJmdYvQoCVspwMTu2c140ZzrlLG403xwAUnhZUaeMnHrrQeQwz6YCxfexMT4JDo7OpDGlabEpEQXderg95u3bHEOm67ubheDWs8yYotIdXV2uhUoK+uJoQjTw1WnoeEhlgWvx8GDt7sS4UXe31uQPwC/MHoLIiACIiACIiACIiACIiACIiACIiAC7zCBm+CWsTtclTBzo2PGxIzK6mp8nstH8QnxjA0NO2FlYW4B87NzyMzKwC237MLBeD+ntHvR19dD14qXq0pDePTRR3H1ylX2u2zmSlI/SktL2R8Th9jYWFfw29jY6EQVu1Y6u2OiOcFtMaSZ2QmcOXMGt91+m3tNf/+Ai3wlUaSxye2fPvEEjjx7GOVlZZjg4pKXUanFxSC7byrpzLkXnRRqTMgxMcZEnphYn3PyWMGwDhEQAREQAREQAREQAREQAREQAREQARH4rQRukmywKmHmRsbK7DEmmJRXlOPhrz2M//7DH1KA6VmZkOK7sI1vj9fDCexC1DF+FAqF0dHZwVjTFQoss+x+mcIY40hXG+pZxDuNIj5vgYXBNrcdYKlvfWM96usbUV1Z5a7R1NKMCAo1JuBgCc6pU8gumpzcHMT57bEljI9z/pqumct05ZgIU8nemtsOMGIVWMD42DjW1a7DrftvxX/69rfwox89hgOMXyVQVEqlA8fiTjpEQAREQAREQAREQAREQAREQAREQAREYK0IRK7mQtRGVg6qRTZzbZ0x9pB1uIRYCBxaYrkunzQ7P4fuzm60s5w3IjISSclJSGHRbyRXkXoonti8ts1Wz3DNaXpqykWZ6rnc9J+/85/dxHVZSQUjRjyPLShNTuLky6/i6tV6rjdVYZkFN5NTk1R/lp3LJoZLTtZrMzg4gAhPBCew0+ClaJSUkIQHHngA7XTJ/MVf/gV+9tOfuZLgSDpkmpqa3IfNetv96hABERABERABERABERABERABERABERCBtSSwSsfMyi1aBOmJxx9HHTtiquhquXHYjJQ5fMKLYbS1t+H1189gCyNLxcXFFGR64aeIYu4U+4hiGe+Vq1cxTeGljPEjv9+PYbporPF43637UFhUwKnsU+gfGqREs+wiSPmF+VgKhdBQ38A40jiFGR+CdNiYSGPz2Sb+bNqwER/96Me4DtXvyoEtznSB61HzjFlNU9ApZ7Rp9549TlRaorNHwsyNfz19FgEREAEREAEREAEREAEREAEREAERWCsCqxJmnD2GwkkgEMA5ih3dnKK+5ZZbkEpBZOOGTWhoaKSAEnBtxeamGaN4Qu0Dubm5jBhdQfW6WgovEWhubnZCjLlXxinM2LJSBWNRH7n3Xp57AZOT4ywPzkJ8fBzeeOMs/uAP/oAOl0ZcvHCRQlAFrFR4bGKSUaRE9FEkiqb7xiJNiUnJSEpNxqbNm5DWnYqnnnoKheyfuZfnXaJ4Mz0z5865f/9+vPjii5hhjErCzFr9yuk6IiACIiACIiACIiACIiACIiACIiACNwisTpgxOwzVmWhGl9KSUhAMBjA+MYH8/Dzcefch3LJrp3OzHDt2nFGnOa4hcaaa8SXrpomklyYzIx3FRfcgzPnrTK4uVVJkCSwEMT077VaWCgsL0N7WiZOvvoaNmzbRWRPFxaUEFLBPJsBrfe+7f49PP/Rp1HJ5ybJYFo0KBoMYHRtFSloyqqoqUFxYTGeOH4mJCc69ExcXi0OHDiHeH4+klCT09fa5zhtTj8JhRq+c2nQDiz6LgAiIgAiIgAiIgAiIgAiIgAiIgAiIwM0nsDphxt1XhIsh5ebnskOmA91d3aipqqYTJREZ6VlITU7BpatX4PPHYP36DcjIyHaulKXlMDwRUSwNjoLPF8MumFRs27od84F5nH/jPGNIyQjT1RLJnhhbTppjSbAvNgY1NdVOwLGQVAkXnObm5l0UysQhu2ZaWgZnsUfQyYLhRDpoCvLzEZ+UgKLYEuzdt4/unBYKQz0UbSpZFpzLye5L6OruQh3nu/Ny87n4tKq6nZv/L6QriIAIiIAIiIAIiIAIiIAIiIAIiIAIfGAJrE6YoXBiHTDW5zI3N4cwy35NFElNTUFGViZmZ2Zd8W5ifDxCdLKsW7eOnS6lnK6Owvz8ApIYNfJ4vM5x43pmuNwU54nnmlIXzp0/5ya17zx0B0KLIQwND7GrZskVC1tUas/uXchgse/Z828wkjTD882zQNjDmFQOtjC6NMOi4EZGqVrZbZOemQ5fTAz2sEvmuRPP4fDhp+mauYOijwfDw6N0yoQpGGUgNy/HiTwf2H9lvTEREAEREAEREAEREAEREAEREAEREIH3JIHVCTOMJFmaaXR0lGLHsygrr0BuTg4XrBlv8kZjHnN0w/hQVV1DcSVI90oBvCz5tR4XixylpqRyxrqKsaV4jHHC2oSS5aWwizqZS2ZhIQA/f+bjqtLjjz/ByFKtW166yrLfQp7LH+dHCsWd/oEBtLW14lpLK9LT0rHAmFNOdg6uXLqMn/zkJ5idnaPgEglfdIxbeWpra8Pzzz/vnDit/DonO5tFwOyX4XuJ5GqUDhEQAREQAREQAREQAREQAREQAREQARFYSwKrE2au36EtHTU1NVMUycBmdsF0dXbCHxvnOl+WKLQkJsYjOysb1VVVdNUsMe7Ug7HRMedUSU5MwvZtOxAIBZwzxiaxPVFcaWKfTEREiJ00rj0Gkyz3NWdOeno6vvnII9i+fTt2smi4n7PY09NTFHoWMTQwiHEWDE9MjON/vPk/GJName0eGRlGK0WbY8eO4ve/8hWUFJegobHBuWxGhoYRDobQN9CH6tlqIG0tsetaIiACIiACIiACIiACIiACIiACIiACIgCsSpgx54sV+a7jutIf/uG/QUtrC+NEURRGptDb3Y3svGwKJRNOtDHXSlFxkXv+ACev+/p6XEnvxOQEY1CzFG+SnLvGZrPj4uJQXFLi1pWsWyYunj020VHObRPBDhgTWaamppzjpbGBAgsFm9LSEhSXFnFtaYmT2Ys4cfw5LjIlobKiCimpqYj19yNEkSiVX3/6oYdw6cKbeOPceRejCvD11i1jLh0dIiACIiACIiACIiACIiACIiACIiACIrDWBFYlzJgoY6tM8XHxeOSbf40f/OM/opnOmbzcPNfxEg4vu3Lek6++igHGjV59+SV8/OOfokCyyNiSF4t0uUxSYLl48SLdNqkoKChyy0iTdMBk0H2TlZ2F3r4+FgGn4ODBQ5igyDM3Pwtz4URyUjsYCGJkbATtrR0UZsrxhc9/EUcOH8bE1CQjUHEUgSLd9HZiQgI7ZNKRTlHm6tV65Bfmc4TJw4iUD1u3bUNxcTELi6so2qSvNXddTwREQAREQAREQAREQAREQAREQAREQARW55hZ4RZBx8o8C3inODsd4tcrfS719Vfx3HPPsXMm16029Q8O4plnnmW8yIfdu3fj0MGD8PtjOaM9j7GRUcaYFumY4fd0r5w+fQZlJaWMQ/nZF7PgzrudAsrx48fcORd4PYspmTsHyxzepkDEehpk57ArhlPbTU2NdPFYr03IuXCsGDiCT4jhtTs62vGLn/0CmzZuQs26agwODsHL2FQWe2ZiY2P1qyACIiACIiACIiACIiACIiACIiACIiACa05g1Y23Vpg7SmHlFz//JY4ePYrGxiYXX1rkXPW58+cpsrxGIWYXNrC4d5HCzQwXlMzxsnHTRieG9Pb04KWXXkIPP89z+tpElGGKJdeam2GRp7mZOfdYgCtOs7ML6OvtQ5CCS0Z6GqezaxhzinPCj9UQh8IhnhuMJIWQwGhUFouIk1OS3MqTj3PalXTFFBSwgNjrpWMGKCoqxtTkJGe+210hsWv/XXP0uqAIiIAIiIAIiIAIiIAIiIAIiIAIiMCHncCqokxUM6hvRFAoCWKQxbsWOzKhxcQPc7jce/fd6Ob3MdE+pHOO2jpc2ts7cObM6+yPieF0djkdNgHUN9SzFJj9LxRbkhOTcQtLfQcpytjcdqwvBvOBBYyOj6Grq4MOl0EXnfJ4opCclITE+AQnznR39+DNNy+wn8bvemNMoYng/dnz5ucWkMq1pgcffJD3VugmvedsqYl9Nl6uR9m1zXVj098WztIhAiIgAiIgAiIgAiIgAiIgAiIgAiIgAmtJYFWOGZNl7D8r8L1y5QpFlSQnlszOzlAcicEDDzyA++69l1GkRRQXFaGmutq5Zbq7uvHMs89geHiYQkk+1teud46VAAWUMN0uf/z1b2BrXR3Onz/nHl9kl8z09DRsXWloaMj1xSwETNBpcNe2wl9bc6rnPdjy09atdRR2hjinHccemRhEREaw3DfCzXLn83p33XUX9u7by7nuFOyimyeZk9sjwyMUctYSua4lAiIgAiIgAiIgAiIgAiIgAiIgAiIgAisEViXM3HCXBOYDuHTlMuooiPzOQ59hd4wfi4wVmSMlkeLHIoWZqsoq/Nmf/jlyc3Pd49bnMjc3h5TkFNx1911IS0tj9Oks2tgBw15eOmVi0cQ4k81r9/T04uc//znFkzH3WhNcwuFFFvleZUypG6PDo5iicGNRqUSeL4PunJHRURYCl2Lv/v3IyMygQyea79T8PayloZPGrh9D0cZmt4uLi3H4yGGKOf36fRABERABERABERABERABERABERABERCBNSewuiiT3SZdJsWlxdi/dx+SGC0qKS5Bf28PxsfGXaxpbHyc89mTyMsvQG5BnosaBRlpqiirwEtcaepj/OnG7HZbeyeG2VcTnxCH8rJy3HXnHS6WNEqR5cSJE3zekrtGWVkpe20mXPdMmF02gcUAZqam3XnKSkqQyzWnhsZm597J50LUYjAIH4WYqGiv65vx8vNyaJnunSX3mkhPpFuNGhkdR1ZWzprD1wVFQAREQAREQAREQAREQAREQAREQAQ+3ARWJ8xct8zkUMx49NHv4ewb59DZ2UGtZtktKU3PTMNiSw2MHBWy2yU21sfS3gwMjYwght0xnZ2dOMu+mWKKKZGMG8X6fPDQLmMOmG1bt+GRR76JU6dOIYFz17Xr16OBS0/RUdEs7S1yYk9nV5eLJF27dg3L7ImxIz0jjQ6YrUhKTsYkI1ZWNjzC69mEtycyEm2tbbjW0orpySn4/D7s2bXbTXqb6HP97Xy4fxP07kVABERABERABERABERABERABERABNacwOqEmRu3SUUjRPdJEmNES1w4iqC4EmBh7yTFjyguIAXYB7PIgmCbv961ew9ee+1VNF9rdp0vEzEmnszyZz6W+sbBz5UlW1qa4mtnOMEdTQEnh+tKD336Ic5gN6G1rQ1D7IMxp8xiMIA9e/biox/9KN688CbG6Hgxd46V/Zp4Mzzs5+MXcPnyZT4+gapwFea47hQIzOPatRa0trbgl7/4JfLy8pGYEH/j3eizCIiACIiACIiACIiACIiACIiACIiACKwpgbclzJgrxZwocyzgDYdt4YhCDftebIo6NTUVf/yNP0ZjQyPaKKrEx8e7/piTJ0/hM5/7DA7cFktXTS8jTJzGnp3nSlIUfHTODAwO4PkTL7E3JogFzmiDLTjZjCiZGPPCiy+6st6xiXHOczfgzjvuRGpyKiYYn7rCrpuBgQHnsmFOiRPbMxjn8yLplrFemagoL50zLAJOSGQfTTLa6J6ZmJhCRWWZE5DWlLouJgIiIAIiIAIiIAIiIAIiIAIiIAIiIAIk8LaEmRCLfkfGRxFcWKAzJgQvF5n6B/oRCi/BQ6Fl48aNuHr5Cv7rf/1v+P3f/31s27GdwssgkiiObNy0ETVcUhoZGcMYJ7Hb6bixzybu2Ly2fTTSKXOtqRlf/PIXXewp9eJFF5GaYH/ND37wA7x68iQyMzORnpaKYHDR9dR0MFI1NjbBj5W1JZ7Odcosgb0yS2DZcCr7ahKxg+W//b297l5tMluHCIiACIiACIiACIiACIiACIiACIiACKw1gbclzFDxoOqxTLcMZQ9GjKJYpjtI18rVK1fR3tqKEMUaWlZc8a83yoP9+27FPXffA+uIsa6XiSk6ayiU2McSz/Gzn/+M7pgcVFSUMcbEFSeeb2xslP0yMdiwaT0WGJOyRSfrsEmm66WzvQ39LBHesGEDNm/c5EShzvZWBCjSTE+yFJiCi4eRqiiuRGWkpvH1Qd6jhzEpH9avW8fJ7XG0tXc4N89ag9f1REAEREAEREAEREAEREAEREAEREAEROBtCTOmy4QYYZqbn0N4OWxDTYwyhemCGXUulZjXYrB582ZGju5Abn4ul5ZMJPEiOysLUxRlXn3lNaRyVjs3L9cVB1sfjHXUxLJfxjppCgoK6K5JoqAy7xw5aYwtZWVnc0Epy/XLTLJX5vU3ztoaNpJTUzDF+NLrZ9+goBPFee4clLJceH5+3sWZ0jmdHccumzh/vIteTU1PIUjhKI79NxZ10iECIiACIiACIiACIiACIiACIiACIiACa01g1cIMa1wQ5iKS9ck8+/QznMXOR1JiIvz+WETSlbLEn82y3NfcKqWl5VxHKqAwE+UKeH2+WMQw9hRFF83g0BAmpyaQTxHm61//Y5bzNmOB0aj4uHiUl5dhx44d+Pa3voV6LjNlZma788XHJzhHTWVlFWYoxkxS0BlnvGlmepoCThCz7KYp8hbgS1/8Eo4eP0pnTSfvK47RqvXIyEiHJyobR549zGntBifi7N69a62563oiIAIiIAIiIAIiIAIiIAIiIAIiIAIisLqOmWWqMtbLYlPTFleyLpiR0RH85V/+lYsVnThxggLLNUSwbHeJz43yelju67m+2hR0BbxZWZkoo/Di5Qz29AxXmPh5+/Zt2LZ9K0t/5xAd42MkKUj3jJ8lvpN4881zKC4qdu6aPXt2c3471cWUkhKTME1BZmZ2Gv39/Sz9ncPevXtd+bBdP5aOmK7uLvbRvMrX5uFf/e7nsa52HeYDc4wxtaGnuxef//zv6ldBBERABERABERABERABERABERABERABNacwKocMytluct0xSwzyhRyUaEwc00pKakopPNlS10dY0qv4tlnn3HT1CXFxa6HJoJCjok6kRGRdM/EcFEpBZUsALaOmfPnzuHIkSOorKxEYlIS8vkzW33q7+3hBHYhi4HHKeKwHyY6GlVV1Yw5FbK/JgL33Hs3zjK+ND4+gQVOYptYVMP+mIryCjppJhmtWnRLTSmMTNWziPiNc2+4ouDWFi5F0UWTmJjgiobXnLwuKAIiIAIiIAIiIAIiIAIiIAIiIAIi8KEnsCphZoVaBDtagq6018vumDCdMx10oIyyrDcnN2/F0cJI04kTz6O58RoKCgtcx0uIa0uJCQnX57Uj4Y/zIz8vHw319XjyyV86B43Fmg4cuJU9MblO0Eli0e+WzZtWumLS0hlHyuTrEijChFFTXeNEmke//33XFVPCXplAIMCoVBSFnTnXeVNRXokvfflLaObCU3Mz3T0jjE8xgmWRqIMH1zlB6UP/myAAIiACIiACIiACIiACIiACIiACIiACa05gVcLMjShTO10nZ8+e5dLRyvLRBJeWnvrVrxg7usDi3WIXQUqhqDIw2IeBgUEM9PXj/IXz2LdvHzIprtg8doDlvFYgvEz3zcT4pBN2Bji5HQwGnDsmjq6WQ4du58+XXGlvRma6K/UtryhnbGnWuV0iWfY7yBnuZDp2drM0+OrVq/DFRHOSezNCXGjyREVw6akSe3fvxde/8XX8lOtPO3fuXHHUUFzysWxYhwiIgAiIgAiIgAiIgAiIgAiIgAiIgAisNYFVCTMRNoPEHaXu/l48e/gwdm7fjii6Zq61XsNiYBFTLOM9ffqMc9T8h//33zuB5fLFS67st5edLmdOn0aQ09XNLS04xa8//uADLorEZW3XXWOOlynrjZmacj0yX/jiFyi6pFH0edKJN15vNCoqy1zkqb29HZ1tbW6tyaJNURRpJsbG0M4Z7I898ABG2X3T3tGOvp5eLjwlcN4bGKKIk0YRx8SiI0eP4dMPfWatuet6IiACIiACIiACIiACIiACIiACIiACIrDK8l+CM3EmlgW9JtDEJcQjJiYGZynGUFlBaVkZJ7TnnQCSyHLeDZs24E/+5E+QlpqGouIiV8gbWFjE8OAQzr7+Op7+1dOMJFUjMysH8QmJbuI6ggLKMv8sWgFwrM/FmMZHx9ld8wouXr7EhaZM57qZYPfMJKevk+nMCVLQGRgcQFx8PM+TAB97bOa40FRf34DnWEjc2dVJQacCHZ0dXH4KUqRZxjBXoaZnpvWrIAIiIAIiIAIiIAIiIAIiIAIiIAIiIAJrTmB1jhlnmFl2U9Z/+G/+0Llcunq6OVM9x+6XOLeclJaajqTkRFe+29nZ5Yp5FyjWpKWlMVrkQXZ2NrtosvAKhZZLl66gr68XOxgvuu22A26G+8UXX2J8iToPbTTh8JLrpAkvhTDBQt+pqWn09fayK2YEDVfr6dRpQW9vH8+Z5USZWE52Z6RnOLHorrvupHOHcSZPJPr7+hhpqkACC3+f4cR3WVkpbrllJ6K9ijKt+W+eLigCIiACIiACIiACIiACIiACIiACIrBKx8z1uWwPy32379gBD8WTuIQ41yGzyBUkc9Mk0kVTW1vLONIM3jx/ni6YBSwEFjhrPYNElu5GRUVha91WRpByXC9Ma2srDSwrAkxGRoZzytjctQkqjz76KG4/eDtSGWeK5LVsFcp6acJ0w/QNDLhS33Oc0y7IL2ARcIFZbVBdWeXcO3W8RlV1FUWaWAwODbioUygUYsRpFAdvvx1f/drDFHHS9asgAiIgAiIgAiIgAiIgAiIgAiIgAiIgAmtOYJWOGbPMAE2NTbh0+TLqtmxhbMiHODpVgqFoFvIyfuT3I4HLSTPTM4wNLSBMwSbA+NAge10CSUEMUxiZZo9MFiNJWTnZTmxpudaCY0ePwu+P52rSBFJSU3iVCPzosR87x8xXH/4aXTkzuHDhAot7pxBPd46JOQEWBU+xeHjYO0jhxkMnzTCKS4oxxTJhE4pCoTBjVH7kbNwMH2NRlSwCHmcPTjrdO7k5jE8x+qRDBERABERABERABERABERABERABERABNaawKqEGXOkUC/B5NQkF5jOo7S0BAlJiRRYcpyAMjg86Ap/U9JTMTw87Fwsd919NyNLl9zP5+ZnMcRul9dZ/Pvl3/s9ZHGhabB/gAJL0JX+NjQ22ukZO6pCeXmpm9c2V05JYTF27dpDEWYGP/qnx5zA4vV63bWs22aZHyYKTc9Ouzns2cCci1CZG+coBZ+amhoXvzLB55Of+ASXm3yI0SLTWv/O6XoiIAIiIAIiIAIiIAIiIAIiIAIiIALXCaxOmFkxzNCJEqLwMc/Ol0kkJiYiNzcXmzZtYi9MBHq6etwCkvXJpMal4EEuL/3ghz9E/ZUGbN60Ec+/8CJG6Jqx6e1UOlcW6XppabnmXp/AczVRnInyRqGqqhqf+MQn0djUhF8f/TXXm8a4ABWFzvZO2MJTXd0WijdljC9xJaq720WkPBEeeCIisRxedo6aSbpjjh8/xo/juPuuexid2sES4lIXebLSYqcC6VdCBERABERABERABERABERABERABERABNaYwKqEGWogZlBx7pcRriIdP/4ccuiWieGM9bat21BsDhp2zPT29qOwsND1uXR19zCi5Ac4if2R+z7Ggt8lTmqfRhunrq0XJrAYwvjYOIWZjdjO+e1vfetbFH6CblY7NjbWdda89NLLjEXNIyU5xZX2DtJ1s8QbSeCS04HbbsOlixfdCtPmLXUsGU5xEScTfqw8eIqRKhNoGpsbkZaexphUOpoHmp2QY9ezzhsdIiACIiACIiACIiACIiACIiACIiACIrCWBFalRpgoY3aTJYor8zOzbsI6ipEic81s3VaHopIihIJhBNkpk5qahInxSPz4R4+x0yUdOVnZGOjvQ6zP72axX6BzprfXRJs4FBYXUvCJdKtJtpQUWgxjnF0zc1x7sn4aK/y18t9oOmmKi2tREax0nTVJjFHZ2lJbexvjTMv43d/9LJZDS+jq6OIEdzodPLzbpbBz5yzxHEG6c8LsnRkcHHQrT5V8bZoKgNfy907XEgEREAEREAEREAEREAEREAEREAERIIFVCTMr5CKcSBLFqFIwuIhJFvmO0z0zPDLKWNKi65Zp7Wjj46mYm52jADKNpKRkTmRn4ydPPMF1pAjGlKpQVFSEkeFR9tN4sfuW3Thy5DAnt+MZY/KwL8YKgxfofOGiE50yJqqkp2dSYFlyhb62CuWlSJOYkIz7738A69atcy6chYUAUnmtgZEhZEVmIiY6hnGrWQQCAVdEPMf41dJyyDl+TKSx+9chAiIgAiIgAiIgAiIgAiIgAiIgAiIgAmtNYHXCDONBy3TNLPFzJMWRBQonFWVlyOTCUv9AvyvwnZ6ZRndnJ7tmupwAshgKIC7O75wyc7OzTnTZvm07Pvu5z7qOmtdOnoKVAo9Q2BltakZ2VhZamq+5jphDBw8hSJEmxIUli0h5o7zo7e+1Whl3reGRQdc9U8TYVCTntU+fOu0KgzfUrnf1MRF02YTCQSfuXL16FQOc2B5hKXF2dvYK7+udOWsNX9cTAREQAREQAREQAREQAREQAREQARH4cBPwPMLjX4yAWSb7E0VRpq2tFb19/a74d/+t+/H0r57Giy+84KJBXjpVhhgXmpyYpCsl6BwtqalpruR3nk4Y647Zt3cfatbVMO40gWss/921azc8Hi/m6JAZ4rrTOB//+Cc+jsuc5X7p5ZdRXFiEeIozM3ToWMRpaGgQzZzZPnXqFCNRvU4EmuRrMinsbNy4ie4bP5hucgLOEuNMk5OTFH9GcPjIEa4yxSCTi1DrN6ynkJP4L8agF4iACIiACIiACIiACIiACIiACIiACIjA2yGwOsfM9SvW1FTjK1/5KrIyczC/MEc3jM/95OVXX0F5RTm+8IUv4t/9P/+OQouHbpoMOl2i3XNszWlmahreyCgEWfBrfS9BumEWKNasX7+B89ulePLJJ9kxY6tPcy4WNUuXTWtLC/rYT2Plwnffe48r+7U57snxTpx89VVOYv/avb5uax2fU+wKigOMNSWyg+bOO+/AgVsPsBlnGceOHcPjP3mc4g6dO2E6cXgdHSIgAiIgAiIgAiIgAiIgAiIgAiIgAiKw1gRWJczY0lGEawCOYLlvCvIKcuDzxdKBEou8nDzGmfrYthvhynqnGGny+WJw330fxQRntYeGBmAiS4DdLnPzM04UWaQoM0oXSz+dN+aCiea6k59umpSUZATj/fjVU09hw4ZNuOfue2Dz1ubU2X3LLmygiJOalkp3Tiq7bCbRzYnuVrpnCosK6IyZwgDdOuakWVqi8ML78UZHYz3dOdZd8zLdN9ExURyJsse9a81d1xMBERABERABERABERABERABERABERCB1ZX/rogyy640d55rScthzmdT+PCw3yUmNgYZjAeFWdTb0d6O9bW1GGbcyBaWPOx66eGEdltrG6ZnZyjqpKKTPTSpKWkuYtTD3piRkWHkF+Qz9lTrOmwsdhSgkJLMdac//7/+3E1zj42OOedMTIyPZcL5qNu8FbExfjz9zNNOdElMSGJR8BJmZ2Zw9coVij7DLBkuRbTPy26ZEYo0y3TRJCHCQ13KBCYrzNEhAiIgAiIgAiIgAiIgAiIgAiIgAiIgAmtMYFWOmZV7jGC/TBu6ujoxzb4XHwUZZ6IxNw3/mKvGnDHrN25EOyNIzx5+BuXl5W5i2xsdhYmecYxTYPmbv/kbbNu2zQkzJqbYuSL5p3ZDLcxJEx8fzxnuBNdHE0vnjY9xqYmJcVy+cpURqBD6+nr5mhlThly58DwXl9bRFWOdNi10z/j5+rHRUbfsNEVXzSkWA6/fsA5f/tKXcPzYccxQvFniypMOERABERABERABERABERABERABERABEVhrAqsWZtiny2LfIVy4cNGV7qbQ/RLHmetMLh0NDA0xcuR1Qk1URCQ8UVHo7u5GNCNDmZlZOHDgNjdh3dXV7Zw1FiuKpJvG1paOHT2KMc5ub9u6FX6/H+ksC87NK2C0yc8emwBqaqrccpPX68XC3DwG+gcwyOuNjY1h7569GGTnjElDFl/q6GzHZgpDdZs3c9EphObmJifQzM7M4c477nDCjK002Xl0iIAIiIAIiIAIiIAIiIAIiIAIiIAIiMBaE1idMHPdFRMdE814kofOlhAFlhBMnNm5fTuSExPpcknktHUiZtkZY+6ZMN0w/OSen5Obi1r2w8RSeLnvvvvQ1dON10+f4Wz2FC5fveJWlq5cvITajRsQx+d4otgEQztOMBB0S0tLjEkFg4sUXjooDF1w13BOmfXrsT8lFWffOEtxBk7MKS+vhM1oJ/PxQn7u7xtAmA6ZIAt/7b4mJiYQ5lqTDhEQAREQAREQAREQAREQAREQAREQARFYawKrEmaWKZJYK4v1yZRwISkvPw+RfGyJ4kx8QgLnr9chh86ZBIozTU1NLnK0EFhw7hgTQey5kZERzjVTU12DHdt3YGJ0HOfOnccnP/EJtLd3uBjTFGe2TZhZ4OJTLyNL1lFTxrUnE33sCNEVs7S0hAm6ZWwqu7OjnbPXCRRdFt1Md5g/S0tPQ29PDxKSEijQFCErO8sVDF+lAGQrUDbZbefQIQIiIAIiIAIiIAIiIAIiIAIiIAIiIAJrTWBVwgzrXNjpAszPzLuC3uLiYkyx52WIkSLnRLF3QRHFHDW2sDQ/v4AJxpNGR0bfEkIsWhRBcSbSCoNZymvPZbUMvvbVr7nXvn7mDK6wR2aOnTGzdh3MYpoT22UVZcjOyoE/zu+6aMytM0mnTUdnF06cOIHpmVmKOXFITU/nHPYUbC67saERR48dRXVVNe65527GqTKdWyZkQhI7aOw+dIiACIiACIiACIiACIiACIiACIiACIjAWhNYlTBjN2lFvcOjwzj/xjnUs6elvKycoouPsSA2vDAiNMKuFxNqrMNldn6Oa0lBN2mNLvC/LvbNMAbF2ev+/n7XQWNOGhppXLwpKyvTLSp1dHSgoaEBr5181c1tp2ZkUGgJuVLhRUaZRkZHMDI07Jw4y3S9DPPrxsYm3LZ/P4uGy5xzxqaxFynAzHC2u6uniz020YxRrUcsS4StSNjuIYI9ODpEQAREQAREQAREQAREQAREQAREQAREYK0JrE6YoYBiLpMwJ7CHhgZw/s0L2LNnN3bu2Imy8goEAvOMEg1jdGwUL5x4HuOT4zTYRKyU9rLg15wqsWEfnTKxeOGF51FcXOKcLa7glw6ZUMhEmpVlp3nGjRobG/HA/Q9i37697hzzCyuumFm6YyYmx5wwU7Ou2nXIRFrnTXgRSYlJyMjMcG4Z7mEjnsXE9jO7pxDXnny+JFiBsEWq/Iwz6RABERABERABERABERABERABERABERCBtSawOmGGd2nCSXVNNRq5dNTd2+/KeGmWQVlZKZKTk/GjH/8Yra1tLqq0wCiTuVK6WNYbooNlw4b1FEZi6aIJMeI0gfOj5zhrHYeioiJcunSJHTTWHxNGFNecrFzY1qwj6GxJSk5Bd1cPOrva2UPTxZ97MMbyXh8Xm+L8sTzvJoo0IYo9L7jnJqckU5Lhi3lffgozUYxN1TfUI5F9MzXV69w9x9C5o0MEREAEREAEREAEREAEREAEREAEREAE3g0CqxJmrlfMYF3NOuTn5eOOgwfR1d3l5rItHmQxJRNgGhopgrCo1wQWW2eampxwbpg4FvQW8HXR0T6KMwHGmQY4dz3q5rHNHRNeXEJOXjbycvMwbt00jCzFcn47zD6ZZf6ZXwjSqTPI802hn6KQOWn237qfzxtzMadpds6MjY6yn2YO42MTLCLOwSbOZndSGJrka5pbWnjtGPd1fIKfPTkq/303fvl0TREQAREQAREQAREQAREQAREQARH4sBNYlTDzVlUuv0hISsT2Hdv5sQMB9r7ERHvNoMIYUToy0tJdzMiWjxISE1gCPAef14coumAsChUX50Nqap6bs3788cfR09uDzZs3YSE4TzdLEDarnc4S370UXqzEd2hw0C0oec1JQwcNB7hdf0ws3TI7dtyCl15+mb0yQdxx190oKi5CHJ00FluyJaY777kLrRRkWltbXTmwiTGhxQW6c3gv1mSsQwREQAREQAREQAREQAREQAREQAREQATWmMCqhJnfvMcVSYN/U40xUca+MKHjnrvuRUF+IaNDDWijGDI6MsKPMTpV2OtiokqIO0tzs+yjWXSlwbau1Nfbh9raDXjjjdfx+uuv444770BaSiqyuMLU0nYNl65cRlZmFjJYAmyxJ8/1fhgvl5/sPthK46a4t1DcMTdPaloq+20mKepQgKEpJj+vgKW/sSguKXGi0WOPPebWmX7z/ehrERABERABERABERABERABERABERABEVgrAm9bmHnrRm+YTpb5Bf/zcaFp27Zt7mOUsaLOrk688vIraGtrRxJdNlHRUXTFhNDe3cFVJo/rpTF3TE9PN6yT5tVXT7oumuSkZK45TXDlaZRTUEuIyIpwc9fZ2dno7enFFAuA09LTsEhnTIAuGw5C0U0T5a5vU9hFhYWY5ZS3FQJbXKqpqXnl9VxmsucF5gMSZ976R9QXIiACIiACIiACIiACIiACIiACIiACa0ngnRNmbtz1DYHmxvf8nJaahrS0NNRt2eImtAe52NTT3Y2+/j5MTkxiYmLMRZ2Kiorxve99jwXCZW5haYltwjajPUfxpb7+CqamZlx0am5uHgWFBcjOycEtO3ciNSWF55ig46YXMTHRruNmMRhgVCnkSorNoQMKRuPjY3j66adcl82e3XuQy9fbfPayqTk6REAEREAEREAEREAEREAEREAEREAERGCNCbzzwsxvewPXxRoLOWUyimQfGzbUcup6kr0vbbh8+RKFmh70D/Tj3Lk3kJKchG1bt7rC4IarTXS2RNLtMoap6WkW9k7w+Zdx5MizKCgogC8m1sWbFjjR3dHRwc6aVIpAqbjW0srn0ynDiJS5aayTJkyhx8p/Z2dn6OpZxs5bdrKY2Msoleayf9s/mx4TAREQAREQAREQAREQAREQAREQARG4uQTWRpi5/h7+ZzNNBCymtHVrHer4MTI8jBaW85pIYpGkTJYHz9AZ09PTgRbObtuHzV13VFU7F0x7ezsGBgadOONlb03OUo4TWazs18tVqKtX63H27Fns2rULI+y3CQQCmKEoYwKNy1pRJLJy4HguRNmKlA4REAEREAEREAEREAEREAEREAEREAERWGsCayrM/K/enAk2GRlcceLHju07MMAp7AnOZJtQ09wcjY7OLvbOzGGajpkzZ87g0MFDeOihz6KxqYET2mHOZ89janyShb5pdMlM4fTp0xgc6EMovIQFCjLtbR04f+EcI0+pyM3N4YT2KLtu2tzPIyMjJcz8r/5h9LgIiIAIiIAIiIAIiIAIiIAIiIAIiMBNJeB5hMdNvcL/9skpz7AHJpKumESW9mZmZqK6uhoV5eXIz8+10SfnlPFzGtsWl+772H1cWcp3s9pdnd1uijuK60wzFG962TVjbhkml7Bv715+342GhkaE2TlTyBntltYWnH/zvJvRznILT8USZ/63/530RBEQAREQAREQAREQAREQAREQAREQgXeKwHvCMfPWmzFthmJKRIQNX68EjjKyMnHnXXfj1gO3o72DkSZ2x3ijouDn6lMKHTIWRRpjVOnipR783r/+167Qt77+qltcmqfLJoEij5+RJQ+dMf44P0WfRATponnx+RewnstMqekZ1+NNb92FvhABERABERABERABERABERABERABERCBNSHw3hJm+Japybjj+idb3uax7NaWqiurUc2OmRuiTZQnGoUsAM7Jy0X/4ADSklOQzWLh8YlxbNy0yfXW1Nc3YHpmBhZZ8sfFobKqEtU11bh44RICCwGKPB4nBLnL6C8REAEREAEREAEREAEREAEREAEREAERWEMC7zlh5re/9xtqzYok476jtSY+IR579+xFUVERzr7xBrKzcxBk+W9iYgIy6ISZZt/MD3/4A0REelzUKYIyj9/vR0lpCWpra/Hoo9+n0yYLFZWVFH5ifvul9agIiIAIiIAIiIAIiIAIiIAIiIAIiIAI3CQC7xNh5sa7v+6foShjcSd38FNBfoFbZ7LvFxYWEEkBxia4g4sBtHGOe35hDqHQEj8WkZefx5LgDPcYG20wOjqOxeAiELdyOv0tAiIgAiIgAiIgAiIgAiIgAiIgAiIgAmtF4H0mzKxgeUuU4bfLLthkfTRUaCjY+Hw+bKnbgtr1tejp7sH62vV47bXX0N/XxyluPztnEpCSmorouRgUlxQjnV9HRF4XedaKuq4jAiIgAiIgAiIgAiIgAiIgAiIgAiIgAiTwvhRmfvNfzgky13WVG90zJtZER0ejpKwUpaWluP/++9Hd3e3WmhITk/gzL90z0ZzPTkE6V5miPO97DL+JRF+LgAiIgAiIgAiIgAiIgAiIgAiIgAi8Twh8oBSJG04aJ9bwH8DpNfwrKSmJH4luhWmlOXjZLTHl5OYiISFR/TLvk19W3aYIiIAIiIAIiIAIiIAIiIAIiIAIfNAIRCzz+KC9qX/+ftxbZCeNCTU3XDX2nOBikNPZQUacYjWZ/c+h6XsREAEREAEREAEREAEREAEREAEREIGbTuBDIcz8Voq/qdD81ifoQREQAREQAREQAREQAREQAREQAREQARG4uQQib+7p38Nnj/jAG4Xew/B1ayIgAiIgAiIgAiIgAiIgAiIgAiIgAkbgwyvMrDTQ6LdABERABERABERABERABERABERABERABN41Ah9iYeZdY64Li4AIiIAIiIAIiIAIiIAIiIAIiIAIiIAjIGFGvwgiIAIiIAIiIAIiIAIiIAIiIAIiIAIi8C4R+P8BFWP6JbSAJlUAAAAASUVORK5CYII=)", "_____no_output_____" ] ], [ [ "from tensorflow.keras.utils import plot_model\n\ndef build_generator(img_shape, z_dim, verbose=False):\n '''\n Genera una imagen de 28x28x1 a partir de un vector aleatorio de 100 entradas (espacio latente)\n '''\n # Batch de datos\n z = Input(shape=(z_dim,))\n # Pasa entrada unidimensional de dimensión 20 en un tensor de (7)(7)(256) tensor via un red Densa\n # luego la reformatea en un tensor de 7x7x128\n X = Dense(256 * 8 * 8, input_dim=z_dim) (z)\n X = Reshape((8, 8, 256))(X)\n\n # Convolución transpuesta, tensor de 7x7x256 a 14x14x128, con normalización por lote y activación ReLU\n X = Conv2DTranspose(filters = 128, \n kernel_size = 3, \n strides = 2, \n padding = 'same')(X)\n X = BatchNormalization()(X)\n X = LeakyReLU(alpha = 0.01)(X)\n \n # Convolución transpuesta, tensor de 14x14x128, a 14x14x64 con normalización por lote y activación ReLU\n X = Conv2DTranspose(filters = 64, \n kernel_size = 3, \n strides = 1, \n padding = 'same')(X)\n X = BatchNormalization()(X)\n X = LeakyReLU(alpha=0.01)(X)\n \n # Convolución transpuesta, tensor de 14x14x128 a 28x28x1, con activación tahn\n Y = Conv2DTranspose(filters = 3, \n kernel_size = 3, \n strides = 2, \n padding = 'same',\n activation = 'tanh')(X)\n\n \n\n generator_model = Model(inputs = z, outputs = [Y], name ='generator')\n \n return generator_model", "_____no_output_____" ], [ "generator = build_generator(img_shape, z_dim)\ngenerator.summary()", "Model: \"generator\"\n_________________________________________________________________\nLayer (type) Output Shape Param # \n=================================================================\ninput_73 (InputLayer) [(None, 100)] 0 \n_________________________________________________________________\ndense_47 (Dense) (None, 16384) 1654784 \n_________________________________________________________________\nreshape_22 (Reshape) (None, 8, 8, 256) 0 \n_________________________________________________________________\nconv2d_transpose_66 (Conv2DT (None, 16, 16, 128) 295040 \n_________________________________________________________________\nbatch_normalization_140 (Bat (None, 16, 16, 128) 512 \n_________________________________________________________________\nleaky_re_lu_188 (LeakyReLU) (None, 16, 16, 128) 0 \n_________________________________________________________________\nconv2d_transpose_67 (Conv2DT (None, 16, 16, 64) 73792 \n_________________________________________________________________\nbatch_normalization_141 (Bat (None, 16, 16, 64) 256 \n_________________________________________________________________\nleaky_re_lu_189 (LeakyReLU) (None, 16, 16, 64) 0 \n_________________________________________________________________\nconv2d_transpose_68 (Conv2DT (None, 32, 32, 3) 1731 \n=================================================================\nTotal params: 2,026,115\nTrainable params: 2,025,731\nNon-trainable params: 384\n_________________________________________________________________\n" ] ], [ [ "## Agregando etiquetas\n", "_____no_output_____" ], [ "Necesitamos que nuestro generador tenga también información de las etiquetas generadas de manera aleatoria, para ello, pedimos que este modelo reciba dos entradas\n1. El lote de imágenes falsas.\n2. Las etiquetas generadas.\n\nLuego vamos a combinar esta información en una capa que codifique dichas etiquetas por categorías para luego pasarlas al resto de las capas. El modelo se ilustra a continuación", "_____no_output_____" ], [ "![generador.jpg](data:image/jpeg;base64,/9j/4AAQSkZJRgABAQEASABIAAD/4SgWRXhpZgAATU0AKgAAAAgABgALAAIAAAAmAAAIYgESAAMAAAABAAEAAAExAAIAAAAmAAAIiAEyAAIAAAAUAAAIrodpAAQAAAABAAAIwuocAAcAAAgMAAAAVgAAEUYc6gAAAAgAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAFdpbmRvd3MgUGhvdG8gRWRpdG9yIDEwLjAuMTAwMTEuMTYzODQAV2luZG93cyBQaG90byBFZGl0b3IgMTAuMC4xMDAxMS4xNjM4NAAyMDIxOjA1OjIzIDIzOjI4OjA5AAAGkAMAAgAAABQAABEckAQAAgAAABQAABEwkpEAAgAAAAMyNQAAkpIAAgAAAAMyNQAAoAEAAwAAAAEAAQAA6hwABwAACAwAAAkQAAAAABzqAAAACAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAMjAyMTowNToyMyAyMjozODowOQAyMDIxOjA1OjIzIDIyOjM4OjA5AAAAAAYBAwADAAAAAQAGAAABGgAFAAAAAQAAEZQBGwAFAAAAAQAAEZwBKAADAAAAAQACAAACAQAEAAAAAQAAEaQCAgAEAAAAAQAAFmoAAAAAAAAAYAAAAAEAAABgAAAAAf/Y/9sAQwAIBgYHBgUIBwcHCQkICgwUDQwLCwwZEhMPFB0aHx4dGhwcICQuJyAiLCMcHCg3KSwwMTQ0NB8nOT04MjwuMzQy/9sAQwEJCQkMCwwYDQ0YMiEcITIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIy/8AAEQgBAADUAwEhAAIRAQMRAf/EAB8AAAEFAQEBAQEBAAAAAAAAAAABAgMEBQYHCAkKC//EALUQAAIBAwMCBAMFBQQEAAABfQECAwAEEQUSITFBBhNRYQcicRQygZGhCCNCscEVUtHwJDNicoIJChYXGBkaJSYnKCkqNDU2Nzg5OkNERUZHSElKU1RVVldYWVpjZGVmZ2hpanN0dXZ3eHl6g4SFhoeIiYqSk5SVlpeYmZqio6Slpqeoqaqys7S1tre4ubrCw8TFxsfIycrS09TV1tfY2drh4uPk5ebn6Onq8fLz9PX29/j5+v/EAB8BAAMBAQEBAQEBAQEAAAAAAAABAgMEBQYHCAkKC//EALURAAIBAgQEAwQHBQQEAAECdwABAgMRBAUhMQYSQVEHYXETIjKBCBRCkaGxwQkjM1LwFWJy0QoWJDThJfEXGBkaJicoKSo1Njc4OTpDREVGR0hJSlNUVVZXWFlaY2RlZmdoaWpzdHV2d3h5eoKDhIWGh4iJipKTlJWWl5iZmqKjpKWmp6ipqrKztLW2t7i5usLDxMXGx8jJytLT1NXW19jZ2uLj5OXm5+jp6vLz9PX29/j5+v/aAAwDAQACEQMRAD8A9/ooAKKACigAooAKKACigAooAKKACigAooAKKACop7iG1haa4mjhiX7zyMFA/E0WuBkyeKtM+ZbRpr9x0FpE0isfTf8AcH4sKvaVqSarYC6SGWH948ZjlxuVkcqc4JHVT0NW6birslSTdkXaKgoKKACigAooAKKACigAooAKKACigAooAKinuIbWB5riaOGJBlnkYKqj3JotcDJfxVpZyLRpr89jZwtIhPpvHyA/VqgfWdXn4ttMhtl/vXc4Lf8AfKZB/wC+hWqpfzEOfYgeHVLr/j71iYKesdpGIV/P5m/8epsei6dHMJmtlmnHPnTkyv8A99Nk1pdLSOhNr7l8ccCl8K/8gaT/AK/bv/0okqJfCNbm3RWJoFFABRQAUUAFFABRQAUUAFFABUU9xDbRNLcTRxRqMs8jBQPxNFrgZL+KtLPFo8183/TpE0i/99j5R+JqB9Z1i44ttMhtl/v3c25h/wAATI/8erVUrfEQ59iB4dUus/a9YmCnrHaRrCv58uP++qbHounJMJ2tlmuByJ7gmWQf8CfJ/WtLpaRJ33L9FSAUUDCl8K/8gaT/AK/bv/0okpS+Ea3NuisSwooAKKACigAooAKKACopriG2jMk8scUY6tIwUD8TQlcDJfxVph4tHmvm/wCnSIyL/wB9/d/WoH1nWLjIttMhtlPR7ubcw/4AmQf++q2VK3xaEOfYgeHVLrP2zWJlRuGis4xCv/fXLj6hhTY9E02OUTNarNOOk1wTNJ/32+T+tWml8OhO+5foqRhRQAUUAFFABS+Ff+QNJ/1+3f8A6USUpfCNbm3RWJYUyWaKCIyzSJHGvVnYAD8TQBXTVNPkdUS/tWZjhVWZSSfQc1bptNbiTTCikMKinuILWIy3E0cMY6vIwUD8TQk2Bkv4q0w5Fo818w6C0iaRT/wP7n/j1QPrOr3GRa6ZDbKRxJeTZYe/lpkH/vsVqqX8xDn2IHg1S65vNZmUHgx2cawoR9TucH6MKamiacknmtarNKOktwTM4+jOSf1rS6WkdCd9y/RUjCigAooAKKACigAooAKXwr/yBpP+v27/APSiSlL4Rrc26KxLCuW+Igz4Gv8AP96H/wBHJV0vjXqTP4WeS6IijxFpJCgH7fb9v+mi19B125j8aMMNszP1PWLfSzEkiTzTzbvKhhjLM+MZ9h1HJIFZr6zrFx/x66ZDbIeA95Plx77EyCP+BiuONNNXbN3LoiB4dUuf+PvWZgP7lnGsK/mdzfk1Nj0XTo5fNNsssw6Szkyv/wB9OSa1ul8JG+5foqRhRQAUUAFFABRQAUUAFFABRQAUvhX/AJA0n/X7d/8ApRJSl8I1ubdFYlhXL/ET/kRr/wD34f8A0clXS+NepM/hZ5Nov/IxaT/1/wBv/wCjVr6BruzH40YYbZnO65/yMWkf9cbj/wBp1NXKvhRq9wooAKKACigAooAKKACigAooAKKACigApfCv/IGk/wCv27/9KJKUvhGtzborEsK5f4if8iNf/wC/D/6OSrpfGvUmfws8m0X/AJGLSf8Ar/t//Rq19A13Zj8aMMNszndc/wCRi0j/AK43H/tOpq5F8KNXuFFMAooAKKACigAooAKKACigAooAKKACl8K/8gaT/r9u/wD0okpS+Ea3NuisSwrl/iJ/yI1//vw/+jkq6Xxr1Jn8LPJtF/5GLSf+v+3/APRq19A13Zj8aMMNszndc/5GLSP+uNx/7TqauRfCjV7hRTAKKACigAooAKKACigAooAKKACigApfCv8AyBpP+v27/wDSiSlL4Rrc26KxLCuX+In/ACI1/wD78P8A6OSrpfGvUmfws8m0X/kYtJ/6/wC3/wDRq19A13Zj8aMMNszndc/5GLSP+uNx/wC06mrkXwo1e4UUwCigAooAKKACigAooAKKACigAooAKXwr/wAgaT/r9u//AEokpS+Ea3NuisSwrl/iJ/yI1/8A78P/AKOSrpfGvUmfws8m0X/kYtJ/6/7f/wBGrX0DXdmPxoww2zOd1z/kYtI/643H/tOpq5F8KNXuFFMAooAKKACigAooAKKACigAooAKKACl8K/8gaT/AK/bv/0okpS+Ea3NuisSwrl/iJ/yI1//AL8P/o5Kul8a9SZ/CzybRf8AkYtJ/wCv+3/9GrX0DXdmPxoww2zOd1z/AJGLSP8Arjcf+06mrkXwo1e4UUwCigAooAKKACigAooAKKACigAooAKXwr/yBpP+v27/APSiSlL4Rrc26KxLCuX+In/IjX/+/D/6OSrpfGvUmfws8m0X/kYtJ/6/7f8A9GrX0DXdmPxoww2zOd1z/kYtI/643H/tOpq5V8KNXuFFABRQAUUAFFABRQAUUAFFABRQAUUAFL4V/wCQNJ/1+3f/AKUSUpfCNbm3RWJYVy/xE/5Ea/8A9+H/ANHJV0vjXqTP4WeQafLJBqthNFCZpI7qF0iBwXIkUhfx6V6v/wAJXr//AEKF1/3/AP8A7CvVxlKE5LmlY5KM5RWiuczrfibWm8SWczaTJbSpCVS0kJffuPLDAB5wB/wGuysZ5rm0SW4tmtpGHMTMCR+IrnrUoQhFxdzSE5Sk7osUVzGoUUAFcDD491W6uNNmt9FtTpupX72VvI96RKCpbLMgQ4GEbjPpnGaTdgSK9/8AFWGzmvLhbayfTLK7+yyk36i6fDBWdIccqDnuCQM1oar401e31XXLPTPD6XkWjxpNPM935YZWjD4UbSS3X2498UuYdiK++IyedZQabBYmSewjv3Oo362qqkgyqA4O5yM+w455oPxBub+y8OyaHo32ubW452jjln8sQtEVDbjg5UZbnvgetFwsO8QeN9U8NL9q1HS9Pjs4vLEqf2iDcNu2hmjj28hST1IJxnim+JPFOpzt4g03SNFW8t9Otit5O9z5Z3PGWxGu07iFOTkj0ouFiv4f8RXtr4b8LaHo+nR3upT6RHct50/lRwxBVXcxwScscAAVI/xGu3trWK20EyarJqMumTWhuQBHMi7sh9vzL0OcDjNFwsdvYvdSWML30McN0UHmxxvvVW7gNgZH4VYqhBRQAUvhX/kDSf8AX7d/+lElKXwjW5t0ViWFcv8AET/kRr//AH4f/RyVdL416kz+Fnk2i/8AIxaT/wBf9v8A+jVr6BruzH416GGG+FnN63Gg8T6TIEUSNBOpbHJA2YGfxP51YrlXwo1e4UUAFFACEkAkDJA6eteFeEbe60nVNOu7aCK41me78u5srjRpBPbRs53sbjgZCk/NjnOOaljR3kPge+068vF06bRXsrq5e4H23TjLNBvOSqsHAYZzjPT3rZHhpxfeJrj7SuNZjjRV2f6rbF5fPPPrTsFzDHgG7spNPudPuNMluIdNhsLiO/sjLHL5YwJFwwKnqPpitqLw3cf2poN/NcWofTI7lJEt7fykcy7cbVydoG31OaLCuczr/wAM7zWH1uKK/wBNWHVLgXBuJ7EyXUf3f3aybhhPl446Eite/wDCGqnVtam0rV4LW01pUF2ktsZHjIXYWjO4AFl/vA4PNKw7kcPgzUtLGiXekanbR6hp+nDTpvtFuXiuIxg9AwKkMMjBosPAUlpdafeS6is13Fqc2pXcnlbRM8iFcKMnaAMdz0p2C521FMQUUAFL4V/5A0n/AF+3f/pRJSl8I1ubdFYlhXL/ABE/5Ea//wB+H/0clXS+NepM/hZ5Nov/ACMWk/8AX/b/APo1a+ga7sx+NGGG2Zzuuf8AIxaR/wBcbj/2nU1ci+FGr3CimAUUAFFABRQAUUAFFABRQAUUAFFABS+Ff+QNJ/1+3f8A6USUpfCNbm3RWJYVy/xE/wCRGv8A/fh/9HJV0vjXqTP4WeTaL/yMWk/9f9v/AOjVr6BruzH40YYbZnO65/yMWkf9cbj/ANp1NXIvhRq9wopgFFABRQAUUAFFABRQAUUAFFABRQAUvhX/AJA0n/X7d/8ApRJSl8I1ubdFYlhXL/ET/kRr/wD34f8A0clXS+NepM/hZ5Nov/IxaT/1/wBv/wCjVr6BruzH40YYbZnO65/yMWkf9cbj/wBp1NXIvhRq9wopgFFABRQAUUAFFABRQAUUAFFABRQAUvhX/kDSf9ft3/6USUpfCNbm3RWJYVy/xE/5Ea//AN+H/wBHJV0vjXqTP4WeTaL/AMjFpP8A1/2//o1a+ga7sx+NGGG2Zzuuf8jFpH/XG4/9p1NXIvhRq9wopgFFABRQAUUAFFABRQAUUAFFABRQAUvhX/kDSf8AX7d/+lElKXwjW5t0ViWFcv8AET/kRr//AH4f/RyVdL416kz+Fnk2i/8AIxaT/wBf9v8A+jVr6BruzH40YYbZnO65/wAjFpH/AFxuP/adTVyL4UavcKKYBRQAUUAFFABRQAUUAFFABRQAUUAFL4V/5A0n/X7d/wDpRJSl8I1ubdFYlhXLfEUgeBdQJOAGh5/7bJV0vjXqTP4WeSaHNG3iPSQsiEm+t8AMP+ei19CV25g05oww2zOd1z/kYtI/643H/tOpq5V8KNXuFFMAooAKKACigAooAKKACigAooAKKACl8K/8gaT/AK/bv/0okpS+Ea3NuisSwooAKKAOd1z/AJGLSP8Arjcf+06mrZfCjN7hRTAKKACigAooAKKACigAooAKKACigApfCv8AyBpP+v27/wDSiSlL4Rrc26KxLCigAooA53XP+Ri0j/rjcf8AtOpq2Xwoze4UUwCigAooAKKACigAooAKKACigAooAKXwr/yBpP8Ar9u//SiSlL4Rrc26KxLCub8eXE9p4MvpraeWCZWiAkicowzKgOCORwSKumrzSfcmXws8v0bWdWbX9LR9X1F0e9gVle7kYMDIoIILYIIr3SuvHwjCaUVYxw8m07nM63cRf8JVpUHmDzVgnYr7Nsx/6C35VarBJqKuaN6sKKACigAooAKKACigAooAKKACigAooAKXwr/yBpP+v27/APSiSlL4Rrc26KxLCuX+In/IjX/+/D/6OSrpfGvUmfws8h06E3Gr6fAJHiMl3CgkjOGTMijIPYivXP8AhDZv+hl1v/wJNepjKyhJJxT9Tko0+Zb2OY17wfct4is7c6lczRzQswuLl/McFTyo/wC+gfxNdbZWps7RIDPNOV/5aTNuY/jXPVrqpCKSsaQpuMmWKK5zUKKACigAooAKKACigAooAKKACigApfCv/IGk/wCv27/9KJKUvhGtzborEsK5f4if8iNf/wC/D/6OSrpfGvUmfws8m0X/AJGLSf8Ar/t//Rq19A13Zj8aMMNszmPElxDZ61pVxcOIoFjnVpG4VSdmMnoOh61PDcQ3MYkgljlQ9GRgw/MVype6mavckooAKKACigAooAKKACigAooAKKACigApfCv/ACBpP+v27/8ASiSlL4Rrc26KxLCuX+In/IjX/wDvw/8Ao5Kul8a9SZ/CzybRf+Ri0n/r/t//AEatfQNd2Y/GjDDbMMVl3PhzR7p/MewiSU9ZYcxP/wB9Jg/rXBGTjsdDSe5Tbw7cw82OsXCc8JdIs6Af+Ov+bVA8WvWpPm2NteRg43Ws21299j4A/wC+zWinF76EOLWxC2s28HF9HcWLd/tUJVB/wPlPyarsM8NzGJIJUljPRkYMPzFW11FckoqRhRQAUUAFFABRQAUUAFFABS+Ff+QNJ/1+3f8A6USUpfCNbm3RWJYVy/xE/wCRGv8A/fh/9HJV0vjXqTP4WeTaL/yMWk/9f9v/AOjVr6BruzH40YYbZhRXnHSFFACYFZlz4c0e6kMr2ESTHrLDmJ/++lwf1qoycdhNJ7lRvDtzFzZaxcp/sXSLOv8ARv8Ax6oHi121/wBZYW94mcZtZtrn32vgD/vs1opxe+hDi1sQNrNtAcXsdzYkHDG6hZEB/wCun3D+DVdguIbmIS280csbdHjYMD+Iq3HqK5JRUjCigAooAKKACigApfCv/IGk/wCv27/9KJKUvhGtzborEsK5f4if8iNf/wC/D/6OSrpfGvUmfws8m0X/AJGLSf8Ar/t//Rq19A13Zj8a9DDDbMKK846QooAKKACigBMVmXPhzSLqVpnsY0nbrNDmKT/vpcH9aqMnHYTSe5Ufw7cw/wDHjrFwg/uXSLOv58N/49UDQ69bf6yxt7xfW1m2N/3y+B/49WqnGW+hDi1sQPrVtb/8f0dxY84LXULImfTf9w/gauwzw3MSywSpLG3R0YMD+IqnHqK5JRUjCigAooAKXwr/AMgaT/r9u/8A0okpS+Ea3NuisSwqpqWm2mr2EljfRebbyY3IGK5wQRyCCOQKabTugauY1v4D8N2t1Dcw2DrLDIskbG5lOGU5BwWweRXSVU6kpu8ncmMVHYKKgoKKACigAooAKKACigBMVl3PhvR7qVpnsIkuGOTPBmKQn/fXDfrVRk47CaT3Kr+HbmHmx1i4Qdo7pBOg/Hh//HqrtFrtt/rLG3vF/vWs2xj/AMBfA/8AHq0VSL30IcWtiFtatoP+P6O4se2bqFkX/vv7v61dhnhuIxJBKksZ6MjBgfxFW49RXJKKkYUvhX/kDSf9ft3/AOlElKXwjW5t0ViWFFABRQAUUAFFABRQAUUAFFABRQAUUAFFACYrLufDej3UhlawijmY5M0GYZD/AMDTDfrVRk47CaT3KreHbmDJsdYuFH8MV0izoPx4c/i1QNFrtr/rLG3vEH8VrNsc/wDAH4H/AH3WinF76EOLWxA2tW0H/H7HcWJ9bqFkX/vv7v61oeEXSTQmkjZXRry6ZWU5BBnkwRTmrQuEXqbtFYGgUUAFFABRQAUUAFFABRQAUUAFFABRQAUUAFFABRQAUyOKOFNkSKi5JwowOeTQA+igD//Z/+EyxWh0dHA6Ly9ucy5hZG9iZS5jb20veGFwLzEuMC8APD94cGFja2V0IGJlZ2luPSfvu78nIGlkPSdXNU0wTXBDZWhpSHpyZVN6TlRjemtjOWQnPz4NCjx4OnhtcG1ldGEgeG1sbnM6eD0iYWRvYmU6bnM6bWV0YS8iPjxyZGY6UkRGIHhtbG5zOnJkZj0iaHR0cDovL3d3dy53My5vcmcvMTk5OS8wMi8yMi1yZGYtc3ludGF4LW5zIyI+PHJkZjpEZXNjcmlwdGlvbiByZGY6YWJvdXQ9InV1aWQ6ZmFmNWJkZDUtYmEzZC0xMWRhLWFkMzEtZDMzZDc1MTgyZjFiIiB4bWxuczpNaWNyb3NvZnRQaG90bz0iaHR0cDovL25zLm1pY3Jvc29mdC5jb20vcGhvdG8vMS4wLyI+PE1pY3Jvc29mdFBob3RvOkl0ZW1TdWJUeXBlPkx1bWlhLkxpdmluZ0ltYWdlPC9NaWNyb3NvZnRQaG90bzpJdGVtU3ViVHlwZT48L3JkZjpEZXNjcmlwdGlvbj48cmRmOkRlc2NyaXB0aW9uIHJkZjphYm91dD0idXVpZDpmYWY1YmRkNS1iYTNkLTExZGEtYWQzMS1kMzNkNzUxODJmMWIiIHhtbG5zOnhtcD0iaHR0cDovL25zLmFkb2JlLmNvbS94YXAvMS4wLyI+PHhtcDpDcmVhdGVEYXRlPjIwMjEtMDUtMjNUMjI6Mzg6MDkuMjQ3PC94bXA6Q3JlYXRlRGF0ZT48eG1wOkNyZWF0b3JUb29sPldpbmRvd3MgUGhvdG8gRWRpdG9yIDEwLjAuMTAwMTEuMTYzODQ8L3htcDpDcmVhdG9yVG9vbD48L3JkZjpEZXNjcmlwdGlvbj48L3JkZjpSREY+PC94OnhtcG1ldGE+DQogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICA8P3hwYWNrZXQgZW5kPSd3Jz8+/9sAQwADAgIDAgIDAwMDBAMDBAUIBQUEBAUKBwcGCAwKDAwLCgsLDQ4SEA0OEQ4LCxAWEBETFBUVFQwPFxgWFBgSFBUU/9sAQwEDBAQFBAUJBQUJFA0LDRQUFBQUFBQUFBQUFBQUFBQUFBQUFBQUFBQUFBQUFBQUFBQUFBQUFBQUFBQUFBQUFBQU/8AAEQgB5gGTAwEiAAIRAQMRAf/EAB8AAAEFAQEBAQEBAAAAAAAAAAABAgMEBQYHCAkKC//EALUQAAIBAwMCBAMFBQQEAAABfQECAwAEEQUSITFBBhNRYQcicRQygZGhCCNCscEVUtHwJDNicoIJChYXGBkaJSYnKCkqNDU2Nzg5OkNERUZHSElKU1RVVldYWVpjZGVmZ2hpanN0dXZ3eHl6g4SFhoeIiYqSk5SVlpeYmZqio6Slpqeoqaqys7S1tre4ubrCw8TFxsfIycrS09TV1tfY2drh4uPk5ebn6Onq8fLz9PX29/j5+v/EAB8BAAMBAQEBAQEBAQEAAAAAAAABAgMEBQYHCAkKC//EALURAAIBAgQEAwQHBQQEAAECdwABAgMRBAUhMQYSQVEHYXETIjKBCBRCkaGxwQkjM1LwFWJy0QoWJDThJfEXGBkaJicoKSo1Njc4OTpDREVGR0hJSlNUVVZXWFlaY2RlZmdoaWpzdHV2d3h5eoKDhIWGh4iJipKTlJWWl5iZmqKjpKWmp6ipqrKztLW2t7i5usLDxMXGx8jJytLT1NXW19jZ2uLj5OXm5+jp6vLz9PX29/j5+v/aAAwDAQACEQMRAD8A/VOiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKSgBaKKSgA/OkzWP4k8ZaD4Psxda9ren6Lbk4E2oXSQKfYFiK8j1b9sj4ex3Uln4fOseN9QRgn2Xw5pslwckZ4dgqEY54Y9DW0KNSp8MbmbqRjuz3Sj24r5yuPjV8YfF2V8N/DjTfC9uzYS+8V6gZGK9ybeEBlPXqT2ry/wCP0Xxc0L4U674m1n4p3ST2phKab4ftRYwrvnjjI81SJGGGzyfauqGCnJqMpJGMsQkrpH2/RSDoKWvPOoKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAoopKAFoopKAD86TNY/iTxloPg+zF1r2t6fotuTgTahdJAp9gWIryPVv2yPh7HdSWfh86x431BGCfZfDmmyXByRnh2CoRjnhj0NbQo1KnwxbM3UjHdnutFfONx8avjD4tyvhv4cab4Xt2bCX3irUDIxXuTBDhlPXqT2rNuPh78TPGOT4s+LOpWtuzZaw8K26aeqj0EwBkI6fez3FdKwcvtySMnWX2Vc+hvEnjLQfB9p9q13WtP0a2PAm1C6SBT7AsRzXkurftkfD2K6ks/D51jxvqCME+y+HNNkuDkjPDsFQjHOQx6H6VzOh/s0/DzR7xr2fQl1zUWbfJe63K15JIcYywclSfT5ewr0mx0+10u1S2sraG0t0zthgjCKOc8AAAVqqNCO95fgQ6k3tocFcfGr4w+Lcr4b+HOm+F7dmwl94q1AyMV7kwQgMp+pPas24+HvxM8Y8+LfizqVrbs2TYeFbdNPVR6CYAyEdPvZ7ivVqK2Uow+CKX4kNOXxM8v0P9mn4eaPeNez6Euuaizb5L3W5WvJJDjGWDkqT6fL2Fek2On2ul2qW1lbQ2lumdsMEYRRzngAACp6KUqk5bsFFLZBXjv7Xn/Ju/iz/ALdP/SyGvYq8d/a8/wCTd/Fn/bp/6WQ1eH/ix9RVPhZ9WL90UtIv3RS14Z3rYKKKKBhRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRSUALRRSUAH50max/EnjLQfB9mLrXtb0/RbcnAm1C6SBT7AsRXkerftkfD2O6ks/D51jxvqCME+y+HNNkuDkjPDsFQjHPDHoa2hRqVPhi2ZupGO7PdaK+cbj41fGHxblfDfw403wvbs2EvvFWoGRivcmCHDKevUntWbcfD34meMcnxZ8WdStbdmy1h4Vt009VHoJgDIR0+9nuK6Vg5fbkkZOsvsq59DeJPGWg+D7T7Vrutafo1seBNqF0kCn2BYjmvJdW/bI+HsV1JZ+HzrHjfUEYJ9l8OabJcHJGeHYKhGOchj0P0rmdD/Zp+Hmj3jXs+hLrmos2+S91uVrySQ4xlg5Kk+ny9hXpNjp9rpdqltZW0NpbpnbDBGEUc54AAArVUaEd7y/Ah1JvbQ4K4+NXxh8W5Xw38OdN8L27NhL7xVqBkYr3JghAZT9Se1Ztx8PfiZ4x58W/FnUrW3Zsmw8K26aeqj0EwBkI6fez3FerUVspRh8EUvxIacviZ5fof7NPw80e8a9n0Jdc1Fm3yXutyteSSHGMsHJUn0+XsK9JsdPtdLtUtrK2htLdM7YYIwijnPAAAFT0UpVJy3YKKWyCiiisygooooAKKKKACiiigArx39rz/k3fxZ/26f+lkNexV47+15/ybv4s/7dP/SyGt8P/Fj6kVPhZ9WL90UtIv3RS14Z3rYKKKKBhRRRQAUUlLQAUUUlAC0UUUAFFFFABRRRQAUUUUAFFFFABRRSUALRRSUAH50max/EnjLQfB9mLrXtb0/RbcnAm1C6SBT7AsRXkerftkfD2O6ks/D51jxvqCME+y+HNNkuDkjPDsFQjHPDHoa2hRqVPhi2ZupGO7PdaK+cbj41fGHxblfDfw403wvbs2EvvFWoGRivcmCHDKevUntWbcfD34meMcnxZ8WdStbdmy1h4Vt009VHoJgDIR0+9nuK6Vg5fbkkZOsvsq59DeJPGWg+D7T7Vrutafo1seBNqF0kCn2BYjmvJdW/bI+HsV1JZ+HzrHjfUEYJ9l8OabJcHJGeHYKhGOchj0P0rmdD/Zp+Hmj3jXs+hLrmos2+S91uVrySQ4xlg5Kk+ny9hXpNjp9rpdqltZW0NpbpnbDBGEUc54AAArVUaEd7y/Ah1JvbQ4K4+NXxh8W5Xw38OdN8L27NhL7xVqBkYr3JghAZT9Se1Ztx8PfiZ4x58W/FnUrW3Zsmw8K26aeqj0EwBkI6fez3FerUVspRh8EUvxIacviZ5fof7NPw80e8a9n0Jdc1Fm3yXutyteSSHGMsHJUn0+XsK9JsdPtdLtUtrK2htLdM7YYIwijnPAAAFT0UpVJy3YKKWyCiiisygooooAKKKKACiiigAooooAKKKKACiiigAooooAK8d/a8/wCTd/Fn/bp/6WQ17FXjv7Xn/Ju/iz/t0/8ASyGt8P8AxY+pFT4WfVi/dFLSL90UteGd62CiiigYUUUUAJR+lJXyp+1F+1F41+DvxQsPDPhqw0G5tZ9Hj1F5NVhndw7TSxkAxyKMYjXt3PNdGHw9TFVFSp7syq1I0Y80tj6rzRX5+f8ADd3xa/6BPgz/AMBrv/4/XvH7JP7Qnir453njG28T2Wj2j6L9j8k6TFKgfzhMW3eZI+ceWuMY6nrXfiMqxWFpurVjZLzOenjKVaXLB6n0bRRRXkHaFFJmigBaKSigA/OkzWP4k8ZaD4Psxda9ren6Lbk4E2oXSQKfYFiK8j1b9sj4ex3Uln4fOseN9QRgn2Xw5pslwckZ4dgqEY54Y9DW0KNSp8MWzN1Ix3Z7rRXzjcfGr4w+Lcr4b+HGm+F7dmwl94q1AyMV7kwQ4ZT16k9qzbj4e/Ezxjk+LPizqVrbs2WsPCtumnqo9BMAZCOn3s9xXSsHL7ckjJ1l9lXPobxJ4y0Hwfafatd1rT9GtjwJtQukgU+wLEc15Lq37ZHw9iupLPw+dY8b6gjBPsvhzTZLg5Izw7BUIxzkMeh+lczof7NPw80e8a9n0Jdc1Fm3yXutyteSSHGMsHJUn0+XsK9JsdPtdLtUtrK2htLdM7YYIwijnPAAAFaqjQjveX4EOpN7aHBXHxq+MPi3K+G/hzpvhe3ZsJfeKtQMjFe5MEIDKfqT2rNuPh78TPGPPi34s6la27Nk2HhW3TT1UegmAMhHT72e4r1aitlKMPgil+JDTl8TPL9D/Zp+Hmj3jXs+hLrmos2+S91uVrySQ4xlg5Kk+ny9hXpNjp9rpdqltZW0NpbpnbDBGEUc54AAAqeilKpOW7BRS2QUUUVmUFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAV47+15/ybv4s/wC3T/0shr2KvHf2vP8Ak3fxZ/26f+lkNb4f+LH1IqfCz6sX7opaRfuilrwzvWwUUUUDCiiigAr8+f27f+Th9J/7FaH/ANK7iv0Gr8+f27f+Th9J/wCxWh/9K7iveyP/AH6HzPOzD+Azwevqn/gnX/yH/in/ANwr/wBBua+Vq+qf+Cdf/If+Kf8A3Cv/AEG5r7TP/wDcZeq/M8LL/wCOj7Vooor8tPrjI8R+KtF8HaadQ17VrHRbHcE+06hcJBHuPRdzEDJwePavINW/bI+Hsd1JZ+HzrHjfUEYJ9l8N6bJcHJGeHYKhGOchj0NZ37ZNpBf6J8L7W6gjubafx5pcUsMyhkkQrOGVlPBBGQQeDmu1sdPtdLtUtrK2htLdM7YYIwijnPAAAFehRo0/Zqc7u5yzqT5nGJwVx8avjD4tyvhv4c6b4Xt2bCX3irUDIxXuTBCAyn6k9qzbj4e/Ezxjz4t+LOpWtuzZNh4Vt009VHoJgDIR0+9nuK9WorqUow+CKX4mTTl8TPL9D/Zp+Hmj3jXs+hLrmos2+S91uVrySQ4xlg5Kk+ny9hXpNjp9rpdqltZW0NpbpnbDBGEUc54AAAqeilKpOW7BRS2QUUUVmUFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFeO/tef8m7+LP+3T/0shr2KvHf2vP+Td/Fn/bp/wClkNb4f+LH1IqfCz6sX7opaRfuilrwzvWwUUUUDCiiigAr8+f27f8Ak4fSf+xWh/8ASu4r9Bq/Pn9u3/k4fSf+xWh/9K7iveyP/fofM87MP4DPB6+qf+Cdf/If+Kf/AHCv/Qbmvlavqn/gnX/yH/in/wBwr/0G5r7TP/8AcZeq/M8LL/46PtWiiivy0+uPn/8AbA/5B/wo/wCygaV/Kau1riv2wP8AkH/Cj/soGlfymrta9Wn/AAY+rOKXxsKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAK8d/a8/wCTd/Fn/bp/6WQ17FXjv7Xn/Ju/iz/t0/8ASyGt8P8AxY+pFT4WfVi/dFLSL90UteGd62CiiigYUUUUAFfnz+3b/wAnD6T/ANitD/6V3FfoNX58/t2/8nD6T/2K0P8A6V3Fe9kf+/Q+Z52YfwGeD19U/wDBOv8A5D/xT/7hX/oNzXytX1T/AME6/wDkP/FP/uFf+g3NfaZ//uMvVfmeFl/8dH2rRRRX5afXHz/+2B/yD/hR/wBlA0r+U1drXFftgf8AIP8AhR/2UDSv5TV2terT/gx9WcUvjYUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAV47+15/wAm7+LP+3T/ANLIa9irx39rz/k3fxZ/26f+lkNb4f8Aix9SKnws+rF+6KWkX7opa8M71sFFFFAwooooAK/Pn9u3/k4fSf8AsVof/Su4r9Bq/Pn9u3/k4fSf+xWh/wDSu4r3sj/36HzPOzD+Azwevqn/AIJ1/wDIf+Kf/cK/9Bua+Vq+qf8AgnX/AMh/4p/9wr/0G5r7TP8A/cZeq/M8LL/46PtWiiivy0+uPn/9sD/kH/Cj/soGlfymrta4r9sD/kH/AAo/7KBpX8pq7WvVp/wY+rOKXxsKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAK8d/a8/5N38Wf9un/AKWQ17FXjv7Xn/Ju/iz/ALdP/SyGt8P/ABY+pFT4WfVi/dFLSL90UteGd62CiiigYUUUUAFfnz+3b/ycPpP/AGK0P/pXcV+g1fnz+3b/AMnD6T/2K0P/AKV3Fe9kf+/Q+Z52YfwGeD19U/8ABOv/AJD/AMU/+4V/6Dc18rV9U/8ABOv/AJD/AMU/+4V/6Dc19pn/APuMvVfmeFl/8dH2rRRRX5afXHz/APtgf8g/4Uf9lA0r+U1drXFftgf8g/4Uf9lA0r+U1drXq0/4MfVnFL42FFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFeO/tef8m7+LP+3T/wBLIa9irx39rz/k3fxZ/wBun/pZDW+H/ix9SKnws+rF+6KWkX7opa8M71sFFFFAwooooAK/Pn9u3/k4fSf+xWh/9K7iv0Gr8+f27f8Ak4fSf+xWh/8ASu4r3sj/AN+h8zzsw/gM8Hr6p/4J1/8AIf8Ain/3Cv8A0G5r5Wr6p/4J1/8AIf8Ain/3Cv8A0G5r7TP/APcZeq/M8LL/AOOj7Vooor8tPrj5/wD2wP8AkH/Cj/soGlfymrta4r9sD/kH/Cj/ALKBpX8pq7WvVp/wY+rOKXxsKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAK8d/a8/5N38Wf9un/pZDXsVeO/tef8m7+LP+3T/0shrfD/xY+pFT4WfVi/dFLSL90UteGd62CiiigYUUUUAFfnz+3b/ycPpP/YrQ/wDpXcV+g1fnz+3b/wAnD6T/ANitD/6V3Fe9kf8Av0PmedmH8Bng9fVP/BOv/kP/ABT/AO4V/wCg3NfK1fVP/BOv/kP/ABT/AO4V/wCg3NfaZ/8A7jL1X5nhZf8Ax0fatFFFflp9cfP/AO2B/wAg/wCFH/ZQNK/lNXa1xX7X/wDyD/hR/wBlA0r+U1drXq0/4MfVnFL42FFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFeO/tef8m7+LP8At0/9LIa9irx39rz/AJN38Wf9un/pZDW+H/ix9SKnws+rF+6KWkX7opa8M71sFFFFAwooooAK/Pn9u3/k4fSf+xWh/wDSu4r9Bq/Pn9u3/k4fSf8AsVof/Su4r3sj/wB+h8zzsw/gM8Hr6o/4J2nGvfFP/uFf+g3NfK9ex/sn/HLQvgnrXjt9ds9RuhqhsVh/s+ON9vlJLu3b3XH+tXGM9D07/d51RqV8JKnTjdtr8z5/AzjTrKUnZH6NfjR+NfN//Dd/gH/oD+I//Aa3/wDj9H/Dd3gH/oD+I/8AwGt//j9fnX9l43/n0/uPpfreH/nRoftfEf2f8KR/1UDSv5TV21fFXx2+NkPxG+JFh4h0D+0LaxsWtbuG01DCqLqBmKSlFdlOA2ATzgkd69V+Hv7WWlat5dp4ptf7Iujx9stwXt2+o+8n/jw75Fe08pxNLDxla/VrqjgWMpTqNXPoCiobO8g1C1hurSaO5tplDxyxsGV1PQgjgipq8c7gooopAFFFFABRRRQAUUUUAFFFfIniz/gpJ4PsPEGuaf4R8FeLfiFY6GzLqWs6DZB7OEAncwfJyg2n5iFU4yCRyU2luNJs+u6K+cv2a/23vDX7T3ia70jw74T8T6bFbW0lw+p6jaxC0DI0YMXmJI2JMSKQPTJ475PxJ/4KBeDvCPjzUvBvhbwv4n+JXiDSyy6hH4YsvOitmVtsiM+ckqRg4UrnjdnOFzK17j5XsfUdFeOfs6/tVeCv2lrDUm8Nm+07V9LcJqGi6tCIbu3ySqsQGZSpKkZBJGOQpwK838Z/8FHPhp4B8WeNvDer2GuDWfDV8mnR2dtbxyy6nMSynyBv+6u3kvt+8uMkgUcyDlex9V0V80/s/wD7eHgv47ePJvA76LrXg3xeqNJFpmuQqhnCqXZUIOQwQbirAZHIzg10n7Qn7YHgj9nXUtM0TVIdT8Q+K9TUPaeH9BtxPdOhJVXYFgFBYED+IkHCnBIOZWuFnse5UV81fBn9uzwh8U/iDD4C1jw94i+H3jKdN1vpviW08n7RwW2o2chtoJAdVBxhSTxW38eP2yvBH7Onj7Q/DHi631KMapYTagNQt40eGFY9+FI3BizMm0AKRll564OZWuFnex71RXyp8NP+Ch/gnx14a8Z+ItU8Pa/4T0Pw1Zw373WoQK/2mGaXyovLVCcszlQOo+b7w5Ncpcf8FQvDljDBq978LfHtj4PnkVY9euLFFiYMeCPm2HPs5PpRzx7hys+1qK5PT/ir4V1L4ZxfEGLWYB4Pk0/+0/7UkJCLb7SxZh1BAByuMgjGM8V8r2//AAVU+G8+qmT/AIRPxenhNbj7K3iY2KG3RyfvFQ5O3HzYzvx/BnihySDlbPo34N/tAeDvjxHr7eEbu4ul0O9+w3n2i2eHEnP3cjkcHn25r0evgr/glDfQano3xevLZ/Mt7jxCs0UmCNysrkHB5HBHXpXp3xU/4KHeBfAfjq/8G+HdB8Q/ETxHp7Ml3D4btRLFA6ttdC5OSyng7VYA8E5yBMZ3SbG462R9T0V4L+zp+2d4D/aR1LUNF0dNS0HxRYIZJ9D1qBYrjYpAZ0KsysAxAIyGB6rjBrhPiD/wUk+Hnw78VeM/DN7ouvXPiDw9qCabDY28UbNqMrFgxiw/CrtGSwB+ZcAk4quZW3Fys+taK84/Z/8AjVZ/H74b2vi6y0fUNBjluJrV7HUgoljeNtrdCePy+gr0eqRIUUUUAFFFFABRRRQAV47+15/ybz4sH/Xp/wClkNexV5F+1myr+z/4nL42B7MtnkY+2Q1vh/4sfUip8LPqdfuilpF+6KWvDO9bBRRRQMKKKKACvz5/bt/5OH0n/sVof/Su4r9Bq/Pn9u3/AJOH0n/sVof/AEruK97I/wDfofM87MP4DPB695/Yv+EvhX4qa58Q18T6X/aS2H9nm3H2iWHy/MSbf/q3XOfLTrnGOO9eDV9Uf8E7f+Q98U/+4V/6Dc19xnlSVPBSlB2d1t6ngYGMZV0pI90/4ZH+FH/QrH/wY3f/AMdo/wCGR/hR/wBCsf8AwY3f/wAdr2KivzX65if+fkvvZ9T7Cl/Kj8/v2jPgXB4N+I2i2vh7R5NI8M6rd2Wkw3Ukryx/bJ2fjLsWOFUk44GMcEjPrnw9/Zr8LeCvKub2L+3tTXnz7xB5at6rH0H1OT710v7X/wDyD/hT/wBlA0r+U1drXuf2liKuGhDm8vX1Z5/1anCq3YFAUADgDoPpRRRXmavVnUFFFFABRRRQAUUUUAFFFFAGb4m8r/hHdV+0Xf2CD7JL5l1nHkLsOX/Ac/hX5d/sP/tZQfs7fBvVfDt78N/FHiX7ZqdxeaTq2g6eXt9RYqsexnfaRhosblDnBwVBXB/R/wCOHiLTfCXwZ8cazrFi+qaVZ6Ldy3VjG203EflNujB7bh8ue2c9q/Pn9i/9mP4reOPgZZeJ/Cfx31HwBo9/dXLWegWMLXsUQWRkYufOQRszKSVCngqTycDGd7qxpHY9n/YV+F/jX4N/s5/ETxFr2kS6BrWvT3esWGiupjkt1WA+XlDzGxYH5TztC5x28A/4J/618dtD+FmuX/wv8B+FPEtjf6zJ9u1bWr4xXbzJFH+7b96uVUPuGR1kY9697/Yx+N3xI1T49fEb4M/ELxBZ+PI/Dlq88WvwRIOUlijaJioAbcJejAsrRsCT2o2P7H/xs/Zx8aeIb74A+MdAXwhrM5uX8OeI0fEDEnCphGBCj5Q+5CVwCDgGotdJorvctfs5/Ar4y6f+2B4g+LXjnw1oXhSx1rSpLa9tdGvVkjkkxCFIUOxyzRB2JPXPc1zf7KXhPR9a/wCChHx/1a/0+C81HSbqV7CeZNxtmkmKuydgxX5d2MgEgYya+h/2fPBfx60zxZqmu/GDxxourWU9n9nsvD2g2+2C3kLq3mlzGhLAKy4y2d3UYrG+A37Nfib4X/tL/F/4hare6XPovi2Xfp8NpNI1wg80ufNVowq4Bxwzc1SjtZE33PHv2qLGHSv+Cif7Pmp2sSw3t4iQTyooBkUSyKM46/LIw+mBXmFnrHxPuv8Ago58WNV8B+F9F8V+KdNt3hht/ENz5aWlqot4hJEd64YrtXjtI3qa+r/jp+zZ4m+Jn7T3wj+Iul3ulwaJ4Tbdfw3U0i3D/vN/7tVQq2fdh+Vc5+0B+x/4y1b4123xj+DHiyx8JeOjEsGoW2pI32S9ULt3MVR+SiopUoQdqsCpGSnF9O41JaHlfxR+Df7S/wAfPil8MvEniTwL4T8NTeE9UjuRqGl6mDJ5XnRSEPmRiyr5ZIAGfmb1qf8AbR8I6T46/b0+Amha5ZpqGk3lsouLWT7kqi4lfaw7qSoyO4yO9er+BPA/7WWveOfD97478eeE9E8Lafdxz32naBbGSe/jRgWiJaIbVYZBIcYznBwK2fjN+zX4m+In7V/wp+Jmm3ulw6B4ViKX0FzNILlzvkYeWojKtneOrDoabjdBzanpf7QXxQ8J/BH4Q6v4l8WWCX+h2SxxrpiwJJ9pkLqIYlRht+8FOTwoXPavj/4wfHj49fGj9nDxfqZ+B+m6H8O9S0Oe4a+1LWEe5S18suJ0iJRiVUB1ynJAIzX1d+1R8A4v2kvg1qvgs6gNKvJJI7uyvHTekU8bZXcvUqQWU45AbPOMH5gf9l39qvxt8Lh8MfE3xI8L6V4PtLD7CsumiV7q+jjj2wwyP5SHyiQqschiuch+hc+a+go2sebeIdavrP8A4I8eGVgeTF3etaTsrY/cjVrhgD7ZRBivuf8AZ98B+Gf+GXfAvhuPTbS98O33hu0+0WskavFdCaBXlZ16Euzsx9ya4T4b/sjOn7FNn8DvHN3avd+RdJNe6UzSxwyvey3MMkZdUJKFkyCBnDDOOa8d8H/sqftU+HfBq/C+P4saBpvw8QG3XUrRJH1GK1ZjuiiJiDKcE8eYMDgPjikk1rboPRmJ/wAE45pPC3wn/aEl0aMrLpt9O1lGjE/NHBN5YBPPYDPWu0/4JK6DpcfwD1/xBGqTa9qOvzRX143zSlY44iiMx5wPMZvrITXof7Ev7Kes/sw6P490fWrzT9U07V9UEun/AGeRpXNqqsi+cGjQBypGQuR1ry2x/Y3+N/7OvjTxDe/AHxvocPhLWZ/tD6D4hVv9HbJ2hf3bq2wHAcFWIwCDjNCi1byB2d0Znx20628L/wDBUb4P32gxLb6nq2nxvqawfKZQftULO+OpMS457Rio/wBlXwXo2v8A/BQ749a1qNhHeahot1NJp8soz9neSba8ijs20EZ6gM3rXq/7On7H/ifwt8XtQ+L/AMX/ABXa+MviJPG0FqtihFpYoyhSyEonzbMoAFVVDN94tkbnwG/Zr8TfC/8AaX+L/wAQtVvdLn0XxbLvsIbSaRrhB5pf96rRhVwDjhm5oUXe4N6WPpSiiitzIKKKKACiiigAooooAK8d/a8/5N38Wf8Abp/6WQ17FXjv7Xn/ACbv4s/7dP8A0shrfD/xY+pFT4WfVi/dFLSL90UteGd62CiiigYUUUUAFfnz+3b/AMnD6T/2K0P/AKV3FfoNX58/t2/8nD6T/wBitD/6V3Fe9kf+/Q+Z52YfwGeD19U/8E6/+Q/8U/8AuFf+g3NfK1fVP/BOv/kP/FP/ALhX/oNzX2mf/wC4y9V+Z4WX/wAdH2rRRRX5afXHz/8Atgf8g/4Uf9lA0r+U1drXFftgf8g/4Uf9lA0r+U1drXq0/wCDH1ZxS+NhRRRQAUUUUAFFFFABRRRQAUUUUAVNX0my1/SrzTNStYr7T72F7e5tZ0DRzROpV0ZTwQQSCD1zXyZc/wDBMH4WRajdTaJ4h8ceFrG5bdLpmj6yq25Hp+8idyP95ia+vqKTimNNrY8u+A/7NXgL9nHRbvT/AAVpTWsl6yteX91KZrm5Kg7d7nsMnCqFXknGSa9RoooSSQrt7hRRRTAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACvHf2vP+Td/Fn/bp/wClkNexV47+15/ybv4s/wC3T/0shrfD/wAWPqRU+Fn1Yv3RS0i/dFLXhnetgooooGFFFFABX58/t2/8nD6T/wBitD/6V3FfoNX58/t2/wDJw+k/9itD/wCldxXvZH/v0PmedmH8Bng9fVP/AATr/wCQ/wDFP/uFf+g3NfK1fVP/AATr/wCQ/wDFP/uFf+g3NfaZ/wD7jL1X5nhZf/HR9q0UUV+Wn1x8/wD7YH/IP+FH/ZQNK/lNXa1xX7YH/IP+FH/ZQNK/lNXa16tP+DH1ZxS+NhRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABXjv7Xn/Ju/iz/t0/8ASyGvYq8d/a8/5N38Wf8Abp/6WQ1vh/4sfUip8LPqxfuilpF+6KWvDO9bBRRRQMKKKKACvz5/bt/5OH0n/sVof/Su4r9Bq/Pn9u3/AJOH0n/sVof/AEruK97I/wDfofM87MP4DPB6+qf+Cdf/ACH/AIp/9wr/ANBua+Vq+qf+Cdf/ACH/AIp/9wr/ANBua+0z/wD3GXqvzPCy/wDjo+1aKKK/LT64+f8A9sD/AJB/wo/7KBpX8pq7WuK/bA/5B/wo/wCygaV/Kau1r1af8GPqzil8bCiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACvHf2vP+Td/Fn/bp/6WQ17FXjv7Xn/Ju/iz/t0/9LIa3w/8WPqRU+Fn1Yv3RS0i/dFLXhnetgooooGFFFFABX58/t2/8nD6T/2K0P8A6V3FfoNX58/t2/8AJw+k/wDYrQ/+ldxXvZH/AL9D5nnZh/AZ4PX1T/wTr/5D/wAU/wDuFf8AoNzXytX1T/wTr/5D/wAU/wDuFf8AoNzX2mf/AO4y9V+Z4WX/AMdH2rRRRX5afXHz/wDtgf8AIP8AhR/2UDSv5TV2tcV+2B/yD/hR/wBlA0r+U1drXq0/4MfVnFL42FFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFeO/tef8m7+LP8At0/9LIa9irx39rz/AJN38Wf9un/pZDW+H/ix9SKnws+rF+6KWkX7opa8M71sFFFFAwooooAK/Pn9u3/k4fSf+xWh/wDSu4r9Bq/Pn9u3/k4fSf8AsVof/Su4r3sj/wB+h8zzsw/gM8Hr6p/4J1/8h/4p/wDcK/8AQbmvlavqn/gnX/yH/in/ANwr/wBBua+0z/8A3GXqvzPCy/8Ajo+1aKKK/LT64+f/ANsD/kH/AAo/7KBpX8pq7WuK/bA/5B/wo/7KBpX8pq7WvVp/wY+rOKXxsKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAK8d/a8/5N38Wf8Abp/6WQ17FXjv7Xn/ACbv4s/7dP8A0shrfD/xY+pFT4WfVi/dFLSL90UteGd62CiiigYUUUUAIK/Pr9u3/k4fSf8AsVof/Su4r9Ba+Ev23vA/irXPjlpep6N4S17XrBfDsVu1xpOmy3MayC5nYqWUEA4IOOuGHrXt5NONPGRlJ2R5+Oi5UGkfN9fVP/BOv/kP/FP/ALhX/oNzXzh/wrzx9/0Tbxp/4Irj/wCJr6k/YD8I+I/DWp/Ee417w5rHh+O8/s37ONWsZLYy7FuA23eBuxlc46bh619fneIo1MFKMJpu62fmeNgKc4102j7Aooor81PqT5//AGwP+Qf8KP8AsoGlfymrta4r9sD/AJB/wo/7KBpX8pq7WvVp/wAGPqzil8bCiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACvHf2vP8Ak3fxZ/26f+lkNexV47+15/ybv4s/7dP/AEshrfD/AMWPqRU+Fn1Yv3RS0i/dFLXhnetgooooGFFFFABRRRQAUUUUAFFFFAHz/wDtgf8AIP8AhR/2UDSv5TV2tcV+2B/yD/hR/wBlA0r+U1drXq0/4MfVnFL42FFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFeO/tef8m7+LP8At0/9LIa9irx39rz/AJN38Wf9un/pZDW+H/ix9SKnws+rF+6KWkX7opa8M71sFFFFAwooooAKKK5rxL8TPB/gy+jsvEHivRNCvJIxMlvqWow28jRkkBwrsCVyrDPTIPpTScnZITaWrOlorhP+F9fDL/oonhT/AMHdt/8AF1u+F/H3hnxv9p/4RzxFpOv/AGXb5/8AZd9Fc+Vuzt3bGO3O1sZ67T6VThOKu0JST2ZvUUUVBR8//tgf8g/4Uf8AZQNK/lNXa1xX7YH/ACD/AIU/9lA0r+U1drXq0/4MfVnFL42FFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFeO/tef8AJu/iz/t0/wDSyGvYq8d/a8/5N38Wf9un/pZDW+H/AIsfUip8LPqxfuilpF+6KWvDO9bBRRRQMKKKKAEXpX57ft6W8Vz+0LpKzRJMv/CLwnDqG/5e7j1r9Cq/Pn9u3/k4fSf+xWh/9K7iveyPXHQuedmH8Bnz5/ZNl/z52/8A36X/AAr6w/4JzwpBrXxSSJFjQf2XhVGAPlue31r5ar3n9i/4seFPhXrnxDfxPqn9mC//ALPFsfs8su/y0n3/AOrVsY8xeuOvfmvt88pyqYKUYRu9NvU8HAzUaylJ6H6CUV49/wANcfCj/oaT/wCC66/+NUf8NcfCj/oaT/4Lrr/41X5t9TxP/PuX3M+o+sUv5kYP7X3/ACD/AIU/9lA0r+U1drXyT+0b8coPGXxG0W58PaxJrHhnSruy1aG1kjeOP7ZAz4OHUMAQxBxwc+oFeufD39pLwt408q2vZf7B1NsL5F4w8pieMLLwD9GwT2Br3P7NxFPDQk43T19Dz/rVKdWSuetUUAhgCDkHpRXmnUFFFFIAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAK8d/a8/5N38Wf8Abp/6WQ17FXjv7Xn/ACbv4s/7dP8A0shrfD/xY+pFT4WfVi/dFLSL90UteGd62CiiigYUUUUAFfnz+3b/AMnD6T/2K0P/AKV3FfoNX58/t2/8nD6T/wBitD/6V3Fe9kf+/Q+Z52YfwGeD17H+yf8AAzQfjZrXjtNcu9RtBpZsWh/s+SNN3mpLu3b0bp5a4xjqa8cr6p/4J2f8h74p/wDcK/8AQbmvu86rVKGDlUpuzTWvzPn8DCNSsoyV0ei/8MH+Av8AoMeI/wDwJt//AIxR/wAMH+Av+gx4j/8AAm3/APjFfR2KMV+d/wBqY3/n6z6b6nQ/kR+cHx2+CcHw5+JFh4e0Eahc2F69raQ3eobWU3U7MEjLqiqCQucdcAntXq3w9/ZM0rSPKu/FFz/a90OfscBKW6n3PDP/AOOj1Br0n9r3P9n/AAp7/wDFf6V/Kau3r2Xm2Jq4eMb26N9zz1g6carZDZ2cGnWkNrawx29tCoSOGJQqoo6AAcAD0qaiivH1erO7bQKKKKQBRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAV47+15/ybv4s/7dP/SyGvYq8d/a8/5N38Wf9un/AKWQ1vh/4sfUip8LPqxfuilpF+6KWvDO9bBRRRQMKKKKACvz5/bt/wCTh9J/7FaH/wBK7iv0Gr8+f27f+Th9J/7FaH/0ruK97I/9+h8zzsw/gM8Hr6p/4J1/8h/4p/8AcK/9Bua+Vq+qf+Cdf/If+Kf/AHCv/QbmvtM//wBxl6r8zwsv/jo+1aKKK/LT64+f/wBr7/kHfCn/ALKBpX8pq7Wt34kfCzwz8XNDh0fxVpp1PT4bhbuONZ5YSkyqyhw0bKeA7cZxz0rye4/ZQv8AQMv4G+J3ibw7824WmpMup2qgdFWOTBA6DljwK9KjUp+zUJOzRyThLmckju6K8yuNJ+PXg3/XaN4a+IFqG4bTbptPutvqwl+TJA/hJ5NZ037R1l4bYR+N/CXibwQ2/Z9o1HTnktif9mWPII69B05rp9nzaxafzMua3xaHr1Fcx4X+J/hLxoo/sPxHpupOTjyobhfMH/ACdw6+ldPWbi46NFXQUUUVIwooopgFFFFIAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAK8d/a8/wCTd/Fn/bp/6WQ17FXjv7Xn/Ju/iz/t0/8ASyGt8P8AxY+pFT4WfVi/dFLSL90UteGd62CiiigYUUUUAFfnz+3b/wAnD6T/ANitD/6V3FfoNX58/t2/8nD6T/2K0P8A6V3Fe9kf+/Q+Z52YfwGeD19U/wDBOv8A5D/xT/7hX/oNzXytX1T/AME6/wDkP/FP/uFf+g3NfaZ//uMvVfmeFl/8dH2rRRRX5afXBRRRQAVG0YkUqVBBGCD3FSUmKAPM/GH7Nvwz8dMz6t4M0tp2bebmzi+yzFvUyRFWPXPJPNcNcfsoX/h/5/A3xO8TeHvm3C01Jk1O1UDoqxyYIHQck8CvoWgV0RxNWGikZOlCW6Pmi40n49eDf9do3hr4gWobhtNum0+62+rCX5MkD+Enk1nTftG2XhthH438JeJvBDb9n2jUdOeS2J/2ZY8gjr0HQZr6nxTWjEilWUFSMEEcEV0LF3+OC/IydH+VniPhf4n+EvGij+w/Eem6k5OPKhuF8wf8AJ3Dr6V09J4w/Zt+Gfjpi+reDNLadm3m5tIvssxbnkyRFWPXPJNcLcfsoX/h/wCbwN8TvE3h75twtNSZdTtVA6KscmCB0HLHgVsqtCXVojkqR6XO7orzK40n49eDf9do3hr4gWobhtNum0+62+rCX5MkD+Enk1nTftHWXhthH438JeJvBDb9n2jUdOeS2J/2ZY8gjr0HTmtfZ83wNP5kc1vi0PXqK5jwv8T/AAl40Uf2H4j03UnJx5UNwvmD/gBO4dfSunrNxcdGiroKKKKkYUUUUwCiiikAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAV47+15/ybv4s/wC3T/0shr2KvHf2vP8Ak3fxZ/26f+lkNb4f+LH1IqfCz6sX7opaRfuilrwzvWwUUUUDCiiigAr8+f27f+Th9J/7FaH/ANK7iv0Fr8+v27f+Th9J/wCxWh/9K7iveyP/AH6HzPOzD+Azwevqn/gnX/yH/in/ANwr/wBBua+Vq+qf+Cdf/If+Kf8A3Cv/AEG5r7TP/wDcZeq/M8LL/wCOj7Vooor8tPrgooooAKKKKACiiigAooooAKKKKACo2jEilSoIIwQe4qSkxQB5n4w/Zt+GfjpmfVvBmltOzbzc2cX2WYt6mSIqx655J5rhrj9lC/8AD/z+Bvid4m8PfNuFpqTJqdqoHRVjkwQOg5J4FfQtArojiasNFIydKEt0fNFxpPx68G/67RvDXxAtQ3DabdNp91t9WEvyZIH8JPJrOm/aNsvDbCPxv4S8TeCG37PtGo6c8lsT/syx5BHXoOgzX1PimtGJFKsoKkYII4IroWLv8cF+Rk6P8rPEfC/xP8JeNFH9h+I9N1JyceVDcL5g/wCAE7h19K6ek8Yfs2/DPx0xfVvBmltOzbzc2kX2WYtzyZIirHrnkmuFuP2UL/w/83gb4neJvD3zbhaaky6naqB0VY5MEDoOWPArZVaEurRHJUj0ud3RXmVxpPx68G/67RvDXxAtQ3DabdNp91t9WEvyZIH8JPJrOm/aOsvDbCPxv4S8TeCG37PtGo6c8lsT/syx5BHXoOnNa+z5vgafzI5rfFoevUVzHhf4n+EvGij+w/Eem6k5OPKhuF8wf8AJ3Dr6V09ZuLjo0VdBRRRUjCiiimAUUUUgCiiigAooooAKKKKACiiigAooooAK8d/a8/5N38Wf9un/AKWQ17FXjv7Xn/Ju/iz/ALdP/SyGt8P/ABY+pFT4WfVi/dFLSL90UteGd62CiiigYUUUUAJ/Ovz6/bt/5OH0n/sVof8A0ruK/QWvnP8AaE/ZIufjl4/s/E9t4w/4R17fTE002/8AZgutwWWSTfuMq4z5mMY/h688etleIp4XFRq1HZI4sXTlVpOMdz4Or6p/4J2f8h74p/8AcK/9BuaT/h3Zq3/RU/8Ay3l/+SK9k/Zr/Zrm/Z9uPE80/ib/AISSXW/suW/s/wCy+V5IlA6SPuz5vtjb3zx9LmubYXF4WVKk9Xbp5nl4PB1aNVTktD3KiiivhD6EKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKjaMSKVKggjBB7ipKTFAHmfjD9m34Z+OmZ9W8GaW07NvNzZxfZZi3qZIirHrnknmuGuP2UL/AMP/AD+Bvid4m8PfNuFpqTJqdqoHRVjkwQOg5J4FfQtArojiasNFIydKEt0fNFxpPx68G/67RvDXxAtQ3DabdNp91t9WEvyZIH8JPJrOm/aNsvDbCPxv4S8TeCG37PtGo6c8lsT/ALMseQR16DoM19T4prRiRSrKCpGCCOCK6Fi7/HBfkZOj/KzxHwv8T/CXjRR/YfiPTdScnHlQ3C+YP+AE7h19K6ek8Yfs2/DPx0xfVvBmltOzbzc2kX2WYtzyZIirHrnkmuFuP2UL/wAP/N4G+J3ibw9824WmpMup2qgdFWOTBA6DljwK2VWhLq0RyVI9Lnd0V5lcaT8evBv+u0bw18QLUNw2m3TafdbfVhL8mSB/CTyazpv2jrLw2wj8b+EvE3ght+z7RqOnPJbE/wCzLHkEdeg6c1r7Pm+Bp/Mjmt8Wh69RXMeF/if4S8aKP7D8R6bqTk48qG4XzB/wAncOvpXT1m4uOjRV0FFFFSMKKKKYBRRRSAKKKKACvHf2vP8Ak3fxZ/26f+lkNexV47+15/ybv4s/7dP/AEshrfD/AMWPqRU+Fn1Yv3RS0i/dFLXhnetgooooGFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAVG0YkUqVBBGCD3FSUmKAPM/GH7Nvwz8dMz6t4M0tp2bebmzi+yzFvUyRFWPXPJPNcNcfsoX/AIf+fwN8TvE3h75twtNSZNTtVA6KscmCB0HJPAr6FoFdEcTVhopGTpQluj5ouNJ+PXg3/XaN4a+IFqG4bTbptPutvqwl+TJA/hJ5NZ037Rtl4bYR+N/CXibwQ2/Z9o1HTnktif8AZljyCOvQdBmvqfFNaMSKVZQVIwQRwRXQsXf44L8jJ0f5WeI+F/if4S8aKP7D8R6bqTk48qG4XzB/wAncOvpXT0njD9m34Z+OmL6t4M0tp2bebm0i+yzFueTJEVY9c8k1wtx+yhf+H/m8DfE7xN4e+bcLTUmXU7VQOirHJggdByx4FbKrQl1aI5Kkelzu6K8yuNJ+PXg3/XaN4a+IFqG4bTbptPutvqwl+TJA/hJ5NZ037R1l4bYR+N/CXibwQ2/Z9o1HTnktif8AZljyCOvQdOa19nzfA0/mRzW+LQ9eormPC/xP8JeNFH9h+I9N1JyceVDcL5g/4ATuHX0rp6zcXHdFXQV47+15/wAm7+LP+3T/ANLIa9irx39rz/k3fxZ/26f+lkNbYf8Aiw9Sanws+rF+6KWkX7opa8I71sFFFFAwooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAqNoxIpUqCCMEHuKkpMUAeZ+MP2bfhn46Zn1bwZpbTs283NnF9lmLepkiKseueSea4a4/ZQv/AA/8/gb4neJvD3zbhaakyanaqB0VY5MEDoOSeBX0LQK6I4mrDRSMnShLdHzRcaT8evBv+u0bwz8QLUNw2m3TafdbfVhL8mSB/CTya8l/aS+L1zrHwb8ReHtb8E+JvCerXRgWH+0bEm1lKXETttnUlT8qN7HHvX3hTWjEilWUMp4II4NdVPG8slKUFp8jKVC6smPHQUtFFeadQUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQB//Z)", "_____no_output_____" ] ], [ [ "def build_generator(img_shape, z_dim, verbose=False):\n '''\n Genera una imagen de 28x28x1 a partir de un vector aleatorio de 100 entradas (espacio latente)\n '''\n # Batch de datos\n z = Input(shape=(z_dim,))\n # batch de etiquetas\n image_class = Input(shape = (1,), dtype = np.float32)\n\n # 10 clases en MNIST\n cls = Flatten()(tf.keras.layers.Embedding(10, z_dim)(image_class))\n # hadamard product between z-space and a class conditional embedding\n h = tf.keras.layers.Multiply()([z, cls])\n\n # Pasa entrada unidimensional de dimensión 20 en un tensor de (7)(7)(256) tensor via un red Densa\n # luego la reformatea en un tensor de 7x7x128\n X = Dense(256 * 7 * 7, input_dim=z_dim) (h)\n X = Reshape((7, 7, 256))(X)\n\n # Convolución transpuesta, tensor de 7x7x256 a 14x14x128, con normalización por lote y activación ReLU\n X = Conv2DTranspose(filters = 128, \n kernel_size = 3, \n strides = 2, \n padding = 'same')(X)\n X = BatchNormalization()(X)\n X = LeakyReLU(alpha = 0.01)(X)\n \n # Convolución transpuesta, tensor de 14x14x128, a 14x14x64 con normalización por lote y activación ReLU\n X = Conv2DTranspose(filters = 64, \n kernel_size = 3, \n strides = 1, \n padding = 'same')(X)\n X = BatchNormalization()(X)\n X = LeakyReLU(alpha=0.01)(X)\n \n # Convolución transpuesta, tensor de 14x14x128 a 28x28x1, con activación tahn\n Y = Conv2DTranspose(filters = 1, \n kernel_size = 3, \n strides = 2, \n padding = 'same',\n activation = 'tanh')(X)\n\n \n\n generator_model = Model(inputs = [z, image_class], outputs = [Y], name ='generator')\n \n return generator_model", "_____no_output_____" ] ], [ [ "Construye Generador", "_____no_output_____" ] ], [ [ "generator = build_generator(img_shape, z_dim)\ngenerator.summary()", "Model: \"generator\"\n__________________________________________________________________________________________________\nLayer (type) Output Shape Param # Connected to \n==================================================================================================\ninput_31 (InputLayer) [(None, 1)] 0 \n__________________________________________________________________________________________________\nembedding (Embedding) (None, 1, 100) 1000 input_31[0][0] \n__________________________________________________________________________________________________\ninput_30 (InputLayer) [(None, 100)] 0 \n__________________________________________________________________________________________________\nflatten_20 (Flatten) (None, 100) 0 embedding[0][0] \n__________________________________________________________________________________________________\nmultiply (Multiply) (None, 100) 0 input_30[0][0] \n flatten_20[0][0] \n__________________________________________________________________________________________________\ndense_19 (Dense) (None, 12544) 1266944 multiply[0][0] \n__________________________________________________________________________________________________\nreshape_9 (Reshape) (None, 7, 7, 256) 0 dense_19[0][0] \n__________________________________________________________________________________________________\nconv2d_transpose_27 (Conv2DTran (None, 14, 14, 128) 295040 reshape_9[0][0] \n__________________________________________________________________________________________________\nbatch_normalization_58 (BatchNo (None, 14, 14, 128) 512 conv2d_transpose_27[0][0] \n__________________________________________________________________________________________________\nleaky_re_lu_78 (LeakyReLU) (None, 14, 14, 128) 0 batch_normalization_58[0][0] \n__________________________________________________________________________________________________\nconv2d_transpose_28 (Conv2DTran (None, 14, 14, 64) 73792 leaky_re_lu_78[0][0] \n__________________________________________________________________________________________________\nbatch_normalization_59 (BatchNo (None, 14, 14, 64) 256 conv2d_transpose_28[0][0] \n__________________________________________________________________________________________________\nleaky_re_lu_79 (LeakyReLU) (None, 14, 14, 64) 0 batch_normalization_59[0][0] \n__________________________________________________________________________________________________\nconv2d_transpose_29 (Conv2DTran (None, 28, 28, 1) 577 leaky_re_lu_79[0][0] \n==================================================================================================\nTotal params: 1,638,121\nTrainable params: 1,637,737\nNon-trainable params: 384\n__________________________________________________________________________________________________\n" ] ], [ [ "# Discriminador", "_____no_output_____" ], [ "En discriminador es una red neuronal secuencial profunda que recibe de entrada un vector datos y produce una clasificación indicando si los datos son reales o falsos (fake).", "_____no_output_____" ], [ "![discriminador.png](data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAAAdgAAAGkCAYAAACSOTQjAAABfGlDQ1BJQ0MgUHJvZmlsZQAAKJFjYGAqSSwoyGFhYGDIzSspCnJ3UoiIjFJgv8PAzcDDIMRgxSCemFxc4BgQ4MOAE3y7xsAIoi/rgsxK8/x506a1fP4WNq+ZclYlOrj1gQF3SmpxMgMDIweQnZxSnJwLZOcA2TrJBUUlQPYMIFu3vKQAxD4BZIsUAR0IZN8BsdMh7A8gdhKYzcQCVhMS5AxkSwDZAkkQtgaInQ5hW4DYyRmJKUC2B8guiBvAgNPDRcHcwFLXkYC7SQa5OaUwO0ChxZOaFxoMcgcQyzB4MLgwKDCYMxgwWDLoMjiWpFaUgBQ65xdUFmWmZ5QoOAJDNlXBOT+3oLQktUhHwTMvWU9HwcjA0ACkDhRnEKM/B4FNZxQ7jxDLX8jAYKnMwMDcgxBLmsbAsH0PA4PEKYSYyjwGBn5rBoZt5woSixLhDmf8xkKIX5xmbARh8zgxMLDe+///sxoDA/skBoa/E////73o//+/i4H2A+PsQA4AJHdp4IxrEg8AAAAJcEhZcwAAFiUAABYlAUlSJPAAAAIEaVRYdFhNTDpjb20uYWRvYmUueG1wAAAAAAA8eDp4bXBtZXRhIHhtbG5zOng9ImFkb2JlOm5zOm1ldGEvIiB4OnhtcHRrPSJYTVAgQ29yZSA1LjQuMCI+CiAgIDxyZGY6UkRGIHhtbG5zOnJkZj0iaHR0cDovL3d3dy53My5vcmcvMTk5OS8wMi8yMi1yZGYtc3ludGF4LW5zIyI+CiAgICAgIDxyZGY6RGVzY3JpcHRpb24gcmRmOmFib3V0PSIiCiAgICAgICAgICAgIHhtbG5zOmV4aWY9Imh0dHA6Ly9ucy5hZG9iZS5jb20vZXhpZi8xLjAvIgogICAgICAgICAgICB4bWxuczp0aWZmPSJodHRwOi8vbnMuYWRvYmUuY29tL3RpZmYvMS4wLyI+CiAgICAgICAgIDxleGlmOlBpeGVsWURpbWVuc2lvbj40MjA8L2V4aWY6UGl4ZWxZRGltZW5zaW9uPgogICAgICAgICA8ZXhpZjpQaXhlbFhEaW1lbnNpb24+NDcyPC9leGlmOlBpeGVsWERpbWVuc2lvbj4KICAgICAgICAgPHRpZmY6T3JpZW50YXRpb24+MTwvdGlmZjpPcmllbnRhdGlvbj4KICAgICAgPC9yZGY6RGVzY3JpcHRpb24+CiAgIDwvcmRmOlJERj4KPC94OnhtcG1ldGE+Crc8k5IAAEAASURBVHgB7J0HfBVFu8ZfUkgl9BYIvXcQBJSiFAHBBiqK2FAUAbHr9WL57HSsVz+72LCCAqIoIh2U3ntNQkshvRLuPgN7sntyEtNO2/O8v184O7O7U/4T8pxp71Q4r5nQSIAESIAESIAEypWAX7mmxsRIgARIgARIgAQUAQosfxFIgARIgARIwAkEKLBOgMokSYAESIAESIACy98BEiABEiABEnACAQqsE6AySRIgARIgARKgwPJ3gARIgARIgAScQIAC6wSoTJIESIAESIAEKLD8HSABEiABEiABJxCgwDoBKpMkARIgARIgAQosfwdIgARIgARIwAkEKLBOgMokSYAESIAESIACy98BEiABEiABEnACAQqsE6AySccEYs5kOL7BWBIgARKwIAEKrAUb1VOr9MD0DbLvWIqnFo/lIgESIIFyJUCBLVecTKwoAgnJ2TJx1kbZsDuhqMd4jwRIgAQsQaACz4O1RDs6pRJthrxSqnR3LZ7s8L1e45aq+MAAP5l8VxsZ0LW2w+cYSQIkQAJWIMAerBVa0cvqkJObJy9+tFO+XXrcy0rO4pIACZBA8QlQYIvPik+WI4G88+flze/2ybvzDoh2SSMBEiAByxGgwFquSb2rQl/+dlRenbNLcs9RZb2r5VhaEiCBfyMQ8G8P8D4J6ASqNeyhX5o+E46uM4VLGli89oScTcmWl8a2l+Ag/5K+zudJgARIwCMJsAfrkc3ie4VauyNeJr2+SZJSc3yv8qwxCZCAJQlQYC3ZrN5RqcrhgaaC7jqcLNgrezIh0xTPAAmQAAl4IwEKrDe2mkXKPHJgA6lVNdhUm2On0mXc1A1yMCbVFM8ACZAACXgbAQqst7WYhcobGhwgN/aPkgZ1Qk21ikvKkgkzNsqWfYmmeAZIgARIwJsIUGC9qbUsWNaKgX5yXd/60rJhJVPtUjNy5dG3tsjyzadN8QyQAAmQgLcQoMB6S0tZuJz+fhVkcM9I6dyyqqmW2Tl58uwHO2Te8mhTPAMkQAIk4A0EKLDe0Eo+UMYKFUT6dqklvTrWNNU2L++8zPx6r3y04JApngESIAES8HQCFFhPbyEfK1/XNtXkqh51xU/r1Rrtk0WHZdoXuwWCSyMBEiABbyBAgfWGVvKxMrZpHCHX9q4nOBTAaD+vipXJ/90uWdrQMY0ESIAEPJ2A+S+Yp5eW5fMZAo0iw2REvygJsfPstHLrGXnkjc2Skp7rMyxYURIgAe8kQIH1znbziVLXqR4sNw9oIBFhZocU2w6cVdt4Tidm+QQHVpIESMA7CVBgvbPdfKbUVSMqChxS1KgSZKrzodhU5fXpyMk0UzwDJEACJOApBCiwntISLEehBMJCAuSm/g2kfi2zQ4pTmkvF8dM3yvaDSYW+yxskQAIk4C4CFFh3kWe+JSIQVNFPbriivjSLMjukSE7LUXOya7bHlSg9PkwCJEACziZAgXU2YaZfbgT8/SvI0MsjpUPzKqY0M7PPydPvbpNFa2JN8QyQAAmQgDsJUGDdSZ95l5gAHFL061pberavYXr3nLY/dsrnu2XO4iOmeAZIgARIwF0EKLDuIs98y0Sge7vqMuDSOlIBinvRzms+KN7/6aDMnrtX8hCgkQAJkIAbCVBg3QifWZeNQLumlWVYr0gJ0IaOjfbDX9Hy/Ic7JCeXDimMXHhNAiTgWgIUWNfyZm7lTKBp/XAZfmWUBFf0N6W8bONpeUw7jSctkw4pTGAYIAEScBkBCqzLUDMjZxGIrBkiNw2IkvDQAFMWm/YmysSZmyQhOdsUzwAJkAAJuIIABdYVlJmH0wlUrxykHFJU0xxTGG3/8RS5f9oGiT6dbozmNQmQAAk4nQAF1umImYGrCFQKDZSbNa9PdWuEmLI8EZch46ZvkN1Hkk3xDJAACZCAMwlQYJ1Jl2m7nADmYkf0qy+N64Wb8j6bkiOTZm+S9bviTfEMkAAJkICzCFBgnUWW6bqNQIC/n1zTO1LaNqlsKkNG1jl56p2t8tv6k6Z4BkiABEjAGQQosM6gyjTdTsBP2x87sHsdubRtdVNZcs+dl5c/3Slzfz9mimeABEiABMqbAAW2vIkyPY8icFmHGnLlJbU0hxT5xYIPird/2K9+6I8inwuvSIAEypcABbZ8eTI1DyTQsUVVufqySIEvY6OhF4veLHq1NBIgARIobwIU2PImyvQ8kkDzBpXk+r71pWKg+Vce87GYl8X8LI0ESIAEypOA+a9NeabMtEjAwwhE1Q5V58qGBZsdUmBlMVYYn02hQwoPazIWhwS8mgAF1qubj4UvKYGaVS84pKhSyeyQAntkx2mHt2PPLI0ESIAEyoMABbY8KDINryIQER6ovD7VrhZsKje8PcEhBbw/0UiABEigrAQosGUlyPe9kkBIkL/c2D9KGtYNM5U/Pilb+S+GH2MaCZAACZSFAAW2LPT4rlcTCAzwk+v61JNWjSJM9cAJPDiJByfy0EiABEigtAQosKUlx/csQcDPr4IM7llXLmlVzVQfnCWLM2VxtiyNBEiABEpDgAJbGmp8x3IEeneuKfgxWp7mhWL23L3y/k8HhQ4pjGR4TQIkUBwCFNjiUOIzPkEAvVj0ZtGrNdqcxUdkyue75VweHVIYufCaBEigaAIU2KL58K6PEcB8LOZlMT9rtEVrYuV/39smmdl0SGHkwmsSIIHCCZj/ihT+HO+QgM8QwMpirDDGSmOjrd4WJ4+8sVmS03KM0bwmARIgAYcEKLAOsTDS1wlgj+xI7fD2ytqeWaNtP5gkD2gOKU4lZBqjeU0CJEACBQhQYAsgYQQJXCAAb083D2gg8P5ktKMn0zSR3SCHYlON0bwmARIgARMBCqwJBwMkYCYQFhKg/BfDj7HRTidmyYQZG2XbgbPGaF6TAAmQgI0ABdaGghck4JgATuC5/or6ghN5jJaSnqvmZFduPWOM5jUJkAAJKAIUWP4ikEAxCPhrW3dwpizOljVaVk6eTP7vdvl5VYwxmtckQAIkIBRY/hKQQDEJVNC2x155SS25rEMN0xt52v7YaV/skU8WHTbFM0ACJODbBCiwvt3+rH0pCFzatroM7F5H/KC4BvtowSGZ+fVegeDSSIAESIACy98BEigFgbZNKsuw3pESYOeQYt7yaHn2gx2SrQ0d00iABHybAAXWt9uftS8DgSb1wmXElfUluKLZIcXyzafl0bc2S6q2CIpGAiTguwQosL7b9qx5ORCoWyNEbtYcUlQKNTuk2LLvrEyYuVHizmaVQy5MggRIwBsJUGC9sdVYZo8iUC2iovL6VL2y2SHFwZhU5ZDi2Kl0jyovC0MCJOAaAhRY13BmLhYnEB4aoHl9ipLImiGmmp6Iz1Qiu+twsimeARIgAesToMBav41ZQxcRCNLmYodfGSVN64ebckxKzZFJr2+StTviTfEMkAAJWJsABdba7cvauZhAgH8FGdarnrRrWtmUc2bWOXn63a2yeO0JUzwDJEAC1iVAgbVu27JmbiKA7bEDLq0j3dtVN5Ug99x5eXXOLvnyt6OmeAZIgASsSYACa812Za08gEDP9jWkX9faUsHgkOK85oPi3XkH5M3v9kkeAjQSIAHLEqDAWrZpWTFPINCheRUZenldwdCx0b5delxe/Hin5OTSIYWRC69JwEoEKLBWak3WxSMJNIuqpE7jCapo/u/2xz+n5Il3tkp65jmPLDcLRQIkUDYC5v/xZUuLb5MACRRCoH6tUHWuLM6XNdqG3QkycdZGSUjONkbzmgRIwAIEKLAWaERWwTsI1KgSpBxSVNUcUxht37EUtVc25kyGMZrXJEACXk6AAuvlDcjiexeBiLBAzSFFA6lTPdhUcIjrA9M3CMSWRgIkYA0CFFhrtCNr4UUEQoL8ZUS/KGkcGWYqNYaJMVyMYWMaCZCA9xOgwHp/G7IGXkggUDvm7pre9aRN4whT6bHgCQuf/thwyhTPAAmQgPcRoMB6X5uxxBYh4OdXQa7qUVe6tqlmqhG27rz40U7BVh4aCZCA9xKgwHpv27HkFiHQq2NN6dulluaQIr9CcEIBZxRwSkF/FPlceEUC3kSAAutNrcWyWpZA55ZVZXDPuuKv9WqNBreKcK8IN4s0EiAB7yJAgfWu9mJpLUygZcMIua5vfakYaP5viQMCcFAADgygkQAJeA8B8/9k7yk3S0oCliTQoE6o3Ng/SkKD/U31w1F3OPIOR9/RSIAEvIMABdY72oml9CECtaoGaw4pGkrl8EBTrXFoO/bKnkzINMUzQAIk4JkEKLCe2S4slY8TgLiOHNhAILZGO3YqXcZN3SAHY1KN0bwmARLwQAIUWA9sFBaJBEAgNDhADRdj2NhocUlZMmHGRtmyL9EYzWsSIAEPI0CB9bAGYXFIwEgAC56w8Kllw0rGaEnNyJVH39oiyzefNsUzQAIk4DkEKLCe0xYsCQk4JICtO4N7Rgq28hgtOydPnv1gh8xbHm2M5jUJkICHEKDAekhDsBgkUBQBOKGAMwo4pTBaXt55mfn1XvlowSFjNK9JgAQ8gAAF1gMagUUggeISgFtFuFeEm0WjfbLosEz7YrdAcGkkQAKeQYAC6xntwFKQQLEJ4ICAa7WDAnBggNF+XhUrk9/fLlna0DGNBEjA/QTM/0PdXx6WgARIoBgEGmlH3eHIOxx9Z7SVW87II29slpT0XGM0r0mABNxAgALrBujMkgTKgwAObcfh7TjE3WjbDpxV23hOJ2YZo3lNAiTgYgIUWBcDZ3YkUJ4EqkZUVA4palQJMiV7KDZVeX06cjLNFM8ACZCA6whQYF3HmjmRgFMIhIUEyE39G0j9WmaHFKc0l4rjp2+U7QeTnJIvEyUBEiiaAAW2aD68SwJeQSCoop/ccEV9aRZldkiRnJaj5mTXbI/zinqwkCRgJQIUWCu1Juvi0wT8/SvI0MsjpUPzKiYOmdnntOPutsmiNbGmeAZIgAScS4AC61y+TJ0EXEoADin6da0tPdvXMOV7TtsfO+Xz3TJn8RFTPAMkQALOI0CBdR5bpkwCbiPQvV11GXBpHakAxb1o5zUfFO//dFBmz90reQjQSIAEnEqAAutUvEycBNxHoF3TyjKsV6QEaEPHRvvhr2j5z4c7JSeXDimMXHhNAuVNgAJb3kSZHgl4EIGm9cNl+JVRElzR7JDiz42n5DHtNJ60TDqk8KDmYlEsRoACa7EGZXVIwJ5AZM0QuWlAlISHBphubdqbKBNnbpKE5GxTPAMkQALlQ4ACWz4cmQoJeDSB6pWDlEOKappjCqPtP54i90/bINGn043RvCYBEigHAhTYcoDIJEjAGwhUCg2Umwc2kLo1QkzFPRGXIeOmb5DdR5JN8QyQAAmUjQAFtmz8+DYJeBUBzMWO6FdfGtcLN5X7bEqOTJq9SdbvijfFM0ACJFB6AhTY0rPjmyTglQQC/P3kmt6R0rZJZVP5M7LOyVPvbJXf1p80xTNAAiRQOgIU2NJx41sk4NUE/LT9sQO715FL21Y31SP33Hl5+dOdMvf3Y6Z4BkiABEpOgAJbcmZ8gwQsQ+CyDjXkyktqaQ4p8qsEHxRv/7Bf/dAfRT4XXpFASQmY1+2X9G0+TwIk4PUEOraoKqHBAfLruhNyTuvB6oZebKK2hefpO9oUcFahP8NP7yWQmpoqy5YtE3xWrlxZBgwYIBUrmleZe2/tPKPkFFjPaAeWggTcSqB5g0oSHOQvC1bGSHZOvocnzMdiAdTL97eXEO0+rXgE8vLyBD9G8/PzE/x4in388cfy0EMP2Yrzww8/yPDhw21hXpSdgOe0dtnrwhRIgATKQCCqdqg6VzZM680aDSuLscL4bAodUhi5FHY9ZcoU1RMMDAwU44+/v78EBQVJmzZtZNKkSbJjx47CknBJvP0XAPuwo0J88MEH8sorr0hCQoKj24yzI0CBtQPCIAn4MoGaVS84pKhSyTxUiD2y47TD27FnllY0gT/++EMbaj/n8KHs7GzZvXu3vPXWW3LJJZfI22+/7fA5T4zctWuX3HffffLMM89IYmKiU4s4Y8YMGTdunNu/hJS1kuavqmVNje+TAAl4PYGI8EDl9Wm+dijAqYRMW33g7QkOKWZM7CTN7Q52tz3Ei2ITgNg+8sgjctlll0mXLl2K/Z67Hvz8889V1pdeeqk0bdpUXX/55Zdy+vRph0UKCAiQ8ePHC3ruJbUPP/xQ9u7dK5dffrm0a9eupK97zPMUWI9pChaEBDyHAOZbb+wfJQtXxcrRE2m2gsUnZSv/xa890EG6tKxqi+dF4QRWrFgh7du3l6SkJPnss8/UECvEFZabmyuffvqpxwssho8hprDbb79dfZ44cUJGjx6trgv756abbpI6deoUdtvy8RRYyzcxK0gCpSMQGOAn1/WpJ0u0hU57DG4UcQIPTuJ57u62aotP6VL3nbciIiKkSpUq6ue5556TrKwsefXVV20Adu7cabs2XqxatUpWrlwpW7dulfj4eGnVqpUS6hEjRkj16ub9y3gPPcmffvpJ1q1bJ7GxsZKenq7ybN26tdx1113qfWP6Jbn+66+/5Pjx42pO+ZZbblGv6l8SEIDo1q5d25QkerDVqlUzxflagALray3O+pJACQj4+VWQwT3rChY+bdyTv7AFZ8k+/+EOSUhpISOuqF+CFPloy5YtTRDs52vT0tLU0DEWFBkNc7uwF154QfV6Bw4caLv9/PPPy2uvvSY5OTm2OP3i559/FsxpPv300/LSSy/p0SX61IeHBw8eLDVq1Cjw7lNPPSVt27YtEO/rERRYX/8NYP1JoBgEeneuKaEh/rJy8xnb03maF4rZc/dKfFKWjL22qclZhe0hXhQgoAulfqNFixb6pfrEsOrixYtNccYAeqfDhg1Tc5SNGjVS24GwstdeqI3v4B6eGTRokPTq1ct461+v0RPGFh7YHXfc8a/PF/YA9tsuXLhQ/vnnH4mLi1M9XvSuR40apVZXF/aefTyG3JcuXSpHjx6VqlWryhVXXCHXXHNNgS1Q27dvl9WrV8u2bdsEvWl8AejXr580b97cPkmnhSmwTkPLhEnAWgQuaVVN9WQxZJyXl++QYs7iI5rIZsuTo1uJv9bjpZkJYMgW21piYmLk+++/V8O4xifGjh1rCy5ZssQkrlgA9d5770nNmjXlnXfekZdfflk9i+HZZ599VtCzrKC54cJ2IIjolVdeKSNHjlTDwejNTps2TX7//Xf1znntCxGEsqQCO3/+fElJSVHDzRCy0hgE8cYbb5SzZ88WeB3CD0aOesb2D2N7E1ZgG+3111+XmTNnyqOPPqqiUW+wmT59eoG9yNgmheF5/VljOs64psA6gyrTJAGLEmjVKEI5nMDiJwwT67ZoTaycTc2WF+5tJzixh5ZPANtNCrOpU6dKt27dbLchCroFBwcrQa5bt66KevHFF+X999+3rdr97bffVDwEFsIZGRkpnTp10l9XnwhDnHU7cOCAflnsT314GD1rCFRpDCIPkb755pvl+uuvV73Xv//+W1CngwcPqt717Nmzi0waz0NcsSr5jTfeUF8UUJ93331XfcnQX8YwOLjC0DseMmSIbYEZes+PPfaYYJh+6NCh+itO+6TAOg0tEyYBaxJoWDdMrTDGNh6cwKPb6m1x8sgbm2Xq+I4SERaoR/PTAQGIBEQHPU6jYa+pbhDWw4cPqx89LioqyiawZ86cUaJVqVIlufrqq/VH1CdWJ+N+ZmamhISESEbGhf3LWCxVEjt58qStB1zU8DBEzSjkyAPDv9imA5swYYKgp65v70EchmvRs8eXCojnv9nmzZvVIx07dlTpIYBrLPrS7dSpUzJr1iwVxLwwnH7odvfdd0ufPn1k48aNai8vBVYnw08SIAGPIlC7WrDaKztPE9mk1PyFNdsPJskDmkOKWZM6CZ6hidrHCaEz9h4xnIuel1FgsbgJ86u6QVyxD7QogwBCYJEetgDNnTtXNm3apIQLQ8Jlta+//lql3bhx4yLL8s033xTICr1nXWDxxcCR6T1uY70dPYc4PQ2sqv7oo49kzJgxpp4rnlm+fLmAI3r/EFijhYaGKg9ad955p1qZjS8d+PLhTGMP1pl0mTYJWJgAvD3dPKCBzF8eLWcSs2w1PXoyTRNZzSHFg52kSaT5YHfbQz508cUXX6ieFuYyMQ8JMYR99913AgG79dZbVRi9r5IYHPOjlwshvffee9XK4pK8X5xn58yZox7DflcMRRdmWNncoEED020swDIavmT8+OOPSgQhqPAGpTupKM6XARxGAIcc+AKB+mJ++YEHHpB77rlHfclAXvqXmGbNmqkFUMb8cQ3vWTDkhy8wcFvpTKPAOpMu0yYBixMICwlQ/otxSMDxU+m22p7WBHfCjI1quLhDsyq2eF++wNzjxIkT1fyhzgGLlrCvFOKFOVR86mKDXqM+3Kk/b/zEEGx4eLh8++23JnHF3lP00jDPiN4tBAjiVlLDKtwtW7ao13TnEoWlgWHaorbpQPiw2hcLvWD4coAVwPrQdWHpGuPxzl/aflzsJcYWpn379qntTNiCtGDBAuncubNNsHE6kCMzziEb9/E6erY84uiLuDwoMg0S8GECFQP95HptLyxO5DFaSnqumpNduTV/a4/xvi9eP/HEE2rLiF53zLnq23YwrFmvXj39ljpGDgt0IMyOfvT9tHBIYTRshYEw33///WqRD4ZGS2PoecN69OhR5q0tGM6FuMLtIeZbIfgY3v7kk09KVDR8YcBiKPSA8QkhRbr4EgGrX//Cnmx4mXJkeo8Z9+x73I6eL2scBbasBPk+CZCA2p5z9WWRgrNljZalHX03+b/b5edVF3ouxnu+eA0Bve6660xVf/PNN21huFTUDYuUnnzySeVOUY/DJ4ZW4bZQX/SDPaVG04egEQefvo62xhifd3TtyDWio+eKE4ee4vr169WjWAyFVdP6cLOxrMVJS38G3rEefvhhm0BjrysWdjVp0kQ9cujQIUEP3N6wTQoGr1Ou8DJFgbVvAYZJgARKRQBTdFdeUksu62D29IM9s9O+2COfLDpcqnSt9hJOpDHaL7/8IhAEGLat6OKDMMQX84nowWL+tmvXrmq1LuZEP/30Uzyi3Ceqi4v/YMgZPbqrrrpKrdwtzjF0xvdx/eeff6qeIYZlddeI9s8UNwwnD9inC4PjB91wXB+2zBTXUF/MW+tD6HhP/3IBsUQ+qLPushEM9Pt4Fh6tsKcY5qp9sBRYhZv/kAAJlBeBS9tWl4Hd64gfFNdgHy04JDO/3mtyUmG47TOXcHGo97RQaQggepowCCh6ZkaDxyL4GMZeV2wx0Xt9EBQYxNbYG8OQKQ5T1x1MGNMq7rW+9xXbf4xpF/d943M4ZF7fRoT5UvRgsWAJdYV/4+Iahr6xjxZz09iTC476HmN9jhhz0nBcgS8p2POKIWOc/gPvTRg5wJwv5mrhsMIVRoF1BWXmQQI+RqBtk8oyrHek1qsw/4mZp604fvaDHZKtDR1b1Yz7QTGvinlDo+GP/0MPPWSMkujoaFsY86dwlYhFTI4MaUKwMK8Jw/YViA+Ew2gQ4Ntuu83Wa8M9Y1mMwgkRxIEEMGxzwWpfmC5cKmD3D3qlem8bPd2iDFuS0AuHbdiwQbk6hOjhSwDytn9f7/Hqn3gPYooDD/CFA0O9mLsGC8xrG/e7oucKfpijxvA0hBaLrNAuOB5w7dq16j2k6WyroHW3y75ZytmlZPpuIdBmyCumfKs17GEK64GEo+v0S/W5a/FkU1gP9Bq3VL9Unw/fanZ6brrJgCUI4ID2n5bHSGZ2vkMKVKxTiyoyZVxHCQ+90AuzRGUNlUhOTlZzgpgr1HuahtvqEkKGHhVEpLBVr/DfC3HA4hykg+FP+C42Co+eLv6UY6gZvUIID/aY4hOG8qDni3wgaLqhDFhwFBYWZnsWi5sgrBBgLBayFz/9XXyi/Mi3uAupMB+M7THoWepfRFAG1M1+hS/SdsQFC5zwg3sNGzYssnxIG6uNcWSe7hHLWH5nX1NgnU3Yi9OnwHpx43lQ0ROSs2XesmhJSc93SIHiNa0XLjO1vbI1qpTO/Z4HVdFSRcGBAPCJjB4jep600hPI/ypT+jT4JgmQAAkUSqBaREXl9al6ZbOQHoxJVQ4pjhn2zxaaCG+4hAB6mPq2oaJcI7qkMBbIxGvHZzBsgnkH4wg3hingmaNDhw4OhxYs0F6sAgl4JQEMBd88IEp+WhEjsWcu+MVFRU7EZyqRnT6hk7RpHOGVdbNSoTEPixXE8JXcs2dPK1XNLXXxWoHFpDb8URZm2KCN4Q2M0ZfVsDEay7179+5tWiRQ1nT5Pgn4EoEg7ZSd4VdGyeI1sXIwOtVWdfgynvT6JnlpbHvp2a66LZ4X7iHQt29f92RswVy9dohYX6qOCX+c/YcN2ViGjT1jMKwig9cQuBErq8ENGE5eKMuy97KWge+TgBUIBPhXkGG96km7pmZXdpnaqTxPv7tVFq917IHHCnVnHXyPgNf2YPWmwnAwNmcbDRuKcTQRjkLCpm58I9M3Hxuf4zUJkIDrCWB77IBL6wj8GK/fkX98Wu658/LqnF2CRVG3DSr7yJPra8YcScBMwOsF1lydC6Frr71Wli1bpvaFJSUlqWOL4AXEaDjyCL1cLGvPyspSm5exT0s/PgnPYrJ/z5496rBehDHni+XhsGHDhkkjw2kRWA6OHi7SPXLkiDr3EG7P8BzmM2gkQAJmAj3b15Cw4ABZtvG0bS0FNg2+O++AxCdnycQbmxdwVmFOgSES8GwClhRYIEfPFsdAwWcnjokyGgQYpy/YG/xk4ggk3X0X5nmXLs3fuwnH1LpzauxDgzNtGHyCwp0Y9lvZW/fu3eWrr74yeW6xf4ZhEvBVAh2aV5HQYH/5VRsaRg9Wt2+XHlc92cl3tpFAO2cV+jP8JAFPJ+C1c7DFATt8+HD1GHqxOLkBhk3VEFe424JLrV9//VUJMBYwwWXZM888I/q5jPAQMnXqVHUkFN4dNWqUCiNu8ODBiFLD0P369VPiCkfe06dPV4ceQ6QhwnByXVZfnioj/kMCFiXQLKqSOo0nqKL5z9Ef/5ySJ97ZKumZZicVFsXAalmQgGV7sGgro7/PvXv3Km8e8GyC453gn9LoDQUCi/MYIcA4mQF+LrHhGj9vvfWWOjoKi5100dZ/F1577TV1WgW8nmC1MdKAjRw5UuWBT7jqgi9R+1M09DT4SQK+TqB+rVB1ruy8v6IlLSPXhmPD7gSZOGujzJjYSbCflkYC3kTA/JXRm0pejLIaBdS4X/byyy83iSuSgkDq5wPq86zFyEIWLVqkHhs7dqxNXPX34JC6adOmKgj/lzQSIIHCCcCj08iBDaSqnZDuO5ai9srGGPbPFp4K75CA5xCwdA/WKJQYEjYaepWYm4VfTAwfY6HTsWPH1CNGMTa+Y3+NIWX9mCmcEGFvcIQNB9wHDx60PWf/DMMkQAL5BCLCAjWHFA00/8XRclJzQqEbxPWB6RtUT7aF3cHu+jP8JAFPI2BpgdUXHWFYGCdO6PbUU0+pxUx6GN5L8AxOXiiJwYE2hBnmyCk14nVH2SVNG+/SSMAXCYQE+cuIflHyy+pYORybZkOA7TsYLn71/g7StXU1WzwvSMBTCVh2iBiOKN5++23FvW3btrYTJOAGDCuFYVg1DN+biYmJ6tQIOKYoiUGYcQoFDKdOOLIzZ86oaH342dEzjCMBEjATwMrha3rXK+A+EQuesPDpjw2nzC8wRAIeSMCyAostNtjDCoOQ6rZy5Up12bFjR7Vi2Njz1L1D6c/qn/rxTjgyyt70hVT2W4HwHJ5fsWKFegU+kmkkQALFJ+DnV0Gu6lFXurYx91ZzcvPkxY92ynd/Hi9+YnySBNxAwDICi3lTzKHCOQT8EGO7DQzOIxDWTT8fEauK4+MveJHBXCq21ezatUt/zPRZq1YtFUbvFyKs/yBSP3ECq4Q//vhj23spKSly5513qiFkvD969GjbPV6QAAkUn0CvjjWlb5da2uHe+e/kaf/f3/h2n3JKAecUNBLwRAJeL7A42T4kJEQd1gvH/theg72t6HWOHz9e7D04Yf8qDvfFdhz4LcbWGaz0nTVrVqHto58q8d1336n9sxBMHEoMmzhxojoAGaJ7zz33qLlePI/tOigH8po9e7ZtL22hmfAGCZBAoQQ6t6wqg3vWFX+tV2u0L387qtwrGp1UGO/zmgTcScBrBdY45AqxhMDBdSEE9oEHHlAOHt55550Ci48wNDxnzhzlmxjzr/BbjJNyHn74YXnwwQdVW0AUjTZ58mTb0U3Hjx9X87Z6rxY94g0bNsiYMWOUiEZHR8u6deuUgHft2lUwJA0HFTQSIIGyEWjZMEKu61tfKgaa/2zhgAAcFIADA2gk4EkEKmhDqz45wAJBhs/g9PR0dYYs/AUjLjU1tYAo6w0GD0/Y0oMVydg3a29AefToUcHq4latWtlWENs/5y3hNkNeMRW1WsMeprAeSDi6Tr9Un7sWTzaF9UCvcfluJxH38K0t9Vv8JIFiEzidmCnzNYcU9h6ecJ4szpWtHB5Y7LT4IAk4k4D5q6Azc/KwtCGoGBqGQ37dGT8+jYue7IuME3nQA3YkrngW+17Ri4YfZH17jn0aDJMACZSNQK2qwZpDioYFhHTX4WS1V/ZkQv7+2bLlxLdJoGwEfFZgy4aNb5MACbiTAHqp8PoEsTXasVPpMm7qBjkYk3+gu/E+r0nAlQQosK6kzbxIgATKjUCodtTdjf2jpEGdUFOacUlZMmHGRtmyL9EUzwAJuJoABdbVxJkfCZBAuRHAgicsfGrZsJIpzVTtwIBH39oiyzcX3LtuepABEnAiAfNyWSdmxKRJgARIwBkEsHVncM9I7VzZ07J5b36vNTsnT579YIc8MrKF3KCJsNXNflGi1epb2OJJT64ne7Ce3DosGwmQQLEIaOsLlTMKOKUwWl7eeZn59V75aMEhYzSvScAlBCiwLsHMTEiABFxBAG4V4V4RbhaN9smiwzLti90CwaWRgKsIUGBdRZr5kAAJuIQA9sNeqx0UgAMDjPbzqliZ/P52ydKGjmkk4AoCnIN1BWXmQQIk4FICjSLD1JF3OFc2w+DhaeWWM/LIG5tl6viOUinU2n/+vHHO0vhLYoU5ZfNXPGPteE0CJEACXkygTvVgdXg7DnE32rYDZ9U2ntOJF85yNt7jNQmUJwFrf4UrT1JMiwRIwOsIVI2oqBxSzNNcK8adzRfUQ7GpyuvTzEmdpFGdC2c6e13lnFxgzFfv2H9C1m05LLsOnJRD0XESn5AmqRkXOIaHBEn1amHSpH4NadOsjvTo1FjaNS84/+3kYnp08hRYj24eFo4ESKCsBMJCAuSm/g1kwcoYiT6dbkvulOZScfz0jfLLzD62OF6IxJw6K9/8skl++mObnElMKxRJYkqG4OfA0ThZslo7e/uzv6Rm1TC5bkAHGXl1F6lXu0qh7/rKDQpsGVp6xowZcuDAAXVkXbt27cqQEl8lARJwJoGgin5ywxX1BSfvHDieYssqOS3Hdu3rF2cSUuXtz5fLvKXbJFc71L40BkH+8Lu18um89XJD/w4y8fa+UrNaeGmSssQ7XiOwELKFCxeKfvgPHOsHBQWpM1379u3rFuf6H374oeDg9ssvv1wosJb4/8BKWJiAv38FGXp5pLwxd6+Fa1m6qn3/2xaZ/sFSSUkvn4MSINDfaWn+unKPPDG2v9w4qFPpCublb3mNwE6ZMkU++ugjh7irVKkib7zxhtxxxx0O7zOSBEiABEAADilo+QQysnLkudcXyaLlO/MjDVdhgf5yeVSY9IgMlpbVgqVBRJBUunhedkpurhxLzpK92lD7uthMWX08TdJyzGfyQrCfe2ORrN9yRF58eKiEBJkXnBmysuSl1wgszmqFtWzZUm699VZtCENr3GPH5Ouvv1YHoN95551St25ddeC6JVuKlSIBEiCBciSQlJwhY5+dqy1kii2QasPKFWVMx2oyrElV0aawHVp1TWirawcudK4VJre0EtHcP8vCQ4ny8dYEOZqUbXoHAn40NlE+fPkWiagUYrpn5YDXbdPB+a3PP/+8vPTSS/LZZ5/J33//LREREaqN3n77bSu3FetGAiRAAuVCIDk1Q+56+osC4hqqHZ7wRM+asnBEU7mpReHi6qgQEGK8g3eRBtIyGoT8zv/5QpC3r1gh3028p/qdOnVSc6CLFy+W/fv3Oyw4RHjlypWya9cuiYyMlM6dO8s111wjgYEFhyu2bt0qSOvQoUOSlZUljRs3luuvv16QT1ktLS1Nfv/9d0EeR44csR34PmzYMNuh72XNg++TAAmQQFEEcnLPyUMv/yB7D5tPGmpWtaK8dVWUNKxUsajX//Wev6ard7etIf0aRMiDS47LgcT83izyRN7vv3yr5mnL/1/T8vYHvF5g0QAVK174hahevbqpPSCQjz/+uDjq2Xbv3l2+//57qV8//5SNa6+9VhYsWGBKAwH0lqdNmyaPPfZYgXvFjdi8ebPccsstsm/fvgKvoCxfffWVNGnSpMA9RpAACZBAeRKY9fGfsn7bUVOS3eqGyLtXNdJ6naboMgUg1HOvbSoPLDki/5zI77Uib5ThqfsGlil9b3jZ3If3hhLblfGHH36wieKgQYNMd1999VUlrsHBwTJ16lRZvXq1GlbGXO369etlwoQJtuczMzNVOuixvvLKK/Lrr7/K/PnzpXfv3pqD8Dx55pln5NSpU7bnS3KRkJAg/fr1U+Jar149mT59usydO1cJNnrRKAvEl0YCJEACziSwZtMhmTP/b1MWXeqGyn8Hla+46hlAsJF259rmeVeUAWWxunldDxYiedNNN6nhWwwJ79mjbXDWrH///vLoo4/a2uv06dMyc+ZMFcZ2mttuu01dX3bZZRIWFiY33nij/Pzzz0r0WrRoIRDhVatWyaWXXmoaOobAYlgZArxt27ZSLaJ67bXX1EKsatWqqTljpAcbOXKkyg+f//zzj/z0009y3XXXqXv8hwRIgATKk0COtsL3pf/7TYznCdWrFChvDYgSba2S0wxpv31VA7l5/iGJSbmw7xhlQFl+fvc+7e+tdYeKva4He+LECTW0i6FcXVxHjRqlepyhoaG2X5J169YJ5jyrVq0quG+0AQMGSMDFpeY7d+YvT8d+Vvt5WYhigwYN1OuxsbHGZIp9vWjRIvXs2LFjlVgbX8SXhaZNm6qotWvXGm/xmgRIgATKjcBXCzdoK3kTbOkFaEf6TbmyrlQNcr4MIA/khTx1Q1lQJiub88mWMz30KJcvXy6ffvqpbeHRt99+q4Z/jVnBMQUsMTFRLVRq1KiR6D8dO3ZU23xwPyYmBh82Q09y8uTJSpQxrAvRxXYgmO7kwvZwMS4wvIwFU7Bu3boVeAMOM7DoCqY/V+AhRpAACZBAGQig9/rJD+tNKQxvWVku0bbYuMqQF/I0GsqEslnVnDgw4BxktWvXlj59+qifm2++Wbp06aJ6spMmTVKrc/Vcz5w5oy79/Pxs23j0e/jEdh/0Vo2i99RTT6nFTPpzcGCBoePs7PxVcPq94n4mJyer4Ww8X7my+ZdLT0NfpFWWfPS0+EkCJEAC9gQWr9wlpxPyXURiC81D3WrbP+b0MPJceCBZ0i+eyYsyoWzX9mvv9LzdkYHX9WCNkEJCQgQLmWCYH126dKnttj6si0/cc/SzceNGwQpe2J9//mkTV6waPnv2rOr9Yki6LG4QIdKY84UhLUemfxnQy+zoGcaRAAmQQGkJLFy2w/Tq9c0jpKrmn9nVhjyRt9Hsy2a85+3XridczsSwKCgqKkql+t5779lS1+c1Mbx78OBBW3xhF9gnC8PwMVYMG3ubuhepwt79t3h9+w1WJdsbFmOtWLFCRbdp08b+NsMkQAIkUCYC8Ni0dvNhUxoj21Y1hV0ZsM8bZUMZrWheL7AYAr777rtV20DATp48qa4xVwvhxRzo/fffL9gqY7SjR4/KF198YYvCUDAMzvvj4+PVNd7F3lc4qCiL6T6SsUr4448/tiWVkpIicPGI/bq1atWS0aNH2+7xggRIgATKg8CGncflnHa2q25wKNG88oW/d3qcKz+RN8qgG8qGMlrRvF5g0ShjxowRCC38E+sHAmD4GPtNsYgIQ8fo0WLR0vDhw6VDhw5q4RPe023w4MFqZTG24zRr1kxtl8E7s2bN0h8p9efEiRMFW4HQE77nnnuU8Pfs2VOtKMZ+W6xonj17toSH++6xTqWGyxdJgASKJLBp5zHT/e6R+bstTDdcGLAvg30ZXVgUp2blNQKrLwTSP41UGjZsKEOGDFFR2CerG/aXLlu2TAkqFhvhet68ebJ9+3bBO/BprBuGhufMmSNYRIX5V+yRjYuLk4cfflgefPBB9Zi+tUd/R9/So3/q8faf6B1v2LBBfRGAiEZHRwu2EUHMu3btqtw42m8lsk+DYRIgARIoDYGDx+NMr3WqY3b6YLrpooB9GezL6KJiOD0br1lF/O677ypvTLpjf3syEESIqL6gSL+Ps2Lh+zcjI8PmqxienGrWrKk/YvvEKT1YmQw/wenp6YI5UX9/f9XzxMIn47wsXsIiKaRrH29L0HBRqVIl1buG0wsMT6OsrVq1srl5NDzKSxIgARIoNwJHYxJNaTWqHGQKuyNgXwb7MrqjTM7I02sEFkPAWJFbmP3bfQwZY2j43wyCqi+Q0p9FnCMRRW/aUY9af8/RJ4asG2l7cmkkQAIk4AoCZ1PSTdlEhrlfYO3LYF9GU4G9OOA1Q8RezJhFJwESIAG3EUjPMO/jD8NxN242+zLYl9HNxSu37N1PutyqwoRIgARIgARIwHMIUGA9py1YEhIgARIodwKhIflbYpB42rm8cs+jpAnal8G+jCVNz1Ofp8B6asuwXCRAAiRQDgSqVDJvy4lNyyqHVMuWhH0Z7MtYttQ9520KrOe0BUtCAiRAAuVOoGE9s9emI0nuF1j7MtiXsdwhuClBCqybwDNbEiABEnAFgaZRNUzZbDnpfreE9mWwL6OpwF4coMB6ceOx6CRAAiTwbwS6tL1wnrX+3PpY87YdPd6Vn/ZlsC+jK8vizLwosM6ky7RJgARIwM0EuraNEn/DQecHErNlf1Km20qFvFEG3VA2lNGKRoG1YquyTiRAAiRwkUDliBDp2bmxicc3O83enUw3nRywzxtlQxmtaBRYK7Yq60QCJEACBgLDrmxnCInM358sidmu366DPJG30ezLZrzn7dcUWG9vQZafBEiABP6FwJDebaRWtUq2p9Jz8uSNf07Zwq66QJ7IWzeUCWWzqlFgrdqyrBcJkAAJXCQQGOgvd4/obuLx494k2Xg6zRTnzADyQp5GQ5lQNqsaBdaiLTvti92yaE2s6aBli1aV1SIBEigGgVHDukrDyGq2J3O1g87/Z9kJSczK71HabpbzBfJAXshTN5QFZbKyUWAt2roxZzLktTm7ZdTzaym0Fm1jVosESkIAPcVnxw+SCoaXYlJyZOKSY5KZa4gs50ukjTyQl24oA8pi5d4r6kqB1Vvcop8UWos2LKtFAqUgcFmXJnLH9Zea3tx8KkPu/007Aztf/0z3yxJAmkgbeRgNZUBZrG4UWKu38MX6UWh9pKFZTRL4FwKPjukn3Ts0ND31z4kMueXng3I0JX9/qumBUgSQFtJE2kZD3iiDLxgF1hda2VBHCq0BBi+dTiAx/rR8+f5rMv+rd5yel7dmsG5HvMz4ao+cSnCN84fAAH9545kR0rJxLRMyOH8Y8eNh+WRnnLZ2w3SrRAG8izSQltGhBBJBnsgbZfAFC/CFSrKOBQnoQjtn8RG5Y0gjGdyjrsnbS8E33B9zMuaIbFz7h5w/n79QIigoROo3aiENm7aW0LD8bQjG0u7auk77Iz9F6tZvLBOfnm285dLr/5v6mBzev1NuHzdZOnTtXe55e0o9jRWLOXZAE9f/k8CKQXL9qAnGW7y+SCAnV9sbuiJGWytxQoZeVlduH9xIalcLdiqfiPAQ+fS10TL22bmyY3+sLS9soZm+9ox8uytJxnSsJsOaVJWQYqpEhjbXuvBQony8NUGOJhXsCbdrHikfvHSLIG9fsWKi8xUcvldPbxLaeVov6M9FcwttpM7dr5Sxj74qNWvXNz2zcc0fsm/nRvVz36OvScUg5/7xMmV+MZCZkSbLFn+rQts2rnSKwHpCPR3VnXHFI+BqoYX3pM+mjZbnXl8ki5bvNBUSAvn8ipMyTRPby6PCpEdksLTURL9BRJBUCrggGym5uXIsOUv2aj3vdbGZsvp4mqTlnDOloweG9m0rLz48VEKCAvUon/ikwGrN3GvcUp9o7KIq6Uhoi3reHffyLo5bRUY1kcuuvEZycrIl9vhBOX54n6B3u3n9Mnn0rgHywJPT1X29jJf2HizHDu+Reg2au0VcUY7gkDDpP/RWiT8TK90uv0ovWrl+ekI9y7VCPpqYK4UWgjf9qeule6dGMv2DpZKSbh6mhmAuOZSsfkrTHJVCg+WJsf3lxkGdSvO6179DgfX6JizfChiFtnxTLr/UGjRpLSPHPG5KcMPq3+WdqY9KavJZ+e+Mp6Rtp55SueqFY7patusqk6d9YXreHYFxT0xzaraeUk+nVtKHEnel0EIA+3ZrJm9/vlzmLd0mudqwdVksIMBPbujfQSbe3ldqVgsvS1Je/S4F1qubj4XXCXS9fKD8Z/a32rflwZKeliJf/PdVmfA/s9TtuFMxsmrpfKleK1J6D7hBf0V97tq6XnZsWiVntGfCwitL2849pWvPAVLBz7z+LykxTv5ZvUSOHdqj/fHJkQaNW8rl/a+TShEXDrM+czJa5dG4RXtppQn6L99/LDFa77puvUYy/PYHxc/PX1b9MV/LJ1p69b9eata5MIxtfK9Tt76y5Z/lsvWfFZKdmSHN23RWzwYEBkpuTo6sWPKD7N25QSKqVJc2HXsIhsSNVlg9V/z+oyScOSmDrrtDsrIyZMOa32X/7s1Ss1Y9aX9JL4EwO7KjB3epUYFTscfUaEGtulFyaa9B0qhZW0ePy4njh1T5j2jv1a7TQLr07OfwOWNkVma6bN2wUo4e2CWnNYZ16jWUBk1aySU9+ytmxmeN9Th8YKes1OoFu3LIzdKi7SXGRy117SqhhRC+8NBQue+Wy+WbXzbJT39skzOJJfP0VLNqmFw3oIOMvLqL1KtdxVLtUJrKUGBLQ83C79SrGWJb9NR+6DKvqikWOvXSRG/l7/Pk75W/2gR2rzb/ikVOtSMbmAT24zefk8U/fmKq46LvP5Q7xz8nw24ea4v/e9Vv8t60JyQl2XwCydcfTpPnZs2Vpi07yL5dm+SrD6ZKx2595Gvt89C+7bb3B91wpxLi7z57XQ1pV69Z1yaw+ntdevST5b9+p4n0T7b3lvz8uaoLhrynTr5HjmiiohsWDl0z8j6544Fn9ShNfB3X84c5b6p8MUz93WezJflsvO2dbz+dJXeMf1aG3ZRfX9yc+r9jlBDbHrx48f2cN+T2+yervI331ixbIO9MeVSys/KHGL/5ZGaRc82H9++Q11+coJXtkDEpdd28dWd56Nm3VZvpN/V6pKaclZ+/+a+cz7vQywoIrGhpgdXr7yqhhTA+enc/efjOK7UFUCdk3ZbDsuvASTkUHSfxCWmSmpGlihQeEiTVq4VJk/o1pE2zOtKjU2Np17yu9sXI6MpCL71vflJgtXZf9V5/y7X+Q7M3yca9ZkEoqpJGYTWeHVnUO554r3vvIUqU0Is9m3BGqlSr6bCYB3ZvUeKKnuXdD74grTt0kxPaPO6Snz4Xo6ub+NOx8sZLE5VwNGraRq667nbx9w+QNX8tUD3NnVvWKoHVM0HvE9ZE68l26NpHE8UdUlFbQftvtmndn+qRZq07qS8JOzatUQKHBVGTRvdRPUiI9yVa73r1nz/L3h0bZOG3H8iAYaMkMqrpvyWv7n/0xjNqDhpzwY2atZFf530mWOX7lfblAz17fUg9JztL5Y0eK55ton2BQNwCTdR2b/tbvv5omvS5arjtefTK33x5kpw7l6ut6G4u/a++Vc6cjlE97i1//+WwbBjKf+GRkZKWmizVatSRoTfdKzW0EYYDe7bILz98onrYr784Xl57b2GB93/6+l1Br77/taNVu9Rv2LzAMyWJ8LY1GK4SWghlh5aR6qckPPlsPgEKbD4Ln7yyirDqjYdeqm5YAFWYwKL3BIPQDNZ6mLCGmoD26HO1utb/+Vbr8aFXhpXJL771g4SEXphP6jf0Flm3fJG2cKqZ/qjts3WHS+XZmV9rbuAq2uKKc9HzimHy8HNvq6HRq4ePUT0I9O6wmOvaW8ap7T1I54rBN8n9N3aTjPRUNYRbXIENCg6Vp6d8quankU43bbh3wi2XqfS3akPTfa4agWi1pealt3+U5q06i//FFaOIb93+UrlvxCVKbI8e3G3rnS787kMlrhBoDNPrQn3VNbfJC4/eItgLa2/zvnxbiWt4RBWZoolo1Rq11SNYvNasVSeZ/cJ4TWy3qmF5+0VhGL5//MX31ZcN+3R9KWwvtL5Ud2+pq3miyVtKzXKWmQCE9ek7WstXL/TU9t5Fevwe2OJW2D8gfxuAcb+s/fuYj4Ud0eb+sPWnsGf379yknkNPURdXFaH906PvUInS5mKNBhF76Jm3SiyuVavX0oZE37LNO0JE0AOGYX4Re2d1QzlatO2igphbLa6Nvv9pm7jiHQxVY1gdZi+Crdp1M4krnoEY1qhdD5eSEJ+f7x6tVwvrPTC/V4twPa1nef/jU3FZwPQeO7jq4qo/hC8adSIbqiC2V9nbkBvu8nlxtWfCsGcSYA/WM9vFaaWyWo/VHlRi3ClbFIY4C7MOl/RWw7iYK313+hPy09x31fBvv6tvsQkpRPdk7FGVRFQTs5AWlm6VajXUYqrC7hcWj725GHo2WmVtMRMsNKzgKkx9cVW2NnRbXHPkiEPvbaI3bG+qB6nNP58+cUwNt6MnjYVUys7nP40FYrCm2rC4veELh71h7vTURa5NW3W0vy0VKlSQRs3bKfZYYGVvmPOmaSMN2kpdo2OKj7XZDZpnETD/j/assrE05UjA6sKqo9IXzMBzUI2aF3qp+j3jJ+bw/vP6t/LNxzPlj4VfqYU2n779gvw897/yP699Io21P/Dp2vwg5h5h9r1XY1pOu9aEplAr6l6hLxW8YS/q+hNYhY25Tt3CwiPU0DFWUBsNX0JSLy7+CgopKKbGZ/XrdE3MIdYwR6KPeLQPLDf3wnMq4KR/3L0GY+WWM/L0e9uKXTt7YS32i3zQ5QQosC5H7toMfUVYQTUv75y2cOdTBThKc59ov9VG3TD8A9G8a+LzMvLux+TPxXPl209mSULcSXlXWzE87YPFElapsloUhDnYsw7mEQ1JWepyx6bVNnG95Z4nZMjwu21C+NiYgWqrkl5h9DaxbQjbmPBTHINYo2eLLTqFcU0+m6CSqqFtJaJdIEBh9b7fBAqs97VZsUoMYb2qex2v8DFcrAoV46H5X/6fWhWLRyEMxbUQbQh26I33qoVM058dK1jAgxWx6N3VqhMl0Uf3q8VEvQZcb0oSQ5152o9xIZDpAS8NYKUwDKumR9w+yVQLfImxN8zlQly3aftZMadqtO0bVxmDtmssRsOeYmyB0hdX6TeR1m7NfzQMq5J93dwlrDm552Tr7hhZv+2oHI2Jl9PxqXImIVX7TFFNUqt6JeVEolb1cGlYr7o6oadj63o+48i/OL+XFNjiUPLCZ54cfWHxihcWvdhFxvAkttHERh/StpC8L/qWEDhDsHfCYJ/oX9qe06DgELVQCb0wWHLShV4TFvPoQ6c9+l4t2Pu5Wtvn2b3PEIE7Qhjmbt+b/qRyyXj9qPHBMZQDAAAv0UlEQVQqzir/YHgdBkcZ2PuL+V58mZjz3isSfWR/gWr2vGKo4oG9x3CUAYcZEGJw+/GLtwo8j4i+g26Uz999Wf5ZtUT+/OUb6Xf1SPUc5oLffu0RNYSM+eE+2sIpXzV3Cevugyflm0WbZPGK3ZKSlr+v2b4dDkfHC350e0e7qBQWLEP6aJ7WhnaR1k3r6Ld89pMC67NN770Vx/aYUVc1U73Hc4Y5QQwJX6XtjRw19n/+tXI4lWfd8l9UrxULbTDfun3zavWesUc17Ob75PcFX6oeGnq3WN3q5+8vJ6IPq5XHcGxhNevc/QqZ+9F0Nf/84Khe0rpjdzl2cI/maem4w6piYdii7z9SC6FeeWK0YHge+5Djz5xQq45tC6MMb2Ml8NKLc9/vTntcG56fqVY1Hzu8V3AwAr7g3DXheeXD2fCaT1y6S1gPHY+XN+f8Jb+v3qP9bpcONQT528Wb5btfN8vAy1vJpDuukCZRFxbrlS5F736LAuvd7edTpY8yDBdi8REEFUO4dbRj6OrUa6R6QY5WmOpDuAEB+ftSB2pCjEMC4GgBjhJg6NEOuv5OGXXvUzaumC/EfOw7Ux6TnZvX2FYVh2hH4w3UhkMHa0IBc5SHumH4J+DinlLjVqKi3tOfN5ZbT06Pgxcj3QpLS0/HmK/+jr6YKMCwvQn7gSf+7+uCRV9JiWcEfp7hBQrD6OiZwvsVvmTohjnYl9+eJ69rDjkO7t0mx4/sUywv63eNjHnwRW3vbFcxpo/30Eue8t9FWh7/kTXLFioxhiBDWNGGYx56SVq06aJnoT6LqofpQS8NuEtYM7Ny5J0vV8pn89aX2Qexjh4CvWTVHvlz3T6584buMuG23hLsYyfpgEUFbZitlN9VdJT8tCqBNkNeMVWtWsMeprAeSDh6Yb5MD+9aPFm/NH3ae8x5+NbibX0xJVLKQEZaqvqjrguKngy29STEn1KLeOBMwv6+/hw+IepwXgGxgKjrw8j6M4Xlod+HP+Hs7EzbgiE9vrD38F8TPcFgbXWufV4Quoz0NAnVFmoZF3M5SquwfJE/5pozM9JVmfShcr1cyAO+krM0v8iYC4XXK8Tpz+vPGT+R/6kTR7VebEvblw68j7QLOyYQ9UQvF3WF447C2qCoehjL8G/Xr3+91/SIu1cRr90RL6u3nSnzObD2/18L+39orPz+I2fkkVd/EPReC1iF8xJWI1ki6p+RylFn1GeE9llZC8OSomtK8nHtR/tMUp81JC2ussj5gqvf0Yud/b8jpHkjx57VCuStRZSmPo7ScWcce7DupM+8XUYAC5kcGZwc2Ds6cPQc4tDrQu+uMCssD/15CIcj8SjsPYgSetCODGLn6J6jtArLF+lCuB2lg3vIo/ZFhw8IwxBX2NYa3Ef+9ocBYGSgKEM99cMPinquqHoU9Z6n3+vZrrrgx9X2z/ZjMv75byQto+BWqBotj0uXMb9JjVYF9yHr5azV5qjgx2hxexrIpo8HSdxe8x50CPioRz+V/3thpHRrn+9tzfiuFa/pycmKrco6kQAJkEARBOC8/wEH4hpW86z0euobuWraB0WKa2FJQ5DxLtJAWkaDkCNP5O0rRoH1lZZmPUmABEhAIxCvHUE34YVvJd2u59qk/2YZ+vbb0uCynWXmhDSQFtI0GvJE3iiDLxgFtohWPnHihNx6661yww03CK5pJEACJODNBPK0+e4nps2XU3EX9rLqdel4+x/SY9I8CQguOFysP1PST6SFNJG20ZA3yoCyWN3cKrDr1q2TF154Qe68804ZM2aMzJo1S44dO2Zivnz5crnsssvUM6YbLghMnDhR5s6dK/Hx8VKzZvEn51E0d5bbBWiYBQmQgBcS+OzHv7XzXY+YSt5m+Cppe+MKU1x5BpA28jAayoCyWN3cssgJ3m9uv/12+eqrrwrwfeqpp2ThwoUyaNAgdQ/Xa9euVT/vvfeehIQUvWCiQIKGiISEBJVOZGSkdO7c2XCn4OWCBQvkxx9/lDp16si3336rrRwtiOrvv/+WuLg46d27t1SqVMmUSHmW25QwAyRAAiRQCgLRJ8/KW3P+Mr1Zt/MB6XjH76Y4ZwSQR+LhOnJiczNb8m99vlzbK9tS6tepYouz2oVberDoqeriCqH9/PPP5Z133pGBAwdq+7ByZc+ePTbOGJ6F2D788MNlElckOG/ePBk2bFixesMPPvigElWIK0TWkY0YMUKGDh0qv/9e8Be0PMvtKG/GkQAJkEBJCMz4aKlkZufaXgmqnCY9H/5R20Ll/KFa5IG8kKdu2H+LMlnZCnbLXFDb+fPnq1xGjRolc+bMseU4fvx4+eeff6RWrVq2OAwP//rrr7awqy7Qgw0PD5fGjRuXKkt3lbtUheVLJEACliaw59Ap5aHJWMlL7lkswVUKHlNofKY8r5EX8lwz60ZbsvAahbK1alLbFmelC7cI7IEDBxRDR+LVrVs3E1/MyaK3GxUVJbfddpvt3hdffCExMTHywAMPSHp6ukAQ169fLw0aNJABAwaoeVv94eTkZNVLXrlypYo6ffq0vK2tcINBzG+++WZ1jX+SkpJk0aJFsnnzZsF7HTp0kP79+0urVq1sz/zxxx+ql41nYRgOjo2NVdfoITdq1EjNJTsqt3pI+wdpYA768OHDan4XvfQrr7xSv237TEtLUz3krVu3ypEjR6Rp06bSvn171RP3N3jTsb3ACxIgARKwI/Dpj+tN7g+rt4iWRn222z3l/CDy3Luwh8Tvq68ywzonlG3K49c6P3M35OAWgW3RooWcOnVKiR4Esl69eoVWfc2aNfL0009LkyZNTAL78ssvy969e1UvEwulzpy54F0ECf3nP/+RmTNnyiOPPKLS3bZtm2DBkm7IG0PAMAz/6gKLvNCrPnrUvHm6YsWKagHWhAkT1DtTpkyRpUvzhzY++eQTwQ8sUHMmcP/990th5YYoY0EX5neNNnXqVJX3l19+aYuGyN9yyy2yb98+W5x+0b17d/XFA1xoJEACJFAYgdT0LFmycrfpdvuRf8GPnynOJQEtT+T910ujbdmhbM+MHyThoUG2OKtcuGUOVu+Jond6ySWXyIcffqidnpFTKqYQztTUVLn33nvlrbfeUj1NuF6DKENIYR07dpTp06er7TYIY0UwBA0/EGIYFkCh9wlx7dOnj+oRL1u2TO666y7NvV22EuRNmzapZ5944gn1LoaQYRBlPb3Bgy+ctqJuOPjnmWeeUeLqp/nRxRcA9HIRV61aNfnpp59sb6A8/fr1U+KKLyAoP1Y0P/bYY0rE0VuH+NJIgARIoCgCK/45aJp7Dat1Vupesr+oV5x6D3mjDLphXhhltKK5pQd73333CcTq/fffVyI4duxYefHFF2Xy5Mlyzz33OFyxWxj8sLAwNUR7xRVXqEeuv/56NW+alaV9a1uyRK1Wxgrfxx9/XD766CO10Am91ieffNKU5GuvvSaJiYnSunVrNXyLnigM6aKs6AXPmDFDCSKGc/EDQYe4Y7HT8OH/fqwWhnixEho2e/ZsmTRpkrrGP5h//vnnn21hlOfs2bNKeLFaGSufYSNHjpRLL71UfWK+GqJ83XXWO9HFBoIXJEACZSKw3m5bDpxAuGJhU2GFRt4ow+75l9seQRmv7lu4G1Lbg1524ZYeLHyP/ve//1Vi16bNBajHjx+XcePGSd++fW09z+KwxHCtLq54vn79+qrHiuuSOIfQh3wxvKuLK9KAXXPNNepz586d6rO0/2zYsEGtkobg40uF0erWrauGlvU4zAPD8Jwurvq9m266Sc3FIowtTDQSIAESKIzAzgMnTLdqdzhsCrsjYF8G+zK6o0zOyNMtAqtXBL3NHTt2qCHTli1bqmjMXcLxRHGtcmXt9AY701chp6SYvZXYPWYK6guvnnvuObVICQuV9B+914lFVWWx/fsvDMtgoVJR+3mxT/jQoUMqK/tFX4jEFxR9H6/+XFnKxXdJgASsS+D4ifzhWNSySsMLU2furLF9GezL6M6ylWfebhVYVARigT2jGILV5y9/++03QY+2tObIKURRaWE4WRdjDDlHRESYftCDxMrdG2/MX15eVHqF3Tt58oKTa6RflGH1MsoEc/QFAvFYeAXD/DCNBEiABBwRgDvC1PTM/Fva8GxI1eJ3PPJfLN8rVQbDIiuU0YquE90yB+uoqSAY6D3qe14PHjyotuY4era844KCgtR2HWzfwdxnSXrQJSmLvlr634auq1SpIhB6bNEp7Fl91TS2JdFIgARIwBEBbIPBj24VtIsK/nl60G2fKAPKohfNVk5EWsjc0oMtbMUw9rPqhrnU8jas3IVBSO0Nw7awP//80/5WoeGi0nP0EoacYRiO1oeLVcTFf/ReK4L69hvdKYfxOZR/xYoVKkqfwzbe5zUJkAAJkID7CbhcYOFgAdtkXnnlFZPQYa/nQw89pIhgSLZZs2blTkefm8X2HQxJw3RRw1YbGLbNfP/99+pa/wfuG7HoyH6Rk54eRPncuXO2H/09+0+4gsSQL7YRYUEXhoJhmZmZ8tJLL5k8WN1xxx3qHlYJf/zxx+oa/2AoGz1slBv5jx6dv5/M9hAvSIAESEAjoM3AqR8dBnqM58+5/M++nr3tE2XQe6+ItC+n7UEvv3A5aQwFw9kC9n5i5Sx6dRBTbI+BgME7EXwTO8O6dOkiGA6GYf9tu3btVL4IQ/DgtQliilW6uIetNxBFfCHAHlnjPlW807NnT3zId999p7YGQfDgYaowq1q1quAwAxhEGUPGWKyEbUMYHkfeumF/LxxyQLixdQmerJAfvnxgGB3zzNjqo+/F1d/jJwmQAAnoBPw05QoPDdaDmrpWkIxE88Ek+Tddd6XKoJVFN5QRZbWauVxg4cQB+18hYFgtC8cOmG9Frw5bdODOEA4WdNO3zOiLeuzj9ft6PD71Z+3vQdDffPNNCQ4OVmIGQa9du7Z6FYKFFczYm4ptNLiHwwHQ40YPc8iQIYJVz0bDvl1dZLEoC/tW9V6tnrdeFv09OMDAFqUaNWqoPbRbtmxRXzjgmWnx4sX6Y6qM2NYDr08Q0ejoaOVaEWXp2rWr4qT3um0v8YIESIAE7AhE1a1iijl79MLfPFOkiwP2ZbAvo4uL47TsKmjCZuypOy0jRwnjqDcIB8QIc6AQPkeGoVTcM4oVVs9mZGQ4XGWLniAcQGA4FquU7Q0iBVFHvug963OpxuewJQf+hdHrbNiwYYG9scZnMeSMFcLoZcIjk26Oyq3fwyfywCImLFTShdl4X79GE+GLCNKDT2QjB/0ZZ3y2GfKKKdlqDXuYwnog4eg6/VJ97lo82RTWA73G5buXRNzDt17YmqXf5ycJuILA61/vNWWz6r3+prC3Buz/v+r/D59/4xf57tfNtmq1vn61dL77N1vYHRebPxlkcjRx0+DO8sJDV5uKUlh9TA95eMCtq4jRi8PPv5mjbS0QmcKEBr1RrMQtzCDWbdu2Ley2isfwrb7qt8gHtZvoBes9YeOzjsptvF/cPPAloZE2lE4jARIggZIS6N6pkUlgj61pK53uWqJ1PtzTtzqvDQ2jDEZDGa1oLh8itiJE1okESIAEPJVAn27a6GDF/L5U2ukqcmJjc7cVF3mjDLqhbCijFY0Ca8VWZZ1IgARI4CIBnFJzVe/WJh7bv7lCLXgyRboioPVeVd6GvFA2K56kgypSYA0NzUsSIAESsCKBu4Z3N23XwXmsR1a0d3lVkad+Fiwy12a/BGWzqlFgrdqyrBcJkAAJXCTQqkltGXh5KxOPjR8NkcyzF47cNN1wUgB5IU+joUwom1WNAmvVlmW9SIAESMBA4PF7+pvmYrOSwmTt68O1LZIFd1oYXiuXS+SBvJCnbsFBgYIyWdkosFZuXdaNBEiABC4SqF+nijx4xxUmHic2N5Otcwaa4pwRQB7Iy2gP3t5XUCYrGwXWyq3LupEACZCAgcCdwy+VHnZbYnb92Et2ft/H8FT5XiJt5GE0lAFlsbpRYK3ewqwfCZAACVwkAHeE05+8XmrXMLtL3Pr5AFn35g2Sm3nhGMzyAIa0kCbSNhryRhms6BrRWE9cU2DtiTBMAiRAAhYmUL1qmLzz/M0SGmIW00NLO8sizQe6vROI0qBAGkgLaRoNeSJvlMEXjALrC63MOpIACSgC7nMM61kN0KZZHXn3hZESZieyaWeqyKqpI2XJk2Mlbk+DEhca7+BdpIG0jIa8kCfy9hXLd+/hKzVmPUmABHySwLlz52Xx2hM+WXdHle7WvoF8NesueeTVH+TQ8XjTI3F7o2TJ/9wjYTWSJKJ+nFSOOqN9aj/aZ2XtE5YUXVOSj2s/2mfSxc+0uAiHDiyaRFWX2f87Qpo3qmnKx+oBCqzVW5j1IwESkKzsPFmwMkaiT6ebaESEBZrCvhaA4H3/1j3yzpcr5bN567VTxvLyEWhba9ALxY/9CuD8h4q+Cgjwkztv6C4Tbust2Jbja0aB9bUWZ31JwMcIpGXkyry/oiXubJap5rWrBcvMSZ1Mcb4YgPA9Nqaf3DCwo7w55y/5ffUebW9s2UjAQxOcSEzStgWh9+qrRoH11ZZnvUnABwgkJmcrcU1OyzHVtklkuMx4sJPUqhpkivflAITw9ckjZPfBk/LNok2yeMVuSUnLLBGSSmHBMqRPaxk5tIu0buo7c62FQaLAFkaG8SRAAl5N4GR8pvy0PFoyss6Z6tGhWRWZOr6jVArlnz8TmIsBCON/Jl0tk8cPkq27Y2T9tqNyNCZeTsenypmEVO0zRT1Zq3olqVktXGpVD5eG9apL9w4NpWPrehIY4O8oWZ+M42+YTzY7K00C1iZwJDZNFq2OlRzjnKJW5d4da8p/7m0nQYHcQPFvvwEQyq7aQij80EpHgAJbOm58iwRIwEMJ7DqcLH/8fVLy8swTidf2ipTHR7USPz/n+971UDQslosJUGBdDJzZkQAJOI/Ahl0JsmrrhW0kxlzuHtpY7rmmiTGK1yTgdAIUWKcjZgYkQALOJoBVrys2n5bNexNNWaG3+sjIFnJD3/qmeAZIwBUEKLCuoMw8SIAEnEbgnDYUvGTdCdl79MLiGz2jito86/Nj2krfzrX0KH6SgEsJUGBdipuZkQAJlCeB7Jw8WbgqRo6dNDuQCA8JkCkPdJBOLaqWZ3ZMiwRKRIACWyJcfJgESMBTCKRn5sr8v2LkdKJ5r2aNykHKgUTTeuGeUlS3lKPNkFfcki8zzSdAgc1nwSsSIAEvIZCUmiM/Ljsu+DRag9qhMuuhzlJH89JEIwF3E6DAursFmD8JkECJCKDHOl9zfZieaXYg0aZxhEyf0Ekqh/uez9sSAeTDLiNAgXUZamZEAiRQVgKYa8WcK+ZejdazXXV5aWx7zaE8vQgZufDavQQosO7lz9xJgASKSWDv0WRttfBJwaphow3pWVeeGt1aAvx924HErsWTjVh47QEEKLAe0AgsAgmQQNEEsL8V+1ztT3m5bVBDGXd9M8HpLTQS8DQCFFhPaxGWhwRIwEQAnpngoclofpqiTryxudzcP8oYzWsS8CgCFFiPag4WhgRIQCcAX8LwKQzfwkYL1A7xnnxXGxnQtbYxmtck4HEEKLAe1yQsEAmQAE7B+UU7DeewdiqO0UKD/eXV+ztI19bVjNG8JgGPJECB9chmYaFIwHcJ4PxWnOOK81yNVi2iosyY2ElaNKhkjOY1CXgsAQqsxzYNC0YCvkcgOS1H5ml7XBOTs02Vr1czRGZN6iz4pJGAtxCgwHpLS7GcJGBxAnFns5S4pmXkmmqKHit6rujB0kjAmwhQYL2ptVhWErAogejT6bJgZYxkZZsdSGCuFXOumHulkYC3EaDAeluLsbwkYDECB46nyK9rT0juObMDiQHdasvkO9sIVg3TSMAbCVBgvbHVWGYSsAiBbfvPyrKNcCBhFlfsb8U+V+x3pZGAtxKgwHpry7HcJODlBNZuj5P1O+JNtYCewjMTPDTRSMDbCVBgvb0FWX4S8DIC6Kwu/eek7DiYZCo5fAnDpzB8C9NIwAoEKLBWaEXWgQS8hADmWReviZWD0ammEuMUHJyGg1NxaCRgFQIUWKu0JOtBAh5OICtbcyCxIkZiz2SYSorzW3GOK85zpZGAlQhQYK3UmqwLCXgogdT0XLXHNT4py1TCutWDZabmQKJB7VBTPAMkYAUCFFgrtCLrQAIeTCBB88o0b1m0pKTnmErZtF64zHywk9SoEmSKZ4AErEKAAmuVlmQ9SMADCZyIy9D8CsdIpjY8bLROLarIlHEdJTyUf4KMXHhtLQL87bZWe7I2JOAxBA7FpMovazQHEtrJOEbr27mWPD+mrVQMpAMJIxdeW48ABdZ6bcoakYDbCew8lCRL/z4leXYOJG7oW18eGdlC/PzoQMLtjcQCOJ0ABdbpiJkBCfgWgb93xsuabXEFKn3PNU3k7qGNC8QzggSsSoACa9WWZb1IwMUE0Fn9a9Np2bov0ZQzequPj2op1/aqZ4pngASsToACa/UWZv1IwAUEzuWdVw779x9LMeUWpM2z/ufedtK7Y01TPAMk4AsEKLC+0MqsIwk4kUB2Tp46au74qXRTLpW0FcJTx3eUDs2qmOIZIAFfIUCB9ZWWZj1JwAkEcDj6/OXRcibR7ECiVtUgmaHtcW0SGe6EXJkkCXgHAQqsd7QTS0kCHkfgbIrmQOKvaElKNTuQaFgnTGZN6iS1qwV7XJlZIBJwJQEKrCtpMy8SsAiBUwmZMl8T14wsswOJ9k0rq2HhiLBAi9SU1SCB0hOgwJaeHd8kAZ8kcPREmixcFSs5dg4kLu9QQ17QFjQFV/T3SS6sNAnYE6DA2hNhmARIoFACe44ky5L1JyVPWzVstKGXRcqTo1uJPx1IGLHw2scJUGB9/BeA1SeB4hLYuCdBVm4+U+DxO4Y0krHXNpUKdM5UgA0jfJsABda325+1J4FiEYCwQmCN5qcp6kOa28MRV9Q3RvOaBEjgIgEKLH8VSIAECiWAoWAMCWNo2GiBAX7y3N1t5cpLahmjeU0CJGAgQIE1wOAlCZBAPgEsYsJiJixqMlpYcIC89kAH6dKyqjGa1yRAAnYEKLB2QBgkARIQtf0G23CwHcdo1StXlBkTO0nzqErGaF6TAAk4IECBdQCFUSTgywSSNccRP2riCkcSRqtfK1Rmaw4k6tYIMUbzmgRIoBACFNhCwDCaBHyRAFwewvUhXCAarXWjCJk+oaNUqVTRGM1rEiCBIghQYIuAw1sk4EsE4Kx/wcoYgfN+o3VvU11evr+9hATRgYSRC69J4N8IUGD/jRDvk4APEMAxc7+uOyHnzpkdSAzqXkeevqONBPhzk6sP/BqwiuVMgAJbzkCZHAl4GwEckI6D0nFgutFuGdhAJgxvTgcSRii8JoESEKDAlgAWHyUBqxFYsy1O/t4Zb6oWPDJBWCGwNBIggdIToMCWnh3fJAGvJZCndVeX/n1Kdh5KMtUBQ8EYEsbQMI0ESKBsBCiwZePHt0nA6wjknsuTX1afkEMxqaayYxETFjNhURONBEig7AQosGVnyBRIwGsIZGafk5+Wx8iJuAxTmatUCtS24XQSbMehkQAJlA8BCmz5cGQqJODxBFLSc2TesmhJSDY7kIDjCDiQgCMJGgmQQPkRoMCWH0umRAIeSyA+KUvmad6ZUtPNDiTg8nDmg52kWgQdSHhs47FgXkuAAuu1TceCk0DxCMSeyZCfV8QIhoeNBmf9cNoP5/00EiCB8ifA/1nlz5QpkoDHEDgYnSqL18RKrp0DCRwzh+PmcOwcjQRIwDkEKLDO4cpUScDtBHYcTJKl/5zSHEiYPUjggHQclI4D02kkQALOI0CBdR5bpkwCbiOwfke8rN0eZ8ofejr22qZyx5BGpngGSIAEnEOAAuscrkyVBNxCAJ3VZRtPybb9Z035+/tVkCdHt5Khl0Wa4hkgARJwHgEKrPPYMmUScCkBOOpfvPaEHDieYso3uKK/vDi2nVzWvoYpngESIAHnEqDAOpcvUycBlxDIys5TR81Fn0435RcRFihTx3eU9k0rm+IZIAEScD4BCqzzGTMHEnAqARyOjj2ucWezTPnUrhYsMzUHEo3qhJniGSABEnANAQqsazgzFxJwCoFEzSsTxDU5LceUfpPIcJmhOZCoVTXIFM8ACZCA6whQYF3HmjmRQLkSOBmfqfkVjpaMLLMDiQ7Nqqhh4Uqh/O9drsCZGAmUkAD/B5YQGB8nAU8gcCQ2TRatjpWc3DxTcXp3rCn/ubedBAXSgYQJDAMk4AYCFFg3QGeWJFAWArsOJ8sff5+UvDyzA4lre0XK46NaiZ+2JYdGAiTgfgIUWPe3AUtAAsUmsGFXgqzaeqbA83cPbSz3XNOkQDwjSIAE3EeAAus+9syZBIpNAA4kVmw+LZv3JpreQW/1Ec3t4Q1965viGSABEnA/AQqs+9uAJSCBIgmc04aCl6w7KXuPJpueq6jNsz4/pq307VzLFM8ACZCAZxCgwHpGO7AUJOCQQHZOnixcFSPHTpodSISHBMgU7ai5Ti2qOnyPkSRAAu4nQIF1fxuwBCTgkEB6Zq7M/ytGTidmmu7XqBykHEg0rRduimeABEjAswhQYD2rPVgaElAEklJz5MdlxwWfRmtQO1RmPdRZ6mhemmgkQAKeTYAC69ntw9L5IAH0WOdr3pnSM80OJNo0jpDpEzpJ5fBAH6TCKpOA9xGgwHpfm7HEFiaAuVbMuWLu1Wg921WXl8a2l+Agf2M0r0mABDyYAAXWgxuHRfMtAlgljNXCWDVstCE968pTo1tLgD8dSBi58JoEPJ0ABdbTW4jl8wkC2N+Kfa7Y72q02wY1lHHXN5MK1FYjFl6TgFcQoMB6RTOxkFYmAM9M8NBkND9NUSfe2Fxu7h9ljOY1CZCAFxGgwHpRY7Go1iIAX8LwKQzfwkYLDPCTyXe1kQFdaxujeU0CJOBlBCiwXtZgLK41COAUnF+003AOa6fiGC002F9evb+DdG1dzRjNaxIgAS8kQIH1wkZjkb2bAM5vxTmuOM/VaNUiKsqMiZ2kRYNKxmhekwAJeCkBCqyXNhyL7Z0EktNyZJ62xzUxOdtUgXo1Q2TWpM6CTxoJkIA1CFBgrdGOrIUXEIg7m6XENS0j11Ra9FjRc0UPlkYCJGAdAhRY67Qla+LBBKJPp8uClTGSlW12IIG5Vsy5Yu6VRgIkYC0CFFhrtSdr44EEDhxPkV/XnpDcc+ZNrgO61ZbJd7YRrBqmkQAJWI8ABdZ6bcoaeRCBbfvPyrKNcCBhFlfsb8U+V+x3pZEACViTAAXWmu3KWnkAgbXb42T9jnhTSaCn8MwED000EiABaxOgwFq7fVk7NxBAZ3XpPydlx8EkU+7wJQyfwvAtTCMBErA+AQqs9duYNXQhAcyzLl4TKwejU0254hQcnIaDU3FoJEACvkGAAusb7cxauoBAVrbmQGJFjMSeyTDlhvNbcY4rznOlkQAJ+A4BCqzvtDVr6kQCqem5ao9rfFKWKZe61YNlpuZAokHtUFM8AyRAAtYnQIG1fhuzhk4mkKB5ZZq3LFpS0nNMOTWtFy4zH+wkNaoEmeIZIAES8A0CFFjfaGfW0kkETsRlaH6FYyRTGx42WqcWVWTKuI4SHsr/YkYuvCYBXyLA//2+1Nqsa7kSOBSTKr+s0RxIaCfjGK1v51ry/Ji2UjGQDiSMXHhNAr5GgALray3O+pYLgZ2HkmTp36ckz86BxA1968sjI1uInx8dSJQLaCZCAl5MgALrxY3HoruHwN874/+/vTv/jaM84wD+9V7ey7u212uvHSdxnIsk2DHQCHEJqhYhhED0UFVRRKGAuNKitqhV1aotatUWkcIv/StatYBatVU5GgSklJsEIsdJHB+xNz6yXu9lr+21+z5j1p7xxm7iMN5jvq+U7NzHZ0Z+NDPv+7w4cnS8YOcP3tmOB+7YVjCdEyhAAWsKMMBa87rzrNchIA+rhz8Yxcc9E4a15Wn1qXt2464bNxmmc4QCFLC2AAOsta8/z/4iBXLzC1rC/pMDScMa1eo76y8fuhI37Q8bpnOEAhSgAAMs7wEK/B+Bmdl5rau5wZGMYckaVUP4mcf3o3NHrWE6RyhAAQqIAAMs7wMKrCEgnaO/+PpZjE0YE0g01lXjkGrj2t7iX2NtzqIABawswABr5avPc19TIJ5UCSQOn8VkyphAYmvEh+e+14Wmevea63MmBShgbQEGWGtff579KgIjsWm8qILrVNaYQKJje1B7LRzwOVdZk5MpQAEKLAowwPJOoMAKgf5oGn97cxizKxJI3NDZgKdVhSa3y75iDY5SgAIUKBRggC004RQLC3T3JfCv/57DvKo1rC93XN+CH917BexMIKFn4TAFKLCGAAPsGjicZS2B97tjeOPDsYKTvu/2Njx813ZUMTlTgQ0nUIACqwswwK5uwzkWEpDAKgFWX2wqoj6p0h5+7ZZW/WQOU4ACFLgoAQbYi2LiQpUqIK+C5ZWwvBrWF6fDhp8/sA9fvKZRP5nDFKAABS5agAH2oqm4YKUJSCUmqcwklZr0xed24LePdeLq3XX6yRymAAUocEkCDLCXxMWFK0VAmt9IMxxpjqMvoaALhw52YefmGv1kDlOAAhS4ZAEG2Esm4wrlLpBQiSP+ooKrJJLQl9ZGL55XCSSaGzz6yRymAAUosC4BBth1sXGlchWQlIeS+lBSIOrLnrYAnn1iP2prXPrJHKYABSiwbgEG2HXTccViC4yNnMXE+VFMZ1Jo3tyO+oYI7PbVb2lJ1v/XN4Ygyfv15dq9Ifz6kQ54qplAQu/CYQpQ4PIEVv9rdHnb5doUMFVgaOC0CqZ2bNuxD5m06kJOGqlKh62rFOlm7p9vR5HLGZe57doIfnLfXjjsbOS6Ch0nU4AC6xRggF0nHFcrFMjNTkH+9fT0aDPn5ubQ3NyMYDAIm81WuMIFpkxOjCN2fgS1dQ0I1IYu+EQ6PHga1dVu1IYa4XA4kYwOoC7UBLsavlCRDtKlo/SV8febt27BE1/dyQQSF0LjNApQ4LIFGGAvm5AbEAEJrFKcnlrs2rVLG5b/+vr64HQ64fcbu3WbzUxgejqjAqVHBbjFp0d5KnU6XSogh7T15+fnVYBd2pQ2IK+FXS43gvVhLbjGY6Nwu71qWrVxwc/GjhwdxzufnjfMk91JYJUAy0IBClDALIGLe6wwa+/cbsUI2J0eyL+qKuMttfLJdS6bRPp8r8r1O4ex6CCm1PfTBRVIB3q7UROoRTiyCalkHE4VRCXY6ksiHsN8LoeaYL02T8anMmk1XqeWNwbYefW4+rJKILEyuMqr4J/dv4/BVQ/LYQpQwBQBPsGawsqN5gXkKTRfsin1mlYFVk9wE2yOajhUAJ2bncHAyBAaws3w1QSRTExo0x0O460pyyXi46gPR1Dt9qjKTSPat9dQuEWNe/O70H7ncvP4+1tR9A6lDNOlEpNUZpJKTSwUoAAFzBYw/hUze2/cvuUE8q9/z5w5gyqbAy5vSP3aMZdNqKfdMMZHhxHZ1AavbzGxw8hQPxqbN8Pt8RmshtR315AKwl5fAKPqyTeXm0O4qRUu9S1WX6Zncnjp9SFExxdfWefn1dY4VTOcLkhzHBYKUIACGyHAALsRyhbdx8DAAAKBAEZGRuDz+eB0B7XgKhzZ1BiSk3Zs2rpzKbj2nfoUQVW5yec3BsHRc4Oq0lMYHq8fg309Kvh6EQ63at9g9bTJzCxe+PdZxBLGBBKSOEISSEgiCRYKUIACGyXAALtR0hbcz8zMjBZcpRZxfX39UnCdig9ibjqhnlyvWwquqUQcOVXr2OPzG2oDZ1VFKPlGK7WF+08dV6+IVa3kOvUUvOJb7/nJLF5Q2ZlSGWMCCUl5eOjgfoSCxm+0FrwcPGUKUGCDBYw1UjZ459xd5QokEgmk02kVGB2oq6tTtXwXKyxJJafMRD/cwVatclJe4MypT7Tvq/6a2vwk7VdeIUtwldfCjS1bUatqD68MrsNjU/jTK4MFwVWS9f/hh1czuBpEOUIBCmyUAJ9gN0raYvvp7+9X7U4XEIlEVDOaxe+kCwvzmBz6CL7wTlT7w0siPZ++r31PldrB+iIVmeSpNjEZQ6S1TXtFrJ8vw6fPpvCPI8OYW5FAQrqZk+7mpNs5FgpQgALFEGCALYZ6he8zmUwiGo2io6PD0P41OXIc/sZdcPmWg6u0Y5VALDWI5UlVXwZ6T2ivkLe07y6oKSzLfXJ6Eq++O6Ktr19POkiXjtKlw3R9yakmPvF4HOPj46oN7rSW/CLfRjccDqs2uXyNrPfiMAUocHkCDLCX58e1LyDQ3d2Nzs5ONDU1Lc2VgCbNc+wun+EVb2/PMWxt36PawBr7Xj1zUio8hdR32m0FwVXazh7+zwm8d2wAntotcFRLEosFZJNRXNMyjnZPCm+9Oar2U6UFXwngUvKvqyW7lFS+WtlGV1uI/1GAAhT4nAQYYD8nSG5mWeDAgQPLI58NyWtip8cYRLuPvasSS2zWUiLqV5C2sPPzOTSpb65SY1hfMukU/vjCyzjZG9USW8g8af5jUw+rP7j/etx9y3YtsMr+JFcxCwUoQIFiCTDAFkve4vudy6aQzU6pJ9StBVmYTnzyHtpUEn+/em0sZUoF1cH+HqTVq+ePerOI5doQ3rl3SdDtsuPph67EDZ0NS9M4QAEKUKDYAgywxb4CFt1/InoULbffqmoFNxoE5PWvZGYaVnmJzw31aU+yknQi3NyOo9EkJpFRr5mXVwn4nHjm8f3o2L4YjJfncIgCFKBAcQUYYIvrb8m9Z2J9qqJTg3pCNTbJEYx4bAw7dneq9rCLmZ1kmnSOLm1cx+NZGV0qTfVu/F4lkGiLGLM+LS3AAQpQgAJFFGCALSK+VXc9nTwHX6i9oNmN9OsqqQ/tqvedfJlQWZkkuCbSs/lJ2m97ix+HvtuFxjrW/DXAcIQCFCgZAQbYkrkU1jiQ1NgJuAMRlTax8OlV+oKVfmClOzop585Pq7zCZzGVzRlwOnfUaq+Fa7y8fQ0wHKEABUpKgK3wS+pyVPrBLGBWpUh0uPyqNx3dh1R12plUQusbVnrYkdI3nMafX1Pd2a0IrjftD+P5J68Cg6vGxP8oQIESFuAjQAlfnEo7tNT4KXgCLXCopP8ri2RrkkT/0q/r8TMJvPLOOVXBabH9an7Zu25swVP3XKHarxoTSOTn85cCFKBAKQkwwJbS1ajwY1nIzcKuetWx2Ze/scopp5OTWs1heXp973gMb348ViDxwB3b8OCd7QXTOYECFKBAqQowwJbqlamw4xodHYXTW7+UHEJ/emn1ergmUI8jx+L48MSEfpb2tPp9lfbwKze3GqZzhAIUoECpCzDAlvoVqpDjk951nNUqPaHDWOs3qV4Nu1Q719c+jOHkYMZwti6nDb/4zj7cfJWxraxhIY5QgAIUKFEBBtgSvTCVdliSurDKVlinLpFI4q3j0xiOGc/Y73Hgd491omuXMb2icSmOUYACFChdAQbY0r02FXVkra2t6un1hOGchoaG8MoHKUykVc5gXc83DapzdEkgsX2TJPFnoQAFKFCeAgyw5Xndyv+oVQ83Lx0ewFTOq77LLt+GW5q8eE41w4moLE0sFKAABcpZYPkvWzmfBY+97ARmplSPOS4JrsuBdO+2AJ59ogtBv7GWcdmdHA+YAhSggBJggOVtUBQBl6pRrC/XXRnCrx7ugLuaXczpXThMAQqUrwADbPleuw0/8lj/26bs8/brmvHje/fAYWcCCVOAuVEKUKAoAgywRWHnTvMC37ptKx69e4e+jlN+Fn8pQAEKlLUAA2xZX77yPXibqjV88Os78Y0vbS7fk+CRU4ACFFhDgAF2DRzOMkfA6bDhp9/eiy8faDJnB9wqBShAgRIQqFpQpQSOg4dgAYEbH30VXrcdv3mkE1/YY6zkZIHT5ylSgAIWE+ATrMUueDFPtz7gwqGDXdi1paaYh8F9U4ACFNgQAT7Bbggzd0IBClCAAlYTKEwOazUBni8FKEABClDABAEGWBNQuUkKUIACFKAAAyzvAQpQgAIUoIAJAgywJqBykxSgAAUoQAEGWN4DFKAABShAARMEGGBNQOUmKUABClCAAgywvAcoQAEKUIACJggwwJqAyk1SgAIUoAAFGGB5D1CAAhSgAAVMEGCANQGVm6QABShAAQowwPIeoAAFKEABCpggwABrAio3SQEKUIACFGCA5T1AAQpQgAIUMEGAAdYEVG6SAhSgAAUowADLe4ACFKAABShgggADrAmo3CQFKEABClCAAZb3AAUoQAEKUMAEAQZYE1C5SQpQgAIUoAADLO8BClCAAhSggAkC/wM/TIgj3q3zaQAAAABJRU5ErkJggg==)", "_____no_output_____" ] ], [ [ "def build_discriminator(img_shape, verbose=False):\n Xin = Input(shape=(img_shape[0],img_shape[1],img_shape[2],))\n # Convolución2D tensor de 28x28x1 a 14x14x32 y activación Leaky ReLU\n X = Conv2D(filters = 32, \n kernel_size = 3, \n strides = 2, \n input_shape = img_shape, \n padding = 'same')(Xin)\n #X = BatchNormalization()(X)\n X = LeakyReLU(alpha = 0.01)(X)\n\n # Convolución2D tensor de 14x14x32 a 7x7x64, con normalización por lote y activación Leaky ReLU\n X = Conv2D(filters = 64, \n kernel_size = 3, \n strides = 2, \n padding = 'same')(X)\n X = BatchNormalization()(X)\n X = LeakyReLU(alpha = 0.01)(X)\n \n # Convolución2D tensor de 7x7x64 a 3x3x128, con normalización por lote y activación Leaky ReLU\n X = Conv2D(filters = 128, \n kernel_size = 3, \n strides = 2, \n padding = 'same')(X)\n X = BatchNormalization()(X)\n X = LeakyReLU(alpha = 0.01)(X)\n\n # Aplanado del tensor, y capa densa de salida de clasificacion con activación sigmoide\n X = Flatten()(X)\n\n fake = Dense(1, activation = 'linear', name = 'generator')(X) \n # aux = Dense(10, activation = 'softmax', name = 'auxiliary')(X)\n\n \n discriminator_model = Model(inputs = Xin, outputs = [fake], name ='discriminator')\n\n return discriminator_model", "_____no_output_____" ] ], [ [ "## Construye y compila el Discriminador\n", "_____no_output_____" ] ], [ [ "# construye el discriminador \ndiscriminator = build_discriminator(img_shape)\ndiscriminator.summary()", "Model: \"discriminator\"\n_________________________________________________________________\nLayer (type) Output Shape Param # \n=================================================================\ninput_74 (InputLayer) [(None, 32, 32, 3)] 0 \n_________________________________________________________________\nconv2d_144 (Conv2D) (None, 16, 16, 32) 896 \n_________________________________________________________________\nleaky_re_lu_190 (LeakyReLU) (None, 16, 16, 32) 0 \n_________________________________________________________________\nconv2d_145 (Conv2D) (None, 8, 8, 64) 18496 \n_________________________________________________________________\nbatch_normalization_142 (Bat (None, 8, 8, 64) 256 \n_________________________________________________________________\nleaky_re_lu_191 (LeakyReLU) (None, 8, 8, 64) 0 \n_________________________________________________________________\nconv2d_146 (Conv2D) (None, 4, 4, 128) 73856 \n_________________________________________________________________\nbatch_normalization_143 (Bat (None, 4, 4, 128) 512 \n_________________________________________________________________\nleaky_re_lu_192 (LeakyReLU) (None, 4, 4, 128) 0 \n_________________________________________________________________\nflatten_50 (Flatten) (None, 2048) 0 \n_________________________________________________________________\ngenerator (Dense) (None, 1) 2049 \n=================================================================\nTotal params: 96,065\nTrainable params: 95,681\nNon-trainable params: 384\n_________________________________________________________________\n" ] ], [ [ "# Crítico Wasserstein", "_____no_output_____" ], [ "De manera similar al generador, necesitamos que el crítico conozca las etiquetas generadas, para ellos se las damos de manera similar a como se hizo en el generador", "_____no_output_____" ], [ "![discriminador.jpg](data:image/jpeg;base64,/9j/4AAQSkZJRgABAQEASABIAAD/4SdsRXhpZgAATU0AKgAAAAgABgALAAIAAAAmAAAIYgESAAMAAAABAAEAAAExAAIAAAAmAAAIiAEyAAIAAAAUAAAIrodpAAQAAAABAAAIwuocAAcAAAgMAAAAVgAAEUYc6gAAAAgAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAFdpbmRvd3MgUGhvdG8gRWRpdG9yIDEwLjAuMTAwMTEuMTYzODQAV2luZG93cyBQaG90byBFZGl0b3IgMTAuMC4xMDAxMS4xNjM4NAAyMDIxOjA1OjIzIDIzOjI3OjU1AAAGkAMAAgAAABQAABEckAQAAgAAABQAABEwkpEAAgAAAAMyNQAAkpIAAgAAAAMyNQAAoAEAAwAAAAEAAQAA6hwABwAACAwAAAkQAAAAABzqAAAACAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAMjAyMTowNToyMyAyMjozODowOQAyMDIxOjA1OjIzIDIyOjM4OjA5AAAAAAYBAwADAAAAAQAGAAABGgAFAAAAAQAAEZQBGwAFAAAAAQAAEZwBKAADAAAAAQACAAACAQAEAAAAAQAAEaQCAgAEAAAAAQAAFb8AAAAAAAAAYAAAAAEAAABgAAAAAf/Y/9sAQwAIBgYHBgUIBwcHCQkICgwUDQwLCwwZEhMPFB0aHx4dGhwcICQuJyAiLCMcHCg3KSwwMTQ0NB8nOT04MjwuMzQy/9sAQwEJCQkMCwwYDQ0YMiEcITIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIy/8AAEQgBAADxAwEhAAIRAQMRAf/EAB8AAAEFAQEBAQEBAAAAAAAAAAABAgMEBQYHCAkKC//EALUQAAIBAwMCBAMFBQQEAAABfQECAwAEEQUSITFBBhNRYQcicRQygZGhCCNCscEVUtHwJDNicoIJChYXGBkaJSYnKCkqNDU2Nzg5OkNERUZHSElKU1RVVldYWVpjZGVmZ2hpanN0dXZ3eHl6g4SFhoeIiYqSk5SVlpeYmZqio6Slpqeoqaqys7S1tre4ubrCw8TFxsfIycrS09TV1tfY2drh4uPk5ebn6Onq8fLz9PX29/j5+v/EAB8BAAMBAQEBAQEBAQEAAAAAAAABAgMEBQYHCAkKC//EALURAAIBAgQEAwQHBQQEAAECdwABAgMRBAUhMQYSQVEHYXETIjKBCBRCkaGxwQkjM1LwFWJy0QoWJDThJfEXGBkaJicoKSo1Njc4OTpDREVGR0hJSlNUVVZXWFlaY2RlZmdoaWpzdHV2d3h5eoKDhIWGh4iJipKTlJWWl5iZmqKjpKWmp6ipqrKztLW2t7i5usLDxMXGx8jJytLT1NXW19jZ2uLj5OXm5+jp6vLz9PX29/j5+v/aAAwDAQACEQMRAD8A9/ooAKD0oA5208Y2E9rFcT219aRSorxvLbllZSMg7k3Ade+K2LPUrHUE3WV5BcDv5UgbH5Vcqco6kqSZaoqCgooAKKACigAooAKKACigAooAKKACigAooAKKACigAooAKD0oA5Hw7/yLOk/9ecP/AKAKmutKsLx99xZwvJ2k24cfRhyPzrpu09DK10MWwurbmw1a9gA6RyOJ0z7+YC2PYMKnXU9etv8AW29jfL6xM0DD/gJ3A/mKTUZb6DTaJ18U2sf/AB/Wt7ZY6tLCXQf8CTcB+JFadnqVjqCb7K8guF7mKQNj8qylTktSlJMtUVBQUUAFFABRQAUUAFFABRQAUUAFFABRQAUUAFFABQelAHI+Hf8AkWdJ/wCvOH/0AVpV0PcyWwUUhhVO60qwvX8y4s4ZJcYEhQb1+jdR+BpptbCsMSwurX/jx1e+hH9yVxOp/wC/mW/Iip11PXrb/W29lfKO8TNA/wCR3A/mKGoy30HdonXxTbR8X1pfWRAyWkgLoPcum5QPqRWlZanY6lGZLG9t7lVOCYZA+PrispU5R1KUky3RUFBRQAUUAFFABRQAUUAFFABRQAUUAFFABQelAHI+Hf8AkWdJ/wCvOH/0AVpV0PcyWwUUhhRQAUUAFU7rSdPvXElzZwySDpIUG4fRuopptbCsMWwurbmx1e9h9Elfz0/8fyfyIqddS162/wBZb2V8gHWNmgc/RTuBP/AhQ1GW+gXaJ08VWqcX1nfWTf8ATWAuv13R7lH4kVp2epWOoqWsry3uVHUwyBsfXFZSpta9C1JMtUVBQUUAFFABRQAUUAFFABRQAUUAFB6UAcj4d/5FnSf+vOH/ANAFaVdD3MlsFFIYUUAFFABRQAUUAFU7rSrC9bfcWkLyDpJtw4+jDkU02thNXGLYXVt/x46tew4+6kjidPx3gtj6MKnXUtetuJLexvl7tG7QN/3ydwP/AH0KGoy30BNonXxTap/x/Wl9ZY/ilh3r/wB9R7gPxIrSs9SsdQXdZ3lvcD1ikDY/KspU2tS1JMt0VBQUUAFFABRQAUUAFFABQelAHI+Hf+RZ0n/rzh/9AFaVdD3MlsFFIYUUAFFABRQAUUAFFABRQAVTudK0+8ffcWcLydpCgDj6MOR+BpptbCsMWwurb/jx1e+gUf8ALORxOp/7+Atj2DCp11PXrb/W29jer6xM0Dfk24H8xQ1GW+gJtE6+KbWPi+tL6yI6tLCXQf8AA03KPxIrUstRstSiMtldw3MYOC0UgYA+hxWUqbjr0LUkyzRUFBRQAUUAFFABQelAHI+Hf+RZ0n/rzh/9AFaVdD3MlsFFIYUUAFFABRQAUUAFFABRQAUUAFFABUGgf8h/WvpB/wCgtQ/hYLco+OfFOoeG209bCO2b7T5m8zozY27cYww/vGuSPxP8QLyYNMIzyBBJ/wDHK6KGEjUp87ZnUrOMuU9dpa4DoCigAooAKD0oA5Hw7/yLOk/9ecP/AKAK0q6HuZLYKKQwooAKKACigAooAKKACigAooAKKACodB/5D+tfSD/0FqH8LBbnLfFj/W6L9J//AGnXnL/cr08H/u/3nLW/iH0jRXincFFABRQAUHpQByPh3/kWdJ/684f/AEAVpV0PcyWwUUhhRQAUUAFFABRQAUUAFFABRQAUUAFQ6D/yH9a+kH/oLUP4WC3OW+LH+t0X6T/+0685f7leng/93+85a38Q+kaK8U7gooAKKACg9KAOR8O/8izpP/XnD/6AK0q6HuZLYKKQwooAKKACigAooAKKACigAooAKKACodB/5D+tfSD/ANBah/CwW5y3xY/1ui/Sf/2nXnL/AHK9PB/7v95y1v4h9I0V4p3BRQAUUAFB6UAcj4d/5FnSf+vOH/0AVpV0PcyWwUUhhRQAUUAFFABRQAUUAFFABRQAUUAFQ6D/AMh/WvpB/wCgtQ/hYLc5b4sf63RfpP8A+0685f7leng/93+85a38Q+kaK8U7gooAKKACg9KAOR8O/wDIs6T/ANecP/oArSroe5ktgopDCigAooAKKACigAooAKKACigAooAKh0H/AJD+tfSD/wBBah/CwW5y3xY/1ui/Sf8A9p15y/3K9PB/7v8Aectb+IfSNFeKdwUUAFFABQelAHI+Hf8AkWdJ/wCvOH/0AVpV0PcyWwUUhhRQAUUAFFABRQAUUAFFABRQAUUAFQ6D/wAh/WvpB/6C1D+Fgtzlvix/rdF+k/8A7Trzl/uV6eD/AN3+85a38Q+kaK8U7gooAKKACg9KAOR8O/8AIs6T/wBecP8A6AK0q6HuZLYKKQwooAKiubmCytZbm5mSGCJS8kjthVUdSTQBg6Z450PVr21tbaS6VrwE2rzWcsST4GTsZlAPHNbGp6hHpWnTXssVxKkQBKW8Rkc5IHCjk9aLhYeL61OoGwFxH9sEQnMG75/LJxux6ZBFR6bqMWqWpniiuIlEjR7biFo2ypxnB5x6GgC5RQAUUAFFABRQAVDoP/If1r6Qf+gtQ/hYLc5b4sf63RfpP/7Trzl/uV6eD/3f7zlrfxD6RorxTuCigAooAKD0oA5Hw7/yLOk/9ecP/oArSroe5ktgopDCigArjPinn/hA7kuM2q3Fu12AMkwiVd2P0/DNJ7Atzo572y/s6NobmALNETalXHz/ACkjZ68c8dq8rS2u7P4K3HiI6zqk2qXVlHule6fCL5igBRnggDGepyc9aTGjfOh203xrd2uL4MNIjuwFu5FBYTkYwD9zgfJ93npzWPaajf3uiaJp0+pXkNvqGvXUFzdJOVfYrSFYg+cjcQBx6YoA07zTo4vFuleE49b1MaVMtxcSqb9jI0iqmIN+d4UA79uc/N6YrZ8EyzQ6r4k0gX9xfWOn3aJbTXEpkdd0YZoy55O08c8ijqB2NFUIKKACigAqDQiBr+s5IGfIA/75ah/CwW5y/wAWP9bov0n/APadecv9yvTwf+7/AHnLW/iH0jRXincFFABRQAUHpQByPh3/AJFnSf8Arzh/9AFaVdD3MlsFFIYUUAFMlijnheGaNZI3Uq6OuQwPUEHqKAMHTfA3hjSL1ryw0a2huCCocAkqDwduT8uQT0xV5vD2lPoA0JrJDpgQRi3ycbQcgZznqPWlYLjdQ8NaPquo2eoX1hHNeWZBgmJIZMHcOh5AIzg0yXwroc+kPpU2mwvYvK0xhbJG8ksWBzkHJPI9aLAVm8C+GG0pNNOjW32VJTMq87g56tuzuycDnPYelaml6Rp+iWK2WmWkVrbqSQkY6k9Se5PuaLBcu0UwCigAooA5m6svFzXczW2qWiQFyY1ZRkLngH5PSsK807xPN/aMTXKXDh4POSBfmc87CuFHTnPSu2nOgmtDCSqdyl4qtddtLTSV1y4ErES+Spbc6D5M7j37dzXNP9w110XB024bamM01L3tz6RorwD0QooAKKACg9KAOR8O/wDIs6T/ANecP/oArSroe5ktgopDCigAooAKKACigAooAKKACigAooAKg0H/AJD+tfSD/wBBah/CwW5y/wAWP9bov0n/APadecv9yvTwf+7/AHnLW/iH0jRXincFFABRQAUHpQByPh3/AJFnSf8Arzh/9AFaVdD3MlsFFIYUUAFFABRQAUUAFFABRQAUUAFFABUOg/8AIf1r6Qf+gtQ/hYLc5b4sf63RfpP/AO0685f7leng/wDd/vOWt/EPpGivFO4KKACigAoPSgDkfDv/ACLOk/8AXnD/AOgCtKuh7mS2CikMKKACigAooAKKACigAooAKKACigAqHQf+Q/rX0g/9Bah/CwW5y3xY/wBbov0n/wDadecv9yvTwf8Au/3nLW/iH0jRXincFFABRQAUHpQByPh3/kWdJ/684f8A0AVpV0PcyWwUUhhRQAUUAFFABRQAUUAFFABRQAUUAFQ6D/yH9a+kH/oLUP4WC3OW+LH+t0X6T/8AtOvOX+5Xp4P/AHf7zlrfxD6RorxTuCigAooAKD0oA5Hw7/yLOk/9ecP/AKAK0q6HuZLYKKQwooAKKACigAooAKKACigAooAKKACodB/5D+tfSD/0FqH8LBbnLfFj/W6L9J//AGnXnL/cr08H/u/3nLW/iH0jRXincFFABRQAUHpQByPh3/kWdJ/684f/AEAVpV0PcyWwUUhhRQAUUAFFABRQAUUAFFABRQAUUAFQ6D/yH9a+kH/oLUP4WC3OW+LH+t0X6T/+0685f7leng/93+85a38Q+kaK8U7gooAKKACg9KAOR8O/8izpP/XnD/6AK0q6HuZLYKKQwooAKKACigAooAKKACigAooAKKACodB/5D+tfSD/ANBah/CwW5y3xY/1ui/Sf/2nXnL/AHK9TB/7v95y1v4h9I0V4h3BRQAUUAFB6UAcj4d/5FnSf+vOH/0AVpV0PcyWwUUhhRQAUUAFFABRQAUUAFFABRQAUUAFQ6D/AMh/WvpB/wCgtQ/hYLcu6z4c0vX/ACP7SgaXyN3l7ZWTG7GfukegrK/4V14Yzn7DL+N1L/8AFUQxFSEeWL0CVKMndnU0tYGgUUAFFABQelAHI+Hf+RZ0n/rzh/8AQBWlXQ9zJbBRSGFFABRQAUUAFFABRQAUUAFFABRQAVDoH/If1r6Qf+gtQ/hYLc6KisDQKKACigArxvVvGfiSDW9Rgh1V44oruaONBBEdqq7ADlM9BXXg6MasmpGNabgk0ei+C9Qu9V8KWl5fTGa4dpQ0m0LnbIyjgADoBW+elc9SKjNpdGaRd4pnI+Hf+RZ0n/rzh/8AQBWlWr3IWwUUhhRQAUUAFFABRQAUUAFFABRQBHNPDbR+ZPNHEg/idgo/M1SXWrafIsY7m+bt9lhZlP8AwPhP/HqpRdriuTLFr10f3Vhb2cZ/ju5tzj32JkH/AL7FaOj6RLp013cXF2Lme5K7isXlqAoIAAyfX1rOco2sikne7NaisiwooAKKACvn7Wv+Rh1b/r+uP/RrV6GXfxH6HNifhR6z8O/+RHsf9+b/ANHPXUHpXJW/iy9WbQ+BHI+Hf+RZ0n/rzh/9AFaVW9yVsFFIYUUAFFABRQAUUAFFAEc1xDbRmSeaOKMdWkYKB+Jqiut20/FjHc3xzjNrCzr/AN9/d/WqUeork6xa9dcRWFtZqej3U29h/wAATj/x+p18OXU3N/rNw2esVoiwIfx5f8QwqXUittRqLe5btvDej2snmJYRSS9pZ8yuP+BPk/rWrWLk5blpJbBRSGFFABRQAUUAFfP2tf8AIw6t/wBf1x/6NavQy7+I/Q5sT8KPT/A2pWNj4J09bu9trdmecqJpVTP75+mTXQHxDouP+Qvp/wD4Ep/jXPVpzdSTS6s1hJKK1OT8K61YXOkWNik6rdQW6RNG/BJVQDj16dq6KrqQlCVmTGSktAorMoKKACigAooAjnnhtoWmuJY4ol+88jBQPxNUV1u2n/48Y7m/ycBrSFnTPpv+5+tUovcVydYteuf9XYW9mufvXU29v++UyP8Ax6p18O3U3/H9rNw6nrHaosC/ny4/76qXUittRqLe5btvDej2siyrYRSTqciafMsn/fb5P61q1lKTluWklsFFSMKKACigAooAKKACigAr5+1r/kYdW/6/rj/0a1ehl38R+hzYn4Uei+EPDula34L02TULXzmjacIfMZcAzP6EVsHwF4ax/wAg3/yPJ/8AFVlUxFWE3GL0uyo0oOKbRy/hrwnp506z1C5BuJJ4kmCuPlXcAcY79e9dhVVqrqSuwhBRWgUViWFFAEc9xDbRmS4mjijHVpGCj8zVEa3bTkrYx3N83QfZYWdT/wAD+5+bVSi3qK5OkWvXQzFYW1mh73c25x/wBMj/AMfqdfDtzN/x/azcOM/ctUEC/ny3/j1S6kY7ajUW9y3beG9HtZhMlhE84ORNPmWTP+82T+tamAOlZSk5astJLYWipGFFABRQAUUAFFABRQAUUAFFABXz9rX/ACMOrf8AX9cf+jWr0Mu/iP0ObE/Cj1n4d/8AIj2P+/N/6OeuoPSuSt/Fl6s2h8COR8O/8izpP/XnD/6AK0qt7krYjmnht4/MnljiQfxOwUfmaorrdtPxYx3N82f+XWEuv/ffC/rTUeoXJ1i166/1Vhb2aHo91Nvdf+AJkH/vsVOnh26mGb7WblueUtEWBP8A2Zv/AB6pdSMdtRqLe5btvDej2snmpYRPN/z1mzK//fT5P61qYAGBWUpOW5aSWwtFSMKKACigAooAKKACigAooAKKACigAooAK+fta/5GHVv+v64/9GtXoZd/Efoc2J+FHrPw7/5Eex/35v8A0c9dOelclb+LL1ZtD4EcdpGn+IYtHsrP7Fa2jQQRxM9xPvOVUA4VOD0/vCtJPDt1Oc32s3DA/eitUECfny4/BqqVSKemolF9S3b+GtHtn8xbCKSUdJZ8yv8A99Pk/rWpWUpOW5aSWwtFSMKKACigAooAKKACigAooAKKACigAooAKKACigArzDUfhtqt3qt7dRXtksdxcSTKG3ZAZy2Dx7104auqMm2jKrT51Y7fwtpE2heHbbTp5I5JYjIWaPO07nZuM/WtisZy5pOXc0irJIKKgYUUAFFABRQAUUAFFABRQAUUAFFABRQAUUAFFABRQAUUAf/ZAP/hMsVodHRwOi8vbnMuYWRvYmUuY29tL3hhcC8xLjAvADw/eHBhY2tldCBiZWdpbj0n77u/JyBpZD0nVzVNME1wQ2VoaUh6cmVTek5UY3prYzlkJz8+DQo8eDp4bXBtZXRhIHhtbG5zOng9ImFkb2JlOm5zOm1ldGEvIj48cmRmOlJERiB4bWxuczpyZGY9Imh0dHA6Ly93d3cudzMub3JnLzE5OTkvMDIvMjItcmRmLXN5bnRheC1ucyMiPjxyZGY6RGVzY3JpcHRpb24gcmRmOmFib3V0PSJ1dWlkOmZhZjViZGQ1LWJhM2QtMTFkYS1hZDMxLWQzM2Q3NTE4MmYxYiIgeG1sbnM6TWljcm9zb2Z0UGhvdG89Imh0dHA6Ly9ucy5taWNyb3NvZnQuY29tL3Bob3RvLzEuMC8iPjxNaWNyb3NvZnRQaG90bzpJdGVtU3ViVHlwZT5MdW1pYS5MaXZpbmdJbWFnZTwvTWljcm9zb2Z0UGhvdG86SXRlbVN1YlR5cGU+PC9yZGY6RGVzY3JpcHRpb24+PHJkZjpEZXNjcmlwdGlvbiByZGY6YWJvdXQ9InV1aWQ6ZmFmNWJkZDUtYmEzZC0xMWRhLWFkMzEtZDMzZDc1MTgyZjFiIiB4bWxuczp4bXA9Imh0dHA6Ly9ucy5hZG9iZS5jb20veGFwLzEuMC8iPjx4bXA6Q3JlYXRlRGF0ZT4yMDIxLTA1LTIzVDIyOjM4OjA5LjI0NzwveG1wOkNyZWF0ZURhdGU+PHhtcDpDcmVhdG9yVG9vbD5XaW5kb3dzIFBob3RvIEVkaXRvciAxMC4wLjEwMDExLjE2Mzg0PC94bXA6Q3JlYXRvclRvb2w+PC9yZGY6RGVzY3JpcHRpb24+PC9yZGY6UkRGPjwveDp4bXBtZXRhPg0KICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAKICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgIAogICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgICAgPD94cGFja2V0IGVuZD0ndyc/Pv/bAEMAAwICAwICAwMDAwQDAwQFCAUFBAQFCgcHBggMCgwMCwoLCw0OEhANDhEOCwsQFhARExQVFRUMDxcYFhQYEhQVFP/bAEMBAwQEBQQFCQUFCRQNCw0UFBQUFBQUFBQUFBQUFBQUFBQUFBQUFBQUFBQUFBQUFBQUFBQUFBQUFBQUFBQUFBQUFP/AABEIAdcBuwMBIgACEQEDEQH/xAAfAAABBQEBAQEBAQAAAAAAAAAAAQIDBAUGBwgJCgv/xAC1EAACAQMDAgQDBQUEBAAAAX0BAgMABBEFEiExQQYTUWEHInEUMoGRoQgjQrHBFVLR8CQzYnKCCQoWFxgZGiUmJygpKjQ1Njc4OTpDREVGR0hJSlNUVVZXWFlaY2RlZmdoaWpzdHV2d3h5eoOEhYaHiImKkpOUlZaXmJmaoqOkpaanqKmqsrO0tba3uLm6wsPExcbHyMnK0tPU1dbX2Nna4eLj5OXm5+jp6vHy8/T19vf4+fr/xAAfAQADAQEBAQEBAQEBAAAAAAAAAQIDBAUGBwgJCgv/xAC1EQACAQIEBAMEBwUEBAABAncAAQIDEQQFITEGEkFRB2FxEyIygQgUQpGhscEJIzNS8BVictEKFiQ04SXxFxgZGiYnKCkqNTY3ODk6Q0RFRkdISUpTVFVWV1hZWmNkZWZnaGlqc3R1dnd4eXqCg4SFhoeIiYqSk5SVlpeYmZqio6Slpqeoqaqys7S1tre4ubrCw8TFxsfIycrS09TV1tfY2dri4+Tl5ufo6ery8/T19vf4+fr/2gAMAwEAAhEDEQA/AP1TooooAKKKKACiiigAqJZVkZgGBKnDDPIOAcH8CPzqSvjbU/hvD8QP2oPjDI2u65oF5pzaM1vdaHfG2kG+yG7dgENyidR29zXRQoqs2m7WRjUm4JWVz7Ipa+ZLfR/jX4Nx/YfxC07xXbBsiz8U6ftYKO3nQncxOB1x3NaNv+0b488MYXxj8JtSlhDY+3eFblL9WHdvKyGUcE8n0rZ4Of2Gn8/8yVWj9pWPorNFeOeGv2tvhd4kuPsr+JY9Dv1YK9prkT2ToTxgmQBev+12Nes6bqlnq9qtzY3cF5bt92a3kEiH6FeK5ZU50/jVjWMoy2ZapaSlrMsKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAK+Y/Cf/J0Hxz/AO4H/wCkTV9OV8x+E/8Ak6D45/8AcD/9Imruwu8vT9Uc1b7Pqeo0UUV0GZl694W0XxTbiDWdJsdWhHSO+tkmUfTcDivObj9mPwfbXTXfhybWPBV8zbjceHtSltyT2ypLLj2AFetUVpGpOOzJ5Ys8ut9H+Nfg3H9h/ELTvFdsGyLPxTp+1go7edCdzE4HXA6mtK3/AGjvHnhjC+MfhNqU0IbH27wrcpfqw7t5XDKOCeT6V39FJuEvjgn+H5AuaPwv9TF8NftbfC7xJcfZX8Sx6HfqwV7TXIXspEJ4wTIAvX/a7GvWdN1Sz1i1W5sbuC9t2+7NbyLIh+hHFeWa94W0XxTbiDWdJsdWhHSO+tkmUfTcDivObj9mPwfbXTXfhybWPBV8zbjceHtSlt8ntlSWXHsAKxeHoy+FtfiX7Sa3Vz6jFLXzJb6P8a/BuP7D+IWneK7YNkWfinT9rBR286E7mJwOuO5rRt/2jfHnhjC+MfhNqUsIbH27wrcpfqw7t5WQyjgnk+lZPBz+w0/68zRV49dD6KzRXjnhr9rb4XeJLj7K/iWPQ79WCvaa5E9k6E8YJkAXr/tdjXrOm6pZ6varc2N3BeW7fdmt5BIh+hXiuWVOdP41Y1jKMtmWqWkpazLCiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAK+Y/Cf/J0Hxz/7gf8A6RNX05XzH4T/AOToPjn/ANwP/wBImruwu8vT9Uc1b7Pqeo0UUVuZhRRRQAUUUUAFFFFABRRRTAy9e8LaL4ptxBrOk2OrQjpHfWyTKPpuBxXnNx+zH4Ptrprvw5NrHgq+ZtxuPD2pS25J7ZUllx7ACvWqK0jUnHZk8sWeXW+j/Gvwbj+w/iFp3iu2DZFn4p0/awUdvOhO5icDrgdTWlb/ALR3jzwxhfGPwm1KaENj7d4VuUv1Yd28rhlHBPJ9K7+ik3CXxwT/AA/IFzR+F/qYvhr9rb4XeJLj7K/iWPQ79WCvaa5C9lIhPGCZAF6/7XY16zpuqWesWq3NjdwXtu33ZreRZEP0I4ryzXvC2i+KbcQazpNjq0I6R31skyj6bgcV5zcfsx+D7a6a78OTax4Kvmbcbjw9qUtvk9sqSy49gBWLw9GXwtr8S/aTW6ufUYpa+ZLfR/jX4Nx/YfxC07xXbBsiz8U6ftYKO3nQncxOB1x3NaNv+0b488MYXxj8JtSlhDY+3eFblL9WHdvKyGUcE8n0rJ4Of2Gn/XmaKvHrofRWaK8c8NftbfC7xJcfZX8Sx6HfqwV7TXInsnQnjBMgC9f9rsa9Z03VLPV7Vbmxu4Ly3b7s1vIJEP0K8Vyypzp/GrGsZRlsy1S0lLWZYUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFfMfhP/k6D45/9wP8A9Imr6cr5j8J/8nQfHP8A7gf/AKRNXdhd5en6o5q32fU9RooorczCiiigAooooAKKKKACiiigAooooAKKKKACiiigAooopgZeveFtF8U24g1nSbHVoR0jvrZJlH03A4rzm4/Zj8H210134cm1jwVfM243Hh7Upbck9sqSy49gBXrVFaRqTjsyeWLPLrfR/jX4Nx/YfxC07xXbBsiz8U6ftYKO3nQncxOB1wOprSt/2jvHnhjC+MfhNqU0IbH27wrcpfqw7t5XDKOCeT6V39FJuEvjgn+H5AuaPwv9TF8NftbfC7xJcfZX8Sx6HfqwV7TXIXspEJ4wTIAvX/a7GvWdN1Sz1i1W5sbuC9t2+7NbyLIh+hHFeWa94W0XxTbiDWdJsdWhHSO+tkmUfTcDivObj9mPwfbXTXfhybWPBV8zbjceHtSlt8ntlSWXHsAKxeHoy+FtfiX7Sa3Vz6jFLXzJb6P8a/BuP7D+IWneK7YNkWfinT9rBR286E7mJwOuO5rRt/2jfHnhjC+MfhNqUsIbH27wrcpfqw7t5WQyjgnk+lZPBz+w0/68zRV49dD6KzRXjnhr9rb4XeJLj7K/iWPQ79WCvaa5E9k6E8YJkAXr/tdjXrOm6pZ6varc2N3BeW7fdmt5BIh+hXiuWVOdP41Y1jKMtmWqWkpazLCiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACvmPwn/wAnQfHP/uB/+kTV9OV8x+E/+ToPjn/3A/8A0iau7C7y9P1RzVvs+p6jRRRW5mFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUwMvXvC2i+KbcQazpNjq0I6R31skyj6bgcV5zcfsx+D7a6a78OTax4Kvmbcbjw9qUtuSe2VJZcewAr1qitI1Jx2ZPLFnl1vo/wAa/BuP7D+IWneK7YNkWfinT9rBR286E7mJwOuB1NaVv+0d488MYXxj8JtSmhDY+3eFblL9WHdvK4ZRwTyfSu/opNwl8cE/w/IFzR+F/qYvhr9rb4XeJLj7K/iWPQ79WCvaa5C9lIhPGCZAF6/7XY16zpuqWesWq3NjdwXtu33ZreRZEP0I4ryzXvC2i+KbcQazpNjq0I6R31skyj6bgcV5zcfsx+D7a6a78OTax4Kvmbcbjw9qUtvk9sqSy49gBWLw9GXwtr8S/aTW6ufUYpa+ZLfR/jX4Nx/YfxC07xXbBsiz8U6ftYKO3nQncxOB1x3NaNv+0b488MYXxj8JtSlhDY+3eFblL9WHdvKyGUcE8n0rJ4Of2Gn/AF5mirx66H0VmivHPDX7W3wu8SXH2V/Eseh36sFe01yJ7J0J4wTIAvX/AGuxr1nTdUs9XtVubG7gvLdvuzW8gkQ/QrxXLKnOn8asaxlGWzLVLSUtZlhRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFfMfhP/AJOg+Of/AHA//SJq+nK+Y/Cf/J0Hxz/7gf8A6RNXdhd5en6o5q32fU9RooorczCiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiimBl694W0XxTbiDWdJsdWhHSO+tkmUfTcDivObj9mPwfbXTXfhybWPBV8zbjceHtSltyT2ypLLj2AFetUVpGpOOzJ5Ys8ut9H+Nfg3H9h/ELTvFVsGyLPxTp+1go7edCdzE4HXA6mtK3/aO8eeGML4x+E2pSwhsG+8K3KX6sO7eVwwHBPJ9K7+j9aTcJ/HFfl+QLmj8LKvw8/aY8AfEzXItD0rVpbbX5N23SdRtZLe4O1SzABhgkBWJAJ+6a9Vr5i8Wgf8ADUXwNbAz/wATwE+v+hCvpznNcGIpxpyXJs0dNKTkncWk59a5H4qfErTPhF4D1PxbrEF3c6dp/lebHYorTN5kqRLtDMoPzOCckcZ+leD/APDw7wH/ANCr4y/8F9v/APJFFHC1q65qcG0E61Om7Sdj6mpBXy3/AMPDvAX/AEKvjP8A8F8H/wAfr2P4J/GvRfjt4XvNd0Ky1GxtrW+ewki1OJI5fMVEckBHcYxIO+cg8U6uEr0Y81SDSCFanUdou56FzS02nVyGwUUUUAFFFFABRRRQAUUUUAFfMfhP/k6D45/9wP8A9Imr6cr5j8J/8nQfHP8A7gf/AKRNXdhd5en6o5q32fU9RooorczCiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKAPLvFn/J0HwM/7jn/pEtfTZr5k8Wf8nQfAz/uOf+kS19NmsMXvH0/zNaP2vU8K/bi/5Nf8Z/8Abl/6WwV+e9foR+3F/wAmv+M/+3L/ANLYK/PevteGf4E/X/I8LNP4i9Ar7Z/4J4f8kj8U/wDY03X/AKT21fE1fbP/AATw/wCSReKf+xpuv/Se2rp4j/3Repnlv8d+h9SfxUtJ/FS1+Zn1IUUUUAFFFFABRRRQAUUUUAFfMfhP/k6D45/9wP8A9Imr6cr5j8J/8nQfHP8A7gf/AKRNXdhd5en6o5q32fU9RooorczCiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKAPLvFn/J0HwM/7jn/pEtfTZr5k8Wf8nQfAz/uOf+kS19NmsMX8UfT/ADNaP2vU8K/bi/5Nf8Z/9uX/AKWwV+e9foR+3F/ya/4z/wC3L/0tgr896+14Z/gT9f8AI8HNP4i9Ar7Z/wCCeH/JIvFP/Y03X/pPbV8TV9s/8E8P+SReKf8Asabr/wBJ7auniP8A3RepGW/x36H1J/FS0n8VLX5mfUhRRRQAUUUUAFFFFABRRRQAV8x+E/8Ak6D45/8AcD/9Imr6cr5j8J/8nQfHP/uB/wDpE1d2F3l6fqjmrfZ9T1GiiitzMKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooA8u8Wf8AJ0HwM/7jn/pEtfTZr5k8Wf8AJ0HwM/7jn/pEtfTZrDF/FH0/zNaP2vU8K/bi/wCTX/Gf/bl/6WwV+e9foR+3F/ya/wCM/wDty/8AS2Cvz3r7Xhn+BP1/yPBzT+IvQK+2f+CeH/JIvFP/AGNN1/6T21fE1fbP/BPD/kkXin/sabr/ANJ7auniP/dF6kZb/HfofUn8VLSfxUtfmZ9SFFFFABRRRQAUUUUAFFFFABXzH4T/AOToPjn/ANwP/wBImr6cr5j8J/8AJ0Hxz/7gf/pE1d2F3l6fqjmrfZ9T1GiiitzMKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooA8u8Wf8nQfAz/ALjn/pEtfTZr5k8Wf8nQfAz/ALjn/pEtfTZrDF/FH0/zNaP2vU8K/bi/5Nf8Z/8Abl/6WwV+e9foR+3F/wAmv+M/+3L/ANLYK/PevteGf4E/X/I8HNP4i9Ar7Z/4J4f8ki8U/wDY03X/AKT21fE1fbP/AATw/wCSReKf+xpuv/Se2rp4j/3RepGW/wAd+h9SfxUtJ/FS1+Zn1IUUUUAFFFFABRRRQAUUUUAFfMfhP/k6D45/9wP/ANImr6cr5j8J/wDJ0Hxz/wC4H/6RNXdhd5en6o5q32fU9RooorczCiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKAPLvFn/J0HwM/7jn/AKRLX02a+ZPFn/J0HwM/7jn/AKRLX02awxfxR9P8zWj9r1PCv24v+TX/ABn/ANuX/pbBX571+hH7cX/Jr/jP/ty/9LYK/PevteGf4E/X/I8HNP4i9Ar7Z/4J4f8AJIvFP/Y03X/pPbV8TV9s/wDBPD/kkXin/sabr/0ntq6eI/8AdF6kZb/HfofUn8VLSfxUtfmZ9SFFFFABRRRQAUUUUAFFFFABXzH4T/5Og+Of/cD/APSJq+nK+Y/Cf/J0Hxz/AO4H/wCkTV3YXeXp+qOat9n1PUaKKK3MwooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigDy7xZ/ydB8DP+45/6RLX02a+ZPFn/J0HwM/7jn/pEtfTZrDF/FH0/wAzWj9r1PCv24v+TX/Gf/bl/wClsFfnvX6Eftxf8mv+M/8Aty/9LYK/PevteGf4E/X/ACPBzT+IvQK+2f8Agnh/ySLxT/2NN1/6T21fE1fbP/BPD/kkXin/ALGm6/8ASe2rp4j/AN0XqRlv8d+h9SfxUtJ/FS1+Zn1IUUUUAFFFFABRRRQAUUUUAFfMfhP/AJOg+Of/AHA//SJq+nK+Y/Cf/J0Hxz/7gf8A6RNXdhd5en6o5q32fU9RooorczCiiigAooooAKKKKACiiigAooooAKKK+ev22f2ox+y/8LodQ06C3vvFusTmz0izuMlNwGZJmUEFlQFeAeS6DoTSbsrsFqz6For4G8MfshftHfErw/beJvGX7Q/iHwj4kuk+0Jo2m+d5VqW+YI4imiRTjAKqhC9MtivoD4A3XxG+Evwx8RH4+eJtJuU0W9YWviVp0RJrHYgV5WwuDvLLlwGJ654JSlfoVZdz3mivNdQ/aV+Fel6pomm3XxA8PxX2tRQzafD9uRmnjlwYmGDwHyNpOAe1eb2vgeFf23rzxGfjL5tw2lAD4beed0aeQqb9nm425Hnf6vO5s9OafMhJH0lRXxp4/wD28tO8OftdeDvBOn+KfCc3w0utOeTW9YacObS7H2oeWZhJsQgxQDaRn5/cY7b9qS88M/HL4A295oXxs07wHoX9rxMfFFneh7eZlEitbF0lTJJbdjd1jHGOi5kPlZ9K0Vi+FIU0bwXo8U2r/wBrx2unwo+rTOP9KCxjM7HJHzY3E5PWuG039qf4P6x4hj0Oy+JXhm51ORhHHDHqUREjE4CK+7azZ4wCSfSnfuKx6nRWD4z8feGvh1pI1PxVr+meHdOL+WLnVLuO3jZsfdUuRk8dBk1gfD/4+fDn4rXjWfhLxtomv3yIZGs7O8Rp9o4LeWTu29OcY5HrRzIVmd7RRRTAKKKKACiiigAooooAKKKKACiiigAooooA8u8Wf8nQfAz/ALjn/pEtfTZr5k8Wf8nQfAz/ALjn/pEtfTZrDF/FH0/zNaP2vU8K/bi/5Nf8Z/8Abl/6WwV+e9foR+3F/wAmv+M/+3L/ANLYK/PevteGf4E/X/I8HNP4i9Ar7Z/4J4f8ki8U/wDY03X/AKT21fE1fbP/AATw/wCSReKf+xpuv/Se2rp4j/3RepGW/wAd+h9SfxUtJ/FS1+Zn1IUUUUAFFFFABRRRQAUUUUAFfMfhP/k6D45/9wP/ANImr6cr5j8J/wDJ0Hxz/wC4H/6RNXdhd5en6o5q32fU9RooorczCiiigAooooAKKKKACiiigAooooAK/PX/AIKIMbj9qn9mu31PH9g/2rEXJ6fNfWwm/wDHBHX6FV8zft3/ALLF5+0x8N9OPh6SG38aeHp2utLaeTy1lVwPNg3fwltkbBum6MZIBJqJ3cdC4OzPpmvAv29P+TQ/iV/14R/+j4q8B8J/tvfHvwPoNr4e8afs6+KPEXii2i8kanYRXEcd6y/LvIW3kXPQlkYgknAUEV0/jiT43/Fj9i/4tTfELwkun+INV2HRPDWkW7S3MdqHhO1kUsxfIY4J3cHhRgVPMnFpDUbO5i/sN/sUfDbVPgf4V8ceK9J/4SjxPq4i1OK8up5U+xLG/wC4iiCsOFCLknr0+6AKXT/+Ut2p/wDYsrn/AMBY6+jP2PdD1Lw1+zJ8OdM1ewutK1K20pEns7yFoZom3E7WRgCDgjgjNeIWPgPxKv8AwVC1HxQfD+qL4abw4Il1k2Ugsy/2eNdgm27N24EYznIpW91BfVnkvxi+Anw9t/8AgpR8MPCMXhPT4/DOtaHJf6jpioRDczn+0WMjAHrmKPp/dFegf8FIPh74b+F/7GkOg+FNHttC0eLxDbSLZ2i4QMwlLN9Se5qD9tjw348+Hv7Vnwv+OHhbwZqXjbS9G0/+zruz0qJ5ZEw9xuDbFZkDJcsA+0jK4PUZP2wNc8Z/tLfsU22qWXwz8SaNrE/iGFk8PNZy3F6LdDKomaNYwwU5B6YGRyQQSntIrdoqf8FBvHGraT+y38K/Buk3bWR8XPZWd0y5BkgSBCYyfQu8ZPqFx0Jz7Fcf8E5/gdd/Du38LS+FfKuIYFjOvWsrJqLSgYMvmHIJJydrAoM4C9K5n9sX9m/xJ8b/ANmDwbb+G7d/+Ey8LxWt9b2EhEUk2LcJLEN2Nrg4YZxzHjuMcbY/tvfHfXfDdp4Z0n9nrxCvxHaNbea/1G1ni0+OTGDOytGgUE87WdVGfvHGC9E9SemhyWk/D/Tf2rf2/vFvh3xm82q+CPhzpwtLDR5JmEcjRGKLD4IJ3SNI7EY3bUByoxTf+ChP7PPhT9nrw54P+LXwv02LwX4i0nW4bdl03McUmUkkR9mcBlaLBxgMrnOcCtn4jeBfi1+y/wDtRXPxq8KeC7r4gaF4psEh8QaToyvJLDOUjMwVUVmC+ZEJFk2sACVOMgnL+LF58V/+CheteFPB1v8ADTXPht8OLG/TUNU1bX43ieXaChMe5FDMqPIFRd2WbLFQOI6NNalrfyP0H8J61/wknhbRtXCeX/aFlDdbMfd3oGx6/wAVatQWFjDpdjbWdunl29vGsMaDsqjAH5AVPXSYBRRRQAUUUUAFFFFABRRRQAUUUUAFFFFOzFdHl3iz/k6D4Gf9xz/0iWvpr+VfMvi3/k6D4G9v+Q5/6RLWxq37XGk+DfihrfhTxNp8lrZ2dwscOqWmZBtKqf3kfXjJ5XPbiqqYariJJUo3aV/xCNaFK/O7XZd/bh/5Nf8AGf8A25f+lsFfnxX3l+154o0jxf8Asm+L9Q0XUbfVLKT7Fia1kDjP22Dg46H2PNfBtfW8NxcaNRSVnf8AyPGzNpzi12Cvtn/gnh/ySLxT/wBjTdf+k9tXxNX2z/wTw/5JF4p/7Gm6/wDSe2rfiP8A3RepOW/x36H1J/FS0n8VLX5mfUhRRRQAUUUUAFFFFABRRRQAV8x+E/8Ak6D45/8AcD/9Imr6cr5j8J/8nQfHP/uB/wDpE1d2F3l6fqjmrfZ9T1GiiitzMKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigD57+NX7PfiP4jeOJdZ0280yC1e3jiCXcsivlRzwI2H61wf/DH/jL/AKCWhf8Af+b/AONV9gUV7FPNMRRgoRtp5HFPCU5ycmfEGpfs5eJdL8eeFfCUt7pbaj4k+1/ZJEll8lPs8QkfzCY8jI6YByeuOtch4o+Hur+FvHF34TkiXUtXtpRCY9OV5RIxUMAg2hjwR2FfXvi3/k6D4G/9xz/0iWvoHTfB+i6PrF/q1npltBqd+2+5vFQGWU4AwWPOOBx0rp/t6rh376vdadNbmf8AZ8Kmztqfnj40/Z68Z+A/gf4n8Wa7Kuk2Kx2yvozSFpLoPcxIvmhThQpYOASTlegPNeVf5FfoN+3B/wAmv+M8f9OX/pbBX5817eS4ypjYTqVN7nn46jGhKMIhX2z/AME8P+SReKf+xpuv/Se2r4mr7Z/4J4f8ki8U/wDY03X/AKT21ZcR/wC6L1Rrln8b5H1J/FS0n8VLX5mfUhRRRQAUUUUAFFFFABRRRQAV8x+E/wDk6D45/wDcD/8ASJq+nK+Y/Cf/ACdB8c/+4H/6RNXdhd5en6o5q32fU9RooorczCiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKAPLvFn/J0HwM/wC45/6RLX03XzJ4s/5Og+Bn/cc/9Ilr6bNYYv4o+n+ZrR+16nhX7cX/ACa/4z/7cv8A0tgr896/Qj9uL/k1/wAZ/wDbl/6WwV+e9fa8M/wJ+v8AkeDmn8RegV9s/wDBPD/kkXin/sabr/0ntq+Jq+2f+CeH/JIvFP8A2NN1/wCk9tXTxH/ui9SMt/jv0PqT+KlpP4qWvzM+pCiiigAooooAKKKKACiiigAr5j8J/wDJ0Hxz/wC4H/6RNX05XzH4T/5Og+Of/cD/APSJq7sLvL0/VHNW+z6nqNFFFbmYUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQB5d4s/wCToPgZ/wBxz/0iWvps18yeLP8Ak6D4Gf8Acc/9Ilr6bNYYv4o+n+ZrR+16nhX7cX/Jr/jP/ty/9LYK/Pev0I/bi/5Nf8Z/9uX/AKWwV+e9fa8M/wACfr/keDmn8RegV9s/8E8P+SReKf8Asabr/wBJ7aviavtn/gnh/wAki8U/9jTdf+k9tXTxH/ui9SMt/jv0PqT+KlpP4qWvzM+pCiiigAooooAKKKKACiiigAr5j8J/8nQfHP8A7gf/AKRNX05XzH4T/wCToPjn/wBwP/0iau7C7y9P1RzVvs+p6jRRRW5mFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAeXeLP+ToPgZ/3HP/AEiWvps18yeLP+ToPgZ/3HP/AEiWvps1hi/ij6f5mtH7XqeFftxf8mv+M/8Aty/9LYK/Pev0I/bi/wCTX/Gf/bl/6WwV+e9fa8M/wJ+v+R4OafxF6BX2z/wTw/5JF4p/7Gm6/wDSe2r4mr7Z/wCCeH/JIvFP/Y03X/pPbV08R/7ovUjLf479D6k/ipaT+Klr8zPqQooooAKKKKACiiigAooooAK+Y/Cf/J0Hxz/7gf8A6RNX05XzH4T/AOToPjn/ANwP/wBImruwu8vT9Uc1b7Pqeo0UUVuZhRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFAHl3iz/k6D4Gf9xz/ANIlr6bNfMniz/k6D4Gf9xz/ANIlr6bNYYv4o+n+ZrR+16nhX7cX/Jr/AIz/AO3L/wBLYK/Pev0I/bi/5Nf8Z/8Abl/6WwV+e9fa8M/wJ+v+R4OafxF6BX2z/wAE8P8AkkXin/sabr/0ntq+Jq+2f+CeH/JIvFP/AGNN1/6T21dPEf8Aui9SMt/jv0PqT+KlpP4qWvzM+pCiiigAooooAKKKKACiiigAr5j8J/8AJ0Hxz/7gf/pE1fTlfMfhP/k6D45/9wP/ANImruwu8vT9Uc1b7Pqeo0UUVuZhRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFAHl3iz/k6D4Gf9xz/0iWvps18yeLP+ToPgZ/3HP/SJa+mzWGL+KPp/ma0ftep4V+3F/wAmv+M/+3L/ANLYK/Pev0I/bi/5Nf8AGf8A25f+lsFfnvX2vDP8Cfr/AJHg5p/EXoFfbP8AwTw/5JF4p/7Gm6/9J7aviavtn/gnh/ySLxT/ANjTdf8ApPbV08R/7ovUjLf479D6k/ipaT+Klr8zPqQooooAKKKKACiiigAooooAK+Y/Cf8AydB8c/8AuB/+kTV9OV8x+E/+ToPjn/3A/wD0iau7C7y9P1RzVvs+p6jRRRW5mFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAeXeLP8Ak6D4Gf8Acc/9Ilr6bNfMniz/AJOg+Bn/AHHP/SJa+mzWGL+KPp/ma0ftep4V+3F/ya/4z/7cv/S2Cvz3r9CP24v+TX/Gf/bl/wClsFfnvX2vDP8AAn6/5Hg5p/EXoFfbP/BPD/kkXin/ALGm6/8ASe2r4mr7Z/4J4f8AJI/FP/Y03X/pPbV08R/7ovUjLf479D6k/ipaTvS1+Zn1IUUUUAFFFFABRRRQAUUUUAFfMfhP/k6D45/9wP8A9Imr6cr5j8J/8nQfHP8A7gf/AKRNXdhd5en6o5q32fU9RooorczCiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKAPLvFn/J0HwM/7jn/pEtfTR/pXzL4s/wCToPgZ/wBxz/0iWvpvOawxfxR9DWj19TyX9qjwPrfxI+A/ibw74dsv7R1m8+y+Ra+akW/ZdRSN8zsqjCqx5Pb1wK+Jf+GW/jX/ANE9/wDK1Zf/AB2v00ox7V04LNK+Bi4UrWfcyr4SniJc0z8zP+GW/jX/ANE9/wDK1Zf/AByvrT9i/wCGfij4WfDXW9N8WaV/Y+o3euz30dv9oinzE0MCht0bMPvIwxkHjpXv+PakxV4zNsRjafsqtrCo4OnQlzRuHSnU3mnV4p3BRRRQAUUUUAFFFFABRRRQAV8x+E/+ToPjn/3A/wD0iavpyvmPwn/ydB8c/wDuB/8ApE1d2F3l6fqjmrfZ9T1GiiitzMKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooA8u8Wf8nQfAz/uOf+kS19OV8x+LP+ToPgZ/3HP/AEiWvpysMXvH0NaP2vUKKKK4ToCiiigAooooAKKKKACmseCadTWPBoE9j887D9vr4r6haJcRaR4PVHzgPa3eeDjtOe4qS5/b0+LNtbyyvpHg0pGpc7bW7zgDP/Pevnbwv/yAbX/gf/obVb1b/kFXv/XF/wD0E1+p08nwTpqTh0PkpYyuptcx+sHwy8SXXjL4c+FNfvUhivNV0q0vp44ARGsksKuwUEkgZY4yScd66iuD+Av/ACQ34d/9i7p3/pNHXeV+XzVpteZ9ZHWKbCvmPwn/AMnQfHP/ALgf/pE1fTlfMfhP/k6D45/9wP8A9Imrrwu8vT9UYVvs+p6jRRRW5mFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUwCisvXvFWjeFbX7TrWrWOkwf89L24SJeuMAsRnn0rzm6/ac8H3F09l4ch1jxrqCts+y+HdNluDn/eIVSMdwTwD6VpGnOWyJ5kj1qivLrfWPjX40/wCQD8PNO8KWrthL3xVqG5tvcmCEb1P1z2rSt/2cfHnij5vGXxZ1KGFmy2n+FbZLBVHoJuXIPuPUUmoQ+OaX4/kC5pfCv0Or17xVo3hW1+061q1jpMH/AD0vbhIl64wCxGefSvObr9pzwfcXbWXhyHWPGuoK2w2vh3TZLg5/3iFUjHcE8A+legeGf2Sfhd4auheSeGo9d1Etue812Z72SQ46sJCU/wDHew9K9Y03S7PR7OO1sbSCytY/uQ28aoi9+FAAFYyxFGOyb/A09nN7ux80+FNB+IPxH+N3gDxfqfgaXwj4c8NnUMyajfxtcz/aLYxgmFRlMMF69jmvqOkpa461V1ndqxvCHIrXCiiisDQKKKKACiiigAooooAKa3Q/SnU1uh+lNbiex+N3hf8A5ANr/wAD/wDQ2q3q3/IKvf8Ari//AKCaqeF/+QDa/wDA/wD0Nqt6t/yCr3/ri/8A6Ca/bKX8Feh8LL42fqb8Bf8Akhnw7/7FzTv/AEmjru64T4C/8kM+Hf8A2Lmnf+k0dd3X4tU+OXqfcQ+FBXzH4T/5Og+Of/cD/wDSJq+nK+Y/Cf8AydB8c/8AuB/+kTV1YXeXp+qMa32fU9RooorczCiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooopgFFZeveKtG8K2v2nWtWsdJg/56XtwkS9cYBYjPPpXnN1+054PuLp7Lw5DrHjXUFbZ9l8O6bLcHP8AvEKpGO4J4B9K0jTnLZE8yR61RXl1vrHxr8af8gH4ead4UtXbCXvirUNzbe5MEI3qfrntWlb/ALOPjzxR83jL4s6lDCzZbT/Ctslgqj0E3LkH3HqKTUIfHNL8fyBc0vhX6HV694q0bwra/ada1ax0mD/npe3CRL1xgFiM8+lec3X7Tng+4unsvDkOseNdQVtn2Xw7pstwc/7xCqRjnIJ4B9K9A8M/sk/C7w3dC8k8NR67qJbc95rsz3skhx1YSEp/472HpXrOm6XZ6PZx2ljaQWVrH9yG3jEaL9FAwKyeIox+FN/gX7Ob3dj5vg1j41+M/wDkA/DzTvClq7YS98Vahubb3JghG9T9c9q0rf8AZx8eeKMN4y+LOpRQM2W0/wAK2yWCqPQTcuQfceor6IFLWLxk/sJL+vM0VCPXU8b8M/skfC7w3dC8fw1Hruolt73muzPeySHHVhISn/jvYeles6bpdno9pHaWNpBZWsf3IbeMRovOeFAAFW8Utcsqk6nxO5rGMY7ISlpM0tZlhRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABTW6H6U6mt0P0prcT2Pxu8L/APIBtf8Agf8A6G1W9W/5BV7/ANcX/wDQTVTwv/yAbX/gf/obVb1b/kFXv/XF/wD0E1+2Uv4K9D4Wfxv1P1M+A3/JDfh1/wBi7p3/AKTR13XNfHviLxh8VPDfwr+EMXw/h1SWzk8KWLXP9n6St4u/yIwMsY22nHbiuS/4Wp+0p/z5+I//AAmU/wDkevy+OWTr3qKpFXb3ev5H1LxcadouL+4+7+a+ULDxVpHhf9qj4vQatqMGnyao2jraee20SGOyG8bjwD+8TGTznjpXn/8AwtT9pT/nz8R/+Eyn/wAj14l8QNU8Ra14y1O98WpcR+I5mjN4t3bC3lBESKm6MKu392E6AZBB75r1sBkzdRqdSLTX2Xd/kcmIx3urli/mj9E1YMoI5B6Y5FFfF/wO8Y/EqG9jsfDMM2s6ajASW14CbaIf9dCR5fHYHn0NfZNi9xJZwtdxRQ3TIDLHDIXRW7gMQMj04H0rixmDeDnytp/12NqFf2yulYnooorzzqCiiikAUUUUAFFFFABRRRQAUUUUAFFFFMAorL17xVo3hW1+061q1jpMH/PS9uEiXrjALEZ59K85uv2nPB9xdPZeHIdY8a6grbPsvh3TZbg5/wB4hVIx3BPAPpWkac5bInmSPWqK8ut9Y+NfjT/kA/DzTvClq7YS98Vahubb3JghG9T9c9q0rf8AZx8eeKPm8ZfFnUoYWbLaf4VtksFUegm5cg+49RSahD45pfj+QLml8K/Q6vXvFWjeFbX7TrWrWOkwf89L24SJeuMAsRnn0rzm6/ac8H3F09l4ch1jxrqCts+y+HdNluDn/eIVSMc5BPAPpXoHhn9kn4XeG7oXknhqPXdRLbnvNdme9kkOOrCQlP8Ax3sPSvWdN0uz0ezjtLG0gsrWP7kNvGI0X6KBgVk8RRj8Kb/Av2c3u7HzfBrHxr8Z/wDIB+HmneFLV2wl74q1Dc23uTBCN6n657VpW/7OPjzxRhvGXxZ1KKBmy2n+FbZLBVHoJuXIPuPUV9EClrF4yf2El/XmaKhHrqeN+Gf2SPhd4buheP4aj13US297zXZnvZJDjqwkJT/x3sPSvWdN0uz0e0jtLG0gsrWP7kNvGI0XnPCgACreKWuWVSdT4nc1jGMdkJS0maWsywooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKa3Q/SnU1uh+lNbiex+N3hf8A5ANr/wAD/wDQ2q3q3/IKvf8Ari//AKCaqeF/+QDa/wDA/wD0Nqt6t/yCr3/ri/8A6Ca/bKX8Feh8LL42fqZ8Bf8Akhnw7/7FzTv/AEmjru64X4C/8kM+Hf8A2Lmnf+k0dd3X4tU+OXqfcQ+FCHoa+Pbz4c6F8Rv2pfis2uW73UWktpTxQCQrHIZbJc7wOTjyhjnuc5FfYVfMnhP/AJOg+Of/AHA//SJq7sDUnTlKUHZ2/wAjnxEVJRT7npGn6baaTZxWllbQ2drENqQQIERfYAdBVmiitW3J3buRtsFFFFSMKKKKACiiimAUVl694q0bwra/ada1ax0mD/npe3CRL1xgFiM8+lec3X7Tng+4unsvDkOseNdQVtn2Xw7pstwc/wC8QqkY7gngH0rSNOctkTzJHrVFeXW+sfGvxp/yAfh5p3hS1dsJe+KtQ3Nt7kwQjep+ue1aVv8As4+PPFHzeMvizqUMLNltP8K2yWCqPQTcuQfceopNQh8c0vx/IFzS+FfodXr3irRvCtr9p1rVrHSYP+el7cJEvXGAWIzz6V5zdftOeD7i6ey8OQ6x411BW2fZfDumy3Bz/vEKpGOcgngH0r0Dwz+yT8LvDd0LyTw1Hruoltz3muzPeySHHVhISn/jvYeles6bpdno9nHaWNpBZWsf3IbeMRov0UDArJ4ijH4U3+Bfs5vd2Pm+DWPjX4z/AOQD8PNO8KWrthL3xVqG5tvcmCEb1P1z2rSt/wBnHx54ow3jL4s6lFAzZbT/AArbJYKo9BNy5B9x6ivogUtYvGT+wkv68zRUI9dTxvwz+yR8LvDd0Lx/DUeu6iW3vea7M97JIcdWEhKf+O9h6V6zpul2ej2kdpY2kFlax/cht4xGi854UAAVbxS1yyqTqfE7msYxjshKWkzS1mWFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAU1uh+lOprdD9Ka3E9j8bvC/wDyAbX/AIH/AOhtVvVv+QVe/wDXF/8A0E1U8L/8gG1/4H/6G1W9W/5BV7/1xf8A9BNftlL+CvQ+Fl8bP1N+Av8AyQz4d/8AYuad/wCk0dd3XCfAX/khnw7/AOxc07/0mjru6/Fqnxy9T7iHwoK+Y/CX/J0Hxz/7gf8A6RNX03XzJ4T/AOToPjn/ANwP/wBImrqwm8vT9UY1vsnqNFZeveKtG8K2v2nWtWsdJg/56XtwkS9cYBYjPPpXnN1+054PuLp7Lw5DrHjXUFbZ9l8O6bLcHP8AvEKpGO4J4B9K7I05y2RhzRR61RXl1vrHxr8af8gH4ead4UtXbCXvirUNzbe5MEI3qfrntWlb/s4+PPFHzeMvizqUMLNltP8ACtslgqj0E3LkH3HqKTUIfHNL8fyGuaXwr9Dq9e8VaN4VtftOtatY6TB/z0vbhIl64wCxGefSvObr9pzwfcXT2XhyHWPGuoK2z7L4d02W4Of94hVIxzkE8A+legeGf2Sfhd4buheSeGo9d1Etue812Z72SQ46sJCU/wDHew9K9Z03S7PR7OO0sbSCytY/uQ28YjRfooGBWTxFGPwpv8C/Zze7sfN8GsfGvxn/AMgH4ead4UtXbCXvirUNzbe5MEI3qfrntWlb/s4+PPFGG8ZfFnUooGbLaf4VtksFUegm5cg+49RX0QKWsXjJ/YSX9eZoqEeup434Z/ZI+F3hu6F4/hqPXdRLb3vNdme9kkOOrCQlP/Hew9K9Z03S7PR7SO0sbSCytY/uQ28YjRec8KAAKt4pa5ZVJ1PidzWMYx2QlLSZpazLCiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACiiigAooooAKKKKACmt0P0p1Nf7pHtTW4nsfjd4X/wCQDa/8D/8AQ2q3q3/IKvf+uL/+gmsTQdesLPSYIZp9kq7srsY9WJHQelWdS8RafLY3UK3GZWjdApRvvYx6etftFKpD2SXMtj4eUZc7dj9XvgL/AMkN+Hf/AGLmnf8ApNHXeVwfwF/5Ib8Ox/1Lmnf+k0dd5X41U+OXqfbQ+FCHpXgHij9le58U/FLxV4q/4WBreg2HiBrU3GnaEBbSkQQLEoM5LHkhjgKOGIOeCPf6SnTqTpNuDCUIz+I8c8M/sk/C7w3dC8fw1Hruoltz3muzPeySHplhISn/AI72HpXrOm6XZ6PaR2ljaQWVrH9yG3jEaLznhQABVqnUp1J1NZu4RhGOyEpaTNLWZYUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFADGjVxgqGHuKFjVSSFAJ5PHWiigQ+iiigYUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAFFFFABRRRQAUUUUAf/2Q==)", "_____no_output_____" ] ], [ [ "def build_Wasserstein_critic(img_shape, verbose = False):\n Xin = Input(shape=(img_shape[0],img_shape[1],img_shape[2],))\n '''# Agrego clase de vector\n image_class = Input(shape = (1,), dtype = np.float32)\n # 10 clases en MNIST\n cls = Flatten()(tf.keras.layers.Embedding(10, z_dim)(image_class))\n # hadamard product between z-space and a class conditional embedding\n h = tf.keras.layers.Multiply()([Xin, cls])'''\n\n # Convolución2D tensor de 28x28x1 a 14x14x32 y activación Leaky ReLU\n X = Conv2D(filters = 32, \n kernel_size = 3, \n strides = 2, \n input_shape = img_shape, \n padding = 'same')(Xin)\n #X = BatchNormalization()(X)\n X = LeakyReLU(alpha = 0.01)(X)\n\n # Convolución2D tensor de 14x14x32 a 7x7x64, con normalización por lote y activación Leaky ReLU\n X = Conv2D(filters = 64, \n kernel_size = 3, \n strides = 2, \n padding = 'same')(X)\n X = BatchNormalization()(X)\n X = LeakyReLU(alpha = 0.01)(X)\n \n # Convolución2D tensor de 7x7x64 a 3x3x128, con normalización por lote y activación Leaky ReLU\n X = Conv2D(filters = 128, \n kernel_size = 3, \n strides = 2, \n padding = 'same')(X)\n X = BatchNormalization()(X)\n X = LeakyReLU(alpha = 0.01)(X)\n\n # Aplanado del tensor, y capa densa de salida de clasificacion con activación sigmoide\n Y_out1 = Flatten()(X)\n Y_out2 = Dense(1, activation='sigmoid')(Y_out1)\n\n critic_model = Model(inputs = Xin, outputs = [Y_out2], name ='critic')\n\n return critic_model", "_____no_output_____" ], [ "# construye el discriminador \nwasserstein_critic = build_Wasserstein_critic(img_shape)\nwasserstein_critic.summary()", "Model: \"critic\"\n_________________________________________________________________\nLayer (type) Output Shape Param # \n=================================================================\ninput_75 (InputLayer) [(None, 32, 32, 3)] 0 \n_________________________________________________________________\nconv2d_147 (Conv2D) (None, 16, 16, 32) 896 \n_________________________________________________________________\nleaky_re_lu_193 (LeakyReLU) (None, 16, 16, 32) 0 \n_________________________________________________________________\nconv2d_148 (Conv2D) (None, 8, 8, 64) 18496 \n_________________________________________________________________\nbatch_normalization_144 (Bat (None, 8, 8, 64) 256 \n_________________________________________________________________\nleaky_re_lu_194 (LeakyReLU) (None, 8, 8, 64) 0 \n_________________________________________________________________\nconv2d_149 (Conv2D) (None, 4, 4, 128) 73856 \n_________________________________________________________________\nbatch_normalization_145 (Bat (None, 4, 4, 128) 512 \n_________________________________________________________________\nleaky_re_lu_195 (LeakyReLU) (None, 4, 4, 128) 0 \n_________________________________________________________________\nflatten_51 (Flatten) (None, 2048) 0 \n_________________________________________________________________\ndense_48 (Dense) (None, 1) 2049 \n=================================================================\nTotal params: 96,065\nTrainable params: 95,681\nNon-trainable params: 384\n_________________________________________________________________\n" ] ], [ [ "# Modelo Completo", "_____no_output_____" ], [ "El modelo completo se ve como", "_____no_output_____" ], [ "![image.png](data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAAA1QAAAJECAYAAAARwGVyAAAgAElEQVR4Aey9ebQ9V1XtDwrSqvSNHSD6sEdEHmIDqLEDaQQBUUCEICCIJEAEpDGgmAAirZgAChGE0ER6EExC8yCGTkhoAyiCDxB5jjDG+4Pfe+ON+o19k3m+dc9d55w6tVbtpurDGIzzvfvUXnvVXHPNVTP33nMv0/E/EAABEAABEAABEAABEAABEACBUQhcZtQuNoEACIAACIAACIAACIAACIAACHQYKkgAAiAAAiAAAiAAAiAAAiAAAiMRwFCNBI5tIAACIAACIAACIAACIAACIIChggMgAAIgAAIgAAIgAAIgAAIgMBIBDNVI4NgGAiAAAiAAAiAAAiAAAiAAAhgqOAACIAACIAACIAACIAACIAACIxHAUI0Ejm0gAAIgAAIgAAIgAAIgAAIggKGCAyAAAiAAAiAAAiAAAiAAAiAwEgEM1Ujg2AYCIAACIBCLwC3v8tRuiv/HZkk0EAABEAABEDiMAIbqMB58BQIgAAIgUAiBKcxUisn/QAAEQAAEQGBKBDBUU6JLbBAAARAAgcEIyFDd8cRzuoj/K97gBLgQBEAABEAABEYggKEaARpbQAAEQAAE4hGQAYowUymG4sVnSkQQAAEQAAEQOIYAhuoYFvwLBEAABECgIAIyQBiqgkXgaBAAARAAgb0RwFDtDRkbQAAEQAAEpkAAQzUFqsQEARAAARCYGgEM1dQIEx8EQAAEQGAQAhiqQTBxEQiAAAiAQGUIYKgqKwjpgAAIgMBSEcBQLbXy3DcIgAAItI0Ahqrt+pE9CIAACMwGAQzVbErJjYAACIDAohDAUC2q3NwsCIAACNSLAIaq3tqQGQiAAAiAwGYEMFSbseEdEAABEACBjAhgqDKCzVEgAAIgAAJhCGCowqAkEAiAAAiAgAcBDJUHPfaCAAiAAAiUQgBDVQp5zgUBEAABEDiEAIbqEBx8AQIgAAIg0AgCGKpGCkWaIAACIDB3BDBUc68w9wcCIAAC80QAQzXPunJXIAACINAcAhiq5kpGwiAAAiAAAl3XYaigAQiAAAiAQBUIYKiqKANJgAAIgAAI7IkAhmpPwLgcBEAABOaAwBf/4+LqbgNDVV1JSAgEQAAEQGAAAhiqASBxCQiAAAi0jkAyUG8858Lu5Ge/uZNxqe2elNcdTzyni/i/4tV2n+QDAiAAAiAwLwQwVPOqJ3cDAiAAAgcIWAZKBkOvtUGlvCLMVIqheLXdJ/mAAAiAAAjMCwEM1bzqyd2AAAgsFIFdBur2xz+/e+Qpr+1OOb1eoyEDhKFaKIm5bRAAARBoFAEMVaOFI20QAIFlI7CPgTrvI//efeYLX1v9X8alNgSVF4aqtsqQDwiAAAiAwDYEMFTb0OE9EAABEKgEAY+B6pup9G8Zl0pubZWG8sJQrSDhHyAAAiAAAg0ggKFqoEikCAIgsDwEIg0Uhmp5/OGOQQAEQAAE8iGAocqHNSeBAAiAwEYEpjRQGKqNsPMGCIAACIAACLgRwFC5ISQACIAACOyPQE4DhaHavz7sAAEQAAEQAIGhCGCohiLFdSAAAiDgQKCkgcJQOQrHVhAAARAAARDYgQCGagdAvA0CIAACYxCoyUBhqMZUkD0gAAIgAAIgMAwBDNUwnLgKBEAABLYiULOBwlBtLR1vggAIgAAIgIALAQyVCz42gwAILBWBlgwUhmqpLOW+QQAEQAAEciCAocqBMmeAAAg0j0DLBgpD1Tz9uAEQAAEQAIGKEcBQVVwcUgMBECiHwJwMFIZqeh598KOfn/4QTgABEAABEKgSAQxVlWUhKRAAgdwIzNlAYaimZ9Mt7/LU7k4PPK07+dlv7t54zoUdBmt6zDkBBEAABGpBAENVSyXIAwRAICsCSzJQGKrpqZUM1fr/MVjT484JIAACIFADAhiqGqpADiAAApMjsGQDhaGanF4rM/WiV7+v++2TXtrd/vjnr9ZktDBY09eBE0AABECgBAIYqhKocyYIgMDkCGCgvtatGyl9rQf8yYuw5wHK644nntNF/F/x9kxj1OU6Sxin1/M+8u8dBmsUnGwCARAAgaYQwFA1VS6SBQEQ2IQABmqzgeo/5Kd/6+F/E5al1pVXhJlKMRQvx/3orHWs+19jsHJUgjNAAARAID8CGKr8mHMiCIBAAAIYqOEGqv9Qj6EKIJ8RYoihWq8DBssAkiUQAAEQaBABDFWDRSNlEFgiAhio8QZq/UFeD/+18Uh5zfU7VOt1WP8ag1UbI8kHBEAABIYhgKEahhNXgQAIZEYAAxVnoNYf3GVcMpd053HKa6mGar1OGKydlOECEAABEKgCAQxVFWUgCRAAAQzUdAZq/UFdxqU21ikvDJXNBQxWbYwlHxAAARC4BAEMFUwAARAoggAGyn5oXjc/U3wt41Kk8FsOVV4YqmHcwGBtIRNvgQAIgEBGBDBUGcHmKBBYMgIYqGEPyVMYqPWYMi618VF5YajGcQWDVRujyQcEQGApCGCollJp7hMEMiOAgRr3ULxufqb4WsYlMyV2Hqe8MFQx3MFg7aQcF4AACIBACAIYqhAYCQICIICBinkInsJArceUcamNtcoLQzUNlzBYtTGefEAABOaCAIZqLpXkPkAgMwIYqGkeetfNzxRfy7hkpszO45QXhioPtzBYOynJBSAAAiAwCAEM1SCYuAgEQAADlechdwoDtR5TxqU2VisvDFUZrmGwausI8gEBEGgFAQxVK5UiTxDIjAAGqsxD7br5meJrGZfMlNp5nPLCUNXBPQzWTspyAQiAAAgcIIChggggAAIHCGCg6niIncJArceUcamN+soLQ1UnFzFYtXUM+YAACNSCAIaqlkqQBwhkRgADVedD67r5meJrGZfMlNt5nPLCULXBTQzWTkpzAQiAwEIQwFAtpNDcJghgoNp4SJ3CQK3HlHGprSuUF4aqTa5isGrrKPIBARDIhQCGKhfSnAMCmRHAQLX5ULpufqb4WsYlMyV3Hqe8MFTz4C4GayfluQAEQGAmCGCoZlJIbgMEMFDzeAidwkCtx5Rxqa1rlBeGap5cxmDV1nHkAwIgEIUAhioKSeKAQGYEMFDzfOhcNz9TfC3jkpmyO49TXhiqZXAbg7WzJbgABECgEQQwVI0UijRBAAO1jIfMKQzUekwZl9q6SnlhqJbJdQxWbR1JPiAAAkMRwFANRYrrQCAzAhioZT5UrpufKb6WcclM6Z3HKS8MFdxPvMdg7WwZLgABEKgEAQxVJYUgDRDAQPEQOYV5smLKuNTWdcoLQ0UvWLzFYNXWseQDAiAgBDBUQoJXEMiMAAaKh0broTHHmoxLZsrvPE55YajojSF9gMHa2VJcAAIgkAkBDFUmoDkGBDBQPCQOeUjMcY2MS21dqbwwVPTKmD7AYNXW0eQDAstBAEO1nFpzp5kRwEDxUDjmoTDHHhmXzC2x8zjlhaGidyL6AIO1s+W4AARAIAgBDFUQkIQBAQwUD4ERD4E5Ysi41Na1ygtDRS9N0QcYrNo6nnxAYD4IYKjmU0vuJDMCGCge+qZ46MsRU8Ylc8vsPE55YajorRx9gMHa2ZJcAAIgMBABDNVAoLgMBDBQPOTleMjLcYaMS21drbwwVPRajj5YPwODVZsikA8ItIMAhqqdWpFpZgQwUDzUrT9wzeVrGZfMLbXzOOWFoaL3aug1DNbOluUCEACBSxHAUEEFELgUAQwUD3E1PMTlyEHGpbbmV14YKnoxRx/sewYGqzbFIB8QqAcBDFU9tSCTzAhgoHho2/eBai7Xy7hkbrmdxykvDBW92UKvYbB2tjQXgMBiEMBQLabU3CgGioe0Fh7ScuQo41KbKigvDBW9mqMPos/AYNWmKOQDAvkQwFDlw5qTMiOAgeKhLPqBaS7xZFwyt+TO45QXhorenUOvYbB2tjwXgMBsEMBQzaaU3AgGioewOTyE5bgHGZfaVEN5Yajo5Rx9kPsMDFZtikM+IBCHAIYqDksiZUYAA8VDV+4HormcJ+OSuWV3Hqe8MFT09lx6bdt9YLB2SgIXgEAzCGComikViWKgeMja9nDCe8P5IeNSm6ooLwzV8FrC+/lghcGqTZHIBwSGI4ChGo4VV2ZGAAM1nwcFHvrqqqWMS+aW3nmc8sJQ1cUX+rdMPTBYOyWDC0CgGgQwVNWUgkQwUGWGNg9Ly8NdxqU21VFeGKrlcRId2l1zDFZtikU+IHAMAQzVMSz4V2YEMFC7BygPGWA0BQdkXDK3/M7jlBeGCt5Pwfu5xcRg7ZQULgCBbAhgqLJBzUEYKB6S5vZA0+r9yLjUpkrKC0OFVrTaWyXzxmDVpmjksyQEMFRLqnbme8VA8VBU8uGCszfzT8YlsyTsPE55Yag21w5eg81QDmCwdkoOF4BAGAIYqjAoCYSBYtAPHfRcV5YrMi61qZbywlCV5Qf9OU/8MVi1KR75zAkBDNWcqpn5XjBQ8xy6PEzNv64yLpklY+dxygtDNX8OojPla4zB2ilJXAACgxHAUA2GigsxUOUHIA8h1CCCAzIutama8sJQwfMInhNjPx5hsGpTRPJpCQEMVUvVypwrBmq/YcTwBq9WOCDjkllSdh6nvDBU9FIrvTTnPDFYOyWLC0BghQCGagUF/8BA8RAz54cD7u0Yv2VcalM95YWhOlYreAsWtXAAg1WbYpJPTQhgqGqqRuZcMFAM6loGNXnk5aKMS2bJ2Xmc8sJQ5eUD/QfeYziAwdopaVywIAQwVAsqNgaKoTlmaLJnfryRcalN/pQXhmp+nENH5l9TDFZtiko+ORHAUOVEO/NZGKj5DzAeUqjxGA7IuGSWpJ3HKS8MFbwew2v21MUbDNZOyeOCGSGQ1VBpWHpeZ4R9+K1goOoaJgx36lErB6TB4SLkDKi8MFT0Tq29Q17juYnBcgok26tGAENVdXm2J4eBGi/sDEWwWzIHZFy2K0z+d5UXhor+XHJ/LuXeMVj5NZYTp0OgiKEaMyw1aKeDov7IGCgeMpYyaLnPableq54qrzEzwtqjeDnUXWfB3Wm5C77zxReDlUOpOGMqBDBUUyEbEBcDNd/BwUMBtS3JAT38B8hUaAjlZZmjMWuKF5rkhmA6q2RdORtdmRMHMFgbxIblKhHAUFVUFgwUw3BOw5B7qZfPevivSP4OUlFeY8yTtUfxctynzoL39fKe2rRdGwxWDiXjjLEIYKjGIhewDwPVtrgznKlfqxzQw3+AjIWGUF6WORqzpnihSW4IprNa5QR5o2etcQCDtUGMWC6CAIYqI+wYKAZWawOLfOfJWT38Z5S/QUcprzHmydqjeIMOd16ks+iZefYMda2/rhgsp4ix3YUAhsoF3/bNGKj6BZghSY2WyAE9/G9XsPzvKi/LHI1ZU7wcd6Kzlsgn7hkdrZEDGKwcyscZQgBDJSQCXjFQDJUahwo5wct1DujhP0D2QkMorzHmydqjeKFJbgims9ax5mv6Dw7UwQEM1gbxYjkEAQyVA0YMVB0iybCiDnBgPw7o4d8hf5NsVV6WORqzpniTJLsWVGfBxf24CF7gVYoDGKw1EeNLFwIYqj3gw0Ah/KWEn3PhXiQH9PC/h/xluVR5jTFP1h7Fy5G8zoqsE7HoeziQjwMYrBxKOd8zMFRbaouByidkDA2whgP5OKCH/y3yV+Qt5WWZozFripfjZnQWPM7HY7AG6yk5gMHKoZzzOQNDZdTy5Ge/udNw7L/e/vjnd4885bXdKaef06VGm7KRic2ggANwYCoOSNcM+Su6pLzGmCdrj+LluCmdNVXNiIsewIGyHMBg5VDSds/AUBm102DEQJUVL4YH+MOBaTggjTPkr+iS8rLM0Zg1xctxUzoLzk7DWXAF19o4gMHKoaztnIGhMmrFYES4axNu8oGTkRyQxhnyV3RJeY0xT9YexctxUzorsk7Eou/hQDscwGDlUNp6z8BQGbVhMLYjYAwbagUH9ueANM6Qv6JLyssyR2PWFC/HTeks+Lg/H8EMzObIAQxWDuWt5wwMlVELBiPiPkdx557gtTggjTPkr+iS8hpjnqw9ipfjpnSWMOaVfoMDcKDPAQxWDiUudwaGysCewYgI9kWQf8OHuXFAGmfIX9El5WWZozFripfjpnTW3LjC/aB/cGAaDmCwcihzvjMwVAbWDMZpxANRBlc4UAcHpHGG/BVdUl5jzJO1R/Fy3JTOguN1cJw6UIfWOIDByqHU052BoTKwZTAixK0JMfnC2X04II0z5K/okvKyzNGYNcXLcVM6a586cC19CwfgwCYOYLByKHfcGRgqA0sGIwK3SeBYhxtz4IA0zpC/okvKa4x5svYoXo6b0llz4Af3gM7Bgfo4gMHKoeTjz8BQGdgxGOsTEsSdmsCBOA5I4wz5K7qkvCxzNGZN8XLclM6Cp3E8BUuwhAObOYDByqHsw8/AUBlYMRg3NzDiBjZwoH0OSOMM+Su6pLzGmCdrj+LluCmdRX+03x/UkBq2yAEMVg6l33wGhsrAhsGImLYopuQMb4dyQBpnyF/RJeVlmaMxa4qX46Z01tAacB39CgfgwJQcwGDlUP5jZ2CojmGx+heDEZGbUuSIDb9Kc0AatxK9Sv6hvMaYJ2uP4uW4PZ1Vuracj77AAThgcQCDNe0kwFAZ+DIYESNLjFiDF3PhgDTOkL+iS8rLMkdj1hQvx03prLlwhPtA7+DAvDmAwYqdDBgqA08G47xFhCFBfZfOAWmcIX9Fl5TXGPNk7VG8HDels5bOLe4ffYUDbXJgH4OVQ1NbOwNDZVSMwdimGCDi1A0ODOOANM6Qv6JLyssyR2PWFC/HTeksODiMg+AETnCgbg5sM1g5NLW1MzBURsUYjHU3OSJMfeCAjwPSOEP+ii4przHmydqjeDluSmfBTR83wQ/84ECdHEgGSzqXQ1NbOwNDZVRMhKGp62xq6kJd4ICPA9I4Q/6KLikvyxyNWVO8HDels+Cmj5vgB35woF4OSOdyaGprZ2CojIqJMDR1vU1NbagNHBjPAWmcIX9Fl5TXGPNk7VG8HDels+DleF6CHdjBgbo5IJ3LoamtnYGhMiomwtDYdTc29aE+cGAcB6RxhvwVXVJeljkas6Z4OW5KZ8HJcZwEN3CDA/VzQDqXQ1NbOwNDZVRMhKG5629uakSN4MD+HJDGGfJXdEl5jTFP1h7Fy3FTOgs+7s9HMAMzONAGB6RzOTS1tTMwVEbFRBgavI0Gp07UCQ7sxwFpnCF/RZeUl2WOxqwpXo6b0llwcT8ughd4wYF2OCCd82qq4kS+enPy7sdQGQiqwDR5O01OragVHBjOAWmcIX9Fl5TXGPNk7VG8HDels+DhcB6CFVjBgbY4IJ3zaqriRL56c/Lux1AZCKrANHpbjU69qBccGMYBaZwhf0WXlJdljsasKV6Om9JZcHAYB8EJnOBAexyQznk1VXHG6Pr6HsXy5uTdj6EyEFRxaPb2mp2aUTM4sJsD0jhD/oouKa/1gTn2a8XLcVM6C/7t5h8YgREcaJMD0jmvpirOWG3v71Msb07e/RgqA0EVh4Zvs+GpG3WDA9s5II0z5K/okvLqD0vPvxUvx03pLLi3nXvgAz5woF0OSOe8mqo4Hn3XXsXy5uTdj6EyEFRxaPp2m57aUTs4sJkD0jhD/oouKS8NSu+r4uW4KZ0F7zbzDmzABg60zQHpnFdTFcer8Wm/Ynlz8u7HUBkIqjg0ftuNT/2oHxywOSCNM+Sv6JLyihiyuQetcodzNufABVzgQPsckM55B4XiRGi9Ynlz8u7HUBkIqjg0f/vNTw2pIRw4ygFpnCF/RZeUV8SQxVAdrTu9ACZwAA54OCCN9g4KxYnQesXy5uTdj6EyEFRxPKRjL6IFB+BArRyQxhnyV3RJeUUMWQwV/Vdr/5EX3GyVA9Jo76BQnAitVyxvTt79GCoDQRWnVcKTN2INB+DANg5I4wz5K7qkvCKGLIaKHtjWA7wHP+DA/hyQRnsHheJEaL1ieXPy7sdQGQiqODTb/s0GZmAGB+rngDTOkL+iS8orYshiqOrnIVpBjeBAWxyQRnsHheJEaL1ieXPy7sdQGQiqODR6W41OvagXHBjGAWmcIX9Fl5RXxJDFUA3jAj0DTnAADgzlgDTaOygUJ0LrFcubk3c/hspAUMUZSjCuQ4zgABxoiQPSOEP+ii4pr4ghi6GiJ1vqSXKFry1wQBrtHRSKE6H1iuXNybsfQ2UgqOK0QG5yRIThABzYlwPSOEP+ii4pr4ghi6GiL/btC66HM3BgOwek0d5BoTgRWq9Y3py8+zFUBoIqDo21vbHAB3zgQJsckMYZ8ld0SXlFDFkMVZvcRFOoGxyolwPSaO+gUJwIrVcsb07e/RgqA0EVh6aut6mpDbWBA+M5II0z5K/okvKKGLIYqvH8oLfADg7AAYsD0mjvoFCcCK1XLG9O3v0YKgNBFcciE2uIDByAA61zQBpnyF/RJeUVMWQxVPRp631K/nC4Ng5Io72DQnEitF6xvDl592OoDARVnNqITD6IKxyAAxEckMYZ8ld0SXlFDFkMFb0S0SvEgEdw4BgHpNHeQaE4EVqvWN6cvPsxVAaCKg5NdKyJwAIs4MB8OCCNM+Sv6JLyihiyGKr58BXtoZZwoA4OSKO9g0JxIrResbw5efdjqAwEVRwauI4Gpg7UAQ7EckAaZ8hf0SXlFTFkMVSxnKEHwRMOwAFptHdQKE6E1iuWNyfvfgyVgaCKg3ggHnAADsyRA9I4Q/6KLimviCGLoaJ359i73BO8LskBabR3UChOhNYrljcn734MlYGgilOStJyNaMIBODAVB6RxhvwVXVJeEUMWQ0X/TNU/xIVbS+WANNo7KBQnQusVy5uTdz+GykBQxVlqw3DfDAs4MG8OSOMM+Su6pLwihiyGat4cRqOoLxzIzwFptHdQKE6E1iuWNyfvfgyVgaCKQ7Pmb1YwB3M4MD0HpHGG/BVdUl4RQxZDNT2P6FUwhgPL4oA02jsoFCdC6xXLm5N3P4bKQFDFQSiWJRTUm3ovhQPSOEP+ii4pr4ghi6Gin5fSz9wnXM/FAWm0d1AoToTWK5Y3J+9+DJWBoIqTi6CcgxjCATiQkwPSOEP+ii4pr4ghi6Gip3L2FGfBtyVwQBrtHRSKE6H1iuXNybsfQ2UgqOIsoTm4R4YAHFgeB6RxhvwVXVJeEUMWQ7U8XqNl1BwOTMsBabR3UChOhNYrljcn734MlYGgikNjTtuY4Au+cKAMB6RxhvwVXVJeEUMWQ1WGW/Q0uMOB+XJAGu0dFIoTofWK5c3Jux9DZSCo4iAK8xUFakttl8wBaZwhf0WXlFfEkMVQ0eNL7nHuHf5PwQFptHdQKE6E1iuWNyfvfgyVgaCKMwUZiYnIwQE4UJoD0jhD/oouKa+IIYuhos9K9xnnw8G5cUAa7R0UihOh9Yrlzcm7H0NlIKjizK0RuB/EHQ7AgcQBaZwhf0WXlFfEkMVQwXX0Dg7AgVgOSKO9g0JxIrResbw5efdjqAwEVRwaMbYRwRM84UAdHJDGGfJXdEl5RQxZDFUdXKPnqQMcmA8HpNHeQaE4EVqvWN6cvPsxVAaCKg4iMB8RoJbUEg4c44A0zpC/okvKK2LIYqiO1RvugwUcgAMRHJBGeweF4kRovWJ5c/Lux1AZCKo4EeQjBiIGB+BAbRyQxhnyV3RJeUUMWQwVfVdb35EPnGydA9Jo76BQnAitVyxvTt79GCoDQRWndeKTP+INB+CAxQFpnCF/RZeUV8SQxVDBfYv7rMELODCeA9Jo76BQnAitVyxvTt79GCoDQRWHphvfdGAHdnCgXg5I4wz5K7qkvCKGLIaqXv6hDdQGDrTJAWm0d1AoToTWK5Y3J+9+DJWBoIpDw7fZ8NSNusGB7RyQxhnyV3RJeUUMWQzVdg7QI+ADB+DAvhyQRnsHheJEaL1ieXPy7sdQGQiqOPsSjesRJzgAB1rggDTOkL+iS8orYshiqOjFFnqRHOFpSxyQRnsHheJEaL1ieXPy7sdQGQiqOC2RnFwRZTgAB4ZyQBpnyF/RJeUVMWQxVPTD0H7gOrgCB4ZxQBrtHRSKE6H1iuXNybsfQ2UgqOLQYMMaDJzACQ60xQFpnCF/RZeUV8SQxVC1xUk0hHrBgfo5II32DgrFidB6xfLm5N2PoTIQVHFo7vqbmxpRIziwPwekcYb8FV1SXhFDFkO1Py/oJTCDA3BgGwek0d5BoTgRWq9Y3py8+zFUBoIqzjZS8R6iAwfgQKsckMYZ8ld0SXlFDFkMFf3Zan+SN9ytlQPSaO+gUJwIrVcsb07e/RgqA0EVp1ZCkxdiCwfggIcD0jhD/oouKa+IIYuhokc8PcJe+AMHjnJAGu0dFIoTofWK5c3Jux9DZSCo4tBMR5sJTMAEDrTPAWmcIX9Fl5RXxJDFULXPU7SGGsKBujggjfYOCsWJ0HrF8ubk3Y+hMhBUcWjkuhqZelAPOBDDAWmcIX9Fl5RXxJDFUMVwhZ4DRzgAB8QBabR3UChOhNYrljcn734MlYGgiiMC8YqYwAE4MCcOSOMM+Su6pLwihiyGip6dU89yL/C5Bg5Io72DQnEitF6xvDl592OoDARVnBrISw6IKByAA9EckMYZ8ld0SXlFDFkMFX0T3TfEg1NL54A02jsoFCdC6xXLm5N3P4bKQFDFWXrjcP8MDzgwTw5I4wz5K7qkvCKGLIZqntxFk6grHCjHAWm0d1AoToTWK5Y3J+9+DJWBoIpD05ZrWrAHezgwHQekcYb8FV1SXhFDFkM1HX/oTbCFA8vkgDTaOygUJ0LrFcubk3c/hspAUMVBMJYpGNSdus+dA9I4Q/6KLimviCGLoaKP597H3B8cz80BabR3UChOhNYrljcn734MlYGgipObqJyHOMIBOJCDA9I4Q/6KLimviCGLoaKXcvQSZ8CzJXFAGu0dFIoTofWK5c3Jux9DZSCo4iypSbhXhgIcWA4HpHGG/BVdUl4RQxZDtfPhk1UAACAASURBVBw+o13UGg7k4YA02jsoFCdC6xXLm5N3P4bKQFDFoUHzNCg4gzMcyMsBaZwhf0WXlFfEkMVQ5eUUPQzecGD+HJBGeweF4kRovWJ5c/Lux1AZCKo4iMP8xYEaU+MlckAaZ8hf0SXlFTFkMVT09hJ7m3uG91NyQBrtHRSKE6H1iuXNybsfQ2UgqOJMSUpiI3pwAA6U4oA0zpC/okvKK2LIYqjor1L9xblwb64ckEZ7B4XiRGi9Ynlz8u7HUBkIqjhzbQjuC7GHA8vmgDTOkL+iS8orYshiqJbNcTSO+sOBeA5Io72DQnEitF6xvDl592OoDARVHJoxvhnBFEzhQHkOSOMM+Su6pLwihiyGqjzP6HVqAAfmxQFptHdQKE6E1iuWNyfvfgyVgaCKgxDMSwioJ/WEA5dwQBpnyF/RJeUVMWQxVPQ7/Q4H4EAsB6TR3kGhOBFar1jenLz7MVQGgioOjRjbiOAJnnCgDg5I4wz5K7qkvCKGLIaqDq7R89QBDsyHA9Jo76BQnAitVyxvTt79GCoDQRUHEZiPCFBLagkHjnFAGmfIX9El5RUxZDFUx+oN98ECDsCBCA5Io72DQnEitF6xvDl592OoDARVnAjyEQMRgwNwoDYOSOMM+Su6pLwihiyGir6rre/IB062zgFptHdQKE6E1iuWNyfvfgyVgaCK0zrxyR/xhgNwwOKANM6Qv6JLyitiyGKo4L7FfdbgBRwYzwFptHdQKE6E1iuWNyfvfgyVgaCKQ9ONbzqwAzs4UC8HpHGG/BVdUl4RQxZDVS//0AZqAwfa5IA02jsoFCdC6xXLm5N3P4bKQFDFoeHbbHjqRt3gwHYOSOMM+Su6pLwihiyGajsH6BHwgQNwYF8OSKO9g0JxIrResbw5efdjqAwEVZx9icb1iBMcgAMtcEAaZ8hf0SXlFTFkMVT0Ygu9SI7wtCUOSKO9g0JxIrResbw5efdjqAwEVZyWSE6uiDIcgANDOSCNM+Sv6JLyihiyGCr6YWg/cB1cgQPDOCCN9g4KxYnQesXy5uTdj6EyEFRxaLBhDQZO4AQH2uKANM6Qv6JLyitiyGKo2uIkGkK94ED9HJBGeweF4kRovWJ5c/Lux1AZCKo4NHf9zU2NqBEc2J8D0jhD/oouKa+IIYuh2p8X9BKYwQE4sI0D0mjvoFCcCK1XLG9O3v0YKgNBFWcbqXgP0YEDcKBVDkjjDPkruqS8IoYshor+bLU/yRvu1soBabR3UChOhNYrljcn734MlYGgilMrockLsYUDcMDDAWmcIX9Fl5RXxJDFUNEjnh5hL/yBA0c5II32DgrFidB6xfLm5N2PoTIQVHFopqPNBCZgAgfa54A0zpC/okvKK2LIYqja5ylaQw3hQF0ckEZ7B4XiRGi9Ynlz8u7HUBkIqjg0cl2NTD2oBxyI4YA0zpC/okvKK2LIYqhiuELPgSMcgAPigDTaOygUJ0LrFcubk3c/hspAUMURgXhFTOAAHJgTB6RxhvwVXVJeEUMWQ0XPzqlnuRf4XAMHpNHeQaE4EVqvWN6cvPsxVAaCKk4N5CUHRBQOwIFoDkjjDPkruqS8IoYshoq+ie4b4sGppXNAGu0dFIoTofWK5c3Jux9DZSCo4iy9cbh/hgccmCcHpHGG/BVdUl4RQxZDNU/uoknUFQ6U44A02jsoFCdC6xXLm5N3P4bKQFDFoWnLNS3Ygz0cmI4D0jhD/oouKa+IIYuhmo4/9CbYwoFlckAa7R0UihOh9Yrlzcm7H0NlIKjiIBjLFAzqTt3nzgFpnCF/RZeUV8SQxVDRx3PvY+4PjufmgDTaOygUJ0LrFcubk3c/hspAUMXJTVTOQxzhABzIwQFpnCF/RZeUV8SQxVDRSzl6iTPg2ZI4II32DgrFidB6xfLm5N2PoTIQVHGW1CTcK0MBDiyHA9I4Q/6KLimviCGLoVoOn9Euag0H8nBAGu0dFIoTofWK5c3Jux9DZSCo4tCgeRoUnMEZDuTlgDTOkL+iS8orYshiqPJyih4Gbzgwfw5Io72DQnEitF6xvDl592OoDARVHMRh/uJAjanxEjkgjTPkr+iS8ooYshgqenuJvc09w/spOSCN9g4KxYnQesXy5uTdj6EyEFRxpiQlsRE9OAAHSnFAGmfIX9El5RUxZDFU9Fep/uJcuDdXDkijvYNCcSK0XrG8OXn3Y6gMBFWcuTYE94XYw4Flc0AaZ8hf0SXlFTFkMVTL5jgaR/3hQDwHpNHeQaE4EVqvWN6cvPsxVAaCKg7NGN+MYAqmcKA8B6RxhvwVXVJeEUMWQ1WeZ/Q6NYAD8+KANNo7KBQnQusVy5uTdz+GykBQxUEI5iUE1JN6woFLOCCNM+Sv6JLyihiyGCr6nX6HA3AglgPSaO+gUJwIrVcsb07e/RgqA0EVh0aMbUTwBE84UAcHpHGG/BVdUl4RQxZDVQfX6HnqAAfmwwFptHdQKE6E1iuWNyfvfgyVgaCKgwjMRwSoJbWEA8c4II0z5K/okvKKGLIYqmP1hvtgAQfgQAQHpNHeQaE4EVqvWN6cvPsxVAaCKk4E+YiBiMEBOFAbB6RxhvwVXVJeEUMWQ0Xf1dZ35AMnW+eANNo7KBQnQusVy5uTdz+GykBQxWmd+OSPeMMBOGBxQBpnyF/I0nkXfKU77ayLuq9e/PW94imviCGLoYL7FvdZgxdwYDwHpNF7CbtxseJEaL1iGcdkXcJQGXCrODTd+KYDu2Vg98a3ndc97o9P6R538qnd2e/+cEfd26i7NM6Qv5Cld33oy10alHc96R17GSvlFTFkMVRtcBHNoE5woB0OSKO9g0JxIrResbw5efdjqAwEVRyavJ0mp1ZlanW337h3d5nLXObg/w948B9gqL5Qpg778l8aZ8hfyJIMlYblUGOlvLTP+6p4ITe1I4jO2rcWXN9Gz1An6gQHvtZJ53bI4c63Fcer8Wm/Yu08dOILMFQGwCoOzYOAwoHtHPj1e9wLQ9WIiepzWRpnyF/I0rqh0tDcZayUl673vipeyE3tCKKz+jjz7+36AT7gAwfa4oB0bocc7nxbcbwan/Yr1s5DJ74AQ3UpwP2f+VdxaPS2Gp165a8Xhio/5hE8l8ZNNV82GSoNz03GSnnpOu+r4k11n/24OiuiPsRos6+oG3WbOwekc33tG/NvxfFqfNqvWGPyiNyDoboUTT0ApEF/3PFndre62zP58aUG/8v73MWstvvDULX5AJEG0M/f72UHv+cUMdDGxlg3VhqMY+Ot71O8yKG5KZbOqq1HyafNHqVu1K1GDkjnNung0HXFWdfsMV8r1tCzp7oOQ3UpsjJUKuYdTji7O/WMj3Yf+MRXMVYYKziwgQMYqjaHfhpANRgq6a2MVfoPWSk3rXtfcw5anVXjQxA5tdmn1I261cYB6ZzXlCiOV+PTfsXy5uTdj6HaYKhU5Ls86lyM1YaH6doanXzyDx8MVX7MI3ieBlBNhkp6e4eHv737qXs+D0OF5vIfseAAHKiQA1HmRXGk/Z5XxfIaIu9+DNUOQ6UiY6zKPjh++vNf695y9vndc087o3v04/6ke/zJT+1O++tXdK//h/d0b3jbe1f/f/u7PrRVhN/9vk8e7Hvko5/YPeWpz+n+/k3v7D72ma9s3fPeD366O+MVb+jO/Pu3dSkPPdCms558yjO7xz7hKd3LXvmm7sOf+J+r93TNpteoPN78j+cfnH/Cox7XPfv5L+k++un/OJTDpz73X92rX3/Owb0+4g+f0D38kX/UPe2Zpx3gtim39fWET4rxhCc97QD31775XV2Km67b11D988e/0L38NW/tnvjkp3cnPfZJ3ekvfmV37nsvPITr+vkW/rvuez0GXx/jbcKiNkOVvkP1V6+56OBHrVNu0l3va85Bq7Pg2mGugQd4wIH5cEA65zUfiuPV+LRfsbw5efc3Zahq+C+qGKv8wpBMz81ufsvVp8npY7o3vf7ju/75kKlIYv66t7y7u/H33MSMcbnLXa77gxMf0130bxcf2Zf23uFOd13tO+uN7+iSobjBDW+8WlMeKc6fnPosM4YGSmQeDzvh0UdySH8TKp31/gv+tTv+gb/ffevVrn7kGuV7y1v9TPemt//T1nyTWbz85S9/JMY1r3XtA5M51FAl83nTm93iSBzlct3rfVv34pe91sxlHf9t9y2ced3ep2kA1aCn+lE//QFgDcaIIZt70Cp3uLede+ADPnCgXQ5I57zmQ3EitF6xvDl592OoTjxn1H8NxVjlEYT0HSjrgV4P4tbruqE6+U//fFCMW/3UbbqPf/Y/jzzU3/4Ov7YyAvd7wEO6K1zxiquvrfP/6Il/diRGGiCRefzy7e9s5pC+c5fO+qEfuZn5/nq+V7vaNboPXPi5I/mm78Td/Z732RrjG77hG7pv/MZvXF2z6e9Qpe+IXfnKV1ldt55D/+v7/+5Dj+TSx3/XfTOoh/Xl1ANo/XdS14fmupHSIFNe69eP/VrxFH/KV50FB4dxEJzACQ60xwHpnFdLFWestvf3KZY3J+9+DNVIQ6Vi3v0x7+zOfv8XjzwEIhR+oUgP9d97k+9fPYin7wq98IxXd+nH5dKP393iJ3569V56KH/cyad2f/a0565+HC3VIP2YYHrw10P7D/7wj3bPO/2l3dnv/nD3irP+obvr3X9r9V66Jv0o2nrt+g/0ipO+G/Vzv3C77klPeUb3O8f/3iFjcY1rXuvIj95NlUcyNHe+yz26e/zmfbvvvMGNumc8+4UH+X/Lt17t4L6SYfrt+z2oO/1vzuxe99b/0b34717X3fmuv3Honi0Tk34UT/eaXq//bd/RPeghJx78yN9P/vRtD2Gq6yxDlX40r49/MlYPeNDDumf95Yu7Zz7vb7pkUK/6zd9y6KyUY78GFv6b7ru/j39v7sGpB9AmQ7XJSGmQKS/pq/dV8RR/yledBe828w5swAYOtM0B6ZxXSxXHq/Fpv2J5c/Lux1CNNFR8h2p6UUjfadLD+uW/6Zu6d/7Txw49aF940Ze6b/+O71pdk36/al2sb/vzv7x6/yd+8tZHjE66/oG/d8Lqmmtd+zpHrll/oE9mJf0YYv+sZzznRasYKefnnv63h96fIo8rXvFK3d+9+i2HzlFOD37oIw5+R+mCT33JfD+ZMGF70x/98SPX/Ph//8nV+9e93vWPYJ9MTzpfMdKrZah+9rhfWV1znetez/zdrWT0kglVrB/4oZse+p2qdfy33bfun9ft/Tn1AFo3VLuMlAaZ8ooYsrkHrXKHe9u5Bz7gAwfa5YB0Tpo99lVxIrRescbmErUPQ7WnocJI5ROC5/zVS1YP2be+7XFHHvqTKP/WfY5fXfPQh//hoWveeu77V+9d9rKX3fj7Qh/62Oe7q1zlqqtr0+9J9QW//0Cfficp/Q5V//307/TdtP7vVaXfPdI1U+SRzMdfPPevV2forKGvL3n561f3m8xMf1/6bprMzbZzXvrKN3VXutKVV9euG6r045r9OKc8/XmHzumfmX73rH9t+hAMvd/Hf1s+up7X3T069QCSoRpqpDTQlFfEkMVQ7eYBvQJGcAAO7MMBabQ0e+yr4kRovWKNzSVqX1OGKoE21f/0ALCpuBip/KKTPoVPD9m/+Ct3XD1g95v/N+99/9U1D/mDkw5dk75jpf3pAyn6+9b//cM3/bHVtelH0frv9x/o06fk9d/r/7t/XfpACL03RR6b8NCZu17f9s4Pru43YdS/Pv1IpHBLRvPCi7586P3+tds+lOLUZzx/FefqV7+m+ftpipU+IbH/HS/96GJ6v4+r97513tJfpx5A513wle60sy7q9GETQ3VbeW3S4X3XFW/o+Z7rdNbSucX955+VYA7muTggnfNoZdqrOPtqunW9Ynlz8u7HUF2K4CZDhZEqJ1Tpx9n0YH/t61y3W//xtU/+6//qbvTd37u6Jn1seF9UHvWYP169l37nJv3I36b/9z/4Iu3rx+k/0J940uMPvde/Lv2ukvK9+z1/e3Vd7jz6OaXvnKXv9qQPhjjpMScf/L7XPe91v0MmZd1QpU8K1H2kT+brx1v/9zZD9ZCHPWoVJ32i4Pre9a+//wd+eHX9wx/x2NX1Q/Ffj8fXm3u3lgG0PsCUlzU0x6wp3vo5U3yts+DdZt6BDdjAgbY5IJ3zaqjijNH19T2K5c3Jux9DdSmC64bqDieczR/0LfxH5T7yyS8e+t2a9HtI6QMpkiCf/+HPHvo48/QhEeu/Y7XrU+pkGtZf1z+lb+gDff93sfqGKnceGljJGF3v+t++Minr99n/WnvSa/qQDb133C/96srY9K/Rv7cZqv7Hnaff2dKeTa//vfchI/f5nd9dXT8U/01xWT86wGsZQOsDTHmtD8yxXyve+jlTfK2z4NtRvoEJmMCBeXBAOufVUMUZq+39fYrlzcm7H0N1KYIyVOln/o87/syDPzCJAJQXgP6HJ+ghv//7Tlpb/65Sql3fyFzlqt/c3e5X77zz/+nT8tIfku3XfugD/RBDlSOPlHv6fTJho9f0Y3f/7ft+4ODvQaVXrafX/v3e+74PWL2XTGz/vfV/3+037r26dv13qPofcX7HX7vb1jgpbvoulnJKn/6ns4bir+t53d23tQyg9QGmvPrD0vNvxVs/Z4qvdRb8280/MAIjONAmB6RzXg1VHI++a69ieXPy7sdQXYpg/2f+VRwavmzDpw+H6H/kth62+6/pQxFOeNTjzD/K+4g/fMLqAf3HfvwnVg/o+9Z16AP9JkOVO49Xve4fV/edsPrFX77DwcfH9+/7Hed99NA1/ff6+VqfANi/dtt3qJLBUq3Sj1r291n/vuGNvmd1ffqbXbpmKP66ntfdfSuN8w6Q6P3KS4PS+6p40Xla8XQW/NvNPzACIzjQJgekc5YG7rOmOF6NT/sVa5/zp7gWQ2WgquLQ8GUbvm9Q0u9HPf+Ff9f93u8/srvL3X7z4G8iPfUv/qp7zwcuWj14r9cr/a0jPdBf81rXPvT3qdav3fb10Af6fr79H/nLnUf/d7Zu8n0/2KXfNVu/v22GKuEq3NLH1acPjFjfr6/734Va/w7Vk/7sL1Zx0ncV049wat/667vf94lDf8ur/7eohuK/HpOvN/evNM6Qv6JLyitiyOYetMod3m3m3Rywef8F/9q96/yPb/2QnTH3mWbZnz/rBQd/S/H8j/zLRq0cEjsy1pDzuGbenO/XVzrnHRSKE6H1iuXNybsfQ2UgqOL0ScS/8wvGT93651YP5OkP6O5bg/T3jWQM0uvTn3X63jHSmUMf6DcZqtx59H936ed/8XbmPfc/8CNh08c2fWewj1sa8v339e/+p/il69cN1cte9eZDcdJHo2vv+mv/u1npu47vu+BfV9cOxX89Jl9v7llpnCF/RZeUV8SQxVBtrj+9MRyb8z70mS591z59wmj/7x4mzUvfVU+/7/mWc9630qux2P7Qj9xspZf3/K3fccWLjDX2ftg3nGMtYSWN9g4KxYnQesXy5uTdj6EyEFRxWiL5HHNNg6r/YP+9N/n+7g53/vUufffnvvd/8MF3q/74T57epe9mbPpo7/SjZoqR/kDtK1/7dnNQXfRvF3cvftlru7960cuPvD/0gX6ToUq1KZXHN13hCt0HLvzcoXtK95l+l0u4pNd1/nzf9//Q6v2rXe0aXfojy/1r0gderP845rqhStf/yI/efBUnGaW//tuzDsVJ1ySjlT5URPms/z2xofj38+Pf24e5NM6Qv6JLyitiyGKotnOAHtmNzxmveEOX/ti7tGnb671++3jzR8+H4py0WvF/8qdve0QnU5xk3NLfEHzdW95tvq+zhsTStbzu5gEYHcNIGu0dFIoTofWK5c3Jux9DZSCo4tBEx5qoBBZvf9eHuute79tWQ0bDxnpNf5w2mav1PNN3h9If9dWe9PHp6Ts4j3z0Ew++Y/W4k0/t0u8BXee61zu4Jn0M+3qMoQ/02wxVzjzSj0bqftPrd9/4v3W/f8Kju3Sv6d4TBv3307/X77n/43rp/W/+lm89+ECP9GERP/BDNz2yP11jGaq/PfONh65NximZy4TV/X/3od3Nb3GrQ++nOq7/iOFQ/Nfvga839680zpC/okvKK2LIYqg215/e2I1N+g87/dkhzUw/Pp40Nf04tNb0+rJXvumIlg7Fuv8f3X73wQ8/Eif9CQydk17Tj0lvir0r1qZ9rO/mxdIxkkZ7B4XiRGi9Ynlz8u7HUBkIqjhLb5zS9//Bj/7bwafSaYikAdb/L29a77+e8vTnHRkyyRyk7470r9v072QW1u976AP9NkOVYubKI52VPp1v0z2m9et/23ccwnL9nj/1uf86+E7gthjJfD76cX+yOscyVCnuY5/wFPPhYz12+jtUbz33/aPxX78Hvt78cCCNM+Sv6JLyihiyGKrN9ac3tmPT/2PsSafSd+PTj/2d+94LV/qUfqohGajb/OwvrDTQ+gmHoVin3zF9+Wve2qX/kGjtSef1NTP9Hqx1XVrbFWvTPta38wJ8vhb2ARCRWq9YRYdX13UYKqMCKg7NU1Zcful2d1oNkPQjaP/0z589GCDpv9SlH/F7/T+85+Bhvf8jGenjwK26nfOeCw5MxvqPumlApe+E/ea979+97Z0fPLK//0l26bs8Vvy09vBH/tEq3/QjidZ1OfJI56bv8iSDd4UrXnGVU7rXdP/pxybTL1b/zG1+/uC9ZDatXHVP3/ldNzwUI303L/3Y5Yc+9vmu/x2o9F2wTXHe+LbzuvR3pq585ascipX+C3AyZsmMfewzXzH3D8V/09msH+1jaZwhf0WXlBeG6mjN4HEeTNKH+Nzghjde6VTSzBee8WpTm1ST553+0u5mN7/lRjOk6zyv+xgqzznszcOzVnGWRnsHheJEaL1ieXPy7sdQGQiqOK0Sfg55p/8SKLOTfkQtGZFN9/WEJz1tdW16QP/op/9j47UpRvrjwOnn0F/wkld1r379OVs/KTBdnwZZMnP9D0rYlEv6g8Pp/8n0bbpG61PmoTMSFm/+x/O701/8yu4Nb3vvoZ/x//hn//Pgb25t+v0zxUiv6W9zpYeKf3jHB47cV/odrYTPkHtO16TavuiM13Tp490v+NSXjsTrn5v+vQ/+63v52uahNM6Qv6JLyitiyPIdKrv29MR2XNKPjmv2pNcnn/LMnRplYZo0M/0O1pl//7aVNiYtTt+xT3/qI31ybX9Wpbnx0le+qUsf5tPXxfQfrpJ2p9+b6ud1+t+cebCe3kv/7+v4plhWnslAvuXs87u/eO5fH/wofLrfpM1Jd63rtfbPH//CwXfUnvjkp3cnPfZJBzMmafuQOaAYvG7nYo34SKO9g0JxIrResbw5efdjqAwEVZwaybyUnJLZ0fD4rht+91ZhTx9KoWu/9WpXR9C/0J5IL4XXtdynNM6Qv6JLyitiyGKo0IEx/Xa963/7ap5c+zrX3fid812x+5+2mj459WEnPHoVV/MqfbiP4vSv7/+h+nve635H9ml///UP/+jJO2PpLL0mU9b/CY9+vPSJhulPfuhavaYfSbzpzW6xMaf00x7pg490Pa/z6kNptHdQKE6E1iuWNyfvfgyVgaCKgxCUE4L0naO+uCfht+qR/h7Ijb/3+1bXpp9nt65jrVwtwb4+7KVxhvwVXVJeEUMWQ1Uf72rXgvTdoP7c6RubfXPv/+5t/+/19eM//uSnruZV//oTT3r8ar3/I8/9vev/Th+0pBw3xdL76fW37/egQ/e6Hk9fpx/r1r6nPfO0Iz+2revWX9OHDmkfr/PpRWm0d1AoToTWK5Y3J+9+DJWBoIqDCJQTgU/8y1e7q1/9mocEP/1XseMf+PsHvxz8gAc9rPuFX/7VQx+3fcUrXung96qoW7m6gX0b2EvjDPkruqS8IoYshqoNLtakGet/g69vJvbNs29qZDbSj6/f+S736O7xm/ftvvMGN+qe8ewXrkxH//q+oUo/6pd+P/cPTnzMoXl4n9954MF6ei/9v/+BGZti6R7SjxYqp/Sacknm8S9f8LKDT4P94Zv+2Op9/Q3I9OOK/T+VkX4fNs3h9F2sZz7vb7r0CbBX/eZvWe1Lcft/oF1n89p2X0qjvYNCcSK0XrG8OXn3Y6gMBFUcGr9s47/mDece+XtJ/SHQ/3cS8jQMqFnZmoF/G/hL4wz5K7qkvCKGLIaqDS7WpBnPeM6LDhmCbb+7uyvvvqlJsyr9B7/0B9U37etf3zdUun6fD6XYFivF+f4f/JHVfaZPV02/c6Vz9PqUpz6nu/o1rnXwwUNp7WeP+5XVnvTBROlDoXStXtOfB0l/+kKzOX1qLr9TNa8+lEZ7B4XiRGi9Ynlz8u7HUBkIqjgSCV7LCUL66PT0YxHpE5TS30KSUKePT0+fDpc+qS79ku/63y6iZuVqBvb1Yy+NM+Sv6JLyihiyGKr6eVibVvS/C5S+G5M+sGFsjn1Tk+ZW+tCHbbH6109pqF71+rNXczTlZf2x9fU8k3nS7E2v1p8n0Z70h9r716Yf39d7vLbfk9Jo76BQnAitVyxvTt79GCoDQRWH5q+v+dMn0yXzxH/1qq829Es7NZHGGfJXdEl5RQxZDFU7fKxFO+7zO7+7MgPpP9p58uobpF/8lTvujNW/fkpD9fRnnb66x/TBE0Nm6anPeP5qT/pR/DSHN2GT5nP6bpxMVf/HGjftYb2dXpVGeweF4kRovWJ5c/Lux1AZCKo4NHk7TU6tqBUcGM4BaZwhf0WXlFfEkMVQDecDvXMJVumT8mQE0uv5H/mXjcZhF2a7DNL6/l3XR/3IX/p7gbrHn7ntcYPu7yEPe9Rqzy1v9TM796QfI9QZD3/EY3dev44FX9fbu9Jo76BQnAitVyxvTt79GCoDQRWHpq63qakNtYED4zkgjTPkr+iS8ooYshiq8fxYam+l38OVEUivf/+md442A7sMFvFYZwAAIABJREFU0jrGu66PMlTpD7vrHu96998adH/9j3RPH6qxnvv61+mPuOuM9F2/9ff5ut3elEZ7B4XiRGi9Ynlz8u7HUBkIqjg0fbtNT+2oHRzYzAFpnCF/RZeUV8SQxVBtrj+9YWPzxredtzICyRB4flxtl0Far8Gu66MM1S/d7k6re7zzXX9jkNnpf+z7HX/tbjv3pO9iyVClT/9bv1e+tvnXAi7SaO+gUJwIrVcsb07e/RgqA0EVpwVyk2O7wkTtqF0pDkjjDPkruqS8IoYshor+2re/LrzoSysjkAxB+jCkfWPo+l0GSdfpddf1UYYq/X0omZ1b/dRtBt3fAx78B6s9P/GTt96554Y3+p7V9Sf/6Z/vvF4Y8Fp/z0qjvYNCcSK0XrG8OXn3Y6gMBFUcmrv+5qZG1AgO7M8BaZwhf0WXlFfEkMVQ7c8Leulr3S16P66WjEf6m01jcNllkNZj7rp+3VC99dz3b8xrW6wnPvnpK7NzzWtdu0t/83E9l/Wvn/Rnf7Hac5WrXLX7yCe/uHHPu9/3iS79vS2ZNv4W1bz6UBrtHRSKE6H1iuXNybsfQ2UgqOKsiwpfz0sYqCf1XCoHpHGG/BVdUl4RQxZDRX+P6e/1jxW/8ffcpDv73R/eaCDSGcncpN9H6v/O1TZTY+U15PrLX/7yK6Py7Oe/ZGNO22Klj0mX2Umvf/6sF2yMk/5u1ns/+OnuZa9686E96aPRrXtIa/3vZl3pSlfu3nfBv268dlMM1uvtXWm0d1AoToTWK5Y3J+9+DJWBoIpDU9fb1NSG2sCB8RyQxhnyV3RJeUUMWQzVeH4svbf6v2eUTEf6G4jPPe2M7qOf/o9D5uDc917Ypd8Ruvw3fdOB4ej/ztU2U2PhO+T69DHnMkO/dZ/jD+XSj7kt1qc+91/dDW5441Wca1/nut1Zb3zHoVj/9M+f7dLvV6WzHvP4Pz1470d+9OarPckoWX+/Khmty13ucqvrHvrwPzwUt58j/26zP6XR/UFx3gVf6U4766Luqxd/vb+89d+KE6H1irX1wAxvYqgMkFUcGr7Nhqdu1A0ObOeANM6Qv6JLyitiyGKotnOAHtmMT/qO1PWu/+0rYyATk36U7Sbf94Nd+tCFZG4ue9nLHrrmGc950cpAbDM1FvZDrr/1bY87dN4P/vCPdve9/4O79F20/neNdsVK35XSPaXXK1zxit3P/cLtuuMf+PtdMpNXvvJVVu8/7uRTD+7pb89842ot7UnGKf0+1QN/74Qu/V7WzW9xq0PvX+Oa1zr4m5HWvbK2mXu1YyON7g+Kd33oy13S27ue9I7BxkpxIrResfo5lfg3hspAXcWpndjk164oUTtqV5ID0jhD/oouKa+IIYuhosc8PfaBCz93YC76xmPbv9Mn4V140ZcnNVQvefnrD5mWfj7pb2jpfncZqvT7WPe81/02xlLc9J2o/nevHvuEp6y+G6drrNf0d6i2/Y6X8uS1vR6VRvcHhQyVdHuIsVIc7fG8KlY/pxL/xlAZqKs4NHt7zU7NqBkc2M0BaZwhf0WXlJdnuPb3Kl6Om9JZ8G83/1rC6NRnPL+77c/9Ynfd613/iAG52tWu0d3p1+7evfK1b1+ZGd3br9/jXqvr9V0evWe9Dr3+z5723O4qV/3mVexkaNKPHP7NS/9+lcPQWM/6yxd333XD7z4UK8VLRir9yJ/1u2Ppo+XT35nqfxcr7UnfrbvRd3/vwe9QfewzX1nlYt0ra+32iHSur6nrhkoavM1YKY6u9bwqVj+nEv/GUBmoqzg0fbtNT+2oHRzYzAFpnCF/RZeUl2e49vcqXo6b0lnwbjPvWscmfcBC+qCG17zh3O7d7/tkl77Ts+me0nvpd5GGfijDPtenT+Z7y9nndy9/zVu7V73uH4986t4+sVL+H/7E/+xecdY/dKe/+JUHH6zx8c/+58b70v1++vNf69LvkL3ojNcc5HDBp760c4/28tpuj0jn+pq6yVBJiy1jpTi6xvOqWP2cSvwbQ2WgruLQ9O02PbWjdnBgMwekcYb8FV1SXp7h2t+reDluSmfBu828AxuwgQNlOPDcV37i4Pec+vqY+999YyW9jMhBsXLo/LYzMFQGOioOjV+m8cEd3OHAtByQxhnyV3RJeUUM2RRD8XLclM6Cu9NyF3zBFw7sz4EaDJV0PRmr444/s7vV3Z4ZYvKkvTl0ftsZGCoDHRWHpt2/acEMzOBA/RyQxhnyV3RJeWnwel8VL8dN6Sz4Xz//qRE1WhoHajJU0vU7PPztGKqxw0kDR2Du86q9Y8/eZ5/OWlrDcb8MGTiwDA5I4/bRxRzXKq99ZsO2axUvZ+700DJ6iDpT55Y4UJOh4jtUARNJw23bANz0nvYGpLEzhM5qqVnIFXGHA3BgKAekcTvFMPMFymvTHNh3XfFy3IbOGloDrqNf4QAcyMWBGgwVv0MVOIk0cPYdiul67Q1MZ2MonZWL6JyDqMIBOJCTA9K4jSJY6A3lNWZGWHsUL8ft6KycdeQsdAMOwIGcHJDO9TWVT/m7BA1+h6rPikv/LcLkJClnIYpwAA7k4oA0zpC/okvKyzJHY9YUL8dN6axcNeQc9AIOwIHcHJDO9TV1k6Hqf0eqf336t+KM0fX1PYq1fkburzFUBuIqTm6ich7iCAfgQA4OSOMM+Su6pLzWB+bYrxUvx03prBz14wx0Ag7AgRIckM71NXXdUG0zUtqnOGO1vb9PsRS71CuGykBexSlBVs5EJOEAHJiaA9I4Q/6KLimv/rD0/FvxctyUzpq6dsRHH+AAHCjFAelcX1NlqIYYKe1THI++a69iKXapVwyVgbyKU4qwnItYwgE4MCUHpHGG/BVdUl4alN5XxctxUzpryroRG12AA3CgJAekc31NPe+Cr3SnnXVR99WLv95f3vpvxfFqfNqvWFsPzPAmhsoAWcUpSVrORjThAByYigPSOEP+ii4pr4ghm3vQKvepakZc9AAOwIHSHJDOeQeF4kRovWJ5c/Lux1AZCKo4pYnL+YgnHIADU3BAGmfIX9El5RUxZDFU9M4UvUNMeLVkDkijvYNCcSK0XrG8OXn3Y6gMBFWcJTcN987QgAPz5YA0zpC/okvKK2LIYqjmy1+0idrCgTIckEZ7B4XiRGi9Ynlz8u7HUBkIqjg0bJmGBXdwhwPTckAaZ8hf0SXlFTFkMVTTcogeBV84sDwOSKO9g0JxIrResbw5efdjqAwEVRzEYnliQc2p+RI4II0z5K/okvKKGLIYKnp5Cb3MPcLznByQRnsHheJEaL1ieXPy7sdQGQiqODlJylmIIhyAA7k4II0z5K/okvKKGLIYKvopVz9xDlxbCgek0d5BoTgRWq9Y3py8+zFUBoIqzlIahPtkGMCBZXFAGmfIX9El5RUxZDFUy+I0Gka94cD0HJBGeweF4kRovWJ5c/Lux1AZCKo4NOf0zQnGYAwH8nNAGmfIX9El5RUxZDFU+XlFL4M5HJg3B6TR3kGhOBFar1jenLz7MVQGgioOwjBvYaC+1HepHJDGGfJXdEl5RQxZDBX9vdT+5r7h/lQckEZ7B4XiRGi9Ynlz8u7HUBkIqjhTEZK4iB0cgAMlOSCNM+Sv6JLyihiyGCp6rGSPcTb8myMHpNHeQaE4EVqvWN6cvPsxVAaCKs4cm4F7QuThAByQxhnyV3RJeUUMWQwVPEfr4AAciOWANNo7KBQnQusVy5uTdz+GykBQxaERYxsRPMETDtTBAWmcIX9Fl5RXxJDFUNXBNXqeOsCB+XBAGu0dFIoTofWK5c3Jux9DZSCo4iAC8xEBakkt4cAxDkjjDPkruqS8IoYshupYveE+WMABOBDBAWm0d1AoToTWK5Y3J+9+DJWBoIoTQT5iIGJwAA7UxgFpnCF/RZeUV8SQxVDRd7X1HfnAydY5II32DgrFidB6xfLm5N2PoTIQVHFaJz75I95wAA5YHJDGGfJXdEl5RQxZDBXct7jPGryAA+M5II32DgrFidB6xfLm5N2PoTIQVHFouvFNB3ZgBwfq5YA0zpC/okvKK2LIYqjq5R/aQG3gQJsckEZ7B4XiRGi9Ynlz8u7HUBkIqjg0fJsNT92oGxzYzgFpnCF/RZeUV8SQxVBt5wA9Aj5wAA7sywFptHdQKE6E1iuWNyfvfgyVgaCKsy/RuB5xggNwoAUOSOMM+Su6pLwihiyGil5soRfJEZ62xAFptHdQKE6E1iuWNyfvfgyVgaCK0xLJyRVRhgNwYCgHpHGG/BVdUl4RQxZDRT8M7QeugytwYBgHpNHeQaE4EVqvWN6cvPsxVAaCKg4NNqzBwAmc4EBbHJDGGfJXdEl5RQxZDFVbnERDqBccqJ8D0mjvoFCcCK1XLG9O3v0YKgNBFYfmrr+5qRE1ggP7c0AaZ8hf0SXlFTFkMVT784JeAjM4AAe2cUAa7R0UihOh9Yrlzcm7H0NlIKjibCMV7yE6cAAOtMoBaZwhf0WXlFfEkMVQ0Z+t9id5w91aOSCN9g4KxYnQesXy5uTdj6EyEFRxaiU0eSG2cAAOeDggjTPkr+iS8ooYshgqesTTI+yFP3DgKAek0d5BoTgRWq9Y3py8+zFUBoIqDs10tJnABEzgQPsckMYZ8ld0SXlFDFkMVfs8RWuoIRyoiwPSaO+gUJwIrVcsb07e/RgqA0EVh0auq5GpB/WAAzEckMYZ8ld0SXlFDFkMVQxX6DlwhANwQByQRnsHheJEaL1ieXPy7sdQGQiqOCIQr4gJHIADc+KANM6Qv6JLyitiyGKo6Nk59Sz3Ap9r4IA02jsoFCdC6xXLm5N3P4bKQFDFqYG85ICIwgE4EM0BaZwhf0WXlFfEkMVQ0TfRfUM8OLV0DkijvYNCcSK0XrG8OXn3Y6gMBFWcpTcO98/wgAPz5IA0zpC/okvKK2LIYqjmyV00ibrCgXIckEZ7B4XiRGi9Ynlz8u7HUBkIqjg0bbmmBXuwhwPTcUAaZ8hf0SXlFTFkMVTT8YfeBFs4sEwOSKO9g0JxIrResbw5efdjqAwEVRwEY5mCQd2p+9w5II0z5K/okvKKGLIYKvp47n3M/cHx3ByQRnsHheJEaL1ieXPy7sdQGQiqOLmJynmIIxyAAzk4II0z5K/okvKKGLIYKnopRy9xBjxbEgek0d5BoTgRWq9Y3py8+zFUBoIqzpKahHtlKMCB5XBAGmfIX9El5RUxZDFUy+Ez2kWt4UAeDkijvYNCcSK0XrG8OXn3Y6gMBFUcGjRPg4IzOMOBvByQxhnyV3RJeUUMWQxVXk7Rw+ANB+bPAWm0d1AoToTWK5Y3J+9+DJWBoIqDOMxfHKgxNV4iB6RxhvwVXVJeEUMWQ0VvL7G3uWd4PyUHpNHeQaE4EVqvWN6cvPsxVAaCKs6UpCQ2ogcH4EApDkjjDPkruqS8IoYshor+KtVfnAv35soBabR3UChOhNYrljcn734MlYGgijPXhuC+EHs4sGwOSOMM+Su6pLwihiyGatkcR+OoPxyI54A02jsoFCdC6xXLm5N3P4bKQFDFoRnjmxFMwRQOlOeANM6Qv6JLyitiyGKoyvOMXqcGcGBeHJBGeweF4kRovWJ5c/Lux1AZCKo4CMG8hIB6Uk84cAkHpHGG/BVdUl4RQxZDRb/T73AADsRyQBrtHRSKE6H1iuXNybsfQ2UgqOLQiLGNCJ7gCQfq4IA0zpC/okvKK2LIYqjq4Bo9Tx3gwHw4II32DgrFidB6xfLm5N2PoTIQVHEQgfmIALWklnDgGAekcYb8FV1SXhFDFkN1rN5wHyzgAByI4IA02jsoFCdC6xXLm5N3P4bKQFDFiSAfMRAxOAAHauOANM6Qv6JLyitiyGKo6Lva+o584GTrHJBGeweF4kRovWJ5c/Lux1AZCKo4rROf/BFvOAAHLA5I4wz5K7qkvCKGLIYK7lvcZw1ewIHxHJBGeweF4kRovWJ5c/Lux1AZCKo4NN34pgM7sIMD9XJAGmfIX9El5RUxZDFU9fIPbaA2cKBNDkijvYNCcSK0XrG8OXn3Y6gMBFUcGr7Nhqdu1A0ObOeANM6Qv6JLyitiyGKotnOAHgEfOAAH9uWANNo7KBQnQusVy5uTdz+GykBQxdmXaFyPOMEBONACB6RxhvwVXVJeEUMWQ0UvttCL5AhPW+KANNo7KBQnQusVy5uTdz+GykBQxWmJ5OSKKMMBODCUA9I4Q/6KLimviCGLoaIfhvYD18EVODCMA9Jo76BQnAitVyxvTt79GCoDQRWHBhvWYOAETnCgLQ5I4wz5K7qkvCKGLIaqLU6iIdQLDtTPAWm0d1AoToTWK5Y3J+9+DJWBoIpDc9ff3NSIGsGB/TkgjTPkr+iS8ooYshiq/XlBL4EZHIAD2zggjfYOCsWJ0HrF8ubk3Y+hMhBUcbaRivcQHTgAB1rlgDTOkL+iS8orYshiqOjPVvuTvOFurRyQRnsHheJEaL1ieXPy7sdQGQiqOLUSmrwQWzgABzwckMYZ8ld0SXlFDFkMFT3i6RH2wh84cJQD0mjvoFCcCK1XLG9O3v0YKgNBFYdmOtpMYAImcKB9DkjjDPkruqS8IoYshqp9nqI11BAO1MUBabR3UChOhNYrljcn734MlYGgikMj19XI1IN6wIEYDkjjDPkruqS8IoYshiqGK/QcOMIBOCAOSKO9g0JxIrResbw5efdjqAwEVRwRiFfEBA7AgTlxQBpnyF/RJeUVMWQxVPTsnHqWe4HPNXBAGu0dFIoTofWK5c3Jux9DZSCo4tRAXnJAROEAHIjmgDTOkL+iS8orYshiqOib6L4hHpxaOgek0d5BoTgRWq9Y3py8+zFUBoIqztIbh/tneMCBeXJAGmfIX9El5RUxZDFU8+QumkRd4UA5DkijvYNCcSK0XrG8OXn3Y6gMBFUcmrZc04I92MOB6TggjTPkr+iS8ooYshiq6fhDb4ItHFgmB6TR3kGhOBFar1jenLz7MVQGgioOgrFMwaDu1H3uHJDGGfJXdEl5RQxZDBV9PPc+5v7geG4OSKO9g0JxIrResbw5efdjqAwEVZzcROU8xBEOwIEcHJDGGfJXdEl5RQxZDBW9lKOXOAOeLYkD0mjvoFCcCK1XLG9O3v0YKgNBFWdJTcK9MhTgwHI4II0z5K/okvKKGLIYquXwGe2i1nAgDwek0d5BoTgRWq9Y3py8+zFUBoIqDg2ap0HBGZzhQF4OSOMM+Su6pLwihiyGKi+n6GHwhgPz54A02jsoFCdC6xXLm5N3P4bKQFDFQRzmLw7UmBovkQPSOEP+ii4pr4ghi6Git5fY29wzvJ+SA9Jo76BQnAitVyxvTt79GCoDQRVnSlISG9GDA3CgFAekcYb8FV1SXhFDFkNFf5XqL86Fe3PlgDTaOygUJ0LrFcubk3c/hspAUMWZa0NwX4g9HFg2B6RxhvwVXVJeEUMWQ7VsjqNx1B8OxHNAGu0dFIoTofWK5c3Jux9DZSCo4tCM8c0IpmAKB8pzQBpnyF/RJeUVMWQxVOV5Rq9TAzgwLw5Io72DQnEitF6xvDl592OoDARVHIRgXkJAPaknHLiEA9I4Q/6KLimviCGLoaLf6Xc4AAdiOSCN9g4KxYnQesXy5uTdj6EyEFRxaMTYRgRP8IQDdXBAGmfIX9El5RUxZDFUdXCNnqcOcGA+HJBGeweF4kRovWJ5c/Lux1AZCKo4iMB8RIBaUks4cIwD0jhD/oouKa+IIYuhOlZvuA8WcAAORHBAGu0dFIoTofWK5c3Jux9DZSCo4kSQjxiIGByAA7VxQBpnyF/RJeUVMWQxVPRdbX1HPnCydQ5Io72DQnEitF6xvDl59xcxVLr5Ma/eGx6yX3m1TnzyR7zhABywOCCNG6KHOa9RXhFDFkMF9y3uswYv4MB4DkijvXNBcSK0XrG8OXn3Y6gMBFUcmm5804Ed2MGBejkgjTPkr+iS8ooYshiqevmHNlAbONAmB6TR3kGhOBFar1jenLz7sxoqb7K59qs4NHybDU/dqBsc2M4BaVwuTR16jvKKGLIYqu0coEfABw7AgX05II0equmbrlOcCK1XrE1n5VrHUBlIqzj7Eo3rESc4AAda4IA0zpC/okvKK2LIYqjoxRZ6kRzhaUsckEZ7B4XiRGi9Ynlz8u7HUBkIqjgtkZxcEWU4AAeGckAaZ8hf0SXlFTFkMVT0w9B+4Dq4AgeGcUAa7R0UihOh9Yrlzcm7H0NlIKji0GDDGgycwAkOtMUBaZwhf0WXlFfEkMVQtcVJNIR6wYH6OSCN9g4KxYnQesXy5uTdj6EyEFRxaO76m5saUSM4sD8HpHGG/BVdUl4RQxZDtT8v6CUwgwNwYBsHpNHeQaE4EVqvWN6cvPsxVAaCKs42UvEeogMH4ECrHJDGGfJXdEl5RQxZDBX92Wp/kjfcrZUD0mjvoFCcCK1XLG9O3v0YKgNBFadWQpMXYgsH4ICHA9I4Q/6KLimviCGLoaJHPD3CXvgDB45yQBrtHRSKE6H1iuXNybsfQ2UgqOLQTEebCUzABA60zwFpnCF/RZeUV8SQxVC1z1O0hhrCgbo4II32DgrFidB6xfLm5N2PoTIQVHFo5LoamXpQDzgQwwFpnCF/RZeUV8SQxVDFcIWeA0c4AAfEAWm0d1AoToTWK5Y3J+9+DJWBoIojAvGKmMABODAnDkjjDPkruqS8IoYshoqenVPPci/wuQYOSKO9g0JxIrResbw5efdjqAwEVZwayEsOiCgcgAPRHJDGGfJXdEl5RQxZDBV9E903xINTS+eANNo7KBQnQusVy5uTdz+GykBQxVl643D/DA84ME8OSOMM+Su6pLwihiyGap7cRZOoKxwoxwFptHdQKE6E1iuWNyfvfgyVgaCKQ9OWa1qwB3s4MB0HpHGG/BVdUl4RQxZDNR1/6E2whQPL5IA02jsoFCdC6xXLm5N3P4bKQFDFQTCWKRjUnbrPnQPSOEP+ii4pr4ghi6Gij+fex9wfHM/NAWm0d1AoToTWK5Y3J+9+DJWBoIqTm6ichzjCATiQgwPSOEP+ii4pr4ghi6Gil3L0EmfAsyVxQBrtHRSKE6H1iuXNybsfQ2UgqOIsqUm4V4YCHFgOB6RxhvwVXVJeEUMWQ7UcPqNd1BoO5OGANNo7KBQnQusVy5uTdz+GykBQxaFB8zQoOIMzHMjLAWmcIX9Fl5RXxJDFUOXlFD0M3nBg/hyQRnsHheJEaL1ieXPy7sdQGQiqOIjD/MWBGlPjJXJAGmfIX9El5RUxZDFU9PYSe5t7hvdTckAa7R0UihOh9Yrlzcm7H0NlIKjiTElKYiN6cAAOlOKANM6Qv6JLyitiyGKo6K9S/cW5cG+uHJBGeweF4kRovWJ5c/Lux1AZCKo4c20I7guxhwPL5oA0zpC/okvKK2LIYqiWzXE0jvrDgXgOSKO9g0JxIrResbw5efdjqAwEVRyaMb4ZwRRM4UB5DkjjDPkruqS8IoYshqo8z+h1agAH5sUBabR3UChOhNYrljcn734MlYGgioMQzEsIqCf1hAOXcEAaZ8hf0SXlFTFkMVT0O/0OB+BALAek0d5BoTgRWq9Y3py8+zFUBoIqDo0Y24jgCZ5woA4OSOMM+Su6pLwihiyGqg6u0fPUAQ7MhwPSaO+gUJwIrVcsb07e/RgqA0EVBxGYjwhQS2oJB45xQBpnyF/RJeUVMWQxVMfqDffBAg7AgQgOSKO9g0JxIrResbw5efdjqAwEVZwI8hEDEYMDcKA2DkjjDPkruqS8IoYshoq+q63vyAdOts4BabR3UChOhNYrljcn734MlYGgitM68ckf8YYDcMDigDTOkL+iS8orYshiqOC+xX3W4AUcGM8BabR3UChOhNYrljcn734MlYGgikPTjW86sAM7OFAvB6RxhvwVXVJe0a85bko5w/t6eU9tqA0c8HFAOufVVMXBUHmRrHy/Ck3j+RoP/MAPDtTJAWlcbVKsvKJfc9yncobzdXKeulAXOODngHTOq6mKg6HyIln5fhWa5vM3HxiCIRyojwPSuMqluKn0hCl8r4/v1ISawIEYDkjnvOKsOBgqL5KV71ehacCYBgRHcIQDdXFAGle5FDeVnjCF63VxnXpQDzgQxwHpnFecFQdD5UWy8v0qNE0Y14RgCZZwoB4OSOMql+Km0hOm8LwenlMLagEHYjkgnfOKs+JgqLxIVr5fhaYRYxsRPMETDtTBAWlc5VLcVHrCFI7XwXHqQB3gQDwHpHNecVYcDNVIJAWg53Xk0XttU340Y3wzgimYwoHyHJDG7SWMXLwVAWEKv8vzmxpQAzgwDQekc1vFcMCbioOhGgCWdYkA9LxacaPXlB8NOU1Dgiu4woGyHJDGRWvnkuMJU7hdltvgD/5wYDoOSOe8Wq84GKqRSHoA1N6RR++1TWfRlNM1JdiCLRwoxwFp3F7CyMVbERCm8Locr8Ee7OHAtByQzm0VwwFvKg6GagBY1iUeALXXihu9prNozGkbE3zBFw6U4YA0Llo7lxxPmMLpMpwGd3CHA9NzQDrn1XrFwVCNRNIDoPaOPHqvbTqL5py+OcEYjOFAfg5I4/YSRi7eioAwhc/5+QzmYA4H8nBAOrdVDAe8qTiRrwOOnfSSy0wafS24gBvjSLV3LeQkX+osGjRPg4IzOMOBvByQxk0ioAsNKkzhcl4ugzd4w4F8HJDOeWVecSJfvTl592OoDARVYJo0X5OCNVjDgXwckMYZ8sfSSASEKTzOx2OwBms4kJcD0rmRMjnrbRgqo7wiDI2at1HBG7zhQB4OSOMM+WNpJALCFA7n4TA4gzMcyM8B6dxImZz1NgyVUV4RhmbN36xgDuZwYHoOSOMM+WNpJALCFP5Oz18wBmM4UIYD0rmRMjnrbRgqo7wiDA1bpmHBHdzhwLQckMYZ8sfSSASEKdydlrvgC75woBwHpHMjZXLW2zBUl5b3vAu+0p3nk6pjAAAe2UlEQVR21kXdVy/+eifC0LTlmhbswR4OTMcBadysp1vmmxOm8HY63oIt2MKBshyQzmWW1yaOw1BdWqZ3fejLXfr0wbue9I7uuOPP7G51t2d2NG7ZxgV/8IcD03CAoRg/n4UpnJ2Gs+AKrnCgPAekc/EK2n5EDNWaodJHut/hhLO7U8/4aPeBT3wVY/WF8k2MkFIDOBDHAYZi/PAWpvA0jqdgCZZwoC4OSOfiFbT9iBiqDYZKxuoujzoXY4WhwlTDgVlxgKEYP7yFKQ+AdT0AUg/qAQfiOCCdi1fQ9iNiqHYYKoxVXCMiamAJB+rgAEMxfngLUzheB8epA3WAA/EckM7FK2j7EZsyVD9/v5cd/J6TTE6JV75jFd+giB6YwoG8HGAoxg9vYQqX83IZvMEbDuTjgHQuXkHbj4ihOvGcUSYNY5WvgRFLsIYDsRxgKMYPb2EKV2O5Cp7gCQfq4YB0Ll5B24+IoRppqPTdsbs/5p3d2e//4qx+vwLxqke8qAW1mIIDDMX44S1Mp6gXMdEBOAAHauCAdC5eQduPiKEaaagOvkP1kgv5FEA+rAAzDQea4wBDMX54C9MaHnrIgYdvOAAHpuCAdC5eQduPiKHa01Dxo36I1BQiRUx4lZMDDMX44S1Mc9aRs9ANOAAHcnJAOhevoO1HbMpQpUJO9T/9YV/9KN/6K0YK0copWpwF36bkAEMxfpII0ynrRmx0AQ7AgZIckM7FK2j7ETFUl9Zwk6HCSCFeJcWLs+HfFBxgKMYPb2E6Rb2IiQ7AAThQAwekc/EK2n5EDNUGQ3WHE87mD/ryuzHN/W5MDYJLDvUPfoZi/PAWpvC/fv5TI2oEB8ZxQDoXr6DtR8RQrRmqu570ju6448/sbnW3Z/IwjaGCA3BglhxgKMYPb2HKg9q4BzVwAzc4UD8HpHPxCtp+RAzVpTU874KvdKeddVH31Yu/3okwNHf9zU2NqBEc2J8D0rj2R1g9dyBM4eP+fAQzMIMDbXBAOleP8taTCYbKqIUIQ4O30eDUiTrBgf04II0z5I+lkQgIU7i4HxfBC7zgQDsckM6NlMlZb8NQGeUVYWjydpqcWlErODCcA9I4Q/5YGomAMIWHw3kIVmAFB9rigHRupEzOehuGyiivCEOjt9Xo1It6wYFhHJDGGfLH0kgEhCkcHMZBcAInONAeB6RzI2Vy1tswVEZ5RRiavb1mp2bUDA7s5oA0zpA/lkYiIEzh327+gREYwYE2OSCdGymTs96GoTLKK8LQ8G02PHWjbnBgOwekcYb8sTQSAWEK97ZzD3zABw60ywHp3EiZnPU2DJVRXhGGpm+36akdtYMDmzkgjTPkj6WRCAhTeLeZd2ADNnCgbQ5I50bK5Ky3YaiM8oowNH7bjU/9qB8csDkgjTPkj6WRCAhTOGdzDlzABQ60zwHp3EiZnPU2DJVRXhGG5m+/+akhNYQDRzkgjTPkj6WRCAhT+HaUb2ACJnBgHhyQzo2UyVlvw1AZ5RVhEIB5CAB1pI5w4DAHpHGG/LE0EgFhCtcOcw08wAMOzIcD0rmRMjnrbRgqo7wiDCIwHxGgltQSDhzjgDTOkD+WRiIgTOHZMZ6BBVjAgXlxQDo3UiZnvQ1DZZRXhEEI5iUE1JN6woFLOCCNM+SvuiXl6nnNcVPKD46hM3AADsyRA+d95N876VwOTW3tDAyVUTERZo4NwT0h9HAADkjjDPmrbkm5el5z3JTyo7/oLzgAB+bAgWSgXvTq93WPPOW1KyMlncuhqa2dgaEyKibCzKEhuAeEHQ7AgXUOSOMM+atuSbne8cRzun3/r705bkpnrWPN1/QfHIADLXBgm4FK+nanB57WnfzsN3cveMV7ckhqc2dgqIySMRgRvxbEjxzh6VgOSOMM+atuSbnua6bS9dqb46Z01tiasI9+hgNwICcH9jFQX/yPi3PIaNNnYKiM8jEYEbWcosZZ8C03B6RxhvxVt6RcMVT0Se4+4Tw4NycOYKCmHW8YKgNfDfA5NRL3wmCAA3BAHJDGGfJX3ZJyxVDBX/GXV7gAB3ZzAAOVd5xhqAy8NcBvf/zzD34ZL/1S3uvP/WRHA+9uYDACIzhQPwekcYb8VbekXDFU9fOK3qdGcKAcBzBQZccXhsrAP/3inYZ4/xWDVU4oEGmwhwNxHJCuGfJX3ZJyxVDF1Z9eAks40D4HMFB1jSsM1ZZ6pF/Ce+M5F3YPevwrDj7dRINdrxis9gWJoUINl8gBadgW+avmLeWKoaJXl9ir3DO8FwcwUNWMJTMRDJUJi72IwULYJGy8woWWOSCTYitdXavKFUNFz7Xcc+QOf/flAAaqrlm0KxsM1S6EtryPwUIg9xVIroczNXBAJmWLvFXzlnLFUNE7NfQOOcDDqTiAgapm7IxKBEM1CjZ7EwYLoZ1KaIkLtyI5IJNiK1ldq8oVQ0UPRPYAseBTaQ5goOqaNd5sMFReBLfsx2Ah2KUFm/PhoMUBmZQt8lXNW8oVQwWXLS6zBi9a4QAGqpqxMkkiGKpJYLWDYrAQ/laEnzznzVWZFFup6lpVrhiqeXMSzaG+c+MABqquWTJ1NhiqqRHeEh+DxQCZ2wDhftrgtEzKFnmq5i3liqFqg1toAHVaKgcwUNWMjSKJYKiKwG4fisFiEC11EHHfebkvk2IrUV2ryhVDlZcj9CR4w4HtHMBA1TUrSmeDoSpdgS3nY7C2ixliDz5wYBwHZFK2yE81bylXDNW4WtMj4AYHYjiAgapmLFSZCIaqyrLYSWGwYkSR4QKOS+eATIqtNHWtKlcMFX279L7l/vP2AAaqrllQezYYqtortCU/DFZecWWYgfdcOCCTskVeqnlLuWKo6L+59B/3USeXMVDVyH6TiWComiybnTQGq06RZnhSl9o4IJNiK0ldq8oVQ0Uf1dZH5NM2JzFQdWl969lgqFqv4Jb8MVhtiz3DmvpNxQGZlC3yUc1byhVDRT9M1Q/EXQa3MFDVyPosE8FQzbKs9k1hsJYxNHg4oM67OCCTYitFXavKFUMFr3fxmvfhSJ8DGKi6tHzu2WCo5l7hLfeHwWL49IcP/14OH2RStshDNW8pVwzVcviJFlHrMRzAQFUj24tMBEO1yLLbN43BYoiNGWLsaY83Mim2EtS1qlwxVO3xDG2gZlNyAANVl1YvPRsM1dIZsOX+MVgMwymHIbHL8UsmZUv7V/OWcsVQleMLvQr2NXAAA1WNLJOIgQCGygCFJRsBDBZDtYahSg5+Hsqk2J1e16pyxVD5607vgGFLHMBA1aXFZLMdAQzVdnx4dwsCGCyGc0vDmVyP8VUmZUt7V/OWcsVQHasfXAaLOXIAA1WN7JLICAQwVCNAY4uNAAaLIT/HIT/He5JJsTu5rlXliqFCX+bYi0u+JwxUXVpLNj4EMFQ+/Ni9BQEMFg9AS35YqPneZVK2tG81bylXDBV6UnNPkdtufmKgqpFVEpkAAQzVBKAS0kYAg7V74DCUwSgHB2RS7E6ta1W5YqjojRy9wRlxPMNA1aWlZDMtAhiqafEl+hYEMFhxg4uHALDchwMyKVvas5q3lCuGCo7vw3Guzc8XDFQ1skkiBRDAUBUAnSNtBDBY+QcgDx3LxFwmxe7EulaVK4ZqmVxFo+qtOwaqLq0km7IIYKjK4s/pWxDAYNU7SHnIabs2Milb2q+at5QrhqptzqEZ7dcPA1WNLJJIhQhgqCosCinZCGCw2h/IPFTVUUOZFLvT6lpVrhiqOrhDDy+nDhiourSQbOpGAENVd33IbgsCGKzlDHYe4mJrLZOypb2qeUu51m6obn2vF3THHX9m94FPfLWDr7F8Bc88eGKgqpE9EmkQAQxVg0UjZRsBDFaeocvDTfs4y6TYnVTXqnKt3VDd5t4v7FKOd3nUud2pZ3wUY/WF9vtk7lqHgapL68imbQQwVG3Xj+y3IIDB4oFm7g9EY+9PJmVL+1TzlnJtxVApT4wV+jO2P6fah4GqRtZIZIYIYKhmWFRuyUYAg8UDzlQPKq3FlUmxO6WuVeUqo7LPq/bmuCN9h2o9P4wVulNKHzBQOTqfM0DgEgQwVDBhsQhgsHjQKfWgU/rcnEbDKzDKdd2oDPlae705DNm/yVApT4wVejN132OghnQq14DANAhgqKbBlagNIoDB4oFn6geeWuLnNBpeKVCuMib7vGqvcvi7t/7Lwe857RMj+lqMFToTpQMYKHU2ryBQHgEMVfkakEGlCGCwePCJevCpLc660ai0BQ/SUq5jjI326v5qMFS6D4wV+rKvLmCg1Mm8gkB9CGCo6qsJGVWKAAaLB6B9H4BqvX7daFTacgdpKVcZkX1etVf3V5Oh0n3c/THv7M5+/xf5qHU+FfAIBzBQ6lxeQaB+BDBU9deIDCtFAIOFwarVMO3Ka91oVNpiB2kpVxmQfV61V/dXk6E6+A7VSy7k49UxUisjhYFSp/IKAu0hgKFqr2ZkXCkCGCwM1i4jU8v760aj0pY6SEu57mOkdK326v5qMFT8qB86IR3AQKkzeQWB9hHAULVfQ+6gUgQwWDw46cGpttd1o1FpCx2kpVxlkvZ51d4c98en/NHvu/ocA5WjEzkDBMoggKEqgzunLhABDBYPXLseuHK9n9NoeFtdue5jpHSt9npzGLJ/k6HiO1LL7XsM1JDO4RoQmAcCGKp51JG7aBABDNZyH7RyGadN5+Q0Gt7WVK4ySfu8aq83hyH71w0VRmp5/Y2BGtIpXAMC80QAQzXPunJXDSKAwVreA9gmwzP1ek6j4W1F5bqPkdK12uvNYch+GSqM1HL6GAM1pDO4BgSWgQCGahl15i4bRACDtZwHs6kN1Hr8nEbD23rKVSZpn1ft9eYwZP+t7/WC7rjjz+RT+2b8qX0YqCGdwDUgsEwEMFTLrDt33SACGCwM1roxGvt1TqPhbTXluo+R0rXa681hyH6dNbYm7KuvvzFQQ5jPNSAAAgkBDBU8AIFGEcBg1fcA1spDsR7+W6C+cpVJ2udVe3Pcp85qhQPkeVQ/MFA5OoUzQGCeCGCo5llX7mqBCGCwjj4g8dBoY6KH/xbaRLnuY6R0rfbmuE+dBedsztWICwYqR2dwBggsAwEM1TLqzF0uEAEMVjsPdrkfNvXw30JbKFeZpH1etTfHfeqs3LXkvOF9joHK0QmcAQLLRABDtcy6c9cLRACDNfzBa+4PqXr4b6ENlOs+RkrXam+O+9RZc+dOS/eHgcrBfM4AARBICGCo4AEILBQBDNZyDZYe/lugvnKVSdrnVXtz3KfOaslwzC1XDFQOpnMGCICAhQCGykKFNRBYIAIYrOUYLD38t0Bz5bqPkdK12pvjPnXW3ExKzfeDgcrBbM4AARAYgkARQ6XBM+Z1yE1xDQiAgB8BDNZ8DZa018+S6SMoV5mkfV61d/osu05n1WxAWs8NA5WDyZwBAiAwBgEM1RjU2AMCC0QAgzUfg6WH/xZorFz3MVK6Vntz3KfOat201JQ/BioHczkDBEAgAoGshioiYWKAAAjUgQAGq12DpYf/Opi0PQvlKpO0z6v2bj8h5l2dVZMhaS0XDFQMF4kCAiCQHwEMVX7MOREEZokABqsdg6WH/xaIqFz3MVK6Vntz3KfOas3ElMwXA5WDmZwBAiCQAwEMVQ6UOQMEFogABqteg6WH/xZoqVxlkvZ51d4c96mzShqU2s/GQOVgImeAAAiUQABDVQJ1zgSBBSKAwarHYOnhvwUaKtd9jJSu1d4c96mzajc1OfPDQOVgHmeAAAjUgACGqoYqkAMILBABDFY5g6WH/xZop1xlkvZ51d4c96mzchqW2s7CQOVgGmeAAAjUiACGqsaqkBMILBABDFY+g6WH/xZoplz3MVK6Vntz3KfOqs3kTJkPBioHszgDBECgBQQwVC1UiRxBYIEIYLCmM1h6+G+BVspVJmmfV+3NcZ86a0oDUzo2BioHkzgDBECgRQQwVC1WjZxBYIEIYLDiDJYe/lugkXLdx0jpWu3NcZ86q7TpiTwfA5WDOZwBAiAwBwQwVHOoIvcAAgtEAIM13mDp4b8F2ihXmaR9XrU3x33qrEhDkzsWBioHUzgDBEBgjghgqOZYVe4JBBaIAAZruMHSw38LNFGu+xgpXau9Oe5TZ+U2QZ7zMFA5mMEZIAACS0AAQ7WEKnOPILBABDBYmw2WHv5boIVylUna51V7c9ynzvIYnKn3YqByMIEzQAAElohAE4ZqfYAusVDcMwiAgA8BDNYxg6WHfx+ieXYr1/U5MORr7c2Rqc6a2hTtEx8DlaPynAECIAACXYehggUgAAKLRGDJBksP/y0UXrkOMVDr12hvjvvUWfsYnuhrMVA5Ks0ZIAACIHAUAQzVUUxYAQEQWCACSzJYevhvoczKdd0sDflae3Pcp86KNknb4mGgclSWM0AABEBgNwIYqt0YcQUIgMACEZizwdLDfwtlVa5DDNT6Ndqb4z511jYD5H0PA5WjkpwBAiAAAvsjgKHaHzN2gAAILBCBORksPfy3UEblum6WhnytvTnuU2d5TVN/PwYqR+U4AwRAAAT8CGCo/BgSAQRAYIEItGyw9PDfQtmU6xADtX6N9ua4T53VN0T7/hsDlaNSnAECIAAC8QhgqOIxJSIIgMACEWjJYOnhv4UyKdd1szTka+3NcZ86ax8ThYHKURnOAAEQAIHpEcBQTY8xJ4AACCwQgZoNlh7+WyiLch1ioNav0d4c96mzthkqDFSOSnAGCIAACORHAEOVH3NOBAEQWCACNRksPfy3UAblum6WhnytvTnuU2f1DRUGKgfynAECIAAC5RHAUJWvARmAAAgsEIGSBksP/y3ArlyHGKj1a7Q3x33qrBe9+n3dI095baev+693euBp3cnPfnP3gle8p0v1538gAAIgAALzQABDNY86chcgAAKNI5DTYOkhvwXIlOu6WRrytfbmuE+d1X/FQOVAnjNAAARAoDwCGKryNSADEAABEDiCwJQGSw/9Rw6tcEG5DjFQ69dob47b0ll8ByoH2pwBAiAAAnUhgKGqqx5kAwIgAAImApEGSw//5kGVLSrXdbM05GvtzXFL/AhfDpQ5AwRAAATqRABDVWddyAoEQAAEtiLgMVg5jcbWmxjwpnIdYqDWr9HeAcdwCQiAAAiAAAiMRgBDNRo6NoIACIBAPQjsY7BaMhrKdd0sDflae+upEpmAAAiAAAjMEQEM1Ryryj2BAAgsHoFdBiuZjRb+J1M0xECtX6O9LdwnOYIACIAACLSLAIaq3dqROQiAAAgMRsAyWIM3F7xQpmjdLA35WnsLps/RIAACIAACC0AAQ7WAInOLIAACILCOQDJYLfxPpmiIgVq/RntbuE9yBAEQAAEQaBcBDFW7tSNzEAABEJg9AjJF62ZpyNfaO3uQuEEQAAEQAIGiCGCoisLP4SAAAiAAAtsQkCkaYqDWr9HebfF5DwRAAARAAAS8CGCovAiyHwRAAARAYDIEZIrWzdKQr7V3suQIDAIgAAIgAAJd12GooAEIgAAIgEC1CMgUDTFQ69dob7U3R2IgAAIgAAKzQABDNYsychMgAAIgME8EZIo8r/NEhrsCARAAARCoBQEMVS2VIA8QAAEQAIEjCHiMlPYeCcoCCIAACIAACAQigKEKBJNQIAACIAACeRB4w7u/0PV/xO8ujzq3+8+Lv57ncE4BARAAARAAgR4CGKoeGPwTBEAABECgfgT+v//7/7r7/vF7DhmqZK7+6jUX1Z88GYIACIAACMwOAQzV7ErKDYEACIDAvBFY/+6UvlPFd6nmXXfuDgRAAARqRQBDVWtlyAsEQAAEQOAIApu+OyVTxXepjkDGAgiAAAiAwMQIYKgmBpjwIAACIAACcQhs+u6UDBXfpYrDmkggAAIgAALDEMBQDcOJq0AABEAABAojsOu7UzJVfJeqcKE4HgRAAAQWhgCGamEF53ZBAARAoFUEdn13SoaK71K1WmHyBgEQAIE2EcBQtVk3sgYBEACBRSHwf/7v/+vu96T3dnd+5LkH/7/TI8458il/ei+9nnYWn/i3KIJwsyAAAiBQEAEMVUHwORoEQAAEQGAcAv/2pf99yFA99KnnjwvELhAAARAAARBwIoChcgLIdhAAARAAgfwIYKjyY86JIAACIAACNgIYKhsXVkEABEAABCpGAENVcXFIDQRAAAQWhgCGamEF53ZBAARAYA4IYKjmUEXuAQRAAATmgQCGah515C5AAARAYFEIYKgWVW5uFgRAAASqRgBDVXV5SA4EQAAEQMBCAENlocIaCIAACIBACQQwVCVQ50wQAAEQAAEXAhgqF3xsBgEQAAEQCEQAQxUIJqFAAARAAATyIIChyoMzp4AACIAACOxGAEO1GyOuAAEQAAEQqAwBDFVlBSEdEAABEFgwAhiqBRefWwcBEACBVhHAULVaOfIGARAAgfkhgKGaX025IxAAARCYPQIYqtmXmBsEARAAgWYQwFA1UyoSBQEQAAEQEAIYKiHBKwiAAAiAQGkEMFSlK8D5IAACIAACeyOAodobMjaAAAiAAAhMhACGaiJgCQsCIAACIDAdAhiq6bAlMgiAAAiAwH4INGmovva//0/H/8EADsABOLBcDnzsXy7u7njiOav/P/Sp5+83/bgaBEAABEAABIIQaNJQ9Yco/z72QAEWYAEH4MBSOYChCnoqIAwIgAAIgMDeCGCoev+Fc6kPItw3D+FwAA60zgEM1d7znw0gAAIgAAJBCGCoMFSrH5lp/YGK/DEFcGC5HMBQBT0VEAYEQAAEQGBvBJowVPd6/P/o+D8YwAE4AAfgwCYOPPo5H9p7ALIBBEAABEAABCIQaMJQRdwoMUAABEAABEAABEAABEAABEAgGgEMVTSixAMBEAABEAABEAABEAABEFgMAhiqxZSaGwUBEAABEAABEAABEAABEIhGAEMVjSjxQAAEQAAEQAAEQAAEQAAEFoMAhmoxpeZGQQAEQAAEQAAEQAAEQAAEohHAUEUjSjwQAAEQAAEQAAEQAAEQAIHFIIChWkypuVEQAAEQAAEQAAEQAAEQAIFoBDBU0YgSDwRAAARAAARAAARAAARAYDEIYKgWU2puFARAAARAAARAAARAAARAIBoBDFU0osQDARAAARAAARAAARAAARBYDAIYqsWUmhsFARAAARAAARAAARAAARCIRgBDFY0o8UAABEAABEAABEAABEAABBaDAIZqMaXmRkEABEAABEAABEAABEAABKIRwFBFI0o8EAABEAABEAABEAABEACBxSCAoVpMqblREAABEAABEAABEAABEACBaAQwVNGIEg8EQAAEQAAEQAAEQAAEQGAxCGCoFlNqbhQEQAAEQAAEQAAEQAAEQCAaAQxVNKLEAwEQAAEQAAEQAAEQAAEQWAwCGKrFlJobBQEQAAEQAAEQAAEQAAEQiEYAQxWNKPFAAARAAARAAARAAARAAAQWgwCGajGl5kZBAARAAARAAARAAARAAASiEcBQRSNKPBAAARAAARAAARAAARAAgcUggKFaTKm5URAAARAAARAAARAAARAAgWgEMFTRiBIPBEAABEAABEAABEAABEBgMQhgqBZTam4UBEAABEAABEAABEAABEAgGgEMVTSixAMBEAABEAABEAABEAABEFgMAhiqxZSaGwUBEAABEAABEAABEAABEIhGAEMVjSjxQAAEQAAEQAAEQAAEQAAEFoMAhmoxpeZGQQAEQAAEQAAEQAAEQAAEohHAUEUjSjwQAAEQAAEQAAEQAAEQAIHFIIChWkypuVEQAAEQAAEQAAEQAAEQAIFoBDBU0YgSDwRAAARAAARAAARAAARAYDEIYKgWU2puFARAAARAAARAAARAAARAIBoBDFU0osQDARAAARAAARAAARAAARBYDAIYqsWUmhsFARAAARAAARAAARAAARCIRgBDFY0o8UAABEAABEAABEAABEAABBaDAIZqMaXmRkEABEAABEAABEAABEAABKIRwFBFI0o8EAABEAABEAABEAABEACBxSCAoVpMqblREAABEAABEAABEAABEACBaAQwVNGIEg8EQAAEQAAEQAAEQAAEQGAxCGCoFlNqbhQEQAAEQAAEQAAEQAAEQCAaAQxVNKLEAwEQAAEQAAEQAAEQAAEQWAwCGKrFlJobBQEQAAEQAAEQAAEQAAEQiEYAQxWNKPFAAARAAARAAARAAARAAAQWgwCGajGl5kZBAARA4P9vv45pAAAAEIb5d42JfdQACeWCAAECBAgQIFALOFS1qDwCBAgQIECAAAECBG4EHKqbqRUlQIAAAQIECBAgQKAWcKhqUXkECBAgQIAAAQIECNwIOFQ3UytKgAABAgQIECBAgEAt4FDVovIIECBAgAABAgQIELgRcKhuplaUAAECBAgQIECAAIFawKGqReURIECAAAECBAgQIHAj4FDdTK0oAQIECBAgQIAAAQK1gENVi8ojQIAAAQIECBAgQOBGwKG6mVpRAgQIECBAgAABAgRqAYeqFpVHgAABAgQIECBAgMCNgEN1M7WiBAgQIECAAAECBAjUAgPASxNqgMhERQAAAABJRU5ErkJggg==)", "_____no_output_____" ], [ "# Funciones de Pérdida", "_____no_output_____" ] ], [ [ "# Funciones clásicas\ncross_entropy = tf.keras.losses.BinaryCrossentropy(from_logits=True)\nmae = tf.keras.losses.MeanAbsoluteError(name='mean_absolute_error')", "_____no_output_____" ], [ "Lambda = [0.5, 1.0]\n\ndef generator_loss_classic(fake_output) :\n return cross_entropy(tf.ones_like(fake_output), fake_output)\n\ndef generator_loss_v2(real_output, fake_output) :\n return - tf.keras.backend.log(tf.reduce_mean(fake_output))\n\ndef generator_loss(real_output, fake_output) :\n return tf.keras.backend.log(1.0 - tf.reduce_mean(fake_output))\n\ndef discriminator_loss_classic(real_output, fake_output) :\n # Pérdida de los verdaderos (1 vs y_hat)\n real_loss = cross_entropy(tf.ones_like(real_output), real_output)\n # Pérdida de los sintéticos (0 vs y_hat)\n fake_loss = cross_entropy(tf.zeros_like(real_output), real_output)\n # Suma de los dos tipos de errores\n total_loss = real_loss + fake_loss\n return total_loss\n\ndef discriminator_loss_v2 (real_output, fake_output) :\n # Pérdida de los verdaderos (1 vs y_hat)\n return tf.keras.backend.log(tf.reduce_mean(fake_output)) - tf.keras.backend.log(tf.reduce_mean(real_output))\n\ndef discriminator_loss(real_output, fake_output) :\n # Pérdida de los valores ( 1 vs y_hat)\n return - tf.keras.backend.log(tf.reduce_mean(real_output) * (1.0 - tf.reduce_mean(fake_output)))", "_____no_output_____" ] ], [ [ "## Implementación de la nueva función de pérdida", "_____no_output_____" ] ], [ [ "def wasserstein_generator_loss (real_output, fake_output) :\n lossEM = tf.reduce_mean(tf.math.abs(tf.reduce_mean(real_output[0], 0) - tf.reduce_mean(fake_output[0], 0)))\n lossC = tf.keras.backend.log(1.0 - tf.reduce_mean(fake_output[1]))\n return Lambda[0] * lossEM + Lambda[1] * lossC\n\ndef wasserstein_critic_loss(real_output, fake_output) :\n lossEM = tf.reduce_mean(tf.math.abs(tf.reduce_mean(real_output[0], 0) - tf.reduce_mean(fake_output[0], 0)))\n lossC = - tf.keras.backend.log( tf.reduce_mean(real_output[1]) * (1.0 - tf.reduce_mean(fake_output[1])))\n return Lambda[0] * lossEM + Lambda[1] * lossC", "_____no_output_____" ] ], [ [ "## Creamos modelos", "_____no_output_____" ] ], [ [ "Losses = {'dcgan' : {'generator' : generator_loss,\n 'discriminator' : discriminator_loss},\n 'wgan' : {'generator' : wasserstein_generator_loss,\n 'discriminator' : wasserstein_critic_loss}\n }\nModels = {'dcgan' : {'generator' : generator,\n 'discriminator' : discriminator},\n 'wgan' : {'generator' : generator,\n 'discriminator' : wasserstein_critic}\n }\n\n# Para el paso de entrenamiento\nmodelname = 'wgan'\nGAN_discriminator = Models[modelname]['discriminator']\nGAN_generator = Models[modelname]['generator']\n# Pérdidas\nLoss_discriminator = Losses[modelname]['discriminator']\nLoss_generator = Losses[modelname]['generator']", "_____no_output_____" ] ], [ [ "# Paso de entrenamiento", "_____no_output_____" ], [ "Definiremos un paso de entrenamiento de la GAN, que corresponde a hacer un paso de entrenamiento en cada modelo.\n\n1. Carga lote de imágenes reales $ X $ y etiquetas $y$.\n\n2. Genera un lote de variables latentes $ Z $\n\n3. Genera un lote de variables sintéticas: $ \\hat{X} = G(Z; \\theta_g) $\n\n4. Calcula las predicciones para $ \\hat {Y} = D([X | \\hat{X}], \\theta_d) $\n\n5. Calcula costos $ J^D $ y $ J^G$.\n\n6. Calcula gradientes $\\nabla_{\\theta_d} J^D$ y $ \\nabla_{\\theta_g} J^G$ .\n\n7. Actualiza pesos con un paso del algoritmo de optimización:\n\n$$ \\theta_d = \\theta_d + \\delta(\\nabla_{\\theta_d} J^D) $$ \n$$ \\theta_g = \\theta_g + \\delta(\\nabla_{\\theta_g} J^G)$$ \n\ndonde $\\delta(\\cdot)$ representa el paso del algoritmo utilizado.", "_____no_output_____" ] ], [ [ "#El decorador `tf.function` indica que la función se \"compila\" para que pueda incluirse en un gráfo de calculo.\[email protected]\ndef train_step(images = None, labels = None, modelname = 'dcgan'):\n '''\n Implementa un paso de entrenamiento para la GAN\n \n Recibe como parámetros un semi-lote de imágenes reales\n '''\n\n # variables latentes (ruido Gaussiano), tantas como imagenes reales\n z = tf.random.normal([images.shape[0], z_dim])\n sampled_labels = np.random.randint(0, 10, images.shape[0])\n sampled_labels = sampled_labels.reshape((-1, 1))\n\n # Los siguientes pasos registaran (en \"TAPE\") para efectos de calcular gradientes\n # son dos \"Tapes\" para registrar los calculos de cada modelo\n with tf.GradientTape() as gen_tape, tf.GradientTape() as disc_tape:\n # genera el semi-lote de datos sintéticos\n # generated_images = GAN_generator([z,sampled_labels] , training=True)\n generated_images = GAN_generator(z, training=True)\n \n # calcula la predicción para datos verdaderos y falsos\n # real_output = GAN_discriminator([images,labels] , training = True)\n real_output = GAN_discriminator(images, training = True)\n # fake_output = GAN_discriminator([generated_images,sampled_labels], training = True)\n fake_output = GAN_discriminator(generated_images, training = True)\n\n # calcula las pérdidas del Discriminador y del Generador\n disc_loss = Loss_discriminator(real_output, fake_output)\n gen_loss = Loss_generator(real_output, fake_output)\n\n # Para cada modelo, calcula el gradiente de su función de costo respecto a sus pesos entrenables, \n # haciendo retropropagación sobre los calculos realizados\n gradients_of_generator = gen_tape.gradient(gen_loss, GAN_generator.trainable_variables)\n gradients_of_discriminator = disc_tape.gradient(disc_loss, GAN_discriminator.trainable_variables)\n\n # Hace el paso de actualizacion de los pesos\n generator_optimizer.apply_gradients (zip(gradients_of_generator , GAN_generator.trainable_variables))\n discriminator_optimizer.apply_gradients(zip(gradients_of_discriminator, GAN_discriminator.trainable_variables))\n \n return disc_loss, gen_loss", "_____no_output_____" ] ], [ [ "# Iteración del entrenamiento", "_____no_output_____" ] ], [ [ "from PIL import Image\nimport imageio\nimport glob\n\n# - - - - - - - - - - - - - - - - - - - - - - - - - - \n# Crea una collage de imágenes con el Generador y la guarda\n# - - - - - - - - - - - - - - - - - - - - - - - - - - \n\nnum_img_rows, num_img_cols = (4,4)\nnum_examples_to_generate = num_img_rows * num_img_cols\n\nseed = tf.random.normal([num_examples_to_generate, z_dim])\nlabels_seed = np.random.randint(0, 10, num_examples_to_generate)\nlabels_seed = labels_seed.reshape((-1, 1))\n\ndef generate_and_save_images(model, epoch, test_input, labels, pathdir=path_results):\n # Se pone entrenable en Falso porque esta en modo inferencia el \n # predictions = model([test_input, labels], training=False)\n predictions = model(test_input, training=False)\n # predictions -> [0,255]\n predictions = predictions - tf.reduce_min(predictions)\n predictions = 255 * predictions / tf.reduce_max(predictions)\n data = tf.cast(predictions, tf.uint8)\n\n fig = plt.figure(figsize=(4,4)) \n\n for i in range(data.shape[0]):\n plt.subplot(4, 4, i+1)\n plt.imshow(data[i, :, :, :])\n plt.axis('off')\n\n plt.savefig(pathdir+'image_epch_{:04d}.png'.format(epoch))\n plt.show()\n \n# - - - - - - - - - - - - - - - - - - - - - - - - - - \n# Despliega la imagen correspondiente a una época\n# - - - - - - - - - - - - - - - - - - - - - - - - - - \n\ndef display_image(epoch, pathdir=path_results, modelname = 'dcgan') :\n filename = '{}{}_image_epoch_{:04d}.png',format(pathdir, modelname, epoch)\n\n return Image.open(filename)\n\n# ----------------------------------------------------\n# Crea un gif animado con las imágenes de las épocas\n# ----------------------------------------------------\n\ndef createGif(anim_file = 'gan.gif', pathdir = path_results, modelname = 'dcgan') :\n gifname = '{}{}_{}'.format(pathdir, modelname, anim_file)\n with imageio.get_writer(gifname, mode = 'I') as writer :\n finelames = glob.glob(pathdir + 'image*.png')\n filenames = sorted(finelames)\n for i, filename in enumerate(filenames) :\n image = imageio.imread(filename)\n if i % 5 == 0 :\n writer.append_data(image)\n\nfrom tensorflow.keras.preprocessing.image import load_img, img_to_array\n\ndef saveGif(anim_file = 'gan.gif', pathdir = path_results, modelname = 'dcgan') :\n gifname = '{}{}_{}'.format(pathdir, modelname, anim_file)\n filenames = glob.glob(pathdir + 'image*.png')\n filenames = sorted(filenames)\n\n print(gifname, filenames, len(filenames))\n \n images = []\n for i, filename in enumerate(filenames) :\n images.append(load_img(filename))\n images[1].save(gifname,\n save_all = True,\n append_images = images[2:],\n optimize = False,\n duration = 40,\n loop = 0)", "_____no_output_____" ] ], [ [ "# Optimizadores para los modelos", "_____no_output_____" ] ], [ [ "generator_optimizer = tf.keras.optimizers.Adam(learning_rate)\ndiscriminator_optimizer = tf.keras.optimizers.Adam(learning_rate)\nlearning_rate", "_____no_output_____" ], [ "checkpoint_prefix = os.path.join(path_checkpoints, 'ckpt')", "_____no_output_____" ], [ "checkpoint = tf.train.Checkpoint(generator_optimizer = generator_optimizer,\n discriminator_optimizer = discriminator_optimizer,\n GAN_generator = GAN_generator,\n GAN_discriminator = GAN_discriminator)", "_____no_output_____" ], [ "from IPython import display\nimport time \nfrom tqdm import tqdm\n\ndef train(dataset, epochs, modelname = 'dcgan'):\n generator_losses = [] \n discriminator_losses = []\n for epoch in range(epochs):\n start = time.time()\n t = 0\n for image_batch, true_label in dataset:\n gen_loss, disc_loss = train_step(images = image_batch, labels = true_label, modelname = modelname)\n\n # solo registramos los costos en el último lote\n generator_losses.append(gen_loss)\n discriminator_losses.append(disc_loss)\n \n # Produce imágenes para crear el GIF\n if (epoch + 1) % each_save == 0:\n display.clear_output(wait = True) # limpia el buffer\n generate_and_save_images(model = generator,\n epoch = epoch + 1,\n test_input = seed,\n labels = labels_seed,\n pathdir = path_results)\n if (epochs + 1) % save_epochs == 0 :\n checkpoint.save(file_prefix = checkpoint_prefix)\n print ('{} : Time for epoch {} is {} sec'.format(modelname, epoch + 1, time.time()-start))\n # Genera conjunto de imágenes para mostrar avances\n display.clear_output(wait = True)\n generate_and_save_images(model = generator,\n epoch = epoch + 1,\n test_input = seed,\n labels = labels_seed,\n pathdir = path_results)\n return np.array(generator_losses), np.array(discriminator_losses)", "_____no_output_____" ] ], [ [ "# Entrenamiento de la WGAN", "_____no_output_____" ], [ "## Creando el modelo", "_____no_output_____" ] ], [ [ "# creo generador\ngenerator = build_generator(img_shape, z_dim)\n# creo discriminador\ndiscriminator = build_discriminator(img_shape)\n# construye el discriminador \nwasserstein_critic = build_Wasserstein_critic(img_shape)\n\nLosses = {'dcgan' : {'generator' : generator_loss,\n 'discriminator' : discriminator_loss},\n 'wgan' : {'generator' : wasserstein_generator_loss,\n 'discriminator' : wasserstein_critic_loss}\n }\nModels = {'dcgan' : {'generator' : generator,\n 'discriminator' : discriminator},\n 'wgan' : {'generator' : generator,\n 'discriminator' : wasserstein_critic}\n }\n\n# Para el paso de entrenamiento\nmodelname = 'wgan'\nGAN_discriminator = Models[modelname]['discriminator']\nGAN_generator = Models[modelname]['generator']\n# Pérdidas\nLoss_discriminator = Losses[modelname]['discriminator']\nLoss_generator = Losses[modelname]['generator']\n\ncheckpoint = tf.train.Checkpoint(generator_optimizer = generator_optimizer,\n discriminator_optimizer = discriminator_optimizer,\n GAN_generator = GAN_generator,\n GAN_discriminator = GAN_discriminator)", "_____no_output_____" ], [ "generator_losses, discriminator_losses = train(train_dataset, num_epochs, modelname = 'wgan')", "_____no_output_____" ] ], [ [ " Guardo Gifs", "_____no_output_____" ] ], [ [ "saveGif('cifar10.gif', pathdir = path_results, modelname = 'wgan')", "dcgan_results/wgan_cifar10.gif ['dcgan_results/image_epch_0005.png', 'dcgan_results/image_epch_0010.png', 'dcgan_results/image_epch_0015.png', 'dcgan_results/image_epch_0020.png', 'dcgan_results/image_epch_0025.png', 'dcgan_results/image_epch_0030.png', 'dcgan_results/image_epch_0035.png', 'dcgan_results/image_epch_0040.png', 'dcgan_results/image_epch_0045.png', 'dcgan_results/image_epch_0050.png', 'dcgan_results/image_epch_0055.png', 'dcgan_results/image_epch_0060.png', 'dcgan_results/image_epch_0065.png', 'dcgan_results/image_epch_0070.png', 'dcgan_results/image_epch_0075.png', 'dcgan_results/image_epch_0080.png', 'dcgan_results/image_epch_0085.png', 'dcgan_results/image_epch_0090.png', 'dcgan_results/image_epch_0095.png', 'dcgan_results/image_epch_0100.png'] 20\n" ] ], [ [ "# Salvar modelos y graficar desempeño", "_____no_output_____" ] ], [ [ "generator.save_weights('dcgan_generator_2_0.h5')\ndiscriminator.save_weights('dcgan_discriminator_2_0.h5')", "_____no_output_____" ], [ "plt.figure(figsize=(12,4))\nplt.subplot(131)\nplt.plot(generator_losses, 'r')\nplt.title('$J^G$')\nplt.subplot(132)\nplt.plot(discriminator_losses, 'g')\nplt.title('$J^D$')\nplt.subplot(133)\nplt.plot(generator_losses+discriminator_losses, 'b')\nplt.title('$J^G+J^D$')\nplt.show()", "_____no_output_____" ], [ "", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code" ]
[ [ "markdown", "markdown", "markdown", "markdown", "markdown", "markdown" ], [ "code", "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code", "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown", "markdown" ], [ "code", "code" ], [ "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown", "markdown" ], [ "code", "code" ], [ "markdown", "markdown", "markdown", "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code", "code", "code" ], [ "markdown", "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code", "code" ] ]
ecea20d843c3932e1dd179586f4d574a91e4c51b
48,504
ipynb
Jupyter Notebook
OLPS_Comparison.ipynb
rosdyana/OLPS-Comparison
b9f0596d07190e43318c3086aefc5822baf34375
[ "MIT" ]
4
2017-10-31T14:06:12.000Z
2020-11-05T20:13:59.000Z
OLPS_Comparison.ipynb
rosdyana/OLPS-Comparison
b9f0596d07190e43318c3086aefc5822baf34375
[ "MIT" ]
null
null
null
OLPS_Comparison.ipynb
rosdyana/OLPS-Comparison
b9f0596d07190e43318c3086aefc5822baf34375
[ "MIT" ]
3
2017-09-17T09:53:41.000Z
2022-02-06T19:56:56.000Z
31.414508
983
0.592467
[ [ [ "# Comparing OLPS algorithms on a diversified set of ETFs", "_____no_output_____" ], [ "Let's compare the state of the art in OnLine Portfolio Selection (OLPS) algorithms and determine if they can enhance a rebalanced passive strategy in practice. [Online Portfolio Selection: A Survey by Bin Li and Steven C. H. Hoi](http://arxiv.org/abs/1212.2129) provides the most comprehensive review of multi-period portfolio allocation optimization algorithms. The authors [developed](http://olps.stevenhoi.org/) the [OLPS Toolbox]( http://www.mysmu.edu.sg/faculty/chhoi/olps/OLPS_toolbox.pdf), but here we use [Mojmir Vinkler's](https://www.linkedin.com/profile/view?id=210899853) [implementation](https://github.com/Marigold/universal-portfolios) and extend [his comparison](http://nbviewer.ipython.org/github/Marigold/universal-portfolios/blob/master/On-line%20portfolios.ipynb) to a more recent timeline with a set of ETFs to avoid survivorship bias (as suggested by [Ernie Chan](http://epchan.blogspot.cz/2007/01/universal-portfolios.html)) and idiosyncratic risk.\n\nVinkler does all the hard work in his [thesis](http://is.muni.cz/th/358102/prif_m/?lang=en;id=183901), and concludes that Universal Portfolios work practically the same as Constant Rebalanced Portfolios, and work better for an uncorrelated set of small and volatile stocks. Here I'm looking to find if any strategy is applicable to a set of ETFs.\n\nThe agorithms compared are:\n\nType | Name | Algo | Reference \n-------------------------|------|------|----\nBenchmark | BAH | Buy and Hold |\nBenchmark | CRP | Constant Rebalanced Portfolio | T. Cover. [Universal Portfolios](http://www-isl.stanford.edu/~cover/papers/paper93.pdf), 1991.\nBenchmark | UCRP | Uniform CRP (UCRP), a special case of CRP with all weights being equal | T. Cover. [Universal Portfolios](http://www-isl.stanford.edu/~cover/papers/paper93.pdf), 1991.\nBenchmark | BCRP | Best Constant Rebalanced Portfolio | T. Cover. [Universal Portfolios](http://www-isl.stanford.edu/~cover/papers/paper93.pdf), 1991.\nFollow-the-Winner | UP | Universal Portfolio | T. Cover. [Universal Portfolios](http://www-isl.stanford.edu/~cover/papers/paper93.pdf), 1991.\nFollow-the-Winner | EG | Exponential Gradient | Helmbold, David P., et al. [On‐Line Portfolio Selection Using Multiplicative Updates](http://www.cis.upenn.edu/~mkearns/finread/helmbold98line.pdf) Mathematical Finance 8.4 (1998): 325-347.\nFollow-the-Winner | ONS | Online Newton Step | A. Agarwal, E. Hazan, S. Kale, R. E. Schapire. [Algorithms for Portfolio Management based on the Newton Method](http://machinelearning.wustl.edu/mlpapers/paper_files/icml2006_AgarwalHKS06.pdf), 2006.\nFollow-the-Loser | Anticor | Anticorrelation | A. Borodin, R. El-Yaniv, and V. Gogan. [Can we learn to beat the best stock](http://arxiv.org/abs/1107.0036), 2005\nFollow-the-Loser |PAMR | Passive Aggressive Mean Reversion | B. Li, P. Zhao, S. C.H. Hoi, and V. Gopalkrishnan. [Pamr: Passive aggressive mean reversion strategy for portfolio selection](http://www.cais.ntu.edu.sg/~chhoi/paper_pdf/PAMR_ML_final.pdf), 2012.\nFollow-the-Loser |CWMR | Confidence Weighted Mean Reversion | B. Li, S. C. H. Hoi, P. L. Zhao, and V. Gopalkrishnan.[Confidence weighted mean reversion strategy for online portfolio selection](http://jmlr.org/proceedings/papers/v15/li11b/li11b.pdf), 2013. \nFollow-the-Loser | OLMAR | Online Moving Average Reversion| Bin Li and Steven C. H. Hoi [On-Line Portfolio Selection with Moving Average Reversion](http://arxiv.org/abs/1206.4626)\nFollow-the-Loser |RMR | Robust Median Reversion | D. Huang, J. Zhou, B. Li, S. C.vH. Hoi, S. Zhou [Robust Median Reversion Strategy for On-Line Portfolio Selection](http://ijcai.org/papers13/Papers/IJCAI13-296.pdf), 2013.\nPattern Matching | Kelly | Kelly fractional betting |[Kelly Criterion](http://en.wikipedia.org/wiki/Kelly_criterion#Application_to_the_stock_market)\nPattern Matching | BNN | nonparametric nearest neighbor log-optimal | L. Gyorfi, G. Lugosi, and F. Udina. [Nonparametric kernel based sequential investment strategies](http://papers.ssrn.com/sol3/papers.cfm?abstract_id=889976). Mathematical Finance 16 (2006) 337–357.\nPattern Matching | CORN | correlation-driven nonparametric learning | B. Li, S. C. H. Hoi, and V. Gopalkrishnan. [Corn: correlation-driven nonparametric learning approach for portfolio selection](http://www.cais.ntu.edu.sg/~chhoi/paper_pdf/TIST-CORN.pdf), 2011.\n\nWe pick 6 ETFs to avoid survivorship bias and capture broad market diversification. We select the longest running ETF per assset class: [VTI](https://www.google.com/finance?q=VTI), [EFA](https://www.google.com/finance?q=EFA), [EEM](https://www.google.com/finance?q=EFA), [TLT](https://www.google.com/finance?q=TLT), [TIP](https://www.google.com/finance?q=TIP), [VNQ](https://www.google.com/finance?q=VNQ) . We train and select the best parameters on market data from 2005-2012 inclusive (8 years), and test on 2013-2014 inclusive (2 years). ", "_____no_output_____" ] ], [ [ "# You will first need to either download or install universal-portfolios from Vinkler\n# one way to do it is uncomment the line below and execute\n#!pip install --upgrade universal-portfolios \n# or\n#!pip install --upgrade -e [email protected]:Marigold/universal-portfolios.git@master#egg=universal-portfolios\n#\n# if the above fail, git clone [email protected]:marigold/universal-portfolios.git and python setup.py install", "_____no_output_____" ], [ "#Initialize and set debugging level to `debug` to track progress.", "_____no_output_____" ], [ "%matplotlib inline\n\nimport numpy as np\nimport pandas as pd\nfrom pandas_datareader import data as pdr\n# data reader now seperated to new package. pip install pandas-datareader\n#from pandas.io.data import DataReader\nimport fix_yahoo_finance as yf\nyf.pdr_override() # <== that's all it takes :-)\nfrom datetime import datetime\nimport six\nimport universal as up\nfrom universal import tools\nfrom universal import algos\nimport logging\n# we would like to see algos progress\nlogging.basicConfig(format='%(asctime)s %(message)s', level=logging.DEBUG)\n \nimport matplotlib as mpl\nimport matplotlib.pyplot as plt\n\n# we need 14 colors for the plot\n#n_lines = 14\n#color_idx = np.linspace(0, 1, n_lines)\n#mpl.rcParams['axes.color_cycle']=[plt.cm.rainbow(i) for i in color_idx]\n#from cycler import cycler\n#mpl.rcParams['axes.prop_cycle'] = cycler(color='bgrcmyk')\n# Generate random colormap\ndef rand_cmap(nlabels, type='bright', first_color_black=True, last_color_black=False, verbose=True):\n \"\"\"\n Creates a random colormap to be used together with matplotlib. Useful for segmentation tasks\n :param nlabels: Number of labels (size of colormap)\n :param type: 'bright' for strong colors, 'soft' for pastel colors\n :param first_color_black: Option to use first color as black, True or False\n :param last_color_black: Option to use last color as black, True or False\n :param verbose: Prints the number of labels and shows the colormap. True or False\n :return: colormap for matplotlib\n \"\"\"\n from matplotlib.colors import LinearSegmentedColormap\n import colorsys\n import numpy as np\n\n if type not in ('bright', 'soft'):\n print ('Please choose \"bright\" or \"soft\" for type')\n return\n\n if verbose:\n print('Number of labels: ' + str(nlabels))\n\n # Generate color map for bright colors, based on hsv\n if type == 'bright':\n randHSVcolors = [(np.random.uniform(low=0.0, high=1),\n np.random.uniform(low=0.2, high=1),\n np.random.uniform(low=0.9, high=1)) for i in range(nlabels)]\n\n # Convert HSV list to RGB\n randRGBcolors = []\n for HSVcolor in randHSVcolors:\n randRGBcolors.append(colorsys.hsv_to_rgb(HSVcolor[0], HSVcolor[1], HSVcolor[2]))\n\n if first_color_black:\n randRGBcolors[0] = [0, 0, 0]\n\n if last_color_black:\n randRGBcolors[-1] = [0, 0, 0]\n\n random_colormap = LinearSegmentedColormap.from_list('new_map', randRGBcolors, N=nlabels)\n\n # Generate soft pastel colors, by limiting the RGB spectrum\n if type == 'soft':\n low = 0.6\n high = 0.95\n randRGBcolors = [(np.random.uniform(low=low, high=high),\n np.random.uniform(low=low, high=high),\n np.random.uniform(low=low, high=high)) for i in range(nlabels)]\n\n if first_color_black:\n randRGBcolors[0] = [0, 0, 0]\n\n if last_color_black:\n randRGBcolors[-1] = [0, 0, 0]\n random_colormap = LinearSegmentedColormap.from_list('new_map', randRGBcolors, N=nlabels)\n\n # Display colorbar\n if verbose:\n from matplotlib import colors, colorbar\n from matplotlib import pyplot as plt\n fig, ax = plt.subplots(1, 1, figsize=(15, 0.5))\n\n bounds = np.linspace(0, nlabels, nlabels + 1)\n norm = colors.BoundaryNorm(bounds, nlabels)\n\n cb = colorbar.ColorbarBase(ax, cmap=random_colormap, norm=norm, spacing='proportional', ticks=None,\n boundaries=bounds, format='%1i', orientation=u'horizontal')\n\n return random_colormap\n\nmpl.rcParams['figure.figsize'] = (16, 10) # increase the size of graphs\nmpl.rcParams['legend.fontsize'] = 12\nmpl.rcParams['lines.linewidth'] = 1\n#default_color_cycle = mpl.rcParams['axes.color_cycle'] # save this as we will want it back later\nnew_cmap = rand_cmap(100, type='bright', first_color_black=True, last_color_black=False, verbose=True)\ndefault_color_cycle = new_cmap", "_____no_output_____" ], [ "# note what versions we are on:\nimport sys\nprint('Python: '+sys.version)\nprint('Pandas: '+pd.__version__)\nimport pkg_resources\nprint('universal-portfolios: '+pkg_resources.get_distribution(\"universal-portfolios\").version)\nprint('Numpy: '+np.__version__)", "_____no_output_____" ] ], [ [ "## Loading the data", "_____no_output_____" ], [ "We want to train on market data from 2005-2012 inclusive (8 years), and test on 2013-2014 inclusive (2 years). But at this point we accept the default parameters for the respective algorithms and we essentially are looking at two independent time periods. In the future we will want to optimize the paramaters on the train set.", "_____no_output_____" ] ], [ [ "# load data from Yahoo\n# Be careful if you cange the order or types of ETFs to also change the CRP weight %'s in the swensen_allocation\netfs = ['VTI', 'EFA', 'EEM', 'TLT', 'TIP', 'VNQ']\n#etfs = ['SCHX','VOO','FNDX','PRF','SCHA','VB','FNDA','PRFZ','SCHF','VEA',\n# 'FNDF','PXF','SCHC','VSS','FNDC','PDN','SCHE','IEMG','FNDE',\n# 'PXH','SCHH','VNQ','VNQI','IFGL','SCHD','VYM','DWX','IQDF','MLPA',\n# 'ZMLP','SCHR','VGIT','ITR','VCIT','VMBS','MBG','SCHP','STIP','SHYG',\n# 'JNK','BNDX','IGOV','EMLC','VWOB','PSK','PFF','BKLN','VTEB','TFI',\n# 'PWZ','CMF','IAU','GLTR']\n# Swensen allocation from http://www.bogleheads.org/wiki/Lazy_portfolios#David_Swensen.27s_lazy_portfolio\n# as later updated here : https://www.yalealumnimagazine.com/articles/2398/david-swensen-s-guide-to-sleeping-soundly \nswensen_allocation = [0.3, 0.15, 0.1, 0.15, 0.15, 0.15] \nbenchmark = ['SPY']\ntrain_start = datetime(2008,1,1)#'2005-01-01'\ntrain_end = datetime(2012,12,31)#'2012-12-31'\ntest_start = datetime(2013,1,1)#'2013-01-01'\ntest_end = datetime(2014,12,31)#'2014-12-31'\n#train = DataReader(etfs, 'yahoo', start=train_start, end=train_end)['Adj Close']\n#test = DataReader(etfs, 'yahoo', start=test_start, end=test_end)['Adj Close']\n#train_b = DataReader(benchmark, 'yahoo', start=train_start, end=train_end)['Adj Close']\n#test_b = DataReader(benchmark, 'yahoo', start=test_start, end=test_end)['Adj Close']\ntrain = pdr.get_data_yahoo(etfs, train_start, train_end)['Adj Close']\ntest = pdr.get_data_yahoo(etfs, test_start, test_end)['Adj Close']\ntrain_b = pdr.get_data_yahoo(benchmark, train_start, train_end)['Adj Close']\ntest_b = pdr.get_data_yahoo(benchmark, test_start, test_end)['Adj Close']\n", "_____no_output_____" ], [ "# plot normalized prices of the train set\n#idx = pd.IndexSlice\n#ax1 = (train / train.iloc[idx[0,:]]).plot()\n#(train_b / train_b.iloc[idx[0,:]]).plot(ax=ax1)\n(train / train.iloc[0,:]).plot()", "_____no_output_____" ], [ "# plot normalized prices of the test set\n#ax2 = (test / test.iloc[0,:]).plot()\n#(test_b / test_b.iloc[0,:]).plot(ax=ax2)\n(test / test.iloc[0,:]).plot()", "_____no_output_____" ] ], [ [ "# Comparing the Algorithms", "_____no_output_____" ], [ "We want to train on market data from a number of years, and test out of sample for a duration smaller than the train set. To get started we accept the default parameters for the respective algorithms and we essentially are just looking at two independent time periods. In the future we will want to optimize the paramaters on the train set.", "_____no_output_____" ] ], [ [ "#list all the algos\nolps_algos = [\nalgos.Anticor(),\nalgos.BAH(),\nalgos.BCRP(),\nalgos.BNN(),\nalgos.CORN(),\nalgos.CRP(b=swensen_allocation), # Non Uniform CRP (the Swensen allocation)\nalgos.CWMR(),\nalgos.EG(),\nalgos.Kelly(),\nalgos.OLMAR(),\nalgos.ONS(),\nalgos.PAMR(),\nalgos.RMR(),\nalgos.UP()\n]", "_____no_output_____" ], [ "# put all the algos in a dataframe\nalgo_names = [a.__class__.__name__ for a in olps_algos]\nalgo_data = ['algo', 'results', 'profit', 'sharpe', 'information', 'annualized_return', 'drawdown_period','winning_pct']\nmetrics = algo_data[2:]\nolps_train = pd.DataFrame(index=algo_names, columns=algo_data)\nolps_train.algo = olps_algos", "_____no_output_____" ] ], [ [ "At this point we could train all the algos to find the best parameters for each.", "_____no_output_____" ] ], [ [ "# run all algos - this takes more than a minute\nfor name, alg in zip(olps_train.index, olps_train.algo):\n olps_train.loc[name,'results'] = alg.run(train)", "_____no_output_____" ], [ "# Let's make sure the fees are set to 0 at first\nfor k, r in olps_train.results.iteritems():\n r.fee = 0.0", "_____no_output_____" ], [ "import cycler\nn_lines = 14\ncolor_idx = np.linspace(0, 1, n_lines)\nmpl.rcParams['axes.prop_cycle']=cycler.cycler(color=[plt.cm.rainbow(i) for i in color_idx])\n# plot as if we had no fees\n# get the first result so we can grab the figure axes from the plot\n#ax = olps_train.results[0].plot(assets=False, weights=False, ucrp=True, portfolio_label=olps_train.index[0])\n#for k, r in olps_train.results.iteritems():\n #if k == olps_train.results.keys()[0]: # skip the first item because we have it already\n #continue\n #r.plot(assets=False, weights=False, ucrp=False, portfolio_label=k, ax=ax[0])\n#ax = olps_train.results[0].plot(assets=False, weights=False, ucrp=True, portfolio_label=olps_train.index[0], color='red')\nfor k, r in olps_train.results.iteritems():\n #if k == olps_train.results.keys()[0]: # skip the first item because we have it already\n #continue\n r.plot(assets=False, weights=False, ucrp=True, portfolio_label=k, color='red')", "_____no_output_____" ], [ "def olps_stats(df):\n for name, r in df.results.iteritems():\n df.ix[name,'profit'] = r.profit_factor\n df.ix[name,'sharpe'] = r.sharpe\n df.ix[name,'information'] = r.information\n df.ix[name,'annualized_return'] = r.annualized_return * 100\n df.ix[name,'drawdown_period'] = r.drawdown_period\n df.ix[name,'winning_pct'] = r.winning_pct * 100\n return df", "_____no_output_____" ], [ "olps_stats(olps_train)\nolps_train[metrics].sort_values('profit', ascending=False)", "_____no_output_____" ], [ "# Let's add fees of 0.1% per transaction (we pay $1 for every $1000 of stocks bought or sold).\nfor k, r in olps_train.results.iteritems():\n r.fee = 0.001", "_____no_output_____" ], [ "# plot with fees\n# get the first result so we can grab the figure axes from the plot\nimport plotly.plotly as py\nimport plotly.graph_objs as go\nax = olps_train.results[0].plot(assets=False, weights=False, ucrp=True, portfolio_label=olps_train.index[0])\nfor k, r in olps_train.results.iteritems():\n if k == olps_train.results.keys()[0]: # skip the first item because we have it already\n continue\n r.plot(assets=False, weights=False, ucrp=False, portfolio_label=k, ax=ax[0])", "_____no_output_____" ] ], [ [ "### Notice how Kelly crashes right away and how RMR and OLMAR float to the top after some high volatility. ", "_____no_output_____" ] ], [ [ "olps_stats(olps_train)\nolps_train[metrics].sort_values('profit', ascending=False)", "_____no_output_____" ] ], [ [ "# Run on the Test Set", "_____no_output_____" ] ], [ [ "# create the test set dataframe\nolps_test = pd.DataFrame(index=algo_names, columns=algo_data)\nolps_test.algo = olps_algos", "_____no_output_____" ], [ "# run all algos\nfor name, alg in zip(olps_test.index, olps_test.algo):\n olps_test.ix[name,'results'] = alg.run(test)", "_____no_output_____" ], [ "# Let's make sure the fees are 0 at first\nfor k, r in olps_test.results.iteritems():\n r.fee = 0.0", "_____no_output_____" ], [ "# plot as if we had no fees\n# get the first result so we can grab the figure axes from the plot\nax = olps_test.results[0].plot(assets=False, weights=False, ucrp=True, portfolio_label=olps_test.index[0])\nfor k, r in olps_test.results.iteritems():\n if k == olps_test.results.keys()[0]: # skip the first item because we have it already\n continue\n r.plot(assets=False, weights=False, ucrp=False, portfolio_label=k, ax=ax[0])", "_____no_output_____" ] ], [ [ "### Kelly went wild and crashed, so let's remove it from the mix", "_____no_output_____" ] ], [ [ "# plot as if we had no fees\n# get the first result so we can grab the figure axes from the plot\nax = olps_test.results[0].plot(assets=False, weights=False, ucrp=True, portfolio_label=olps_test.index[0])\nfor k, r in olps_test.results.iteritems():\n if k == olps_test.results.keys()[0] or k == 'Kelly': # skip the first item because we have it already\n continue\n r.plot(assets=False, weights=False, ucrp=False, portfolio_label=k, ax=ax[0])", "_____no_output_____" ], [ "olps_stats(olps_test)\nolps_test[metrics].sort_values('profit', ascending=False)", "_____no_output_____" ] ], [ [ "### Wow, ONS and OLMAR are at the bottom of the list. Remember, we really didn't do any training, but if we had selected ONS or OLMAR at the beginning of 2013 based on past performance, we would not have beat BAH. Hm.", "_____no_output_____" ], [ "# Focusing on OLMAR", "_____no_output_____" ], [ "Instead of using the default parameters, we will test several `window` parameters to see if we can get OLMAR to improve.", "_____no_output_____" ] ], [ [ "# we need need fewer colors so let's reset the colors_cycle\nmpl.rcParams['axes.prop_cycle']= default_color_cycle", "_____no_output_____" ], [ "train_olmar = algos.OLMAR.run_combination(train, window=[3,5,10,15], eps=10)\ntrain_olmar.plot()", "_____no_output_____" ], [ "print(train_olmar.summary())", "_____no_output_____" ], [ "train_olmar = algos.OLMAR.run_combination(train, window=5, eps=[3,5,10,15])\ntrain_olmar.plot()", "_____no_output_____" ], [ "print(train_olmar.summary())", "_____no_output_____" ] ], [ [ "### We find that a window of 5 and eps are 5 are optimal over the train time period, but the default of w=5 and eps=10 were also fine for our purposes.", "_____no_output_____" ] ], [ [ "# OLMAR vs UCRP\nbest_olmar = train_olmar[1]\nax1 = best_olmar.plot(ucrp=True, bah=True, weights=False, assets=False, portfolio_label='OLMAR')\nolps_train.loc['CRP'].results.plot(ucrp=False, bah=False, weights=False, assets=False, ax=ax1[0], portfolio_label='CRP')", "_____no_output_____" ] ], [ [ "### On the train set OLMAR really delivers over CRP !", "_____no_output_____" ] ], [ [ "# let's print the stats\nprint(best_olmar.summary())", "_____no_output_____" ] ], [ [ "### Let's see how individual ETFs contribute to portfolio equity.", "_____no_output_____" ] ], [ [ "best_olmar.plot_decomposition(legend=True, logy=True)", "_____no_output_____" ] ], [ [ "### Let's highlight the magnitude of the highest contributing ETF by removing the log scale and looking at it directly.", "_____no_output_____" ] ], [ [ "best_olmar.plot_decomposition(legend=True, logy=False)", "_____no_output_____" ] ], [ [ "### So VNQ (Real Estate) is the big driver after the market crash of 2008, which makes sense.", "_____no_output_____" ], [ "### Let's look at portfolio allocations", "_____no_output_____" ] ], [ [ "best_olmar.plot(weights=True, assets=True, ucrp=False, logy=True, portfolio_label='OLMAR')", "_____no_output_____" ] ], [ [ "### VNQ is the big driver of wealth (log scale). Let's test the strategy by removing the most profitable stock and comparing Total Wealth.", "_____no_output_____" ] ], [ [ "# find the name of the most profitable asset\nmost_profitable = best_olmar.equity_decomposed.iloc[-1].argmax()\n\n# rerun algorithm on data without it\nresult_without = algos.OLMAR().run(train.drop([most_profitable], 1))\n\n# and print results\nprint(result_without.summary())\nresult_without.plot(weights=False, assets=False, bah=True, ucrp=True, logy=True, portfolio_label='OLMAR-VNQ')", "_____no_output_____" ], [ "result_without.plot_decomposition(legend=True, logy=False)", "_____no_output_____" ] ], [ [ "### Let's add fees of 0.1% per transaction (we pay \\$1 for every \\$1000 of stocks bought or sold).", "_____no_output_____" ] ], [ [ "best_olmar.fee = 0.001\nprint(best_olmar.summary())\nbest_olmar.plot(weights=False, assets=False, bah=True, ucrp=True, logy=True, portfolio_label='OLMAR')", "_____no_output_____" ] ], [ [ "### The results now fall, with a Sharpe Ratio below the ~0.5 market Sharpe, and an annualized return that has been cut in half due to fees. It's as if all the trading makes OLMAR underperform for the first 4 years until it can grab some volatility in 2008 to beat UCRP.", "_____no_output_____" ], [ "### Let's look at OLMAR in the test time frame", "_____no_output_____" ] ], [ [ "test_olmar = algos.OLMAR(window=5, eps=5).run(test)\n#print(train_olmar.summary())\ntest_olmar.plot(ucrp=True, bah=True, weights=False, assets=False, portfolio_label='OLMAR')", "_____no_output_____" ] ], [ [ "### With fees", "_____no_output_____" ] ], [ [ "test_olmar.fee = 0.001\nprint(test_olmar.summary())\ntest_olmar.plot(weights=False, assets=False, bah=True, ucrp=True, logy=True, portfolio_label='OLMAR')", "_____no_output_____" ] ], [ [ "# OLMAR Starting in 2010", "_____no_output_____" ], [ "The 2008-2009 recession was unique. Let's try it all again starting in 2010, with a train set from 2010-2013 inclusive, and a test set of 2014.", "_____no_output_____" ] ], [ [ "# set train and test time periods\ntrain_start_2010= datetime(2010,1,1)\ntrain_end_2010 = datetime(2013,12,31)\ntest_start_2010 = datetime(2014,1,1)\ntest_end_2010 = datetime(2014,12,31)", "_____no_output_____" ], [ "# load data from Yahoo\n#train_2010 = DataReader(etfs, 'yahoo', start=train_start_2010, end=train_end_2010)['Adj Close']\n#test_2010 = DataReader(etfs, 'yahoo', start=test_start_2010, end=test_end_2010)['Adj Close']\ntrain_2010 = pdr.get_data_yahoo(etfs, train_start_2010, train_end_2010)['Adj Close']\ntest_2010 = pdr.get_data_yahoo(etfs, train_start_2010, train_end_2010)['Adj Close']", "_____no_output_____" ], [ "# plot normalized prices of these stocks\n(train_2010 / train_2010.iloc[0,:]).plot()", "_____no_output_____" ], [ "# plot normalized prices of these stocks\n(test_2010 / test_2010.iloc[0,:]).plot()", "_____no_output_____" ], [ "train_olmar_2010 = algos.OLMAR().run(train_2010)\ntrain_crp_2010 = algos.CRP(b=swensen_allocation).run(train_2010)\nax1 = train_olmar_2010.plot(assets=True, weights=False, ucrp=True, bah=True, portfolio_label='OLMAR')\ntrain_crp_2010.plot(ucrp=False, bah=False, weights=False, assets=False, ax=ax1[0], portfolio_label='CRP')", "_____no_output_____" ], [ "print(train_olmar_2010.summary())", "_____no_output_____" ], [ "train_olmar_2010.plot_decomposition(legend=True, logy=True)", "_____no_output_____" ] ], [ [ "Not bad, with a Sharpe at 1 and no one ETF dominating the portfolio. Now let's see how it fairs in 2014. ", "_____no_output_____" ] ], [ [ "test_olmar_2010 = algos.OLMAR().run(test_2010)\ntest_crp_2010 = algos.CRP(b=swensen_allocation).run(test_2010)\nax1 = test_olmar_2010.plot(assets=True, weights=False, ucrp=True, bah=True, portfolio_label='OLMAR')\ntest_crp_2010.plot(ucrp=False, bah=False, weights=False, assets=False, ax=ax1[0], portfolio_label='CRP')", "_____no_output_____" ], [ "print(test_olmar_2010.summary())", "_____no_output_____" ] ], [ [ "We just happen to be looking at a different time period and now the Sharpe drops below 0.5 and OLMAR fails to beat BAH. Not good.", "_____no_output_____" ] ], [ [ "test_olmar_2010.plot_decomposition(legend=True, logy=True)", "_____no_output_____" ] ], [ [ "# SPY / TLT portfolio comparison", "_____no_output_____" ], [ "Let's step back and simplify this by looking at OLMAR on a SPY and TLT portfolio. We should also compare this portfolio to a rebalanced 70/30 mix of SPY and TLT.", "_____no_output_____" ] ], [ [ "# load data from Yahoo\n#spy_tlt_data = DataReader(['SPY', 'TLT'], 'yahoo', start=datetime(2010,1,1))['Adj Close']\nspy_tlt_data = pdr.get_data_yahoo(['SPY', 'TLT'], datetime(2010,1,1))['Adj Close']\n\n# plot normalized prices of these stocks\n(spy_tlt_data / spy_tlt_data.iloc[0,:]).plot()", "_____no_output_____" ], [ "import plotly.plotly as py\nimport plotly.graph_objs as go\n\nimport pandas as pd\n\nTLT = pd.read_csv(\"stock_datas/TLT.csv\")\nSPY = pd.read_csv(\"stock_datas/SPY.csv\")\ndf = pd.concat([TLT,SPY.rename(columns={'date':'datex'})], ignore_index=True)\n\ntrace_spy = go.Scatter(\n x=df.Date,\n y=df['SPY.Adjusted'],\n name = \"SPY Adjusted\",\n line = dict(color = '#17BECF'),\n opacity = 0.8)\n\ntrace_tlt = go.Scatter(\n x=df.Date,\n y=df['TLT.Adjusted'],\n name = \"TLT Adjusted\",\n line = dict(color = '#7F7F7F'),\n opacity = 0.8)\n\ndata = [trace_spy,trace_tlt]\n\nlayout = dict(\n title = \"SPY / TLT portfolio comparison\",\n xaxis = dict(\n range = ['2010-01-01','2017-07-17'])\n)\n\nfig = dict(data=data, layout=layout)\n\npy.iplot(fig)", "_____no_output_____" ], [ "spy_tlt_olmar_2010 = algos.OLMAR().run(spy_tlt_data)\nspy_tlt_olmar_2010.plot(assets=True, weights=True, ucrp=True, bah=True, portfolio_label='OLMAR')", "_____no_output_____" ], [ "spy_tlt_olmar_2010.plot_decomposition(legend=True, logy=True)", "_____no_output_____" ], [ "print(spy_tlt_olmar_2010.summary())", "_____no_output_____" ], [ "spy_tlt_2010 = algos.CRP(b=[0.7, 0.3]).run(spy_tlt_data)\n\nax1 = spy_tlt_olmar_2010.plot(assets=False, weights=False, ucrp=True, bah=True, portfolio_label='OLMAR')\nspy_tlt_2010.plot(assets=False, weights=False, ucrp=False, bah=False, portfolio_label='CRP', ax=ax1[0])", "_____no_output_____" ] ], [ [ "## Now OLMAR looks better!", "_____no_output_____" ], [ "# OLMAR Market Sectors comparison", "_____no_output_____" ], [ "Let's look at algo behavior on market sectors:\n\n- XLY Consumer Discrectionary SPDR Fund \n- XLF Financial SPDR Fund \n- XLK Technology SPDR Fund \n- XLE Energy SPDR Fund \n- XLV Health Care SPRD Fund \n- XLI Industrial SPDR Fund \n- XLP Consumer Staples SPDR Fund \n- XLB Materials SPDR Fund \n- XLU Utilities SPRD Fund ", "_____no_output_____" ] ], [ [ "sectors = ['XLY','XLF','XLK','XLE','XLV','XLI','XLP','XLB','XLU']\n#train_sectors = DataReader(sectors, 'yahoo', start=train_start_2010, end=train_end_2010)['Adj Close']\n#test_sectors = DataReader(sectors, 'yahoo', start=test_start_2010, end=test_end_2010)['Adj Close']\ntrain_sectors = pdr.get_data_yahoo(sectors, train_start_2010, train_end_2010)['Adj Close']\ntest_sectors = pdr.get_data_yahoo(sectors, test_start_2010, test_end_2010)['Adj Close']", "_____no_output_____" ], [ "# plot normalized prices of these stocks\n(train_sectors / train_sectors.iloc[0,:]).plot()", "_____no_output_____" ], [ "# plot normalized prices of these stocks\n(test_sectors / test_sectors.iloc[0,:]).plot()", "_____no_output_____" ], [ "train_olmar_sectors = algos.OLMAR().run(train_sectors)\ntrain_olmar_sectors.plot(assets=True, weights=False, ucrp=True, bah=True, portfolio_label='OLMAR')", "_____no_output_____" ], [ "print(train_olmar_sectors.summary())", "_____no_output_____" ], [ "train_olmar_sectors.plot(assets=False, weights=False, ucrp=True, bah=True, portfolio_label='OLMAR')", "_____no_output_____" ], [ "test_olmar_sectors = algos.OLMAR().run(test_sectors)\ntest_olmar_sectors.plot(assets=True, weights=False, ucrp=True, bah=True, portfolio_label='OLMAR')", "_____no_output_____" ], [ "test_olmar_sectors = algos.OLMAR().run(test_sectors)\ntest_olmar_sectors.plot(assets=False, weights=False, ucrp=True, bah=True, portfolio_label='OLMAR')", "_____no_output_____" ] ], [ [ "# All OLPS Algos Market Sectors comparison", "_____no_output_____" ] ], [ [ "#list all the algos\nolps_algos_sectors = [\nalgos.Anticor(),\nalgos.BAH(),\nalgos.BCRP(),\nalgos.BNN(),\nalgos.CORN(),\nalgos.CRP(), # removed weights, and thus equivalent to UCRP\nalgos.CWMR(),\nalgos.EG(),\nalgos.Kelly(),\nalgos.OLMAR(),\nalgos.ONS(),\nalgos.PAMR(),\nalgos.RMR(),\nalgos.UP()\n]", "_____no_output_____" ], [ "olps_sectors_train = pd.DataFrame(index=algo_names, columns=algo_data)\nolps_sectors_train.algo = olps_algos_sectors", "_____no_output_____" ], [ "# run all algos - this takes more than a minute\nfor name, alg in zip(olps_sectors_train.index, olps_sectors_train.algo):\n olps_sectors_train.ix[name,'results'] = alg.run(train_sectors)", "_____no_output_____" ], [ "# we need 14 colors for the plot\nn_lines = 14\ncolor_idx = np.linspace(0, 1, n_lines)\nmpl.rcParams['axes.color_cycle']=[plt.cm.rainbow(i) for i in color_idx]", "_____no_output_____" ], [ "# plot as if we had no fees\n# get the first result so we can grab the figure axes from the plot\nolps_df = olps_sectors_train\nax = olps_df.results[0].plot(assets=False, weights=False, ucrp=True, portfolio_label=olps_df.index[0])\nfor k, r in olps_df.results.iteritems():\n if k == olps_df.results.keys()[0]: # skip the first item because we have it already\n continue\n r.plot(assets=False, weights=False, ucrp=False, portfolio_label=k, ax=ax[0])", "_____no_output_____" ], [ "# Kelly went wild, so let's remove it\n# get the first result so we can grab the figure axes from the plot\nolps_df = olps_sectors_train\nax = olps_df.results[0].plot(assets=False, weights=False, ucrp=True, portfolio_label=olps_df.index[0])\nfor k, r in olps_df.results.iteritems():\n if k == olps_df.results.keys()[0] or k == 'Kelly' : # skip the first item because we have it already\n continue\n r.plot(assets=False, weights=False, ucrp=False, portfolio_label=k, ax=ax[0])", "_____no_output_____" ], [ "olps_stats(olps_sectors_train)\nolps_sectors_train[metrics].sort_values('profit', ascending=False)", "_____no_output_____" ], [ "# create the test set dataframe\nolps_sectors_test = pd.DataFrame(index=algo_names, columns=algo_data)\nolps_sectors_test.algo = olps_algos_sectors", "_____no_output_____" ], [ "# run all algos\nfor name, alg in zip(olps_sectors_test.index, olps_sectors_test.algo):\n olps_sectors_test.ix[name,'results'] = alg.run(test_sectors)", "_____no_output_____" ], [ "# plot as if we had no fees\n# get the first result so we can grab the figure axes from the plot\nolps_df = olps_sectors_test\nax = olps_df.results[0].plot(assets=False, weights=False, ucrp=True, portfolio_label=olps_df.index[0])\nfor k, r in olps_df.results.iteritems():\n if k == olps_df.results.keys()[0] : #or k == 'Kelly': # skip the first item because we have it already\n continue\n r.plot(assets=False, weights=False, ucrp=False, portfolio_label=k, ax=ax[0])", "_____no_output_____" ], [ "# drop Kelly !\n# get the first result so we can grab the figure axes from the plot\nolps_df = olps_sectors_test\nax = olps_df.results[0].plot(assets=False, weights=False, ucrp=True, portfolio_label=olps_df.index[0])\nfor k, r in olps_df.results.iteritems():\n if k == olps_df.results.keys()[0] or k == 'Kelly': # skip the first item because we have it already\n continue\n r.plot(assets=False, weights=False, ucrp=False, portfolio_label=k, ax=ax[0])", "_____no_output_____" ], [ "olps_stats(olps_sectors_test)\nolps_sectors_test[metrics].sort_values('profit', ascending=False)", "_____no_output_____" ] ], [ [ "# Further work", "_____no_output_____" ], [ "- More algo's could be optimized for parameters before they are run against the test set\n- In addition to the BAH, CRP and BCRP benchmarks, we could consider holding [SPY](https://www.google.com/finance?q=SPY) at 100% as a benchmark.\n- Could look into BAH(OLMAR) and other combinations as this framework supports combining approaches directly\n- Experiment with the ```run_subsets``` feature", "_____no_output_____" ], [ "# Conclusion", "_____no_output_____" ], [ "RMR and OLMAR do add value to a Lazy Portfolio if tested or run over a long enough period of time. This gives RMR and OLMAR a chance to grab onto a period of volatility. But in an up market (2013-1014) you want to Follow-the-Leader, not Follow-the-Looser. Of the other algo's, CRP or BAH are decent, and maybe it's worth understanding what ONS is doing.", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown" ]
[ [ "markdown", "markdown" ], [ "code", "code", "code", "code" ], [ "markdown", "markdown" ], [ "code", "code", "code" ], [ "markdown", "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code", "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown", "markdown", "markdown" ], [ "code", "code", "code", "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code", "code", "code", "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code", "code", "code", "code", "code", "code" ], [ "markdown", "markdown", "markdown" ], [ "code", "code", "code", "code", "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code" ], [ "markdown", "markdown", "markdown", "markdown" ] ]
ecea20decb8ca0307e3ca0e098b290b308eaca9b
21,945
ipynb
Jupyter Notebook
courses/machine_learning/deepdive/06_structured/labs/5_train.ipynb
Glairly/introduction_to_tensorflow
aa0a44d9c428a6eb86d1f79d73f54c0861b6358d
[ "Apache-2.0" ]
2
2022-01-06T11:52:57.000Z
2022-01-09T01:53:56.000Z
courses/machine_learning/deepdive/06_structured/labs/5_train.ipynb
Glairly/introduction_to_tensorflow
aa0a44d9c428a6eb86d1f79d73f54c0861b6358d
[ "Apache-2.0" ]
null
null
null
courses/machine_learning/deepdive/06_structured/labs/5_train.ipynb
Glairly/introduction_to_tensorflow
aa0a44d9c428a6eb86d1f79d73f54c0861b6358d
[ "Apache-2.0" ]
null
null
null
38.165217
554
0.547368
[ [ [ "<h1>Training on Cloud AI Platform</h1>\n\nThis notebook illustrates distributed training on Cloud AI Platform (formerly known as Cloud ML Engine).", "_____no_output_____" ] ], [ [ "!sudo chown -R jupyter:jupyter /home/jupyter/training-data-analyst", "_____no_output_____" ], [ "# Ensure the right version of Tensorflow is installed.\n!pip freeze | grep tensorflow==2.1", "_____no_output_____" ], [ "# change these to try this notebook out\nBUCKET = 'cloud-training-demos-ml'\nPROJECT = 'cloud-training-demos'\nREGION = 'us-central1'", "_____no_output_____" ], [ "import os\nos.environ['BUCKET'] = BUCKET\nos.environ['PROJECT'] = PROJECT\nos.environ['REGION'] = REGION\nos.environ['TFVERSION'] = '2.1'", "_____no_output_____" ], [ "%%bash\ngcloud config set project $PROJECT\ngcloud config set compute/region $REGION", "_____no_output_____" ], [ "%%bash\nif ! gsutil ls | grep -q gs://${BUCKET}/babyweight/preproc; then\n gsutil mb -l ${REGION} gs://${BUCKET}\n # copy canonical set of preprocessed files if you didn't do previous notebook\n gsutil -m cp -R gs://cloud-training-demos/babyweight gs://${BUCKET}\nfi", "_____no_output_____" ], [ "%%bash\ngsutil ls gs://${BUCKET}/babyweight/preproc/*-00000*", "_____no_output_____" ] ], [ [ "Now that we have the TensorFlow code working on a subset of the data, we can package the TensorFlow code up as a Python module and train it on Cloud AI Platform.\n<p>\n<h2> Train on Cloud AI Platform</h2>\n<p>\nTraining on Cloud AI Platform requires:\n<ol>\n<li> Making the code a Python package\n<li> Using gcloud to submit the training code to Cloud AI Platform\n</ol>\n\nEnsure that the AI Platform API is enabled by going to this [link](https://console.developers.google.com/apis/library/ml.googleapis.com).", "_____no_output_____" ], [ "## Lab Task 1\n\nThe following code edits babyweight/trainer/task.py. You should use add hyperparameters needed by your model through the command-line using the `parser` module. Look at how `batch_size` is passed to the model in the code below. Do this for the following hyperparameters (defaults in parentheses): `train_examples` (5000), `eval_steps` (None), `pattern` (of).", "_____no_output_____" ] ], [ [ "%%writefile babyweight/trainer/task.py\nimport argparse\nimport json\nimport os\n\nfrom . import model\n\nimport tensorflow.compat.v1 as tf\ntf.disable_v2_behavior()\n\nif __name__ == '__main__':\n parser = argparse.ArgumentParser()\n parser.add_argument(\n '--bucket',\n help = 'GCS path to data. We assume that data is in gs://BUCKET/babyweight/preproc/',\n required = True\n )\n parser.add_argument(\n '--output_dir',\n help = 'GCS location to write checkpoints and export models',\n required = True\n )\n parser.add_argument(\n '--batch_size',\n help = 'Number of examples to compute gradient over.',\n type = int,\n default = 512\n )\n parser.add_argument(\n '--job-dir',\n help = 'this model ignores this field, but it is required by gcloud',\n default = 'junk'\n )\n parser.add_argument(\n '--nnsize',\n help = 'Hidden layer sizes to use for DNN feature columns -- provide space-separated layers',\n nargs = '+',\n type = int,\n default=[128, 32, 4]\n )\n parser.add_argument(\n '--nembeds',\n help = 'Embedding size of a cross of n key real-valued parameters',\n type = int,\n default = 3\n )\n\n ## TODOs after this line\n ################################################################################\n \n ## TODO 1: add the new arguments here \n\n ## parse all arguments\n args = parser.parse_args()\n arguments = args.__dict__\n\n # unused args provided by service\n arguments.pop('job_dir', None)\n arguments.pop('job-dir', None)\n\n ## assign the arguments to the model variables\n output_dir = arguments.pop('output_dir')\n model.BUCKET = arguments.pop('bucket')\n model.BATCH_SIZE = arguments.pop('batch_size')\n model.TRAIN_STEPS = (arguments.pop('train_examples') * 100) / model.BATCH_SIZE\n model.EVAL_STEPS = arguments.pop('eval_steps') \n print (\"Will train for {} steps using batch_size={}\".format(model.TRAIN_STEPS, model.BATCH_SIZE))\n model.PATTERN = arguments.pop('pattern')\n model.NEMBEDS= arguments.pop('nembeds')\n model.NNSIZE = arguments.pop('nnsize')\n print (\"Will use DNN size of {}\".format(model.NNSIZE))\n\n # Append trial_id to path if we are doing hptuning\n # This code can be removed if you are not using hyperparameter tuning\n output_dir = os.path.join(\n output_dir,\n json.loads(\n os.environ.get('TF_CONFIG', '{}')\n ).get('task', {}).get('trial', '')\n )\n\n # Run the training job\n model.train_and_evaluate(output_dir)", "_____no_output_____" ] ], [ [ "## Lab Task 2\n\nAddress all the TODOs in the following code in `babyweight/trainer/model.py` with the cell below. This code is similar to the model training code we wrote in Lab 3. \n\nAfter addressing all TODOs, run the cell to write the code to the model.py file.", "_____no_output_____" ] ], [ [ "%%writefile babyweight/trainer/model.py\nimport shutil\nimport numpy as np\n\nimport tensorflow.compat.v1 as tf\ntf.disable_v2_behavior()\n\ntf.logging.set_verbosity(tf.logging.INFO)\n\nBUCKET = None # set from task.py\nPATTERN = 'of' # gets all files\n\n# Determine CSV, label, and key columns\nCSV_COLUMNS = 'weight_pounds,is_male,mother_age,plurality,gestation_weeks,key'.split(',')\nLABEL_COLUMN = 'weight_pounds'\nKEY_COLUMN = 'key'\n\n# Set default values for each CSV column\nDEFAULTS = [[0.0], ['null'], [0.0], ['null'], [0.0], ['nokey']]\n\n# Define some hyperparameters\nTRAIN_STEPS = 10000\nEVAL_STEPS = None\nBATCH_SIZE = 512\nNEMBEDS = 3\nNNSIZE = [64, 16, 4]\n\n# Create an input function reading a file using the Dataset API\n# Then provide the results to the Estimator API\ndef read_dataset(prefix, mode, batch_size):\n def _input_fn():\n def decode_csv(value_column):\n columns = tf.decode_csv(value_column, record_defaults=DEFAULTS)\n features = dict(zip(CSV_COLUMNS, columns))\n label = features.pop(LABEL_COLUMN)\n return features, label\n \n # Use prefix to create file path\n file_path = 'gs://{}/babyweight/preproc/{}*{}*'.format(BUCKET, prefix, PATTERN)\n\n # Create list of files that match pattern\n file_list = tf.gfile.Glob(file_path)\n\n # Create dataset from file list\n dataset = (tf.data.TextLineDataset(file_list) # Read text file\n .map(decode_csv)) # Transform each elem by applying decode_csv fn\n \n if mode == tf.estimator.ModeKeys.TRAIN:\n num_epochs = None # indefinitely\n dataset = dataset.shuffle(buffer_size = 10 * batch_size)\n else:\n num_epochs = 1 # end-of-input after this\n \n dataset = dataset.repeat(num_epochs).batch(batch_size)\n return dataset.make_one_shot_iterator().get_next()\n return _input_fn\n\n# Define feature columns\ndef get_wide_deep():\n # Define column types\n is_male,mother_age,plurality,gestation_weeks = \\\n [\\\n tf.feature_column.categorical_column_with_vocabulary_list('is_male', \n ['True', 'False', 'Unknown']),\n tf.feature_column.numeric_column('mother_age'),\n tf.feature_column.categorical_column_with_vocabulary_list('plurality',\n ['Single(1)', 'Twins(2)', 'Triplets(3)',\n 'Quadruplets(4)', 'Quintuplets(5)','Multiple(2+)']),\n tf.feature_column.numeric_column('gestation_weeks')\n ]\n\n # Discretize\n age_buckets = tf.feature_column.bucketized_column(mother_age, \n boundaries=np.arange(15,45,1).tolist())\n gestation_buckets = tf.feature_column.bucketized_column(gestation_weeks, \n boundaries=np.arange(17,47,1).tolist())\n \n # Sparse columns are wide, have a linear relationship with the output\n wide = [is_male,\n plurality,\n age_buckets,\n gestation_buckets]\n \n # Feature cross all the wide columns and embed into a lower dimension\n crossed = tf.feature_column.crossed_column(wide, hash_bucket_size=20000)\n embed = tf.feature_column.embedding_column(crossed, NEMBEDS)\n \n # Continuous columns are deep, have a complex relationship with the output\n deep = [mother_age,\n gestation_weeks,\n embed]\n return wide, deep\n\n# Create serving input function to be able to serve predictions later using provided inputs\ndef serving_input_fn():\n feature_placeholders = {\n 'is_male': tf.placeholder(tf.string, [None]),\n 'mother_age': tf.placeholder(tf.float32, [None]),\n 'plurality': tf.placeholder(tf.string, [None]),\n 'gestation_weeks': tf.placeholder(tf.float32, [None]),\n KEY_COLUMN: tf.placeholder_with_default(tf.constant(['nokey']), [None])\n }\n features = {\n key: tf.expand_dims(tensor, -1)\n for key, tensor in feature_placeholders.items()\n }\n return tf.estimator.export.ServingInputReceiver(features, feature_placeholders)\n\n# create metric for hyperparameter tuning\ndef my_rmse(labels, predictions):\n pred_values = predictions['predictions']\n return {'rmse': tf.metrics.root_mean_squared_error(labels, pred_values)}\n\ndef forward_features(estimator, key):\n def new_model_fn(features, labels, mode, config):\n spec = estimator.model_fn(features, labels, mode, config)\n predictions = spec.predictions\n predictions[key] = features[key]\n spec = spec._replace(predictions=predictions)\n return spec\n return tf.estimator.Estimator(model_fn=new_model_fn, model_dir=estimator.model_dir, config=estimator.config)\n\n## TODOs after this line\n################################################################################\n\n# Create estimator to train and evaluate\ndef train_and_evaluate(output_dir):\n tf.summary.FileWriterCache.clear() # ensure filewriter cache is clear for TensorBoard events file\n wide, deep = get_wide_deep()\n EVAL_INTERVAL = 300 # seconds\n\n ## TODO 2a: set the save_checkpoints_secs to the EVAL_INTERVAL\n run_config = tf.estimator.RunConfig(save_checkpoints_secs = None,\n keep_checkpoint_max = 3)\n \n ## TODO 2b: change the dnn_hidden_units to NNSIZE\n estimator = tf.estimator.DNNLinearCombinedRegressor(\n model_dir = output_dir,\n linear_feature_columns = wide,\n dnn_feature_columns = deep,\n dnn_hidden_units = None,\n config = run_config)\n \n # illustrates how to add an extra metric\n estimator = tf.estimator.add_metrics(estimator, my_rmse)\n # for batch prediction, you need a key associated with each instance\n estimator = forward_features(estimator, KEY_COLUMN)\n\n ## TODO 2c: Set the third argument of read_dataset to BATCH_SIZE \n ## TODO 2d: and set max_steps to TRAIN_STEPS\n train_spec = tf.estimator.TrainSpec(\n input_fn = read_dataset('train', tf.estimator.ModeKeys.TRAIN, None),\n max_steps = None)\n \n exporter = tf.estimator.LatestExporter('exporter', serving_input_fn, exports_to_keep=None)\n\n ## TODO 2e: Lastly, set steps equal to EVAL_STEPS\n eval_spec = tf.estimator.EvalSpec(\n input_fn = read_dataset('eval', tf.estimator.ModeKeys.EVAL, 2**15), # no need to batch in eval\n steps = None,\n start_delay_secs = 60, # start evaluating after N seconds\n throttle_secs = EVAL_INTERVAL, # evaluate every N seconds\n exporters = exporter)\n tf.estimator.train_and_evaluate(estimator, train_spec, eval_spec)", "_____no_output_____" ] ], [ [ "## Lab Task 3\n\nAfter moving the code to a package, make sure it works standalone. (Note the --pattern and --train_examples lines so that I am not trying to boil the ocean on the small notebook VM. Change as appropriate for your model).\n<p>\nEven with smaller data, this might take <b>3-5 minutes</b> in which you won't see any output ...", "_____no_output_____" ] ], [ [ "%%bash\necho \"bucket=${BUCKET}\"\nrm -rf babyweight_trained\nexport PYTHONPATH=${PYTHONPATH}:${PWD}/babyweight\npython -m trainer.task \\\n --bucket=${BUCKET} \\\n --output_dir=babyweight_trained \\\n --job-dir=./tmp \\\n --pattern=\"00000-of-\" --train_examples=1 --eval_steps=1", "_____no_output_____" ] ], [ [ "## Lab Task 4\n\nThe JSON below represents an input into your prediction model. Write the input.json file below with the next cell, then run the prediction locally to assess whether it produces predictions correctly.", "_____no_output_____" ] ], [ [ "%%writefile inputs.json\n{\"key\": \"b1\", \"is_male\": \"True\", \"mother_age\": 26.0, \"plurality\": \"Single(1)\", \"gestation_weeks\": 39}\n{\"key\": \"g1\", \"is_male\": \"False\", \"mother_age\": 26.0, \"plurality\": \"Single(1)\", \"gestation_weeks\": 39}", "_____no_output_____" ], [ "%%bash\nsudo find \"/usr/lib/google-cloud-sdk/lib/googlecloudsdk/command_lib/ml_engine\" -name '*.pyc' -delete", "_____no_output_____" ], [ "%%bash\nMODEL_LOCATION=$(ls -d $(pwd)/babyweight_trained/export/exporter/* | tail -1)\necho $MODEL_LOCATION\ngcloud ai-platform local predict --model-dir=$MODEL_LOCATION --json-instances=inputs.json", "_____no_output_____" ] ], [ [ "## Lab Task 5\n\nOnce the code works in standalone mode, you can run it on Cloud AI Platform. \nChange the parameters to the model (-train_examples for example may not be part of your model) appropriately.\n\nBecause this is on the entire dataset, it will take a while. The training run took about <b> 2 hours </b> for me. You can monitor the job from the GCP console in the Cloud AI Platform section.", "_____no_output_____" ] ], [ [ "%%bash\nOUTDIR=gs://${BUCKET}/babyweight/trained_model\nJOBNAME=babyweight_$(date -u +%y%m%d_%H%M%S)\necho $OUTDIR $REGION $JOBNAME\ngsutil -m rm -rf $OUTDIR\ngcloud ai-platform jobs submit training $JOBNAME \\\n --region=$REGION \\\n --module-name=trainer.task \\\n --package-path=$(pwd)/babyweight/trainer \\\n --job-dir=$OUTDIR \\\n --staging-bucket=gs://$BUCKET \\\n --scale-tier=STANDARD_1 \\\n --runtime-version=2.1 \\\n --python-version=3.7 \\\n -- \\\n --bucket=${BUCKET} \\\n --output_dir=${OUTDIR} \\\n --train_examples=20000", "_____no_output_____" ] ], [ [ "When I ran it, I used train_examples=20000. When training finished, I filtered in the Stackdriver log on the word \"dict\" and saw that the last line was:\n<pre>\nSaving dict for global step 5714290: average_loss = 1.06473, global_step = 5714290, loss = 34882.4, rmse = 1.03186\n</pre>\nThe final RMSE was 1.03 pounds.", "_____no_output_____" ], [ "<h2> Repeat training </h2>\n<p>\nThis time with tuned parameters (note last line)", "_____no_output_____" ] ], [ [ "%%bash\nOUTDIR=gs://${BUCKET}/babyweight/trained_model_tuned\nJOBNAME=babyweight_$(date -u +%y%m%d_%H%M%S)\necho $OUTDIR $REGION $JOBNAME\ngsutil -m rm -rf $OUTDIR\ngcloud ai-platform jobs submit training $JOBNAME \\\n --region=$REGION \\\n --module-name=trainer.task \\\n --package-path=$(pwd)/babyweight/trainer \\\n --job-dir=$OUTDIR \\\n --staging-bucket=gs://$BUCKET \\\n --scale-tier=STANDARD_1 \\\n --runtime-version=2.1 \\\n --python-version=3.7 \\\n -- \\\n --bucket=${BUCKET} \\\n --output_dir=${OUTDIR} \\\n --train_examples=2000 --batch_size=35 --nembeds=16 --nnsize=281", "_____no_output_____" ] ], [ [ "Copyright 2017 Google Inc. Licensed under the Apache License, Version 2.0 (the \"License\"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown" ]
[ [ "markdown" ], [ "code", "code", "code", "code", "code", "code", "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown" ] ]
ecea237a6672e9ea510bbcca0bec494cc7cb94e4
5,493
ipynb
Jupyter Notebook
STRINGS.ipynb
dejanBpetrovic/pythonOsnove
d1f67c148cde85ab7813b997f273794f2fc05ddb
[ "MIT" ]
null
null
null
STRINGS.ipynb
dejanBpetrovic/pythonOsnove
d1f67c148cde85ab7813b997f273794f2fc05ddb
[ "MIT" ]
null
null
null
STRINGS.ipynb
dejanBpetrovic/pythonOsnove
d1f67c148cde85ab7813b997f273794f2fc05ddb
[ "MIT" ]
null
null
null
18.31
592
0.458948
[ [ [ "string = \"123XYZ+-*/\"", "_____no_output_____" ], [ "print (string)", "123XYZ+-*/\n" ], [ "String = '123xZy+-*/'", "_____no_output_____" ], [ "print (String)", "123xZy+-*/\n" ], [ "string = \"OVO JE NOVI STRING\"", "_____no_output_____" ], [ "print (string)", "OVO JE NOVI STRING\n" ], [ "#korišćenje indeksa", "_____no_output_____" ], [ "a = \"ABCDEFG\"[6]", "_____no_output_____" ], [ "b = \"ABCDEFG\"[1]", "_____no_output_____" ], [ "c = \"ABCDEFG\"[3]", "_____no_output_____" ], [ "print (b+a+c)", "BGD\n" ], [ "d=\"ABCDEFG\"[0:4]", "_____no_output_____" ], [ "print (d)", "ABCD\n" ], [ "#citanje unazad putem negativnog indeksa\ne =\"ABCDEFG\"[-1]", "_____no_output_____" ], [ "print(e)", "G\n" ], [ "#funkcija str()\n\ns = \"broj\"\nss= 1\nsss = (s+ss)", "_____no_output_____" ], [ "s = \"broj\"\nss= str(1)\nsss = str((s+ss))", "_____no_output_____" ], [ "print (sss)", "broj1\n" ], [ "# funkcija input()", "_____no_output_____" ], [ "s = input(\"Unesi reč BROJ \")\nss = str (1) \nsss = (s+ss)\nprint (sss)", "Unesi reč BROJ BROJ\nBROJ1\n" ] ] ]
[ "code" ]
[ [ "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code" ] ]
ecea2a184313df03eaa2621a54328b4071e1fa08
28,229
ipynb
Jupyter Notebook
winter21/3_beatmap_features/1_beatmaps_analysis.ipynb
the-data-science-union/dsu-mlpp
fb1bd4022b20a74b90e98bc8798a8586fcf8f197
[ "MIT" ]
null
null
null
winter21/3_beatmap_features/1_beatmaps_analysis.ipynb
the-data-science-union/dsu-mlpp
fb1bd4022b20a74b90e98bc8798a8586fcf8f197
[ "MIT" ]
37
2021-01-27T19:10:52.000Z
2021-03-08T01:09:06.000Z
winter21/3_beatmap_features/1_beatmaps_analysis.ipynb
the-data-science-union/dsu-mlpp
fb1bd4022b20a74b90e98bc8798a8586fcf8f197
[ "MIT" ]
2
2021-03-04T02:02:40.000Z
2021-04-21T04:55:41.000Z
26.936069
1,689
0.53027
[ [ [ "## Notebook Objective: Analyze the Most Popular Beatmaps and their Attributes", "_____no_output_____" ] ], [ [ "import sys\nsys.path.append('..')\nfrom pymongo import UpdateOne\nimport pandas as pd\nimport matplotlib.pyplot as plt\nfrom pymongo import MongoClient\nimport seaborn as sns\nfrom mlpp.data_collection.sample import ScoresSubset, get_more_recent_than", "_____no_output_____" ], [ "client = MongoClient('localhost', 27017, username='admin', password=\"dsumlppftw\")\ndb = client.osu_random_db\ndata = client[\"osu_random_db\"]\n\nosu_subset = ScoresSubset(data['osu_scores_high'], data['osu_user_stats'])", "_____no_output_____" ], [ "new_subset, user_ids = osu_subset.init_random_sample(data['sample_scores_1M'], data['sample_users_1M'], 1000)\n#creating a new collection of 1 million scores", "_____no_output_____" ], [ "collection = data[\"sample_scores_1M\"]\n#got 500k sample scores more recent than 12/1/2018 (last 2 years)\n#get_more_recent_than function is in the file Sample.py", "_____no_output_____" ], [ "collection2 = data[\"sample_scores_3k\"]", "_____no_output_____" ], [ "\"\"\"\nmax est user pp in collection of 500k\n\"\"\"\ncursor=db.scores_sample_3k.aggregate(\n [\n {\n \"$group\":\n {\n \"_id\": {},\n \"max\": { \"$max\": \"$mlpp.est_user_pp\" }\n }\n }\n ]\n)\nfor document in cursor:\n print(document)\nprint(document['max'])\nmax_pp = document['max']", "{'_id': {}, 'max': 8858.10306730833}\n8858.10306730833\n" ] ], [ [ "## Objective 1: Create a \"Uniform\" Collection ", "_____no_output_____" ] ], [ [ "a = 0\nb = 100\n\nwhile b <= max_pp:\n db.uniform_collection2.insert_many(\n collection2.aggregate([\n {\n '$match': {\n 'mlpp.est_user_pp' : {\n '$gt': a,\n '$lt': b,\n }\n }\n },\n {'$sample': {\n 'size': 800\n }\n}\n \n])\n )\n a = b\n b += 100\n# creates a new \"uniform\" collection", "_____no_output_____" ] ], [ [ "## Objective 2: Create a Collection of the 1000 Most Popular Beatmaps", "_____no_output_____" ] ], [ [ "#once you have a uniform collection, use this to have a collection with the most 1000 popular maps:\ndb.uniform_collection2.aggregate([\n {\n '$group': {\n '_id': '$beatmap_id', \n 'count': {\n '$sum': 1\n }\n }\n }, {\n '$sort': {\n 'count': -1\n }\n }, {\n '$limit': 1000\n }, {\n '$out': 'oneThousand_most_popular_maps2'\n }\n])", "_____no_output_____" ] ], [ [ "## Objective 3: Relationship as Score Count Decreases", "_____no_output_____" ] ], [ [ "g=[]\nh=[]\nx=db.oneThousand_most_popular_maps2\nx1=x.find({},{ \"_id\": 0, \"count\": 1}) #finds only the count column in the collection\ny1=x.find({},{ \"_id\": 1, \"count\": 0})\n\n# for i in x.find({},{ \"_id\": 0, \"count\": 1}):\n# print(i)\n\nfor i in x1:\n g.append(i['count'])\nfor i in x.find({},{ \"_id\": 1, \"count\": 0}):\n h.append(i['_id'])\n\nplt.plot(g,'ro') #automatically index x \nplt.xlabel('index')\nplt.ylabel('count')\nplt.title('Relationship as Score Count Decreases')\n\n#Score count seems to decrease exponentially", "_____no_output_____" ] ], [ [ "## Objective 4 & 6: Feature Distribution Analysis", "_____no_output_____" ] ], [ [ "collection = data[\"osu_beatmap_attribs\"]", "_____no_output_____" ], [ "db.osu_beatmaps_attribs_modZero.insert_many(\n collection.aggregate([\n {\n '$match': {\n 'mods': 0\n }\n }\n]))\n\n#creating a collection from the beatmap_attrib with only documents containing mod zero", "_____no_output_____" ], [ "collection = data['oneThousand_most_popular_maps2']", "_____no_output_____" ], [ "cursor = collection.find({},{\"_id\":1})\nl = []\nfor el in cursor:\n l.append(el)\nlistOfIds = []\nfor i in range(1000):\n listOfIds.append(l[i]['_id'])\n#list of id's among the 1k most popular beatmaps", "_____no_output_____" ], [ "collection2 = data['osu_beatmap_attribs']", "_____no_output_____" ], [ "for _id in listOfIds:\n db.oneThousand_Beatmaps_attribs.insert_many(\n collection2.aggregate([\n {'$match' : {\n \"_id\" : _id }\n }\n ] ))\n#from the beatmap attrib collection, creating a new collection with beatmap attrib for the 1k most popular maps", "_____no_output_____" ], [ "collection = data[\"osu_beatmaps_attribs_modZero\"]", "_____no_output_____" ] ], [ [ "Distribution for attribute 5", "_____no_output_____" ] ], [ [ "db.attrib_5.insert_many(\n collection.aggregate([\n {'$match':{\n \"beatmap_id\": {\n '$in': listOfIds\n },\n \"attrib_id\": 5,\n } \n }\n])\n )", "_____no_output_____" ], [ "c1=[]\nd=[]\nt=db.attrib_5\nx=t.find({},{ \"_id\": 1, \"count\": 0})\ny=t.find({},{ \"_id\": 0, \"value\": 1})\n\nfor i in y:\n d.append(i['value'])\nfor i in x:\n c1.append(i[\"beatmap_id\"])\n\nplt.hist(d)\nplt.show()", "_____no_output_____" ] ], [ [ "Distribution for attribute 17", "_____no_output_____" ] ], [ [ "db.attrib_17.insert_many(\n collection.aggregate([\n {'$match':{\n \"beatmap_id\": {\n '$in': listOfIds\n },\n \"attrib_id\": 17,\n } \n }\n])\n )", "_____no_output_____" ], [ "j=[]\nk=[]\nt=db.attrib_17\nx=t.find({},{ \"_id\": 1, \"count\": 0})\ny=t.find({},{ \"_id\": 0, \"value\": 1})\n\nfor i in y:\n k.append(i['value'])\nfor i in x:\n j.append(i[\"beatmap_id\"])\n \nk\nplt.hist(k)\nplt.show()", "_____no_output_____" ] ], [ [ "Distribution for attribute 1", "_____no_output_____" ] ], [ [ "db.attrib_1.insert_many(\n collection.aggregate([\n {'$match':{\n \"beatmap_id\": {\n '$in': listOfIds\n },\n \"attrib_id\": 1,\n } \n }\n])\n )", "_____no_output_____" ], [ "j=[]\nk=[]\nt=db.attrib_1\nx=t.find({},{ \"_id\": 1, \"count\": 0})\ny=t.find({},{ \"_id\": 0, \"value\": 1})\n\nfor i in y:\n k.append(i['value'])\nfor i in x:\n j.append(i[\"beatmap_id\"])\n \nk\nplt.hist(k)\nplt.show()", "_____no_output_____" ] ], [ [ "Distribution for attribute 3", "_____no_output_____" ] ], [ [ "db.attrib_3.insert_many(\n collection.aggregate([\n {'$match':{\n \"beatmap_id\": {\n '$in': listOfIds\n },\n \"attrib_id\": 3,\n } \n }\n])\n )", "_____no_output_____" ], [ "j=[]\nk=[]\nt=db.attrib_3\nx=t.find({},{ \"_id\": 1, \"count\": 0})\ny=t.find({},{ \"_id\": 0, \"value\": 1})\n\nfor i in y:\n k.append(i['value'])\nfor i in x:\n j.append(i[\"beatmap_id\"])\n \nk\nplt.hist(k)\nplt.show()", "_____no_output_____" ] ], [ [ "Distribution for Attribute 7", "_____no_output_____" ] ], [ [ "db.attrib_7.insert_many(\n collection.aggregate([\n {'$match':{\n \"beatmap_id\": {\n '$in': listOfIds\n },\n \"attrib_id\": 7,\n } \n }\n])\n )", "_____no_output_____" ], [ "j=[]\nk=[]\nt=db.attrib_7\nx=t.find({},{ \"_id\": 1, \"count\": 0})\ny=t.find({},{ \"_id\": 0, \"value\": 1})\n\nfor i in y:\n k.append(i['value'])\nfor i in x:\n j.append(i[\"beatmap_id\"])\n \nk\nplt.hist(k)\nplt.show()", "_____no_output_____" ] ], [ [ "Distribution for attribute 9", "_____no_output_____" ] ], [ [ "db.attrib_9.insert_many(\n collection.aggregate([\n {'$match':{\n \"beatmap_id\": {\n '$in': listOfIds\n },\n \"attrib_id\": 9,\n } \n }\n])\n )", "_____no_output_____" ], [ "j=[]\nk=[]\nt=db.attrib_9\nx=t.find({},{ \"_id\": 1, \"count\": 0})\ny=t.find({},{ \"_id\": 0, \"value\": 1})\n\nfor i in y:\n k.append(i['value'])\nfor i in x:\n j.append(i[\"beatmap_id\"])\n \nk\nplt.hist(k)\nplt.show()", "_____no_output_____" ] ], [ [ "Distribution for attribute 11", "_____no_output_____" ] ], [ [ "db.attrib_11.insert_many(\n collection.aggregate([\n {'$match':{\n \"beatmap_id\": {\n '$in': listOfIds\n },\n \"attrib_id\": 11,\n } \n }\n])\n )", "_____no_output_____" ], [ "j=[]\nk=[]\nt=db.attrib_11\nx=t.find({},{ \"_id\": 1, \"count\": 0})\ny=t.find({},{ \"_id\": 0, \"value\": 1})\n\nfor i in y:\n k.append(i['value'])\nfor i in x:\n j.append(i[\"beatmap_id\"])\n \nk\nplt.hist(k)\nplt.show()", "_____no_output_____" ] ], [ [ "## Objective 5: Star/OD Correlation with Popularity of Beatmap", "_____no_output_____" ] ], [ [ "df = pd.DataFrame(list(db.oneThousand_most_popular_maps2.find({})))\ndf.sort_values([\"_id\"], inplace = True)\ndf.reset_index(inplace = True)\na = df[\"count\"]", "_____no_output_____" ], [ "df1 = pd.DataFrame()\ndf1[\"beatmap_id\"] = j\ndf1['Star Difficulty'] = k\ndf1[\"count\"] = a\ndf1.sort_values(by = [\"count\"], axis = 0, ascending = False, inplace = True)\nc = df1[\"count\"]\nsd = df1[\"Star Difficulty\"]\ndf1", "_____no_output_____" ], [ "plt.scatter(c, sd)", "_____no_output_____" ], [ "df2 = pd.DataFrame()\ndf2[\"beatmap_id\"] = c1\ndf2['OD'] = d\ndf2[\"count\"] = a\ndf2\n\n#recall d is the list of OD values ", "_____no_output_____" ], [ "c2 = df2[\"count\"]\nod = df2[\"OD\"]", "_____no_output_____" ], [ "plt.scatter(c2, od)", "_____no_output_____" ] ], [ [ "## Objective 7: Heatmap of Correlation Between Attributes", "_____no_output_____" ] ], [ [ "df3 = pd.DataFrame(list(db.osu_beatmaps_attribs_modZero.find({})))\ndf3.drop([\"_id\",\"mods\"], axis = 1, inplace = True)", "_____no_output_____" ], [ "df3.head(20)", "_____no_output_____" ], [ "col = df3['attrib_id'].unique()\nind = df3['beatmap_id'].unique()\nDF = pd.DataFrame(columns=col, index=ind)\n\n\ng=df3.groupby(['beatmap_id', 'attrib_id'])\n\nfor name, group in g:\n bmap = name[0]\n attr = name[1]\n val = float(group['value'])\n DF.at[bmap, attr] = val\n\n\nDF.head()", "_____no_output_____" ], [ "DF.columns = [\"Aim\", \"Speed\", \"OD\", \"AR\", \"Max_Combo\", \"Strain\", \"Star Difficulty\"]", "_____no_output_____" ], [ "DF.reset_index()", "_____no_output_____" ], [ "correlation = DF.astype('float64').corr()\nf, ax = plt.subplots(figsize = (14, 12))\nplt.title(\"Correlation of Attributes\")\nsns.heatmap(correlation, annot = True)\nplt.show()\n\n", "_____no_output_____" ], [ "# collections = db.list_collection_names() \n# print (\"collections:\", collections, \"\\n\")\n\n#All of the collection in the database", "_____no_output_____" ] ], [ [ "## Objective 8: Conclusion", "_____no_output_____" ] ], [ [ "# Attribute AR is left skewed\n# Attrib Max Combo is right skewed\n# Aim and Star Difficulty, Strain and Star Dificulty are highly correlated \n# Speed and OD, AR and OD are highly correlated \n# Players tend ", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code" ]
[ [ "markdown" ], [ "code", "code", "code", "code", "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code", "code", "code" ], [ "markdown" ], [ "code" ] ]
ecea2d1de6c6459aed5dd91347d7cf4b819052f9
95,457
ipynb
Jupyter Notebook
persistence-example.ipynb
scotthellman/discrete-topology
6182fe607868d88c462c185be8629a35ad2d7c37
[ "MIT" ]
null
null
null
persistence-example.ipynb
scotthellman/discrete-topology
6182fe607868d88c462c185be8629a35ad2d7c37
[ "MIT" ]
null
null
null
persistence-example.ipynb
scotthellman/discrete-topology
6182fe607868d88c462c185be8629a35ad2d7c37
[ "MIT" ]
null
null
null
667.531469
30,768
0.938904
[ [ [ "import numpy as np\nimport matplotlib.pyplot as plt\nimport networkx as nx\nimport graph\nimport morsesmale\nimport scipy.spatial\n%matplotlib inline\n\nfunc = lambda x : x**3 + x**2 - 1*x\npoints = np.linspace(-2,2,50)\nfunc_vals = [func(x) for x in points]\nprint(points)\nprint(func_vals)", "[-2. -1.91836735 -1.83673469 -1.75510204 -1.67346939 -1.59183673\n -1.51020408 -1.42857143 -1.34693878 -1.26530612 -1.18367347 -1.10204082\n -1.02040816 -0.93877551 -0.85714286 -0.7755102 -0.69387755 -0.6122449\n -0.53061224 -0.44897959 -0.36734694 -0.28571429 -0.20408163 -0.12244898\n -0.04081633 0.04081633 0.12244898 0.20408163 0.28571429 0.36734694\n 0.44897959 0.53061224 0.6122449 0.69387755 0.7755102 0.85714286\n 0.93877551 1.02040816 1.10204082 1.18367347 1.26530612 1.34693878\n 1.42857143 1.51020408 1.59183673 1.67346939 1.75510204 1.83673469\n 1.91836735 2. ]\n[-2.0, -1.4613468877763509, -0.98606872986595717, -0.57090158012392789, -0.21258149240537572, 0.092155479434588905, 0.34657328154085487, 0.55393586005830886, 0.7175071611318411, 0.84055113090633982, 0.92633171552669391, 0.97811286113779139, 0.99915851388452093, 0.99273261991177153, 0.96209912536443154, 0.91052197638738974, 0.84126511912553448, 0.75759249972375453, 0.6627680643269388, 0.56005575907997529, 0.45271953012775312, 0.34402332361516047, 0.23723108568708645, 0.1356067624884191, 0.04241430016404734, -0.03908235514114021, -0.10561925728225463, -0.153932460114408, -0.18075801749271136, -0.18283198327227601, -0.15689041130821363, -0.099669355455635267, -0.0079048695696524174, 0.12166699249462365, 0.2923101768820815, 0.50728862973760824, 0.76986629720609501, 1.0833071254324294, 1.4508750605615002, 1.8758340487381933, 2.3614480361074017, 2.9109809688140129, 3.5276967930029142, 4.2148594548189955, 4.9757329004071389, 5.8135810759122437, 6.7316679274791928, 7.7332574012528763, 8.8216134433781832, 10.0]\n" ], [ "pairs = scipy.spatial.distance.pdist(points.reshape(len(points), 1))\npdist = scipy.spatial.distance.squareform(pairs)\nG = graph.generate_knn_graph(pdist, 2)\nmorsesmale.find_extrema(G, pdist, func_vals)\nfiltrations = morsesmale.get_filtrations(pdist, func_vals, 2)", "_____no_output_____" ], [ "colors = ['r','b','k']\ncolor_map = {crystal:colors[i] for i,crystal in enumerate(filtrations[0])}\nfor filtration in filtrations:\n reverse_lookup = {}\n for crystal,vertices in filtration.items():\n for v in vertices:\n reverse_lookup[v] = crystal\n c = [color_map[reverse_lookup[v]] for v in range(len(points))]\n plt.scatter(points, func_vals, c=c, s=200)\n plt.gcf().set_size_inches(12,8)\n plt.show()", "_____no_output_____" ] ] ]
[ "code" ]
[ [ "code", "code", "code" ] ]
ecea3458292a22164b0bfd44d8558e8385e6541c
232,495
ipynb
Jupyter Notebook
notebooks/2020_03_17_Replay_Distance_on_Linear_Track.ipynb
edeno/pose_analysis
eb3e435e01f3a084a32eaefaab85738e03522155
[ "MIT" ]
1
2020-08-08T17:07:41.000Z
2020-08-08T17:07:41.000Z
notebooks/2020_03_17_Replay_Distance_on_Linear_Track.ipynb
edeno/pose_analysis
eb3e435e01f3a084a32eaefaab85738e03522155
[ "MIT" ]
null
null
null
notebooks/2020_03_17_Replay_Distance_on_Linear_Track.ipynb
edeno/pose_analysis
eb3e435e01f3a084a32eaefaab85738e03522155
[ "MIT" ]
null
null
null
384.925497
153,844
0.911929
[ [ [ "%matplotlib inline\n%reload_ext autoreload\n%autoreload 2\n%config InlineBackend.figure_format = 'retina'", "_____no_output_____" ], [ "import matplotlib.pyplot as plt\nimport numpy as np\nimport xarray as xr\nimport seaborn as sns\nimport logging\n\nFORMAT = '%(asctime)s %(message)s'\n\nlogging.basicConfig(level='INFO', format=FORMAT, datefmt='%d-%b-%y %H:%M:%S')\nsns.set_context(\"talk\")", "_____no_output_____" ], [ "from src.load_data import load_data\n\nepoch_key = ('Jaq', 1, 4) # animal, day, epoch\n\ndata = load_data(epoch_key)", "_____no_output_____" ], [ "fig, ax = plt.subplots(figsize=(30, 10))\n\nfor edge_label, df in data['position_info'].groupby('track_segment_id'):\n ax.scatter(df.index / np.timedelta64(1, 's'), df.linear_position, s=1)\n \nax.set_ylabel('Position [cm]')\nax.set_xlabel('Time [s]');", "_____no_output_____" ], [ "from src.load_data import make_track_graph\nfrom src.parameters import ANIMALS\n\ntrack_graph, center_well_id = make_track_graph(epoch_key, ANIMALS)", "_____no_output_____" ], [ "from loren_frank_data_processing.track_segment_classification import plot_track\n\nplt.plot(data['position_info'].tail_x, data['position_info'].tail_y, color=\"lightgrey\", alpha=0.7, zorder=-1)\nplot_track(track_graph)\nsns.despine(left=True, bottom=True)\n\ncenter_well_id", "_____no_output_____" ], [ "is_running = np.abs(data[\"position_info\"].tail_vel) > 4\nEDGE_ORDER = [0]\nEDGE_SPACING = 0", "_____no_output_____" ], [ "from replay_trajectory_classification import ClusterlessClassifier\nfrom src.parameters import classifier_parameters, discrete_state_transition\n\nfrom sklearn.model_selection import KFold\nfrom tqdm.auto import tqdm\n\ncv = KFold()\ncv_classifier_clusterless_results = []\n\nfor fold_ind, (train, test) in enumerate(cv.split(data[\"position_info\"].index)):\n logging.info(f\"Fold #{fold_ind}\")\n cv_classifier = ClusterlessClassifier(**classifier_parameters)\n\n cv_classifier.fit(\n position=data[\"position_info\"].iloc[train].linear_position,\n multiunits=data[\"multiunits\"].isel(time=train),\n is_training=is_running.iloc[train],\n track_graph=track_graph,\n center_well_id=center_well_id,\n edge_order=EDGE_ORDER,\n edge_spacing=EDGE_SPACING,\n )\n cv_classifier.discrete_state_transition_ = discrete_state_transition\n logging.info('Predicting posterior...')\n cv_classifier_clusterless_results.append(\n cv_classifier.predict(\n data[\"multiunits\"].isel(time=test),\n time=data[\"position_info\"].iloc[test].index / np.timedelta64(1, \"s\"),\n )\n )\nlogging.info(\"Done...\")", "17-Mar-20 11:28:02 Fold #0\n17-Mar-20 11:28:02 Fitting initial conditions...\n17-Mar-20 11:28:03 Fitting state transition...\n17-Mar-20 11:28:03 Fitting multiunits...\n17-Mar-20 11:28:04 Predicting posterior...\n17-Mar-20 11:38:17 Fold #1\n17-Mar-20 11:38:17 Fitting initial conditions...\n17-Mar-20 11:38:17 Fitting state transition...\n17-Mar-20 11:38:17 Fitting multiunits...\n17-Mar-20 11:38:18 Predicting posterior...\n17-Mar-20 11:47:52 Fold #2\n17-Mar-20 11:47:52 Fitting initial conditions...\n17-Mar-20 11:47:52 Fitting state transition...\n17-Mar-20 11:47:52 Fitting multiunits...\n17-Mar-20 11:47:53 Predicting posterior...\n17-Mar-20 11:57:16 Fold #3\n17-Mar-20 11:57:16 Fitting initial conditions...\n17-Mar-20 11:57:16 Fitting state transition...\n17-Mar-20 11:57:16 Fitting multiunits...\n17-Mar-20 11:57:16 Predicting posterior...\n17-Mar-20 12:06:37 Fold #4\n17-Mar-20 12:06:37 Fitting initial conditions...\n17-Mar-20 12:06:37 Fitting state transition...\n17-Mar-20 12:06:38 Fitting multiunits...\n17-Mar-20 12:06:38 Predicting posterior...\n17-Mar-20 12:15:56 Done...\n" ], [ "# concatenate cv classifier results \ncv_classifier_clusterless_results = xr.concat(\n cv_classifier_clusterless_results, dim=\"time\"\n)\ncv_classifier_clusterless_results", "_____no_output_____" ], [ "# save the results as .nc format. ncread matlab can read these\ncv_classifier_clusterless_results.to_netcdf(\n f\"{epoch_key[0]}_{epoch_key[1]:02d}_{epoch_key[2]:02d}_cv_classifier_clusterless_vel_4_tail_alltime_results.nc\"\n)", "_____no_output_____" ], [ "from src.analysis import calculate_replay_distance\n\ncalculate_replay_distance(cv_classifier_clusterless_results.isel(time=slice(0, 1000)).acausal_posterior.sum(\"state\"),\n track_graph,\n cv_classifier,\n data[\"position_info\"].iloc[slice(0, 1000)].loc[:, [\"tail_x\", \"tail_y\"]],\n data[\"position_info\"].iloc[slice(0, 1000)].track_segment_id\n )", "_____no_output_____" ] ] ]
[ "code" ]
[ [ "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code" ] ]
ecea39b50b645e3360cefffcc61cb9525bfef9ee
1,837
ipynb
Jupyter Notebook
notebooks/test_shapefile.ipynb
DFS-UCU/FoodSecurity
eba0b543635cd62482eb2846d63f489881630317
[ "MIT" ]
17
2017-11-27T19:58:38.000Z
2022-02-07T01:13:44.000Z
notebooks/test_shapefile.ipynb
DFS-UCU/FoodSecurity
eba0b543635cd62482eb2846d63f489881630317
[ "MIT" ]
null
null
null
notebooks/test_shapefile.ipynb
DFS-UCU/FoodSecurity
eba0b543635cd62482eb2846d63f489881630317
[ "MIT" ]
6
2019-08-21T11:42:08.000Z
2021-03-26T09:59:40.000Z
1,837
1,837
0.606968
[ [ [ "# pip install pyshp \nimport shapefile\nimport matplotlib.pyplot as plt", "_____no_output_____" ], [ "#make sure you have file \"data/yield/2017_wheat.shp\"\nsf = shapefile.Reader(\"../data/yield/2017_wheat.shp\")", "_____no_output_____" ], [ "shapes = sf.shapes()\nlen(shapes)", "_____no_output_____" ], [ "bbox = shapes[3].bbox\n['%.3f' % coord for coord in bbox]", "_____no_output_____" ], [ "len(shapes[3].points)", "_____no_output_____" ], [ "plt.figure()\nfor shape in sf.shapeRecords():\n x = [i[0] for i in shape.shape.points[:]]\n y = [i[1] for i in shape.shape.points[:]]\n plt.plot(x,y)\nplt.show()", "_____no_output_____" ], [ "sf.fields", "_____no_output_____" ], [ "records = sf.records()\nrecords[3]", "_____no_output_____" ], [ "help(sf)", "_____no_output_____" ] ] ]
[ "code" ]
[ [ "code", "code", "code", "code", "code", "code", "code", "code", "code" ] ]
ecea40faf0abcd7ec55daf5fd129ecac1e267aae
20,338
ipynb
Jupyter Notebook
Samples/K_Fold_test.ipynb
davidfreire/KFold_project
4a69f264bec6f7e7ecc090ee2d1457453fb3e94c
[ "MIT" ]
null
null
null
Samples/K_Fold_test.ipynb
davidfreire/KFold_project
4a69f264bec6f7e7ecc090ee2d1457453fb3e94c
[ "MIT" ]
null
null
null
Samples/K_Fold_test.ipynb
davidfreire/KFold_project
4a69f264bec6f7e7ecc090ee2d1457453fb3e94c
[ "MIT" ]
1
2019-07-04T13:04:23.000Z
2019-07-04T13:04:23.000Z
37.873371
132
0.52011
[ [ [ "Interesting lecture: https://machinelearningmastery.com/k-fold-cross-validation/", "_____no_output_____" ] ], [ [ "import sys\nsys.path.insert(0, '../') #to load KFold", "_____no_output_____" ], [ "from keras import Input, optimizers\nfrom keras.layers import Conv2D, MaxPool2D, Flatten, Dense, Dropout\nfrom keras.models import Model\nfrom KFold import K_Fold", "Using TensorFlow backend.\n" ] ], [ [ "### First, it is important to understand differences between ShuffleSplit and KFold", "_____no_output_____" ] ], [ [ "from sklearn.model_selection import StratifiedShuffleSplit, StratifiedKFold\n\nsplits = 5\n\ntx = range(10)\nty = [0] * 5 + [1] * 5\n\nkfold = StratifiedKFold(n_splits=splits, shuffle=True, random_state=42)\nshufflesplit = StratifiedShuffleSplit(n_splits=splits, random_state=42, test_size=2)\n\nprint(\"KFold\")\nfor train_index, test_index in kfold.split(tx, ty):\n print(\"TRAIN:\", train_index, \"TEST:\", test_index)\n\nprint(\"Shuffle Split\")\nfor train_index, test_index in shufflesplit.split(tx, ty):\n print(\"TRAIN:\", train_index, \"TEST:\", test_index)", "KFold\nTRAIN: [0 2 3 4 5 7 8 9] TEST: [1 6]\nTRAIN: [0 1 2 3 5 6 7 8] TEST: [4 9]\nTRAIN: [0 1 3 4 5 6 8 9] TEST: [2 7]\nTRAIN: [1 2 3 4 6 7 8 9] TEST: [0 5]\nTRAIN: [0 1 2 4 5 6 7 9] TEST: [3 8]\nShuffle Split\nTRAIN: [8 4 1 0 6 5 7 2] TEST: [3 9]\nTRAIN: [7 0 3 9 4 5 1 6] TEST: [8 2]\nTRAIN: [1 2 5 6 4 8 9 0] TEST: [3 7]\nTRAIN: [4 6 7 8 3 5 1 2] TEST: [9 0]\nTRAIN: [7 2 6 5 4 3 0 9] TEST: [1 8]\n" ], [ "# In KFolds, each test set should not overlap, even with shuffle. \n# With KFolds and shuffle, the data is shuffled once at the start, and then divided into the number of desired splits. \n# The test data is always one of the splits, the train data is the rest.\n\n# In ShuffleSplit, the data is shuffled every time, and then split. \n# This means the test sets may overlap between the splits:\n# Test, first row 3 and third row 3, first row 9 and fourth row 9.", "_____no_output_____" ], [ "# Thus, in ShuffleSplit test_size can be specified, for instance .2 means 1-.2 = .8 for training.\n# While, KFolds depends on the size of the data and K (test set should not overlap), thus, if data is len 10, \n# and K is 5, then test size is 10/5=2 in order to never overlap.", "_____no_output_____" ] ], [ [ "### Second, let's create the K-Fold cross validation", "_____no_output_____" ] ], [ [ "DB_Path = '/floyd/input/db1/small_dataset/train'", "_____no_output_____" ] ], [ [ "The general procedure is as follows:\n\nShuffle the dataset randomly. ok\nSplit the dataset into k groups ok\nFor each unique group:\nTake the group as a hold out or test data set\nTake the remaining groups as a training data set\nFit a model on the training set and evaluate it on the test set\nRetain the evaluation score and discard the model\nSummarize the skill of the model using the sample of model evaluation scores", "_____no_output_____" ] ], [ [ "#To save history dict\nimport pickle \ndef save_obj(obj, name):\n with open(name, 'wb') as f:\n pickle.dump(obj, f, pickle.HIGHEST_PROTOCOL)\n\ndef load_obj(name):\n with open(name, 'rb') as f:\n return pickle.load(f)", "_____no_output_____" ], [ "def get_model():\n \n entrada= Input(shape=(150,150,3))\n \n conv = Conv2D(filters=32, kernel_size=3, activation='relu', name='conv_1')(entrada)\n maxpool = MaxPool2D(pool_size=2, strides=2, name='maxpool_1')(conv)\n \n conv = Conv2D(filters=64, kernel_size=3, activation='relu', name='conv_2')(maxpool)\n maxpool = MaxPool2D(pool_size=2, strides=2, name='maxpool_2')(conv) \n \n conv = Conv2D(filters=128, kernel_size=3, activation='relu', name='conv_3')(maxpool)\n maxpool = MaxPool2D(pool_size=2, strides=2, name='maxpool_3')(conv)\n \n conv = Conv2D(filters=128, kernel_size=3, activation='relu', name='conv_4')(maxpool)\n maxpool = MaxPool2D(pool_size=2, strides=2, name='maxpool_4')(conv)\n \n flat = Flatten(name='flatten')(maxpool)\n #drop = Dropout(rate=.5, name='dropout')(flat)\n \n dense = Dense(units=512, activation='relu', name='Dense1')(flat)#(drop)\n output = Dense(units=1, activation='sigmoid', name='output')(dense)\n #output = Dense(units=2, activation='softmax', name='output')(dense)\n \n model = Model(entrada, output)\n \n #model.compile(optimizer=optimizers.RMSprop(lr=0.001), loss='categorical_crossentropy', metrics=['acc'])\n model.compile(optimizer=optimizers.RMSprop(lr=0.001), loss='binary_crossentropy', metrics=['acc'])\n \n return model", "_____no_output_____" ], [ "#Generator parameters\ntraingen_params = {\n 'rescale': 1./255\n}\n\ntestgen_params = {\n 'rescale': 1./255\n}\n\n# Train parameters\ntrain_params = {\n 'batch_size': 20,\n 'target_size': (150,150),\n 'class_mode': 'binary', #'categorical',\n}\n\ntest_params = {\n 'batch_size': 20,\n 'target_size': (150,150),\n 'class_mode': 'binary', #'categorical',\n}\n\nfit_params = {\n 'epochs':10,\n 'shuffle':True,\n 'verbose':1\n}", "_____no_output_____" ], [ "KF = K_Fold(DB_Path, 4)", "_____no_output_____" ], [ "model = get_model()", "_____no_output_____" ], [ "KF.Check_Folds()", "There are 4 Folds\n\nFold 0\nFor training, 1500 samples: {'cats': 750, 'dogs': 750}\nFor testing, 500 samples: {'cats': 250, 'dogs': 250}\nFirst five X_train images: \n['/floyd/input/db1/small_dataset/train/cats/cat.932.jpg'\n '/floyd/input/db1/small_dataset/train/cats/cat.901.jpg'\n '/floyd/input/db1/small_dataset/train/cats/cat.922.jpg'\n '/floyd/input/db1/small_dataset/train/cats/cat.974.jpg'\n '/floyd/input/db1/small_dataset/train/cats/cat.492.jpg']\nFirst five X_val images: \n['/floyd/input/db1/small_dataset/train/cats/cat.391.jpg'\n '/floyd/input/db1/small_dataset/train/cats/cat.980.jpg'\n '/floyd/input/db1/small_dataset/train/cats/cat.524.jpg'\n '/floyd/input/db1/small_dataset/train/cats/cat.571.jpg'\n '/floyd/input/db1/small_dataset/train/cats/cat.699.jpg']\n\nFold 1\nFor training, 1500 samples: {'cats': 750, 'dogs': 750}\nFor testing, 500 samples: {'cats': 250, 'dogs': 250}\nFirst five X_train images: \n['/floyd/input/db1/small_dataset/train/cats/cat.932.jpg'\n '/floyd/input/db1/small_dataset/train/cats/cat.901.jpg'\n '/floyd/input/db1/small_dataset/train/cats/cat.922.jpg'\n '/floyd/input/db1/small_dataset/train/cats/cat.974.jpg'\n '/floyd/input/db1/small_dataset/train/cats/cat.492.jpg']\nFirst five X_val images: \n['/floyd/input/db1/small_dataset/train/cats/cat.621.jpg'\n '/floyd/input/db1/small_dataset/train/cats/cat.416.jpg'\n '/floyd/input/db1/small_dataset/train/cats/cat.591.jpg'\n '/floyd/input/db1/small_dataset/train/cats/cat.690.jpg'\n '/floyd/input/db1/small_dataset/train/cats/cat.577.jpg']\n\nFold 2\nFor training, 1500 samples: {'cats': 750, 'dogs': 750}\nFor testing, 500 samples: {'cats': 250, 'dogs': 250}\nFirst five X_train images: \n['/floyd/input/db1/small_dataset/train/cats/cat.901.jpg'\n '/floyd/input/db1/small_dataset/train/cats/cat.974.jpg'\n '/floyd/input/db1/small_dataset/train/cats/cat.492.jpg'\n '/floyd/input/db1/small_dataset/train/cats/cat.391.jpg'\n '/floyd/input/db1/small_dataset/train/cats/cat.621.jpg']\nFirst five X_val images: \n['/floyd/input/db1/small_dataset/train/cats/cat.932.jpg'\n '/floyd/input/db1/small_dataset/train/cats/cat.922.jpg'\n '/floyd/input/db1/small_dataset/train/cats/cat.326.jpg'\n '/floyd/input/db1/small_dataset/train/cats/cat.348.jpg'\n '/floyd/input/db1/small_dataset/train/cats/cat.453.jpg']\n\nFold 3\nFor training, 1500 samples: {'cats': 750, 'dogs': 750}\nFor testing, 500 samples: {'cats': 250, 'dogs': 250}\nFirst five X_train images: \n['/floyd/input/db1/small_dataset/train/cats/cat.932.jpg'\n '/floyd/input/db1/small_dataset/train/cats/cat.922.jpg'\n '/floyd/input/db1/small_dataset/train/cats/cat.326.jpg'\n '/floyd/input/db1/small_dataset/train/cats/cat.348.jpg'\n '/floyd/input/db1/small_dataset/train/cats/cat.391.jpg']\nFirst five X_val images: \n['/floyd/input/db1/small_dataset/train/cats/cat.901.jpg'\n '/floyd/input/db1/small_dataset/train/cats/cat.974.jpg'\n '/floyd/input/db1/small_dataset/train/cats/cat.492.jpg'\n '/floyd/input/db1/small_dataset/train/cats/cat.279.jpg'\n '/floyd/input/db1/small_dataset/train/cats/cat.724.jpg']\n" ], [ "hist = KF.Apply_KFold(model, traingen_params, testgen_params, train_params, test_params, fit_params)", "\nFold 0\nFound 1500 images belonging to 2 classes.\nFound 500 images belonging to 2 classes.\nTraining\nEpoch 1/10\n75/75 [==============================] - 12s 154ms/step - loss: 0.7573 - acc: 0.4993 - val_loss: 0.6928 - val_acc: 0.5000\nEpoch 2/10\n75/75 [==============================] - 10s 136ms/step - loss: 0.6962 - acc: 0.5240 - val_loss: 0.6901 - val_acc: 0.6120\nEpoch 3/10\n75/75 [==============================] - 10s 136ms/step - loss: 0.6957 - acc: 0.6060 - val_loss: 0.6578 - val_acc: 0.6340\nEpoch 4/10\n75/75 [==============================] - 10s 135ms/step - loss: 0.6377 - acc: 0.6427 - val_loss: 0.6534 - val_acc: 0.6240\nEpoch 5/10\n75/75 [==============================] - 10s 137ms/step - loss: 0.5843 - acc: 0.6853 - val_loss: 0.5846 - val_acc: 0.7280\nEpoch 6/10\n75/75 [==============================] - 10s 137ms/step - loss: 0.5369 - acc: 0.7347 - val_loss: 0.5735 - val_acc: 0.7260\nEpoch 7/10\n75/75 [==============================] - 10s 137ms/step - loss: 0.4900 - acc: 0.7680 - val_loss: 0.5950 - val_acc: 0.7100\nEpoch 8/10\n75/75 [==============================] - 10s 136ms/step - loss: 0.4404 - acc: 0.7867 - val_loss: 0.6886 - val_acc: 0.7080\nEpoch 9/10\n75/75 [==============================] - 10s 137ms/step - loss: 0.3889 - acc: 0.8173 - val_loss: 0.6381 - val_acc: 0.7400\nEpoch 10/10\n75/75 [==============================] - 10s 137ms/step - loss: 0.3176 - acc: 0.8547 - val_loss: 0.7398 - val_acc: 0.7040\n\nFold 1\nFound 1500 images belonging to 2 classes.\nFound 500 images belonging to 2 classes.\nTraining\nEpoch 1/10\n75/75 [==============================] - 11s 153ms/step - loss: 0.7556 - acc: 0.4980 - val_loss: 0.6921 - val_acc: 0.6140\nEpoch 2/10\n75/75 [==============================] - 10s 138ms/step - loss: 0.6943 - acc: 0.5660 - val_loss: 0.6786 - val_acc: 0.5260\nEpoch 3/10\n75/75 [==============================] - 10s 137ms/step - loss: 0.6748 - acc: 0.6207 - val_loss: 0.6589 - val_acc: 0.6160\nEpoch 4/10\n75/75 [==============================] - 10s 137ms/step - loss: 0.6264 - acc: 0.6633 - val_loss: 0.5771 - val_acc: 0.7180\nEpoch 5/10\n75/75 [==============================] - 10s 137ms/step - loss: 0.5782 - acc: 0.7093 - val_loss: 0.6399 - val_acc: 0.6800\nEpoch 6/10\n75/75 [==============================] - 10s 137ms/step - loss: 0.5206 - acc: 0.7427 - val_loss: 0.6180 - val_acc: 0.6980\nEpoch 7/10\n75/75 [==============================] - 10s 137ms/step - loss: 0.4661 - acc: 0.7827 - val_loss: 0.5611 - val_acc: 0.7000\nEpoch 8/10\n75/75 [==============================] - 10s 137ms/step - loss: 0.4075 - acc: 0.8147 - val_loss: 0.5899 - val_acc: 0.7140\nEpoch 9/10\n75/75 [==============================] - 10s 137ms/step - loss: 0.3498 - acc: 0.8420 - val_loss: 0.6385 - val_acc: 0.6840\nEpoch 10/10\n75/75 [==============================] - 10s 138ms/step - loss: 0.2767 - acc: 0.8813 - val_loss: 0.8172 - val_acc: 0.6680\n\nFold 2\nFound 1500 images belonging to 2 classes.\nFound 500 images belonging to 2 classes.\nTraining\nEpoch 1/10\n75/75 [==============================] - 11s 151ms/step - loss: 0.7338 - acc: 0.4987 - val_loss: 0.6909 - val_acc: 0.6180\nEpoch 2/10\n75/75 [==============================] - 10s 136ms/step - loss: 0.7080 - acc: 0.5407 - val_loss: 0.6876 - val_acc: 0.5640\nEpoch 3/10\n75/75 [==============================] - 10s 137ms/step - loss: 0.6912 - acc: 0.5920 - val_loss: 0.7144 - val_acc: 0.6300\nEpoch 4/10\n75/75 [==============================] - 10s 137ms/step - loss: 0.6697 - acc: 0.6280 - val_loss: 0.6691 - val_acc: 0.6060\nEpoch 5/10\n75/75 [==============================] - 10s 136ms/step - loss: 0.6073 - acc: 0.6773 - val_loss: 0.5641 - val_acc: 0.6940\nEpoch 6/10\n75/75 [==============================] - 10s 135ms/step - loss: 0.5780 - acc: 0.7233 - val_loss: 0.5822 - val_acc: 0.7020\nEpoch 7/10\n75/75 [==============================] - 10s 138ms/step - loss: 0.5881 - acc: 0.7560 - val_loss: 0.6802 - val_acc: 0.6660\nEpoch 8/10\n75/75 [==============================] - 10s 137ms/step - loss: 0.4517 - acc: 0.7800 - val_loss: 0.6135 - val_acc: 0.6800\nEpoch 9/10\n75/75 [==============================] - 10s 137ms/step - loss: 0.4113 - acc: 0.8113 - val_loss: 0.6037 - val_acc: 0.7180\nEpoch 10/10\n75/75 [==============================] - 10s 138ms/step - loss: 0.3498 - acc: 0.8540 - val_loss: 0.7174 - val_acc: 0.6800\n\nFold 3\nFound 1500 images belonging to 2 classes.\nFound 500 images belonging to 2 classes.\nTraining\nEpoch 1/10\n75/75 [==============================] - 11s 151ms/step - loss: 0.7114 - acc: 0.5080 - val_loss: 0.6907 - val_acc: 0.5980\nEpoch 2/10\n75/75 [==============================] - 10s 137ms/step - loss: 0.7049 - acc: 0.5613 - val_loss: 0.6759 - val_acc: 0.5920\nEpoch 3/10\n75/75 [==============================] - 10s 137ms/step - loss: 0.6653 - acc: 0.6227 - val_loss: 0.6750 - val_acc: 0.5600\nEpoch 4/10\n75/75 [==============================] - 10s 137ms/step - loss: 0.6402 - acc: 0.6433 - val_loss: 0.6386 - val_acc: 0.6460\nEpoch 5/10\n75/75 [==============================] - 10s 140ms/step - loss: 0.5794 - acc: 0.7007 - val_loss: 0.6564 - val_acc: 0.6420\nEpoch 6/10\n75/75 [==============================] - 10s 135ms/step - loss: 0.5122 - acc: 0.7553 - val_loss: 0.6319 - val_acc: 0.6700\nEpoch 7/10\n75/75 [==============================] - 10s 138ms/step - loss: 0.4448 - acc: 0.7940 - val_loss: 0.6115 - val_acc: 0.7080\nEpoch 8/10\n75/75 [==============================] - 10s 137ms/step - loss: 0.3882 - acc: 0.8180 - val_loss: 0.8759 - val_acc: 0.6860\nEpoch 9/10\n75/75 [==============================] - 10s 136ms/step - loss: 0.3494 - acc: 0.8480 - val_loss: 0.6936 - val_acc: 0.6980\nEpoch 10/10\n75/75 [==============================] - 10s 137ms/step - loss: 0.2821 - acc: 0.8820 - val_loss: 0.7675 - val_acc: 0.7200\n" ], [ "import numpy as np\nfor key, val in hist.items():\n print('{0}: {1}'.format(key, np.mean(np.array(val))))", "val_loss: 0.7604647065699102\nval_acc: 0.6929999971389771\nloss: 0.30655239490171277\nacc: 0.8679999987284343\n" ], [ "save_obj(hist,'hist_config1.pkl')", "_____no_output_____" ], [ "hist_retr = load_obj('hist_config1.pkl')", "_____no_output_____" ], [ "for key, val in hist_retr.items():\n print('{0}: {1}'.format(key, np.mean(np.array(val))))", "val_loss: 0.7604647065699102\nval_acc: 0.6929999971389771\nloss: 0.30655239490171277\nacc: 0.8679999987284343\n" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code" ]
[ [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code" ] ]
ecea4a69dd6e0915d637f068a6f71ec630e52608
8,115
ipynb
Jupyter Notebook
ray-rllib/explore-rllib/extras/Extra-Application-Taxi.ipynb
ewbolme/academy
87bda18d4122dbaa3f67d3a1f6c4b534bd279992
[ "Apache-2.0" ]
null
null
null
ray-rllib/explore-rllib/extras/Extra-Application-Taxi.ipynb
ewbolme/academy
87bda18d4122dbaa3f67d3a1f6c4b534bd279992
[ "Apache-2.0" ]
null
null
null
ray-rllib/explore-rllib/extras/Extra-Application-Taxi.ipynb
ewbolme/academy
87bda18d4122dbaa3f67d3a1f6c4b534bd279992
[ "Apache-2.0" ]
null
null
null
31.823529
384
0.580283
[ [ [ "# Ray RLlib - Extra Application Example - Taxi-v3\n\n© 2019-2020, Anyscale. All Rights Reserved\n\n![Anyscale Academy](../../../images/AnyscaleAcademyLogo.png)\n\nThis example uses [RLlib](https://ray.readthedocs.io/en/latest/rllib.html) to train a policy with the `Taxi-v3` environment ([gym.openai.com/envs/Taxi-v3/](https://gym.openai.com/envs/Taxi-v3/)). The goal is to pick up passengers as fast as possible, negotiating the available paths. This is one of OpenAI Gym's [\"toy text\"](https://gym.openai.com/envs/#toy_text) problems.\n\nFor more background about this problem, see:\n\n* [\"Hierarchical Reinforcement Learning with the MAXQ Value Function Decomposition\"](https://arxiv.org/abs/cs/9905014), [Thomas G. Dietteric](https://twitter.com/tdietterich)\n* [\"Reinforcement Learning: let’s teach a taxi-cab how to drive\"](https://towardsdatascience.com/reinforcement-learning-lets-teach-a-taxi-cab-how-to-drive-4fd1a0d00529), [Valentina Alto](https://twitter.com/AltoValentina)", "_____no_output_____" ] ], [ [ "import pandas as pd\nimport json\nimport os\nimport shutil\nimport sys\nimport ray\nimport ray.rllib.agents.ppo as ppo", "_____no_output_____" ], [ "info = ray.init(ignore_reinit_error=True)", "_____no_output_____" ], [ "print(\"Dashboard URL: http://{}\".format(info[\"webui_url\"]))", "_____no_output_____" ] ], [ [ "Set up the checkpoint location:", "_____no_output_____" ] ], [ [ "checkpoint_root = \"tmp/ppo/taxi\"\nshutil.rmtree(checkpoint_root, ignore_errors=True, onerror=None) # clean up old runs", "_____no_output_____" ] ], [ [ "Next we'll train an RLlib policy with the `Taxi-v3` environment.\n\nBy default, training runs for `10` iterations. Increase the `N_ITER` setting if you want to see the resulting rewards improve.\nAlso note that *checkpoints* get saved after each iteration into the `/tmp/ppo/taxi` directory.\n\n> **Note:** If you prefer to use a different directory root than `/tmp`, change it in the next cell **and** in the `rllib rollout` command below.", "_____no_output_____" ] ], [ [ "SELECT_ENV = \"Taxi-v3\"\nN_ITER = 10\n\nconfig = ppo.DEFAULT_CONFIG.copy()\nconfig[\"log_level\"] = \"WARN\"\n\nagent = ppo.PPOTrainer(config, env=SELECT_ENV)", "_____no_output_____" ], [ "results = []\nepisode_data = []\nepisode_json = []\n\nfor n in range(N_ITER):\n result = agent.train()\n results.append(result)\n \n episode = {'n': n, \n 'episode_reward_min': result['episode_reward_min'], \n 'episode_reward_mean': result['episode_reward_mean'], \n 'episode_reward_max': result['episode_reward_max'], \n 'episode_len_mean': result['episode_len_mean']\n }\n \n episode_data.append(episode)\n episode_json.append(json.dumps(episode))\n file_name = agent.save(checkpoint_root)\n \n print(f'{n+1:3d}: Min/Mean/Max reward: {result[\"episode_reward_min\"]:8.4f}/{result[\"episode_reward_mean\"]:8.4f}/{result[\"episode_reward_max\"]:8.4f}, len mean: {result[\"episode_len_mean\"]:8.4f}. Checkpoint saved to {file_name}')", "_____no_output_____" ] ], [ [ "Do the episode rewards increase after multiple iterations?\n\nAlso, print out the policy and model to see the results of training in detail…", "_____no_output_____" ] ], [ [ "import pprint\n\npolicy = agent.get_policy()\nmodel = policy.model\n\npprint.pprint(model.variables())\npprint.pprint(model.value_function())\n\nprint(model.base_model.summary())", "_____no_output_____" ] ], [ [ "## Rollout\n\nNext we'll use the [`rollout` script](https://ray.readthedocs.io/en/latest/rllib-training.html#evaluating-trained-policies) to evaluate the trained policy.\n\nThe output from the following command visualizes the \"taxi\" agent operating within its simulation: picking up a passenger, driving, turning, dropping off a passenger (\"put-down\"), and so on. \n\nA 2-D map of the *observation space* is visualized as text, which needs some decoding instructions:\n\n * `R` -- R(ed) location in the Northwest corner\n * `G` -- G(reen) location in the Northeast corner\n * `Y` -- Y(ellow) location in the Southwest corner\n * `B` -- B(lue) location in the Southeast corner\n * `:` -- cells where the taxi can drive\n * `|` -- obstructions (\"walls\") which the taxi must avoid\n * blue letter represents the current passenger’s location for pick-up\n * purple letter represents the drop-off location\n * yellow rectangle is the current location of our taxi/agent\n\nThat allows for a total of 500 states, and these known states are numbered between 0 and 499.\n\nThe *action space* for the taxi/agent is defined as:\n\n * move the taxi one square North\n * move the taxi one square South\n * move the taxi one square East\n * move the taxi one square West\n * pick-up the passenger\n * put-down the passenger\n\nThe *rewards* are structured as −1 for each action plus:\n\n * +20 points when the taxi performs a correct drop-off for the passenger\n * -10 points when the taxi attempts illegal pick-up/drop-off actions\n\nAdmittedly it'd be better if these state visualizations showed the *reward* along with observations.", "_____no_output_____" ] ], [ [ "!rllib rollout \\\n tmp/ppo/taxi/checkpoint_10/checkpoint-10 \\\n --config \"{\\\"env\\\": \\\"Taxi-v3\\\"}\" \\\n --run PPO \\\n --steps 2000", "_____no_output_____" ], [ "ray.shutdown() # \"Undo ray.init()\".", "_____no_output_____" ] ], [ [ "## Exercise (\"Homework\")\n\nIn addition to _Taxi_, there are other so-called [\"toy text\"](https://gym.openai.com/envs/#toy_text) problems you can try.", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown" ]
[ [ "markdown" ], [ "code", "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ] ]
ecea58f0fe3e2bb3d6aec5dc8b96cb7b05604660
2,213
ipynb
Jupyter Notebook
DAY 101 ~ 200/DAY108_[SW Expert Academy] 4522번 세상의 모든 팰린드롬 (Python).ipynb
SOMJANG/CODINGTEST_PRACTICE
1a7304e9063579441b8a67765175c82b0ad93ac9
[ "MIT" ]
15
2020-03-17T01:18:33.000Z
2021-12-24T06:31:06.000Z
DAY 101 ~ 200/DAY108_[SW Expert Academy] 4522번 세상의 모든 팰린드롬 (Python).ipynb
SOMJANG/CODINGTEST_PRACTICE
1a7304e9063579441b8a67765175c82b0ad93ac9
[ "MIT" ]
null
null
null
DAY 101 ~ 200/DAY108_[SW Expert Academy] 4522번 세상의 모든 팰린드롬 (Python).ipynb
SOMJANG/CODINGTEST_PRACTICE
1a7304e9063579441b8a67765175c82b0ad93ac9
[ "MIT" ]
10
2020-03-17T01:18:34.000Z
2022-03-30T10:53:07.000Z
27.6625
170
0.503389
[ [ [ "## 2020년 5월 24일 일요일\n### SW Expert Academy - 세상의 모든 팰린드롬\n### 문제 : https://swexpertacademy.com/main/code/problem/problemDetail.do?contestProbId=AWO6Oao6N4QDFAWw&categoryId=AWO6Oao6N4QDFAWw&categoryType=CODE\n### 블로그 : https://somjang.tistory.com/entry/SWExpertAcademy-4522%EB%B2%88-%EC%84%B8%EC%83%81%EC%9D%98-%EB%AA%A8%EB%93%A0-%ED%8C%B0%EB%A6%B0%EB%93%9C%EB%A1%AC-Python", "_____no_output_____" ], [ "### 첫번째 시도", "_____no_output_____" ] ], [ [ "def check_palindrome(string):\n isPalindrome = \"Exist\"\n last_index = len(string) - 1\n for i in range(len(string) // 2):\n if string[i] != string[last_index - i]:\n isPalindrome = \"Not exist\"\n break\n return isPalindrome\n\ndef change_string(string):\n string = list(string)\n last_index = len(string) - 1 \n for i in range(len(string) // 2):\n if string[i] == '?' and string[last_index - i] != '?':\n string[last_index - i] = '?'\n elif string[i] != '?' and string[last_index - i] == '?':\n string[i] = '?'\n return string\n\nT = int(input())\n \nfor i in range(T):\n input_str = str(input())\n \n change_str = change_string(input_str)\n \n check = check_palindrome(change_str)\n \n print(\"#{} {}\".format(i+1, check))", "_____no_output_____" ] ] ]
[ "markdown", "code" ]
[ [ "markdown", "markdown" ], [ "code" ] ]
ecea5b8aa54b2fdbd140ee7af962466145fb0f3d
9,222
ipynb
Jupyter Notebook
Notebook/RNN.ipynb
huiwenzhang/100-Days-ML-Note
1503dff1960bf19b05812fd0d7fc3f633874654f
[ "MIT" ]
1
2019-01-08T04:55:57.000Z
2019-01-08T04:55:57.000Z
Notebook/RNN.ipynb
huiwenzhang/100-Days-ML-Note
1503dff1960bf19b05812fd0d7fc3f633874654f
[ "MIT" ]
null
null
null
Notebook/RNN.ipynb
huiwenzhang/100-Days-ML-Note
1503dff1960bf19b05812fd0d7fc3f633874654f
[ "MIT" ]
2
2019-09-09T07:00:19.000Z
2019-11-17T04:00:21.000Z
34.539326
287
0.553025
[ [ [ "## RNN教程\n- RNN原理简介\n- Basic RNN\n- LSTM\n- CharRNN项目", "_____no_output_____" ], [ "## 案例讲解\n通过一个案例讲解如何使用RNN.项目目的是用LSTM学习句子生成.训练数据来自于莎士比亚的著作,以文本的形式存储.我们使用多层多步的N*N模型来对这个问题建模.\n比如,输入: how are you?对应的标签就是:ow are you? h.通过上一个字符预测下一个字符.字符构成单词,单词构成句子.\n所以训练集样本就是一个个字符,对应标签也是字符,并且是每个训练样本对应的下一个字符.", "_____no_output_____" ], [ "## 代码流程\n### 数据输入\n\n```python\nself.inputs = tf.placeholder(tf.int32, shape=(self.batch_size, self.num_steps),\n name='inputs')\nself.targets = tf.placeholder(tf.int32, shape=(self.batch_size, self.num_steps),\n name='labels')\nself.keep_prob = tf.placeholder(tf.float32, name='keep_prob')\n\nif not self.use_embedding:\n self.rnn_inputs = tf.one_hot(self.inputs, self.n_classes)\nelse:\n embedding = tf.get_variable('embedding', [self.n_classes, self.embedding_size])\n self.rnn_inputs = tf.nn.embedding_lookup(embedding, self.inputs)\n \n```\n**Note:注意这里的维度匹配问题**\n- batch_size:表示一次输入多少个句子,每个句子相当于普通意义下的一个样本.实际上,在模型内部还是一个一个字符输入的.只是使用`tf.nn.rnn_cell.dynamic()`可以自动的把多个字符串起来.看起来的效果就是输入了一个句子\n- num_steps:句子的长度,有多少个字符,或者叫做RNN的展开长度\n- rnn_inputs.shape: (batch_size, num_steps, dimension of each char).由于每个字符被表示成了一个one-hot的向量,因此输入句子中每个字符是多维的\n- 对于汉字的话我们会显式的embedding一下,因此显然每个汉子对应的也是个多维向量\n\n> 为什么汉字不用one-hot编码?\n我们要对输入进行特征表示,计算机只认识数字.你输入字符a,它并不知道什么意思?我们需要把它表示成一段数字,这段数字能够很好的刻画这个字符或者标识这个字符.这就叫做特征提取.对于英文字符,我们可以使用one-hot的形式.什么是one-hot呢?比如我们有猫,狗老鼠三个动物,那么one-hot表示方法就是:猫=100,狗=010,老鼠=001.即用每一位表示一个动物,是哪个动物,就激活哪个位置.对于英文字符,总共也就26个英文字母,加上标点符号等等,顶多也就100个对象.如果用one-hot表示,每个对象的向量长度肯定是小于100的.这个是可以接受的.\n如果是汉字呢?你可能晕了,中国文化如此博大精深,那汉字还能少的了.姑且不说那些生僻字,仅仅是常用的汉字也有3000多个.如果用one-hot表示,每个样本都是3000多维的,这肯定不行啊.所以需要引入embedding,你可以理解它赋予了每个汉字一个更加紧凑的表征.\n\n### 模型搭建\n- 利用`tf.nn.rnn_cell.BasicLSTMCell(size)`创建一个基本的LSTM模块\n**Note:**这里的size其实相当于MLP中隐含层神经元的个数.这个size容易和num_step参数混淆.num_step是在一个时间维度上将下面函数定义的cell复制num_steps次.所以每个cell都是一样的,包括权重都是一样的.而且每个cell里面有size个神经元,对应每个cell是size维的输出.\n\n```python\ndef create_cell(lstm_size, keep_prob):\n lstm = tf.nn.rnn_cell.BasicLSTMCell(lstm_size)\n drop = tf.nn.rnn_cell.DeviceWrapper(lstm, output_keep_prob=keep_prob)\n return drop\n```\n\n- 堆叠多层的cell\n\n```python\ncell = tf.nn.rnn_cell.MultiRNNCell(\n [create_cell(self.lstm_size, self.keep_prob) for _ in range(self.n_layers)])\nself.init_state = cell.zero_state(self.batch_size,\n tf.float32) # state is the hidden state\n```\n\n- 在时间维度上展开,构成一个句子\n\n```python\nself.lstm_outputs, self.final_state = tf.nn.dynamic_rnn(cell, self.rnn_inputs,\n initial_state=self.init_state)\n```\n\n- 输出.上面的输出的是多步隐藏层输出.每个step是个多维的(本例128维),为了得到最后的输出,还需要经过全连接和softmax才能得到每个字符出现的概率(所以最后输出的维度是:(num_steps, num_chars))\n\n```python\nseq_output = tf.concat(self.lstm_outputs, 1)\nx = tf.reshape(seq_output, [-1, self.lstm_size]) # row: bath_size * num_steps\n\nwith tf.variable_scope('output'):\n w = tf.Variable(tf.truncated_normal([self.lstm_size, self.n_classes], stddev=0.1))\n b = tf.Variable(tf.zeros(self.n_classes))\n self.logits = tf.matmul(x, w) + b\n self.preds = tf.nn.softmax(self.logits, name='prob_pred')\n```\n### 计算损失\n\n```python\nlabel = tf.one_hot(self.targets, self.n_classes)\ny = tf.reshape(label, self.logits.get_shape())\nloss = tf.nn.softmax_cross_entropy_with_logits(logits=self.logits, labels=y)\nself.loss = tf.reduce_mean(loss)\n```\n\n### 设置优化器\n\n```python\ntrain_vars = tf.trainable_variables()\ngrads, _ = tf.clip_by_global_norm(tf.gradients(loss, train_vars), self.grid_clip)\noptimizer = tf.train.AdamOptimizer(self.lr)\nself.optimizer = optimizer.apply_gradients(zip(grads, train_vars))\n```\n", "_____no_output_____" ], [ "## 训练模型\n```\nself.session = tf.Session()\nwith self.session as sess:\n sess.run(tf.global_variables_initializer())\n step = 0\n new_state = sess.run(self.init_state)\n for x, y in batch_generator:\n start = time.time()\n feed_dict = {self.inputs: x, self.targets: y, self.keep_prob: self.train_keep_prob,\n self.init_state: new_state}\n batch_loss, new_state, prob, _ = sess.run(\n [self.loss, self.final_state, self.preds, self.optimizer], feed_dict=feed_dict)\n step += 1\n end = time.time()\n\n # print out\n if step % log_interval == 0:\n print('Step: {}, Loss: {:.4f}, Time: {:.4f}'.format(step, batch_loss,\n end - start))\n if step % save_interval == 0:\n self.saver.save(sess, os.path.join(save_path, 'model'), global_step=step)\n\n if step > max_steps:\n return\n```\n主要步骤:\n- 准备数据\n- 准备输入:`self.inputs, selt.targets, self.keep_prob, self.init_state`\n- 计算损失,计算隐状态,计算预测输出,执行梯度更新\n", "_____no_output_____" ], [ "## 预测模型\n```\nsamples = [c for c in prime]\nnew_state = self.session.run(self.init_state)\npreds = np.ones((vocab_size,))\nfor c in prime:\n x = np.zeros((1, 1))\n x[0, 0] = c\n\n feed_dict = {self.inputs: x,\n self.keep_prob: 1.,\n self.init_state: new_state}\n preds, new_state = self.session.run([self.preds, self.final_state], feed_dict=feed_dict)\n\nc = pick_top_n(preds, vocab_size)\nsamples.append(c)\n\nfor _ in range(n_samples):\n x = np.zeros((1, 1))\n x[0, 0] = c\n feed = {self.inputs: x,\n self.keep_prob: 1.,\n self.init_state: new_state}\n preds, new_state = self.session.run([self.preds, self.final_state],\n feed_dict=feed)\n\n c = pick_top_n(preds, vocab_size)\n samples.append(c)\n\nreturn np.array(samples)\n```\n- 逐个字符的预测\n- 将字符转成对应的索引,预测出来的也是数字,最后把数字串转化称字符", "_____no_output_____" ], [ "## 运行\n- 训练莎士比亚\n```bash\npython Notebook/rnn_train.py \\\n --input_file datasets/shakespeare.txt \\\n --name shakespeare \\\n --num_steps 50 \\\n --batch_size 32 \\\n --learning_rate 0.01 \\\n --max_steps 20000\n```\n- 测试莎士比亚\n```bash\npython Notebook/predict.py \\\n --converter_path RNNmodel/shakespeare/vocab.pkl \\\n --checkpoint_path RNNmodel/shakespeare/ \\\n --max_length 1000\n```\n- 训练诗句\n```bash\npython Notebook/rnn_train.py \\\n --use_embedding \\\n --input_file datasets/poetry.txt \\\n --name poetry \\\n --learning_rate 0.005 \\\n --num_steps 26 \\\n --batch_size 32 \\\n --max_steps 80000\n```\n- 测试生成诗句\n```bash\npython Notebook/predict.py \\\n --use_embedding \\\n --converter_path RNNmodel/poetry/vocab.pkl \\\n --checkpoint_path RNNmodel/poetry/ \\\n --max_length 300\n```", "_____no_output_____" ] ] ]
[ "markdown" ]
[ [ "markdown", "markdown", "markdown", "markdown", "markdown", "markdown" ] ]
ecea6929a1bfbdab5f881afff7c3343c8d399568
889,575
ipynb
Jupyter Notebook
notebooks/oscillators.ipynb
sbernasek/genessa
b00d3173e9940943a513ab138d14d770bfd3c132
[ "MIT" ]
2
2020-02-22T09:53:23.000Z
2020-02-24T19:01:28.000Z
notebooks/oscillators.ipynb
sebastianbernasek/genessa
b00d3173e9940943a513ab138d14d770bfd3c132
[ "MIT" ]
null
null
null
notebooks/oscillators.ipynb
sebastianbernasek/genessa
b00d3173e9940943a513ab138d14d770bfd3c132
[ "MIT" ]
null
null
null
2,207.382134
153,448
0.959124
[ [ [ "import numpy as np\nimport matplotlib.pyplot as plt\n \n%reload_ext autoreload\n%autoreload 2\n%matplotlib inline", "_____no_output_____" ] ], [ [ "# Example 3: Network of Coupled Circadian Oscillators", "_____no_output_____" ], [ "In this example we will simulate a population of coupled circadian oscillators. This simulation is considerably more complex than the previous two examples. We will overcome this problem by using an existing template of a single circadian oscillator available within GeneSSA. This oscillator is based upon a proposed model of the Per-Tim clock network in Drosophila (Gonze et al. 2003).\n\n[1] Gonze, Didier, et al. \"Stochastic models for circadian rhythms: effect of molecular noise on periodic and chaotic behaviour.\" Comptes rendus biologies 326.2 (2003): 189-203.", "_____no_output_____" ], [ "![image.png](attachment:image.png)", "_____no_output_____" ], [ "![image.png](attachment:image.png)", "_____no_output_____" ], [ "![image.png](attachment:image.png)", "_____no_output_____" ], [ "![image.png](attachment:image.png)", "_____no_output_____" ], [ "## Example 3.1: Simulating an individual oscillator\n\nFirst we will simulate a single oscillator operating in isolation.\n\n### Creating an individual Per-Tim oscillator", "_____no_output_____" ], [ "We can use GeneSSA to instantiate an individual oscillator. The `transcription` parameter indicates whether or not we should include constitutive transcription of per and tim transcripts, while the `omega` parameter sets the approximate magnitude of Per and Tim protein levels.", "_____no_output_____" ] ], [ [ "from genessa.demo.oscillators import Oscillator\n\noscillator = Oscillator(transcription=True, omega=100)", "_____no_output_____" ] ], [ [ "The `Oscillator` class inherits from the `Network` class. All `Network` instances provide a `__repr__` attribute that provide a quick summary of all the reactions in the network. This allows us to quickly inspect each of the 30 reactions in our individual oscillator.", "_____no_output_____" ] ], [ [ "oscillator", " Rxn Reactants Products Propensity Parameter\n--------------------------- ----------- ---------- ------------------------- ---------------------\n p decay 0 [0] 0.01000\n P0 translation 1 [0] 0.90000\n P0 decay 1 [1] 0.01000\n t decay 2 [2] 0.01000\n T0 translation 3 [2] 0.90000\n T0 decay 3 [3] 0.01000\n P0 transcript deg. 0 1 / (1 + 20.0/[0]) 70.00000\n T0 transcript deg. 2 1 / (1 + 20.0/[2]) 70.00000\n P0 phosphorylation 1 4 1 / (1 + 200.0/[1]) 800.00000\n P1 dephosphorylation 4 1 1 / (1 + 200.0/[4]) 100.00000\n P1 decay 4 [4] 0.01000\n P1 phosphorylation 4 5 1 / (1 + 200.0/[4]) 800.00000\n P2 dephosphorylation 5 4 1 / (1 + 200.0/[5]) 100.00000\n P2 decay 5 [5] 0.01000\n T0 phosphorylation 3 6 1 / (1 + 200.0/[3]) 800.00000\n T1 dephosphorylation 6 3 1 / (1 + 200.0/[6]) 100.00000\n T1 decay 6 [6] 0.01000\n T1 phosphorylation 6 7 1 / (1 + 200.0/[6]) 800.00000\n T2 dephosphorylation 7 6 1 / (1 + 200.0/[7]) 100.00000\n T2 decay 7 [7] 0.01000\n P2 deg. 5 1 / (1 + 20.0/[5]) 200.00000\n T2 deg. 7 1 / (1 + 20.0/[7]) 200.00000 + 2.0[IN_0]\n C association 5, 7 8 [5][7] 0.01200\n C dissociation 8 5, 7 [8] 0.60000\n C decay 8 [8] 0.01000\n C import 8 9 [8] 0.60000\n C export 9 8 [9] 0.20000\n nC decay 9 [9] 0.01000\n T0 transcription 2 100.00000\nT0 transcription repression 1 / (1 + ([9]/100.0)^4.0)\n P0 transcription 0 100.00000\nP0 transcription repression 1 / (1 + ([9]/100.0)^4.0)\n" ] ], [ [ "### Simulating an individual Per-Tim oscillator", "_____no_output_____" ], [ "We can now run our simulation as we did before, then plot the resultant dynamics of the nuclear Per-Tim complex ($nC$). The `Oscillator.proteins` attribute provides a dictionary that maps species names, such as $nC$, to dimensions of the state space.", "_____no_output_____" ] ], [ [ "from genessa.solver.stochastic import MonteCarloSimulation\n\n# define an initial condition\nic = np.ones(oscillator.N)*100\n\n# run simulation\nsimulation = MonteCarloSimulation(oscillator, ic=ic)\ntimeseries = simulation.run(N=1, duration=240, dt=1)\n\n# create figure\nfig, ax = plt.subplots()\nax.set_xlabel('Time (h)')\nax.set_ylabel('Level')\n\n# plot trajectories\nspecies = 'nC'\ndimension = oscillator.proteins[species]\nfor trajectory in timeseries.states[:, dimension, :]:\n ax.plot(timeseries.t, trajectory, '-k')", "_____no_output_____" ] ], [ [ "## Example 3.2: Simulating a network of coupled oscillators", "_____no_output_____" ], [ "We will now extend our simulation to a network of coupled oscillators. Oscillators are coupled by their transcription rates as detailed by Gonze et al.:", "_____no_output_____" ], [ "![image.png](attachment:image.png)", "_____no_output_____" ], [ "![image.png](attachment:image.png)", "_____no_output_____" ], [ "![image.png](attachment:image.png)", "_____no_output_____" ], [ "### Creating a population of Per-Tim oscillators\n\nGeneSSA offers a `CoupledOscillators` template for this purpose. The `replicates` parameter defines the number of oscillators in the network.", "_____no_output_____" ] ], [ [ "from genessa.demo.oscillators import CoupledOscillators\n\n# instantiate system of coupled oscillators\nnetwork = CoupledOscillators(omega=100, replicates=5)", "_____no_output_____" ] ], [ [ "### Adding transcriptional coupling to the network\n\nOscillators are coupled by providing an adjacency matrix for the network. GeneSSA also provides methods for quick addition of some common coupling strategies. These include a random coupling in which edges are randomly generated, and a dense coupling in which the network is fully connected. We will add a random coupling using the `CoupledOscillators.add_random_coupling` method. The `a` parameter determines the coupling strength, and is equivalent to $\\alpha$ in the model above.\n", "_____no_output_____" ] ], [ [ "network.add_random_coupling(a=0.01)", "_____no_output_____" ] ], [ [ "### Run stochastic simulation\n\nWe will now run our simulation as we did before.", "_____no_output_____" ] ], [ [ "from genessa.solver.stochastic import MonteCarloSimulation\n\n# run simulation\nic = np.ones(network.N)*100\nsimulation = MonteCarloSimulation(network, ic=ic)\ntimeseries = simulation.run(N=1, duration=24*10, dt=1)", "_____no_output_____" ] ], [ [ "### Plot simulated dynamics of nuclear Per-Tim complex\n\nNote that the `CoupledOscillators.proteins` attribute provides a dictionary of dictionaries mapping species names to their state space dimensions. This is because each cell in the network has its own list of species.", "_____no_output_____" ] ], [ [ "# create figure\nfig, ax = plt.subplots()\nax.set_xlabel('Time (h)')\nax.set_ylabel('Level')\n\n# plot trajectories\nspecies = 'nC'\nfor p in network.proteins.values():\n ax.plot(timeseries.t, timeseries.states[0, p[species], :], '-k')", "_____no_output_____" ] ] ]
[ "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code" ]
[ [ "code" ], [ "markdown", "markdown", "markdown", "markdown", "markdown", "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown", "markdown", "markdown", "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ] ]
ecea6989c3326ee4c5d052a8e925e6cd53e18d39
46,469
ipynb
Jupyter Notebook
Mini Project 2 Code.ipynb
ryanvolpi/income_prediction
0b5ca5d46c8f955d7a0f8f82bce498e6480919f2
[ "MIT" ]
null
null
null
Mini Project 2 Code.ipynb
ryanvolpi/income_prediction
0b5ca5d46c8f955d7a0f8f82bce498e6480919f2
[ "MIT" ]
null
null
null
Mini Project 2 Code.ipynb
ryanvolpi/income_prediction
0b5ca5d46c8f955d7a0f8f82bce498e6480919f2
[ "MIT" ]
null
null
null
44.214082
7,260
0.560244
[ [ [ "import pandas as pd\nimport numpy as np\nimport math\nfrom sklearn.impute import SimpleImputer\nimport rpy2.robjects as robjects\nfrom rpy2.robjects.packages import importr\nfrom rpy2.robjects import pandas2ri\nimport rpy2\nfrom matplotlib import pyplot as plt\nfrom sklearn.compose import ColumnTransformer\nfrom sklearn.pipeline import Pipeline\nfrom sklearn.preprocessing import MinMaxScaler\nfrom sklearn.preprocessing import PowerTransformer\nfrom sklearn.preprocessing import OneHotEncoder\nfrom sklearn.preprocessing import StandardScaler\nfrom sklearn.model_selection import GridSearchCV\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.ensemble import RandomForestClassifier\nfrom sklearn.neighbors import KNeighborsClassifier\nfrom sklearn.linear_model import LogisticRegression\nfrom mord import LogisticIT\nimport warnings \nimport copy \nwarnings.simplefilter('ignore')\n\nSEED = 777", "_____no_output_____" ] ], [ [ "## Load Data", "_____no_output_____" ] ], [ [ "col_names = ['income','sex','married','age','education','occupation',\n 'time_in_area','dual_income','num_people_household',\n 'num_children_household','home_status','home_type','ethnicity',\n 'language']\ndf = pd.read_csv('marketing.data', delimiter=\" \",header=None, names=col_names)\n\ndf.head()", "_____no_output_____" ], [ "Y = np.array(df['income'])\nX = np.array(df.drop('income',axis=1))\n\nplt.bar(pd.value_counts(Y).index,pd.value_counts(Y).values/len(Y))\nplt.show()\n\nprint(f\"Baseline Accuracy: {pd.value_counts(Y).values[0]/len(Y)}\")", "_____no_output_____" ] ], [ [ "Split the data 80-20 into a train and a test set using stratified random sampling to preserve the proportion of classes in the response variable", "_____no_output_____" ] ], [ [ "x_train,x_test,y_train,y_test = train_test_split(X,Y, test_size=.2, stratify = Y, random_state = SEED)", "_____no_output_____" ] ], [ [ "### Define a KNN Imputer\nDefine a class to impute missing values in each feature using K-Nearest-Neighbors Classifier trained to predict the non-missing values using the other features. Feature classifiers are only fit on the training data in each step and do not see the test or validation data during training. This is to prevent data leakage.", "_____no_output_____" ] ], [ [ "class KNNImputer():\n def __init__(self):\n # Try an array of n values for each classifier\n self.model = GridSearchCV(KNeighborsClassifier(),{'n_neighbors':[5,10,15,20]}, cv=3, n_jobs=-1)\n self.model_list = [None]\n self.X = np.array([])\n \n def fit(self, X, verbose = 0):\n # create deep copy to avoid altering the original data\n self.X = copy.deepcopy(X)\n\n if len(self.X.shape)==1 or self.X.shape[1]==0:\n self.X = self.X.reshape(-1,1)\n\n self.model_list = [None]*(self.X.shape[1])\n \n # count of missing values per column\n col_nan_count = np.sum(np.isnan(self.X),axis=0)\n \n # Indexes of columns with at least 1 missing value sorted from most to least missing values\n col_nan_count_sorted = sorted(enumerate(col_nan_count),key=lambda x: x[1], reverse=True)\n col_ix_sorted = [tup[0] for tup in col_nan_count_sorted if tup[1]>0]\n\n tot_nan = sum(np.sum(np.isnan(self.X),axis=0))\n \n # for each column with missing values, sorted from most missing values to least..\n for i in col_ix_sorted:\n target = self.X[:,i].reshape(-1,1) # feature to fill missing values for\n features = np.delete(self.X, i, axis=1) # features used to predict missing values\n \n # scale features so that each has equal sway in KNN classification\n scaler = StandardScaler()\n features = scaler.fit_transform(features)\n \n # seperate missing values (to be predicted) and existing values (used to train classifier)\n target_train = target[[not na for na in np.isnan(target)]]\n target_missing = target[np.isnan(target)]\n \n # isolate observations with corresponding target values for training\n features_train = features[[not na for na in np.isnan(target).flatten()]]\n # temporarily fill in missing values in observation features used to train imputer\n mean_imputer = SimpleImputer(strategy='most_frequent')\n features_train = mean_imputer.fit_transform(features_train)\n \n # isolate feature observations with no corresponding target feature for prediction\n features_missing = features[np.isnan(target).flatten()]\n features_missing = mean_imputer.fit_transform(features_missing)\n \n # train model and store it in mode list corresponding to feature index\n self.model_list[i] = self.model.fit(features_train, target_train)\n\n in_sample_acc = self.model_list[i].score(features_train, target_train)\n #if False:\n # print(f\"Column {i} - filled {col_nan_count[i]} NaN values - in sample accuracy: {in_sample_acc}\")\n \n # predict in missing values\n predictions_missing = self.model.predict(features_missing)\n # fill missing values in with predictions\n target[np.isnan(target)] = predictions_missing\n self.X[:,i] = target.flatten()\n \n # for each column without missing values \n non_na_col_ix = set(range(0,self.X.shape[1])).difference(col_ix_sorted)\n #train model to predict the column features using other features. No missing values to fill in.\n for i in non_na_col_ix:\n target = self.X[:,i].reshape(-1,1)\n features = np.delete(self.X, i, axis=1)\n \n self.model_list[i] = self.model.fit(features, target)\n \n def fit_transform(self, X, verbose = 0):\n X_backup = copy.deepcopy(X)\n \n self.fit(X, verbose = verbose)\n assert sum(sum(np.isnan(X))) == sum(sum(np.isnan(X_backup)))\n return self.X\n\n def transform(self, X):\n self.X = copy.deepcopy(X)\n col_nan_count = np.sum(np.isnan(X),axis=0)\n col_ixs = [i for i, count in enumerate(col_nan_count) if count>0]\n\n for i in col_ixs:\n target = self.X[:,i].reshape(-1,1)\n features = np.delete(self.X, i, axis=1)\n features_missing = features[np.isnan(target).flatten()]\n if len(features_missing.shape)==1:\n features_missing = features_missing.reshape(-1,1)\n features_missing = mean_imputer.fit_transform(features_missing)\n\n predictions_missing = self.model_list[i].predict(features_missing)\n\n target[np.isnan(target)] = predictions_missing\n \n # Replace features \n self.X[:,i] = target.flatten()\n\n return self.X", "_____no_output_____" ] ], [ [ "### Preprocessing", "_____no_output_____" ] ], [ [ "from sklearn.feature_extraction.text import CountVectorizer\nfrom sklearn.base import BaseEstimator, TransformerMixin\n\ndef col_ix(label):\n col_list = ['sex','married','age','education','occupation',\n 'time_in_area','dual_income','num_people_household',\n 'num_children_household','home_status','home_type','ethnicity',\n 'language']\n\n return col_list.index(label)\n\n# Artificial predictor to satisfy Pipeline's requirement that the last class needs to be an estimator \n# (able to take a Y parameter)\nclass LastEstimator():\n def __init__(self):\n self.encoder=None\n def fit(self,X,Y=None):\n return self\n def transform(self, X, Y=None):\n return X\n def fit_transform(self, X, Y=None):\n #print(\"LastEstimator.fit_transform\")\n return X\n\n# Class to convert 0 dimensional NP arrays to 1 dimensional NP arrays for compatibility \nclass reshaper(BaseEstimator, TransformerMixin):\n def __init__(self):\n self.x = 0\n\n def fit(self, X=None, Y=None):\n return self\n\n def transform(self, X, Y=None):\n if type(X) is pd.Series:\n X=X.values\n return X.reshape(-1, 1)\n\n# Class to convert sparse scipy arrays NP arrays for compatibility \nclass ToDenseArray():\n def __init__(self):\n None\n def fit(self, X, Y=None):\n return self\n def transform(self, X, Y=None):\n return X.toarray()\n\n# Define pipeline to preprocess numeric features:\n # age, time_in_area, num_people_household, num_child_household\nnumeric_pipeline = Pipeline([\n ('reshape', reshaper()),\n ('scale', MinMaxScaler()),\n ('deskew',PowerTransformer()),\n ('predict',LastEstimator())\n ]) # scale and deskew each numeric feature\n\n# Define pipeline to preprocess binary features:\n # sex, married\nbinary_pipeline = Pipeline([\n ('reshape', reshaper()),\n ('OHE', OneHotEncoder(handle_unknown='ignore')),\n ('to_dense',ToDenseArray()),\n ('predict',LastEstimator())\n]) # # Make dummy variables for both possible values of each binary feature\n\n# Define pipeline to preprocess numeric features:\n # dual_income, education, occupation, home_status, home_type, ethnicity, language\ncategorical_pipeline = Pipeline([\n ('reshape', reshaper()),\n ('OHE', OneHotEncoder(handle_unknown='ignore')),\n ('to_dense',ToDenseArray()),\n ('predict',LastEstimator())\n]) # Make dummy variables for each possible value of each categorical feature\n\n# define columnransformer to apply preprocessing steps to each feature\nct = ColumnTransformer([\n ('sex', binary_pipeline, col_ix('sex')),\n ('married', binary_pipeline, col_ix('married')),\n ('dual_income', categorical_pipeline, col_ix('dual_income')),\n ('education', categorical_pipeline, col_ix('education')),\n ('occupation', categorical_pipeline, col_ix('occupation')),\n ('home_status', categorical_pipeline, col_ix('home_status')),\n ('home_type', categorical_pipeline, col_ix('home_type')),\n ('ethnicicty', categorical_pipeline, col_ix('ethnicity')),\n ('language', categorical_pipeline, col_ix('language')),\n ('age', numeric_pipeline, col_ix('age')),\n ('time_in_area', numeric_pipeline, col_ix('time_in_area')),\n ('num_people_household', numeric_pipeline, col_ix('num_people_household')),\n ('num_children_household', numeric_pipeline, col_ix('num_children_household')), \n ])", "_____no_output_____" ], [ "def report_metrics(pipeline_gridsearch):\n estimator = pipeline_gridsearch.best_estimator_\n \n best_params = estimator.named_steps['clf'].best_params_\n best_score_cv = pipeline_gridsearch.best_score_\n score_std = pipeline_gridsearch.cv_results_['std_test_score'][pipeline_gridsearch.best_index_]\n train_score = accuracy_score(estimator.predict(x_train), y_train)\n test_score = accuracy_score(estimator.predict(x_test), y_test)\n \n estimator.fit(x_train, y_train)\n test_score_refit = accuracy_score(estimator.predict(x_test), y_test)\n\n print('{:<40}: {:<50}'.format('Best parameters', repr(best_params)))\n print('{:<40}: {:<50}'.format('Best validation score', '{} +-{}'.format(round(best_score_cv,3),round(score_std,3))))\n print('{:<40}: {:<50}'.format('Train score', round(train_score,3)))\n print('{:<40}: {:<50}'.format('Test score', round(test_score,3)))\n print('{:<40}: {:<50}'.format('Test score (all training data)', test_score_refit))", "_____no_output_____" ] ], [ [ "# Classification", "_____no_output_____" ], [ "### 1. K-Nearest-Neighbors Classifier", "_____no_output_____" ] ], [ [ "# Try odd values between 5 and 29 for N\nKNN_param_grid = {\n 'n_neighbors': [x*2+1 for x in range(2,14)]\n}\n\nKNN_gridsearch = GridSearchCV(KNeighborsClassifier(), KNN_param_grid, cv=3, n_jobs=-1)\n\nKNN_pipeline = Pipeline([\n ('impute', KNNImputer()),\n ('preprocess', ct),\n ('clf', KNN_gridsearch)\n ])\n\nKNN_pipeline_gridsearch = GridSearchCV(KNN_pipeline, param_grid={}, cv=3, n_jobs=-1)\n\nKNN_pipeline_gridsearch.fit(x_train, y_train)", "_____no_output_____" ], [ "print(\"{:^60}\".format(\"KNN CLASSIFIER RESULTS\"))\nreport_metrics(KNN_pipeline_gridsearch)", " KNN CLASSIFIER RESULTS \nBest parameters : {'n_neighbors': 23} \nBest validation score : 0.324 +-0.007 \nTrain score : 0.394 \nTest score : 0.326 \nTest score (all training data) : 0.3257365202890495 \n" ] ], [ [ "### 2a. Logistic Regression - No Regularization", "_____no_output_____" ] ], [ [ "LR_param_grid = {\n}\n\nLR_gridsearch = GridSearchCV(LogisticRegression(), LR_param_grid, cv=3, n_jobs=-1)\n\nLR_pipeline = Pipeline([\n ('impute', KNNImputer()),\n ('preprocess', ct),\n ('clf', LR_gridsearch)\n ])\n\nLR_pipeline_gridsearch = GridSearchCV(LR_pipeline, param_grid={}, cv=3, n_jobs=-1)\nLR_pipeline_gridsearch.fit(x_train,y_train)", "_____no_output_____" ], [ "print(\"{:^60}\".format(\"LOGISTIC REGRESSION RESULTS\"))\nreport_metrics(LR_pipeline_gridsearch)", " LOGISTIC REGRESSION RESULTS \nBest parameters : {} \nBest validation score : 0.334 +-0.004 \nTrain score : 0.362 \nTest score : 0.342 \nTest score (all training data) : 0.3424124513618677 \n" ] ], [ [ "### 2b. Logistic Regression - L1 Regularization", "_____no_output_____" ] ], [ [ "LR1_param_grid = {\n 'penalty': ['l1'],\n 'C':[12,11,10,9,8,7,6]\n}\n\nLR1_gridsearch = GridSearchCV(LogisticRegression(), LR1_param_grid, cv=3, n_jobs=-1)\n\nLR1_pipeline = Pipeline([\n ('impute', KNNImputer()),\n ('preprocess', ct),\n ('clf', LR1_gridsearch)\n ])\n\nLR1_pipeline_gridsearch = GridSearchCV(LR1_pipeline, param_grid={}, cv=3, n_jobs=-1)\nLR1_pipeline_gridsearch.fit(x_train,y_train)", "_____no_output_____" ], [ "print(\"{:^60}\".format(\"LOGISTIC REGRESSION (L1 PENALTY) RESULTS\"))\nreport_metrics(LR1_pipeline_gridsearch)", " LOGISTIC REGRESSION (L1 PENALTY) RESULTS \nBest parameters : {'C': 9, 'penalty': 'l1'} \nBest validation score : 0.334 +-0.004 \nTrain score : 0.363 \nTest score : 0.34 \nTest score (all training data) : 0.33852140077821014 \n" ] ], [ [ "### 2c. Logistic Regression - L2 Regularization", "_____no_output_____" ] ], [ [ "LR2_param_grid = {\n 'penalty': ['l2'],\n 'C':[10, 8, 6, 5, 4, 2]\n}\n\nLR2_gridsearch = GridSearchCV(LogisticRegression(), LR2_param_grid, cv=3, n_jobs=-1)\n\nLR2_pipeline = Pipeline([\n ('impute', KNNImputer()),\n ('preprocess', ct),\n ('clf', LR1_gridsearch)\n ])\n\nLR2_pipeline_gridsearch = GridSearchCV(LR2_pipeline, param_grid={}, cv=3, n_jobs=-1)\nLR2_pipeline_gridsearch.fit(x_train,y_train)", "_____no_output_____" ], [ "print(\"{:^60}\".format(\"LOGISTIC REGRESSION (L2 PENALTY) RESULTS\"))\nreport_metrics(LR2_pipeline_gridsearch)", " LOGISTIC REGRESSION (L2 PENALTY) RESULTS \nBest parameters : {'C': 11, 'penalty': 'l1'} \nBest validation score : 0.335 +-0.004 \nTrain score : 0.363 \nTest score : 0.337 \nTest score (all training data) : 0.33740967204002226 \n" ] ], [ [ "### 3. LogisticIT", "_____no_output_____" ] ], [ [ "LRIT_param_grid = {\n 'alpha':[0.05,0.075,0.1,0.2,0.3,0.4]\n}\n\nLRIT_gridsearch = GridSearchCV(LogisticIT(), LRIT_param_grid, cv=3, n_jobs=-1)\n\nLRIT_pipeline = Pipeline([\n ('impute', KNNImputer()),\n ('preprocess', ct),\n ('clf', LRIT_gridsearch)\n ])\n\nLRIT_pipeline_gridsearch = GridSearchCV(LRIT_pipeline, param_grid={}, cv=3, n_jobs=-1)\nLRIT_pipeline_gridsearch.fit(x_train,y_train)", "_____no_output_____" ], [ "print(\"{:^60}\".format(\"ORDINAL LOGISTIC (IMMEDIATE THRESHOLD) RESULTS\"))\nreport_metrics(LRIT_pipeline_gridsearch)", " ORDINAL LOGISTIC (IMMEDIATE THRESHOLD) RESULTS \nBest parameters : {'alpha': 0.05} \nBest validation score : 0.325 +-0.006 \nTrain score : 0.327 \nTest score : 0.32 \nTest score (all training data) : 0.32017787659811003 \n" ] ], [ [ "### 4. Random Forest Classifier", "_____no_output_____" ] ], [ [ "RF_param_grid = {\n 'n_estimators': [x*100 for x in range(1,10)]\n}\n\nRF_gridsearch = GridSearchCV(RandomForestClassifier(), RF_param_grid, cv=3, n_jobs=-1)\n\nRF_pipeline = Pipeline([\n ('impute', KNNImputer()),\n ('preprocess', ct),\n ('clf', RF_gridsearch)\n ])\n\nRF_pipeline_gridsearch = GridSearchCV(RF_pipeline, param_grid={}, cv=3, n_jobs=-1)\nRF_pipeline_gridsearch.fit(x_train,y_train)", "_____no_output_____" ], [ "print(\"{:^60}\".format(\"RANDOM FOREST RESULTS\"))\nreport_metrics(RF_pipeline_gridsearch)", " RANDOM FOREST RESULTS \nBest parameters : {'n_estimators': 200} \nBest validation score : 0.317 +-0.008 \nTrain score : 0.87 \nTest score : 0.31 \nTest score (all training data) : 0.31017231795441913 \n" ], [ "from xgboost.sklearn import XGBClassifier\n\nXGBC1_param_grid = {\n 'learning_rate':[0.1],\n 'n_estimators':[1000],\n 'max_depth':[5],\n 'min_child_weight':[1],\n 'gamma':[0],\n 'subsample':[0.8],\n 'colsample_bytree':[0.8],\n 'scale_pos_weight':[1],\n 'max_depth': range(3,10,2),\n 'min_child_weight': range(1,6,2)\n}\n\nXGBC1_gridsearch = GridSearchCV(XGBClassifier(), XGBC1_param_grid, cv=3, n_jobs=-1)\n\nXGBC1_pipeline = Pipeline([\n ('impute', KNNImputer()),\n ('preprocess', ct),\n ('clf', XGBC1_gridsearch)\n ])\n\nXGBC1_pipeline_gridsearch = GridSearchCV(XGBC1_pipeline, param_grid={}, cv=3, n_jobs=-1)\nXGBC1_pipeline_gridsearch.fit(x_train,y_train)", "_____no_output_____" ], [ "print(\"{:^60}\".format(\"BOOSTED FOREST CLASSIFIER #1 RESULTS\"))\nreport_metrics(XGBC1_pipeline_gridsearch)", " BOOSTED FOREST CLASSIFIER #1 RESULTS \nBest parameters : {'colsample_bytree': 0.8, 'gamma': 0, 'learning_rate': 0.1, 'max_depth': 3, 'min_child_weight': 3, 'n_estimators': 1000, 'scale_pos_weight': 1, 'subsample': 0.8}\nBest validation score : 0.329 +-0.007 \nTrain score : 0.557 \nTest score : 0.338 \nTest score (all training data) : 0.3379655364091162 \n" ] ] ]
[ "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code" ]
[ [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown", "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code" ] ]
ecea6abc41fe95a0740e80d80b4d27de6f532585
3,491
ipynb
Jupyter Notebook
diffprog/julia_dp/repository/tf2_2-Copy1.2.ipynb
jskDr/keraspp_2021
dc46ebb4f4dea48612135136c9837da7c246534a
[ "MIT" ]
4
2021-09-21T15:35:04.000Z
2021-12-14T12:14:44.000Z
diffprog/julia_dp/repository/tf2_2-Copy1.2.ipynb
jskDr/keraspp_2021
dc46ebb4f4dea48612135136c9837da7c246534a
[ "MIT" ]
null
null
null
diffprog/julia_dp/repository/tf2_2-Copy1.2.ipynb
jskDr/keraspp_2021
dc46ebb4f4dea48612135136c9837da7c246534a
[ "MIT" ]
null
null
null
21.549383
68
0.454025
[ [ [ "## Math based differentiation", "_____no_output_____" ] ], [ [ "f = lambda x, n: x**n \ndf_th = lambda x, n: n*f(x,n-1)\n\ndef df_ad(x, n):\n x_tf = tf.Variable(x)\n with tf.GradientTape() as tape:\n y = f(x_tf,n)\n return tape.gradient(y, x_tf)\n\nx = float(2.0)\nprint(\"f(x,10) = 2^10 = \", f(x,10))\nprint(\"Theory: df(x,10)/dx = 10 x 2^9 = \", df_th(x,10))\nprint(\"AD: df(x,10) = \", df_ad(x,10))", "f(x,10) = 2^10 = 1024.0\nTheory: df(x,10)/dx = 10 x 2^9 = 5120.0\nAD: df(x,10) = tf.Tensor(5120.0, shape=(), dtype=float32)\n" ], [ "x_all = np.random.randn(1000)\n%time y_all = list(map(lambda x: df_ad(x,10), x_all))", "CPU times: user 511 ms, sys: 2.65 ms, total: 514 ms\nWall time: 514 ms\n" ] ], [ [ "## Code based differentiation", "_____no_output_____" ] ], [ [ "def f(x,n): \n r = 1\n for m in range(n):\n r *= x\n return r\ndf_th = lambda x, n: n*f(x,n-1)\n\ndef df_ad(x, n):\n x_tf = tf.Variable(x)\n with tf.GradientTape() as tape:\n y = f(x_tf,n)\n return tape.gradient(y, x_tf)\n\nx = float(2.0)\nprint(\"f(x,10) = 2^10 = \", f(x,10))\nprint(\"Theory: df(x,10)/dx = 10 x 2^9 = \", df_th(x,10))\nprint(\"AD: df(x,10) = \", df_ad(x,10))", "f(x,10) = 2^10 = 1024.0\nTheory: df(x,10)/dx = 10 x 2^9 = 5120.0\nAD: df(x,10) = tf.Tensor(5120.0, shape=(), dtype=float32)\n" ], [ "x_all = np.random.randn(1000)\n%time y_all = list(map(lambda x: df_ad(x,10), x_all))\nlen(y_all)", "CPU times: user 2.24 s, sys: 5.18 ms, total: 2.25 s\nWall time: 2.25 s\n" ] ] ]
[ "markdown", "code", "markdown", "code" ]
[ [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code" ] ]
ecea6efd439a806c70ede3438ca7043005888d28
5,531
ipynb
Jupyter Notebook
1-tutorials/1-getting-started.ipynb
anitagraser/movingpandas-examples
093d87a89f197f322f250d0a308cc8e07b55b525
[ "BSD-3-Clause" ]
77
2020-10-11T20:45:19.000Z
2022-03-29T16:06:58.000Z
1-tutorials/1-getting-started.ipynb
anitagraser/movingpandas-examples
093d87a89f197f322f250d0a308cc8e07b55b525
[ "BSD-3-Clause" ]
8
2021-01-20T15:33:44.000Z
2022-03-24T18:44:18.000Z
1-tutorials/1-getting-started.ipynb
anitagraser/movingpandas-examples
093d87a89f197f322f250d0a308cc8e07b55b525
[ "BSD-3-Clause" ]
18
2021-02-07T14:54:04.000Z
2022-03-29T16:06:52.000Z
24.473451
252
0.578919
[ [ [ "# Getting started with MovingPandas\n\n<img align=\"right\" src=\"https://anitagraser.github.io/movingpandas/pics/movingpandas.png\">\n\n[![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/anitagraser/movingpandas-examples/main?filepath=1-tutorials/1-getting-started.ipynb)\n\nMovingPandas provides a trajectory datatype based on GeoPandas.\nThe project home is at https://github.com/anitagraser/movingpandas\n\nThe documentation is available at https://movingpandas.readthedocs.io/en/master/", "_____no_output_____" ] ], [ [ "import pandas as pd\nimport geopandas as gpd\nfrom geopandas import GeoDataFrame, read_file\nfrom shapely.geometry import Point, LineString, Polygon\nfrom datetime import datetime, timedelta\nimport movingpandas as mpd\n\nprint(mpd.__version__)", "_____no_output_____" ] ], [ [ "## Creating a trajectory from scratch\n\nTrajectory objects consist of a trajectory ID and a GeoPandas GeoDataFrame with a DatetimeIndex. The data frame therefore represents the trajectory data as a Pandas time series with associated point locations (and optional further attributes).\n\nLet's create a small toy trajectory to see how this works:", "_____no_output_____" ] ], [ [ "df = pd.DataFrame([\n {'geometry':Point(0,0), 't':datetime(2018,1,1,12,0,0)},\n {'geometry':Point(6,0), 't':datetime(2018,1,1,12,6,0)},\n {'geometry':Point(6,6), 't':datetime(2018,1,1,12,10,0)},\n {'geometry':Point(9,9), 't':datetime(2018,1,1,12,15,0)}\n]).set_index('t')\ngdf = GeoDataFrame(df, crs=31256)\ntoy_traj = mpd.Trajectory(gdf, 1)\ntoy_traj", "_____no_output_____" ], [ "toy_traj.plot()", "_____no_output_____" ] ], [ [ "We can also access the trajectory's GeoDataFrame:", "_____no_output_____" ] ], [ [ "toy_traj.df", "_____no_output_____" ], [ "toy_traj.df.plot()", "_____no_output_____" ] ], [ [ "## Loading trajectory data from a GeoPackage\n\nThe MovingPandas repository contains a demo GeoPackage file that can be loaded as follows:", "_____no_output_____" ] ], [ [ "%%time\ngdf = read_file('../data/geolife_small.gpkg')\ngdf['t'] = pd.to_datetime(gdf['t'])\ngdf = gdf.set_index('t').tz_localize(None)\nprint(\"Finished reading {} rows\".format(len(df)))", "_____no_output_____" ] ], [ [ "After reading the trajectory point data from file, we want to construct the trajectories.", "_____no_output_____" ], [ "### Creating trajectories with TrajectoryCollection\n\nTrajectoryCollection is a convenience class that takes care of creating trajectories from a GeoDataFrame:", "_____no_output_____" ] ], [ [ "traj_collection = mpd.TrajectoryCollection(gdf, 'trajectory_id')\nprint(traj_collection)", "_____no_output_____" ], [ "traj_collection.plot(column='trajectory_id', legend=True, figsize=(9,5))", "_____no_output_____" ] ], [ [ "### Accessing individual trajectories", "_____no_output_____" ] ], [ [ "my_traj = traj_collection.trajectories[1]\nprint(my_traj)", "_____no_output_____" ], [ "my_traj.plot(linewidth=5, capstyle='round', figsize=(9,3))", "_____no_output_____" ] ], [ [ "To visualize trajectories in their geographical context, we can also create interactive plots with basemaps:", "_____no_output_____" ] ], [ [ "my_traj.hvplot(width=500, height=300, line_width=7.0, tiles='StamenTonerBackground')", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code" ]
[ [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ] ]
ecea73d14c11330327b0ad3d22eea8c702506053
1,015,013
ipynb
Jupyter Notebook
.ipynb_checkpoints/Baseball_Stats-checkpoint.ipynb
bibinmjose/baseball_d3
d104e640cf84f0069eeb81ff2fe0c964b8118d64
[ "MIT" ]
null
null
null
.ipynb_checkpoints/Baseball_Stats-checkpoint.ipynb
bibinmjose/baseball_d3
d104e640cf84f0069eeb81ff2fe0c964b8118d64
[ "MIT" ]
null
null
null
.ipynb_checkpoints/Baseball_Stats-checkpoint.ipynb
bibinmjose/baseball_d3
d104e640cf84f0069eeb81ff2fe0c964b8118d64
[ "MIT" ]
null
null
null
648.985294
209,166
0.93091
[ [ [ "# Dataset Exploration\n\nThe dataset contains career statistics of baseball players. Each dataum has the name of player(`name`), `handednes`(**L** for Left and **R** for Right), `height` (in inchs) amd `weight` (in lbs), `avg`(Career average) and `HR` (Total Home runs). Body metrics include `height` and `weight` while perfomance is given by `HR` and `avg`.\n\n## Questions to explore:\n* Is there a correlation between body metrics and performance(avg or HR)?\n* Can you identify high avg/HR players from body metrics ? \n", "_____no_output_____" ], [ "# Data loading and summary", "_____no_output_____" ] ], [ [ "import csv\nimport pandas as pd\nimport numpy as np\nimport seaborn as sns\nimport matplotlib.pyplot as plt\nplt.style.use('ggplot')", "_____no_output_____" ], [ "baseball_lst=[]\nwith open(\"baseball_data.csv\") as f:\n records = csv.DictReader(f)\n for row in records:\n baseball_lst.append(row)\nbaseball = pd.read_csv('baseball_data.csv')\n\n# Checking for missing values in each column\nfor key in baseball.keys():\n if baseball[key].isnull().sum():\n print (key)", "_____no_output_____" ], [ "baseball.head(5)\n# 'HR' is home runs", "_____no_output_____" ], [ "baseball.describe()", "_____no_output_____" ], [ "# Players with top 5 averages\nbaseball.sort_values(by ='avg', ascending=False)[0:5]", "_____no_output_____" ], [ "# top 5 Home Run getters\nbaseball.sort_values(by ='HR', ascending=False)[0:5]", "_____no_output_____" ], [ "# Players with top 5 weights\nbaseball.sort_values(by ='weight', ascending=False)[0:5]", "_____no_output_____" ] ], [ [ "# Univariate Exploration and grid plots\n\nHeight in inchs and Weight in lbs.", "_____no_output_____" ] ], [ [ "# Histograms of columns.\nfig = plt.figure(figsize=(8,5))\nfor i,key in enumerate(baseball.keys()[2:]):\n ax = fig.add_subplot(2,2,i+1)\n ax.set_title(key)\n baseball[key].hist(bins=10)\nplt.tight_layout()\nplt.show()", "_____no_output_____" ], [ "# count of players based on handedness\nbaseball['handedness'].value_counts().plot.barh()\nplt.show()", "_____no_output_____" ], [ "bball_B = baseball[baseball['handedness']=='B']\nbball_R = baseball[baseball['handedness']=='R']\nbball_L = baseball[baseball['handedness']=='L']\n# plt.show()", "_____no_output_____" ] ], [ [ "# Paired Plots and Bivariate Exploration of dataset", "_____no_output_____" ] ], [ [ "# Paired Plots\n# baseball[(baseball[\"avg\"]==0)&(baseball[\"HR\"]==0)]\ng = sns.PairGrid(baseball)\ng = g.map_diag(plt.hist)\ng = g.map_offdiag(plt.scatter)\nplt.show()", "_____no_output_____" ], [ "# Paired Plots by group\ng = sns.PairGrid(baseball, hue = \"handedness\", palette='husl')\ng = g.map_diag(plt.hist,histtype=\"step\", linewidth=1)\ng = g.map_offdiag(plt.scatter)\nplt.legend()\nplt.show()", "_____no_output_____" ] ], [ [ "\n## Based on `Handedness`\n\nMedian `Height` and `weight` do not vary significantly based on handedness of the player. `avg` and `HR` is slightly higher for Left handed players.\n\n### Height", "_____no_output_____" ] ], [ [ "# Height based on handedness\nsns.boxplot(y = 'height', x='handedness', data = baseball)\nsns.stripplot(y = 'height', x='handedness', data = baseball, color='gray', jitter=True)\nplt.show()", "_____no_output_____" ] ], [ [ "### Weight", "_____no_output_____" ] ], [ [ "# Weight based on handedness\nplt.figure()\nsns.boxplot(y = 'weight', x='handedness', data = baseball)\nplt.show()", "_____no_output_____" ] ], [ [ "### Avg", "_____no_output_____" ] ], [ [ "# avg based on handedness\nsns.boxplot(y = 'avg', x='handedness', data = baseball)\nplt.show()", "_____no_output_____" ], [ "# avg based on handedness with avg>0\nsns.lvplot(y = 'avg', x='handedness', data = baseball[baseball['avg']!=0])\nplt.show()", "_____no_output_____" ] ], [ [ "### HR", "_____no_output_____" ] ], [ [ "# HR based on handedness\nplt.figure()\nsns.boxplot(y = 'HR', x='handedness', data = baseball)\nplt.show()", "_____no_output_____" ], [ "# HR based on handedness with HR >0\nplt.figure()\nsns.lvplot(y = 'HR', x='handedness', \\\n data = baseball[baseball['HR']!=0])\nplt.show()", "_____no_output_____" ], [ "bball_median=baseball.groupby('handedness').median()\nbball_median[\"Count\"]=baseball.groupby('handedness')[\"name\"].count()", "_____no_output_____" ], [ "bball_median.to_csv(\"baseball_median.csv\")\nbball_median", "_____no_output_____" ] ], [ [ "## Based on `weight`\n\nExploring the difference in average and home run based on three levels of weight - Low(`L`) Medium(`M`) and High(`H`).", "_____no_output_____" ] ], [ [ "pd.qcut(baseball['weight'], q=3, labels=['Low','Medium','High'])\\\n .cat.reorder_categories(['Low','Medium','High'],ordered=True)\\\n .value_counts()", "_____no_output_____" ], [ "# Defining a new dataframe with height and weight categorized\nbaseball_new = baseball.copy(deep=True)\nbaseball_new['weight_cat'] = \\\n pd.qcut(baseball_new['weight'], q=3, labels=['Low','Medium','High'])\\\n .cat.reorder_categories(['Low','Medium','High'],ordered=True)\nbaseball_new['height_cat'] = \\\n pd.qcut(baseball_new['height'], q=3, labels=['Low','Medium','High'])\\\n .cat.reorder_categories(['Low','Medium','High'],ordered=True)\nbaseball_new.head(5)", "_____no_output_____" ], [ "# Mean avg based on weight category\nbaseball_new.groupby('height_cat').mean()[['avg']].plot(kind='bar')\nplt.show()", "_____no_output_____" ], [ "# Mean HR based on weight category\nbaseball_new.groupby('weight_cat').mean()[['HR']].plot(kind='bar')\nplt.show()", "_____no_output_____" ], [ "# Paired Plots by group\ng = sns.PairGrid(baseball_new, hue = \"weight_cat\")\ng = g.map_diag(plt.hist,bins=20)\ng = g.map_offdiag(plt.scatter,alpha=0.4)\n# g = g.map_lower(plt.hexbin,gridsize=100)\nplt.legend()\nplt.show()", "_____no_output_____" ], [ "# Seperating into 10 Categories\nbaseball_new['weight_10cat'] = pd.qcut(baseball['weight'], q=10)\nbaseball_new.groupby('weight_10cat').median()[['HR']].plot(kind='bar')\nplt.show()\nbaseball_new.groupby('weight_10cat').median()[['avg']].plot(kind='bar')\nplt.show()", "_____no_output_____" ] ], [ [ "### `HR` vs. `avg` filtered by `weight_cat`", "_____no_output_____" ] ], [ [ "print(baseball_new.groupby('weight_cat').median())\nbaseball_new.groupby('weight_cat').mean().plot(x='avg',y='HR',kind='scatter', legend=True)\nplt.show()", " height weight avg HR\nweight_cat \nLow 71.0 170.0 0.2440 13.0\nMedium 73.0 185.0 0.2365 17.0\nHigh 74.0 200.0 0.2270 19.5\n" ] ], [ [ "## Based on `height`", "_____no_output_____" ] ], [ [ "bb_height_avg", "_____no_output_____" ], [ "baseball_new.groupby('height_cat').mean().plot(x='avg',y='HR',\\\n c='height',kind='scatter', legend=True)\nplt.show()", "_____no_output_____" ], [ "# Paired Plots by group\ng = sns.PairGrid(baseball_new, hue = \"height_cat\")\ng = g.map_diag(plt.hist,bins=20)\ng = g.map_offdiag(plt.scatter,alpha=0.4)\nplt.legend()\nplt.show()", "_____no_output_____" ] ], [ [ "## Home Run vs. Average", "_____no_output_____" ] ], [ [ "def plot_bball_cat(bball,_cat):\n COLORMAP ='rgbyo'\n for i,l in enumerate([\"Low\",\"Medium\",\"High\"]):\n d=bball[bball[_cat]==l]\n# print(bball)\n plt.scatter(x = d[\"avg\"],y=d[\"HR\"],alpha=0.4, label = l, color = COLORMAP[i])\n plt.scatter(x = d[\"avg\"].median(),y=d[\"HR\"].median(),s=5000,\\\n color = COLORMAP[i], marker ='s')\n \n plt.xlabel(\"Average\")\n plt.ylabel(\"Home Runs\")\n plt.legend()\n plt.show()\nplot_bball_cat(baseball_new,'height_cat')", "_____no_output_____" ], [ "plot_bball_cat(baseball_new,'weight_cat')", "_____no_output_____" ] ], [ [ "## Dumping Data Frame to csv", "_____no_output_____" ] ], [ [ "baseball_new.to_csv('baseball_data_modified')", "_____no_output_____" ], [ "baseball_new['height_10cat'] = pd.qcut(baseball['height'], q=5)\nbaseball_new.groupby('height_10cat').mean()[['HR']].plot(kind='bar')\nplt.show()", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code" ]
[ [ "markdown", "markdown" ], [ "code", "code", "code", "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code" ] ]
ecea78c8b25c732708fd9d0df67194de54612b96
1,187
ipynb
Jupyter Notebook
3. Functions exercise.ipynb
muticodes/Python-for-DS-and-ML
d7289bb2d82df1cdcd5188e4e5ba8fd08de6f08e
[ "MIT" ]
null
null
null
3. Functions exercise.ipynb
muticodes/Python-for-DS-and-ML
d7289bb2d82df1cdcd5188e4e5ba8fd08de6f08e
[ "MIT" ]
null
null
null
3. Functions exercise.ipynb
muticodes/Python-for-DS-and-ML
d7289bb2d82df1cdcd5188e4e5ba8fd08de6f08e
[ "MIT" ]
null
null
null
20.465517
177
0.526537
[ [ [ "## LESSER OF TWO EVENS: Write a function that returns the lesser of two given numbers *if* both numbers are even, but returns the greater if one or both numbers are odd ##", "_____no_output_____" ] ], [ [ "def lesser_of_two_evens (arg1,arg2):\n if arg1%2 == 0 and arg2%2 == 0:\n print (min(arg1,arg2))\n else: \n print (max(arg1,arg2))", "_____no_output_____" ] ] ]
[ "markdown", "code" ]
[ [ "markdown" ], [ "code" ] ]
ecea80eef150cffd5ad12bc40dcd0fdb44fc1640
37,853
ipynb
Jupyter Notebook
notebooks/3-fundamentals/PY0101EN-3-1-Conditions.ipynb
raphtrajano/introduction-to-python
be9b1f4cfea185ddad26686804b7607b9e12b7a4
[ "MIT" ]
8
2020-04-02T02:10:47.000Z
2021-09-08T11:50:53.000Z
notebooks/3-fundamentals/PY0101EN-3-1-Conditions.ipynb
raphtrajano/introduction-to-python
be9b1f4cfea185ddad26686804b7607b9e12b7a4
[ "MIT" ]
null
null
null
notebooks/3-fundamentals/PY0101EN-3-1-Conditions.ipynb
raphtrajano/introduction-to-python
be9b1f4cfea185ddad26686804b7607b9e12b7a4
[ "MIT" ]
19
2020-01-16T18:33:42.000Z
2022-02-15T22:24:45.000Z
30.749797
771
0.50638
[ [ [ "<a href=\"https://cognitiveclass.ai/\">\n <img src=\"https://s3-api.us-geo.objectstorage.softlayer.net/cf-courses-data/CognitiveClass/PY0101EN/Ad/CCLog.png\" width=\"200\" align=\"center\">\n</a>", "_____no_output_____" ], [ "<h1>Conditions in Python</h1>", "_____no_output_____" ], [ "<p><strong>Welcome!</strong> This notebook will teach you about the condition statements in the Python Programming Language. By the end of this lab, you'll know how to use the condition statements in Python, including operators, and branching.</p>", "_____no_output_____" ], [ "<h2>Table of Contents</h2>\n<div class=\"alert alert-block alert-info\" style=\"margin-top: 20px\">\n <ul>\n <li>\n <a href=\"#cond\">Condition Statements</a>\n <ul>\n <li><a href=\"comp\">Comparison Operators</a></li>\n <li><a href=\"branch\">Branching</a></li>\n <li><a href=\"logic\">Logical operators</a></li>\n </ul>\n </li>\n <li>\n <a href=\"#quiz\">Quiz on Condition Statement</a>\n </li>\n </ul>\n <p>\n Estimated time needed: <strong>20 min</strong>\n </p>\n</div>\n\n<hr>", "_____no_output_____" ], [ "<h2 id=\"cond\">Condition Statements</h2>", "_____no_output_____" ], [ "<h3 id=\"comp\">Comparison Operators</h3>", "_____no_output_____" ], [ "Comparison operations compare some value or operand and, based on a condition, they produce a Boolean. When comparing two values you can use these operators:\n\n<ul>\n <li>equal: <b>==</b></li>\n <li>not equal: <b>!=</b></li>\n <li>greater than: <b>></b></li>\n <li>less than: <b>&lt;</b></li>\n <li>greater than or equal to: <b>>=</b></li>\n <li>less than or equal to: <b>&lt;=</b></li>\n</ul>", "_____no_output_____" ], [ "Let's assign <code>a</code> a value of 5. Use the equality operator denoted with two equal <b>==</b> signs to determine if two values are equal. The case below compares the variable <code>a</code> with 6.", "_____no_output_____" ] ], [ [ "# Condition Equal\n\na = 5\na == 6", "_____no_output_____" ] ], [ [ "The result is <b>False</b>, as 5 does not equal to 6.", "_____no_output_____" ], [ "Consider the following equality comparison operator <code>i > 5</code>. If the value of the left operand, in this case the variable <b>i</b>, is greater than the value of the right operand, in this case 5, then the statement is <b>True</b>. Otherwise, the statement is <b>False</b>. If <b>i</b> is equal to 6, because 6 is larger than 5, the output is <b>True</b>.", "_____no_output_____" ] ], [ [ "# Greater than Sign\n\ni = 6\ni > 5", "_____no_output_____" ] ], [ [ "Set <code>i = 2</code>. The statement is false as 2 is not greater than 5:", "_____no_output_____" ] ], [ [ "# Greater than Sign\n\ni = 2\ni > 5", "_____no_output_____" ] ], [ [ " Let's display some values for <code>i</code> in the figure. Set the values greater than 5 in green and the rest in red. The green region represents where the condition is **True**, the red where the statement is **False**. If the value of <code>i</code> is 2, we get **False** as the 2 falls in the red region. Similarly, if the value for <code>i</code> is 6 we get a **True** as the condition falls in the green region. ", "_____no_output_____" ], [ "<img src=\"https://s3-api.us-geo.objectstorage.softlayer.net/cf-courses-data/CognitiveClass/PY0101EN/Chapter%203/Images/CondsGreater.gif\" width=\"650\" />", "_____no_output_____" ], [ "The inequality test uses an exclamation mark preceding the equal sign, if two operands are not equal then the condition becomes **True**. For example, the following condition will produce **True** as long as the value of <code>i</code> is not equal to 6:", "_____no_output_____" ] ], [ [ "# Inequality Sign\n\ni = 2\ni != 6", "_____no_output_____" ] ], [ [ "When <code>i</code> equals 6 the inequality expression produces <b>False</b>. ", "_____no_output_____" ] ], [ [ "# Inequality Sign\n\ni = 6\ni != 6", "_____no_output_____" ] ], [ [ "See the number line below. when the condition is **True** the corresponding numbers are marked in green and for where the condition is **False** the corresponding number is marked in red. If we set <code>i</code> equal to 2 the operator is true as 2 is in the green region. If we set <code>i</code> equal to 6, we get a **False** as the condition falls in the red region.", "_____no_output_____" ], [ "<img src=\"https://s3-api.us-geo.objectstorage.softlayer.net/cf-courses-data/CognitiveClass/PY0101EN/Chapter%203/Images/CondsIneq.gif\" width=\"650\" />", "_____no_output_____" ], [ " We can apply the same methods on strings. For example, use an equality operator on two different strings. As the strings are not equal, we get a **False**.", "_____no_output_____" ] ], [ [ "# Use Equality sign to compare the strings\n\n\"ACDC\" == \"Michael Jackson\"", "_____no_output_____" ] ], [ [ " If we use the inequality operator, the output is going to be **True** as the strings are not equal.", "_____no_output_____" ] ], [ [ "# Use Inequality sign to compare the strings\n\n\"ACDC\" != \"Michael Jackson\"", "_____no_output_____" ] ], [ [ "Inequality operation is also used to compare the letters/words/symbols according to the ASCII value of letters. The decimal value shown in the following table represents the order of the character:\n", "_____no_output_____" ], [ "For example, the ASCII code for <b>!</b> is 33, while the ASCII code for <b>+</b> is 43. Therefore <b>+</b> is larger than <b>!</b> as 43 is greater than 33.", "_____no_output_____" ], [ "Similarly, the value for <b>A</b> is 65, and the value for <b>B</b> is 66 therefore:", "_____no_output_____" ] ], [ [ "# Compare characters\n\n'B' > 'A'", "_____no_output_____" ] ], [ [ " When there are multiple letters, the first letter takes precedence in ordering:", "_____no_output_____" ] ], [ [ "# Compare characters\n\n'BA' > 'AB'", "_____no_output_____" ] ], [ [ "<b>Note</b>: Upper Case Letters have different ASCII code than Lower Case Letters, which means the comparison between the letters in python is case-sensitive.", "_____no_output_____" ], [ "<h3 id=\"branch\">Branching</h3>", "_____no_output_____" ], [ " Branching allows us to run different statements for different inputs. It is helpful to think of an **if statement** as a locked room, if the statement is **True** we can enter the room and your program will run some predefined tasks, but if the statement is **False** the program will ignore the task.\n", "_____no_output_____" ], [ "For example, consider the blue rectangle representing an ACDC concert. If the individual is older than 18, they can enter the ACDC concert. If they are 18 or younger than 18 they cannot enter the concert.\n\nUse the condition statements learned before as the conditions need to be checked in the **if statement**. The syntax is as simple as <code> if <i>condition statement</i> :</code>, which contains a word <code>if</code>, any condition statement, and a colon at the end. Start your tasks which need to be executed under this condition in a new line with an indent. The lines of code after the colon and with an indent will only be executed when the **if statement** is **True**. The tasks will end when the line of code does not contain the indent.\n\nIn the case below, the tasks executed <code>print(“you can enter”)</code> only occurs if the variable <code>age</code> is greater than 18 is a True case because this line of code has the indent. However, the execution of <code>print(“move on”)</code> will not be influenced by the if statement.", "_____no_output_____" ] ], [ [ "# If statement example\n\nage = 19\n#age = 18\n\n#expression that can be true or false\nif age > 18:\n \n #within an indent, we have the expression that is run if the condition is true\n print(\"you can enter\" )\n\n#The statements after the if statement will run regardless if the condition is true or false \nprint(\"move on\")", "_____no_output_____" ] ], [ [ "<i>Try uncommenting the age variable</i>", "_____no_output_____" ], [ "It is helpful to use the following diagram to illustrate the process. On the left side, we see what happens when the condition is <b>True</b>. The person enters the ACDC concert representing the code in the indent being executed; they then move on. On the right side, we see what happens when the condition is <b>False</b>; the person is not granted access, and the person moves on. In this case, the segment of code in the indent does not run, but the rest of the statements are run. ", "_____no_output_____" ], [ "<img src=\"https://s3-api.us-geo.objectstorage.softlayer.net/cf-courses-data/CognitiveClass/PY0101EN/Chapter%203/Images/CondsIf.gif\" width=\"650\" />", "_____no_output_____" ], [ "The <code>else</code> statement runs a block of code if none of the conditions are **True** before this <code>else</code> statement. Let's use the ACDC concert analogy again. If the user is 17 they cannot go to the ACDC concert, but they can go to the Meatloaf concert.\nThe syntax of the <code>else</code> statement is similar as the syntax of the <code>if</code> statement, as <code>else :</code>. Notice that, there is no condition statement for <code>else</code>.\nTry changing the values of <code>age</code> to see what happens: ", "_____no_output_____" ] ], [ [ "# Else statement example\n\nage = 18\n# age = 19\n\nif age > 18:\n print(\"you can enter\" )\nelse:\n print(\"go see Meat Loaf\" )\n \nprint(\"move on\")", "_____no_output_____" ] ], [ [ "The process is demonstrated below, where each of the possibilities is illustrated on each side of the image. On the left is the case where the age is 17, we set the variable age to 17, and this corresponds to the individual attending the Meatloaf concert. The right portion shows what happens when the individual is over 18, in this case 19, and the individual is granted access to the concert.", "_____no_output_____" ], [ "<img src=\"https://s3-api.us-geo.objectstorage.softlayer.net/cf-courses-data/CognitiveClass/PY0101EN/Chapter%203/Images/CondsElse.gif\" width=\"650\" />", "_____no_output_____" ], [ "The <code>elif</code> statement, short for else if, allows us to check additional conditions if the condition statements before it are <b>False</b>. If the condition for the <code>elif</code> statement is <b>True</b>, the alternate expressions will be run. Consider the concert example, where if the individual is 18 they will go to the Pink Floyd concert instead of attending the ACDC or Meat-loaf concert. The person of 18 years of age enters the area, and as they are not older than 18 they can not see ACDC, but as they are 18 years of age, they attend Pink Floyd. After seeing Pink Floyd, they move on. The syntax of the <code>elif</code> statement is similar in that we merely change the <code>if</code> in <code>if</code> statement to <code>elif</code>.", "_____no_output_____" ] ], [ [ "# Elif statment example\n\nage = 18\n\nif age > 18:\n print(\"you can enter\" )\nelif age == 18:\n print(\"go see Pink Floyd\")\nelse:\n print(\"go see Meat Loaf\" )\n \nprint(\"move on\")", "_____no_output_____" ] ], [ [ "The three combinations are shown in the figure below. The left-most region shows what happens when the individual is less than 18 years of age. The central component shows when the individual is exactly 18. The rightmost shows when the individual is over 18.", "_____no_output_____" ], [ "<img src =\"https://s3-api.us-geo.objectstorage.softlayer.net/cf-courses-data/CognitiveClass/PY0101EN/Chapter%203/Images/CondsElif.gif\" width=\"650\" />", "_____no_output_____" ], [ " Look at the following code:\n", "_____no_output_____" ] ], [ [ "# Condition statement example\n\nalbum_year = 1983\nalbum_year = 1970\n\nif album_year > 1980:\n print(\"Album year is greater than 1980\")\n \nprint('do something..')", "_____no_output_____" ] ], [ [ "Feel free to change <code>album_year</code> value to other values -- you'll see that the result changes!", "_____no_output_____" ], [ "Notice that the code in the above <b>indented</b> block will only be executed if the results are <b>True</b>. ", "_____no_output_____" ], [ "As before, we can add an <code>else</code> block to the <code>if</code> block. The code in the <code>else</code> block will only be executed if the result is <b>False</b>.\n\n\n<b>Syntax:</b> \n\nif (condition):\n # do something\nelse:\n # do something else", "_____no_output_____" ], [ "If the condition in the <code>if</code> statement is <b>False</b>, the statement after the <code>else</code> block will execute. This is demonstrated in the figure: ", "_____no_output_____" ], [ "<img src=\"https://s3-api.us-geo.objectstorage.softlayer.net/cf-courses-data/CognitiveClass/PY0101EN/Chapter%203/Images/CondsLogicMap.png\" width=\"650\" />", "_____no_output_____" ] ], [ [ "# Condition statement example\n\nalbum_year = 1983\n#album_year = 1970\n\nif album_year > 1980:\n print(\"Album year is greater than 1980\")\nelse:\n print(\"less than 1980\")\n\nprint('do something..')", "_____no_output_____" ] ], [ [ "Feel free to change the <code>album_year</code> value to other values -- you'll see that the result changes based on it!", "_____no_output_____" ], [ "<h3 id=\"logic\">Logical operators</h3>", "_____no_output_____" ], [ "\nSometimes you want to check more than one condition at once. For example, you might want to check if one condition and another condition is **True**. Logical operators allow you to combine or modify conditions.\n<ul>\n <li><code>and</code></li>\n <li><code>or</code></li>\n <li><code>not</code></li>\n</ul>\n\nThese operators are summarized for two variables using the following truth tables: ", "_____no_output_____" ], [ "<img src=\"https://s3-api.us-geo.objectstorage.softlayer.net/cf-courses-data/CognitiveClass/PY0101EN/Chapter%203/Images/CondsTable.png\" width=\"650\" />", "_____no_output_____" ], [ "The <code>and</code> statement is only **True** when both conditions are true. The <code>or</code> statement is true if one condition is **True**. The <code>not</code> statement outputs the opposite truth value.", "_____no_output_____" ], [ "Let's see how to determine if an album was released after 1979 (1979 is not included) and before 1990 (1990 is not included). The time periods between 1980 and 1989 satisfy this condition. This is demonstrated in the figure below. The green on lines <strong>a</strong> and <strong>b</strong> represents periods where the statement is **True**. The green on line <strong>c</strong> represents where both conditions are **True**, this corresponds to where the green regions overlap. \n\n", "_____no_output_____" ], [ "<img src=\"https://s3-api.us-geo.objectstorage.softlayer.net/cf-courses-data/CognitiveClass/PY0101EN/Chapter%203/Images/CondsEgOne.png\" width=\"650\" />", "_____no_output_____" ], [ " The block of code to perform this check is given by:", "_____no_output_____" ] ], [ [ "# Condition statement example\n\nalbum_year = 1980\n\nif(album_year > 1979) and (album_year < 1990):\n print (\"Album year was in between 1980 and 1989\")\n \nprint(\"\")\nprint(\"Do Stuff..\")", "_____no_output_____" ] ], [ [ "To determine if an album was released before 1980 (~ - 1979) or after 1989 (1990 - ~ ), an **or** statement can be used. Periods before 1980 (~ - 1979) or after 1989 (1990 - ~) satisfy this condition. This is demonstrated in the following figure, the color green in <strong>a</strong> and <strong>b</strong> represents periods where the statement is true. The color green in **c** represents where at least one of the conditions \nare true. \n", "_____no_output_____" ], [ "<img src=\"https://s3-api.us-geo.objectstorage.softlayer.net/cf-courses-data/CognitiveClass/PY0101EN/Chapter%203/Images/CondsEgTwo.png\" width=\"650\" />", "_____no_output_____" ], [ "The block of code to perform this check is given by:", "_____no_output_____" ] ], [ [ "# Condition statement example\n\nalbum_year = 1990\n\nif(album_year < 1980) or (album_year > 1989):\n print (\"Album was not made in the 1980's\")\nelse:\n print(\"The Album was made in the 1980's \")", "_____no_output_____" ] ], [ [ "The <code>not</code> statement checks if the statement is false:", "_____no_output_____" ] ], [ [ "# Condition statement example\n\nalbum_year = 1983\n\nif not (album_year == '1984'):\n print (\"Album year is not 1984\")", "_____no_output_____" ] ], [ [ "<hr>", "_____no_output_____" ], [ "<h2 id=\"quiz\">Quiz on Conditions</h2>", "_____no_output_____" ], [ "Write an if statement to determine if an album had a rating greater than 8. Test it using the rating for the album <b>“Back in Black”</b> that had a rating of 8.5. If the statement is true print \"This album is Amazing!\"", "_____no_output_____" ] ], [ [ "# Write your code below and press Shift+Enter to execute", "This album is Amazing!\n" ] ], [ [ "Double-click __here__ for the solution.\n\n<!-- \nrating = 8.5\nif rating > 8:\n print (\"This album is Amazing!\")\n -->", "_____no_output_____" ], [ "<hr>", "_____no_output_____" ], [ "Write an if-else statement that performs the following. If the rating is larger then eight print “this album is amazing”. If the rating is less than or equal to 8 print “this album is ok”.", "_____no_output_____" ] ], [ [ "# Write your code below and press Shift+Enter to execute", "this album is ok\n" ] ], [ [ "Double-click __here__ for the solution.\n\n<!-- \nrating = 8.5\nif rating > 8:\n print (\"this album is amazing\")\nelse:\n print (\"this album is ok\")\n-->", "_____no_output_____" ], [ "<hr>", "_____no_output_____" ], [ "Write an if statement to determine if an album came out before 1980 or in the years: 1991 or 1993. If the condition is true print out the year the album came out.", "_____no_output_____" ] ], [ [ "# Write your code below and press Shift+Enter to execute", "this album came out already\n" ] ], [ [ "Double-click __here__ for the solution.\n\n<!-- \nalbum_year = 1979\n\nif album_year < 1980 or album_year == 1991 or album_year == 1993:\n print (\"this album came out already\")\n-->", "_____no_output_____" ], [ "<hr>\n<h2>The last exercise!</h2>\n<p>Congratulations, you have completed your first lesson and hands-on lab in Python. However, there is one more thing you need to do. The Data Science community encourages sharing work. The best way to share and showcase your work is to share it on GitHub. By sharing your notebook on GitHub you are not only building your reputation with fellow data scientists, but you can also show it off when applying for a job. Even though this was your first piece of work, it is never too early to start building good habits. So, please read and follow <a href=\"https://cognitiveclass.ai/blog/data-scientists-stand-out-by-sharing-your-notebooks/\" target=\"_blank\">this article</a> to learn how to share your work.\n<hr>", "_____no_output_____" ], [ "<h3>About the Authors:</h3> \n<p><a href=\"https://www.linkedin.com/in/joseph-s-50398b136/\" target=\"_blank\">Joseph Santarcangelo</a> is a Data Scientist at IBM, and holds a PhD in Electrical Engineering. His research focused on using Machine Learning, Signal Processing, and Computer Vision to determine how videos impact human cognition. Joseph has been working for IBM since he completed his PhD.</p>", "_____no_output_____" ], [ "<p>Other contributors: <a href=\"https://www.linkedin.com/in/jiahui-mavis-zhou-a4537814a\">Mavis Zhou</a>, <a href=\"https://github.com/computationalcore\" target=\"_blank\">Vin Busquet</a>, <a href=\"https://github.com/raphtrajano\" target=\"_blank\">Raph Trajano</a></p>", "_____no_output_____" ], [ "<hr>", "_____no_output_____" ], [ "<p>Copyright &copy; 2018 IBM Developer Skills Network. This notebook and its source code are released under the terms of the <a href=\"https://cognitiveclass.ai/mit-license/\">MIT License</a>.</p>", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown" ]
[ [ "markdown", "markdown", "markdown", "markdown", "markdown", "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown", "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown", "markdown", "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown", "markdown", "markdown", "markdown", "markdown", "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown", "markdown", "markdown", "markdown", "markdown", "markdown" ] ]
ecea89e0b1e23474bd32e126b8a04dc85679979e
8,766
ipynb
Jupyter Notebook
project-proposal.ipynb
dillbeck/bike-path
5deee812f20a81f1c6c2fdcc40def1a7fd496281
[ "MIT" ]
null
null
null
project-proposal.ipynb
dillbeck/bike-path
5deee812f20a81f1c6c2fdcc40def1a7fd496281
[ "MIT" ]
null
null
null
project-proposal.ipynb
dillbeck/bike-path
5deee812f20a81f1c6c2fdcc40def1a7fd496281
[ "MIT" ]
null
null
null
60.455172
1,033
0.728154
[ [ [ "# Project Title: Bikeshare Demand Prediction\nDSCI 521: Data Analysis and Interpretation <br> Summer Term - July 28, 2019 <br> Term Project Phase 1: Scoping an analytics project <br> __Team Members: Stephanie Dillbeck, Christopher Spencer, Zexi Yu__ \n\n\n## Team Background\n\n\n__Stephanie Dillbeck:__ I am currently focused on database management and big data with an interest in cybersecurity as well as HCI. I have a solid background in SQL, with some experience in HTML, Javascript and Java. I’m interested in gaining a solid foundation in Python and basic statistical analysis. \n\n__Christopher Spencer:__ With a background in Mechanical Engineering and transitioning into Aerospace Systems Engineering, I have experience in both mathematics and statistics with a primary focus on computational mechanics. While new to Python in particular, there is definite overlap with my past programming experiences.\n\n__Zexi Yu:__ I am presently focused on the area of machine learning with knowledge of common algorithms and skills mostly in image processing using python on platforms of Tensorflow and Keras. And I would like to grow more into the computer vision area and gain more data analysis skills.", "_____no_output_____" ], [ "## Project Domain\n\nFor our project, we would like our analyze transportation data in the United States. The transportation sector is immense, where data may be analyze with respect to the mode of transportation, geographic location, emissions, age of driver or rider, infrastructure (e.g. bike lanes, pedestrian sidewalks), ridesharing, safety, and more. Over the past several years and even decades, electric vehicles, ridesharing, and population migration patterns that have impacted the transportation sector provide ample problems or trends waiting to be discovered. For instance, the social conscious to protect the quality of air and reduce national dependency of foreign oil has been cited as a boon to electric vehicle industry. Also, one could consider how population migration patterns have impacted the sale or ownership of vehicles to ridesharing. Lastly, one could also evaluate how are cities, regions or states are adjusting their transportation budgets in response to those changes to using alternate modes of transportation. \n\nAs we explored possible datasets within the broad category of transportation, our group quickly scoped down our focus to the the subdomain of of Bicycle Transportation. With strong personal interest and an abundance of quality datasets, we further narrowed our topic to the Bike Sharing in Washington D.C. Dataset found on Kaggle.com. This dataset provides a count of bikes rented in the Capital Bikeshare system, along with associated date/time, weather/conditions, and holiday data for a two-year period. Our initial exploration will look at the features of the data to determine which features have the most impact on rental demand.", "_____no_output_____" ], [ "## Analytic Tools\n\n__Data cleaning:__ excel, python, pandas, numpy\n\n__Data visualization:__ matplotlib, bokeh, seaborn\n\n__Data analysis:__ sklearn\n", "_____no_output_____" ], [ "## Analysis and Exploration of Dataset\nIn choosing a dataset we explored numerous datasets of bike usage data throughout the country and even internationally. There were many pre-processed datasets available, including some with overall bike lane usage counts, journey durations, routes, bike-related incidents, as well as biker demographics, but the most rich data sources seemed to originate from the formalized bikeshare systems within major citites or regions. The Capital Bikeshare dataset was particularly attractive because it already included weather and condition data, as well workday/weekend/holiday information.\n\n#### The visual analysis and exploration phase of our chosen dataset is included here: <br> [Phase 1 Exploration and Analysis](phase-1.ipynb).\n", "_____no_output_____" ], [ "## Audience\nOur data analysis will potentially have different applications for different stakeholder groups. Public interest groups or nonprofits may have an interest in our data on bike-share usage to focus their efforts and further their specific missions, as well as engage supporters in the community. Local and state municipalities may use our data to help with infrastructure planning, as well as citizen engagement and social responsibility. Existing businesses may use the data to support employee-engagement and work/life balance efforts, as well as from within their social responsibility platform. Finally, entrepreneurs and start-up business may leverage the data to identify new markets or develop new-to-market products or concepts.\n\n\n## Application\nAt present, the goal for our analysis will be to construct a prediction model for the Capital Bikeshare System in Washington, D.C. As the analysis will take into consideration such things as bike usage, membership, weather, air quality, and holidays, the model created will potentially have application beyond the local environment of Washington, D.C. once these factors are adjusted for in a new location. The prediction model could support business decisions, infrastructure planning, maintenance schedules and community involvement, as well as present new opportunities for products or business concepts. As the project will be presented in the form a of a Jupyter Notebook that contains background information, discussion and code for reproducible results, different stakeholder groups can utilize the prediction model in a way that suits their mission and needs, as well as adapt it to other applications. \n", "_____no_output_____" ], [ "## Limits and Improvements\nIf we use linear regression to build our prediction model, in order to improve the prediction accuracy, we may want to do some feature crosses. For example, we could combine weather and season and build a new feature. However, the performance may be limited if we just randomly combine them. In this case, we can apply different feature selection models and make comparisons between them. We can keep tuning the model until we achieve an optimal accuracy.\n\n## Continued Analysis\nFrom the result of data visulization above, we can draw some conclusions for continued analysis:\n- The datetime column need to be split into multiple features, including year, month, dayofweek, and hour\n- All columns need to be standardized before creating the feature vectors\n- The atemp column can be dropped due to its high correlation coefficient with temp\n- The demands for casual users and for registered users need to be predicted seperately \n- The outliers need to be removed before creating the feature vectors for hours and workingday\n- The column windspeed can be dropped due to low correlation coefficient with demand\n", "_____no_output_____" ] ] ]
[ "markdown" ]
[ [ "markdown", "markdown", "markdown", "markdown", "markdown", "markdown" ] ]
ecea8f4062b077689e6e96a408531ebd83b40984
7,322
ipynb
Jupyter Notebook
pelops/analysis/makeFeaturesResNet50.ipynb
dave-lab41/pelops
292af80dba190f9506519c8e13432fef648a2291
[ "Apache-2.0" ]
48
2016-12-11T15:43:28.000Z
2021-02-28T14:16:42.000Z
pelops/analysis/makeFeaturesResNet50.ipynb
dave-lab41/pelops
292af80dba190f9506519c8e13432fef648a2291
[ "Apache-2.0" ]
100
2016-11-15T19:23:52.000Z
2017-08-22T20:22:13.000Z
pelops/analysis/makeFeaturesResNet50.ipynb
agude/pelops
292af80dba190f9506519c8e13432fef648a2291
[ "Apache-2.0" ]
25
2016-11-15T17:49:32.000Z
2020-02-17T13:04:26.000Z
26.722628
100
0.550123
[ [ [ "cd 'deep-learning-models/'", "_____no_output_____" ], [ "import numpy as np\nfrom resnet50 import ResNet50\nfrom keras.preprocessing import image\nfrom imagenet_utils import preprocess_input, decode_predictions\nfrom keras.models import Model\nimport scipy.spatial.distance\nimport time\nimport json\n", "_____no_output_____" ], [ "def load_image(img_path):\n data = image.load_img(img_path, target_size=(224, 224))\n x = image.img_to_array(data)\n x = np.expand_dims(x, axis=0)\n x = preprocess_input(x)\n return x", "_____no_output_____" ], [ "#def get_base_model():\n# base_model = ResNet50(weights='imagenet',include_top=False)\n# return base_model\n\ndef get_models():\n # include_top needs to be True for this to work\n base_model = ResNet50(weights='imagenet',include_top=True)\n model = Model(input=base_model.input, output=base_model.get_layer('flatten_1').output)\n return (model, base_model)\n\ndef image_features(img,model):\n features = np.zeros((1,2048),dtype=np.float16)\n #model = Model(input=base_model.input, output=base_model.get_layer('flatten_1').output)\n predictions = model.predict(img)\n return predictions", "_____no_output_____" ], [ "model, base_model = get_models()", "_____no_output_____" ], [ "# model = Model(input=bm.input, output=bm.get_layer('flatten_1').output)", "_____no_output_____" ], [ "def save_model_workaround(model, model_output_file, weights_output_file):\n print('saving model to {}'.format(model_output_file))\n print('saving weignts to {}'.format(weights_output_file))\n # serialize model to JSON\n model_json = model.to_json()\n with open(model_output_file, 'w') as json_file:\n json_file.write(model_json)\n # serialize weights to HDF5\n model.save_weights(weights_output_file)\n\n\ndef load_model_workaround(model_output_file, weights_output_file):\n # load json and create model\n json_file = open(model_output_file, 'r')\n loaded_model_json = json_file.read()\n json_file.close()\n loaded_model = model_from_json(loaded_model_json)\n # load weights into new model\n loaded_model.load_weights(weights_output_file)\n return loaded_model", "_____no_output_____" ], [ "save_model_workaround(model,\n '/local_data/dgrossman/model_save_dir/dgcars_resenet.model.json',\n '/local_data/dgrossman/model_save_dir/dgcars_resenet.weights.hdf5')", "_____no_output_____" ], [ "# print(scipy.spatial.distance.cosine(feature1,feature2),'good')\n# print(scipy.spatial.distance.cosine(feature1,feature3),'bad')\n# print(scipy.spatial.distance.cosine(feature2,feature3),'bad')", "_____no_output_____" ], [ "prefix = '/local_data/dgrossman/VeRi/'\n", "_____no_output_____" ], [ "tLines = open(prefix + 'trainingLines','r')\ntrainingList = list()\nfor tLine in tLines:\n tLine = tLine.strip()\n tLine = tLine.replace('\"','')\n parts = tLine.split(' ')\n ldict = dict()\n for part in parts:\n l, r = part.split('=')\n ldict[l]=str(r)\n trainingList.append(ldict)\ntLines.close()", "_____no_output_____" ], [ "trainingList = list()\ntLines = open(prefix + 'test_features.json','r')\nfor line in tLines:\n line = line.strip()\n line = json.loads(line)\n trainingList.append(line)", "_____no_output_____" ], [ "trainingList[0]", "_____no_output_____" ], [ "import json\n#outFile = open('/local_data/dgrossman/VeRi/training_features.json','w')\noutFile = open('/local_data/dgrossman/VeRi/test_features.json','w')\n\nbatchSize = 1000\nstart = time.time()\nfor idx,line in enumerate(trainingList):\n tempd = dict()\n if idx % batchSize == 0:\n end = time.time() - start\n start = time.time()\n print ('total {0} batch {1} images in {2} seconds'.format(idx,batchSize,end))\n #img = load_image(prefix + 'image_train/'+line['imageName'])\n img = load_image(prefix + 'image_test/'+line['imageName'])\n feature = image_features(img, model) \n tempd['resnet50'] = feature.tolist()[0]\n tempd.update(line)\n outFile.write(json.dumps(tempd)+'\\n')\noutFile.close()", "_____no_output_____" ], [ "funtime = open('/local_data/dgrossman/VeRi/training_features.json','r')\nwork = list()\nfor line in funtime:\n line = line.strip()\n line = json.loads(line)\n work.append(line)\nfuntime.close()", "_____no_output_____" ], [ "work[0]\n", "_____no_output_____" ] ] ]
[ "code" ]
[ [ "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code" ] ]
ecea9a1687e92801094c7dc748da10adbf344085
2,096
ipynb
Jupyter Notebook
amz_reports/amz_returns_catalog.ipynb
dyslab/jnb-sample
38af701866b8496729d63f844e56137d125c4223
[ "MIT" ]
null
null
null
amz_reports/amz_returns_catalog.ipynb
dyslab/jnb-sample
38af701866b8496729d63f844e56137d125c4223
[ "MIT" ]
2
2020-03-24T17:57:15.000Z
2020-03-31T10:21:49.000Z
amz_reports/amz_returns_catalog.ipynb
dyslab/jnb-sample
38af701866b8496729d63f844e56137d125c4223
[ "MIT" ]
null
null
null
21.833333
140
0.582538
[ [ [ "Doc title: **Amazon Returns Report Analyzation Catalog**\n\nArticle notes: The links related to the analyzation of the reports that exported from 'Reports/Fulfillment' @Amazon Seller Central.\n\n文章备注:亚马逊后台退货报告分析链接目录\n\nLast modified date: 2019-09-15 03:20:46 ", "_____no_output_____" ], [ "# 亚马逊退货报表目录:", "_____no_output_____" ], [ "## 特定时段的FBA退货情况分析,点击:[amz_returns.ipynb](amz_returns.ipynb)", "_____no_output_____" ], [ "## FBA退货产品款式占比分析,点击:[amz_returns_model.ipynb](amz_returns_model.ipynb)", "_____no_output_____" ] ] ]
[ "markdown" ]
[ [ "markdown", "markdown", "markdown", "markdown" ] ]
eceaa02f44a82e9f569db8f4b57e0c783d8dc5e6
15,879
ipynb
Jupyter Notebook
module1-rnn-and-lstm/LS_DS_431_RNN_and_LSTM_Assignment.ipynb
ToddMG/DS-Unit-4-Sprint-3-Deep-Learning
93ce8a0254b2def6bf241a42a9e62429f18d42a7
[ "MIT" ]
null
null
null
module1-rnn-and-lstm/LS_DS_431_RNN_and_LSTM_Assignment.ipynb
ToddMG/DS-Unit-4-Sprint-3-Deep-Learning
93ce8a0254b2def6bf241a42a9e62429f18d42a7
[ "MIT" ]
null
null
null
module1-rnn-and-lstm/LS_DS_431_RNN_and_LSTM_Assignment.ipynb
ToddMG/DS-Unit-4-Sprint-3-Deep-Learning
93ce8a0254b2def6bf241a42a9e62429f18d42a7
[ "MIT" ]
null
null
null
33.785106
318
0.55501
[ [ [ "<img align=\"left\" src=\"https://lever-client-logos.s3.amazonaws.com/864372b1-534c-480e-acd5-9711f850815c-1524247202159.png\" width=200>\n<br></br>\n<br></br>\n\n## *Data Science Unit 4 Sprint 3 Assignment 1*\n\n# Recurrent Neural Networks and Long Short Term Memory (LSTM)\n\n![Monkey at a typewriter](https://upload.wikimedia.org/wikipedia/commons/thumb/3/3c/Chimpanzee_seated_at_typewriter.jpg/603px-Chimpanzee_seated_at_typewriter.jpg)\n\nIt is said that [infinite monkeys typing for an infinite amount of time](https://en.wikipedia.org/wiki/Infinite_monkey_theorem) will eventually type, among other things, the complete works of Wiliam Shakespeare. Let's see if we can get there a bit faster, with the power of Recurrent Neural Networks and LSTM.\n\nThis text file contains the complete works of Shakespeare: https://www.gutenberg.org/files/100/100-0.txt\n\nUse it as training data for an RNN - you can keep it simple and train character level, and that is suggested as an initial approach.\n\nThen, use that trained RNN to generate Shakespearean-ish text. Your goal - a function that can take, as an argument, the size of text (e.g. number of characters or lines) to generate, and returns generated text of that size.\n\nNote - Shakespeare wrote an awful lot. It's OK, especially initially, to sample/use smaller data and parameters, so you can have a tighter feedback loop when you're trying to get things running. Then, once you've got a proof of concept - start pushing it more!", "_____no_output_____" ] ], [ [ "# TODO - Words, words, mere words, no matter from the heart.\n\nfrom tensorflow.keras.preprocessing import sequence\nfrom tensorflow.keras.models import Sequential\nfrom tensorflow.keras.layers import Dense, Embedding\nfrom tensorflow.keras.layers import LSTM\nfrom tensorflow.keras.datasets import imdb\nfrom tensorflow.keras.callbacks import LambdaCallback\nfrom tensorflow.keras.optimizers import RMSprop\n\nimport numpy as np\nimport random\nimport sys\nimport os", "_____no_output_____" ], [ "data = []\n\nwith open('100-0.txt', 'r') as f:\n data.append(f.read())", "_____no_output_____" ], [ "text = \" \".join(data)\n\nchars = list(set(text))\n\nchar_int = {c:i for i,c in enumerate(chars)}\nint_char = {i:c for i,c in enumerate(chars)}", "_____no_output_____" ], [ "len(chars)", "_____no_output_____" ], [ "maxlen = 40\nstep = 5\n\nencoded = [char_int[c] for c in text]\n\nsequences = []\nnext_chars = []\n\nfor i in range(0, len(encoded) - maxlen, step):\n sequences.append(encoded[i : i + maxlen])\n next_chars.append(encoded[i + maxlen])\n \nprint('sequences:', len(sequences))", "sequences: 1114623\n" ], [ "x = np.zeros((len(sequences), maxlen, len(chars)), dtype=np.bool)\ny = np.zeros((len(sequences), len(chars)), dtype=np.bool)\n\nfor i, sequence in enumerate(sequences):\n for t, char in enumerate(sequence):\n x[i,t,char] = 1\n \n y[i, next_chars[i]] = 1", "_____no_output_____" ], [ "x.shape", "_____no_output_____" ], [ "y.shape", "_____no_output_____" ], [ "model = Sequential()\nmodel.add(LSTM(128, input_shape=(maxlen, len(chars))))\nmodel.add(Dense(len(chars), activation='softmax'))\n\nmodel.compile(loss='categorical_crossentropy', optimizer='adam')", "_____no_output_____" ], [ "# Sample temperature function for lecture\ndef sample(preds, temperature=1.0):\n # helper function to sample an index from a probability array\n preds = np.asarray(preds).astype('float64')\n preds = np.log(preds) / temperature\n exp_preds = np.exp(preds)\n preds = exp_preds / np.sum(exp_preds)\n probas = np.random.multinomial(1, preds, 1)\n return np.argmax(probas)", "_____no_output_____" ], [ "def on_epoch_end(epoch, _):\n # Function invoked at end of each epoch. Prints generated text.\n \n print()\n print('----- Generating text after Epoch: %d' % epoch)\n \n start_index = random.randint(0, len(text) - maxlen - 1)\n \n generated = ''\n \n sentence = text[start_index: start_index + maxlen]\n generated += sentence\n print('----- Generating with seed: \"' + sentence + '\"')\n sys.stdout.write(generated)\n \n for i in range(400):\n x_pred = np.zeros((1, maxlen, len(chars)))\n for t, char in enumerate(sentence):\n x_pred[0, t, char_int[char]] = 1\n \n preds = model.predict(x_pred, verbose=0)[0]\n next_index = sample(preds, temperature=1.0)\n next_char = int_char[next_index]\n \n sentence = sentence[1:] + next_char\n \n sys.stdout.write(next_char)\n sys.stdout.flush()\n print()\n \nprint_callback = LambdaCallback(on_epoch_end=on_epoch_end)", "_____no_output_____" ], [ "model.fit(x, y,\n batch_size=128,\n epochs=5,\n callbacks=[print_callback])", "Train on 1114623 samples\nEpoch 1/5\n1114496/1114623 [============================>.] - ETA: 0s - loss: 2.2517\n----- Generating text after Epoch: 0\n----- Generating with seed: \"nded are you, and have fought\n Not as\"\nnded are you, and have fought\n Not as sorres harid comsiqurlad is pornew\nLet be quoader in the bridise thesengly wee’t on myip.\n\n [_Exgutt to d’st Soor ANs wath Ciong the Verchiopstt tien,\nHe Is the with ned Gut\n To best’s buck off jold, mars it leve thes\n sfeely at uveet see thyow ane pray ate: I eple wher'd offry roplonduch to wich;\nThou bord,\n Sike awe Feanpy, for didgasse,\n Weat As bave athand then simot; Had speed ulc\n1114623/1114623 [==============================] - 235s 211us/sample - loss: 2.2517\nEpoch 2/5\n1114368/1114623 [============================>.] - ETA: 0s - loss: 1.9045\n----- Generating text after Epoch: 1\n----- Generating with seed: \"efore follow me, and I'll direct you how\"\nefore follow me, and I'll direct you how say!\n KENCENTE. Higat well thit goof come?\n\nREDICETIUS.\nHo! tell will daye for you. Herr-hall knest\nTo clury wares then marchin, araster made my yourr stous at that shand,\n Lere of gringlads woundstion, their wordd,\nBet, and weramile reglmour, know buldeen,\nWMats Eporst baty nation! Alle.\n\nFIMI.\nWith in a to8lat! O there woddsed,\nWhat net and a shand blearena. I pringain,\n Than howrumonier\n1114623/1114623 [==============================] - 228s 205us/sample - loss: 1.9045\nEpoch 3/5\n1114496/1114623 [============================>.] - ETA: 0s - loss: 1.7956\n----- Generating text after Epoch: 2\n----- Generating with seed: \"The way to dusty death. Out, out, brief \"\nThe way to dusty death. Out, out, brief eatharess, me\nHorw? Hulais lovers me, mane here terced;\n Sears a everuchy for you know as all of you hers,\nFood wixh all this woul writ his deesues\n What than Biders unot borts a todder, Or thee’s newn;\nHe withing on this I veiced Loks a gan by notes\nSave the word her wan me to be faik’d this!\n And thou baskss morn’d facking thou but in sir\nIver befunce withable of be they will kneir\nmy H\n1114623/1114623 [==============================] - 230s 206us/sample - loss: 1.7956\nEpoch 4/5\n1114496/1114623 [============================>.] - ETA: 0s - loss: 1.7291\n----- Generating text after Epoch: 3\n----- Generating with seed: \"King;—\nShall it, for shame, be spoken in\"\nKing;—\nShall it, for shame, be spoken in their shing\n Thy frown from marronious.\n\nLASCHAST.\nOul'd. Brest MERe a dought hears to ALIAGH ERVICar Gurst inchand\n\n\n Exeunt\n\n\n\n\n FIRSA.\n Remy that I done shall. Drealizal.\n DOMARD. Their world from to head mis agatsst to it?\n [Twould of followick, 'Tonars. I woom bad who mance\n But coundose I can it that weneswer flous\nToom, By To ghelte\n1114623/1114623 [==============================] - 231s 208us/sample - loss: 1.7291\nEpoch 5/5\n1114496/1114623 [============================>.] - ETA: 0s - loss: 1.6817\n----- Generating text after Epoch: 4\n----- Generating with seed: \"this difference.\nI give consent; are you\"\nthis difference.\nI give consent; are you shall the prome of dead,\n 'Tis visment-mouging shall letter wouldstides\n This frarse acome croend though a chasue that is sonce\n Are more wroth ophanted of your and earthan\n But that thee be homcemped it eyounat\n Thou sweak I cornort for murchous of wark.\n DOCLERANI. Aland.\n ARTZAN. No, a strings, in Majeston's prother liend\n Om. So fissince and shall upword to streaved, I,\nKi\n1114623/1114623 [==============================] - 235s 211us/sample - loss: 1.6817\n" ], [ "from tensorflow.python.client import device_lib\nprint(device_lib.list_local_devices())", "[name: \"/device:CPU:0\"\ndevice_type: \"CPU\"\nmemory_limit: 268435456\nlocality {\n}\nincarnation: 789254733241210422\n, name: \"/device:XLA_CPU:0\"\ndevice_type: \"XLA_CPU\"\nmemory_limit: 17179869184\nlocality {\n}\nincarnation: 14510598408632905384\nphysical_device_desc: \"device: XLA_CPU device\"\n, name: \"/device:XLA_GPU:0\"\ndevice_type: \"XLA_GPU\"\nmemory_limit: 17179869184\nlocality {\n}\nincarnation: 11034020051727156322\nphysical_device_desc: \"device: XLA_GPU device\"\n]\n" ] ], [ [ "# Resources and Stretch Goals", "_____no_output_____" ], [ "## Stretch goals:\n- Refine the training and generation of text to be able to ask for different genres/styles of Shakespearean text (e.g. plays versus sonnets)\n- Train a classification model that takes text and returns which work of Shakespeare it is most likely to be from\n- Make it more performant! Many possible routes here - lean on Keras, optimize the code, and/or use more resources (AWS, etc.)\n- Revisit the news example from class, and improve it - use categories or tags to refine the model/generation, or train a news classifier\n- Run on bigger, better data\n\n## Resources:\n- [The Unreasonable Effectiveness of Recurrent Neural Networks](https://karpathy.github.io/2015/05/21/rnn-effectiveness/) - a seminal writeup demonstrating a simple but effective character-level NLP RNN\n- [Simple NumPy implementation of RNN](https://github.com/JY-Yoon/RNN-Implementation-using-NumPy/blob/master/RNN%20Implementation%20using%20NumPy.ipynb) - Python 3 version of the code from \"Unreasonable Effectiveness\"\n- [TensorFlow RNN Tutorial](https://github.com/tensorflow/models/tree/master/tutorials/rnn) - code for training a RNN on the Penn Tree Bank language dataset\n- [4 part tutorial on RNN](http://www.wildml.com/2015/09/recurrent-neural-networks-tutorial-part-1-introduction-to-rnns/) - relates RNN to the vanishing gradient problem, and provides example implementation\n- [RNN training tips and tricks](https://github.com/karpathy/char-rnn#tips-and-tricks) - some rules of thumb for parameterizing and training your RNN", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown" ]
[ [ "markdown" ], [ "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code" ], [ "markdown", "markdown" ] ]
eceaacb72619732b97ee1ca42deb5602c96eacbe
20,718
ipynb
Jupyter Notebook
pydhamed/tests/test_rna/RNA_test.ipynb
bio-phys/pyDHAMed
d42f2a67a2650a0f8b09a798c2afe3b6d85aab08
[ "BSD-3-Clause" ]
21
2017-11-14T18:57:23.000Z
2020-10-17T08:31:07.000Z
pydhamed/tests/test_rna/RNA_test.ipynb
bio-phys/pyDHAMed
d42f2a67a2650a0f8b09a798c2afe3b6d85aab08
[ "BSD-3-Clause" ]
11
2017-11-07T23:28:21.000Z
2018-05-24T07:00:55.000Z
pydhamed/tests/test_rna/RNA_test.ipynb
bio-phys/pyDHAMed
d42f2a67a2650a0f8b09a798c2afe3b6d85aab08
[ "BSD-3-Clause" ]
4
2018-01-04T15:18:49.000Z
2020-05-08T12:41:52.000Z
83.540323
15,898
0.850565
[ [ [ "# Table of Contents\n <p>", "_____no_output_____" ] ], [ [ "pwd", "_____no_output_____" ], [ "import numpy as np\n%matplotlib inline\nimport sys\nsys.path.append(\"../\")\nfrom optimize_dhamed import *", "_____no_output_____" ] ], [ [ "this should form as the basis of a test. ", "_____no_output_____" ] ], [ [ "c_l = [np.genfromtxt(\"count_matrix_1.txt\")]", "_____no_output_____" ], [ "v_ar = np.genfromtxt(\"wfile.txt\")[:,1].reshape((9,1))", "_____no_output_____" ], [ "og = run_dhamed(c_l, -np.log(v_ar), g_init=-(np.zeros(9))*-1.0, numerical_gradients=False, maxiter=10000)", "36\nloglike-start 307329.006201\n307329.006201\nOptimization terminated successfully.\n Current function value: 288165.226259\n Iterations: 22\n Function evaluations: 27\n Gradient evaluations: 27\ntime elapsed 0.00710487365723 s\n" ], [ "ref_f = np.genfromtxt(\"us_dt-e4_2_p1.out\")", "_____no_output_____" ], [ "py_rna = og*-1 - (og[-1]*-1)\nf_rna = ref_f[:,-1] - ref_f[-1,-1]", "_____no_output_____" ], [ "fig, ax = plt.subplots(figsize=(5,3))\nplt.plot(py_rna, \"o\")\nplt.plot(f_rna)\nax.set_xlabel(\"Reaction coordinate\")\nax.set_ylabel(\"PMF ($\\mathrm{k_BT}$)\")\n\nax.legend()\nfig.tight_layout()", "_____no_output_____" ], [ "np.testing.assert_almost_equal(f_rna, py_rna)", "_____no_output_____" ], [ "f_rna - py_rna", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code" ]
[ [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code", "code", "code", "code" ] ]
eceac6b66b11b2aebe1bdad9e06c2b6cd8cb2d35
21,636
ipynb
Jupyter Notebook
Wi19_content/DSMCER/L6_MLIntro_filled.ipynb
ShahResearchGroup/UWDIRECT.github.io
d4db958a6bfe151b6f7b1eb4772d8fd1b9bb0c3e
[ "BSD-3-Clause" ]
1
2021-01-26T19:55:02.000Z
2021-01-26T19:55:02.000Z
Wi19_content/DSMCER/L6_MLIntro_filled.ipynb
ShahResearchGroup/UWDIRECT.github.io
d4db958a6bfe151b6f7b1eb4772d8fd1b9bb0c3e
[ "BSD-3-Clause" ]
null
null
null
Wi19_content/DSMCER/L6_MLIntro_filled.ipynb
ShahResearchGroup/UWDIRECT.github.io
d4db958a6bfe151b6f7b1eb4772d8fd1b9bb0c3e
[ "BSD-3-Clause" ]
null
null
null
46.730022
8,676
0.691486
[ [ [ "import numpy as np\nimport matplotlib.pyplot as plt \nimport pandas as pd \n%matplotlib inline\n# this is a new library you haven't seen before, what do you think it does? \nfrom sklearn.neighbors import KNeighborsClassifier\nfrom sklearn.model_selection import train_test_split", "_____no_output_____" ] ], [ [ "<img src=\"http://www.nature.com/article-assets/npg/srep/2015/150825/srep13285/images/w926/srep13285-f4.jpg\" width=\"300\" height=\"300\" />\n\nFrom this article in [Scientific Reports](http://www.nature.com/articles/srep13285)", "_____no_output_____" ], [ "#### Read in the data\n* elemental data: [https://raw.githubusercontent.com/UWDIRECT/UWDIRECT.github.io/master/Wi18_content/DSMCER/atomsradii.csv](https://raw.githubusercontent.com/UWDIRECT/UWDIRECT.github.io/master/Wi18_content/DSMCER/atomsradii.csv)\n* testing data: [https://raw.githubusercontent.com/UWDIRECT/UWDIRECT.github.io/master/Wi18_content/DSMCER/testing.csv](https://raw.githubusercontent.com/UWDIRECT/UWDIRECT.github.io/master/Wi18_content/DSMCER/testing.csv)", "_____no_output_____" ] ], [ [ "d_train = pd.read_csv('https://raw.githubusercontent.com/UWDIRECT/UWDIRECT.github.io/master/Wi18_content/DSMCER/atomsradii.csv')\nd_test = pd.read_csv('https://raw.githubusercontent.com/UWDIRECT/UWDIRECT.github.io/master/Wi18_content/DSMCER/testing.csv')", "_____no_output_____" ], [ "d_test", "_____no_output_____" ] ], [ [ "Take 1-2 min and look @ the data in elements using pandas and Python you and your partner decide what to do. E.g. you could recreate the above plot with plt.scatter(elements.rWC,elements.rCh)", "_____no_output_____" ] ], [ [ "fig, ax = plt.subplots(figsize=(5, 5))\ncolors = ['b', 'r', 'k']\nshapes = ['s', '^', 'o']\n\nfor typ, color, shape in zip(d_train.Type.unique(), colors, shapes):\n ax.scatter(d_train[d_train.Type==typ].rWC, d_train[d_train.Type==typ].rCh, s=200, marker=shape, color=color)\n \nax.set_xlim([0, 1.5])\nax.set_ylim([0, 1.5])", "_____no_output_____" ] ], [ [ "#### Now, let's make a new classifier object\n\nWe'll use `KNeighborsClassifier(n_neighbors=k)` where `k` is the number of neighbors to use.\n\n\n\nThen 'train' it using the `.fit` function on the object returned by the `KNeighborsClassifier` call.", "_____no_output_____" ] ], [ [ "inputs = ['rWC', 'rCh']\nX_train = d_train[inputs]\ny_train = d_train['Type']\nX_test = d_test[inputs]\ny_test = d_test['Type']\n#X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.5)", "_____no_output_____" ], [ "KNNmodel = KNeighborsClassifier(n_neighbors=3)", "_____no_output_____" ], [ "KNNmodel.fit(X_train, y_train)", "_____no_output_____" ], [ "y_train", "_____no_output_____" ] ], [ [ "### You can use the following function to see how your model is doing: \n\n`knn.predict(X) `\n\n#### As a function of K, you and your partner should determine: \n* Testing error rate\n* Training error rate \n\n#### Need not be quantitative but spend (1/2 - 2/3 of remaining time exploring this) \n\n", "_____no_output_____" ] ], [ [ "rate = KNNmodel.predict(X_train) == y_train\nprint('Training Error Rate:', 1 - np.mean(rate))\n\nrate = KNNmodel.predict(X_test) == y_test\nprint('Testing Error Rate:', 1 - np.mean(rate))", "Training Error Rate: 0.0\nTesting Error Rate: 0.4\n" ], [ "rate", "_____no_output_____" ] ], [ [ "#### With remaining time go through the cell below and look at graphs of the decision boundary vs K. \n* See if you can use the graph to determine your **testing** error rate \n* Could you also use the graph to determine your **training** error rate? (_open ended_)", "_____no_output_____" ] ], [ [ "# additional library we will use \nfrom matplotlib.colors import ListedColormap\n\n# just for convenience and similarity with sklearn tutorial\n# I am going to assign our X and Y data to specific vectors\n# this is not strictly needed and you could use elements df for the whole thing!\nd1 = d_train\nelements = d1\nX=elements[['rWC','rCh']]\n\n#this is a trick to turn our strings (type of element / class) into unique \n#numbers. Play with this in a separate cell and make sure you know wth is \n#going on!\nlevels,labels=pd.factorize(elements.Type)\ny=levels\n\n#This determines levelspacing for our color map and the colors themselves\nh=0.02\ncmap_light = ListedColormap(['#FFAAAA', '#AAFFAA', '#AAAAFF'])\ncmap_bold = ListedColormap(['#FF0000', '#00FF00', '#0000FF'])\n\n# in the sklearn tutorial two different weights are compared\n# the decision between \"uniform\" and \"distance\" determines the probability\n# weight. \"uniform\" is the version presented in class, you can change to \n# distance\nweights='uniform'\n\n# I am actually refitting the KNN here. If you had a big data set you would\n# not do this, but I want you to have the convenience of changing K or \n# weights here in this cell. Large training sets with many features can take \n# awhile for KNN training! \n\nK=5\nclf = KNeighborsClassifier(n_neighbors=K, weights=weights)\nclf.fit(X,y)\n\n# Straight from the tutorial - quickly read and see if you know what these \n# things are going - if you are < 5 min until end then you should skip this part \n\n# Plot the decision boundary. For that, we will assign a color to each\n# point in the mesh [x_min, x_max]x[y_min, y_max].\nx_min, x_max = elements.rWC.min() - 0.1 , elements.rWC.max() + 0.1\ny_min, y_max = elements.rCh.min() - 0.1 , elements.rCh.max() + 0.1 \nxx, yy = np.meshgrid(np.arange(x_min, x_max, h), \n np.arange(y_min, y_max, h)) \nZ = clf.predict(np.c_[xx.ravel(), yy.ravel()])\n\n# Put the result into a color plot\nZ = Z.reshape(xx.shape)\nplt.figure(figsize=(4,4));\n#plt.pcolormesh(xx, yy, Z, cmap=cmap_light)\nplt.pcolormesh(xx, yy, Z, cmap=cmap_light,vmin=0,vmax=np.max(y))\n# Plot also the training points\n# This may be the 1st time you have seen how to color points by a 3rd vector\n# In this case y ( see c=y in below statement ). This is very useful! \nplt.scatter(X.rWC, X.rCh, c=y, cmap=cmap_bold)\n\n# Set limits and lebels \nplt.xlim(xx.min(), xx.max())\nplt.ylim(yy.min(), yy.max())\nplt.xlabel('rWC')\nplt.ylabel('rCh')", "_____no_output_____" ] ] ]
[ "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code" ]
[ [ "code" ], [ "markdown", "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ] ]
eceac83d7e90eb30de2ec088f04b172fc5f6782f
3,648
ipynb
Jupyter Notebook
experiments/Atlas/t01_preprocessing.ipynb
gao-lab/GLUE
e84cb6483971dcb1e2485080f812899baaf31b5b
[ "MIT" ]
41
2021-08-23T07:29:55.000Z
2022-03-12T00:29:52.000Z
experiments/Atlas/t01_preprocessing.ipynb
gao-lab/GLUE
e84cb6483971dcb1e2485080f812899baaf31b5b
[ "MIT" ]
7
2021-11-25T21:25:50.000Z
2022-02-15T02:22:57.000Z
experiments/Atlas/t01_preprocessing.ipynb
gao-lab/GLUE
e84cb6483971dcb1e2485080f812899baaf31b5b
[ "MIT" ]
8
2021-10-05T07:24:14.000Z
2022-03-27T22:46:16.000Z
21.458824
92
0.536732
[ [ [ "import os\n\nimport anndata\nimport networkx as nx\nfrom networkx.algorithms.bipartite import biadjacency_matrix", "_____no_output_____" ], [ "PATH = \"t01_preprocessing\"\nos.makedirs(PATH, exist_ok=True)", "_____no_output_____" ] ], [ [ "# Read data", "_____no_output_____" ] ], [ [ "rna = anndata.read_h5ad(\"../../data/dataset/Cao-2020.h5ad\")\natac = anndata.read_h5ad(\"../../data/dataset/Domcke-2020.h5ad\")", "_____no_output_____" ], [ "rna_pp = anndata.read_h5ad(\"s01_preprocessing/rna.h5ad\", backed=\"r\")\natac_pp = anndata.read_h5ad(\"s01_preprocessing/atac.h5ad\", backed=\"r\")", "_____no_output_____" ], [ "graph = nx.read_graphml(\"s01_preprocessing/full.graphml.gz\")", "_____no_output_____" ] ], [ [ "# Update meta", "_____no_output_____" ] ], [ [ "rna.var[\"highly_variable\"] = [item in rna_pp.var_names for item in rna.var_names]\natac.var[\"highly_variable\"] = [item in atac_pp.var_names for item in atac.var_names]", "_____no_output_____" ] ], [ [ "# Subsample", "_____no_output_____" ] ], [ [ "rna = rna[rna_pp.obs[\"mask\"], :]\natac = atac[atac_pp.obs[\"mask\"], :]", "_____no_output_____" ] ], [ [ "# Convert data", "_____no_output_____" ] ], [ [ "atac2rna = anndata.AnnData(\n X=atac.X @ biadjacency_matrix(graph, atac.var_names, rna.var_names),\n obs=atac.obs, var=rna.var\n)", "_____no_output_____" ] ], [ [ "# Save data", "_____no_output_____" ] ], [ [ "rna.write(f\"{PATH}/rna.h5ad\", compression=\"gzip\")\natac.write(f\"{PATH}/atac.h5ad\", compression=\"gzip\")\natac2rna.write(f\"{PATH}/atac2rna.h5ad\", compression=\"gzip\")", "_____no_output_____" ] ] ]
[ "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code" ]
[ [ "code", "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ] ]
eceacfbebfed2c28d6002e504012df9af7ea3736
5,556
ipynb
Jupyter Notebook
examples/reference/elements/bokeh/RGB.ipynb
pyviz/holoviews
22db2d1dd3b5d0b42a08f00645e04f693ad63dbb
[ "BSD-3-Clause" ]
304
2019-02-07T16:41:57.000Z
2019-11-08T23:43:33.000Z
examples/reference/elements/bokeh/RGB.ipynb
pyviz/holoviews
22db2d1dd3b5d0b42a08f00645e04f693ad63dbb
[ "BSD-3-Clause" ]
689
2019-02-07T00:01:16.000Z
2019-11-11T22:42:58.000Z
examples/reference/elements/bokeh/RGB.ipynb
pyviz/holoviews
22db2d1dd3b5d0b42a08f00645e04f693ad63dbb
[ "BSD-3-Clause" ]
61
2019-02-08T17:06:30.000Z
2019-11-02T21:32:50.000Z
31.748571
505
0.573794
[ [ [ "<div class=\"contentcontainer med left\" style=\"margin-left: -50px;\">\n<dl class=\"dl-horizontal\">\n <dt>Title</dt> <dd> RGB Element</dd>\n <dt>Dependencies</dt> <dd>Bokeh</dd>\n <dt>Backends</dt>\n <dd><a href='./RGB.ipynb'>Bokeh</a></dd>\n <dd><a href='../matplotlib/RGB.ipynb'>Matplotlib</a></dd>\n <dd><a href='../plotly/RGB.ipynb'>Plotly</a></dd>\n</dl>\n</div>", "_____no_output_____" ] ], [ [ "import numpy as np\nimport holoviews as hv\nfrom holoviews import opts\nhv.extension('bokeh')", "_____no_output_____" ] ], [ [ "``RGB`` represents a regularly spaced 2D grid of an underlying continuous space of RGB(A) (red, green, blue, and alpha) color space values. The definition of the grid closely matches the semantics of an Image and in the simplest case the grid may be specified as a ``NxMx3`` or ``NxMx4`` array of values along with a bounds. An RGB may also be defined through explicit and regularly spaced x/y-coordinate arrays. The two most basic supported constructors of an ``RGB`` element therefore include:\n\n RGB((X, Y, R, G, B))\n\nwhere ``X`` is a 1D array of shape ``M``, ``Y`` is a 1D array of shape ``N`` and ``R``/``G``/``B`` are 2D array of shape ``NxM``, or equivalently:\n\n RGB(Z, bounds=(x0, y0, x1, y1))\n\nwhere Z is a 3D array of stacked R/G/B arrays with shape NxMx3/4 and the bounds define the (left, bottom, right, top) edges of the four corners of the grid. Other gridded formats which support declaring of explicit x/y-coordinate arrays such as xarray are also supported. See the [Gridded Datasets](../../../user_guide/09-Gridded_Datasets.ipynb) user guide for all the other accepted data formats.\n\nOne of the simplest ways of creating an ``RGB`` element is to load an image file (such as PNG) off disk, using the ``load_image`` classmethod:", "_____no_output_____" ] ], [ [ "hv.RGB.load_image('../assets/penguins.png')", "_____no_output_____" ] ], [ [ "If you have ``PIL`` or [``pillow``](https://python-pillow.org) installed, you can also pass in a PIL Image as long as you convert it to Numpy arrays first:\n\n```\nfrom PIL import Image\nhv.RGB(np.array(Image.open('../assets/penguins.png')))\n```\n\nThis Numpy-based method for constructing an ``RGB`` can be used to stack up arbitrary 2D arrays into a color image:", "_____no_output_____" ] ], [ [ "x,y = np.mgrid[-50:51, -50:51] * 0.1\n\nr = 0.5*np.sin(np.pi +3*x**2+y**2)+0.5\ng = 0.5*np.sin(x**2+2*y**2)+0.5\nb = 0.5*np.sin(np.pi/2+x**2+y**2)+0.5\n\nhv.RGB(np.dstack([r,g,b]))", "_____no_output_____" ] ], [ [ "You can see how the RGB object is created from the original channels as gray `Image` elements:", "_____no_output_____" ] ], [ [ "opts.defaults(opts.Image(cmap='gray'))", "_____no_output_____" ] ], [ [ "Now we can index the components:", "_____no_output_____" ] ], [ [ "hv.Image(r,label=\"R\") + hv.Image(g,label=\"G\") + hv.Image(b,label=\"B\")", "_____no_output_____" ] ], [ [ "``RGB`` also supports an optional alpha channel, which will be used as a mask revealing or hiding any ``Element``s it is overlaid on top of:", "_____no_output_____" ] ], [ [ "mask = 0.5*np.sin(0.2*(x**2+y**2))+0.5\nrgba = hv.RGB(np.dstack([r,g,b,mask]))\n\nbg = hv.Image(0.5*np.cos(x*3)+0.5, label=\"Background\") * hv.VLine(x=0,label=\"Background\")\noverlay = (bg*rgba).relabel(\"RGBA Overlay\")\nbg + hv.Image(mask,label=\"Mask\") + overlay", "_____no_output_____" ] ], [ [ "RGB elements can be positioned in the plot space using their bounds, with transparency if defined for that image:", "_____no_output_____" ] ], [ [ "hv.Overlay([hv.RGB.load_image('../assets/penguins.png', bounds=(2*i,2*i,3*i,3*i))\n for i in range(1,8)])", "_____no_output_____" ] ], [ [ "One additional way to create RGB objects is via the separate [ImaGen](https://github.com/pyviz-topics/imagen) library, which creates parameterized streams of images for experiments, simulations, or machine-learning applications.\n\nFor full documentation and the available style and plot options, use ``hv.help(hv.RGB).``", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown" ]
[ [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ] ]
eceae72484c16ac4f2dec907091b74f48b29fe34
12,657
ipynb
Jupyter Notebook
posts/wip/partitions_1.ipynb
willboulton/blog
304946d1e9c7edcf7a9af684e7fb27b6fb546153
[ "MIT" ]
null
null
null
posts/wip/partitions_1.ipynb
willboulton/blog
304946d1e9c7edcf7a9af684e7fb27b6fb546153
[ "MIT" ]
null
null
null
posts/wip/partitions_1.ipynb
willboulton/blog
304946d1e9c7edcf7a9af684e7fb27b6fb546153
[ "MIT" ]
null
null
null
74.452941
677
0.6624
[ [ [ "# Partition Numbers 1\n\nThis is based on a lecture I gave for some 6th formers in 2017.\n\n\nMathematicians are often stereotyped as people who like to count things; generally speaking this is an incorrect judgement and mathematics as a subject goes far further than 'accounting-like' tasks. However, in this talk I am going to talk about a very classical subject, dating back to Euler and even before - and this subject area called arithmetic number theory is in fact to do with ways of counting things. \n\n#### Ways of counting stuff\n\n\nHere's a pretty simple question; if I've got 5 identical beads, how many ways can I split them up into different piles? Well, here are some different combinations:\n\n- 5\n- 1 + 1 + 1 + 1 + 1\n- 2 + 3\n- 4 + 1\n- some other ways?\n\nAfter listing a few, we should probably try to approach this problem at least a little bit methodically, so I'm going to start with 1 pile (5) and work my way up to 5 piles (1+1+1+1+1):\n\n- 5\n- 4 + 1\n- 3 + 2\n- 3 + 1 + 1\n- 2 + 2 + 1\n- 2 + 1 + 1 + 1\n- 1 + 1 + 1 + 1 + 1\n\nWe've done this systematically, counting down from the largest number, and always using the largest values where possible, so we've definitely not left any possibilities out. So that's the answer; there are 7 ways to split 5 beads into various piles. If you wanted to write this using mathematical notation, you could summarise this by writing that p(5) = 7, where we are defining p as a function which takes a natural number n as input and returns the number of ways of writing n.\n\nBy the way, these 'possible ways' are called partitions, and the number of ways (as described by this function p) is called the partition function. \n\nYou might reasonably ask \"What is the point of this?\". That's a completely fair question. In many branches of mathematics, and science more generally, it's common to start off with some daft question, and generalise it, or ask some follow up question, until things become both more difficult, but also more interesting and applicable to the real world. In this case, there are a lot of different follow-up questions:\n\n1. This counting process worked for 5, but what about 10, or 200, or a million? Is there a general pattern / a formula? If you graph these, what's the general shape of the graph?\n2. Is there a way of speeding up this counting process? Or say someone wants to know the same problem but with 6 instead of 5, do I have to do all of the work again, or does knowing the answer with 5 help?\n3. What if the order doesn't matter? So that you're allowed 4 + 1 and 1 + 4 as being different and 2 + 2 + 1 is different to 2 + 1 + 2 which is different to 1 + 2 + 2. Is there a formula then?\n4. What if you're only allowed to use different numbers (no repeats)? What if you're only allowed to use odd numbers? Or only allowed to use even numbers? Or square numbers? Or some other restriction?\n\nTo use more technical terminology, we want to generalise this question in different ways: using any number (call it n) instead of 5 (Q1), having more (or less) restrictive rules on how to divide up n (Q3 and Q4), and asking for a closed formula (Q1), asymptotic equation (Q1), and an efficient algorithm to compute these numbers, or a recursive algorithm; we'd also like to know the runtime of the algorithm (Q2). \n\nBefore reading on, you should try to investigate some of these questions for yourself. You could try computing some of these numbers by hand and drawing out a table, or writing a computer program to do it for you, and graphing the results. In particular, you could try Q3, and draw out a table for Q4 using either only even numbers, only odd numbers, or only distinct numbers. \n\n\n#### Partitions\n\n\nIf you're still reading this I'm assuming you either got bored of trying the question yourself, or you did it. Either way, here's how I approached the problem of counting the partitions of a natural number, where you are allowed repeated entries in the sum, and the order of the summands matters (i.e. question 3 above). \n\nTo get started, just work out the first few possibilities. I'll write n for the number we're testing, and r(n) for the number of ways of making n when you're allowed repeats:", "_____no_output_____" ], [ "| n | r(n) | Options |\n|-----|------|---------|\n| 1 | 1 | 1 |\n| 2 | 2 | 2, 1 + 1|\n| 3 | 4 | 3, 2 + 1, 1 + 2, 1 + 1 + 1|\n| 4 | 8 | 4, 3 + 1, 2 + 2, 2 + 1 + 1, 1 + 3 ...etc.|", "_____no_output_____" ], [ "It looks like the pattern here is powers of two - $ r(n) = 2 ^ {n-1} $. Great! Are we finished? Unfortunately, just working out the first 4 terms in a sequence and spotting a pattern doesn't really constitute a proof. That kind of logic can be pretty persuasive to set up a hypothesis, but for a mathematician, there's still plenty of work left to do. To be sure, we need to come up with a foolproof argument for why our guess for this formula of r(n) is right. \n\n\n#### Back to Partitions\nLet's define a set of functions:\n Given an input n (the symbol means \"is an element of\" and the N is the set of natural numbers, so 1, 2, 3, 4, ...etc.) let's define-\n \n - p(n) is the number of partitions of n\n - r(n) is the number of partitions of n where re-orderings are allowed\n - e(n) is the number of partitions of n using only even numbers\n - o(n) is the number of partitions of n using only odd numbers\n - d(n) is the number of partitions of n using only distinct numbers (no repeats allowed)\n\nWe've already worked out an exact formula for r(n) - it's 2 ** (n-1). So that's an open and shut case. Are there any other easy things we can think of? Well, e(n) is pretty obviously going to be 0 whenever n is odd (you can't make an odd number by adding together a bunch of even numbers), and when n is even, e(n) is just the same thing as p(n/2) (divide all your piles in half). So we have a complete formula for e(n) too: \n\ne(n) = { 0 if n is odd; p(n/2) if n is even;\n\nThis is another general mathematical principle - reducing one problem to another. I don't have a formula for e(n) but I do know that it's almost the same as p(n/2); so actually the e(n) problem is no harder or easier than working things out for p(n). \n\nWhat about these other ones though? I've drawn out a table below. \n\nHow weird. It looks like there are just as many ways of splitting a number up into odd parts as there are for splitting it up into distinct (ie. different) parts. That's pretty crazy. In fact it's so crazy that it's probably worth investigating a bit more. I've written a computer program which does this, and prints the results. The code is in the appendix, but the results are given here below. If you're keen, and know some coding, I'd suggest you write your own code to check this too. \n\n#### Summary\n\n#### Appendix - Some coding\nBeing able to code is a pretty useful skill for almost every technical or scientific field. If you know a little bit of coding, I would strongly encourage you to try to write programs that can calculate partition numbers. I've put some code in a github repository which you are free to mess with as much as you like. However, if you don't know any coding, you can still get started a bit. Just make sure you do the following steps:\n\n1. Open a new tab in your browser.\n2. Go on the three dots next to the url bar, click it and look for \"Developer Tools\". In Chrome, it's under \"More Tools\" > \"Developer Tools\" and the shortcut is Control + Shift + I. \n3. Click the Console tab. This opens an interactive console where you can type in any valid javascript code and the console will evaluate it for you. \n4. Try typing in some maths operations like 2 + 2, or 6 ** 3 (to run that line of code you need to hit enter). \n5. Type in the command: \n\n{% highlight javascript %}\nfetch().then(response => response.text).then()\n{% endhighlight %}\n\nThis copies the code from http://sekjsjf directly into the console. You can now use that code even if you don't understand exactly what it does! (This is 99% of programming.)\n6. Still in the console, type in Partitions.partition(5) and hit enter. The console should return 7 as the answer. But there are a whole load of other functions that you can try using. There's a handy autocomplete function, and a drop down, showing you a range of available options. Try typing in Partitions.partition.text(), Partitions.odd.text(), Partitions.distinct.text() to get a short description of what each function does. \n7. Try using some other commands. Writing Partitions.odd(5, list=True) will return [1,2,3,4,5]; i.e. rather than just calculating the answer for 5 it will run through all the numbers up to 5 and return the answers for each in an array. \n8. What about if we want to check that d(n) = o(n) for every single number up to 100? We could try writing some code like this:\n\n{% highlight javascript %}\nfor (let i = 1; i < 100; i = i + 1) {\n if (Partitions.odd(i) === Partitions.distinct(i)) {\n console.log(\"Same\")\n }\n else {\n console.log(\"These two are different!\")\n }\n}\n{% endhighlight %}\n\nThere are loads of good javascript tutorials (here's just one) but as a super crash course:\n\n- A single equals sign assigns a variable a value. i = 1 assigns the variable i a value of 1. \n- triple equals checks for equality. So something like (2 + 2 === 4) evaluates as true. \n- for (let i = 1; i < 100; i++) {do some stuff} looks more intimidating, but it's still pretty simple. We start by creating a new variable called i and giving it the value 1 (let i = 1;). Next we do whatever is in the curly brackets. Once that's done, we add 1 to the value of i (that's the i = i + 1, which looks weird, but remember that a single equals sign in javascript is an assignment, so really this is saying \"I want you to assign the variable i a new value, by taking its current value and adding 1 to it\"). Then we evaluate what's in the curly brackets, using the updated value for i. We keep doing this until i reaches 100 ( i < 100;), when we can stop. \n- if (something) {do this} else {do that} is pretty self explanatory. If the condition inside the if (something) is evaluated as true, we do what's inside the first set of brackets. Otherwise, we do whatever the else part says. \n- console.log(\"whatever\") just tells the console to print out that line. \n\nSo in the code here, we assign a new variable called i the value 1. Then we check if Partitions.odd(i) and Partitions.distinct(i) are equal, and if they are we print \"Same\", otherwise we print \"These two are different!\". Then we add 1 to the value of i (that's the i = i + 1 part) and do it again. We keep on doing this until i hits 100, when we can stop. One thing to note is that in javascript (and in programming in general) adding 1 to a variable is so common that it has a shorthand - rather than writing i = i + 1, you can write i++, which is more compact but means the same thing. ", "_____no_output_____" ] ] ]
[ "markdown" ]
[ [ "markdown", "markdown", "markdown" ] ]
eceaf3e199f8d6e090e416572afab9b5f5d82d59
899,481
ipynb
Jupyter Notebook
legacy/unidad_2/03_debug.ipynb
LVFerrero/diseno_sci_sfw
d782c0b6ccd5dbfd7c59fd2ac408207470e22301
[ "BSD-3-Clause" ]
1
2020-10-15T12:48:48.000Z
2020-10-15T12:48:48.000Z
legacy/unidad_2/03_debug.ipynb
LVFerrero/diseno_sci_sfw
d782c0b6ccd5dbfd7c59fd2ac408207470e22301
[ "BSD-3-Clause" ]
null
null
null
legacy/unidad_2/03_debug.ipynb
LVFerrero/diseno_sci_sfw
d782c0b6ccd5dbfd7c59fd2ac408207470e22301
[ "BSD-3-Clause" ]
null
null
null
5,518.288344
893,228
0.963797
[ [ [ "\n# Diseño de software para cómputo científico\n\n----\n\n## Unidad 2: Depuración de código\n", "_____no_output_____" ], [ "### Agenda de la Unidad 2\n---\n\n**Clase 1**\n \n - Calidad de software.\n - Principios de diseño: DRY y KISS \n - Refactoreo.\n\n**Clase 2**\n \n - Depuración de código.\n - Pruebas unitarias y funcionales con pytest.\n - Testing basados en propiedades (Hypothesis).\n\n**Clase 3**\n\n - Cobertura de código (codecov).\n - Perfilado de código\n", "_____no_output_____" ], [ "## Depuración\n\nLa depuración de programas es el proceso de identificar y corregir errores de programación. En inglés se conoce como debugging, porque se asemeja a la eliminación de bichos (bugs), manera en que se conoce informalmente a los errores de programación.\n\n![image.png](attachment:image.png)", "_____no_output_____" ], [ "## Dos depuradores principales en Python\n\n- Lo que vamos a ver es **Interactive debugging**\n- ``pdb`` (Python debugger) viene con python y es bastante tosco.\n- ``ipdb`` (IPython debugger) hay que instalarlo y es ultra cómodo.\n- En pytest siempre ejecutar con `-s`.", "_____no_output_____" ] ], [ [ "!pip install ipdb", "Requirement already satisfied: ipdb in /home/juan/proyectos/diseno_sci_sfw/lib/python3.7/site-packages (0.12.2)\nRequirement already satisfied: ipython>=5.1.0; python_version >= \"3.4\" in /home/juan/proyectos/diseno_sci_sfw/lib/python3.7/site-packages (from ipdb) (7.5.0)\nRequirement already satisfied: setuptools in /home/juan/proyectos/diseno_sci_sfw/lib/python3.7/site-packages (from ipdb) (41.0.1)\nRequirement already satisfied: pexpect; sys_platform != \"win32\" in /home/juan/proyectos/diseno_sci_sfw/lib/python3.7/site-packages (from ipython>=5.1.0; python_version >= \"3.4\"->ipdb) (4.7.0)\nRequirement already satisfied: pickleshare in /home/juan/proyectos/diseno_sci_sfw/lib/python3.7/site-packages (from ipython>=5.1.0; python_version >= \"3.4\"->ipdb) (0.7.5)\nRequirement already satisfied: prompt-toolkit<2.1.0,>=2.0.0 in /home/juan/proyectos/diseno_sci_sfw/lib/python3.7/site-packages (from ipython>=5.1.0; python_version >= \"3.4\"->ipdb) (2.0.9)\nRequirement already satisfied: pygments in /home/juan/proyectos/diseno_sci_sfw/lib/python3.7/site-packages (from ipython>=5.1.0; python_version >= \"3.4\"->ipdb) (2.4.2)\nRequirement already satisfied: traitlets>=4.2 in /home/juan/proyectos/diseno_sci_sfw/lib/python3.7/site-packages (from ipython>=5.1.0; python_version >= \"3.4\"->ipdb) (4.3.2)\nRequirement already satisfied: decorator in /home/juan/proyectos/diseno_sci_sfw/lib/python3.7/site-packages (from ipython>=5.1.0; python_version >= \"3.4\"->ipdb) (4.4.0)\nRequirement already satisfied: jedi>=0.10 in /home/juan/proyectos/diseno_sci_sfw/lib/python3.7/site-packages (from ipython>=5.1.0; python_version >= \"3.4\"->ipdb) (0.13.3)\nRequirement already satisfied: backcall in /home/juan/proyectos/diseno_sci_sfw/lib/python3.7/site-packages (from ipython>=5.1.0; python_version >= \"3.4\"->ipdb) (0.1.0)\nRequirement already satisfied: ptyprocess>=0.5 in /home/juan/proyectos/diseno_sci_sfw/lib/python3.7/site-packages (from pexpect; sys_platform != \"win32\"->ipython>=5.1.0; python_version >= \"3.4\"->ipdb) (0.6.0)\nRequirement already satisfied: six>=1.9.0 in /home/juan/proyectos/diseno_sci_sfw/lib/python3.7/site-packages (from prompt-toolkit<2.1.0,>=2.0.0->ipython>=5.1.0; python_version >= \"3.4\"->ipdb) (1.12.0)\nRequirement already satisfied: wcwidth in /home/juan/proyectos/diseno_sci_sfw/lib/python3.7/site-packages (from prompt-toolkit<2.1.0,>=2.0.0->ipython>=5.1.0; python_version >= \"3.4\"->ipdb) (0.1.7)\nRequirement already satisfied: ipython-genutils in /home/juan/proyectos/diseno_sci_sfw/lib/python3.7/site-packages (from traitlets>=4.2->ipython>=5.1.0; python_version >= \"3.4\"->ipdb) (0.2.0)\nRequirement already satisfied: parso>=0.3.0 in /home/juan/proyectos/diseno_sci_sfw/lib/python3.7/site-packages (from jedi>=0.10->ipython>=5.1.0; python_version >= \"3.4\"->ipdb) (0.4.0)\n\u001b[33mWARNING: You are using pip version 19.1.1, however version 19.2.3 is available.\nYou should consider upgrading via the 'pip install --upgrade pip' command.\u001b[0m\n" ] ], [ [ "## IPDB \nPara poner un break point hay que poner `import ipdb; ipdb.set_trace()`\n\nO si quieren que se active en las exceptions usar\n\n```python\nfrom ipdb import launch_ipdb_on_exception\n\nwith launch_ipdb_on_exception():\n ...\n```", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown" ]
[ [ "markdown", "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown" ] ]
eceb022e312c94df1abcfb4dc44570ee047f0ff6
28,440
ipynb
Jupyter Notebook
pml1/figure_notebooks/chapter1_introduction_figures.ipynb
hanchenresearch/pml-book
2c1b327b983bb7d9c57dd99491e24bd77b870d5a
[ "MIT" ]
1
2021-08-30T23:47:45.000Z
2021-08-30T23:47:45.000Z
pml1/figure_notebooks/chapter1_introduction_figures.ipynb
luaburto/pml-book
d28d516434f5fef847d1402ee3c39660c60815e1
[ "MIT" ]
null
null
null
pml1/figure_notebooks/chapter1_introduction_figures.ipynb
luaburto/pml-book
d28d516434f5fef847d1402ee3c39660c60815e1
[ "MIT" ]
1
2021-11-05T20:05:38.000Z
2021-11-05T20:05:38.000Z
31.529933
860
0.585267
[ [ [ "# Copyright 2021 Google LLC\n# Use of this source code is governed by an MIT-style\n# license that can be found in the LICENSE file or at\n# https://opensource.org/licenses/MIT.\n# Notebook authors: Kevin P. Murphy ([email protected])\n# and Mahmoud Soliman ([email protected])\n\n# This notebook reproduces figures for chapter 1 from the book\n# \"Probabilistic Machine Learning: An Introduction\"\n# by Kevin Murphy (MIT Press, 2021).\n# Book pdf is available from http://probml.ai", "_____no_output_____" ] ], [ [ "<a href=\"https://opensource.org/licenses/MIT\" target=\"_parent\"><img src=\"https://img.shields.io/github/license/probml/pyprobml\"/></a>", "_____no_output_____" ], [ "<a href=\"https://colab.research.google.com/github/probml/pml-book/blob/main/pml1/figure_notebooks/chapter1_introduction_figures.ipynb\" target=\"_parent\"><img src=\"https://colab.research.google.com/assets/colab-badge.svg\" alt=\"Open In Colab\"/></a>", "_____no_output_____" ], [ "## Figure 1.1:<a name='1.1'></a> <a name='iris'></a> ", "_____no_output_____" ], [ "\n Three types of Iris flowers: Setosa, Versicolor and Virginica. Used with kind permission of Dennis Kramb and SIGNA", "_____no_output_____" ] ], [ [ "#@title Click me to run setup { display-mode: \"form\" }\ntry:\n if PYPROBML_SETUP_ALREADY_RUN:\n print('skipping setup')\nexcept:\n PYPROBML_SETUP_ALREADY_RUN = True\n print('running setup...')\n !git clone --depth 1 https://github.com/probml/pyprobml /pyprobml &> /dev/null \n %cd -q /pyprobml/scripts\n %reload_ext autoreload \n %autoreload 2\n !pip install superimport deimport -qqq\n import superimport\n from deimport.deimport import deimport\n print('finished!')", "_____no_output_____" ] ], [ [ "<img src=\"https://raw.githubusercontent.com/probml/pml-book/main/pml1/figures/images/Figure_1.1_A.png\" width=\"256\"/>", "_____no_output_____" ], [ "<img src=\"https://raw.githubusercontent.com/probml/pml-book/main/pml1/figures/images/Figure_1.1_B.png\" width=\"256\"/>", "_____no_output_____" ], [ "<img src=\"https://raw.githubusercontent.com/probml/pml-book/main/pml1/figures/images/Figure_1.1_C.png\" width=\"256\"/>", "_____no_output_____" ], [ "## Figure 1.2:<a name='1.2'></a> <a name='cat'></a> ", "_____no_output_____" ], [ "\n Illustration of the image classification problem. From https://cs231n.github.io/ . Used with kind permission of Andrej Karpathy", "_____no_output_____" ] ], [ [ "#@title Click me to run setup { display-mode: \"form\" }\ntry:\n if PYPROBML_SETUP_ALREADY_RUN:\n print('skipping setup')\nexcept:\n PYPROBML_SETUP_ALREADY_RUN = True\n print('running setup...')\n !git clone --depth 1 https://github.com/probml/pyprobml /pyprobml &> /dev/null \n %cd -q /pyprobml/scripts\n %reload_ext autoreload \n %autoreload 2\n !pip install superimport deimport -qqq\n import superimport\n from deimport.deimport import deimport\n print('finished!')", "_____no_output_____" ] ], [ [ "<img src=\"https://raw.githubusercontent.com/probml/pml-book/main/pml1/figures/images/Figure_1.2.png\" width=\"256\"/>", "_____no_output_____" ], [ "## Figure 1.3:<a name='1.3'></a> <a name='irisPairs'></a> ", "_____no_output_____" ], [ "\n Visualization of the Iris data as a pairwise scatter plot. On the diagonal we plot the marginal distribution of each feature for each class. The off-diagonals contain scatterplots of all possible pairs of features. \nFigure(s) generated by [iris_plot.py](https://github.com/probml/pyprobml/blob/master/scripts/iris_plot.py) ", "_____no_output_____" ] ], [ [ "#@title Click me to run setup { display-mode: \"form\" }\ntry:\n if PYPROBML_SETUP_ALREADY_RUN:\n print('skipping setup')\nexcept:\n PYPROBML_SETUP_ALREADY_RUN = True\n print('running setup...')\n !git clone --depth 1 https://github.com/probml/pyprobml /pyprobml &> /dev/null \n %cd -q /pyprobml/scripts\n %reload_ext autoreload \n %autoreload 2\n !pip install superimport deimport -qqq\n import superimport\n from deimport.deimport import deimport\n print('finished!')", "_____no_output_____" ], [ "deimport(superimport)\n%run iris_plot.py", "_____no_output_____" ] ], [ [ "## Figure 1.4:<a name='1.4'></a> <a name='dtreeIrisDepth2'></a> ", "_____no_output_____" ], [ "\n Example of a decision tree of depth 2 applied to the Iris data, using just the petal length and petal width features. Leaf nodes are color coded according to the predicted class. The number of training samples that pass from the root to a node is shown inside each box; we show how many values of each class fall into this node. This vector of counts can be normalized to get a distribution over class labels for each node. We can then pick the majority class. Adapted from Figures 6.1 and 6.2 of <a href='#Geron2019'>[Aur19]</a> . ", "_____no_output_____" ], [ "To reproduce this figure, click the open in colab button: <a href=\"https://colab.research.google.com/github/probml/probml-notebooks/blob/master/notebooks/iris_dtree.ipynb\" target=\"_parent\"><img src=\"https://colab.research.google.com/assets/colab-badge.svg\" alt=\"Open In Colab\"/></a>", "_____no_output_____" ] ], [ [ "#@title Click me to run setup { display-mode: \"form\" }\ntry:\n if PYPROBML_SETUP_ALREADY_RUN:\n print('skipping setup')\nexcept:\n PYPROBML_SETUP_ALREADY_RUN = True\n print('running setup...')\n !git clone --depth 1 https://github.com/probml/pyprobml /pyprobml &> /dev/null \n %cd -q /pyprobml/scripts\n %reload_ext autoreload \n %autoreload 2\n !pip install superimport deimport -qqq\n import superimport\n from deimport.deimport import deimport\n print('finished!')", "_____no_output_____" ] ], [ [ "<img src=\"https://raw.githubusercontent.com/probml/pml-book/main/pml1/figures/images/Figure_1.4_A.png\" width=\"256\"/>", "_____no_output_____" ], [ "<img src=\"https://raw.githubusercontent.com/probml/pml-book/main/pml1/figures/images/Figure_1.4_B.png\" width=\"256\"/>", "_____no_output_____" ], [ "## Figure 1.5:<a name='1.5'></a> <a name='linreg'></a> ", "_____no_output_____" ], [ "\n(a) Linear regression on some 1d data. (b) The vertical lines denote the residuals between the observed output value for each input (blue circle) and its predicted value (red cross). The goal of least squares regression is to pick a line that minimizes the sum of squared residuals. \nFigure(s) generated by [linreg_residuals_plot.py](https://github.com/probml/pyprobml/blob/master/scripts/linreg_residuals_plot.py) ", "_____no_output_____" ] ], [ [ "#@title Click me to run setup { display-mode: \"form\" }\ntry:\n if PYPROBML_SETUP_ALREADY_RUN:\n print('skipping setup')\nexcept:\n PYPROBML_SETUP_ALREADY_RUN = True\n print('running setup...')\n !git clone --depth 1 https://github.com/probml/pyprobml /pyprobml &> /dev/null \n %cd -q /pyprobml/scripts\n %reload_ext autoreload \n %autoreload 2\n !pip install superimport deimport -qqq\n import superimport\n from deimport.deimport import deimport\n print('finished!')", "_____no_output_____" ], [ "deimport(superimport)\n%run linreg_residuals_plot.py", "_____no_output_____" ] ], [ [ "## Figure 1.6:<a name='1.6'></a> <a name='polyfit2d'></a> ", "_____no_output_____" ], [ "\n Linear and polynomial regression applied to 2d data. Vertical axis is temperature, horizontal axes are location within a room. Data was collected by some remote sensing motes at Intel's lab in Berkeley, CA (data courtesy of Romain Thibaux). (a) The fitted plane has the form $ f ( \\bm x ) = w_0 + w_1 x_1 + w_2 x_2$. (b) Temperature data is fitted with a quadratic of the form $ f ( \\bm x ) = w_0 + w_1 x_1 + w_2 x_2 + w_3 x_1^2 + w_4 x_2^2$. \nFigure(s) generated by [linreg_2d_surface_demo.py](https://github.com/probml/pyprobml/blob/master/scripts/linreg_2d_surface_demo.py) ", "_____no_output_____" ] ], [ [ "#@title Click me to run setup { display-mode: \"form\" }\ntry:\n if PYPROBML_SETUP_ALREADY_RUN:\n print('skipping setup')\nexcept:\n PYPROBML_SETUP_ALREADY_RUN = True\n print('running setup...')\n !git clone --depth 1 https://github.com/probml/pyprobml /pyprobml &> /dev/null \n %cd -q /pyprobml/scripts\n %reload_ext autoreload \n %autoreload 2\n !pip install superimport deimport -qqq\n import superimport\n from deimport.deimport import deimport\n print('finished!')", "_____no_output_____" ], [ "deimport(superimport)\n%run linreg_2d_surface_demo.py", "_____no_output_____" ] ], [ [ "## Figure 1.7:<a name='1.7'></a> <a name='linregPoly'></a> ", "_____no_output_____" ], [ "\n(a-c) Polynomials of degrees 2, 14 and 20 fit to 21 datapoints (the same data as in \\cref fig:linreg ). (d) MSE vs degree. \nFigure(s) generated by [linreg_poly_vs_degree.py](https://github.com/probml/pyprobml/blob/master/scripts/linreg_poly_vs_degree.py) ", "_____no_output_____" ] ], [ [ "#@title Click me to run setup { display-mode: \"form\" }\ntry:\n if PYPROBML_SETUP_ALREADY_RUN:\n print('skipping setup')\nexcept:\n PYPROBML_SETUP_ALREADY_RUN = True\n print('running setup...')\n !git clone --depth 1 https://github.com/probml/pyprobml /pyprobml &> /dev/null \n %cd -q /pyprobml/scripts\n %reload_ext autoreload \n %autoreload 2\n !pip install superimport deimport -qqq\n import superimport\n from deimport.deimport import deimport\n print('finished!')", "_____no_output_____" ], [ "deimport(superimport)\n%run linreg_poly_vs_degree.py", "_____no_output_____" ] ], [ [ "## Figure 1.8:<a name='1.8'></a> <a name='eqn:irisClustering'></a> ", "_____no_output_____" ], [ "\n(a) A scatterplot of the petal features from the iris dataset. (b) The result of unsupervised clustering using $K=3$. \nFigure(s) generated by [iris_kmeans.py](https://github.com/probml/pyprobml/blob/master/scripts/iris_kmeans.py) ", "_____no_output_____" ] ], [ [ "#@title Click me to run setup { display-mode: \"form\" }\ntry:\n if PYPROBML_SETUP_ALREADY_RUN:\n print('skipping setup')\nexcept:\n PYPROBML_SETUP_ALREADY_RUN = True\n print('running setup...')\n !git clone --depth 1 https://github.com/probml/pyprobml /pyprobml &> /dev/null \n %cd -q /pyprobml/scripts\n %reload_ext autoreload \n %autoreload 2\n !pip install superimport deimport -qqq\n import superimport\n from deimport.deimport import deimport\n print('finished!')", "_____no_output_____" ], [ "deimport(superimport)\n%run iris_kmeans.py", "_____no_output_____" ] ], [ [ "## Figure 1.9:<a name='1.9'></a> <a name='pcaDemo'></a> ", "_____no_output_____" ], [ "\n(a) Scatterplot of iris data (first 3 features). Points are color coded by class. (b) We fit a 2d linear subspace to the 3d data using PCA. The class labels are ignored. Red dots are the original data, black dots are points generated from the model using $ \\bm x = \\mathbf W \\bm z + \\bm \\mu $, where $ \\bm z $ are latent points on the underlying inferred 2d linear manifold. \nFigure(s) generated by [iris_pca.py](https://github.com/probml/pyprobml/blob/master/scripts/iris_pca.py) ", "_____no_output_____" ] ], [ [ "#@title Click me to run setup { display-mode: \"form\" }\ntry:\n if PYPROBML_SETUP_ALREADY_RUN:\n print('skipping setup')\nexcept:\n PYPROBML_SETUP_ALREADY_RUN = True\n print('running setup...')\n !git clone --depth 1 https://github.com/probml/pyprobml /pyprobml &> /dev/null \n %cd -q /pyprobml/scripts\n %reload_ext autoreload \n %autoreload 2\n !pip install superimport deimport -qqq\n import superimport\n from deimport.deimport import deimport\n print('finished!')", "_____no_output_____" ], [ "deimport(superimport)\n%run iris_pca.py", "_____no_output_____" ] ], [ [ "## Figure 1.10:<a name='1.10'></a> <a name='humanoid'></a> ", "_____no_output_____" ], [ "\n Examples of some control problems. (a) Space Invaders Atari game. From https://gym.openai.com/envs/SpaceInvaders-v0/ . (b) Controlling a humanoid robot in the MuJuCo simulator so it walks as fast as possible without falling over. From https://gym.openai.com/envs/Humanoid-v2/ ", "_____no_output_____" ] ], [ [ "#@title Click me to run setup { display-mode: \"form\" }\ntry:\n if PYPROBML_SETUP_ALREADY_RUN:\n print('skipping setup')\nexcept:\n PYPROBML_SETUP_ALREADY_RUN = True\n print('running setup...')\n !git clone --depth 1 https://github.com/probml/pyprobml /pyprobml &> /dev/null \n %cd -q /pyprobml/scripts\n %reload_ext autoreload \n %autoreload 2\n !pip install superimport deimport -qqq\n import superimport\n from deimport.deimport import deimport\n print('finished!')", "_____no_output_____" ] ], [ [ "<img src=\"https://raw.githubusercontent.com/probml/pml-book/main/pml1/figures/images/Figure_1.10_A.png\" width=\"256\"/>", "_____no_output_____" ], [ "<img src=\"https://raw.githubusercontent.com/probml/pml-book/main/pml1/figures/images/Figure_1.10_B.png\" width=\"256\"/>", "_____no_output_____" ], [ "## Figure 1.11:<a name='1.11'></a> <a name='cake'></a> ", "_____no_output_____" ], [ "\n The three types of machine learning visualized as layers of a chocolate cake. This figure (originally from https://bit.ly/2m65Vs1 ) was used in a talk by Yann LeCun at NIPS'16, and is used with his kind permission", "_____no_output_____" ] ], [ [ "#@title Click me to run setup { display-mode: \"form\" }\ntry:\n if PYPROBML_SETUP_ALREADY_RUN:\n print('skipping setup')\nexcept:\n PYPROBML_SETUP_ALREADY_RUN = True\n print('running setup...')\n !git clone --depth 1 https://github.com/probml/pyprobml /pyprobml &> /dev/null \n %cd -q /pyprobml/scripts\n %reload_ext autoreload \n %autoreload 2\n !pip install superimport deimport -qqq\n import superimport\n from deimport.deimport import deimport\n print('finished!')", "_____no_output_____" ] ], [ [ "<img src=\"https://raw.githubusercontent.com/probml/pml-book/main/pml1/figures/images/Figure_1.11.png\" width=\"256\"/>", "_____no_output_____" ], [ "## Figure 1.12:<a name='1.12'></a> <a name='emnist'></a> ", "_____no_output_____" ], [ "\n(a) Visualization of the MNIST dataset. Each image is $28 \\times 28$. There are 60k training examples and 10k test examples. We show the first 25 images from the training set. \nFigure(s) generated by [mnist_viz_tf.py](https://github.com/probml/pyprobml/blob/master/scripts/mnist_viz_tf.py) [emnist_viz_pytorch.py](https://github.com/probml/pyprobml/blob/master/scripts/emnist_viz_pytorch.py) ", "_____no_output_____" ] ], [ [ "#@title Click me to run setup { display-mode: \"form\" }\ntry:\n if PYPROBML_SETUP_ALREADY_RUN:\n print('skipping setup')\nexcept:\n PYPROBML_SETUP_ALREADY_RUN = True\n print('running setup...')\n !git clone --depth 1 https://github.com/probml/pyprobml /pyprobml &> /dev/null \n %cd -q /pyprobml/scripts\n %reload_ext autoreload \n %autoreload 2\n !pip install superimport deimport -qqq\n import superimport\n from deimport.deimport import deimport\n print('finished!')", "_____no_output_____" ], [ "deimport(superimport)\n%run mnist_viz_tf.py", "_____no_output_____" ], [ "deimport(superimport)\n%run emnist_viz_pytorch.py", "_____no_output_____" ] ], [ [ "## Figure 1.13:<a name='1.13'></a> <a name='CIFAR'></a> ", "_____no_output_____" ], [ "\n(a) Visualization of the Fashion-MNIST dataset <a href='#fashion'>[XRV17]</a> . The dataset has the same size as MNIST, but is harder to classify. There are 10 classes: T-shirt/top, Trouser, Pullover, Dress, Coat, Sandal, Shirt, Sneaker, Bag, Ankle-boot. We show the first 25 images from the training set. \nFigure(s) generated by [fashion_viz_tf.py](https://github.com/probml/pyprobml/blob/master/scripts/fashion_viz_tf.py) [cifar_viz_tf.py](https://github.com/probml/pyprobml/blob/master/scripts/cifar_viz_tf.py) ", "_____no_output_____" ] ], [ [ "#@title Click me to run setup { display-mode: \"form\" }\ntry:\n if PYPROBML_SETUP_ALREADY_RUN:\n print('skipping setup')\nexcept:\n PYPROBML_SETUP_ALREADY_RUN = True\n print('running setup...')\n !git clone --depth 1 https://github.com/probml/pyprobml /pyprobml &> /dev/null \n %cd -q /pyprobml/scripts\n %reload_ext autoreload \n %autoreload 2\n !pip install superimport deimport -qqq\n import superimport\n from deimport.deimport import deimport\n print('finished!')", "_____no_output_____" ], [ "deimport(superimport)\n%run fashion_viz_tf.py", "_____no_output_____" ], [ "deimport(superimport)\n%run cifar_viz_tf.py", "_____no_output_____" ] ], [ [ "## Figure 1.14:<a name='1.14'></a> <a name='imagenetError'></a> ", "_____no_output_____" ], [ "\n(a) Sample images from the \\bf ImageNet dataset <a href='#ILSVRC15'>[Rus+15]</a> . This subset consists of 1.3M color training images, each of which is $256 \\times 256$ pixels in size. There are 1000 possible labels, one per image, and the task is to minimize the top-5 error rate, i.e., to ensure the correct label is within the 5 most probable predictions. Below each image we show the true label, and a distribution over the top 5 predicted labels. If the true label is in the top 5, its probability bar is colored red. Predictions are generated by a convolutional neural network (CNN) called ``AlexNet'' (\\cref sec:alexNet ). From Figure 4 of <a href='#Krizhevsky12'>[KSH12]</a> . Used with kind permission of Alex Krizhevsky. (b) Misclassification rate (top 5) on the ImageNet competition over time. Used with kind permission of Andrej Karpathy", "_____no_output_____" ] ], [ [ "#@title Click me to run setup { display-mode: \"form\" }\ntry:\n if PYPROBML_SETUP_ALREADY_RUN:\n print('skipping setup')\nexcept:\n PYPROBML_SETUP_ALREADY_RUN = True\n print('running setup...')\n !git clone --depth 1 https://github.com/probml/pyprobml /pyprobml &> /dev/null \n %cd -q /pyprobml/scripts\n %reload_ext autoreload \n %autoreload 2\n !pip install superimport deimport -qqq\n import superimport\n from deimport.deimport import deimport\n print('finished!')", "_____no_output_____" ] ], [ [ "<img src=\"https://raw.githubusercontent.com/probml/pml-book/main/pml1/figures/images/Figure_1.14_A.png\" width=\"256\"/>", "_____no_output_____" ], [ "<img src=\"https://raw.githubusercontent.com/probml/pml-book/main/pml1/figures/images/Figure_1.14_B.png\" width=\"256\"/>", "_____no_output_____" ], [ "## Figure 1.15:<a name='1.15'></a> <a name='termDoc'></a> ", "_____no_output_____" ], [ "\n Example of a term-document matrix, where raw counts have been replaced by their TF-IDF values (see \\cref sec:tfidf ). Darker cells are larger values. From https://bit.ly/2kByLQI . Used with kind permission of Christoph Carl Kling", "_____no_output_____" ] ], [ [ "#@title Click me to run setup { display-mode: \"form\" }\ntry:\n if PYPROBML_SETUP_ALREADY_RUN:\n print('skipping setup')\nexcept:\n PYPROBML_SETUP_ALREADY_RUN = True\n print('running setup...')\n !git clone --depth 1 https://github.com/probml/pyprobml /pyprobml &> /dev/null \n %cd -q /pyprobml/scripts\n %reload_ext autoreload \n %autoreload 2\n !pip install superimport deimport -qqq\n import superimport\n from deimport.deimport import deimport\n print('finished!')", "_____no_output_____" ] ], [ [ "<img src=\"https://raw.githubusercontent.com/probml/pml-book/main/pml1/figures/images/Figure_1.15.png\" width=\"256\"/>", "_____no_output_____" ], [ "## References:\n <a name='Geron2019'>[Aur19]</a> G. Aur'elien \"Hands-On Machine Learning with Scikit-Learn and TensorFlow: Concepts, Tools, and Techniques for BuildingIntelligent Systems (2nd edition)\". (2019). \n\n<a name='Krizhevsky12'>[KSH12]</a> A. Krizhevsky, I. Sutskever and G. Hinton. \"Imagenet classification with deep convolutional neural networks\". (2012). \n\n<a name='ILSVRC15'>[Rus+15]</a> O. Russakovsky, J. Deng, H. Su, J. Krause, S. Satheesh, S. Ma, Z. Huang, A. Karpathy, A. Khosla, M. Bernstein, A. Berg and L. Fei-Fei. \"ImageNet Large Scale Visual Recognition Challenge\". In: ijcv (2015). \n\n<a name='fashion'>[XRV17]</a> H. Xiao, K. Rasul and R. Vollgraf. \"Fashion-MNIST: a Novel Image Dataset for Benchmarking Machine Learning Algorithms\". abs/1708.07747 (2017). arXiv: 1708.07747 \n\n", "_____no_output_____" ] ] ]
[ "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown" ]
[ [ "code" ], [ "markdown", "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown", "markdown", "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown", "markdown", "markdown" ], [ "code", "code" ], [ "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown", "markdown", "markdown", "markdown" ], [ "code", "code" ], [ "markdown", "markdown" ], [ "code", "code" ], [ "markdown", "markdown" ], [ "code", "code" ], [ "markdown", "markdown" ], [ "code", "code" ], [ "markdown", "markdown" ], [ "code", "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown", "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown", "markdown", "markdown" ], [ "code", "code", "code" ], [ "markdown", "markdown" ], [ "code", "code", "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown", "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown", "markdown" ] ]
eceb0fc2aebb27f7e9669f84aa89cad2910eccc1
782,457
ipynb
Jupyter Notebook
notebooks/.ipynb_checkpoints/Math156_classification_MNIST-checkpoint_BACKUP_721.ipynb
tylerwu2222/Math156_UCLA_SP21
9eb64b1d1bb47837d7a67a0dc5c3f4ed25d6dfd1
[ "MIT" ]
null
null
null
notebooks/.ipynb_checkpoints/Math156_classification_MNIST-checkpoint_BACKUP_721.ipynb
tylerwu2222/Math156_UCLA_SP21
9eb64b1d1bb47837d7a67a0dc5c3f4ed25d6dfd1
[ "MIT" ]
null
null
null
notebooks/.ipynb_checkpoints/Math156_classification_MNIST-checkpoint_BACKUP_721.ipynb
tylerwu2222/Math156_UCLA_SP21
9eb64b1d1bb47837d7a67a0dc5c3f4ed25d6dfd1
[ "MIT" ]
null
null
null
420.449758
157,976
0.92818
[ [ [ "empty" ] ] ]
[ "empty" ]
[ [ "empty" ] ]
eceb23e18fd06fa7851c2e85ed96621760319ee1
73,952
ipynb
Jupyter Notebook
titanic/Titanic-Derived.ipynb
Ravirajadrangi/projects6
bc7dc98a01dabe4b64d7de7738521746b09426cd
[ "BSD-2-Clause" ]
24
2015-12-16T14:44:05.000Z
2021-04-03T12:11:47.000Z
titanic/Titanic-Derived.ipynb
Ravirajadrangi/projects6
bc7dc98a01dabe4b64d7de7738521746b09426cd
[ "BSD-2-Clause" ]
null
null
null
titanic/Titanic-Derived.ipynb
Ravirajadrangi/projects6
bc7dc98a01dabe4b64d7de7738521746b09426cd
[ "BSD-2-Clause" ]
8
2017-03-31T17:36:03.000Z
2020-01-11T15:59:13.000Z
41.546067
13,299
0.42936
[ [ [ "empty" ] ] ]
[ "empty" ]
[ [ "empty" ] ]
eceb2a85706c1694a13b414e4b8eac9150ed522c
24,173
ipynb
Jupyter Notebook
Survey1Theory_results.ipynb
christopherkullenberg/KT2102Evaluations
dbd7789de37268e8240091609a1115019940879c
[ "MIT" ]
null
null
null
Survey1Theory_results.ipynb
christopherkullenberg/KT2102Evaluations
dbd7789de37268e8240091609a1115019940879c
[ "MIT" ]
null
null
null
Survey1Theory_results.ipynb
christopherkullenberg/KT2102Evaluations
dbd7789de37268e8240091609a1115019940879c
[ "MIT" ]
null
null
null
50.677149
4,600
0.663137
[ [ [ "import pandas as pd\nfrom collections import Counter", "_____no_output_____" ] ], [ [ "## Survey 1, Theory\n\n**Coding**\n\n*1. Vilket/vilka moment upplevde du vara svårast i Inlämningsuppgift 1? Ange korta nyckelord, ex. \"litteraturen\", \"föreläsningarna\", osv.*\n\n str\n\n*2. Vilka insikter/kunskaper upplever du att du fått under de teoretiska momenten? Ange korta nyckelord.*\n\n str\n\n*3. Upplevde du att du hade all information som krävdes för att lösa inlämningsuppgift 1? Om inte, vilken information tycker du saknades?*7\n\n str\n\n*4. Hur relevant upplever du att de teoretiska momenten var för kursen Kommunikation i nya och sociala medier? Ringa in en siffra på skalan 0-10 (0-inte alls relevant, 10-mycket relevant)*\n\n int 0-10\n\n*5. Hur relevant upplever du att de teoretiska momenten var för yrkesrollen kommunikatör i offentlig förvaltning? Ringa in en siffra på skalan 0-10 (0-inte alls relevant, 10-mycket relevant)*\n\n int 0-10\n \n\nThe scanned questionnaires can be found on [https://github.com/christopherkullenberg/KT2102Evaluations/tree/master/raw_survey_data/Survey1Theory](https://github.com/christopherkullenberg/KT2102Evaluations/tree/master/raw_survey_data/Survey1Theory)", "_____no_output_____" ] ], [ [ "df1 = pd.read_csv(\"data/survey_theory.csv\")\nprint(\"N =\", str(len(df1)))", "N = 17\n" ], [ "df1.head()", "_____no_output_____" ], [ "Q1results = []\n\nprint(\"Showing results for: \")\nprint(\"Vilket/vilka moment upplevde du vara svårast i Inlämningsuppgift 1?\")\nfor row in df1.iloc[:, 1:4]:\n print(row)\n for n in df1[row]:\n if type(n) == str:\n Q1results.append(n)\n \nprint(\"\\nResult: \", Q1results)\nprint(\"\\nKeyword frequencies:\")\nfor freq in Counter(Q1results).most_common():\n print(freq[0], \" - \", freq[1])", "Showing results for: \nVilket/vilka moment upplevde du vara svårast i Inlämningsuppgift 1?\nQ1A\nQ1B\nQ1C\n\nResult: ['hålla det kort', 'föreläsningar svävande', 'formulera fråga', 'artiklarna', 'källhänvisningar', 'hålla det kort', 'beskrivning av uppgift', 'artiklarna', 'upplägget', 'begrepp', 'nivå', 'litteratur', 'exemplifiera', 'hålla det kort', 'litteratur', 'välja fråga', 'hinna med att läsa', 'exemplifiera', 'hålla det kort', 'akademisk nivå', 'litteratur', 'hålla det kort', 'svårt att hitta djup I litteratur', 'målgrupp', 'stor valfrihet']\n\nKeyword frequencies:\nhålla det kort - 5\nlitteratur - 3\nartiklarna - 2\nexemplifiera - 2\nföreläsningar svävande - 1\nformulera fråga - 1\nkällhänvisningar - 1\nbeskrivning av uppgift - 1\nupplägget - 1\nbegrepp - 1\nnivå - 1\nvälja fråga - 1\nhinna med att läsa - 1\nakademisk nivå - 1\nsvårt att hitta djup I litteratur - 1\nmålgrupp - 1\nstor valfrihet - 1\n" ], [ "Q2results = []\n\nprint(\"Showing results for: \")\nprint(\"Vilka insikter/kunskaper upplever du att du fått under de teoretiska momenten?\")\nfor row in df1.iloc[:, 4:7]:\n print(row)\n for n in df1[row]:\n if type(n) == str:\n Q2results.append(n)\n \nprint(\"\\nResult: \", Q2results)\nprint(\"\\nKeyword frequencies:\")\nfor freq in Counter(Q2results).most_common():\n print(freq[0], \" - \", freq[1])", "Showing results for: \nVilka insikter/kunskaper upplever du att du fått under de teoretiska momenten?\nQ2A\nQ2B\nQ2C\n\nResult: ['begrepp', 'begrepp', 'förankring I forskning', 'begrepp', 'sociala medier I organisationer', 'digitalt samhälle', 'begrepp', 'perspektiv', 'anpassning till teknologi', 'begrepp', 'förståelse', 'digitalt samhälle', 'begrepp', 'insikt I huvudämnet', 'digitalt samhälle', 'Lindgrens bok', 'digital forskning', 'fördelar och nackdelar med sociala medier', 'kommunikation I sociala medier', 'samverkan teknik och människa', 'källhänvisningar', 'hur man påverkas av sociala medier', 'perspektiv', 'perspektiv', 'seminarium', 'förståelse för forskning']\n\nKeyword frequencies:\nbegrepp - 6\ndigitalt samhälle - 3\nperspektiv - 3\nförankring I forskning - 1\nsociala medier I organisationer - 1\nanpassning till teknologi - 1\nförståelse - 1\ninsikt I huvudämnet - 1\nLindgrens bok - 1\ndigital forskning - 1\nfördelar och nackdelar med sociala medier - 1\nkommunikation I sociala medier - 1\nsamverkan teknik och människa - 1\nkällhänvisningar - 1\nhur man påverkas av sociala medier - 1\nseminarium - 1\nförståelse för forskning - 1\n" ], [ "Q3results = []\n\nprint(\"Showing results for: \")\nprint(\"Upplevde du att du hade all information som krävdes för att lösa inlämningsuppgift 1?\")\nfor row in df1.iloc[:, 7:10]:\n print(row)\n for n in df1[row]:\n if type(n) == str:\n Q3results.append(n)\n \nprint(\"\\nResult: \", Q3results)\nprint(\"\\nKeyword frequencies:\")\nfor freq in Counter(Q3results).most_common():\n print(freq[0], \" - \", freq[1])", "Showing results for: \nUpplevde du att du hade all information som krävdes för att lösa inlämningsuppgift 1?\nQ3A\nQ3B\nQ3C\n\nResult: ['ja', 'källhänvisningar', 'ja', 'ja', 'tydlig information om examination', 'källhänvisningar', 'ja', 'ja', 'tidsbrist på föreläsningar', 'oklart om exemplifiering', 'ja', 'ja', 'ja', 'ja', 'antal tecken ospecificerat', 'mer info om föreläsningar', 'ja', 'oklart om målgrupp', 'mer genomgång av begrepp']\n\nKeyword frequencies:\nja - 10\nkällhänvisningar - 2\ntydlig information om examination - 1\ntidsbrist på föreläsningar - 1\noklart om exemplifiering - 1\nantal tecken ospecificerat - 1\nmer info om föreläsningar - 1\noklart om målgrupp - 1\nmer genomgång av begrepp - 1\n" ], [ "print(\"Hur relevant upplever du att de teoretiska momenten var för kursen Kommunikation i nya och sociala medier?\")\ndf1.Q4.value_counts().plot.bar(x=\"Score\", rot=0, subplots=True)\nprint(\"Mean = \" + str(df1.Q4.mean()))\nprint(\"Median = \" + str(df1.Q4.median()))\nprint(\"Std = \" + str(df1.Q4.std()))", "Hur relevant upplever du att de teoretiska momenten var för kursen Kommunikation i nya och sociala medier?\nMean = 9.352941176470589\nMedian = 10.0\nStd = 0.9963167462326074\n" ], [ "print(\"Hur relevant upplever du att de teoretiska momenten var för yrkesrollen kommunikatör i offentlig förvaltning?\")\ndf1.Q5.value_counts().plot.bar(x=\"Score\", rot=0, subplots=True)\nprint(\"Mean = \" + str(df1.Q5.mean()))\nprint(\"Median = \" + str(df1.Q5.median()))\nprint(\"Std = \" + str(df1.Q5.std()))", "Hur relevant upplever du att de teoretiska momenten var för yrkesrollen kommunikatör i offentlig förvaltning?\nMean = 8.764705882352942\nMedian = 10.0\nStd = 1.8550408272025267\n" ] ] ]
[ "code", "markdown", "code" ]
[ [ "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code", "code", "code" ] ]
eceb2bd6750c94f4db2d8b51baef2992b1e37c27
70,366
ipynb
Jupyter Notebook
python-for-data-visualization-projects/geographical-plotting/.ipynb_checkpoints/Choropleth Maps-checkpoint.ipynb
niccololampa/data-science-projects
f1d7c208ad59c5e54614af17210cad997bb6fee5
[ "MIT" ]
null
null
null
python-for-data-visualization-projects/geographical-plotting/.ipynb_checkpoints/Choropleth Maps-checkpoint.ipynb
niccololampa/data-science-projects
f1d7c208ad59c5e54614af17210cad997bb6fee5
[ "MIT" ]
null
null
null
python-for-data-visualization-projects/geographical-plotting/.ipynb_checkpoints/Choropleth Maps-checkpoint.ipynb
niccololampa/data-science-projects
f1d7c208ad59c5e54614af17210cad997bb6fee5
[ "MIT" ]
null
null
null
43.302154
7,472
0.470682
[ [ [ "___\n\n<a href='http://www.pieriandata.com'> <img src='../Pierian_Data_Logo.png' /></a>\n___", "_____no_output_____" ], [ "# Choropleth Maps", "_____no_output_____" ], [ "Focus will be on plotly for plotting\n\nBUt matplotlib has a basemap extension which provides for static plotting. ", "_____no_output_____" ], [ "## Offline Plotly Usage", "_____no_output_____" ], [ "Get imports and set everything up to be working offline.", "_____no_output_____" ] ], [ [ "import plotly.plotly as py\nimport plotly.graph_objs as go \nfrom plotly.offline import download_plotlyjs, init_notebook_mode, plot, iplot #tobe able to do it offline. ", "_____no_output_____" ] ], [ [ "Now set up everything so that the figures show up in the notebook:", "_____no_output_____" ] ], [ [ "init_notebook_mode(connected=True) #to see everything in the notebook. ", "_____no_output_____" ] ], [ [ "More info on other options for Offline Plotly usage can be found [here](https://plot.ly/python/offline/).", "_____no_output_____" ], [ "https://plot.ly/python/reference/#choropleth #reference", "_____no_output_____" ], [ "## Choropleth US Maps\n\nPlotly's mapping can be a bit hard to get used to at first, remember to reference the cheat sheet in the data visualization folder, or [find it online here](https://images.plot.ly/plotly-documentation/images/python_cheat_sheet.pdf).", "_____no_output_____" ] ], [ [ "import pandas as pd #important to see cheat sheet above link. ", "_____no_output_____" ] ], [ [ "Now we need to begin to build our data dictionary. Easiest way to do this is to use the **dict()** function of the general form:\n\n* type = 'choropleth',\n* locations = list of states\n* locationmode = 'USA-states'\n* colorscale= \n\nEither a predefined string:\n\n 'pairs' | 'Greys' | 'Greens' | 'Bluered' | 'Hot' | 'Picnic' | 'Portland' | 'Jet' | 'RdBu' | 'Blackbody' | 'Earth' | 'Electric' | 'YIOrRd' | 'YIGnBu'\n\nor create a [custom colorscale](https://plot.ly/python/heatmap-and-contour-colorscales/)\n\n* text= list or array of text to display per point\n* z= array of values on z axis (color of state)\n* colorbar = {'title':'Colorbar Title'})\n\nHere is a simple example:", "_____no_output_____" ] ], [ [ "data = dict(type = 'choropleth', #clarifies which type of geogrhapical plot we are doing. \n locations = ['AZ','CA','NY'], # actual abbreviation of states since we are doing this only on a nationwide scale. \n locationmode = 'USA-states', # essential. let's plotly know that we are doing this on a US location level. check documentation. This can go down into county level. Check documentation. \n colorscale= 'Portland', #can change the coloring style\n text= ['text1','text2','text3'], #text is what shows when mouse hovers on a particular location in graph.\n z=[1.0,2.0,3.0], #z actual values that will be shown in the color scale. \n colorbar = {'title':'Colorbar Title'}) #colorbar config here it is jsut the title. \n\n\n#documentattion likes to call dict rather than typing the actual dictionary to save time. ", "_____no_output_____" ], [ "data #preview of data", "_____no_output_____" ] ], [ [ "Then we create the layout nested dictionary:", "_____no_output_____" ] ], [ [ "layout = dict(geo = {'scope':'usa'})", "_____no_output_____" ] ], [ [ "Then we use: \n\n go.Figure(data = [data],layout = layout)\n \nto set up the object that finally gets passed into iplot()", "_____no_output_____" ] ], [ [ "choromap = go.Figure(data = [data],layout = layout)", "_____no_output_____" ], [ "iplot(choromap)\n\n#if you use plot(choromap) instead it will open in a new tab as an html where you can save it as png etc. ", "_____no_output_____" ] ], [ [ "### Real Data US Map Choropleth\n\nNow let's show an example with some real data as well as some other options we can add to the dictionaries in data and layout.", "_____no_output_____" ] ], [ [ "df = pd.read_csv('2011_US_AGRI_Exports')\ndf.head()", "_____no_output_____" ] ], [ [ "Now out data dictionary with some extra marker and colorbar arguments:", "_____no_output_____" ] ], [ [ "data = dict(type='choropleth',\n colorscale = 'YIOrRd',\n locations = df['code'], #referencing a column in a dataframe. list of state codes. \n z = df['total exports'],\n locationmode = 'USA-states',\n text = df['text'],\n marker = dict(line = dict(color = 'rgb(255,255,255)',width = 2)), #space width between the states and the color. \n colorbar = {'title':\"Millions USD\"}\n ) ", "_____no_output_____" ] ], [ [ "And our layout dictionary with some more arguments:", "_____no_output_____" ] ], [ [ "layout = dict(title = '2011 US Agriculture Exports by State', # you can set the ttitle. \n geo = dict(scope='usa',\n showlakes = True, #you can add to show lakes (optional)\n lakecolor = 'rgb(85,173,240)') #color of the lakes (optional)\n )", "_____no_output_____" ], [ "choromap = go.Figure(data = [data],layout = layout)", "_____no_output_____" ], [ "iplot(choromap)", "_____no_output_____" ] ], [ [ "# World Choropleth Map\n\nNow let's see an example with a World Map:", "_____no_output_____" ] ], [ [ "df = pd.read_csv('2014_World_GDP')\ndf.head()", "_____no_output_____" ], [ "data = dict(\n type = 'choropleth',\n locations = df['CODE'],\n z = df['GDP (BILLIONS)'],\n text = df['COUNTRY'],\n colorbar = {'title' : 'GDP Billions US'},\n ) ", "_____no_output_____" ], [ "layout = dict(\n title = '2014 Global GDP',\n geo = dict(\n showframe = False, # for world no scope were passed unlike above\n projection = {'type':'Mercator'} #reference the documentation for the projection type. For now mercator was used for map graph. \n )\n)", "_____no_output_____" ], [ "choromap = go.Figure(data = [data],layout = layout)\niplot(choromap)", "_____no_output_____" ] ], [ [ "# Great Job!", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown" ]
[ [ "markdown", "markdown", "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code" ], [ "markdown" ] ]
eceb34feb76f8498a804b6be5dac404077dea5d7
27,081
ipynb
Jupyter Notebook
notebooks/scratch03-essentiality_range.ipynb
pritchardlabatpsu/cga
0a71c672b1348cebc724560643fd908d636fc133
[ "MIT" ]
null
null
null
notebooks/scratch03-essentiality_range.ipynb
pritchardlabatpsu/cga
0a71c672b1348cebc724560643fd908d636fc133
[ "MIT" ]
null
null
null
notebooks/scratch03-essentiality_range.ipynb
pritchardlabatpsu/cga
0a71c672b1348cebc724560643fd908d636fc133
[ "MIT" ]
1
2022-02-08T01:06:20.000Z
2022-02-08T01:06:20.000Z
45.822335
12,488
0.705513
[ [ [ "Getting processed genomic/ceres datasets", "_____no_output_____" ] ], [ [ "import pickle\nimport sys\nimport pandas as pd\nimport seaborn as sns\nimport matplotlib.pyplot as plt", "_____no_output_____" ], [ "dm_data = pickle.load(open('../out/20.0817 proc_data/gene_effect/dm_data.pkl','rb'))", "_____no_output_____" ], [ "dm_data.df_crispr.to_csv('ceres_processed.csv') # for analyses in R later", "_____no_output_____" ], [ "#get gene names\ndm_data.df_crispr.columns = dm_data.df_crispr.columns.str.extract('^(.*)\\s\\(').squeeze().values", "_____no_output_____" ], [ "df = dm_data.df_crispr", "_____no_output_____" ], [ "df_genedep = pd.read_csv('%s/%s' % (dm_data.dir_datasets, dm_data.fname_gene_dependency), header=0, index_col=0)\ndf_genedep.columns = df_genedep.columns.str.extract('^(.*)\\s').squeeze().values", "_____no_output_____" ], [ "# get which ones are selective essential, and which ones are common essential, common nonessential\ndef classifyDep(x):\n if all(x>0.5):\n return 'common_essential'\n elif all(x<0.5):\n return 'common_nonessential'\n else:\n return 'selective_essential'\n\ndep_class = df_genedep.apply(lambda x: classifyDep(x), axis=0)", "_____no_output_____" ], [ "dep_class.to_csv('ceres_class.csv') # for analyses in R later", "_____no_output_____" ], [ "df_common_essential = df.loc[:,df.columns.isin(dep_class[dep_class=='common_essential'].index)]\ndf_common_nonessential = df.loc[:,df.columns.isin(dep_class[dep_class=='common_nonessential'].index)]\ndf_selective_essential = df.loc[:,df.columns.isin(dep_class[dep_class=='selective_essential'].index)]", "_____no_output_____" ], [ "df1 = -1*df_common_essential.kurtosis()\ndf1.name = 'neg_kurtosis'\ndf2 = (df_common_essential.max()-df_common_essential.min())\ndf2.name = 'range'\ndf3 = pd.merge(df1,df2, left_index=True, right_index=True)", "_____no_output_____" ], [ "df3.sort_values(['range', 'neg_kurtosis'])", "_____no_output_____" ], [ "# common essential, high kurtosis\ndf_common_essential.kurtosis().sort_values()[-5:]", "_____no_output_____" ], [ "# common essential, low kurtosis\ndf_common_essential.kurtosis().sort_values()[:5]", "_____no_output_____" ], [ "# common selective-essential, high kurtosis\ndf_selective_essential.kurtosis().sort_values()[-5:]", "_____no_output_____" ], [ "# common selective-essential, low kurtosis\ndf_selective_essential.kurtosis().sort_values()[:5]", "_____no_output_____" ], [ "# common essential, high range\n(df_common_essential.max()-df_common_essential.min()).sort_values()[-5:]", "_____no_output_____" ], [ "# common essential, tight range\n(df_common_essential.max()-df_common_essential.min()).sort_values()[:5]", "_____no_output_____" ], [ "# common selective-essential, high range\n(df_selective_essential.max()-df_selective_essential.min()).sort_values()[-5:]", "_____no_output_____" ], [ "# common selective-essential, tight range\n(df_selective_essential.max()-df_selective_essential.min()).sort_values()[:5]", "_____no_output_____" ], [ "sns.distplot(df_range_merged.range)", "_____no_output_____" ], [ "df_range = (df_selective_essential.max()-df_selective_essential.min())\ndf_range.name = 'range'\ndep_class.name = 'dep_class'\ndf_range_merged = pd.merge(df_range,dep_class, left_index=True, right_index=True)", "_____no_output_____" ], [ "sns.distplot(data=df_range_merged, x='range', hue='dep_class')", "_____no_output_____" ] ] ]
[ "markdown", "code" ]
[ [ "markdown" ], [ "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code" ] ]
eceb39ba77c64ca784232e2c7bdbf5509f4ccc5a
806,753
ipynb
Jupyter Notebook
codici/mnist_keras.ipynb
tvml/ml2021
d72a6762af9cd12019d87237d061bbb39f560da9
[ "MIT" ]
null
null
null
codici/mnist_keras.ipynb
tvml/ml2021
d72a6762af9cd12019d87237d061bbb39f560da9
[ "MIT" ]
null
null
null
codici/mnist_keras.ipynb
tvml/ml2021
d72a6762af9cd12019d87237d061bbb39f560da9
[ "MIT" ]
null
null
null
264.943514
367,438
0.888687
[ [ [ "%tensorflow_version 2.x", "_____no_output_____" ], [ "%matplotlib inline\nimport pandas as pd\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom pylab import subplot,imshow,title,gray,NullLocator\nimport scipy.misc as mi\nimport scipy.special as sp\nfrom PIL import Image\nfrom itertools import chain\nfrom sklearn.metrics import confusion_matrix, accuracy_score, precision_recall_fscore_support\n\nfrom keras.datasets import mnist\nfrom keras.models import Sequential\nfrom keras.layers.core import Dense, Dropout, Activation\nfrom keras.utils import np_utils\nfrom keras.models import model_from_json\n\nfrom keras.layers import Flatten\nfrom keras.layers.convolutional import Conv2D\nfrom keras.layers.convolutional import MaxPooling2D\nfrom keras.utils import np_utils\nfrom keras import backend as K", "_____no_output_____" ], [ "from google.colab import drive\ndrive.mount('/gdrive')", "Go to this URL in a browser: https://accounts.google.com/o/oauth2/auth?client_id=947318989803-6bn6qk8qdgf4n4g3pfee6491hc0brc4i.apps.googleusercontent.com&redirect_uri=urn%3aietf%3awg%3aoauth%3a2.0%3aoob&response_type=code&scope=email%20https%3a%2f%2fwww.googleapis.com%2fauth%2fdocs.test%20https%3a%2f%2fwww.googleapis.com%2fauth%2fdrive%20https%3a%2f%2fwww.googleapis.com%2fauth%2fdrive.photos.readonly%20https%3a%2f%2fwww.googleapis.com%2fauth%2fpeopleapi.readonly\n\nEnter your authorization code:\n··········\nMounted at /gdrive\n" ], [ "# visualizza dati\ndef displayData(X, t, rows=10, cols=10, img_ind=None, size =16, class_value = False):\n if len(X)>rows*cols:\n img_ind = np.random.permutation(len(X))[0:rows * cols]\n else:\n img_ind = range(rows*cols)\n fig = plt.figure(figsize = (size,size))\n fig.patch.set_facecolor('white')\n ax = fig.gca()\n for i in range(100):\n plt.subplot(10,10,i+1)\n plt.imshow([255-x for x in X[img_ind[i]]], cmap='gray', interpolation='gaussian')\n if class_value:\n plt.title(\"{}\".format(t[img_ind[i]]),fontsize = 16, color='b')\n plt.gca().xaxis.set_major_locator(plt.NullLocator())\n plt.gca().yaxis.set_major_locator(plt.NullLocator())\n plt.axis('off')\n plt.subplots_adjust(top=1)\n plt.show()", "_____no_output_____" ], [ "def plotData(X, Y, c, npixel=28):\n m, n = X.shape\n image = np.array(X[c,:])\n plt.figure(figsize = (6,6))\n plt.imshow((image.reshape(npixel, npixel)), cmap='Greys', interpolation='quadric')\n plt.show()", "_____no_output_____" ], [ "def plotAccuracy(acc_history_train, acc_history_test):\n plt.figure(figsize = (12,8))\n plt.plot(acc_history_train, marker='o', markersize=5, label='Train')\n plt.plot(acc_history_test, marker='o', markersize=5, label='Test')\n plt.legend()\n plt.gca().xaxis.set_major_locator(plt.NullLocator())\n plt.show()", "_____no_output_____" ], [ "def save_model(m,filename):\n model_json = m.to_json()\n with open(\"/gdrive/My Drive/colab_data/\"+filename+\".json\", \"w\") as json_file:\n json_file.write(model_json)\n # serialize weights to HDF5\n m.save_weights(\"/gdrive/My Drive/colab_data/\"+filename+\".h5\")\n print(\"Saved model to disk\")", "_____no_output_____" ], [ "def load_model_weights(filename, model):\n model.load_weights(\"/gdrive/My Drive/colab_data/\"+filename+\".h5\")\n print(\"Loaded weights from disk\")\n return model", "_____no_output_____" ], [ "def load_model(filename):\n json_file = open(\"/gdrive/My Drive/colab_data/\"+filename+'.json', 'r')\n loaded_model_json = json_file.read()\n json_file.close()\n m = model_from_json(loaded_model_json)\n # load weights into new model\n m.load_weights(\"/gdrive/My Drive/colab_data/\"+filename+\".h5\")\n print(\"Loaded model from disk\")\n return m", "_____no_output_____" ] ], [ [ "Fissa il numero di classi pari a 10 (corrispondenti alle cifre 0,1,...,9) e leggi i dati, suddivisi in training e test set, $(X_{train}, t_{train}), (X_{test}, t_{test})$. Le matrici $X$ rappresentano le immagini, mentre i vettori $t$ specificano le relative classi.", "_____no_output_____" ] ], [ [ "nb_classes = 10\n\n(X_train, t_train), (X_test, t_test) = mnist.load_data()", "Downloading data from https://s3.amazonaws.com/img-datasets/mnist.npz\n11493376/11490434 [==============================] - 0s 0us/step\n" ], [ "X_train.shape, X_test.shape", "_____no_output_____" ] ], [ [ "Visualizza la dimensione di $X_{train}$ e $X_{test}$. Le immagini rappresentate risultano indicizzate dalla prima dimensione e definite come matrici $28\\times 28$. I valori nelle matrici sono interi compresi tra $0$ e $255$.\n", "_____no_output_____" ] ], [ [ "X_train[0,:,:]", "_____no_output_____" ] ], [ [ "Esempio: i primi 100 elementi nel training set con la relativa classe", "_____no_output_____" ] ], [ [ "displayData(X_train[0:100], t_train[0:100], class_value=True)", "_____no_output_____" ] ], [ [ "Modifica la rappresentazione degli elementi da matrici $28\\times 28$ a vettori di dimensione $784$, con elementi reali compresi tra $0$ e $1$.", "_____no_output_____" ] ], [ [ "X_train = X_train.reshape(X_train.shape[0], 784)\nX_test = X_test.reshape(X_test.shape[0], 784)\nX_train = X_train.astype('float32')\nX_test = X_test.astype('float32')\nX_train /= 255\nX_test /= 255", "_____no_output_____" ], [ "X_train[0,:].shape", "_____no_output_____" ] ], [ [ "Le classi vengono codificate nella forma $\\textit{one-hot}$, come vettori di dimensione $10$ (il numero delle classi) con elementi pari a $0$, eccetto quello di indice pari al valore da codificare.\n", "_____no_output_____" ] ], [ [ "T_train = np_utils.to_categorical(t_train, nb_classes)\nT_test = np_utils.to_categorical(t_test, nb_classes)", "_____no_output_____" ], [ "t_train[0:5]", "_____no_output_____" ], [ "T_train[0:5,:]", "_____no_output_____" ] ], [ [ "Costruisci il modello (insieme delle possibili funzioni)", "_____no_output_____" ] ], [ [ "# softmax regression\nmodel0 = Sequential()\nmodel0.add(Dense(10, input_shape=(784,)))\nmodel0.add(Activation('softmax'))\nmodel0.compile(loss='categorical_crossentropy', metrics=['accuracy'], optimizer='sgd')", "_____no_output_____" ], [ "model0.summary()", "Model: \"sequential_2\"\n_________________________________________________________________\nLayer (type) Output Shape Param # \n=================================================================\ndense_2 (Dense) (None, 10) 7850 \n_________________________________________________________________\nactivation_2 (Activation) (None, 10) 0 \n=================================================================\nTotal params: 7,850\nTrainable params: 7,850\nNon-trainable params: 0\n_________________________________________________________________\n" ], [ "#model0 = load_model_weights('softmax', model0)", "_____no_output_____" ] ], [ [ "Cerca la migliore funzione possibile, rispetto ai dati disponibili", "_____no_output_____" ] ], [ [ "history0 = model0.fit(X_train, T_train, batch_size=128, epochs=50, verbose=1, validation_data=(X_test, T_test))", "Train on 60000 samples, validate on 10000 samples\nEpoch 1/50\n60000/60000 [==============================] - 1s 16us/step - loss: 0.3090 - accuracy: 0.9139 - val_loss: 0.2978 - val_accuracy: 0.9178\nEpoch 2/50\n60000/60000 [==============================] - 1s 16us/step - loss: 0.3082 - accuracy: 0.9144 - val_loss: 0.2972 - val_accuracy: 0.9173\nEpoch 3/50\n60000/60000 [==============================] - 1s 16us/step - loss: 0.3075 - accuracy: 0.9146 - val_loss: 0.2964 - val_accuracy: 0.9181\nEpoch 4/50\n60000/60000 [==============================] - 1s 16us/step - loss: 0.3068 - accuracy: 0.9147 - val_loss: 0.2960 - val_accuracy: 0.9173\nEpoch 5/50\n60000/60000 [==============================] - 1s 15us/step - loss: 0.3061 - accuracy: 0.9148 - val_loss: 0.2954 - val_accuracy: 0.9170\nEpoch 6/50\n60000/60000 [==============================] - 1s 15us/step - loss: 0.3055 - accuracy: 0.9148 - val_loss: 0.2952 - val_accuracy: 0.9179\nEpoch 7/50\n60000/60000 [==============================] - 1s 15us/step - loss: 0.3048 - accuracy: 0.9153 - val_loss: 0.2943 - val_accuracy: 0.9176\nEpoch 8/50\n60000/60000 [==============================] - 1s 15us/step - loss: 0.3042 - accuracy: 0.9152 - val_loss: 0.2941 - val_accuracy: 0.9176\nEpoch 9/50\n60000/60000 [==============================] - 1s 16us/step - loss: 0.3036 - accuracy: 0.9154 - val_loss: 0.2933 - val_accuracy: 0.9174\nEpoch 10/50\n60000/60000 [==============================] - 1s 15us/step - loss: 0.3030 - accuracy: 0.9157 - val_loss: 0.2929 - val_accuracy: 0.9180\nEpoch 11/50\n60000/60000 [==============================] - 1s 15us/step - loss: 0.3024 - accuracy: 0.9157 - val_loss: 0.2925 - val_accuracy: 0.9181\nEpoch 12/50\n60000/60000 [==============================] - 1s 16us/step - loss: 0.3018 - accuracy: 0.9159 - val_loss: 0.2922 - val_accuracy: 0.9186\nEpoch 13/50\n60000/60000 [==============================] - 1s 16us/step - loss: 0.3013 - accuracy: 0.9159 - val_loss: 0.2919 - val_accuracy: 0.9184\nEpoch 14/50\n60000/60000 [==============================] - 1s 16us/step - loss: 0.3007 - accuracy: 0.9163 - val_loss: 0.2913 - val_accuracy: 0.9181\nEpoch 15/50\n60000/60000 [==============================] - 1s 15us/step - loss: 0.3002 - accuracy: 0.9165 - val_loss: 0.2906 - val_accuracy: 0.9187\nEpoch 16/50\n60000/60000 [==============================] - 1s 15us/step - loss: 0.2997 - accuracy: 0.9165 - val_loss: 0.2905 - val_accuracy: 0.9187\nEpoch 17/50\n60000/60000 [==============================] - 1s 16us/step - loss: 0.2992 - accuracy: 0.9168 - val_loss: 0.2902 - val_accuracy: 0.9185\nEpoch 18/50\n60000/60000 [==============================] - 1s 16us/step - loss: 0.2987 - accuracy: 0.9167 - val_loss: 0.2900 - val_accuracy: 0.9192\nEpoch 19/50\n60000/60000 [==============================] - 1s 16us/step - loss: 0.2982 - accuracy: 0.9168 - val_loss: 0.2897 - val_accuracy: 0.9185\nEpoch 20/50\n60000/60000 [==============================] - 1s 16us/step - loss: 0.2977 - accuracy: 0.9171 - val_loss: 0.2892 - val_accuracy: 0.9191\nEpoch 21/50\n60000/60000 [==============================] - 1s 16us/step - loss: 0.2972 - accuracy: 0.9170 - val_loss: 0.2888 - val_accuracy: 0.9191\nEpoch 22/50\n60000/60000 [==============================] - 1s 16us/step - loss: 0.2968 - accuracy: 0.9176 - val_loss: 0.2885 - val_accuracy: 0.9198\nEpoch 23/50\n60000/60000 [==============================] - 1s 17us/step - loss: 0.2964 - accuracy: 0.9176 - val_loss: 0.2880 - val_accuracy: 0.9191\nEpoch 24/50\n60000/60000 [==============================] - 1s 16us/step - loss: 0.2959 - accuracy: 0.9178 - val_loss: 0.2877 - val_accuracy: 0.9197\nEpoch 25/50\n60000/60000 [==============================] - 1s 17us/step - loss: 0.2955 - accuracy: 0.9177 - val_loss: 0.2878 - val_accuracy: 0.9192\nEpoch 26/50\n60000/60000 [==============================] - 1s 21us/step - loss: 0.2951 - accuracy: 0.9179 - val_loss: 0.2871 - val_accuracy: 0.9198\nEpoch 27/50\n60000/60000 [==============================] - 1s 20us/step - loss: 0.2946 - accuracy: 0.9178 - val_loss: 0.2873 - val_accuracy: 0.9196\nEpoch 28/50\n60000/60000 [==============================] - 1s 20us/step - loss: 0.2942 - accuracy: 0.9178 - val_loss: 0.2869 - val_accuracy: 0.9195\nEpoch 29/50\n60000/60000 [==============================] - 1s 18us/step - loss: 0.2938 - accuracy: 0.9181 - val_loss: 0.2866 - val_accuracy: 0.9198\nEpoch 30/50\n60000/60000 [==============================] - 1s 18us/step - loss: 0.2934 - accuracy: 0.9182 - val_loss: 0.2861 - val_accuracy: 0.9199\nEpoch 31/50\n60000/60000 [==============================] - 1s 18us/step - loss: 0.2930 - accuracy: 0.9181 - val_loss: 0.2859 - val_accuracy: 0.9196\nEpoch 32/50\n60000/60000 [==============================] - 1s 18us/step - loss: 0.2927 - accuracy: 0.9184 - val_loss: 0.2856 - val_accuracy: 0.9199\nEpoch 33/50\n60000/60000 [==============================] - 1s 19us/step - loss: 0.2923 - accuracy: 0.9186 - val_loss: 0.2851 - val_accuracy: 0.9202\nEpoch 34/50\n60000/60000 [==============================] - 1s 18us/step - loss: 0.2920 - accuracy: 0.9183 - val_loss: 0.2852 - val_accuracy: 0.9205\nEpoch 35/50\n60000/60000 [==============================] - 1s 18us/step - loss: 0.2916 - accuracy: 0.9187 - val_loss: 0.2850 - val_accuracy: 0.9199\nEpoch 36/50\n60000/60000 [==============================] - 1s 16us/step - loss: 0.2912 - accuracy: 0.9189 - val_loss: 0.2847 - val_accuracy: 0.9205\nEpoch 37/50\n60000/60000 [==============================] - 1s 16us/step - loss: 0.2909 - accuracy: 0.9190 - val_loss: 0.2844 - val_accuracy: 0.9204\nEpoch 38/50\n60000/60000 [==============================] - 1s 15us/step - loss: 0.2906 - accuracy: 0.9191 - val_loss: 0.2843 - val_accuracy: 0.9200\nEpoch 39/50\n60000/60000 [==============================] - 1s 15us/step - loss: 0.2902 - accuracy: 0.9192 - val_loss: 0.2839 - val_accuracy: 0.9204\nEpoch 40/50\n60000/60000 [==============================] - 1s 15us/step - loss: 0.2899 - accuracy: 0.9191 - val_loss: 0.2836 - val_accuracy: 0.9211\nEpoch 41/50\n60000/60000 [==============================] - 1s 15us/step - loss: 0.2896 - accuracy: 0.9191 - val_loss: 0.2835 - val_accuracy: 0.9207\nEpoch 42/50\n60000/60000 [==============================] - 1s 15us/step - loss: 0.2893 - accuracy: 0.9194 - val_loss: 0.2835 - val_accuracy: 0.9200\nEpoch 43/50\n60000/60000 [==============================] - 1s 15us/step - loss: 0.2889 - accuracy: 0.9194 - val_loss: 0.2829 - val_accuracy: 0.9207\nEpoch 44/50\n60000/60000 [==============================] - 1s 15us/step - loss: 0.2886 - accuracy: 0.9199 - val_loss: 0.2828 - val_accuracy: 0.9201\nEpoch 45/50\n60000/60000 [==============================] - 1s 15us/step - loss: 0.2883 - accuracy: 0.9200 - val_loss: 0.2827 - val_accuracy: 0.9207\nEpoch 46/50\n60000/60000 [==============================] - 1s 15us/step - loss: 0.2880 - accuracy: 0.9198 - val_loss: 0.2825 - val_accuracy: 0.9211\nEpoch 47/50\n60000/60000 [==============================] - 1s 15us/step - loss: 0.2877 - accuracy: 0.9199 - val_loss: 0.2823 - val_accuracy: 0.9208\nEpoch 48/50\n60000/60000 [==============================] - 1s 15us/step - loss: 0.2874 - accuracy: 0.9202 - val_loss: 0.2821 - val_accuracy: 0.9207\nEpoch 49/50\n60000/60000 [==============================] - 1s 15us/step - loss: 0.2872 - accuracy: 0.9202 - val_loss: 0.2820 - val_accuracy: 0.9210\nEpoch 50/50\n60000/60000 [==============================] - 1s 15us/step - loss: 0.2868 - accuracy: 0.9206 - val_loss: 0.2818 - val_accuracy: 0.9210\n" ], [ "save_model(model0, 'softmax')", "Saved model to disk\n" ], [ "history0_df = pd.DataFrame(history0.history)", "_____no_output_____" ], [ "history0_df.head()", "_____no_output_____" ], [ "save_model(model0,'softmax')", "Saved model to disk\n" ] ], [ [ "Accuracy: frazione di risposte corrette", "_____no_output_____" ] ], [ [ "plotAccuracy(history0_df.accuracy, history0_df.val_accuracy)", "_____no_output_____" ], [ "predictions_train = model0.predict_classes(X_train, verbose=0)", "_____no_output_____" ], [ "confusion_matrix(t_train, predictions_train)", "_____no_output_____" ], [ "accuracy_score(t_train, predictions_train)", "_____no_output_____" ], [ "meas = precision_recall_fscore_support(t_train, predictions_train, average=None)\nprint('Training set')\nfor i in range(10):\n print('Class {0:d}: precision={1:5.2f}, recall={2:5.2f}, f-measure={2:5.2f}'.format(i, meas[1][i], meas[2][i], meas[3][i]))", "Training set\nClass 0: precision= 0.97, recall= 0.96, f-measure= 0.96\nClass 1: precision= 0.97, recall= 0.96, f-measure= 0.96\nClass 2: precision= 0.89, recall= 0.91, f-measure= 0.91\nClass 3: precision= 0.89, recall= 0.90, f-measure= 0.90\nClass 4: precision= 0.93, recall= 0.93, f-measure= 0.93\nClass 5: precision= 0.87, recall= 0.88, f-measure= 0.88\nClass 6: precision= 0.96, recall= 0.95, f-measure= 0.95\nClass 7: precision= 0.93, recall= 0.93, f-measure= 0.93\nClass 8: precision= 0.88, recall= 0.89, f-measure= 0.89\nClass 9: precision= 0.90, recall= 0.90, f-measure= 0.90\n" ], [ "predictions_test = model0.predict_classes(X_test, verbose=0)", "_____no_output_____" ], [ "confusion_matrix(t_test, predictions_test)", "_____no_output_____" ], [ "accuracy_score(t_test, predictions_test)", "_____no_output_____" ], [ "meas = precision_recall_fscore_support(t_test, predictions_test, average=None)\nprint('Test set')\nfor i in range(10):\n print('Class {0:d}: precision={1:5.2f}, recall={2:5.2f}, f-measure={2:5.2f}'.format(i, meas[1][i], meas[2][i], meas[3][i]))", "Test set\nClass 0: precision= 0.98, recall= 0.96, f-measure= 0.96\nClass 1: precision= 0.98, recall= 0.97, f-measure= 0.97\nClass 2: precision= 0.89, recall= 0.91, f-measure= 0.91\nClass 3: precision= 0.91, recall= 0.90, f-measure= 0.90\nClass 4: precision= 0.93, recall= 0.92, f-measure= 0.92\nClass 5: precision= 0.86, recall= 0.88, f-measure= 0.88\nClass 6: precision= 0.95, recall= 0.94, f-measure= 0.94\nClass 7: precision= 0.92, recall= 0.92, f-measure= 0.92\nClass 8: precision= 0.89, recall= 0.89, f-measure= 0.89\nClass 9: precision= 0.89, recall= 0.90, f-measure= 0.90\n" ], [ "w0 = model0.layers[0].get_weights()\nw = w0[0]", "_____no_output_____" ], [ "w0[1]", "_____no_output_____" ], [ "fig = plt.figure(figsize=(16,16))\nfig.patch.set_facecolor('white')\nfor i in range(10):\n ax = subplot(2,5,i+1,frame_on=False, facecolor=\"#F8F8F8\")\n ax.xaxis.set_major_locator(NullLocator())\n ax.yaxis.set_major_locator(NullLocator())\n imshow(w[:,i].reshape(28,28), interpolation='quadric', \n cmap=plt.get_cmap('Greys'), vmin=np.mean(w[:,i])-3.5*np.std(w[:,i]), \n vmax=np.mean(w[:,i])+3.5*np.std(w[:,i]), aspect='auto')\n plt.title(i)\nplt.subplots_adjust(top=0.4)\nplt.show()", "_____no_output_____" ], [ "c = np.random.randint(0, X_test.shape[0])\np=model0.predict_classes(X_test[c:c+1,:], verbose=0)\nprint(\"Elemento \"+str(c))\nplotData(X_test, t_test, c)\nprint(\"Un \"+str(t_test[c])+\", classificato come \" + str(p[0]))", "Elemento 7917\n" ], [ "# 3 layer NN\nmodel1 = Sequential()\nmodel1.add(Dense(512, input_shape=(784,)))\nmodel1.add(Activation('relu'))\nmodel1.add(Dropout(rate=0.2))\nmodel1.add(Dense(10))\nmodel1.add(Activation('softmax'))\nmodel1.compile(loss='categorical_crossentropy', metrics=['accuracy'], optimizer='adam')", "_____no_output_____" ], [ "model1.summary()", "Model: \"sequential_3\"\n_________________________________________________________________\nLayer (type) Output Shape Param # \n=================================================================\ndense_3 (Dense) (None, 512) 401920 \n_________________________________________________________________\nactivation_3 (Activation) (None, 512) 0 \n_________________________________________________________________\ndropout_1 (Dropout) (None, 512) 0 \n_________________________________________________________________\ndense_4 (Dense) (None, 10) 5130 \n_________________________________________________________________\nactivation_4 (Activation) (None, 10) 0 \n=================================================================\nTotal params: 407,050\nTrainable params: 407,050\nNon-trainable params: 0\n_________________________________________________________________\n" ], [ "#model1 = load_model_weights('nn3', model1)", "_____no_output_____" ], [ "history1 = model1.fit(X_train, T_train, batch_size=1024, epochs=10, verbose=1, validation_data=(X_test, T_test))", "Train on 60000 samples, validate on 10000 samples\nEpoch 1/10\n60000/60000 [==============================] - 3s 50us/step - loss: 0.5838 - accuracy: 0.8402 - val_loss: 0.2569 - val_accuracy: 0.9255\nEpoch 2/10\n60000/60000 [==============================] - 3s 47us/step - loss: 0.2396 - accuracy: 0.9313 - val_loss: 0.1899 - val_accuracy: 0.9447\nEpoch 3/10\n60000/60000 [==============================] - 3s 47us/step - loss: 0.1817 - accuracy: 0.9485 - val_loss: 0.1513 - val_accuracy: 0.9558\nEpoch 4/10\n60000/60000 [==============================] - 3s 47us/step - loss: 0.1454 - accuracy: 0.9587 - val_loss: 0.1284 - val_accuracy: 0.9628\nEpoch 5/10\n60000/60000 [==============================] - 3s 47us/step - loss: 0.1229 - accuracy: 0.9653 - val_loss: 0.1127 - val_accuracy: 0.9665\nEpoch 6/10\n60000/60000 [==============================] - 3s 46us/step - loss: 0.1059 - accuracy: 0.9703 - val_loss: 0.1007 - val_accuracy: 0.9710\nEpoch 7/10\n60000/60000 [==============================] - 3s 47us/step - loss: 0.0924 - accuracy: 0.9738 - val_loss: 0.0917 - val_accuracy: 0.9733\nEpoch 8/10\n60000/60000 [==============================] - 3s 47us/step - loss: 0.0808 - accuracy: 0.9773 - val_loss: 0.0862 - val_accuracy: 0.9738\nEpoch 9/10\n60000/60000 [==============================] - 3s 47us/step - loss: 0.0713 - accuracy: 0.9794 - val_loss: 0.0805 - val_accuracy: 0.9763\nEpoch 10/10\n60000/60000 [==============================] - 3s 47us/step - loss: 0.0639 - accuracy: 0.9823 - val_loss: 0.0751 - val_accuracy: 0.9776\n" ], [ "save_model(model1,'nn3')", "Saved model to disk\n" ], [ "history1_df = pd.DataFrame(history1.history)", "_____no_output_____" ], [ "history1_df.head()", "_____no_output_____" ], [ "plotAccuracy(history1_df.accuracy, history1_df.val_accuracy)", "_____no_output_____" ], [ "predictions_train = model1.predict_classes(X_train, verbose=0)", "_____no_output_____" ], [ "confusion_matrix(t_train, predictions_train)", "_____no_output_____" ], [ "accuracy_score(t_train, predictions_train)", "_____no_output_____" ], [ "meas = precision_recall_fscore_support(t_train, predictions_train, average=None)\nprint('Training set')\nfor i in range(10):\n print('Class {0:d}: precision={1:5.2f}, recall={2:5.2f}, f-measure={2:5.2f}'.format(i, meas[1][i], meas[2][i], meas[3][i]))", "Training set\nClass 0: precision= 0.99, recall= 0.99, f-measure= 0.99\nClass 1: precision= 0.99, recall= 0.99, f-measure= 0.99\nClass 2: precision= 0.99, recall= 0.99, f-measure= 0.99\nClass 3: precision= 0.98, recall= 0.99, f-measure= 0.99\nClass 4: precision= 0.99, recall= 0.99, f-measure= 0.99\nClass 5: precision= 0.99, recall= 0.99, f-measure= 0.99\nClass 6: precision= 0.99, recall= 0.99, f-measure= 0.99\nClass 7: precision= 0.99, recall= 0.99, f-measure= 0.99\nClass 8: precision= 0.98, recall= 0.98, f-measure= 0.98\nClass 9: precision= 0.98, recall= 0.98, f-measure= 0.98\n" ], [ "predictions_test = model1.predict_classes(X_test, verbose=0)", "_____no_output_____" ], [ "confusion_matrix(t_test, predictions_test)", "_____no_output_____" ], [ "accuracy_score(t_test, predictions_test)", "_____no_output_____" ], [ "meas = precision_recall_fscore_support(t_test, predictions_test, average=None)\nprint('Test set')\nfor i in range(10):\n print('Class {0:d}: precision={1:5.2f}, recall={2:5.2f}, f-measure={2:5.2f}'.format(i, meas[1][i], meas[2][i], meas[3][i]))", "Test set\nClass 0: precision= 0.99, recall= 0.99, f-measure= 0.99\nClass 1: precision= 0.99, recall= 0.99, f-measure= 0.99\nClass 2: precision= 0.98, recall= 0.97, f-measure= 0.97\nClass 3: precision= 0.98, recall= 0.97, f-measure= 0.97\nClass 4: precision= 0.98, recall= 0.98, f-measure= 0.98\nClass 5: precision= 0.98, recall= 0.98, f-measure= 0.98\nClass 6: precision= 0.97, recall= 0.98, f-measure= 0.98\nClass 7: precision= 0.97, recall= 0.98, f-measure= 0.98\nClass 8: precision= 0.96, recall= 0.97, f-measure= 0.97\nClass 9: precision= 0.96, recall= 0.97, f-measure= 0.97\n" ], [ "# 4 layer NN\nmodel2 = Sequential()\nmodel2.add(Dense(512, input_shape=(784,)))\nmodel2.add(Activation('relu'))\nmodel2.add(Dropout(0.2))\nmodel2.add(Dense(512))\nmodel2.add(Activation('relu'))\nmodel2.add(Dropout(0.2))\nmodel2.add(Dense(10))\nmodel2.add(Activation('softmax'))", "_____no_output_____" ], [ "model2.summary()", "Model: \"sequential_4\"\n_________________________________________________________________\nLayer (type) Output Shape Param # \n=================================================================\ndense_5 (Dense) (None, 512) 401920 \n_________________________________________________________________\nactivation_5 (Activation) (None, 512) 0 \n_________________________________________________________________\ndropout_2 (Dropout) (None, 512) 0 \n_________________________________________________________________\ndense_6 (Dense) (None, 512) 262656 \n_________________________________________________________________\nactivation_6 (Activation) (None, 512) 0 \n_________________________________________________________________\ndropout_3 (Dropout) (None, 512) 0 \n_________________________________________________________________\ndense_7 (Dense) (None, 10) 5130 \n_________________________________________________________________\nactivation_7 (Activation) (None, 10) 0 \n=================================================================\nTotal params: 669,706\nTrainable params: 669,706\nNon-trainable params: 0\n_________________________________________________________________\n" ], [ "model2.compile(loss='categorical_crossentropy', metrics=['accuracy'], optimizer='adam')", "_____no_output_____" ], [ "#model2 = load_model_weights('nn4', model2)", "_____no_output_____" ], [ "history2 = model2.fit(X_train, T_train, batch_size=1024, epochs=10, verbose=1, validation_data=(X_test, T_test))", "Train on 60000 samples, validate on 10000 samples\nEpoch 1/10\n60000/60000 [==============================] - 5s 88us/step - loss: 0.5063 - accuracy: 0.8563 - val_loss: 0.1870 - val_accuracy: 0.9435\nEpoch 2/10\n60000/60000 [==============================] - 5s 84us/step - loss: 0.1752 - accuracy: 0.9492 - val_loss: 0.1212 - val_accuracy: 0.9627\nEpoch 3/10\n60000/60000 [==============================] - 5s 84us/step - loss: 0.1217 - accuracy: 0.9638 - val_loss: 0.1025 - val_accuracy: 0.9677\nEpoch 4/10\n60000/60000 [==============================] - 5s 85us/step - loss: 0.0933 - accuracy: 0.9722 - val_loss: 0.0808 - val_accuracy: 0.9747\nEpoch 5/10\n60000/60000 [==============================] - 5s 85us/step - loss: 0.0746 - accuracy: 0.9779 - val_loss: 0.0720 - val_accuracy: 0.9770\nEpoch 6/10\n60000/60000 [==============================] - 5s 85us/step - loss: 0.0579 - accuracy: 0.9824 - val_loss: 0.0654 - val_accuracy: 0.9791\nEpoch 7/10\n60000/60000 [==============================] - 5s 86us/step - loss: 0.0502 - accuracy: 0.9840 - val_loss: 0.0623 - val_accuracy: 0.9806\nEpoch 8/10\n60000/60000 [==============================] - 5s 85us/step - loss: 0.0420 - accuracy: 0.9874 - val_loss: 0.0605 - val_accuracy: 0.9803\nEpoch 9/10\n60000/60000 [==============================] - 5s 85us/step - loss: 0.0355 - accuracy: 0.9896 - val_loss: 0.0596 - val_accuracy: 0.9807\nEpoch 10/10\n60000/60000 [==============================] - 5s 85us/step - loss: 0.0305 - accuracy: 0.9907 - val_loss: 0.0577 - val_accuracy: 0.9817\n" ], [ "save_model(model2,'nn4')", "Saved model to disk\n" ], [ "history2_df = pd.DataFrame(history2.history)", "_____no_output_____" ], [ "history2_df.head()", "_____no_output_____" ], [ "plotAccuracy(history2_df.accuracy, history2_df.val_accuracy)", "_____no_output_____" ], [ "predictions_train = model2.predict_classes(X_train, verbose=0)", "_____no_output_____" ], [ "confusion_matrix(t_train, predictions_train)", "_____no_output_____" ], [ "accuracy_score(t_train, predictions_train)", "_____no_output_____" ], [ "meas = precision_recall_fscore_support(t_train, predictions_train, average=None)\nprint('Training set')\nfor i in range(10):\n print('Class {0:d}: precision={1:5.2f}, recall={2:5.2f}, f-measure={2:5.2f}'.format(i, meas[1][i], meas[2][i], meas[3][i]))", "Training set\nClass 0: precision= 1.00, recall= 1.00, f-measure= 1.00\nClass 1: precision= 1.00, recall= 1.00, f-measure= 1.00\nClass 2: precision= 1.00, recall= 1.00, f-measure= 1.00\nClass 3: precision= 1.00, recall= 1.00, f-measure= 1.00\nClass 4: precision= 1.00, recall= 1.00, f-measure= 1.00\nClass 5: precision= 1.00, recall= 1.00, f-measure= 1.00\nClass 6: precision= 1.00, recall= 1.00, f-measure= 1.00\nClass 7: precision= 1.00, recall= 1.00, f-measure= 1.00\nClass 8: precision= 1.00, recall= 1.00, f-measure= 1.00\nClass 9: precision= 0.99, recall= 1.00, f-measure= 1.00\n" ], [ "predictions_test = model2.predict_classes(X_test, verbose=0)", "_____no_output_____" ], [ "confusion_matrix(t_test, predictions_test)", "_____no_output_____" ], [ "accuracy_score(t_test, predictions_test)", "_____no_output_____" ], [ "meas = precision_recall_fscore_support(t_test, predictions_test, average=None)\nprint('Test set')\nfor i in range(10):\n print('Class {0:d}: precision={1:5.2f}, recall={2:5.2f}, f-measure={2:5.2f}'.format(i, meas[1][i], meas[2][i], meas[3][i]))", "Test set\nClass 0: precision= 0.99, recall= 0.99, f-measure= 0.99\nClass 1: precision= 0.99, recall= 0.99, f-measure= 0.99\nClass 2: precision= 0.98, recall= 0.98, f-measure= 0.98\nClass 3: precision= 0.99, recall= 0.98, f-measure= 0.98\nClass 4: precision= 0.98, recall= 0.98, f-measure= 0.98\nClass 5: precision= 0.98, recall= 0.98, f-measure= 0.98\nClass 6: precision= 0.97, recall= 0.98, f-measure= 0.98\nClass 7: precision= 0.98, recall= 0.98, f-measure= 0.98\nClass 8: precision= 0.98, recall= 0.98, f-measure= 0.98\nClass 9: precision= 0.97, recall= 0.97, f-measure= 0.97\n" ], [ "model3 = Sequential()\nmodel3.add(Conv2D(30, (5, 5), input_shape=(28, 28,1), activation='relu'))\nmodel3.add(MaxPooling2D(pool_size=(2, 2)))\nmodel3.add(Conv2D(15, (3, 3), activation='relu'))\nmodel3.add(MaxPooling2D(pool_size=(2, 2)))\nmodel3.add(Dropout(0.2))\nmodel3.add(Flatten())\nmodel3.add(Dense(128, activation='relu'))\nmodel3.add(Dense(50, activation='relu'))\nmodel3.add(Dense(10, activation='softmax'))", "_____no_output_____" ], [ "model3.summary()", "Model: \"sequential_7\"\n_________________________________________________________________\nLayer (type) Output Shape Param # \n=================================================================\nconv2d_4 (Conv2D) (None, 24, 24, 30) 780 \n_________________________________________________________________\nmax_pooling2d_3 (MaxPooling2 (None, 12, 12, 30) 0 \n_________________________________________________________________\nconv2d_5 (Conv2D) (None, 10, 10, 15) 4065 \n_________________________________________________________________\nmax_pooling2d_4 (MaxPooling2 (None, 5, 5, 15) 0 \n_________________________________________________________________\ndropout_5 (Dropout) (None, 5, 5, 15) 0 \n_________________________________________________________________\nflatten_2 (Flatten) (None, 375) 0 \n_________________________________________________________________\ndense_11 (Dense) (None, 128) 48128 \n_________________________________________________________________\ndense_12 (Dense) (None, 50) 6450 \n_________________________________________________________________\ndense_13 (Dense) (None, 10) 510 \n=================================================================\nTotal params: 59,933\nTrainable params: 59,933\nNon-trainable params: 0\n_________________________________________________________________\n" ], [ "model3.compile(loss='categorical_crossentropy', optimizer='adam', metrics=['accuracy'])", "_____no_output_____" ], [ "X_train_c = X_train.reshape(X_train.shape[0], 28, 28,1).astype('float32')\nX_test_c = X_test.reshape(X_test.shape[0], 28, 28, 1).astype('float32')", "_____no_output_____" ], [ "#model3 = load_model_weights('cnn', model3)", "_____no_output_____" ], [ "history3 = model3.fit(X_train_c, T_train, batch_size=1024, epochs=10, verbose=1, validation_data=(X_test_c, T_test))", "Train on 60000 samples, validate on 10000 samples\nEpoch 1/10\n60000/60000 [==============================] - 33s 556us/step - loss: 1.0933 - accuracy: 0.6511 - val_loss: 0.2442 - val_accuracy: 0.9319\nEpoch 2/10\n60000/60000 [==============================] - 33s 547us/step - loss: 0.2142 - accuracy: 0.9366 - val_loss: 0.1038 - val_accuracy: 0.9687\nEpoch 3/10\n60000/60000 [==============================] - 33s 553us/step - loss: 0.1253 - accuracy: 0.9624 - val_loss: 0.0693 - val_accuracy: 0.9782\nEpoch 4/10\n60000/60000 [==============================] - 33s 545us/step - loss: 0.0976 - accuracy: 0.9699 - val_loss: 0.0532 - val_accuracy: 0.9830\nEpoch 5/10\n60000/60000 [==============================] - 33s 545us/step - loss: 0.0831 - accuracy: 0.9745 - val_loss: 0.0469 - val_accuracy: 0.9857\nEpoch 6/10\n60000/60000 [==============================] - 33s 545us/step - loss: 0.0705 - accuracy: 0.9783 - val_loss: 0.0437 - val_accuracy: 0.9852\nEpoch 7/10\n60000/60000 [==============================] - 33s 543us/step - loss: 0.0617 - accuracy: 0.9808 - val_loss: 0.0384 - val_accuracy: 0.9872\nEpoch 8/10\n60000/60000 [==============================] - 33s 544us/step - loss: 0.0567 - accuracy: 0.9825 - val_loss: 0.0328 - val_accuracy: 0.9887\nEpoch 9/10\n60000/60000 [==============================] - 33s 545us/step - loss: 0.0528 - accuracy: 0.9829 - val_loss: 0.0321 - val_accuracy: 0.9894\nEpoch 10/10\n60000/60000 [==============================] - 33s 545us/step - loss: 0.0464 - accuracy: 0.9852 - val_loss: 0.0299 - val_accuracy: 0.9902\n" ], [ "save_model(model3,'cnn')", "Saved model to disk\n" ], [ "history3_df = pd.DataFrame(history3.history) ", "_____no_output_____" ], [ "history3_df.head()", "_____no_output_____" ], [ "plotAccuracy(history3_df.accuracy, history3_df.val_accuracy)", "_____no_output_____" ], [ "predictions_train = model3.predict_classes(X_train_c, verbose=0)", "_____no_output_____" ], [ "confusion_matrix(t_train, predictions_train)", "_____no_output_____" ], [ "accuracy_score(t_train, predictions_train)", "_____no_output_____" ], [ "meas = precision_recall_fscore_support(t_train, predictions_train, average=None)\nprint('Training set')\nfor i in range(10):\n print('Class {0:d}: precision={1:5.2f}, recall={2:5.2f}, f-measure={2:5.2f}'.format(i, meas[1][i], meas[2][i], meas[3][i]))", "Training set\nClass 0: precision= 1.00, recall= 1.00, f-measure= 1.00\nClass 1: precision= 0.99, recall= 0.99, f-measure= 0.99\nClass 2: precision= 0.99, recall= 0.99, f-measure= 0.99\nClass 3: precision= 0.99, recall= 0.99, f-measure= 0.99\nClass 4: precision= 0.99, recall= 0.99, f-measure= 0.99\nClass 5: precision= 0.99, recall= 0.99, f-measure= 0.99\nClass 6: precision= 0.99, recall= 0.99, f-measure= 0.99\nClass 7: precision= 0.99, recall= 0.99, f-measure= 0.99\nClass 8: precision= 0.99, recall= 0.99, f-measure= 0.99\nClass 9: precision= 0.99, recall= 0.99, f-measure= 0.99\n" ], [ "predictions_test = model3.predict_classes(X_test_c, verbose=0)", "_____no_output_____" ], [ "confusion_matrix(t_test, predictions_test)", "_____no_output_____" ], [ "accuracy_score(t_test, predictions_test)", "_____no_output_____" ], [ "meas = precision_recall_fscore_support(t_test, predictions_test, average=None)\nprint('Test set')\nfor i in range(10):\n print('Class {0:d}: precision={1:5.2f}, recall={2:5.2f}, f-measure={2:5.2f}'.format(i, meas[1][i], meas[2][i], meas[3][i]))", "Test set\nClass 0: precision= 1.00, recall= 0.99, f-measure= 0.99\nClass 1: precision= 0.99, recall= 0.99, f-measure= 0.99\nClass 2: precision= 0.99, recall= 0.99, f-measure= 0.99\nClass 3: precision= 1.00, recall= 0.99, f-measure= 0.99\nClass 4: precision= 0.99, recall= 0.99, f-measure= 0.99\nClass 5: precision= 0.99, recall= 0.99, f-measure= 0.99\nClass 6: precision= 0.98, recall= 0.99, f-measure= 0.99\nClass 7: precision= 0.98, recall= 0.99, f-measure= 0.99\nClass 8: precision= 0.99, recall= 0.99, f-measure= 0.99\nClass 9: precision= 0.98, recall= 0.98, f-measure= 0.98\n" ] ] ]
[ "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code" ]
[ [ "code", "code", "code", "code", "code", "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code" ] ]
ecebaf4710e4a9483f67585dd2aaec9a6310cd19
1,160
ipynb
Jupyter Notebook
index.ipynb
RobInLabUJI/ROS-Notebooks
b92c1f0cb88fd52b24d0f5d24d9278894b27d1f3
[ "MIT" ]
2
2021-05-23T18:03:41.000Z
2021-12-23T21:44:59.000Z
index.ipynb
RobInLabUJI/ROS-Notebooks
b92c1f0cb88fd52b24d0f5d24d9278894b27d1f3
[ "MIT" ]
null
null
null
index.ipynb
RobInLabUJI/ROS-Notebooks
b92c1f0cb88fd52b24d0f5d24d9278894b27d1f3
[ "MIT" ]
null
null
null
17.058824
59
0.506034
[ [ [ "# ROS Notebooks", "_____no_output_____" ], [ "## [Core ROS Tutorials](Tutorials/index.ipynb)\n", "_____no_output_____" ], [ "## [Stage](Tutorials/Stage/index.ipynb)\n", "_____no_output_____" ], [ "## [Gazebo](Gazebo/index.ipynb)\n", "_____no_output_____" ], [ "## [Try a TurtleBot](try-a-turtlebot/default.ipynb)\n", "_____no_output_____" ] ] ]
[ "markdown" ]
[ [ "markdown", "markdown", "markdown", "markdown", "markdown" ] ]
ecebcc77c161d9f8452ca716c148a9b5755c4b3f
74,387
ipynb
Jupyter Notebook
notebooks/autoencoder/01_basic_autoencoder.ipynb
tingsyo/course_2020Q3_representation_learning
642b8ffdfaf6bfb2b2456af31d9dd8cf66716b64
[ "Apache-2.0" ]
null
null
null
notebooks/autoencoder/01_basic_autoencoder.ipynb
tingsyo/course_2020Q3_representation_learning
642b8ffdfaf6bfb2b2456af31d9dd8cf66716b64
[ "Apache-2.0" ]
null
null
null
notebooks/autoencoder/01_basic_autoencoder.ipynb
tingsyo/course_2020Q3_representation_learning
642b8ffdfaf6bfb2b2456af31d9dd8cf66716b64
[ "Apache-2.0" ]
null
null
null
121.151466
22,968
0.767083
[ [ [ "import numpy as np\nimport pandas as pd", "_____no_output_____" ] ], [ [ "## Model", "_____no_output_____" ] ], [ [ "# Let's build the simplest possible autoencoder\nfrom keras.layers import Input, Dense\nfrom keras.models import Model\n\n# this is the size of our encoded representations\nencoding_dim = 32 # 32 floats -> compression of factor 24.5, assuming the input is 784 floats\n\n# this is our input placeholder\ninput_img = Input(shape=(784,))\n# \"encoded\" is the encoded representation of the input\nencoded = Dense(encoding_dim, activation='relu')(input_img)\n# \"decoded\" is the lossy reconstruction of the input\ndecoded = Dense(784, activation='sigmoid')(encoded)\n# this model maps an input to its reconstruction\nautoencoder = Model(input_img, decoded)\n# Configure our model to use a per-pixel binary crossentropy loss, and the Adadelta optimizer:\nautoencoder.compile(optimizer='adadelta', loss='binary_crossentropy')\n\n\n# Create a separate encoder model:\n# this model maps an input to its encoded representation\nencoder = Model(input_img, encoded)\n\n\n# Create a separate decoder model:\n# create a placeholder for an encoded (32-dimensional) input\nencoded_input = Input(shape=(encoding_dim,))\n# retrieve the last layer of the autoencoder model\ndecoder_layer = autoencoder.layers[-1]\n# create the decoder model\ndecoder = Model(encoded_input, decoder_layer(encoded_input))", "_____no_output_____" ] ], [ [ "## Data", "_____no_output_____" ] ], [ [ "# Data\n# Let's prepare our input data. We're using MNIST digits, and we're discarding the labels \n# (since we're only interested in encoding/decoding the input images).\nfrom keras.datasets import mnist\nimport numpy as np\n(x_train, _), (x_test, _) = mnist.load_data()\n\n# We will normalize all values between 0 and 1 and we will flatten the 28x28 images into vectors of size 784.\nx_train = x_train.astype('float32') / 255.\nx_test = x_test.astype('float32') / 255.\nx_train = x_train.reshape((len(x_train), np.prod(x_train.shape[1:])))\nx_test = x_test.reshape((len(x_test), np.prod(x_test.shape[1:])))\nprint(x_train.shape)\nprint(x_test.shape)", "(60000, 784)\n(10000, 784)\n" ] ], [ [ "## Training\n\nNow let's train our autoencoder for 50 epochs:", "_____no_output_____" ] ], [ [ "autoencoder.fit(x_train, x_train,\n epochs=50,\n batch_size=256,\n shuffle=True,\n validation_data=(x_test, x_test))", "Train on 60000 samples, validate on 10000 samples\nEpoch 1/50\n60000/60000 [==============================] - 1s 12us/step - loss: 0.1042 - val_loss: 0.1023\nEpoch 2/50\n60000/60000 [==============================] - 1s 12us/step - loss: 0.1039 - val_loss: 0.1020\nEpoch 3/50\n60000/60000 [==============================] - 1s 11us/step - loss: 0.1036 - val_loss: 0.1017\nEpoch 4/50\n60000/60000 [==============================] - 1s 12us/step - loss: 0.1033 - val_loss: 0.1014\nEpoch 5/50\n60000/60000 [==============================] - 1s 11us/step - loss: 0.1030 - val_loss: 0.1012\nEpoch 6/50\n60000/60000 [==============================] - 1s 12us/step - loss: 0.1027 - val_loss: 0.1009\nEpoch 7/50\n60000/60000 [==============================] - 1s 11us/step - loss: 0.1025 - val_loss: 0.1007\nEpoch 8/50\n60000/60000 [==============================] - 1s 12us/step - loss: 0.1022 - val_loss: 0.1004\nEpoch 9/50\n60000/60000 [==============================] - 1s 12us/step - loss: 0.1020 - val_loss: 0.1002\nEpoch 10/50\n60000/60000 [==============================] - 1s 12us/step - loss: 0.1018 - val_loss: 0.1000\nEpoch 11/50\n60000/60000 [==============================] - 1s 12us/step - loss: 0.1015 - val_loss: 0.0998\nEpoch 12/50\n60000/60000 [==============================] - 1s 12us/step - loss: 0.1013 - val_loss: 0.0996\nEpoch 13/50\n60000/60000 [==============================] - 1s 11us/step - loss: 0.1011 - val_loss: 0.0994\nEpoch 14/50\n60000/60000 [==============================] - 1s 12us/step - loss: 0.1009 - val_loss: 0.0992\nEpoch 15/50\n60000/60000 [==============================] - 1s 12us/step - loss: 0.1008 - val_loss: 0.0990\nEpoch 16/50\n60000/60000 [==============================] - 1s 12us/step - loss: 0.1006 - val_loss: 0.0989\nEpoch 17/50\n60000/60000 [==============================] - 1s 12us/step - loss: 0.1004 - val_loss: 0.0987\nEpoch 18/50\n60000/60000 [==============================] - 1s 12us/step - loss: 0.1003 - val_loss: 0.0986\nEpoch 19/50\n60000/60000 [==============================] - 1s 12us/step - loss: 0.1001 - val_loss: 0.0985\nEpoch 20/50\n60000/60000 [==============================] - 1s 12us/step - loss: 0.1000 - val_loss: 0.0983\nEpoch 21/50\n60000/60000 [==============================] - 1s 11us/step - loss: 0.0999 - val_loss: 0.0982\nEpoch 22/50\n60000/60000 [==============================] - 1s 12us/step - loss: 0.0997 - val_loss: 0.0981\nEpoch 23/50\n60000/60000 [==============================] - 1s 11us/step - loss: 0.0996 - val_loss: 0.0980\nEpoch 24/50\n60000/60000 [==============================] - 1s 12us/step - loss: 0.0995 - val_loss: 0.0979\nEpoch 25/50\n60000/60000 [==============================] - 1s 12us/step - loss: 0.0994 - val_loss: 0.0978\nEpoch 26/50\n60000/60000 [==============================] - 1s 12us/step - loss: 0.0993 - val_loss: 0.0977\nEpoch 27/50\n60000/60000 [==============================] - 1s 12us/step - loss: 0.0992 - val_loss: 0.0976\nEpoch 28/50\n60000/60000 [==============================] - 1s 12us/step - loss: 0.0991 - val_loss: 0.0975\nEpoch 29/50\n60000/60000 [==============================] - 1s 12us/step - loss: 0.0991 - val_loss: 0.0975\nEpoch 30/50\n60000/60000 [==============================] - 1s 12us/step - loss: 0.0990 - val_loss: 0.0974\nEpoch 31/50\n60000/60000 [==============================] - 1s 12us/step - loss: 0.0989 - val_loss: 0.0973\nEpoch 32/50\n60000/60000 [==============================] - 1s 12us/step - loss: 0.0988 - val_loss: 0.0972\nEpoch 33/50\n60000/60000 [==============================] - 1s 12us/step - loss: 0.0988 - val_loss: 0.0972\nEpoch 34/50\n60000/60000 [==============================] - 1s 12us/step - loss: 0.0987 - val_loss: 0.0971\nEpoch 35/50\n60000/60000 [==============================] - 1s 12us/step - loss: 0.0986 - val_loss: 0.0971\nEpoch 36/50\n60000/60000 [==============================] - 1s 12us/step - loss: 0.0986 - val_loss: 0.0970\nEpoch 37/50\n60000/60000 [==============================] - 1s 12us/step - loss: 0.0985 - val_loss: 0.0969\nEpoch 38/50\n60000/60000 [==============================] - 1s 12us/step - loss: 0.0985 - val_loss: 0.0969\nEpoch 39/50\n60000/60000 [==============================] - 1s 12us/step - loss: 0.0984 - val_loss: 0.0968\nEpoch 40/50\n60000/60000 [==============================] - 1s 12us/step - loss: 0.0983 - val_loss: 0.0968\nEpoch 41/50\n60000/60000 [==============================] - 1s 12us/step - loss: 0.0983 - val_loss: 0.0967\nEpoch 42/50\n60000/60000 [==============================] - 1s 12us/step - loss: 0.0982 - val_loss: 0.0967\nEpoch 43/50\n60000/60000 [==============================] - 1s 12us/step - loss: 0.0982 - val_loss: 0.0966\nEpoch 44/50\n60000/60000 [==============================] - 1s 12us/step - loss: 0.0981 - val_loss: 0.0966\nEpoch 45/50\n60000/60000 [==============================] - 1s 12us/step - loss: 0.0981 - val_loss: 0.0966\nEpoch 46/50\n60000/60000 [==============================] - 1s 12us/step - loss: 0.0981 - val_loss: 0.0965\nEpoch 47/50\n60000/60000 [==============================] - 1s 13us/step - loss: 0.0980 - val_loss: 0.0965\nEpoch 48/50\n60000/60000 [==============================] - 1s 12us/step - loss: 0.0980 - val_loss: 0.0964\nEpoch 49/50\n60000/60000 [==============================] - 1s 12us/step - loss: 0.0979 - val_loss: 0.0964\nEpoch 50/50\n60000/60000 [==============================] - 1s 12us/step - loss: 0.0979 - val_loss: 0.0964\n" ] ], [ [ "## Visualize what the autoencoder do", "_____no_output_____" ] ], [ [ "# encode and decode some digits\n# note that we take them from the *test* set\nencoded_imgs = encoder.predict(x_test)\ndecoded_imgs = decoder.predict(encoded_imgs)", "_____no_output_____" ], [ "# use Matplotlib (don't ask)\n%matplotlib inline\nimport matplotlib.pyplot as plt\n\nn = 10 # how many digits we will display\nplt.figure(figsize=(20, 4))\nfor i in range(n):\n # display original\n ax = plt.subplot(2, n, i + 1)\n plt.imshow(x_test[i].reshape(28, 28))\n plt.gray()\n ax.get_xaxis().set_visible(False)\n ax.get_yaxis().set_visible(False)\n\n # display reconstruction\n ax = plt.subplot(2, n, i + 1 + n)\n plt.imshow(decoded_imgs[i].reshape(28, 28))\n plt.gray()\n ax.get_xaxis().set_visible(False)\n ax.get_yaxis().set_visible(False)\nplt.show()", "_____no_output_____" ] ], [ [ "## Adding a sparsity constraint on the encoded representations\n\nIn the previous example, the representations were only constrained by the size of the hidden layer (32). In such a situation, what typically happens is that the hidden layer is learning an approximation of PCA (principal component analysis). But another way to constrain the representations to be compact is to add a sparsity contraint on the activity of the hidden representations, so fewer units would \"fire\" at a given time. In Keras, this can be done by adding an activity_regularizer to our Dense layer:", "_____no_output_____" ] ], [ [ "# New version of model\nfrom keras import regularizers\n\nencoding_dim = 32\n\ninput_img = Input(shape=(784,))\n# add a Dense layer with a L1 activity regularizer\nencoded = Dense(encoding_dim, activation='relu',\n activity_regularizer=regularizers.l1(10e-5))(input_img)\ndecoded = Dense(784, activation='sigmoid')(encoded)\nautoencoder = Model(input_img, decoded)\nautoencoder.compile(optimizer='adadelta', loss='binary_crossentropy')\n\n# Training\n# Let's train this model for 100 epochs (with the added regularization the model is \n# less likely to overfit and can be trained longer). The models ends with a train \n# loss of 0.11 and test loss of 0.10. The difference between the two is mostly due \n# to the regularization term being added to the loss during training (worth about 0.01).\nautoencoder.fit(x_train, x_train,\n epochs=100,\n batch_size=256,\n shuffle=True,\n validation_data=(x_test, x_test))\n\n\n# encode and decode some digits\n# note that we take them from the *test* set\nencoded_imgs = encoder.predict(x_test)\ndecoded_imgs = decoder.predict(encoded_imgs)\n\n\n# Show the results\nn = 10 # how many digits we will display\nplt.figure(figsize=(20, 4))\nfor i in range(n):\n # display original\n ax = plt.subplot(2, n, i + 1)\n plt.imshow(x_test[i].reshape(28, 28))\n plt.gray()\n ax.get_xaxis().set_visible(False)\n ax.get_yaxis().set_visible(False)\n\n # display reconstruction\n ax = plt.subplot(2, n, i + 1 + n)\n plt.imshow(decoded_imgs[i].reshape(28, 28))\n plt.gray()\n ax.get_xaxis().set_visible(False)\n ax.get_yaxis().set_visible(False)\nplt.show()", "Train on 60000 samples, validate on 10000 samples\nEpoch 1/100\n60000/60000 [==============================] - 1s 16us/step - loss: 0.6732 - val_loss: 0.6485\nEpoch 2/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.6284 - val_loss: 0.6090\nEpoch 3/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.5916 - val_loss: 0.5749\nEpoch 4/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.5598 - val_loss: 0.5454\nEpoch 5/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.5323 - val_loss: 0.5198\nEpoch 6/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.5084 - val_loss: 0.4975\nEpoch 7/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.4875 - val_loss: 0.4780\nEpoch 8/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.4692 - val_loss: 0.4609\nEpoch 9/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.4531 - val_loss: 0.4457\nEpoch 10/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.4389 - val_loss: 0.4324\nEpoch 11/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.4262 - val_loss: 0.4205\nEpoch 12/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.4150 - val_loss: 0.4098\nEpoch 13/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.4049 - val_loss: 0.4003\nEpoch 14/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.3959 - val_loss: 0.3918\nEpoch 15/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.3877 - val_loss: 0.3840\nEpoch 16/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.3804 - val_loss: 0.3771\nEpoch 17/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.3737 - val_loss: 0.3707\nEpoch 18/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.3676 - val_loss: 0.3649\nEpoch 19/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.3621 - val_loss: 0.3596\nEpoch 20/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.3570 - val_loss: 0.3548\nEpoch 21/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.3524 - val_loss: 0.3503\nEpoch 22/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.3481 - val_loss: 0.3463\nEpoch 23/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.3442 - val_loss: 0.3425\nEpoch 24/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.3406 - val_loss: 0.3390\nEpoch 25/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.3372 - val_loss: 0.3357\nEpoch 26/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.3341 - val_loss: 0.3327\nEpoch 27/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.3312 - val_loss: 0.3299\nEpoch 28/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.3285 - val_loss: 0.3273\nEpoch 29/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.3259 - val_loss: 0.3249\nEpoch 30/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.3236 - val_loss: 0.3226\nEpoch 31/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.3213 - val_loss: 0.3204\nEpoch 32/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.3193 - val_loss: 0.3184\nEpoch 33/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.3173 - val_loss: 0.3165\nEpoch 34/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.3155 - val_loss: 0.3147\nEpoch 35/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.3138 - val_loss: 0.3131\nEpoch 36/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.3121 - val_loss: 0.3115\nEpoch 37/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.3106 - val_loss: 0.3100\nEpoch 38/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.3091 - val_loss: 0.3085\nEpoch 39/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.3077 - val_loss: 0.3072\nEpoch 40/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.3064 - val_loss: 0.3059\nEpoch 41/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.3052 - val_loss: 0.3047\nEpoch 42/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.3040 - val_loss: 0.3035\nEpoch 43/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.3029 - val_loss: 0.3024\nEpoch 44/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.3018 - val_loss: 0.3014\nEpoch 45/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.3008 - val_loss: 0.3004\nEpoch 46/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2998 - val_loss: 0.2994\nEpoch 47/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2989 - val_loss: 0.2985\nEpoch 48/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2980 - val_loss: 0.2976\nEpoch 49/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2971 - val_loss: 0.2968\nEpoch 50/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2963 - val_loss: 0.2960\nEpoch 51/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2955 - val_loss: 0.2952\nEpoch 52/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2948 - val_loss: 0.2945\nEpoch 53/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2941 - val_loss: 0.2938\nEpoch 54/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2934 - val_loss: 0.2931\nEpoch 55/100\n60000/60000 [==============================] - 1s 13us/step - loss: 0.2927 - val_loss: 0.2925\nEpoch 56/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2921 - val_loss: 0.2918\nEpoch 57/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2915 - val_loss: 0.2912\nEpoch 58/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2909 - val_loss: 0.2906\nEpoch 59/100\n60000/60000 [==============================] - 1s 13us/step - loss: 0.2903 - val_loss: 0.2901\nEpoch 60/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2898 - val_loss: 0.2895\nEpoch 61/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2892 - val_loss: 0.2890\nEpoch 62/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2887 - val_loss: 0.2885\nEpoch 63/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2882 - val_loss: 0.2880\nEpoch 64/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2877 - val_loss: 0.2875\nEpoch 65/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2873 - val_loss: 0.2871\nEpoch 66/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2868 - val_loss: 0.2867\nEpoch 67/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2864 - val_loss: 0.2862\nEpoch 68/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2860 - val_loss: 0.2858\nEpoch 69/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2856 - val_loss: 0.2854\nEpoch 70/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2852 - val_loss: 0.2850\nEpoch 71/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2848 - val_loss: 0.2846\nEpoch 72/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2844 - val_loss: 0.2843\nEpoch 73/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2841 - val_loss: 0.2839\nEpoch 74/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2837 - val_loss: 0.2836\nEpoch 75/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2834 - val_loss: 0.2832\nEpoch 76/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2831 - val_loss: 0.2829\nEpoch 77/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2828 - val_loss: 0.2826\nEpoch 78/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2825 - val_loss: 0.2823\nEpoch 79/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2822 - val_loss: 0.2820\nEpoch 80/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2819 - val_loss: 0.2817\nEpoch 81/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2816 - val_loss: 0.2814\nEpoch 82/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2813 - val_loss: 0.2812\nEpoch 83/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2810 - val_loss: 0.2809\nEpoch 84/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2808 - val_loss: 0.2806\nEpoch 85/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2805 - val_loss: 0.2804\nEpoch 86/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2803 - val_loss: 0.2801\nEpoch 87/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2800 - val_loss: 0.2799\nEpoch 88/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2798 - val_loss: 0.2796\nEpoch 89/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2796 - val_loss: 0.2794\nEpoch 90/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2793 - val_loss: 0.2792\nEpoch 91/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2791 - val_loss: 0.2790\nEpoch 92/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2789 - val_loss: 0.2788\nEpoch 93/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2787 - val_loss: 0.2785\nEpoch 94/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2785 - val_loss: 0.2783\nEpoch 95/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2783 - val_loss: 0.2781\nEpoch 96/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2781 - val_loss: 0.2779\nEpoch 97/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2779 - val_loss: 0.2778\nEpoch 98/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2777 - val_loss: 0.2776\nEpoch 99/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2775 - val_loss: 0.2774\nEpoch 100/100\n60000/60000 [==============================] - 1s 12us/step - loss: 0.2774 - val_loss: 0.2772\n" ] ] ]
[ "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code" ]
[ [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ] ]
ecebe23ed72180374d51374d3613bc4f5ddad496
230,156
ipynb
Jupyter Notebook
seaborn.ipynb
m-rafiul-islam/data-science
670de5ad6f7136f80e107e5d78c11143ac010e01
[ "MIT" ]
null
null
null
seaborn.ipynb
m-rafiul-islam/data-science
670de5ad6f7136f80e107e5d78c11143ac010e01
[ "MIT" ]
null
null
null
seaborn.ipynb
m-rafiul-islam/data-science
670de5ad6f7136f80e107e5d78c11143ac010e01
[ "MIT" ]
null
null
null
181.225197
41,342
0.825214
[ [ [ "<a href=\"https://colab.research.google.com/github/m-rafiul-islam/data-science/blob/main/seaborn.ipynb\" target=\"_parent\"><img src=\"https://colab.research.google.com/assets/colab-badge.svg\" alt=\"Open In Colab\"/></a>", "_____no_output_____" ] ], [ [ "import seaborn as sns\ntips = sns.load_dataset(\"tips\")\n%matplotlib inline \ntips.head()", "_____no_output_____" ], [ "sns.(x='total_bill',data=tips)", "_____no_output_____" ], [ "# tips.head()\nsns.factorplot('day','total_bill', data=tips,kind='violin',hue='sex')", "/usr/local/lib/python3.7/dist-packages/seaborn/categorical.py:3717: UserWarning: The `factorplot` function has been renamed to `catplot`. The original name will be removed in a future release. Please update your code. Note that the default `kind` in `factorplot` (`'point'`) has changed `'strip'` in `catplot`.\n warnings.warn(msg)\n/usr/local/lib/python3.7/dist-packages/seaborn/_decorators.py:43: FutureWarning: Pass the following variables as keyword args: x, y. From version 0.12, the only valid positional argument will be `data`, and passing other arguments without an explicit keyword will result in an error or misinterpretation.\n FutureWarning\n" ], [ "flights = sns.load_dataset('flights')\nflights.head()", "_____no_output_____" ], [ "tcor = tips.corr()\nsns.heatmap(tcor)\nsns.heatmap(tcor,annot=True,cmap='coolwarm')", "_____no_output_____" ], [ "fp = flights.pivot_table(values='passengers',index='month',columns='year')\nsns.heatmap(fp,cmap='magma',linecolor='white',linewidths=2)", "_____no_output_____" ], [ "", "_____no_output_____" ] ], [ [ "`Grides`\n\n```\n`# This is formatted as code`\n```\n\n", "_____no_output_____" ] ], [ [ "iris = sns.load_dataset('iris')\niris.head()", "_____no_output_____" ], [ "iris['species'].unique()", "_____no_output_____" ], [ "# sns.pairplot(iris)\n# g = sns.PairGrid(iris)\nimport matplotlib.pyplot as plt\ng = sns.FacetGrid(data=tips,col='time',row='smoker')\ng.map(plt.scatter,'total_bill', 'tip')\n", "_____no_output_____" ], [ "sns.lmplot(x='total_bill',y='tip',data=tips,hue='sex',markers=['o','v'])", "_____no_output_____" ], [ "from plotly import __version__\nprint(__version__)", "5.5.0\n" ], [ "import seaborn as sns\n# import cufflinks as cf \n", "_____no_output_____" ], [ "titanic = sns.load_dataset('titanic')\ntitanic.head()", "_____no_output_____" ], [ "# CODE HERE\n# REPLICATE EXERCISE PLOT IMAGE BELOW\n# BE CAREFUL NOT TO OVERWRITE CELL BELOW\n# THAT WOULD REMOVE THE EXERCISE PLOT IMAGE!\n# sns.boxplot(x='fare',y='age',data=titanic)\n# sns.displot(x='fare',y='age',data=titanic)\nsns.jointplot(x='fare',y='age',data=titanic)\n", "_____no_output_____" ], [ "from numpy.lib.function_base import corrcoef\n# CODE HERE\n# REPLICATE EXERCISE PLOT IMAGE BELOW\n# BE CAREFUL NOT TO OVERWRITE CELL BELOW\n# THAT WOULD REMOVE THE EXERCISE PLOT IMAGE!\n# sns.boxplot(x='fare',y='age',data=titanic)\n# sns.displot(x='fare',y='age',data=titanic)\nsns.jointplot(x='fare',y='age',data=titanic,corrcoef = True )\n", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code" ]
[ [ "markdown" ], [ "code", "code", "code", "code", "code", "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code", "code", "code", "code", "code", "code" ] ]
ecec095206c7b50a7594129218e33b2b3332f657
522,745
ipynb
Jupyter Notebook
titanic.ipynb
7wikd/Titanic
327ea77ec1b9505000acb34e715b6ad8caaf06d5
[ "MIT" ]
null
null
null
titanic.ipynb
7wikd/Titanic
327ea77ec1b9505000acb34e715b6ad8caaf06d5
[ "MIT" ]
null
null
null
titanic.ipynb
7wikd/Titanic
327ea77ec1b9505000acb34e715b6ad8caaf06d5
[ "MIT" ]
null
null
null
44.165681
28,740
0.504005
[ [ [ "# data analysis and wrangling\nimport pandas as pd\nimport numpy as np\nimport random as rnd\n\n# visualization\nimport seaborn as sns\nimport matplotlib.pyplot as plt\n%matplotlib inline", "_____no_output_____" ], [ "train_df = pd.read_csv('./train.csv')\ntest_df = pd.read_csv('./test.csv')\ncombine = [train_df, test_df]", "_____no_output_____" ], [ "print(train_df.columns.values)", "['PassengerId' 'Survived' 'Pclass' 'Name' 'Sex' 'Age' 'SibSp' 'Parch'\n 'Ticket' 'Fare' 'Cabin' 'Embarked']\n" ], [ "train_df.head()", "_____no_output_____" ], [ "train_df.info()\nprint('*'*40)\ntest_df.info()", "<class 'pandas.core.frame.DataFrame'>\nRangeIndex: 891 entries, 0 to 890\nData columns (total 12 columns):\n # Column Non-Null Count Dtype \n--- ------ -------------- ----- \n 0 PassengerId 891 non-null int64 \n 1 Survived 891 non-null int64 \n 2 Pclass 891 non-null int64 \n 3 Name 891 non-null object \n 4 Sex 891 non-null object \n 5 Age 714 non-null float64\n 6 SibSp 891 non-null int64 \n 7 Parch 891 non-null int64 \n 8 Ticket 891 non-null object \n 9 Fare 891 non-null float64\n 10 Cabin 204 non-null object \n 11 Embarked 889 non-null object \ndtypes: float64(2), int64(5), object(5)\nmemory usage: 83.7+ KB\n****************************************\n<class 'pandas.core.frame.DataFrame'>\nRangeIndex: 418 entries, 0 to 417\nData columns (total 11 columns):\n # Column Non-Null Count Dtype \n--- ------ -------------- ----- \n 0 PassengerId 418 non-null int64 \n 1 Pclass 418 non-null int64 \n 2 Name 418 non-null object \n 3 Sex 418 non-null object \n 4 Age 332 non-null float64\n 5 SibSp 418 non-null int64 \n 6 Parch 418 non-null int64 \n 7 Ticket 418 non-null object \n 8 Fare 417 non-null float64\n 9 Cabin 91 non-null object \n 10 Embarked 418 non-null object \ndtypes: float64(2), int64(4), object(5)\nmemory usage: 36.0+ KB\n" ], [ "train_df.describe()", "_____no_output_____" ], [ "train_df.describe(include=['O'])", "_____no_output_____" ], [ "train_df[['Pclass', 'Survived']].groupby(['Pclass'], as_index=False).mean().sort_values(by='Survived', ascending=False)", "_____no_output_____" ], [ "train_df[[\"Sex\", \"Survived\"]].groupby(['Sex'], as_index=False).mean().sort_values(by='Survived', ascending=False)", "_____no_output_____" ], [ "train_df[[\"SibSp\", \"Survived\"]].groupby(['SibSp'], as_index=False).mean().sort_values(by='Survived', ascending=False)", "_____no_output_____" ], [ "train_df[[\"Parch\", \"Survived\"]].groupby(['Parch'], as_index=False).mean().sort_values(by='Survived', ascending=False)", "_____no_output_____" ], [ "graph = sns.FacetGrid(train_df, col='Survived')\ngraph.map(plt.hist, 'Age', bins=20)", "_____no_output_____" ], [ "grid = sns.FacetGrid(train_df, col='Survived', row='Pclass', size=2.2, aspect=1.6)\ngrid.map(plt.hist, 'Age', alpha=.5, bins=20)\ngrid.add_legend()", "/home/satwikd/.local/lib/python3.8/site-packages/seaborn/axisgrid.py:316: UserWarning: The `size` parameter has been renamed to `height`; please update your code.\n warnings.warn(msg, UserWarning)\n" ], [ "grid = sns.FacetGrid(train_df, row='Embarked', size=2.2, aspect=1.6)\ngrid.map(sns.pointplot, 'Pclass', 'Survived', 'Sex', palette='deep')\ngrid.add_legend()", "/home/satwikd/.local/lib/python3.8/site-packages/seaborn/axisgrid.py:316: UserWarning: The `size` parameter has been renamed to `height`; please update your code.\n warnings.warn(msg, UserWarning)\n/home/satwikd/.local/lib/python3.8/site-packages/seaborn/axisgrid.py:643: UserWarning: Using the pointplot function without specifying `order` is likely to produce an incorrect plot.\n warnings.warn(warning)\n/home/satwikd/.local/lib/python3.8/site-packages/seaborn/axisgrid.py:648: UserWarning: Using the pointplot function without specifying `hue_order` is likely to produce an incorrect plot.\n warnings.warn(warning)\n" ], [ "grid = sns.FacetGrid(train_df, row='Embarked', col='Survived', size=2.2, aspect=1.6)\ngrid.map(sns.barplot, 'Sex', 'Fare', alpha=.5, ci=None)\ngrid.add_legend()", "/home/satwikd/.local/lib/python3.8/site-packages/seaborn/axisgrid.py:316: UserWarning: The `size` parameter has been renamed to `height`; please update your code.\n warnings.warn(msg, UserWarning)\n/home/satwikd/.local/lib/python3.8/site-packages/seaborn/axisgrid.py:643: UserWarning: Using the barplot function without specifying `order` is likely to produce an incorrect plot.\n warnings.warn(warning)\n" ], [ "print(\"Before\", train_df.shape, test_df.shape, combine[0].shape, combine[1].shape)\n\ntrain_df = train_df.drop(['Ticket', 'Cabin'], axis=1)\ntest_df = test_df.drop(['Ticket', 'Cabin'], axis=1)\ncombine = [train_df, test_df]\n\n\"After\", train_df.shape, test_df.shape, combine[0].shape, combine[1].shape", "Before (891, 12) (418, 11) (891, 12) (418, 11)\n" ], [ "for dataset in combine:\n dataset['Title'] = dataset.Name.str.extract(' ([A-Za-z]+)\\.', expand=False)\n\npd.crosstab(train_df['Title'], train_df['Sex'])", "_____no_output_____" ], [ "for dataset in combine:\n dataset['Title'] = dataset['Title'].replace(['Lady', 'Countess','Capt', 'Col','Don', 'Dr', 'Major', 'Rev', 'Sir', 'Jonkheer', 'Dona'], 'Rare')\n\n dataset['Title'] = dataset['Title'].replace('Mlle', 'Miss')\n dataset['Title'] = dataset['Title'].replace('Ms', 'Miss')\n dataset['Title'] = dataset['Title'].replace('Mme', 'Mrs')\n \ntrain_df[['Title', 'Survived']].groupby(['Title'], as_index=False).mean()", "_____no_output_____" ], [ "title_mapping = {\"Mr\": 1, \"Miss\": 2, \"Mrs\": 3, \"Master\": 4, \"Rare\": 5}\nfor dataset in combine:\n dataset['Title'] = dataset['Title'].map(title_mapping)\n dataset['Title'] = dataset['Title'].fillna(0)", "_____no_output_____" ], [ "train_df = train_df.drop(['Name', 'PassengerId'], axis=1)\ntest_df = test_df.drop(['Name', 'PassengerId'], axis=1)\ncombine = [train_df, test_df]\ntrain_df.shape, test_df.shape", "_____no_output_____" ], [ "for dataset in combine:\n dataset['Sex'] = dataset['Sex'].map( {'female': 1, 'male': 0} ).astype(int)\n\ntrain_df.head()", "_____no_output_____" ], [ "grid = sns.FacetGrid(train_df, row='Pclass', col='Sex', size=2.2, aspect=1.6)\ngrid.map(plt.hist, 'Age', alpha=.5, bins=20)\ngrid.add_legend()", "/home/satwikd/.local/lib/python3.8/site-packages/seaborn/axisgrid.py:316: UserWarning: The `size` parameter has been renamed to `height`; please update your code.\n warnings.warn(msg, UserWarning)\n" ], [ "for dataset in combine:\n age_avg = dataset['Age'].mean()\n age_std = dataset['Age'].std()\n age_null_count = dataset['Age'].isnull().sum()\n \n age_null_random_list = np.random.randint(age_avg - age_std, age_avg + age_std, size=age_null_count)\n dataset['Age'][np.isnan(dataset['Age'])] = age_null_random_list\n dataset['Age'] = dataset['Age'].astype(int)\n \ntrain_df['CategoricalAge'] = pd.cut(train_df['Age'], 5)\n\nprint (train_df[['CategoricalAge', 'Survived']].groupby(['CategoricalAge'], as_index=False).mean())", " CategoricalAge Survived\n0 (-0.08, 16.0] 0.518182\n1 (16.0, 32.0] 0.352550\n2 (32.0, 48.0] 0.380000\n3 (48.0, 64.0] 0.434783\n4 (64.0, 80.0] 0.090909\n<ipython-input-153-d29f5f4de75f>:7: SettingWithCopyWarning: \nA value is trying to be set on a copy of a slice from a DataFrame\n\nSee the caveats in the documentation: https://pandas.pydata.org/pandas-docs/stable/user_guide/indexing.html#returning-a-view-versus-a-copy\n dataset['Age'][np.isnan(dataset['Age'])] = age_null_random_list\n" ], [ "for dataset in combine: \n dataset.loc[ dataset['Age'] <= 16, 'Age'] = 0\n dataset.loc[(dataset['Age'] > 16) & (dataset['Age'] <= 32), 'Age'] = 1\n dataset.loc[(dataset['Age'] > 32) & (dataset['Age'] <= 48), 'Age'] = 2\n dataset.loc[(dataset['Age'] > 48) & (dataset['Age'] <= 64), 'Age'] = 3\n dataset.loc[ dataset['Age'] > 64, 'Age']=4\ntrain_df.head()", "_____no_output_____" ], [ "train_df = train_df.drop(['CategoricalAge'], axis=1)\ncombine = [train_df, test_df]\ntrain_df.head()", "_____no_output_____" ], [ "for dataset in combine:\n dataset['FamilySize'] = dataset['SibSp'] + dataset['Parch'] + 1\n\ntrain_df[['FamilySize', 'Survived']].groupby(['FamilySize'], as_index=False).mean().sort_values(by='Survived', ascending=False)", "_____no_output_____" ], [ "for dataset in combine:\n dataset['IsAlone'] = 0\n dataset.loc[dataset['FamilySize'] == 1, 'IsAlone'] = 1\n\ntrain_df[['IsAlone', 'Survived']].groupby(['IsAlone'], as_index=False).mean()", "_____no_output_____" ], [ "train_df = train_df.drop(['Parch', 'SibSp', 'FamilySize'], axis=1)\ntest_df = test_df.drop(['Parch', 'SibSp', 'FamilySize'], axis=1)\ncombine = [train_df, test_df]", "_____no_output_____" ], [ "for dataset in combine:\n dataset['Age*Class'] = dataset.Age * dataset.Pclass\n\ntrain_df.loc[:, ['Age*Class', 'Age', 'Pclass']].head(10)", "_____no_output_____" ], [ "freq_port = train_df.Embarked.dropna().mode()[0]\nfreq_port", "_____no_output_____" ], [ "for dataset in combine:\n dataset['Embarked'] = dataset['Embarked'].fillna(freq_port)\n \ntrain_df[['Embarked', 'Survived']].groupby(['Embarked'], as_index=False).mean().sort_values(by='Survived', ascending=False)", "_____no_output_____" ], [ "for dataset in combine:\n dataset['Embarked'] = dataset['Embarked'].map( {'S': 0, 'C': 1, 'Q': 2} ).astype(int)\n\ntrain_df.head()\n", "_____no_output_____" ], [ "test_df['Fare'].fillna(test_df['Fare'].dropna().median(), inplace=True)\ntest_df.head()", "_____no_output_____" ], [ "train_df['FareBand'] = pd.qcut(train_df['Fare'], 4)\ntrain_df[['FareBand', 'Survived']].groupby(['FareBand'], as_index=False).mean().sort_values(by='FareBand', ascending=True)", "_____no_output_____" ], [ "for dataset in combine:\n dataset.loc[ dataset['Fare'] <= 7.91, 'Fare'] = 0\n dataset.loc[(dataset['Fare'] > 7.91) & (dataset['Fare'] <= 14.454), 'Fare'] = 1\n dataset.loc[(dataset['Fare'] > 14.454) & (dataset['Fare'] <= 31), 'Fare'] = 2\n dataset.loc[ dataset['Fare'] > 31, 'Fare'] = 3\n dataset['Fare'] = dataset['Fare'].astype(int)\n\ntrain_df = train_df.drop(['FareBand'], axis=1)\ncombine = [train_df, test_df]\n \ntrain_df.head(10)", "_____no_output_____" ], [ "from sklearn.linear_model import LogisticRegression\nfrom sklearn.svm import SVC, LinearSVC\nfrom sklearn.ensemble import RandomForestClassifier\nfrom sklearn.neighbors import KNeighborsClassifier\nfrom sklearn.naive_bayes import GaussianNB\nfrom sklearn.linear_model import Perceptron\nfrom sklearn.linear_model import SGDClassifier\nfrom sklearn.tree import DecisionTreeClassifier", "_____no_output_____" ], [ "X_train = train_df.drop(\"Survived\", axis=1)\nY_train = train_df[\"Survived\"]\nX_test = test_df\nX_train.shape, Y_train.shape, X_test.shape", "_____no_output_____" ], [ "# Logistic Regression\nlogreg = LogisticRegression()\nlogreg.fit(X_train, Y_train)\nY_pred = logreg.predict(X_test)\nacc_log = round(logreg.score(X_train, Y_train) * 100, 2)", "_____no_output_____" ], [ "coeff_df = pd.DataFrame(train_df.columns.delete(0))\ncoeff_df.columns = ['Feature']\ncoeff_df[\"Correlation\"] = pd.Series(logreg.coef_[0])\n\ncoeff_df.sort_values(by='Correlation', ascending=False)", "_____no_output_____" ], [ "# Support Vector Machines\n\nsvc = SVC()\nsvc.fit(X_train, Y_train)\nY_pred = svc.predict(X_test)\nacc_svc = round(svc.score(X_train, Y_train) * 100, 2)", "_____no_output_____" ], [ "#KNN\nknn = KNeighborsClassifier(n_neighbors = 3)\nknn.fit(X_train, Y_train)\nY_pred = knn.predict(X_test)\nacc_knn = round(knn.score(X_train, Y_train) * 100, 2)", "_____no_output_____" ], [ "# Gaussian Naive Bayes\n\ngaussian = GaussianNB()\ngaussian.fit(X_train, Y_train)\nY_pred = gaussian.predict(X_test)\nacc_gaussian = round(gaussian.score(X_train, Y_train) * 100, 2)", "_____no_output_____" ], [ "# Perceptron\n\nperceptron = Perceptron()\nperceptron.fit(X_train, Y_train)\nY_pred = perceptron.predict(X_test)\nacc_perceptron = round(perceptron.score(X_train, Y_train) * 100, 2)", "_____no_output_____" ], [ "# Linear SVC\n\nlinear_svc = LinearSVC()\nlinear_svc.fit(X_train, Y_train)\nY_pred = linear_svc.predict(X_test)\nacc_linear_svc = round(linear_svc.score(X_train, Y_train) * 100, 2)", "/home/satwikd/.local/lib/python3.8/site-packages/sklearn/svm/_base.py:985: ConvergenceWarning: Liblinear failed to converge, increase the number of iterations.\n warnings.warn(\"Liblinear failed to converge, increase \"\n" ], [ "# Stochastic Gradient Descent\n\nsgd = SGDClassifier()\nsgd.fit(X_train, Y_train)\nY_pred = sgd.predict(X_test)\nacc_sgd = round(sgd.score(X_train, Y_train) * 100, 2)", "_____no_output_____" ], [ "# Decision Tree\n\ndecision_tree = DecisionTreeClassifier()\ndecision_tree.fit(X_train, Y_train)\nY_pred = decision_tree.predict(X_test)\nacc_decision_tree = round(decision_tree.score(X_train, Y_train) * 100, 2)", "_____no_output_____" ], [ "# Random Forest\n\nrandom_forest = RandomForestClassifier(n_estimators=100)\nrandom_forest.fit(X_train, Y_train)\nY_pred = random_forest.predict(X_test)\nrandom_forest.score(X_train, Y_train)\nacc_random_forest = round(random_forest.score(X_train, Y_train) * 100, 2)", "_____no_output_____" ], [ "models = pd.DataFrame({\n 'Model': ['Support Vector Machines', 'KNN', 'Logistic Regression', \n 'Random Forest', 'Naive Bayes', 'Perceptron', \n 'Stochastic Gradient Decent', 'Linear SVC', \n 'Decision Tree'],\n 'Accuracy': [acc_svc, acc_knn, acc_log, \n acc_random_forest, acc_gaussian, acc_perceptron, \n acc_sgd, acc_linear_svc, acc_decision_tree]})\nmodels.sort_values(by='Accuracy', ascending=False)", "_____no_output_____" ] ] ]
[ "code" ]
[ [ "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code", "code" ] ]
ecec0eb019e91e39570089d443133398738657c7
52,214
ipynb
Jupyter Notebook
anticipy_hello.ipynb
edwinnglabs/timeseries-notebooks
9df64fb3a7b70cbfc572e7a926c0ec03bf6351f0
[ "MIT" ]
null
null
null
anticipy_hello.ipynb
edwinnglabs/timeseries-notebooks
9df64fb3a7b70cbfc572e7a926c0ec03bf6351f0
[ "MIT" ]
null
null
null
anticipy_hello.ipynb
edwinnglabs/timeseries-notebooks
9df64fb3a7b70cbfc572e7a926c0ec03bf6351f0
[ "MIT" ]
null
null
null
143.445055
33,920
0.840081
[ [ [ "!pip install -U anticipy\nimport logging, sys\nlogging.disable(sys.maxsize)", "Collecting anticipy\n Downloading anticipy-0.2.1-py3-none-any.whl (74 kB)\n\u001b[?25l\r\u001b[K |████▍ | 10 kB 20.1 MB/s eta 0:00:01\r\u001b[K |████████▉ | 20 kB 3.5 MB/s eta 0:00:01\r\u001b[K |█████████████▏ | 30 kB 4.9 MB/s eta 0:00:01\r\u001b[K |█████████████████▋ | 40 kB 6.2 MB/s eta 0:00:01\r\u001b[K |██████████████████████ | 51 kB 7.3 MB/s eta 0:00:01\r\u001b[K |██████████████████████████▍ | 61 kB 8.4 MB/s eta 0:00:01\r\u001b[K |██████████████████████████████▊ | 71 kB 9.5 MB/s eta 0:00:01\r\u001b[K |████████████████████████████████| 74 kB 1.6 MB/s \n\u001b[?25hRequirement already satisfied: pandas>=0.23.0 in /usr/local/lib/python3.7/dist-packages (from anticipy) (1.1.5)\nRequirement already satisfied: plotly>=3.5.0 in /usr/local/lib/python3.7/dist-packages (from anticipy) (4.4.1)\nRequirement already satisfied: numpy>=1.15.1 in /usr/local/lib/python3.7/dist-packages (from anticipy) (1.19.5)\nRequirement already satisfied: scipy>=1.0.0 in /usr/local/lib/python3.7/dist-packages (from anticipy) (1.4.1)\nRequirement already satisfied: python-dateutil>=2.7.3 in /usr/local/lib/python3.7/dist-packages (from pandas>=0.23.0->anticipy) (2.8.2)\nRequirement already satisfied: pytz>=2017.2 in /usr/local/lib/python3.7/dist-packages (from pandas>=0.23.0->anticipy) (2018.9)\nRequirement already satisfied: six in /usr/local/lib/python3.7/dist-packages (from plotly>=3.5.0->anticipy) (1.15.0)\nRequirement already satisfied: retrying>=1.3.3 in /usr/local/lib/python3.7/dist-packages (from plotly>=3.5.0->anticipy) (1.3.3)\nInstalling collected packages: anticipy\nSuccessfully installed anticipy-0.2.1\n" ], [ "!pip install microprediction", "Collecting microprediction\n Downloading microprediction-0.18.8-py3-none-any.whl (73 kB)\n\u001b[?25l\r\u001b[K |████▍ | 10 kB 20.4 MB/s eta 0:00:01\r\u001b[K |████████▉ | 20 kB 25.5 MB/s eta 0:00:01\r\u001b[K |█████████████▎ | 30 kB 31.6 MB/s eta 0:00:01\r\u001b[K |█████████████████▊ | 40 kB 31.4 MB/s eta 0:00:01\r\u001b[K |██████████████████████▏ | 51 kB 20.5 MB/s eta 0:00:01\r\u001b[K |██████████████████████████▋ | 61 kB 22.8 MB/s eta 0:00:01\r\u001b[K |███████████████████████████████ | 71 kB 19.9 MB/s eta 0:00:01\r\u001b[K |████████████████████████████████| 73 kB 2.2 MB/s \n\u001b[?25hRequirement already satisfied: scikit-learn in /usr/local/lib/python3.7/dist-packages (from microprediction) (1.0.1)\nRequirement already satisfied: requests in /usr/local/lib/python3.7/dist-packages (from microprediction) (2.23.0)\nCollecting microconventions>=0.5.0\n Downloading microconventions-0.5.4-py3-none-any.whl (16 kB)\nCollecting pytz>=2021.3\n Downloading pytz-2021.3-py2.py3-none-any.whl (503 kB)\n\u001b[K |████████████████████████████████| 503 kB 46.9 MB/s \n\u001b[?25hCollecting contexttimer\n Downloading contexttimer-0.3.3.tar.gz (4.9 kB)\nRequirement already satisfied: pathlib in /usr/local/lib/python3.7/dist-packages (from microprediction) (1.0.1)\nCollecting tdigest\n Downloading tdigest-0.5.2.2-py3-none-any.whl (9.4 kB)\nRequirement already satisfied: statsmodels in /usr/local/lib/python3.7/dist-packages (from microprediction) (0.10.2)\nCollecting copulas\n Downloading copulas-0.6.0-py2.py3-none-any.whl (50 kB)\n\u001b[K |████████████████████████████████| 50 kB 5.3 MB/s \n\u001b[?25hRequirement already satisfied: pandas in /usr/local/lib/python3.7/dist-packages (from microprediction) (1.1.5)\nCollecting apscheduler\n Downloading APScheduler-3.8.1-py2.py3-none-any.whl (59 kB)\n\u001b[K |████████████████████████████████| 59 kB 5.5 MB/s \n\u001b[?25hRequirement already satisfied: hyperopt in /usr/local/lib/python3.7/dist-packages (from microprediction) (0.1.2)\nCollecting pycoingecko\n Downloading pycoingecko-2.2.0-py3-none-any.whl (8.3 kB)\nCollecting getjson\n Downloading getjson-1.0.0-py3-none-any.whl (2.6 kB)\nCollecting genson\n Downloading genson-1.2.2.tar.gz (34 kB)\nCollecting numpy>=1.20.1\n Downloading numpy-1.21.5-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl (15.7 MB)\n\u001b[K |████████████████████████████████| 15.7 MB 46.9 MB/s \n\u001b[?25hRequirement already satisfied: scipy in /usr/local/lib/python3.7/dist-packages (from microconventions>=0.5.0->microprediction) (1.4.1)\nCollecting muid>=0.5.3\n Downloading muid-0.5.3-py3-none-any.whl (173 kB)\n\u001b[K |████████████████████████████████| 173 kB 57.1 MB/s \n\u001b[?25hCollecting pymorton\n Downloading pymorton-1.0.5-py2.py3-none-any.whl (4.4 kB)\nCollecting schema\n Downloading schema-0.7.5-py2.py3-none-any.whl (17 kB)\nCollecting deepdiff\n Downloading deepdiff-5.7.0-py3-none-any.whl (68 kB)\n\u001b[K |████████████████████████████████| 68 kB 6.8 MB/s \n\u001b[?25hCollecting tzlocal!=3.*,>=2.0\n Downloading tzlocal-4.1-py3-none-any.whl (19 kB)\nRequirement already satisfied: setuptools>=0.7 in /usr/local/lib/python3.7/dist-packages (from apscheduler->microprediction) (57.4.0)\nRequirement already satisfied: six>=1.4.0 in /usr/local/lib/python3.7/dist-packages (from apscheduler->microprediction) (1.15.0)\nCollecting backports.zoneinfo\n Downloading backports.zoneinfo-0.2.1-cp37-cp37m-manylinux1_x86_64.whl (70 kB)\n\u001b[K |████████████████████████████████| 70 kB 8.8 MB/s \n\u001b[?25hCollecting pytz-deprecation-shim\n Downloading pytz_deprecation_shim-0.1.0.post0-py2.py3-none-any.whl (15 kB)\nCollecting scipy\n Downloading scipy-1.7.3-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl (38.1 MB)\n\u001b[K |████████████████████████████████| 38.1 MB 1.3 MB/s \n\u001b[?25hRequirement already satisfied: matplotlib<4,>=3.2.0 in /usr/local/lib/python3.7/dist-packages (from copulas->microprediction) (3.2.2)\nRequirement already satisfied: python-dateutil>=2.1 in /usr/local/lib/python3.7/dist-packages (from matplotlib<4,>=3.2.0->copulas->microprediction) (2.8.2)\nRequirement already satisfied: kiwisolver>=1.0.1 in /usr/local/lib/python3.7/dist-packages (from matplotlib<4,>=3.2.0->copulas->microprediction) (1.3.2)\nRequirement already satisfied: cycler>=0.10 in /usr/local/lib/python3.7/dist-packages (from matplotlib<4,>=3.2.0->copulas->microprediction) (0.11.0)\nRequirement already satisfied: pyparsing!=2.0.4,!=2.1.2,!=2.1.6,>=2.0.1 in /usr/local/lib/python3.7/dist-packages (from matplotlib<4,>=3.2.0->copulas->microprediction) (3.0.6)\nCollecting ordered-set==4.0.2\n Downloading ordered-set-4.0.2.tar.gz (10 kB)\nCollecting backoff\n Downloading backoff-1.11.1-py2.py3-none-any.whl (13 kB)\nRequirement already satisfied: pymongo in /usr/local/lib/python3.7/dist-packages (from hyperopt->microprediction) (3.12.1)\nRequirement already satisfied: networkx in /usr/local/lib/python3.7/dist-packages (from hyperopt->microprediction) (2.6.3)\nRequirement already satisfied: tqdm in /usr/local/lib/python3.7/dist-packages (from hyperopt->microprediction) (4.62.3)\nRequirement already satisfied: future in /usr/local/lib/python3.7/dist-packages (from hyperopt->microprediction) (0.16.0)\nCollecting tzdata\n Downloading tzdata-2021.5-py2.py3-none-any.whl (339 kB)\n\u001b[K |████████████████████████████████| 339 kB 49.2 MB/s \n\u001b[?25hRequirement already satisfied: urllib3!=1.25.0,!=1.25.1,<1.26,>=1.21.1 in /usr/local/lib/python3.7/dist-packages (from requests->microprediction) (1.24.3)\nRequirement already satisfied: idna<3,>=2.5 in /usr/local/lib/python3.7/dist-packages (from requests->microprediction) (2.10)\nRequirement already satisfied: certifi>=2017.4.17 in /usr/local/lib/python3.7/dist-packages (from requests->microprediction) (2021.10.8)\nRequirement already satisfied: chardet<4,>=3.0.2 in /usr/local/lib/python3.7/dist-packages (from requests->microprediction) (3.0.4)\nRequirement already satisfied: contextlib2>=0.5.5 in /usr/local/lib/python3.7/dist-packages (from schema->microconventions>=0.5.0->microprediction) (0.5.5)\nRequirement already satisfied: threadpoolctl>=2.0.0 in /usr/local/lib/python3.7/dist-packages (from scikit-learn->microprediction) (3.0.0)\nRequirement already satisfied: joblib>=0.11 in /usr/local/lib/python3.7/dist-packages (from scikit-learn->microprediction) (1.1.0)\nRequirement already satisfied: patsy>=0.4.0 in /usr/local/lib/python3.7/dist-packages (from statsmodels->microprediction) (0.5.2)\nCollecting accumulation-tree\n Downloading accumulation_tree-0.6.2.tar.gz (12 kB)\nCollecting pyudorandom\n Downloading pyudorandom-1.0.0.tar.gz (1.6 kB)\nBuilding wheels for collected packages: contexttimer, ordered-set, genson, accumulation-tree, pyudorandom\n Building wheel for contexttimer (setup.py) ... \u001b[?25l\u001b[?25hdone\n Created wheel for contexttimer: filename=contexttimer-0.3.3-py3-none-any.whl size=5818 sha256=e79ad9c1024c3b80c422e692ed6e73dc356901c82d6a26aa197af71d44f28010\n Stored in directory: /root/.cache/pip/wheels/03/8c/3b/8eba5888c3218e78f7fc4442198abb6db2bd1125c1bfcff183\n Building wheel for ordered-set (setup.py) ... \u001b[?25l\u001b[?25hdone\n Created wheel for ordered-set: filename=ordered_set-4.0.2-py2.py3-none-any.whl size=8219 sha256=df9f7efdadeb4657bc41fdc28b5df159c16ea42a3c3b7c10d10efde91a38221c\n Stored in directory: /root/.cache/pip/wheels/73/2b/f6/26e9f84153c25050fe7c09e88f8e32a6be3c7034a38c418319\n Building wheel for genson (setup.py) ... \u001b[?25l\u001b[?25hdone\n Created wheel for genson: filename=genson-1.2.2-py2.py3-none-any.whl size=21291 sha256=b4e95bfdc76bba35b331d34c7442deacb3391bbab3c8376cb0b6a117fbc52b65\n Stored in directory: /root/.cache/pip/wheels/2e/34/be/0194d05d18bc4695b5c4969178790d535bdd23eabdb9d3b1e3\n Building wheel for accumulation-tree (setup.py) ... \u001b[?25l\u001b[?25hdone\n Created wheel for accumulation-tree: filename=accumulation_tree-0.6.2-cp37-cp37m-linux_x86_64.whl size=234425 sha256=6c109737dd76c58ef0075f13e90112c0f872410d96648f892f21e4eec750157e\n Stored in directory: /root/.cache/pip/wheels/42/32/0e/08020ae396bf92a3fd00971d0b81a6fb8f3e0681fd8912760d\n Building wheel for pyudorandom (setup.py) ... \u001b[?25l\u001b[?25hdone\n Created wheel for pyudorandom: filename=pyudorandom-1.0.0-py3-none-any.whl size=2221 sha256=6f74e6d8fda0e21b3f97deff9ad73068f7b640d2f8f03449d05a6b2df12b7a86\n Stored in directory: /root/.cache/pip/wheels/9b/d0/30/b2916c3efec5b42e574e99d479c032cfa4f13bd0713de1d194\nSuccessfully built contexttimer ordered-set genson accumulation-tree pyudorandom\nInstalling collected packages: tzdata, backports.zoneinfo, pyudorandom, pytz-deprecation-shim, pytz, ordered-set, numpy, contexttimer, backoff, accumulation-tree, tzlocal, tdigest, scipy, schema, pymorton, muid, getjson, deepdiff, pycoingecko, microconventions, genson, copulas, apscheduler, microprediction\n Attempting uninstall: pytz\n Found existing installation: pytz 2018.9\n Uninstalling pytz-2018.9:\n Successfully uninstalled pytz-2018.9\n Attempting uninstall: numpy\n Found existing installation: numpy 1.19.5\n Uninstalling numpy-1.19.5:\n Successfully uninstalled numpy-1.19.5\n Attempting uninstall: tzlocal\n Found existing installation: tzlocal 1.5.1\n Uninstalling tzlocal-1.5.1:\n Successfully uninstalled tzlocal-1.5.1\n Attempting uninstall: scipy\n Found existing installation: scipy 1.4.1\n Uninstalling scipy-1.4.1:\n Successfully uninstalled scipy-1.4.1\n\u001b[31mERROR: pip's dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflicts.\nyellowbrick 1.3.post1 requires numpy<1.20,>=1.16.0, but you have numpy 1.21.5 which is incompatible.\ndatascience 0.10.6 requires folium==0.2.1, but you have folium 0.8.3 which is incompatible.\nalbumentations 0.1.12 requires imgaug<0.2.7,>=0.2.5, but you have imgaug 0.2.9 which is incompatible.\u001b[0m\nSuccessfully installed accumulation-tree-0.6.2 apscheduler-3.8.1 backoff-1.11.1 backports.zoneinfo-0.2.1 contexttimer-0.3.3 copulas-0.6.0 deepdiff-5.7.0 genson-1.2.2 getjson-1.0.0 microconventions-0.5.4 microprediction-0.18.8 muid-0.5.3 numpy-1.21.5 ordered-set-4.0.2 pycoingecko-2.2.0 pymorton-1.0.5 pytz-2021.3 pytz-deprecation-shim-0.1.0.post0 pyudorandom-1.0.0 schema-0.7.5 scipy-1.7.3 tdigest-0.5.2.2 tzdata-2021.5 tzlocal-4.1\n" ] ], [ [ "### anticipy hello world\nSee https://www.microprediction.com/blog/popular-timeseries-packages for more packages", "_____no_output_____" ] ], [ [ "from microprediction import MicroReader\nmr = MicroReader()\nYS = mr.get_lagged_values(name='btc_raw.json')[:400]", "_____no_output_____" ], [ "import pandas as pd \nimport datetime \nimport numpy as np\nfrom anticipy import forecast\n\n\ndef anticipy_next(ys:[float])->float:\n \"\"\" Predict the next point in a series \"\"\"\n\n df = pd.DataFrame({'y': ys,\n 'date':pd.date_range(start='2021-01-01', periods=len(ys), freq='D')})\n df_forecast = forecast.run_forecast(df, extrapolate_years=1)\n return (df_forecast.q80.tail(1) + df_forecast.q20.tail(1)) / 2\n \ndef run(ys):\n \"\"\" Slow, see river package or others if you don't like \"\"\"\n burnin = 10\n y_hats = list()\n for t in range(len(ys)):\n if t>burnin:\n y_hat = anticipy_next(ys[:t])\n elif t>=1:\n y_hat = ys[t-1]\n else:\n y_hat = 0 \n y_hats.append(y_hat)\n return y_hats ", "_____no_output_____" ], [ "XS = run(YS)", "_____no_output_____" ], [ "import matplotlib.pyplot as plt\nplt.plot(YS[200:400],'*b')\nplt.plot(XS[200:400],'g')", "_____no_output_____" ] ] ]
[ "code", "markdown", "code" ]
[ [ "code", "code" ], [ "markdown" ], [ "code", "code", "code", "code" ] ]
ecec1aa75ea3cb64f6396ff6542b9ecad697bb70
3,680
ipynb
Jupyter Notebook
Jahr 1/M/Gleichungen/Linearegleichungssystem.ipynb
BackInBash/Technikerschule
6e30654417732fae065e36a40789866ccca6aa7e
[ "MIT" ]
2
2021-01-20T16:16:41.000Z
2022-01-12T15:37:53.000Z
Jahr 1/M/Gleichungen/Linearegleichungssystem.ipynb
BackInBash/Technikerschule
6e30654417732fae065e36a40789866ccca6aa7e
[ "MIT" ]
2
2020-06-17T21:55:24.000Z
2021-09-08T20:40:41.000Z
Jahr 1/M/Gleichungen/Linearegleichungssystem.ipynb
BackInBash/Technikerschule
6e30654417732fae065e36a40789866ccca6aa7e
[ "MIT" ]
1
2020-12-28T13:03:34.000Z
2020-12-28T13:03:34.000Z
20.674157
108
0.447011
[ [ [ "# Linearegleichungssystem (LGS)\n\nI\n\\begin{equation*}\n5x+7y=6\n\\end{equation*}\n\nII\n\\begin{equation*}\n4x+y=-1\n\\end{equation*}\n\n## Struktur\nI\n\\begin{equation*}\na*x+b*y=c\n\\end{equation*}\n\nII\n\\begin{equation*}\nd*x+e*y=f\n\\end{equation*}\n\nX = Variable\nd, e, f = Koeffizient (Zahlen die vor der Varaiblen stehen)\n\n## Lösungsverfahren\n\n### Einsetzungsverfahren (EV)\n\nBsp:\nI\n\\begin{equation*}\n8x+2y=6\n\\end{equation*}\nII\n\\begin{equation*}\n4x+y=-1\n\\end{equation*}\nNach Y umformen\n\n\\begin{equation*}\ny=-1-4x\n\\end{equation*}\n\nI\n\\begin{equation*}\n8x+2(-1-4x)=6\n\\end{equation*}\n\nI\n\\begin{equation*}\n8x-2-8=6\n\\end{equation*}\n\nErgebnis:\n\\begin{equation*}\n-2=6\n\\end{equation*}\n\n### Gleichungsverfahren (GV)\n\nI\n\\begin{equation*}\n6x+3y=15\n\\end{equation*}\nII\n\\begin{equation*}\n8x-2y=14\n\\end{equation*}\n\nI\n\\begin{equation*}\ny=\\frac{15-6x}{3} = 5-2x\n\\end{equation*}\nII\n\\begin{equation*}\ny=\\frac{14-8x}{-2}=-7+4x\n\\end{equation*}\n\nI=II\n\\begin{equation*}\n5-2x=-7+4x\n\\end{equation*}\n\\begin{equation*}\n12-2x=4x\n\\end{equation*}\n\\begin{equation*}\n12=6x\n\\end{equation*}\n\\begin{equation*}\nx=2\n\\end{equation*}\n\nX in II\n\\begin{equation*}\n4=-7+4*2=-7+8=1\n\\end{equation*}\n\nL = {(2;1)}\n\n### Additionsverfahren (AV)\nZiel des Additionsverfahrens ist, den Koeffizienten einer Variablen zur Gegenzahl werden zu lassen,\nsodass bei einer Addition beider Gleichungen eine Gleichung mit nur einer Variablen übrig bleibt.\n\nI\n\\begin{equation*}\n3x+2y=12\n\\end{equation*}\n\nII\n\\begin{equation*}\n2x+3y=13\n\\end{equation*}\n\nI*2\n\\begin{equation*}\n6x+4y=24\n\\end{equation*}\n\nII*(-3)\n\\begin{equation*}\n-6x-9y=-39\n\\end{equation*}\n\nI+II\n\\begin{equation*}\n0x-5y=-15\n\\end{equation*}\n\n\\begin{equation*}\ny=3\n\\end{equation*}\n\nY in I\n\\begin{equation*}\n3x+2*3=12\n\\end{equation*}\n\n\\begin{equation*}\nx=2\n\\end{equation*}\n\nL = {(2;3)}", "_____no_output_____" ] ] ]
[ "markdown" ]
[ [ "markdown" ] ]
ecec1daea84d611c8afae05a94019f49eafdb5e1
138,060
ipynb
Jupyter Notebook
projects/Recurrent-Neural-Network/rnn-234450/RNN_project.ipynb
DMeechan/AI-Nanodegree
455f09b14965aed9d0a73fa507fdf35289834dd6
[ "MIT" ]
16
2017-12-28T05:53:33.000Z
2019-10-15T23:29:51.000Z
projects/Recurrent-Neural-Network/rnn-234450/RNN_project.ipynb
DMeechan/AI-Nanodegree
455f09b14965aed9d0a73fa507fdf35289834dd6
[ "MIT" ]
null
null
null
projects/Recurrent-Neural-Network/rnn-234450/RNN_project.ipynb
DMeechan/AI-Nanodegree
455f09b14965aed9d0a73fa507fdf35289834dd6
[ "MIT" ]
2
2019-07-13T06:25:09.000Z
2021-02-03T01:03:46.000Z
76.108049
39,422
0.775301
[ [ [ "# Artificial Intelligence Nanodegree\n## Recurrent Neural Network Projects\n\nWelcome to the Recurrent Neural Network Project in the Artificial Intelligence Nanodegree! In this notebook, some template code has already been provided for you, and you will need to implement additional functionality to successfully complete this project. You will not need to modify the included code beyond what is requested. Sections that begin with **'Implementation'** in the header indicate that the following block of code will require additional functionality which you must provide. Instructions will be provided for each section and the specifics of the implementation are marked in the code block with a 'TODO' statement. Please be sure to read the instructions carefully!\n\n>**Note:** Code and Markdown cells can be executed using the **Shift + Enter** keyboard shortcut. In addition, Markdown cells can be edited by typically double-clicking the cell to enter edit mode.", "_____no_output_____" ], [ "### Implementation TODOs in this notebook\n\nThis notebook contains two problems, cut into a variety of TODOs. Make sure to complete each section containing a TODO marker throughout the notebook. For convenience we provide links to each of these sections below.\n\n[TODO #1: Implement a function to window time series](#TODO_1)\n\n[TODO #2: Create a simple RNN model using keras to perform regression](#TODO_2)\n\n[TODO #3: Finish cleaning a large text corpus](#TODO_3)\n\n[TODO #4: Implement a function to window a large text corpus](#TODO_4)\n\n[TODO #5: Create a simple RNN model using keras to perform multiclass classification](#TODO_5)\n\n[TODO #6: Generate text using a fully trained RNN model and a variety of input sequences](#TODO_6)\n", "_____no_output_____" ], [ "# Problem 1: Perform time series prediction \n\nIn this project you will perform time series prediction using a Recurrent Neural Network regressor. In particular you will re-create the figure shown in the notes - where the stock price of Apple was forecasted (or predicted) 7 days in advance. In completing this exercise you will learn how to construct RNNs using Keras, which will also aid in completing the second project in this notebook.\n\nThe particular network architecture we will employ for our RNN is known as [Long Term Short Memory (LSTM)](https://en.wikipedia.org/wiki/Long_short-term_memory), which helps significantly avoid technical problems with optimization of RNNs. ", "_____no_output_____" ], [ "## 1.1 Getting started\n\nFirst we must load in our time series - a history of around 140 days of Apple's stock price. Then we need to perform a number of pre-processing steps to prepare it for use with an RNN model. First off, it is good practice to normalize time series - by normalizing its range. This helps us avoid serious numerical issues associated how common activation functions (like tanh) transform very large (positive or negative) numbers, as well as helping us to avoid related issues when computing derivatives.\n\nHere we normalize the series to lie in the range [0,1] [using this scikit function](http://scikit-learn.org/stable/modules/generated/sklearn.preprocessing.MinMaxScaler.html), but it is also commonplace to normalize by a series standard deviation.", "_____no_output_____" ] ], [ [ "### Load in necessary libraries for data input and normalization\n%matplotlib inline\nimport numpy as np\nimport matplotlib.pyplot as plt\n\n%load_ext autoreload\n%autoreload 2\n\nfrom my_answers import *\n\n%load_ext autoreload\n%autoreload 2\n\nfrom my_answers import *\n\n### load in and normalize the dataset\ndataset = np.loadtxt('datasets/normalized_apple_prices.csv')", "Using TensorFlow backend.\n" ] ], [ [ "Lets take a quick look at the (normalized) time series we'll be performing predictions on.", "_____no_output_____" ] ], [ [ "# lets take a look at our time series\nplt.plot(dataset)\nplt.xlabel('time period')\nplt.ylabel('normalized series value')", "_____no_output_____" ] ], [ [ "## 1.2 Cutting our time series into sequences\n\nRemember, our time series is a sequence of numbers that we can represent in general mathematically as \n\n$$s_{0},s_{1},s_{2},...,s_{P}$$\n\nwhere $s_{p}$ is the numerical value of the time series at time period $p$ and where $P$ is the total length of the series. In order to apply our RNN we treat the time series prediction problem as a regression problem, and so need to use a sliding window to construct a set of associated input/output pairs to regress on. This process is animated in the gif below.\n\n<img src=\"images/timeseries_windowing_training.gif\" width=600 height=600/>\n\nFor example - using a window of size T = 5 (as illustrated in the gif above) we produce a set of input/output pairs like the one shown in the table below\n\n$$\\begin{array}{c|c}\n\\text{Input} & \\text{Output}\\\\\n\\hline \\color{CornflowerBlue} {\\langle s_{1},s_{2},s_{3},s_{4},s_{5}\\rangle} & \\color{Goldenrod}{ s_{6}} \\\\\n\\ \\color{CornflowerBlue} {\\langle s_{2},s_{3},s_{4},s_{5},s_{6} \\rangle } & \\color{Goldenrod} {s_{7} } \\\\\n\\color{CornflowerBlue} {\\vdots} & \\color{Goldenrod} {\\vdots}\\\\\n\\color{CornflowerBlue} { \\langle s_{P-5},s_{P-4},s_{P-3},s_{P-2},s_{P-1} \\rangle } & \\color{Goldenrod} {s_{P}}\n\\end{array}$$\n\nNotice here that each input is a sequence (or vector) of length 5 (and in general has length equal to the window size T) while each corresponding output is a scalar value. Notice also how given a time series of length P and window size T = 5 as shown above, we created P - 5 input/output pairs. More generally, for a window size T we create P - T such pairs.", "_____no_output_____" ], [ "Now its time for you to window the input time series as described above! \n\n<a id='TODO_1'></a>\n\n**TODO:** Implement the function called **window_transform_series** in my_answers.py so that it runs a sliding window along the input series and creates associated input/output pairs. Note that this function should input a) the series and b) the window length, and return the input/output subsequences. Make sure to format returned input/output as generally shown in table above (where window_size = 5), and make sure your returned input is a numpy array.\n\n-----", "_____no_output_____" ], [ "You can test your function on the list of odd numbers given below", "_____no_output_____" ] ], [ [ "odd_nums = np.array([1,3,5,7,9,11,13])", "_____no_output_____" ] ], [ [ "Here is a hard-coded solution for odd_nums. You can compare its results with what you get from your **window_transform_series** implementation.", "_____no_output_____" ] ], [ [ "# run a window of size 2 over the odd number sequence and display the results\nwindow_size = 2\n\nX = []\nX.append(odd_nums[0:2])\nX.append(odd_nums[1:3])\nX.append(odd_nums[2:4])\nX.append(odd_nums[3:5])\nX.append(odd_nums[4:6])\n\ny = odd_nums[2:]\n\nX = np.asarray(X)\ny = np.asarray(y)\ny = np.reshape(y, (len(y),1)) #optional\n\nassert(type(X).__name__ == 'ndarray')\nassert(type(y).__name__ == 'ndarray')\nassert(X.shape == (5,2))\nassert(y.shape in [(5,1), (5,)])\n\n# print out input/output pairs --> here input = X, corresponding output = y\nprint ('--- the input X will look like ----')\nprint (X)\n\nprint ('--- the associated output y will look like ----')\nprint (y)", "--- the input X will look like ----\n[[ 1 3]\n [ 3 5]\n [ 5 7]\n [ 7 9]\n [ 9 11]]\n--- the associated output y will look like ----\n[[ 5]\n [ 7]\n [ 9]\n [11]\n [13]]\n" ] ], [ [ "Again - you can check that your completed **window_transform_series** function works correctly by trying it on the odd_nums sequence - you should get the above output.", "_____no_output_____" ] ], [ [ "### DONE: implement the function window_transform_series in the file my_answers.py\nfrom my_answers import window_transform_series\n# X,y = window_transform_series(series = odd_nums,window_size = 2)\n# print(X)\n# print(y)", "_____no_output_____" ] ], [ [ "With this function in place apply it to the series in the Python cell below. We use a window_size = 7 for these experiments.", "_____no_output_____" ] ], [ [ "# window the data using your windowing function\nwindow_size = 7\nX,y = window_transform_series(series = dataset,window_size = window_size)", "_____no_output_____" ] ], [ [ "## 1.3 Splitting into training and testing sets\n\nIn order to perform proper testing on our dataset we will lop off the last 1/3 of it for validation (or testing). This is that once we train our model we have something to test it on (like any regression problem!). This splitting into training/testing sets is done in the cell below.\n\nNote how here we are **not** splitting the dataset *randomly* as one typically would do when validating a regression model. This is because our input/output pairs *are related temporally*. We don't want to validate our model by training on a random subset of the series and then testing on another random subset, as this simulates the scenario that we receive new points *within the timeframe of our training set*. \n\nWe want to train on one solid chunk of the series (in our case, the first full 2/3 of it), and validate on a later chunk (the last 1/3) as this simulates how we would predict *future* values of a time series.", "_____no_output_____" ] ], [ [ "# split our dataset into training / testing sets\ntrain_test_split = int(np.ceil(2*len(y)/float(3))) # set the split point\n\n# partition the training set\nX_train = X[:train_test_split,:]\ny_train = y[:train_test_split]\n\n# keep the last chunk for testing\nX_test = X[train_test_split:,:]\ny_test = y[train_test_split:]\n\n# X_test = np.delete(X_test, 0, 0)\n\n\n# NOTE: to use keras's RNN LSTM module our input must be reshaped to [samples, window size, stepsize] \nX_train = np.asarray(np.reshape(X_train, (X_train.shape[0], window_size, 1)))\nX_test = np.asarray(np.reshape(X_test, (X_test.shape[0], window_size, 1)))\n\nif len(X_test) != len(y_test):\n print('Warning: X_test is not the same size as y_test:', len(X_test), \"vs\", len(y_test))", "_____no_output_____" ] ], [ [ "<a id='TODO_2'></a>\n\n## 1.4 Build and run an RNN regression model\n\nHaving created input/output pairs out of our time series and cut this into training/testing sets, we can now begin setting up our RNN. We use Keras to quickly build a two hidden layer RNN of the following specifications\n\n- layer 1 uses an LSTM module with 5 hidden units (note here the input_shape = (window_size,1))\n- layer 2 uses a fully connected module with one unit\n- the 'mean_squared_error' loss should be used (remember: we are performing regression here)\n\nThis can be constructed using just a few lines - see e.g., the [general Keras documentation](https://keras.io/getting-started/sequential-model-guide/) and the [LSTM documentation in particular](https://keras.io/layers/recurrent/) for examples of how to quickly use Keras to build neural network models. Make sure you are initializing your optimizer given the [keras-recommended approach for RNNs](https://keras.io/optimizers/) \n\n(given in the cell below). (remember to copy your completed function into the script *my_answers.py* function titled *build_part1_RNN* before submitting your project)", "_____no_output_____" ] ], [ [ "### DONE: create required RNN model\n# import keras network libraries\nfrom keras.models import Sequential\nfrom keras.layers import Dense\nfrom keras.layers import LSTM\nimport keras\n\n# given - fix random seed - so we can all reproduce the same results on our default time series\nnp.random.seed(0)\n\n# DONE: implement build_part1_RNN in my_answers.py\nfrom my_answers import build_part1_RNN\nmodel = build_part1_RNN(window_size)\n\n# build model using keras documentation recommended optimizer initialization\noptimizer = keras.optimizers.RMSprop(lr=0.001, rho=0.9, epsilon=1e-08, decay=0.0)\n\n# compile the model\nmodel.compile(loss='mean_squared_error', optimizer=optimizer)", "_____no_output_____" ] ], [ [ "With your model built you can now fit the model by activating the cell below! Note: the number of epochs (np_epochs) and batch_size are preset (so we can all produce the same results). You can choose to toggle the verbose parameter - which gives you regular updates on the progress of the algorithm - on and off by setting it to 1 or 0 respectively.", "_____no_output_____" ] ], [ [ "# run your model!\nmodel.fit(X_train, y_train, epochs=1000, batch_size=256, verbose=0)", "_____no_output_____" ] ], [ [ "## 1.5 Checking model performance\n\nWith your model fit we can now make predictions on both our training and testing sets.", "_____no_output_____" ] ], [ [ "# generate predictions for training\ntrain_predict = model.predict(X_train)\ntest_predict = model.predict(X_test)", "_____no_output_____" ] ], [ [ "In the next cell we compute training and testing errors using our trained model - you should be able to achieve at least\n\n*training_error* < 0.02\n\nand \n\n*testing_error* < 0.02\n\nwith your fully trained model. \n\nIf either or both of your accuracies are larger than 0.02 re-train your model - increasing the number of epochs you take (a maximum of around 1,000 should do the job) and/or adjusting your batch_size.", "_____no_output_____" ] ], [ [ "# print out training and testing errors\ntraining_error = model.evaluate(X_train, y_train, verbose=0)\nprint('training error = ' + str(training_error))\n\ntesting_error = model.evaluate(X_test, y_test, verbose=0)\nprint('testing error = ' + str(testing_error))", "training error = 0.0158862119371\ntesting error = 0.0146829202381\n" ] ], [ [ "Activating the next cell plots the original data, as well as both predictions on the training and testing sets. ", "_____no_output_____" ] ], [ [ "### Plot everything - the original series as well as predictions on training and testing sets\nimport matplotlib.pyplot as plt\n%matplotlib inline\n\n# plot original series\nplt.plot(dataset,color = 'k')\n\n# plot training set prediction\nsplit_pt = train_test_split + window_size \nplt.plot(np.arange(window_size,split_pt,1),train_predict,color = 'b')\n\n# plot testing set prediction\nplt.plot(np.arange(split_pt,split_pt + len(test_predict),1),test_predict,color = 'r')\n\n# pretty up graph\nplt.xlabel('day')\nplt.ylabel('(normalized) price of Apple stock')\nplt.legend(['original series','training fit','testing fit'],loc='center left', bbox_to_anchor=(1, 0.5))\nplt.show()", "_____no_output_____" ] ], [ [ "**Note:** you can try out any time series for this exercise! If you would like to try another see e.g., [this site containing thousands of time series](https://datamarket.com/data/list/?q=provider%3Atsdl) and pick another one!", "_____no_output_____" ], [ "# Problem 2: Create a sequence generator", "_____no_output_____" ], [ "## 2.1 Getting started\n\nIn this project you will implement a popular Recurrent Neural Network (RNN) architecture to create an English language sequence generator capable of building semi-coherent English sentences from scratch by building them up character-by-character. This will require a substantial amount amount of parameter tuning on a large training corpus (at least 100,000 characters long). In particular for this project we will be using a complete version of Sir Arthur Conan Doyle's classic book The Adventures of Sherlock Holmes.\n\nHow can we train a machine learning model to generate text automatically, character-by-character? *By showing the model many training examples so it can learn a pattern between input and output.* With this type of text generation each input is a string of valid characters like this one\n\n*dogs are grea*\n\nwhile the corresponding output is the next character in the sentence - which here is 't' (since the complete sentence is 'dogs are great'). We need to show a model many such examples in order for it to make reasonable predictions.\n\n**Fun note:** For those interested in how text generation is being used check out some of the following fun resources:\n\n- [Generate wacky sentences](http://www.cs.toronto.edu/~ilya/rnn.html) with this academic RNN text generator\n\n- Various twitter bots that tweet automatically generated text like[this one](http://tweet-generator-alex.herokuapp.com/).\n\n- the [NanoGenMo](https://github.com/NaNoGenMo/2016) annual contest to automatically produce a 50,000+ novel automatically\n\n- [Robot Shakespeare](https://github.com/genekogan/RobotShakespeare) a text generator that automatically produces Shakespear-esk sentences", "_____no_output_____" ], [ "## 2.2 Preprocessing a text dataset\n\nOur first task is to get a large text corpus for use in training, and on it we perform a several light pre-processing tasks. The default corpus we will use is the classic book Sherlock Holmes, but you can use a variety of others as well - so long as they are fairly large (around 100,000 characters or more). ", "_____no_output_____" ] ], [ [ "# read in the text, transforming everything to lower case\ntext = open('datasets/holmes.txt').read().lower()\nprint('our original text has ' + str(len(text)) + ' characters')", "our original text has 581864 characters\n" ] ], [ [ "Next, lets examine a bit of the raw text. Because we are interested in creating sentences of English words automatically by building up each word character-by-character, we only want to train on valid English words. In other words - we need to remove all of the other characters that are not part of English words.", "_____no_output_____" ] ], [ [ "### print out the first 1000 characters of the raw text to get a sense of what we need to throw out\ntext[:2000]", "_____no_output_____" ] ], [ [ "Wow - there's a lot of junk here (i.e., weird uncommon character combinations - as this first character chunk contains the title and author page, as well as table of contents)! To keep things simple, we want to train our RNN on a large chunk of more typical English sentences - we don't want it to start thinking non-english words or strange characters are valid! - so lets clean up the data a bit.\n\nFirst, since the dataset is so large and the first few hundred characters contain a lot of junk, lets cut it out. Lets also find-and-replace those newline tags with empty spaces.", "_____no_output_____" ] ], [ [ "### find and replace '\\n' and '\\r' symbols - replacing them \ntext = text[1302:]\ntext = text.replace('\\n',' ') # replacing '\\n' with '' simply removes the sequence\ntext = text.replace('\\r',' ')", "_____no_output_____" ] ], [ [ "Lets see how the first 1000 characters of our text looks now!", "_____no_output_____" ] ], [ [ "### print out the first 1000 characters of the raw text to get a sense of what we need to throw out\ntext[:1000]", "_____no_output_____" ] ], [ [ "<a id='TODO_3'></a>\n\n#### TODO: finish cleaning the text\n\nLets make sure we haven't left any other atypical characters (commas, periods, etc., are ok) lurking around in the depths of the text. You can do this by enumerating all the text's unique characters, examining them, and then replacing any unwanted characters with empty spaces! Once we find all of the text's unique characters, we can remove all of the atypical ones in the next cell. Note: don't remove the punctuation marks given in my_answers.py.", "_____no_output_____" ] ], [ [ "### DONE: implement cleaned_text in my_answers.py\nfrom my_answers import cleaned_text\n\ntext = cleaned_text(text)\n\n# shorten any extra dead space created above\ntext = text.replace(' ',' ')", "_____no_output_____" ] ], [ [ "With your chosen characters removed print out the first few hundred lines again just to double check that everything looks good.", "_____no_output_____" ] ], [ [ "### print out the first 2000 characters of the raw text to get a sense of what we need to throw out\ntext[:2000]", "_____no_output_____" ] ], [ [ "Now that we have thrown out a good number of non-English characters/character sequences lets print out some statistics about the dataset - including number of total characters and number of unique characters.", "_____no_output_____" ] ], [ [ "# count the number of unique characters in the text\nchars = sorted(list(set(text)))\n\n# print some of the text, as well as statistics\nprint (\"this corpus has \" + str(len(text)) + \" total number of characters\")\nprint (\"this corpus has \" + str(len(chars)) + \" unique characters\")\nprint(chars)", "this corpus has 569169 total number of characters\nthis corpus has 33 unique characters\n[' ', '!', ',', '.', ':', ';', '?', 'a', 'b', 'c', 'd', 'e', 'f', 'g', 'h', 'i', 'j', 'k', 'l', 'm', 'n', 'o', 'p', 'q', 'r', 's', 't', 'u', 'v', 'w', 'x', 'y', 'z']\n" ] ], [ [ "## 2.3 Cutting data into input/output pairs\n\nNow that we have our text all cleaned up, how can we use it to train a model to generate sentences automatically? First we need to train a machine learning model - and in order to do that we need a set of input/output pairs for a model to train on. How can we create a set of input/output pairs from our text to train on?\n\nRemember in part 1 of this notebook how we used a sliding window to extract input/output pairs from a time series? We do the same thing here! We slide a window of length $T$ along our giant text corpus - everything in the window becomes one input while the character following becomes its corresponding output. This process of extracting input/output pairs is illustrated in the gif below on a small example text using a window size of T = 5.\n\n<img src=\"images/text_windowing_training.gif\" width=400 height=400/>\n\nNotice one aspect of the sliding window in this gif that does not mirror the analogous gif for time series shown in part 1 of the notebook - we do not need to slide the window along one character at a time but can move by a fixed step size $M$ greater than 1 (in the gif indeed $M = 1$). This is done with large input texts (like ours which has over 500,000 characters!) when sliding the window along one character at a time we would create far too many input/output pairs to be able to reasonably compute with.\n\nMore formally lets denote our text corpus - which is one long string of characters - as follows\n\n$$s_{0},s_{1},s_{2},...,s_{P}$$\n\nwhere $P$ is the length of the text (again for our text $P \\approx 500,000!$). Sliding a window of size T = 5 with a step length of M = 1 (these are the parameters shown in the gif above) over this sequence produces the following list of input/output pairs\n\n\n$$\\begin{array}{c|c}\n\\text{Input} & \\text{Output}\\\\\n\\hline \\color{CornflowerBlue} {\\langle s_{1},s_{2},s_{3},s_{4},s_{5}\\rangle} & \\color{Goldenrod}{ s_{6}} \\\\\n\\ \\color{CornflowerBlue} {\\langle s_{2},s_{3},s_{4},s_{5},s_{6} \\rangle } & \\color{Goldenrod} {s_{7} } \\\\\n\\color{CornflowerBlue} {\\vdots} & \\color{Goldenrod} {\\vdots}\\\\\n\\color{CornflowerBlue} { \\langle s_{P-5},s_{P-4},s_{P-3},s_{P-2},s_{P-1} \\rangle } & \\color{Goldenrod} {s_{P}}\n\\end{array}$$\n\nNotice here that each input is a sequence (or vector) of 5 characters (and in general has length equal to the window size T) while each corresponding output is a single character. We created around P total number of input/output pairs (for general step size M we create around ceil(P/M) pairs).", "_____no_output_____" ], [ "<a id='TODO_4'></a>\n\nNow its time for you to window the input time series as described above! \n\n**TODO:** Create a function that runs a sliding window along the input text and creates associated input/output pairs. A skeleton function has been provided for you. Note that this function should input a) the text b) the window size and c) the step size, and return the input/output sequences. Note: the return items should be *lists* - not numpy arrays.\n\n(remember to copy your completed function into the script *my_answers.py* function titled *window_transform_text* before submitting your project)", "_____no_output_____" ] ], [ [ "### DONE: implement window_transform_series in my_answers.py\nfrom my_answers import window_transform_text", "_____no_output_____" ] ], [ [ "With our function complete we can now use it to produce input/output pairs! We employ the function in the next cell, where the window_size = 50 and step_size = 5.", "_____no_output_____" ] ], [ [ "# run your text window-ing function \nwindow_size = 100\nstep_size = 5\ninputs, outputs = window_transform_text(text,window_size,step_size)", "_____no_output_____" ], [ "# from my_answers import window_transform_text\n# nums = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13]\n# print(nums)\n# a,b = window_transform_text(text = nums,window_size = 5, step_size=2)\n# print(a)\n# print(b)", "_____no_output_____" ] ], [ [ "Lets print out a few input/output pairs to verify that we have made the right sort of stuff!", "_____no_output_____" ] ], [ [ "# print out a few of the input/output pairs to verify that we've made the right kind of stuff to learn from\nprint('input =', inputs[2])\nprint('output =', outputs[2])\nprint('--------------')\nprint('input =', inputs[100])\nprint('output =', outputs[100])", "input = e eclipses and predominates the whole of her sex. it was not that he felt any emotion akin to love f\noutput = o\n--------------\ninput = erexcellent for drawing the veil from mens motives and actions. but for the trained reasoner to admi\noutput = t\n" ] ], [ [ "Looks good!", "_____no_output_____" ], [ "## 2.4 Wait, what kind of problem is text generation again?\n\nIn part 1 of this notebook we used the same pre-processing technique - the sliding window - to produce a set of training input/output pairs to tackle the problem of time series prediction *by treating the problem as one of regression*. So what sort of problem do we have here now, with text generation? Well, the time series prediction was a regression problem because the output (one value of the time series) was a continuous value. Here - for character-by-character text generation - each output is a *single character*. This isn't a continuous value - but a distinct class - therefore **character-by-character text generation is a classification problem**. \n\nHow many classes are there in the data? Well, the number of classes is equal to the number of unique characters we have to predict! How many of those were there in our dataset again? Lets print out the value again.", "_____no_output_____" ] ], [ [ "# print out the number of unique characters in the dataset\nchars = sorted(list(set(text)))\nprint (\"this corpus has \" + str(len(chars)) + \" unique characters\")\nprint ('and these characters are ')\nprint (chars)", "this corpus has 33 unique characters\nand these characters are \n[' ', '!', ',', '.', ':', ';', '?', 'a', 'b', 'c', 'd', 'e', 'f', 'g', 'h', 'i', 'j', 'k', 'l', 'm', 'n', 'o', 'p', 'q', 'r', 's', 't', 'u', 'v', 'w', 'x', 'y', 'z']\n" ] ], [ [ "Rockin' - so we have a multiclass classification problem on our hands!", "_____no_output_____" ], [ "## 2.5 One-hot encoding characters\n\nThe last issue we have to deal with is representing our text data as numerical data so that we can use it as an input to a neural network. One of the conceptually simplest ways of doing this is via a 'one-hot encoding' scheme. Here's how it works.\n\nWe transform each character in our inputs/outputs into a vector with length equal to the number of unique characters in our text. This vector is all zeros except one location where we place a 1 - and this location is unique to each character type. e.g., we transform 'a', 'b', and 'c' as follows\n\n$$a\\longleftarrow\\left[\\begin{array}{c}\n1\\\\\n0\\\\\n0\\\\\n\\vdots\\\\\n0\\\\\n0\n\\end{array}\\right]\\,\\,\\,\\,\\,\\,\\,b\\longleftarrow\\left[\\begin{array}{c}\n0\\\\\n1\\\\\n0\\\\\n\\vdots\\\\\n0\\\\\n0\n\\end{array}\\right]\\,\\,\\,\\,\\,c\\longleftarrow\\left[\\begin{array}{c}\n0\\\\\n0\\\\\n1\\\\\n\\vdots\\\\\n0\\\\\n0 \n\\end{array}\\right]\\cdots$$\n\nwhere each vector has 32 entries (or in general: number of entries = number of unique characters in text).", "_____no_output_____" ], [ "The first practical step towards doing this one-hot encoding is to form a dictionary mapping each unique character to a unique integer, and one dictionary to do the reverse mapping. We can then use these dictionaries to quickly make our one-hot encodings, as well as re-translate (from integers to characters) the results of our trained RNN classification model.", "_____no_output_____" ] ], [ [ "# this dictionary is a function mapping each unique character to a unique integer\nchars_to_indices = dict((c, i) for i, c in enumerate(chars)) # map each unique character to unique integer\n\n# this dictionary is a function mapping each unique integer back to a unique character\nindices_to_chars = dict((i, c) for i, c in enumerate(chars)) # map each unique integer back to unique character", "_____no_output_____" ] ], [ [ "Now we can transform our input/output pairs - consisting of characters - to equivalent input/output pairs made up of one-hot encoded vectors. In the next cell we provide a function for doing just this: it takes in the raw character input/outputs and returns their numerical versions. In particular the numerical input is given as $\\bf{X}$, and numerical output is given as the $\\bf{y}$", "_____no_output_____" ] ], [ [ "# transform character-based input/output into equivalent numerical versions\ndef encode_io_pairs(text,window_size,step_size):\n # number of unique chars\n chars = sorted(list(set(text)))\n num_chars = len(chars)\n \n # cut up text into character input/output pairs\n inputs, outputs = window_transform_text(text,window_size,step_size)\n \n # create empty vessels for one-hot encoded input/output\n X = np.zeros((len(inputs), window_size, num_chars), dtype=np.bool)\n y = np.zeros((len(inputs), num_chars), dtype=np.bool)\n \n # loop over inputs/outputs and transform and store in X/y\n for i, sentence in enumerate(inputs):\n for t, char in enumerate(sentence):\n X[i, t, chars_to_indices[char]] = 1\n current_output = outputs[i]\n current_char = chars_to_indices[current_output]\n y[i, current_char] = 1\n \n return X,y", "_____no_output_____" ] ], [ [ "Now run the one-hot encoding function by activating the cell below and transform our input/output pairs!", "_____no_output_____" ] ], [ [ "# use your function\nwindow_size = 100\nstep_size = 5\nX,y = encode_io_pairs(text,window_size,step_size)", "_____no_output_____" ] ], [ [ "<a id='TODO_5'></a>\n\n## 2.6 Setting up our RNN\n\nWith our dataset loaded and the input/output pairs extracted / transformed we can now begin setting up our RNN for training. Again we will use Keras to quickly build a single hidden layer RNN - where our hidden layer consists of LSTM modules.\n\nTime to get to work: build a 3 layer RNN model of the following specification\n\n- layer 1 should be an LSTM module with 200 hidden units --> note this should have input_shape = (window_size,len(chars)) where len(chars) = number of unique characters in your cleaned text\n- layer 2 should be a linear module, fully connected, with len(chars) hidden units --> where len(chars) = number of unique characters in your cleaned text\n- layer 3 should be a softmax activation ( since we are solving a *multiclass classification*)\n- Use the **categorical_crossentropy** loss \n\nThis network can be constructed using just a few lines - as with the RNN network you made in part 1 of this notebook. See e.g., the [general Keras documentation](https://keras.io/getting-started/sequential-model-guide/) and the [LSTM documentation in particular](https://keras.io/layers/recurrent/) for examples of how to quickly use Keras to build neural network models.", "_____no_output_____" ] ], [ [ "### necessary functions from the keras library\nfrom keras.models import Sequential\nfrom keras.layers import Dense, Activation, LSTM\nfrom keras.optimizers import RMSprop\nfrom keras.utils.data_utils import get_file\nimport keras\nimport random\n\n# DONE implement build_part2_RNN in my_answers.py\nfrom my_answers import build_part2_RNN\n\nmodel = build_part2_RNN(window_size, len(chars))\n\n# initialize optimizer\noptimizer = keras.optimizers.RMSprop(lr=0.001, rho=0.9, epsilon=1e-08, decay=0.0)\n\n# compile model --> make sure initialized optimizer and callbacks - as defined above - are used\nmodel.compile(loss='categorical_crossentropy', optimizer=optimizer)", "_____no_output_____" ] ], [ [ "## 2.7 Training our RNN model for text generation\n\nWith our RNN setup we can now train it! Lets begin by trying it out on a small subset of the larger version. In the next cell we take the first 10,000 input/output pairs from our training database to learn on.", "_____no_output_____" ] ], [ [ "# a small subset of our input/output pairs\nXsmall = X[:10000,:,:]\nysmall = y[:10000,:]", "_____no_output_____" ] ], [ [ "Now lets fit our model!", "_____no_output_____" ] ], [ [ "# train the model\nmodel.fit(Xsmall, ysmall, batch_size=500, epochs=40,verbose = 1)\n\n# save weights\nmodel.save_weights('model_weights/best_RNN_small_textdata_weights.hdf5')", "_____no_output_____" ] ], [ [ "How do we make a given number of predictions (characters) based on this fitted model? \n\nFirst we predict the next character after following any chunk of characters in the text of length equal to our chosen window size. Then we remove the first character in our input sequence and tack our prediction onto the end. This gives us a slightly changed sequence of inputs that still has length equal to the size of our window. We then feed in this updated input sequence into the model to predict the another character. Together then we have two predicted characters following our original input sequence. Repeating this process N times gives us N predicted characters.\n\nIn the next Python cell we provide you with a completed function that does just this - it makes predictions when given a) a trained RNN model, b) a subset of (window_size) characters from the text, and c) a number of characters to predict (to follow our input subset).", "_____no_output_____" ] ], [ [ "# function that uses trained model to predict a desired number of future characters\ndef predict_next_chars(model,input_chars,num_to_predict): \n # create output\n predicted_chars = ''\n for i in range(num_to_predict):\n # convert this round's predicted characters to numerical input \n x_test = np.zeros((1, window_size, len(chars)))\n for t, char in enumerate(input_chars):\n x_test[0, t, chars_to_indices[char]] = 1.\n\n # make this round's prediction\n test_predict = model.predict(x_test,verbose = 0)[0]\n\n # translate numerical prediction back to characters\n r = np.argmax(test_predict) # predict class of each test input\n d = indices_to_chars[r] \n\n # update predicted_chars and input\n predicted_chars+=d\n input_chars+=d\n input_chars = input_chars[1:]\n return predicted_chars", "_____no_output_____" ], [ "# import os\n# os.remove('model_weights/best_RNN_small_textdata_weights.hdf5')\n# f = open('model_weights/best_RNN_small_textdata_weights.hdf5', 'w')\n# f.close()\n# model.save_weights('model_weights/best_RNN_small_textdata_weights.hdf5')", "_____no_output_____" ] ], [ [ "<a id='TODO_6'></a>\n\nWith your trained model try a few subsets of the complete text as input - note the length of each must be exactly equal to the window size. For each subset use the function above to predict the next 100 characters that follow each input.", "_____no_output_____" ] ], [ [ "# DONE: choose an input sequence and use the prediction function in the previous Python cell to predict 100 characters following it\n# get an appropriately sized chunk of characters from the text\nstart_inds = [0, 1, 2, 3, 4, 5]\n\n# load in weights\nmodel.load_weights('model_weights/best_RNN_small_textdata_weights.hdf5')\nfor s in start_inds:\n start_index = s\n input_chars = text[start_index: start_index + window_size]\n\n # use the prediction function\n predict_input = predict_next_chars(model,input_chars,num_to_predict = 100)\n\n # print out input characters\n print('------------------')\n input_line = 'input chars = ' + '\\n' + input_chars + '\"' + '\\n'\n print(input_line)\n\n # print out predicted characters\n line = 'predicted chars = ' + '\\n' + predict_input + '\"' + '\\n'\n print(line)", "------------------\ninput chars = \nis eyes she eclipses and predominates the whole of her sex it was not that he felt any emotion akin \"\n\npredicted chars = \nthe rast in i shand your se the erade a dore wioh hur wat the werl what the enter wat the kend out a\"\n\n------------------\ninput chars = \ns eyes she eclipses and predominates the whole of her sex it was not that he felt any emotion akin t\"\n\npredicted chars = \nhe rast in i shand your se the erade a dore wioh hur wat the werl what the enter wat the kend out a \"\n\n------------------\ninput chars = \n eyes she eclipses and predominates the whole of her sex it was not that he felt any emotion akin to\"\n\npredicted chars = \n eat a cous and she sing thand y uren be in not and recone forn he parked werthen i surmand and sere\"\n\n------------------\ninput chars = \neyes she eclipses and predominates the whole of her sex it was not that he felt any emotion akin to \"\n\npredicted chars = \neat a cous and she sing thand y uren be in not and recone forn he parked werthen i surmand and seres\"\n\n------------------\ninput chars = \nyes she eclipses and predominates the whole of her sex it was not that he felt any emotion akin to l\"\n\npredicted chars = \nath beet ant in in the forn of mas of the ofrthar and i sound he she wald i mald you tay westly thin\"\n\n------------------\ninput chars = \nes she eclipses and predominates the whole of her sex it was not that he felt any emotion akin to lo\"\n\npredicted chars = \nve wist out ave the your and hersend how selleeted and wis leat he west cuand it hes latt of hall we\"\n\n" ] ], [ [ "This looks ok, but not great. Now lets try the same experiment with a larger chunk of the data - with the first 100,000 input/output pairs. \n\nTuning RNNs for a typical character dataset like the one we will use here is a computationally intensive endeavour and thus timely on a typical CPU. Using a reasonably sized cloud-based GPU can speed up training by a factor of 10. Also because of the long training time it is highly recommended that you carefully write the output of each step of your process to file. This is so that all of your results are saved even if you close the web browser you're working out of, as the processes will continue processing in the background but variables/output in the notebook system will not update when you open it again.\n\nIn the next cell we show you how to create a text file in Python and record data to it. This sort of setup can be used to record your final predictions.", "_____no_output_____" ] ], [ [ "### A simple way to write output to file\nf = open('my_test_output.txt', 'w') # create an output file to write too\nf.write('this is only a test ' + '\\n') # print some output text\nx = 2\nf.write('the value of x is ' + str(x) + '\\n') # record a variable value\nf.close() \n\n# print out the contents of my_test_output.txt\nf = open('my_test_output.txt', 'r') # create an output file to write too\nf.read()", "_____no_output_____" ] ], [ [ "With this recording devices we can now more safely perform experiments on larger portions of the text. In the next cell we will use the first 100,000 input/output pairs to train our RNN model.", "_____no_output_____" ], [ "First we fit our model to the dataset, then generate text using the trained model in precisely the same generation method applied before on the small dataset.\n\n**Note:** your generated words should be - by and large - more realistic than with the small dataset, but you won't be able to generate perfect English sentences even with this amount of data. A rule of thumb: your model is working well if you generate sentences that largely contain real English words.", "_____no_output_____" ] ], [ [ "# a small subset of our input/output pairs\nXlarge = X[:100000,:,:]\nylarge = y[:100000,:]\n\n# DONE: fit to our larger dataset\nmodel.fit(Xlarge, ylarge, batch_size=500, epochs=30, verbose=1)\n\n# save weights\nmodel.save_weights('model_weights/best_RNN_large_textdata_weights.hdf5')", "Epoch 1/30\n100000/100000 [==============================] - 45s - loss: 1.9659 \nEpoch 2/30\n100000/100000 [==============================] - 45s - loss: 1.8361 \nEpoch 3/30\n100000/100000 [==============================] - 45s - loss: 1.7759 \nEpoch 4/30\n100000/100000 [==============================] - 45s - loss: 1.7261 \nEpoch 5/30\n100000/100000 [==============================] - 45s - loss: 1.6827 \nEpoch 6/30\n100000/100000 [==============================] - 45s - loss: 1.6443 \nEpoch 7/30\n100000/100000 [==============================] - 45s - loss: 1.6072 \nEpoch 8/30\n100000/100000 [==============================] - 45s - loss: 1.5736 \nEpoch 9/30\n100000/100000 [==============================] - 45s - loss: 1.5099 \nEpoch 11/30\n100000/100000 [==============================] - 45s - loss: 1.4795 \nEpoch 12/30\n100000/100000 [==============================] - 45s - loss: 1.4496 \nEpoch 13/30\n100000/100000 [==============================] - 45s - loss: 1.4202 \nEpoch 14/30\n100000/100000 [==============================] - 45s - loss: 1.3912 \nEpoch 15/30\n100000/100000 [==============================] - 45s - loss: 1.3615 \nEpoch 16/30\n100000/100000 [==============================] - 45s - loss: 1.3341 \nEpoch 17/30\n100000/100000 [==============================] - 45s - loss: 1.3052 \nEpoch 18/30\n100000/100000 [==============================] - 45s - loss: 1.2783 \nEpoch 19/30\n100000/100000 [==============================] - 45s - loss: 1.2501 \nEpoch 20/30\n100000/100000 [==============================] - 45s - loss: 1.2210 \nEpoch 21/30\n100000/100000 [==============================] - 45s - loss: 1.1942 \nEpoch 22/30\n100000/100000 [==============================] - 45s - loss: 1.1659 \nEpoch 23/30\n100000/100000 [==============================] - 45s - loss: 1.1374 \nEpoch 24/30\n100000/100000 [==============================] - 45s - loss: 1.1111 \nEpoch 25/30\n100000/100000 [==============================] - 45s - loss: 1.0837 \nEpoch 26/30\n100000/100000 [==============================] - 45s - loss: 1.0580 \nEpoch 27/30\n100000/100000 [==============================] - 45s - loss: 1.0310 \nEpoch 28/30\n100000/100000 [==============================] - 45s - loss: 1.0057 \nEpoch 29/30\n100000/100000 [==============================] - 45s - loss: 0.9826 \nEpoch 30/30\n100000/100000 [==============================] - 45s - loss: 0.9573 \n" ], [ "# DONE: choose an input sequence and use the prediction function in the previous Python cell to predict 100 characters following it\n# get an appropriately sized chunk of characters from the text\nstart_inds = [0, 2, 10, 12, 20, 22, 1000, 1002]\n\n# save output\nf = open('text_gen_output/RNN_large_textdata_output.txt', 'w') # create an output file to write too\n\n# load weights\nmodel.load_weights('model_weights/best_RNN_large_textdata_weights.hdf5')\nfor s in start_inds:\n start_index = s\n input_chars = text[start_index: start_index + window_size]\n\n # use the prediction function\n predict_input = predict_next_chars(model,input_chars,num_to_predict = 100)\n\n # print out input characters\n line = '-------------------' + '\\n'\n print(line)\n f.write(line)\n\n input_line = 'input chars = ' + '\\n' + input_chars + '\"' + '\\n'\n print(input_line)\n f.write(input_line)\n\n # print out predicted characters\n predict_line = 'predicted chars = ' + '\\n' + predict_input + '\"' + '\\n'\n print(predict_line)\n f.write(predict_line)\nf.close()", "-------------------\n\ninput chars = \nis eyes she eclipses and predominates the whole of her sex it was not that he felt any emotion akin \"\n\npredicted chars = \nthen i have are morning beatul solly of the business he was a small dood and ras elles he sherked hi\"\n\n-------------------\n\ninput chars = \n eyes she eclipses and predominates the whole of her sex it was not that he felt any emotion akin to\"\n\npredicted chars = \n and heard it is a shill contrceet when i am all dishe for arl to be the door what a come is the loo\"\n\n-------------------\n\ninput chars = \ne eclipses and predominates the whole of her sex it was not that he felt any emotion akin to love fo\"\n\npredicted chars = \nr a from assion was the sell and of the roon which he had not heard but i could not bes it the stree\"\n\n-------------------\n\ninput chars = \neclipses and predominates the whole of her sex it was not that he felt any emotion akin to love for \"\n\npredicted chars = \na from assion was the sell and of the roon which he had not heard but i could not bes it the street \"\n\n-------------------\n\ninput chars = \n and predominates the whole of her sex it was not that he felt any emotion akin to love for irene ad\"\n\npredicted chars = \nlerest into the last he was maded by and she of carder instrance it is an ins and he as it is in pre\"\n\n-------------------\n\ninput chars = \nnd predominates the whole of her sex it was not that he felt any emotion akin to love for irene adle\"\n\npredicted chars = \nrest into the last he was maded by and she of carder instrance it is an ins and he as it is in preca\"\n\n-------------------\n\ninput chars = \nstionable memory i had seen little of holmes lately my marriage had drifted us away from each other \"\n\npredicted chars = \nmorther and i have no dound me morner a stall man who should he as in the string of it was the paste\"\n\n-------------------\n\ninput chars = \nionable memory i had seen little of holmes lately my marriage had drifted us away from each other my\"\n\npredicted chars = \n hander them is a recoly to be offer there ase the onder of i have and i have no doubt that i have n\"\n\n" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code" ]
[ [ "markdown", "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code", "code" ] ]
ecec372c19e8f41f58307e621b4ad6694c78747b
6,193
ipynb
Jupyter Notebook
content/lessons/01-Intro/SmallGroup-Intro.ipynb
IST256/learn-python
b2373b5cf596a318976a816e3102cc704d6b9b57
[ "MIT" ]
14
2017-02-23T21:00:46.000Z
2021-03-19T09:29:40.000Z
content/lessons/01-Intro/SmallGroup-Intro.ipynb
IST256/learn-python
b2373b5cf596a318976a816e3102cc704d6b9b57
[ "MIT" ]
null
null
null
content/lessons/01-Intro/SmallGroup-Intro.ipynb
IST256/learn-python
b2373b5cf596a318976a816e3102cc704d6b9b57
[ "MIT" ]
38
2017-02-03T13:49:19.000Z
2021-08-15T16:47:56.000Z
27.043668
298
0.590182
[ [ [ "# Now You Code In Class: Gathering Student Data\n\n\n## The Problem\n\nWrite a program which gathers student data such as name, email, gpa and major then prints out a student profile card with a layout EXACTLY like this:\n\n### Sample Output:\n\n```\n===== STUDENT PROFILE =====\nNAME : Michael Fudge\nEMAIL : [email protected]\nGPA : 4.0\nMAJOR : Information Studies\n```\n\nOf course your actual output of name, email, GPA and major, will vary based on the inputs, which is the point of programming in the first place! ", "_____no_output_____" ], [ "## Problem Analysis: \"Thinking About The Problem\"\n\nThe first thing we should do before writing ANY CODE is have a good understanding of the problem. After all you can't solve a problem in code unless you understand it first! \n\nAs part of your **practice** routine in this course, you will formalize your **understanding** of the problem by articulating:\n\n1. the data inputs into the program\n2. the information output from the program\n3. the algorithm - the high level steps from input to output. \n\nAgain, the purpose of this is to communicate your understanding of the problem **to your professor and more importantly yourself.** \n\nOf course, defining a problem this way is typically not an easy thing to do, which is why we will learn several **problem-solving approaches** in this class.\n", "_____no_output_____" ], [ "## Problem Solving Approach: Work Backwards \n\nThe **work-backwards** approach to problem solving is often used in science and mathematics. It works well when we have an idea of the program's output but are unclear how to start the program. It can be used in this case because we expect out of the program itself is clearly understood. \n\nThe approach is as follows:\n\n1. start with the output\n2. figure out which variables are used in the output (what part of the output changes versus stays the same - the changing parts use variables)\n3. keep working backwards until the variables become inputs.", "_____no_output_____" ], [ "### PROMPT 1: Start with the output\n\nWrite Python code to print the example output under the problem section exactly as shown:", "_____no_output_____" ] ], [ [ "# Code here\n", "_____no_output_____" ] ], [ [ "### PROMPT 2: Which aspects of the output should change?\n\nWhich aspects of the output should change and therefore be written using variables? Write a Python assignment statement for each variable assigning it the **exact** value from the output.\n", "_____no_output_____" ] ], [ [ "# Code here\n", "_____no_output_____" ] ], [ [ "### PROMPT 3: Re-write the code from prompt 1 to use your variables\n\nNow we re-write the code from prompt 1 but use variables instead. Try using f-strings as they provide the best formatting options.", "_____no_output_____" ] ], [ [ "# Code here\n", "_____no_output_____" ] ], [ [ "### PROMPT 4: What transformations are required to get the variables from input to output?\n\nAre there any transformations required of the variables from input to output? For example, you might input first name and last name separarely, then combine them to full name.", "_____no_output_____" ] ], [ [ "# Code here", "_____no_output_____" ] ], [ [ "### PROMPT 5: Finalize the inputs\n\ncombine prompts 3 and 4 into a single program, replace the initialized variables with Python inputs and prompts. ", "_____no_output_____" ] ], [ [ "# Code here... now its written!", "_____no_output_____" ] ], [ [ "## Step 1: Problem Analysis\n\nInputs: PROMPT 6\n\nOutputs: PROMPT 7\n\nAlgorithm (Steps in Program):\n\nPROMPT 8\n", "_____no_output_____" ] ], [ [ "# Step 2: Write code here\n\n\n", "_____no_output_____" ], [ "# run this code to turn in your work!\nfrom coursetools.submission import Submission\nSubmission().submit_now()", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code" ]
[ [ "markdown", "markdown", "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ] ]
ecec3d153b6a5ec700f9c00fea21bab697d4cbbe
18,048
ipynb
Jupyter Notebook
tensorflow/lite/g3doc/models/style_transfer/overview.ipynb
ZhuBaohe/tensorflow
1114a0364ef38d81d7a34e262994cf771e3c9460
[ "Apache-2.0" ]
7
2018-04-12T07:48:57.000Z
2021-12-03T12:35:02.000Z
tensorflow/lite/g3doc/models/style_transfer/overview.ipynb
jatinarora1/tensorflow
f91c1776370e0c81a2326eb33e86aadbb1063e72
[ "Apache-2.0" ]
6
2022-01-15T07:17:47.000Z
2022-02-14T15:28:22.000Z
tensorflow/lite/g3doc/models/style_transfer/overview.ipynb
I-Hong/tensorflow
4a1ae31d56c3c7f40232aace615945c29dcf9c38
[ "Apache-2.0" ]
2
2019-06-18T01:54:19.000Z
2019-06-19T11:17:05.000Z
36.534413
366
0.561946
[ [ [ "##### Copyright 2019 The TensorFlow Authors.", "_____no_output_____" ] ], [ [ "#@title Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# https://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.", "_____no_output_____" ] ], [ [ "# Artistic Style Transfer with TensorFlow Lite", "_____no_output_____" ], [ "<table class=\"tfo-notebook-buttons\" align=\"left\">\n <td>\n <a target=\"_blank\" href=\"https://www.tensorflow.org/lite/models/style_transfer/overview\"><img src=\"https://www.tensorflow.org/images/tf_logo_32px.png\" />View on TensorFlow.org</a>\n </td>\n <td>\n <a target=\"_blank\" href=\"https://colab.research.google.com/github/tensorflow/tensorflow/blob/master/tensorflow/lite/g3doc/models/style_transfer/overview.ipynb\"><img src=\"https://www.tensorflow.org/images/colab_logo_32px.png\" />Run in Google Colab</a>\n </td>\n <td>\n <a target=\"_blank\" href=\"https://github.com/tensorflow/tensorflow/blob/master/tensorflow/lite/g3doc/models/style_transfer/overview.ipynb\"><img src=\"https://www.tensorflow.org/images/GitHub-Mark-32px.png\" />View source on GitHub</a>\n </td>\n <td>\n <a href=\"https://storage.googleapis.com/tensorflow_docs/tensorflow/tensorflow/lite/g3doc/models/style_transfer/overview.ipynb\"><img src=\"https://www.tensorflow.org/images/download_logo_32px.png\" />Download notebook</a>\n </td>\n</table>", "_____no_output_____" ], [ "One of the most exciting developments in deep learning to come out recently is [artistic style transfer](https://arxiv.org/abs/1508.06576), or the ability to create a new image, known as a [pastiche](https://en.wikipedia.org/wiki/Pastiche), based on two input images: one representing the artistic style and one representing the content.\n\n![Style transfer example](https://storage.googleapis.com/download.tensorflow.org/models/tflite/arbitrary_style_transfer/formula.png)\n\nUsing this technique, we can generate beautiful new artworks in a range of styles.\n\n![Style transfer example](https://storage.googleapis.com/download.tensorflow.org/models/tflite/arbitrary_style_transfer/table.png)\n\nThis tutorial shows how to use a pre-trained TensorFlow Lite model to apply style transfer on any pair of content and style image. You can use the pre-trained model to add style transfer to your own mobile applications.\n\nThe model is open-sourced on [GitHub](https://github.com/tensorflow/magenta/tree/master/magenta/models/arbitrary_image_stylization#train-a-model-on-a-large-dataset-with-data-augmentation-to-run-on-mobile). You can retrain the model with different parameters (e.g. increase content layers' weights to make the output image look more like the content image).", "_____no_output_____" ], [ "## Understand the model architecture", "_____no_output_____" ], [ "![Model Architecture](https://storage.googleapis.com/download.tensorflow.org/models/tflite/arbitrary_style_transfer/architecture.png)\n\nThis Artistic Style Transfer model consists of two submodels:\n1. **Style Prediciton Model**: A MobilenetV2-based neural network that takes an input style image to a 100-dimension style bottleneck vector.\n1. **Style Transform Model**: A neural network that takes apply a style bottleneck vector to a content image and creates a stylized image.\n\nIf your app only needs to support a fixed set of style images, you can compute their style bottleneck vectors in advance, and exclude the Style Prediction Model from your app's binary.", "_____no_output_____" ], [ "## Setup", "_____no_output_____" ], [ "Import dependencies.", "_____no_output_____" ] ], [ [ "from __future__ import absolute_import, division, print_function, unicode_literals", "_____no_output_____" ], [ "try:\n # %tensorflow_version only exists in Colab.\n import tensorflow.compat.v2 as tf\nexcept Exception:\n pass\ntf.enable_v2_behavior()", "_____no_output_____" ], [ "import IPython.display as display\n\nimport matplotlib.pyplot as plt\nimport matplotlib as mpl\nmpl.rcParams['figure.figsize'] = (12,12)\nmpl.rcParams['axes.grid'] = False\n\nimport numpy as np\nimport time\nimport functools", "_____no_output_____" ] ], [ [ "Download the content and style images, and the pre-trained TensorFlow Lite models.", "_____no_output_____" ] ], [ [ "content_path = tf.keras.utils.get_file('belfry.jpg','https://storage.googleapis.com/khanhlvg-public.appspot.com/arbitrary-style-transfer/belfry-2611573_1280.jpg')\nstyle_path = tf.keras.utils.get_file('style23.jpg','https://storage.googleapis.com/khanhlvg-public.appspot.com/arbitrary-style-transfer/style23.jpg')\n\nstyle_predict_path = tf.keras.utils.get_file('style_predict.tflite', 'https://storage.googleapis.com/download.tensorflow.org/models/tflite/arbitrary_style_transfer/style_predict_quantized_256.tflite')\nstyle_transform_path = tf.keras.utils.get_file('style_transform.tflite', 'https://storage.googleapis.com/download.tensorflow.org/models/tflite/arbitrary_style_transfer/style_transfer_quantized_dynamic.tflite')", "_____no_output_____" ] ], [ [ "## Pre-process the inputs\n\n* The content image and the style image must be RGB images with pixel values being float32 numbers between [0..1].\n* The style image size must be (1, 256, 256, 3). We central crop the image and resize it.\n* The content image can be any size. However, as we trained the model using square-cropped data, cropping the content image to a square results in better stylized image.", "_____no_output_____" ] ], [ [ "# Function to load an image from a file, and add a batch dimension.\ndef load_img(path_to_img):\n img = tf.io.read_file(path_to_img)\n img = tf.image.decode_image(img, channels=3)\n img = tf.image.convert_image_dtype(img, tf.float32)\n img = img[tf.newaxis, :]\n\n return img\n\n# Function to pre-process style image input.\ndef preprocess_style_image(style_image):\n # Resize the image so that the shorter dimension becomes 256px.\n target_dim = 256\n shape = tf.cast(tf.shape(style_image)[1:-1], tf.float32)\n short_dim = min(shape)\n scale = target_dim / short_dim\n new_shape = tf.cast(shape * scale, tf.int32)\n style_image = tf.image.resize(style_image, new_shape)\n\n # Central crop the image.\n style_image = tf.image.resize_with_crop_or_pad(style_image, target_dim, target_dim)\n\n return style_image\n\n# Function to pre-process content image input.\ndef preprocess_content_image(content_image):\n # Central crop the image.\n shape = tf.shape(content_image)[1:-1]\n short_dim = min(shape)\n content_image = tf.image.resize_with_crop_or_pad(content_image, short_dim, short_dim)\n\n return content_image\n\n# Load the input images.\ncontent_image = load_img(content_path)\nstyle_image = load_img(style_path)\n\n# Preprocess the input images.\npreprocessed_content_image = preprocess_content_image(content_image)\npreprocessed_style_image = preprocess_style_image(style_image)\n\nprint('Style Image Shape:', preprocessed_style_image.shape)\nprint('Content Image Shape:', preprocessed_content_image.shape)", "_____no_output_____" ] ], [ [ "## Visualize the inputs", "_____no_output_____" ] ], [ [ "def imshow(image, title=None):\n if len(image.shape) > 3:\n image = tf.squeeze(image, axis=0)\n\n plt.imshow(image)\n if title:\n plt.title(title)\n\nplt.subplot(1, 2, 1)\nimshow(preprocessed_content_image, 'Content Image')\n\nplt.subplot(1, 2, 2)\nimshow(preprocessed_style_image, 'Style Image')", "_____no_output_____" ] ], [ [ "## Run style transfer with TensorFlow Lite", "_____no_output_____" ], [ "### Style prediction", "_____no_output_____" ] ], [ [ "# Function to run style prediction on preprocessed style image.\ndef run_style_predict(preprocessed_style_image):\n # Load the model.\n interpreter = tf.lite.Interpreter(model_path=style_predict_path)\n\n # Set model input.\n interpreter.allocate_tensors()\n input_details = interpreter.get_input_details()\n interpreter.set_tensor(input_details[0][\"index\"], preprocessed_style_image)\n\n # Calculate style bottleneck.\n interpreter.invoke()\n style_bottleneck = interpreter.tensor(\n interpreter.get_output_details()[0][\"index\"]\n )()\n\n return style_bottleneck\n\n# Calculate style bottleneck for the preprocessed style image.\nstyle_bottleneck = run_style_predict(preprocessed_style_image)\nprint('Style Bottleneck Shape:', style_bottleneck.shape)", "_____no_output_____" ] ], [ [ "### Style transform", "_____no_output_____" ] ], [ [ "# Run style transform on preprocessed style image\ndef run_style_transform(style_bottleneck, preprocessed_content_image):\n # Load the model.\n interpreter = tf.lite.Interpreter(model_path=style_transform_path)\n\n # Set model input.\n input_details = interpreter.get_input_details()\n interpreter.resize_tensor_input(input_details[0][\"index\"],\n preprocessed_content_image.shape)\n interpreter.allocate_tensors()\n\n # Set model inputs.\n interpreter.set_tensor(input_details[0][\"index\"], preprocessed_content_image)\n interpreter.set_tensor(input_details[1][\"index\"], style_bottleneck)\n interpreter.invoke()\n\n # Transform content image.\n stylized_image = interpreter.tensor(\n interpreter.get_output_details()[0][\"index\"]\n )()\n\n return stylized_image\n\n# Stylize the content image using the style bottleneck.\nstylized_image = run_style_transform(style_bottleneck, preprocessed_content_image)\n\n# Visualize the output.\nimshow(stylized_image, 'Stylized Image')", "_____no_output_____" ] ], [ [ "### Style blending\n\nWe can blend the style of content image into the stylized output, which in turn making the output look more like the content image.", "_____no_output_____" ] ], [ [ "# Calculate style bottleneck of the content image.\nstyle_bottleneck_content = run_style_predict(\n preprocess_style_image(content_image)\n )", "_____no_output_____" ], [ "# Define content blending ratio between [0..1].\n# 0.0: 0% style extracts from content image.\n# 1.0: 100% style extracted from content image.\ncontent_blending_ratio = 0.5 #@param {type:\"slider\", min:0, max:1, step:0.01}\n\n# Blend the style bottleneck of style image and content image\nstyle_bottleneck_blended = content_blending_ratio * style_bottleneck_content \\\n + (1 - content_blending_ratio) * style_bottleneck\n\n# Stylize the content image using the style bottleneck.\nstylized_image_blended = run_style_transform(style_bottleneck_blended,\n preprocessed_content_image)\n\n# Visualize the output.\nimshow(stylized_image_blended, 'Blended Stylized Image')", "_____no_output_____" ] ] ]
[ "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code", "markdown", "code" ]
[ [ "markdown" ], [ "code" ], [ "markdown", "markdown", "markdown", "markdown", "markdown", "markdown", "markdown" ], [ "code", "code", "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown", "markdown" ], [ "code" ], [ "markdown" ], [ "code" ], [ "markdown" ], [ "code", "code" ] ]