**Neural Network**

Create a simple feed-forward neural network in python. You must implement the following components:

- An input layer that receives data and passes it on
- Some hidden layers (at least a 2-layer neural network)
- An output layer
- Weights and biases between each layer
- Use the Sigmoid activation function for each hidden layer
- Implement a simple feed-forward neural network
- Train the feed-forward neurons using back-propagation.

```
{
"nbformat": 4,
"nbformat_minor": 0,
"metadata": {
"colab": {
"name": "Tiana.ipynb",
"provenance": []
},
"kernelspec": {
"name": "python3",
"display_name": "Python 3"
},
"language_info": {
"name": "python"
}
},
"cells": [
{
"cell_type": "code",
"metadata": {
"id": "L5AWaGEqD_fx"
},
"source": [
"import numpy as np\n",
"import tensorflow as tf\n",
"import matplotlib.pyplot as plt\n",
"import pandas as pd\n",
"from keras.datasets import boston_housing\n",
"import sys\n",
"import time"
],
"execution_count": 81,
"outputs": []
},
{
"cell_type": "markdown",
"metadata": {
"id": "VMkrkrGTFRb7"
},
"source": [
"# Read DATA"
]
},
{
"cell_type": "code",
"metadata": {
"id": "2CPpkcZzHR5r"
},
"source": [
"(train_data, train_targets), (test_data, test_targets) = boston_housing.load_data()"
],
"execution_count": 57,
"outputs": []
},
{
"cell_type": "code",
"metadata": {
"colab": {
"base_uri": "https://localhost:8080/"
},
"id": "I8Su1eJPHm3q",
"outputId": "60eba728-b37f-46a6-cfe4-1bdddf375a11"
},
"source": [
"print(f'Training data : {train_data.shape}')\n",
"print(f'Test data : {test_data.shape}')\n",
"print(f'Training sample : {train_data[0]}')\n",
"print(f'Training target sample : {train_targets[0]}')"
],
"execution_count": 58,
"outputs": [
{
"output_type": "stream",
"text": [
"Training data : (404, 13)\n",
"Test data : (102, 13)\n",
"Training sample : [ 1.23247 0. 8.14 0. 0.538 6.142 91.7\n",
" 3.9769 4. 307. 21. 396.9 18.72 ]\n",
"Training target sample : 15.2\n"
],
"name": "stdout"
}
]
},
{
"cell_type": "markdown",
"metadata": {
"id": "_9FZvnIPH0Lw"
},
"source": [
"# Build Model"
]
},
{
"cell_type": "code",
"metadata": {
"id": "OWHDS9M4HznA"
},
"source": [
"model = tf.keras.models.Sequential()\n",
"\n",
"# Add input layer\n",
"model.add(tf.keras.layers.Dense(10, input_dim = train_data.shape[1]))\n",
"\n",
"# Add two hidden layers\n",
"model.add(tf.keras.layers.Dense(64, activation='sigmoid'))\n",
"model.add(tf.keras.layers.Dropout(0.1))\n",
"model.add(tf.keras.layers.Dense(32, activation='sigmoid'))\n",
"model.add(tf.keras.layers.Dropout(0.1))\n",
"model.add(tf.keras.layers.Dense(16, activation='sigmoid'))\n",
"model.add(tf.keras.layers.Dropout(0.1))\n",
"model.add(tf.keras.layers.Dense(8, activation='sigmoid'))\n",
"model.add(tf.keras.layers.Dropout(0.1))\n",
"# Add output layer\n",
"model.add(tf.keras.layers.Dense(1))\n",
"\n",
"# compile\n",
"#model.compile(optimizer = tf.keras.optimizers.Adam(lr=1e-3), loss = 'mse')"
],
"execution_count": 146,
"outputs": []
},
{
"cell_type": "code",
"metadata": {
"colab": {
"base_uri": "https://localhost:8080/"
},
"id": "EMRRvdaZYuU3",
"outputId": "db1c285a-a978-4e10-bda9-e19995185816"
},
"source": [
"for i in range(12):\n",
" print(model.trainable_weights[i].shape)"
],
"execution_count": 136,
"outputs": [
{
"output_type": "stream",
"text": [
"(13, 10)\n",
"(10,)\n",
"(10, 64)\n",
"(64,)\n",
"(64, 32)\n",
"(32,)\n",
"(32, 16)\n",
"(16,)\n",
"(16, 8)\n",
"(8,)\n",
"(8, 1)\n",
"(1,)\n"
],
"name": "stdout"
}
]
},
{
"cell_type": "markdown",
"metadata": {
"id": "HlL7XJMsJFaJ"
},
"source": [
"# Run Model"
]
},
{
"cell_type": "code",
"metadata": {
"colab": {
"base_uri": "https://localhost:8080/",
"height": 357
},
"id": "J24adzVrJEjL",
"outputId": "ca75e583-6c65-440c-e5f5-938b33fa0ef5"
},
"source": [
"EPOCHS = 1000\n",
"history = model.fit(train_data, train_targets, validation_data = (test_data, test_targets), epochs = EPOCHS, steps_per_epoch=20, verbose = 1)"
],
"execution_count": 137,
"outputs": [
{
"output_type": "error",
"ename": "RuntimeError",
"evalue": "ignored",
"traceback": [
"\u001b[0;31m---------------------------------------------------------------------------\u001b[0m",
"\u001b[0;31mRuntimeError\u001b[0m Traceback (most recent call last)",
"\u001b[0;32m
```\u001b[0m in \u001b[0;36m\u001b[0;34m()\u001b[0m\n\u001b[1;32m 1\u001b[0m \u001b[0mEPOCHS\u001b[0m \u001b[0;34m=\u001b[0m \u001b[0;36m1000\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0;32m----> 2\u001b[0;31m \u001b[0mhistory\u001b[0m \u001b[0;34m=\u001b[0m \u001b[0mmodel\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mfit\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0mtrain_data\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0mtrain_targets\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0mvalidation_data\u001b[0m \u001b[0;34m=\u001b[0m \u001b[0;34m(\u001b[0m\u001b[0mtest_data\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0mtest_targets\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0mepochs\u001b[0m \u001b[0;34m=\u001b[0m \u001b[0mEPOCHS\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0msteps_per_epoch\u001b[0m\u001b[0;34m=\u001b[0m\u001b[0;36m20\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0mverbose\u001b[0m \u001b[0;34m=\u001b[0m \u001b[0;36m1\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0m",
"\u001b[0;32m/usr/local/lib/python3.7/dist-packages/tensorflow/python/keras/engine/training.py\u001b[0m in \u001b[0;36mfit\u001b[0;34m(self, x, y, batch_size, epochs, verbose, callbacks, validation_split, validation_data, shuffle, class_weight, sample_weight, initial_epoch, steps_per_epoch, validation_steps, validation_batch_size, validation_freq, max_queue_size, workers, use_multiprocessing)\u001b[0m\n\u001b[1;32m 1103\u001b[0m \u001b[0;31m# Legacy graph support is contained in `training_v1.Model`.\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m 1104\u001b[0m \u001b[0mversion_utils\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mdisallow_legacy_graph\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0;34m'Model'\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0;34m'fit'\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0;32m-> 1105\u001b[0;31m \u001b[0mself\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0m_assert_compile_was_called\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0m\u001b[1;32m 1106\u001b[0m \u001b[0mself\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0m_check_call_args\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0;34m'fit'\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m 1107\u001b[0m \u001b[0m_disallow_inside_tf_function\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0;34m'fit'\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n",
"\u001b[0;32m/usr/local/lib/python3.7/dist-packages/tensorflow/python/keras/engine/training.py\u001b[0m in \u001b[0;36m_assert_compile_was_called\u001b[0;34m(self)\u001b[0m\n\u001b[1;32m 2691\u001b[0m \u001b[0;31m# (i.e. whether the model is built and its inputs/outputs are set).\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m 2692\u001b[0m \u001b[0;32mif\u001b[0m \u001b[0;32mnot\u001b[0m \u001b[0mself\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0m_is_compiled\u001b[0m\u001b[0;34m:\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0;32m-> 2693\u001b[0;31m raise RuntimeError('You must compile your model before '\n\u001b[0m\u001b[1;32m 2694\u001b[0m \u001b[0;34m'training/testing. '\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m 2695\u001b[0m 'Use `model.compile(optimizer, loss)`.')\n",
"\u001b[0;31mRuntimeError\u001b[0m: You must compile your model before training/testing. Use `model.compile(optimizer, loss)`."
]
}
]
},
{
"cell_type": "markdown",
"metadata": {
"id": "djasrd4OTK9_"
},
"source": [
"# Backpropagation"
]
},
{
"cell_type": "code",
"metadata": {
"id": "cmtFGf0NZjBA"
},
"source": [
"weights = tf.random.uniform(shape=(8,1))\n",
"offsets = tf.random.uniform(shape=(1,))"
],
"execution_count": 147,
"outputs": []
},
{
"cell_type": "code",
"metadata": {
"id": "FS6BEJlfZt-b"
},
"source": [
"model.layers[9].set_weights([weights, offsets])"
],
"execution_count": 148,
"outputs": []
},
{
"cell_type": "code",
"metadata": {
"colab": {
"base_uri": "https://localhost:8080/"
},
"id": "q_FW6rvvTKnK",
"outputId": "58e33216-bb83-4110-c909-5cb11e6664ad"
},
"source": [
"EPOCHS = 1000\n",
"BATCH_SIZE = 64\n",
"LEARNING_RATE = 1E-3\n",
"\n",
"opt = tf.keras.optimizers.RMSprop(learning_rate=LEARNING_RATE, decay=LEARNING_RATE / EPOCHS)\n",
"\n",
"# calculate number of iteratios to do\n",
"num_iters = int(train_data.shape[0] / BATCH_SIZE)\n",
"\n",
"loss_fn = loss = tf.keras.losses.mean_squared_error\n",
"losses = []\n",
"val_losses = []\n",
"for epoch in range(0, EPOCHS):\n",
"\tsys.stdout.flush()\n",
"\tepoch_start = time.time()\n",
"\tfor i in range(0, num_iters):\n",
"\t\t# determine starting and ending index of batch\n",
"\t\tstart = i * BATCH_SIZE\n",
"\t\tend = start + BATCH_SIZE\n",
"\t\t# step\n",
"\t\twith tf.GradientTape() as tape:\n",
" # predict\n",
"\t\t\tX_ = train_data[start:end]\n",
"\t\t\ty_ = train_targets[start:end]\n",
"\t\t\ty_pred = model(X_) # this returns an array of size (N, 13) where N is the size of the test dataset.\n",
" # Pick the label with the highest probability\n",
"\t\t\tloss = loss_fn(y_, y_pred) # calculate loss\n",
"\t\t\tXtest_ = test_data\n",
"\t\t\tytest_ = test_targets\n",
"\t\t\tytest_pred = model(Xtest_)\n",
"\t\t\tval_loss = np.mean(loss_fn(ytest_, ytest_pred))\n",
"\n",
" # calculate the gradients using our tape and then update the\n",
" # model weights\n",
"\t\tgrads = tape.gradient(loss, model.trainable_weights)\n",
"\t\topt.apply_gradients(zip(grads, model.trainable_weights))\n",
"\t# show timing information for the epoch\n",
" # show the current epoch number\n",
"\tprint(\"Epoch {}/{} - loss: {:.4f} - val_loss: {:.4f}\".format(\n",
"\t\tepoch + 1, EPOCHS, np.mean(loss), np.mean(val_loss)), end=\"\")\n",
"\tepoch_end = time.time()\n",
"\telapsed = (epoch_end - epoch_start)\n",
"\tprint(\"-> took {:.2f} seconds\".format(elapsed))\n",
"\tlosses.append(np.mean(loss.numpy()))\n",
"\tval_losses.append(val_loss)"
],
"execution_count": 150,
"outputs": [
{
"output_type": "stream",
"text": [
"Epoch 1/1000 - loss: 598.2261 - val_loss: 491.6591-> took 0.13 seconds\n",
"Epoch 2/1000 - loss: 590.0509 - val_loss: 484.2758-> took 0.13 seconds\n",
"Epoch 3/1000 - loss: 583.1237 - val_loss: 478.0358-> took 0.14 seconds\n",
"Epoch 4/1000 - loss: 576.7520 - val_loss: 472.2795-> took 0.13 seconds\n",
"Epoch 5/1000 - loss: 570.6929 - val_loss: 466.8140-> took 0.14 seconds\n",
"Epoch 6/1000 - loss: 564.8884 - val_loss: 461.5770-> took 0.13 seconds\n",
"Epoch 7/1000 - loss: 559.2507 - val_loss: 456.4980-> took 0.13 seconds\n",
"Epoch 8/1000 - loss: 553.7295 - val_loss: 451.5350-> took 0.13 seconds\n",
"Epoch 9/1000 - loss: 548.3182 - val_loss: 446.6741-> took 0.15 seconds\n",
"Epoch 10/1000 - loss: 543.0594 - val_loss: 441.9456-> took 0.14 seconds\n",
"Epoch 11/1000 - loss: 537.9497 - val_loss: 437.3539-> took 0.16 seconds\n",
"Epoch 12/1000 - loss: 532.9792 - val_loss: 432.8968-> took 0.15 seconds\n",
"Epoch 13/1000 - loss: 528.1459 - val_loss: 428.5631-> took 0.15 seconds\n",
"Epoch 14/1000 - loss: 523.4313 - val_loss: 424.3358-> took 0.13 seconds\n",
"Epoch 15/1000 - loss: 518.7852 - val_loss: 420.1700-> took 0.13 seconds\n",
"Epoch 16/1000 - loss: 514.2212 - val_loss: 416.0862-> took 0.15 seconds\n",
"Epoch 17/1000 - loss: 509.7957 - val_loss: 412.1268-> took 0.14 seconds\n",
"Epoch 18/1000 - loss: 505.4937 - val_loss: 408.2802-> took 0.13 seconds\n",
"Epoch 19/1000 - loss: 501.3298 - val_loss: 404.5589-> took 0.12 seconds\n",
"Epoch 20/1000 - loss: 497.2869 - val_loss: 400.9485-> took 0.14 seconds\n",
"Epoch 21/1000 - loss: 493.3611 - val_loss: 397.4438-> took 0.13 seconds\n",
"Epoch 22/1000 - loss: 489.5528 - val_loss: 394.0469-> took 0.14 seconds\n",
"Epoch 23/1000 - loss: 485.8574 - val_loss: 390.7527-> took 0.13 seconds\n",
"Epoch 24/1000 - loss: 482.2721 - val_loss: 387.5583-> took 0.13 seconds\n",
"Epoch 25/1000 - loss: 478.7928 - val_loss: 384.4598-> took 0.14 seconds\n",
"Epoch 26/1000 - loss: 475.4146 - val_loss: 381.4541-> took 0.13 seconds\n",
"Epoch 27/1000 - loss: 472.1335 - val_loss: 378.5358-> took 0.12 seconds\n",
"Epoch 28/1000 - loss: 468.9500 - val_loss: 375.7068-> took 0.13 seconds\n",
"Epoch 29/1000 - loss: 465.8613 - val_loss: 372.9629-> took 0.14 seconds\n",
"Epoch 30/1000 - loss: 462.8603 - val_loss: 370.2983-> took 0.13 seconds\n",
"Epoch 31/1000 - loss: 459.9394 - val_loss: 367.7050-> took 0.13 seconds\n",
"Epoch 32/1000 - loss: 457.0928 - val_loss: 365.1799-> took 0.15 seconds\n",
"Epoch 33/1000 - loss: 454.3147 - val_loss: 362.7173-> took 0.12 seconds\n",
"Epoch 34/1000 - loss: 451.6016 - val_loss: 360.3132-> took 0.14 seconds\n",
"Epoch 35/1000 - loss: 448.9493 - val_loss: 357.9639-> took 0.13 seconds\n",
"Epoch 36/1000 - loss: 446.3535 - val_loss: 355.6659-> took 0.13 seconds\n",
"Epoch 37/1000 - loss: 443.8101 - val_loss: 353.4152-> took 0.13 seconds\n",
"Epoch 38/1000 - loss: 441.3170 - val_loss: 351.2101-> took 0.12 seconds\n",
"Epoch 39/1000 - loss: 438.8707 - val_loss: 349.0476-> took 0.13 seconds\n",
"Epoch 40/1000 - loss: 436.4684 - val_loss: 346.9247-> took 0.13 seconds\n",
"Epoch 41/1000 - loss: 434.1071 - val_loss: 344.8392-> took 0.12 seconds\n",
"Epoch 42/1000 - loss: 431.7858 - val_loss: 342.7897-> took 0.13 seconds\n",
"Epoch 43/1000 - loss: 429.5008 - val_loss: 340.7735-> took 0.13 seconds\n",
"Epoch 44/1000 - loss: 427.2509 - val_loss: 338.7893-> took 0.14 seconds\n",
"Epoch 45/1000 - loss: 425.0349 - val_loss: 336.8357-> took 0.13 seconds\n",
"Epoch 46/1000 - loss: 422.8495 - val_loss: 334.9099-> took 0.13 seconds\n",
"Epoch 47/1000 - loss: 420.6929 - val_loss: 333.0105-> took 0.13 seconds\n",
"Epoch 48/1000 - loss: 418.5632 - val_loss: 331.1355-> took 0.13 seconds\n",
"Epoch 49/1000 - loss: 416.4581 - val_loss: 329.2831-> took 0.12 seconds\n",
"Epoch 50/1000 - loss: 414.3763 - val_loss: 327.4519-> took 0.13 seconds\n",
"Epoch 51/1000 - loss: 412.3162 - val_loss: 325.6407-> took 0.13 seconds\n",
"Epoch 52/1000 - loss: 410.2764 - val_loss: 323.8481-> took 0.13 seconds\n",
"Epoch 53/1000 - loss: 408.2556 - val_loss: 322.0730-> took 0.13 seconds\n",
"Epoch 54/1000 - loss: 406.2526 - val_loss: 320.3145-> took 0.13 seconds\n",
"Epoch 55/1000 - loss: 404.2665 - val_loss: 318.5715-> took 0.14 seconds\n",
"Epoch 56/1000 - loss: 402.2963 - val_loss: 316.8432-> took 0.13 seconds\n",
"Epoch 57/1000 - loss: 400.3410 - val_loss: 315.1287-> took 0.13 seconds\n",
"Epoch 58/1000 - loss: 398.4000 - val_loss: 313.4276-> took 0.13 seconds\n",
"Epoch 59/1000 - loss: 396.4724 - val_loss: 311.7390-> took 0.13 seconds\n",
"Epoch 60/1000 - loss: 394.5578 - val_loss: 310.0626-> took 0.12 seconds\n",
"Epoch 61/1000 - loss: 392.6555 - val_loss: 308.3977-> took 0.14 seconds\n",
"Epoch 62/1000 - loss: 390.7649 - val_loss: 306.7437-> took 0.14 seconds\n",
"Epoch 63/1000 - loss: 388.8855 - val_loss: 305.1003-> took 0.13 seconds\n",
"Epoch 64/1000 - loss: 387.0168 - val_loss: 303.4672-> took 0.13 seconds\n",
"Epoch 65/1000 - loss: 385.1586 - val_loss: 301.8438-> took 0.13 seconds\n",
"Epoch 66/1000 - loss: 383.3103 - val_loss: 300.2300-> took 0.14 seconds\n",
"Epoch 67/1000 - loss: 381.4717 - val_loss: 298.6254-> took 0.14 seconds\n",
"Epoch 68/1000 - loss: 379.6424 - val_loss: 297.0295-> took 0.13 seconds\n",
"Epoch 69/1000 - loss: 377.8222 - val_loss: 295.4425-> took 0.13 seconds\n",
"Epoch 70/1000 - loss: 376.0107 - val_loss: 293.8637-> took 0.14 seconds\n",
"Epoch 71/1000 - loss: 374.2077 - val_loss: 292.2930-> took 0.13 seconds\n",
"Epoch 72/1000 - loss: 372.4130 - val_loss: 290.7305-> took 0.15 seconds\n",
"Epoch 73/1000 - loss: 370.6264 - val_loss: 289.1757-> took 0.13 seconds\n",
"Epoch 74/1000 - loss: 368.8479 - val_loss: 287.6287-> took 0.13 seconds\n",
"Epoch 75/1000 - loss: 367.0769 - val_loss: 286.0890-> took 0.14 seconds\n",
"Epoch 76/1000 - loss: 365.3135 - val_loss: 284.5566-> took 0.16 seconds\n",
"Epoch 77/1000 - loss: 363.5576 - val_loss: 283.0315-> took 0.17 seconds\n",
"Epoch 78/1000 - loss: 361.8090 - val_loss: 281.5136-> took 0.13 seconds\n",
"Epoch 79/1000 - loss: 360.0676 - val_loss: 280.0026-> took 0.13 seconds\n",
"Epoch 80/1000 - loss: 358.3332 - val_loss: 278.4985-> took 0.13 seconds\n",
"Epoch 81/1000 - loss: 356.6059 - val_loss: 277.0013-> took 0.13 seconds\n",
"Epoch 82/1000 - loss: 354.8853 - val_loss: 275.5107-> took 0.13 seconds\n",
"Epoch 83/1000 - loss: 353.1716 - val_loss: 274.0269-> took 0.13 seconds\n",
"Epoch 84/1000 - loss: 351.4646 - val_loss: 272.5495-> took 0.12 seconds\n",
"Epoch 85/1000 - loss: 349.7642 - val_loss: 271.0786-> took 0.14 seconds\n",
"Epoch 86/1000 - loss: 348.0704 - val_loss: 269.6144-> took 0.13 seconds\n",
"Epoch 87/1000 - loss: 346.3832 - val_loss: 268.1565-> took 0.12 seconds\n",
"Epoch 88/1000 - loss: 344.7024 - val_loss: 266.7048-> took 0.13 seconds\n",
"Epoch 89/1000 - loss: 343.0279 - val_loss: 265.2596-> took 0.14 seconds\n",
"Epoch 90/1000 - loss: 341.3599 - val_loss: 263.8206-> took 0.13 seconds\n",
"Epoch 91/1000 - loss: 339.6981 - val_loss: 262.3878-> took 0.13 seconds\n",
"Epoch 92/1000 - loss: 338.0426 - val_loss: 260.9612-> took 0.14 seconds\n",
"Epoch 93/1000 - loss: 336.3933 - val_loss: 259.5408-> took 0.12 seconds\n",
"Epoch 94/1000 - loss: 334.7502 - val_loss: 258.1266-> took 0.14 seconds\n",
"Epoch 95/1000 - loss: 333.1133 - val_loss: 256.7184-> took 0.14 seconds\n",
"Epoch 96/1000 - loss: 331.4826 - val_loss: 255.3163-> took 0.14 seconds\n",
"Epoch 97/1000 - loss: 329.8579 - val_loss: 253.9201-> took 0.13 seconds\n",
"Epoch 98/1000 - loss: 328.2395 - val_loss: 252.5301-> took 0.15 seconds\n",
"Epoch 99/1000 - loss: 326.6269 - val_loss: 251.1462-> took 0.14 seconds\n",
"Epoch 100/1000 - loss: 325.0205 - val_loss: 249.7681-> took 0.13 seconds\n",
"Epoch 101/1000 - loss: 323.4201 - val_loss: 248.3961-> took 0.12 seconds\n",
"Epoch 102/1000 - loss: 321.8257 - val_loss: 247.0300-> took 0.13 seconds\n",
"Epoch 103/1000 - loss: 320.2373 - val_loss: 245.6699-> took 0.13 seconds\n",
"Epoch 104/1000 - loss: 318.6548 - val_loss: 244.3157-> took 0.12 seconds\n",
"Epoch 105/1000 - loss: 317.0784 - val_loss: 242.9673-> took 0.15 seconds\n",
"Epoch 106/1000 - loss: 315.5078 - val_loss: 241.6250-> took 0.14 seconds\n",
"Epoch 107/1000 - loss: 313.9434 - val_loss: 240.2886-> took 0.13 seconds\n",
"Epoch 108/1000 - loss: 312.3847 - val_loss: 238.9581-> took 0.13 seconds\n",
"Epoch 109/1000 - loss: 310.8320 - val_loss: 237.6333-> took 0.13 seconds\n",
"Epoch 110/1000 - loss: 309.2853 - val_loss: 236.3146-> took 0.13 seconds\n",
"Epoch 111/1000 - loss: 307.7444 - val_loss: 235.0017-> took 0.14 seconds\n",
"Epoch 112/1000 - loss: 306.2095 - val_loss: 233.6947-> took 0.13 seconds\n",
"Epoch 113/1000 - loss: 304.6804 - val_loss: 232.3934-> took 0.14 seconds\n",
"Epoch 114/1000 - loss: 303.1572 - val_loss: 231.0981-> took 0.13 seconds\n",
"Epoch 115/1000 - loss: 301.6399 - val_loss: 229.8087-> took 0.13 seconds\n",
"Epoch 116/1000 - loss: 300.1285 - val_loss: 228.5250-> took 0.12 seconds\n",
"Epoch 117/1000 - loss: 298.6230 - val_loss: 227.2473-> took 0.13 seconds\n",
"Epoch 118/1000 - loss: 297.1233 - val_loss: 225.9752-> took 0.13 seconds\n",
"Epoch 119/1000 - loss: 295.6295 - val_loss: 224.7091-> took 0.13 seconds\n",
"Epoch 120/1000 - loss: 294.1416 - val_loss: 223.4488-> took 0.14 seconds\n",
"Epoch 121/1000 - loss: 292.6595 - val_loss: 222.1944-> took 0.13 seconds\n",
"Epoch 122/1000 - loss: 291.1833 - val_loss: 220.9457-> took 0.15 seconds\n",
"Epoch 123/1000 - loss: 289.7129 - val_loss: 219.7029-> took 0.13 seconds\n",
"Epoch 124/1000 - loss: 288.2484 - val_loss: 218.4657-> took 0.12 seconds\n",
"Epoch 125/1000 - loss: 286.7898 - val_loss: 217.2346-> took 0.13 seconds\n",
"Epoch 126/1000 - loss: 285.3369 - val_loss: 216.0093-> took 0.13 seconds\n",
"Epoch 127/1000 - loss: 283.8899 - val_loss: 214.7895-> took 0.13 seconds\n",
"Epoch 128/1000 - loss: 282.4487 - val_loss: 213.5759-> took 0.12 seconds\n",
"Epoch 129/1000 - loss: 281.0135 - val_loss: 212.3678-> took 0.13 seconds\n",
"Epoch 130/1000 - loss: 279.5840 - val_loss: 211.1657-> took 0.14 seconds\n",
"Epoch 131/1000 - loss: 278.1604 - val_loss: 209.9692-> took 0.13 seconds\n",
"Epoch 132/1000 - loss: 276.7426 - val_loss: 208.7788-> took 0.14 seconds\n",
"Epoch 133/1000 - loss: 275.3306 - val_loss: 207.5940-> took 0.13 seconds\n",
"Epoch 134/1000 - loss: 273.9244 - val_loss: 206.4150-> took 0.13 seconds\n",
"Epoch 135/1000 - loss: 272.5242 - val_loss: 205.2419-> took 0.14 seconds\n",
"Epoch 136/1000 - loss: 271.1296 - val_loss: 204.0744-> took 0.13 seconds\n",
"Epoch 137/1000 - loss: 269.7411 - val_loss: 202.9128-> took 0.13 seconds\n",
"Epoch 138/1000 - loss: 268.3584 - val_loss: 201.7571-> took 0.13 seconds\n",
"Epoch 139/1000 - loss: 266.9813 - val_loss: 200.6071-> took 0.13 seconds\n",
"Epoch 140/1000 - loss: 265.6102 - val_loss: 199.4629-> took 0.13 seconds\n",
"Epoch 141/1000 - loss: 264.2449 - val_loss: 198.3245-> took 0.13 seconds\n",
"Epoch 142/1000 - loss: 262.8855 - val_loss: 197.1919-> took 0.12 seconds\n",
"Epoch 143/1000 - loss: 261.5318 - val_loss: 196.0650-> took 0.13 seconds\n",
"Epoch 144/1000 - loss: 260.1841 - val_loss: 194.9440-> took 0.13 seconds\n",
"Epoch 145/1000 - loss: 258.8420 - val_loss: 193.8288-> took 0.15 seconds\n",
"Epoch 146/1000 - loss: 257.5058 - val_loss: 192.7193-> took 0.13 seconds\n",
"Epoch 147/1000 - loss: 256.1755 - val_loss: 191.6157-> took 0.13 seconds\n",
"Epoch 148/1000 - loss: 254.8510 - val_loss: 190.5178-> took 0.13 seconds\n",
"Epoch 149/1000 - loss: 253.5323 - val_loss: 189.4258-> took 0.13 seconds\n",
"Epoch 150/1000 - loss: 252.2195 - val_loss: 188.3395-> took 0.13 seconds\n",
"Epoch 151/1000 - loss: 250.9124 - val_loss: 187.2590-> took 0.13 seconds\n",
"Epoch 152/1000 - loss: 249.6112 - val_loss: 186.1842-> took 0.12 seconds\n",
"Epoch 153/1000 - loss: 248.3159 - val_loss: 185.1152-> took 0.13 seconds\n",
"Epoch 154/1000 - loss: 247.0264 - val_loss: 184.0521-> took 0.13 seconds\n",
"Epoch 155/1000 - loss: 245.7426 - val_loss: 182.9946-> took 0.13 seconds\n",
"Epoch 156/1000 - loss: 244.4647 - val_loss: 181.9431-> took 0.13 seconds\n",
"Epoch 157/1000 - loss: 243.1927 - val_loss: 180.8972-> took 0.13 seconds\n",
"Epoch 158/1000 - loss: 241.9265 - val_loss: 179.8573-> took 0.13 seconds\n",
"Epoch 159/1000 - loss: 240.6660 - val_loss: 178.8229-> took 0.13 seconds\n",
"Epoch 160/1000 - loss: 239.4115 - val_loss: 177.7946-> took 0.12 seconds\n",
"Epoch 161/1000 - loss: 238.1627 - val_loss: 176.7718-> took 0.13 seconds\n",
"Epoch 162/1000 - loss: 236.9199 - val_loss: 175.7550-> took 0.12 seconds\n",
"Epoch 163/1000 - loss: 235.6828 - val_loss: 174.7438-> took 0.13 seconds\n",
"Epoch 164/1000 - loss: 234.4515 - val_loss: 173.7384-> took 0.13 seconds\n",
"Epoch 165/1000 - loss: 233.2261 - val_loss: 172.7389-> took 0.12 seconds\n",
"Epoch 166/1000 - loss: 232.0066 - val_loss: 171.7450-> took 0.13 seconds\n",
"Epoch 167/1000 - loss: 230.7928 - val_loss: 170.7571-> took 0.13 seconds\n",
"Epoch 168/1000 - loss: 229.5849 - val_loss: 169.7749-> took 0.13 seconds\n",
"Epoch 169/1000 - loss: 228.3828 - val_loss: 168.7984-> took 0.14 seconds\n",
"Epoch 170/1000 - loss: 227.1865 - val_loss: 167.8277-> took 0.13 seconds\n",
"Epoch 171/1000 - loss: 225.9961 - val_loss: 166.8628-> took 0.14 seconds\n",
"Epoch 172/1000 - loss: 224.8115 - val_loss: 165.9037-> took 0.13 seconds\n",
"Epoch 173/1000 - loss: 223.6327 - val_loss: 164.9503-> took 0.12 seconds\n",
"Epoch 174/1000 - loss: 222.4598 - val_loss: 164.0028-> took 0.14 seconds\n",
"Epoch 175/1000 - loss: 221.2927 - val_loss: 163.0609-> took 0.13 seconds\n",
"Epoch 176/1000 - loss: 220.1315 - val_loss: 162.1249-> took 0.13 seconds\n",
"Epoch 177/1000 - loss: 218.9760 - val_loss: 161.1947-> took 0.12 seconds\n",
"Epoch 178/1000 - loss: 217.8264 - val_loss: 160.2702-> took 0.13 seconds\n",
"Epoch 179/1000 - loss: 216.6827 - val_loss: 159.3515-> took 0.12 seconds\n",
"Epoch 180/1000 - loss: 215.5447 - val_loss: 158.4385-> took 0.13 seconds\n",
"Epoch 181/1000 - loss: 214.4126 - val_loss: 157.5314-> took 0.12 seconds\n",
"Epoch 182/1000 - loss: 213.2864 - val_loss: 156.6299-> took 0.14 seconds\n",
"Epoch 183/1000 - loss: 212.1660 - val_loss: 155.7343-> took 0.13 seconds\n",
"Epoch 184/1000 - loss: 211.0514 - val_loss: 154.8445-> took 0.13 seconds\n",
"Epoch 185/1000 - loss: 209.9426 - val_loss: 153.9603-> took 0.12 seconds\n",
"Epoch 186/1000 - loss: 208.8398 - val_loss: 153.0820-> took 0.12 seconds\n",
"Epoch 187/1000 - loss: 207.7428 - val_loss: 152.2095-> took 0.13 seconds\n",
"Epoch 188/1000 - loss: 206.6515 - val_loss: 151.3427-> took 0.13 seconds\n",
"Epoch 189/1000 - loss: 205.5661 - val_loss: 150.4817-> took 0.12 seconds\n",
"Epoch 190/1000 - loss: 204.4866 - val_loss: 149.6264-> took 0.13 seconds\n",
"Epoch 191/1000 - loss: 203.4129 - val_loss: 148.7770-> took 0.12 seconds\n",
"Epoch 192/1000 - loss: 202.3451 - val_loss: 147.9332-> took 0.13 seconds\n",
"Epoch 193/1000 - loss: 201.2831 - val_loss: 147.0953-> took 0.13 seconds\n",
"Epoch 194/1000 - loss: 200.2269 - val_loss: 146.2631-> took 0.14 seconds\n",
"Epoch 195/1000 - loss: 199.1766 - val_loss: 145.4366-> took 0.14 seconds\n",
"Epoch 196/1000 - loss: 198.1322 - val_loss: 144.6160-> took 0.13 seconds\n",
"Epoch 197/1000 - loss: 197.0936 - val_loss: 143.8011-> took 0.12 seconds\n",
"Epoch 198/1000 - loss: 196.0609 - val_loss: 142.9921-> took 0.12 seconds\n",
"Epoch 199/1000 - loss: 195.0340 - val_loss: 142.1886-> took 0.12 seconds\n",
"Epoch 200/1000 - loss: 194.0129 - val_loss: 141.3911-> took 0.13 seconds\n",
"Epoch 201/1000 - loss: 192.9977 - val_loss: 140.5992-> took 0.13 seconds\n",
"Epoch 202/1000 - loss: 191.9884 - val_loss: 139.8131-> took 0.13 seconds\n",
"Epoch 203/1000 - loss: 190.9849 - val_loss: 139.0328-> took 0.13 seconds\n",
"Epoch 204/1000 - loss: 189.9873 - val_loss: 138.2582-> took 0.12 seconds\n",
"Epoch 205/1000 - loss: 188.9955 - val_loss: 137.4894-> took 0.12 seconds\n",
"Epoch 206/1000 - loss: 188.0096 - val_loss: 136.7263-> took 0.14 seconds\n",
"Epoch 207/1000 - loss: 187.0296 - val_loss: 135.9690-> took 0.13 seconds\n",
"Epoch 208/1000 - loss: 186.0554 - val_loss: 135.2175-> took 0.13 seconds\n",
"Epoch 209/1000 - loss: 185.0871 - val_loss: 134.4717-> took 0.13 seconds\n",
"Epoch 210/1000 - loss: 184.1246 - val_loss: 133.7317-> took 0.13 seconds\n",
"Epoch 211/1000 - loss: 183.1681 - val_loss: 132.9974-> took 0.12 seconds\n",
"Epoch 212/1000 - loss: 182.2174 - val_loss: 132.2689-> took 0.13 seconds\n",
"Epoch 213/1000 - loss: 181.2725 - val_loss: 131.5461-> took 0.13 seconds\n",
"Epoch 214/1000 - loss: 180.3335 - val_loss: 130.8291-> took 0.12 seconds\n",
"Epoch 215/1000 - loss: 179.4004 - val_loss: 130.1178-> took 0.13 seconds\n",
"Epoch 216/1000 - loss: 178.4732 - val_loss: 129.4124-> took 0.12 seconds\n",
"Epoch 217/1000 - loss: 177.5518 - val_loss: 128.7126-> took 0.15 seconds\n",
"Epoch 218/1000 - loss: 176.6363 - val_loss: 128.0186-> took 0.13 seconds\n",
"Epoch 219/1000 - loss: 175.7268 - val_loss: 127.3303-> took 0.12 seconds\n",
"Epoch 220/1000 - loss: 174.8230 - val_loss: 126.6478-> took 0.13 seconds\n",
"Epoch 221/1000 - loss: 173.9252 - val_loss: 125.9710-> took 0.12 seconds\n",
"Epoch 222/1000 - loss: 173.0333 - val_loss: 125.3000-> took 0.13 seconds\n",
"Epoch 223/1000 - loss: 172.1472 - val_loss: 124.6347-> took 0.13 seconds\n",
"Epoch 224/1000 - loss: 171.2671 - val_loss: 123.9752-> took 0.13 seconds\n",
"Epoch 225/1000 - loss: 170.3928 - val_loss: 123.3214-> took 0.13 seconds\n",
"Epoch 226/1000 - loss: 169.5244 - val_loss: 122.6733-> took 0.14 seconds\n",
"Epoch 227/1000 - loss: 168.6620 - val_loss: 122.0310-> took 0.13 seconds\n",
"Epoch 228/1000 - loss: 167.8054 - val_loss: 121.3944-> took 0.13 seconds\n",
"Epoch 229/1000 - loss: 166.9547 - val_loss: 120.7636-> took 0.13 seconds\n",
"Epoch 230/1000 - loss: 166.1099 - val_loss: 120.1385-> took 0.13 seconds\n",
"Epoch 231/1000 - loss: 165.2711 - val_loss: 119.5191-> took 0.14 seconds\n",
"Epoch 232/1000 - loss: 164.4381 - val_loss: 118.9055-> took 0.13 seconds\n",
"Epoch 233/1000 - loss: 163.6111 - val_loss: 118.2976-> took 0.13 seconds\n",
"Epoch 234/1000 - loss: 162.7900 - val_loss: 117.6954-> took 0.13 seconds\n",
"Epoch 235/1000 - loss: 161.9748 - val_loss: 117.0990-> took 0.13 seconds\n",
"Epoch 236/1000 - loss: 161.1655 - val_loss: 116.5083-> took 0.12 seconds\n",
"Epoch 237/1000 - loss: 160.3622 - val_loss: 115.9233-> took 0.13 seconds\n",
"Epoch 238/1000 - loss: 159.5648 - val_loss: 115.3440-> took 0.12 seconds\n",
"Epoch 239/1000 - loss: 158.7733 - val_loss: 114.7705-> took 0.12 seconds\n",
"Epoch 240/1000 - loss: 157.9877 - val_loss: 114.2027-> took 0.15 seconds\n",
"Epoch 241/1000 - loss: 157.2081 - val_loss: 113.6406-> took 0.13 seconds\n",
"Epoch 242/1000 - loss: 156.4345 - val_loss: 113.0843-> took 0.13 seconds\n",
"Epoch 243/1000 - loss: 155.6668 - val_loss: 112.5336-> took 0.13 seconds\n",
"Epoch 244/1000 - loss: 154.9050 - val_loss: 111.9887-> took 0.12 seconds\n",
"Epoch 245/1000 - loss: 154.1492 - val_loss: 111.4495-> took 0.13 seconds\n",
"Epoch 246/1000 - loss: 153.3994 - val_loss: 110.9160-> took 0.13 seconds\n",
"Epoch 247/1000 - loss: 152.6555 - val_loss: 110.3882-> took 0.13 seconds\n",
"Epoch 248/1000 - loss: 151.9176 - val_loss: 109.8661-> took 0.13 seconds\n",
"Epoch 249/1000 - loss: 151.1857 - val_loss: 109.3498-> took 0.13 seconds\n",
"Epoch 250/1000 - loss: 150.4597 - val_loss: 108.8391-> took 0.13 seconds\n",
"Epoch 251/1000 - loss: 149.7398 - val_loss: 108.3341-> took 0.13 seconds\n",
"Epoch 252/1000 - loss: 149.0258 - val_loss: 107.8349-> took 0.12 seconds\n",
"Epoch 253/1000 - loss: 148.3178 - val_loss: 107.3413-> took 0.13 seconds\n",
"Epoch 254/1000 - loss: 147.6159 - val_loss: 106.8535-> took 0.13 seconds\n",
"Epoch 255/1000 - loss: 146.9199 - val_loss: 106.3713-> took 0.12 seconds\n",
"Epoch 256/1000 - loss: 146.2300 - val_loss: 105.8948-> took 0.14 seconds\n",
"Epoch 257/1000 - loss: 145.5460 - val_loss: 105.4240-> took 0.13 seconds\n",
"Epoch 258/1000 - loss: 144.8681 - val_loss: 104.9590-> took 0.13 seconds\n",
"Epoch 259/1000 - loss: 144.1963 - val_loss: 104.4995-> took 0.12 seconds\n",
"Epoch 260/1000 - loss: 143.5304 - val_loss: 104.0458-> took 0.12 seconds\n",
"Epoch 261/1000 - loss: 142.8707 - val_loss: 103.5977-> took 0.15 seconds\n",
"Epoch 262/1000 - loss: 142.2169 - val_loss: 103.1553-> took 0.13 seconds\n",
"Epoch 263/1000 - loss: 141.5692 - val_loss: 102.7186-> took 0.14 seconds\n",
"Epoch 264/1000 - loss: 140.9276 - val_loss: 102.2875-> took 0.13 seconds\n",
"Epoch 265/1000 - loss: 140.2921 - val_loss: 101.8622-> took 0.12 seconds\n",
"Epoch 266/1000 - loss: 139.6627 - val_loss: 101.4424-> took 0.13 seconds\n",
"Epoch 267/1000 - loss: 139.0393 - val_loss: 101.0283-> took 0.13 seconds\n",
"Epoch 268/1000 - loss: 138.4221 - val_loss: 100.6199-> took 0.12 seconds\n",
"Epoch 269/1000 - loss: 137.8109 - val_loss: 100.2170-> took 0.14 seconds\n",
"Epoch 270/1000 - loss: 137.2059 - val_loss: 99.8199-> took 0.14 seconds\n",
"Epoch 271/1000 - loss: 136.6071 - val_loss: 99.4284-> took 0.12 seconds\n",
"Epoch 272/1000 - loss: 136.0143 - val_loss: 99.0425-> took 0.13 seconds\n",
"Epoch 273/1000 - loss: 135.4277 - val_loss: 98.6622-> took 0.13 seconds\n",
"Epoch 274/1000 - loss: 134.8473 - val_loss: 98.2875-> took 0.12 seconds\n",
"Epoch 275/1000 - loss: 134.2730 - val_loss: 97.9185-> took 0.13 seconds\n",
"Epoch 276/1000 - loss: 133.7049 - val_loss: 97.5550-> took 0.12 seconds\n",
"Epoch 277/1000 - loss: 133.1431 - val_loss: 97.1972-> took 0.13 seconds\n",
"Epoch 278/1000 - loss: 132.5873 - val_loss: 96.8449-> took 0.12 seconds\n",
"Epoch 279/1000 - loss: 132.0379 - val_loss: 96.4982-> took 0.14 seconds\n",
"Epoch 280/1000 - loss: 131.4946 - val_loss: 96.1571-> took 0.13 seconds\n",
"Epoch 281/1000 - loss: 130.9576 - val_loss: 95.8216-> took 0.13 seconds\n",
"Epoch 282/1000 - loss: 130.4268 - val_loss: 95.4916-> took 0.12 seconds\n",
"Epoch 283/1000 - loss: 129.9023 - val_loss: 95.1671-> took 0.12 seconds\n",
"Epoch 284/1000 - loss: 129.3840 - val_loss: 94.8482-> took 0.12 seconds\n",
"Epoch 285/1000 - loss: 128.8721 - val_loss: 94.5348-> took 0.14 seconds\n",
"Epoch 286/1000 - loss: 128.3665 - val_loss: 94.2269-> took 0.14 seconds\n",
"Epoch 287/1000 - loss: 127.8672 - val_loss: 93.9246-> took 0.14 seconds\n",
"Epoch 288/1000 - loss: 127.3742 - val_loss: 93.6276-> took 0.13 seconds\n",
"Epoch 289/1000 - loss: 126.8875 - val_loss: 93.3362-> took 0.12 seconds\n",
"Epoch 290/1000 - loss: 126.4072 - val_loss: 93.0502-> took 0.14 seconds\n",
"Epoch 291/1000 - loss: 125.9333 - val_loss: 92.7697-> took 0.13 seconds\n",
"Epoch 292/1000 - loss: 125.4658 - val_loss: 92.4947-> took 0.13 seconds\n",
"Epoch 293/1000 - loss: 125.0046 - val_loss: 92.2249-> took 0.14 seconds\n",
"Epoch 294/1000 - loss: 124.5499 - val_loss: 91.9607-> took 0.12 seconds\n",
"Epoch 295/1000 - loss: 124.1016 - val_loss: 91.7018-> took 0.13 seconds\n",
"Epoch 296/1000 - loss: 123.6597 - val_loss: 91.4482-> took 0.13 seconds\n",
"Epoch 297/1000 - loss: 123.2244 - val_loss: 91.2000-> took 0.13 seconds\n",
"Epoch 298/1000 - loss: 122.7954 - val_loss: 90.9571-> took 0.13 seconds\n",
"Epoch 299/1000 - loss: 122.3730 - val_loss: 90.7195-> took 0.13 seconds\n",
"Epoch 300/1000 - loss: 121.9570 - val_loss: 90.4871-> took 0.14 seconds\n",
"Epoch 301/1000 - loss: 121.5476 - val_loss: 90.2599-> took 0.12 seconds\n",
"Epoch 302/1000 - loss: 121.1446 - val_loss: 90.0380-> took 0.13 seconds\n",
"Epoch 303/1000 - loss: 120.7482 - val_loss: 89.8213-> took 0.12 seconds\n",
"Epoch 304/1000 - loss: 120.3583 - val_loss: 89.6097-> took 0.13 seconds\n",
"Epoch 305/1000 - loss: 119.9750 - val_loss: 89.4033-> took 0.13 seconds\n",
"Epoch 306/1000 - loss: 119.5983 - val_loss: 89.2019-> took 0.13 seconds\n",
"Epoch 307/1000 - loss: 119.2280 - val_loss: 89.0055-> took 0.13 seconds\n",
"Epoch 308/1000 - loss: 118.8644 - val_loss: 88.8142-> took 0.14 seconds\n",
"Epoch 309/1000 - loss: 118.5074 - val_loss: 88.6278-> took 0.13 seconds\n",
"Epoch 310/1000 - loss: 118.1569 - val_loss: 88.4464-> took 0.13 seconds\n",
"Epoch 311/1000 - loss: 117.8130 - val_loss: 88.2698-> took 0.13 seconds\n",
"Epoch 312/1000 - loss: 117.4757 - val_loss: 88.0981-> took 0.13 seconds\n",
"Epoch 313/1000 - loss: 117.1450 - val_loss: 87.9312-> took 0.12 seconds\n",
"Epoch 314/1000 - loss: 116.8208 - val_loss: 87.7690-> took 0.13 seconds\n",
"Epoch 315/1000 - loss: 116.5032 - val_loss: 87.6115-> took 0.14 seconds\n",
"Epoch 316/1000 - loss: 116.1921 - val_loss: 87.4587-> took 0.15 seconds\n",
"Epoch 317/1000 - loss: 115.8876 - val_loss: 87.3103-> took 0.13 seconds\n",
"Epoch 318/1000 - loss: 115.5896 - val_loss: 87.1666-> took 0.13 seconds\n",
"Epoch 319/1000 - loss: 115.2981 - val_loss: 87.0272-> took 0.13 seconds\n",
"Epoch 320/1000 - loss: 115.0131 - val_loss: 86.8923-> took 0.13 seconds\n",
"Epoch 321/1000 - loss: 114.7345 - val_loss: 86.7616-> took 0.13 seconds\n",
"Epoch 322/1000 - loss: 114.4624 - val_loss: 86.6352-> took 0.13 seconds\n",
"Epoch 323/1000 - loss: 114.1966 - val_loss: 86.5130-> took 0.13 seconds\n",
"Epoch 324/1000 - loss: 113.9372 - val_loss: 86.3949-> took 0.13 seconds\n",
"Epoch 325/1000 - loss: 113.6841 - val_loss: 86.2809-> took 0.14 seconds\n",
"Epoch 326/1000 - loss: 113.4373 - val_loss: 86.1707-> took 0.12 seconds\n",
"Epoch 327/1000 - loss: 113.1967 - val_loss: 86.0645-> took 0.13 seconds\n",
"Epoch 328/1000 - loss: 112.9622 - val_loss: 85.9620-> took 0.12 seconds\n",
"Epoch 329/1000 - loss: 112.7338 - val_loss: 85.8632-> took 0.13 seconds\n",
"Epoch 330/1000 - loss: 112.5115 - val_loss: 85.7680-> took 0.13 seconds\n",
"Epoch 331/1000 - loss: 112.2952 - val_loss: 85.6765-> took 0.13 seconds\n",
"Epoch 332/1000 - loss: 112.0847 - val_loss: 85.5883-> took 0.13 seconds\n",
"Epoch 333/1000 - loss: 111.8801 - val_loss: 85.5035-> took 0.12 seconds\n",
"Epoch 334/1000 - loss: 111.6812 - val_loss: 85.4219-> took 0.12 seconds\n",
"Epoch 335/1000 - loss: 111.4879 - val_loss: 85.3436-> took 0.14 seconds\n",
"Epoch 336/1000 - loss: 111.3003 - val_loss: 85.2683-> took 0.13 seconds\n",
"Epoch 337/1000 - loss: 111.1181 - val_loss: 85.1960-> took 0.13 seconds\n",
"Epoch 338/1000 - loss: 110.9413 - val_loss: 85.1266-> took 0.14 seconds\n",
"Epoch 339/1000 - loss: 110.7698 - val_loss: 85.0600-> took 0.14 seconds\n",
"Epoch 340/1000 - loss: 110.6035 - val_loss: 84.9962-> took 0.13 seconds\n",
"Epoch 341/1000 - loss: 110.4422 - val_loss: 84.9349-> took 0.13 seconds\n",
"Epoch 342/1000 - loss: 110.2860 - val_loss: 84.8763-> took 0.13 seconds\n",
"Epoch 343/1000 - loss: 110.1346 - val_loss: 84.8201-> took 0.13 seconds\n",
"Epoch 344/1000 - loss: 109.9880 - val_loss: 84.7662-> took 0.13 seconds\n",
"Epoch 345/1000 - loss: 109.8461 - val_loss: 84.7146-> took 0.13 seconds\n",
"Epoch 346/1000 - loss: 109.7087 - val_loss: 84.6653-> took 0.13 seconds\n",
"Epoch 347/1000 - loss: 109.5758 - val_loss: 84.6180-> took 0.12 seconds\n",
"Epoch 348/1000 - loss: 109.4472 - val_loss: 84.5728-> took 0.13 seconds\n",
"Epoch 349/1000 - loss: 109.3228 - val_loss: 84.5295-> took 0.13 seconds\n",
"Epoch 350/1000 - loss: 109.2025 - val_loss: 84.4881-> took 0.12 seconds\n",
"Epoch 351/1000 - loss: 109.0863 - val_loss: 84.4485-> took 0.13 seconds\n",
"Epoch 352/1000 - loss: 108.9739 - val_loss: 84.4107-> took 0.12 seconds\n",
"Epoch 353/1000 - loss: 108.8653 - val_loss: 84.3745-> took 0.14 seconds\n",
"Epoch 354/1000 - loss: 108.7604 - val_loss: 84.3399-> took 0.17 seconds\n",
"Epoch 355/1000 - loss: 108.6590 - val_loss: 84.3068-> took 0.13 seconds\n",
"Epoch 356/1000 - loss: 108.5611 - val_loss: 84.2752-> took 0.12 seconds\n",
"Epoch 357/1000 - loss: 108.4665 - val_loss: 84.2449-> took 0.13 seconds\n",
"Epoch 358/1000 - loss: 108.3751 - val_loss: 84.2159-> took 0.13 seconds\n",
"Epoch 359/1000 - loss: 108.2870 - val_loss: 84.1883-> took 0.12 seconds\n",
"Epoch 360/1000 - loss: 108.2019 - val_loss: 84.1619-> took 0.13 seconds\n",
"Epoch 361/1000 - loss: 108.1197 - val_loss: 84.1366-> took 0.13 seconds\n",
"Epoch 362/1000 - loss: 108.0404 - val_loss: 84.1125-> took 0.13 seconds\n",
"Epoch 363/1000 - loss: 107.9639 - val_loss: 84.0894-> took 0.13 seconds\n",
"Epoch 364/1000 - loss: 107.8900 - val_loss: 84.0673-> took 0.13 seconds\n",
"Epoch 365/1000 - loss: 107.8186 - val_loss: 84.0462-> took 0.13 seconds\n",
"Epoch 366/1000 - loss: 107.7498 - val_loss: 84.0260-> took 0.13 seconds\n",
"Epoch 367/1000 - loss: 107.6834 - val_loss: 84.0067-> took 0.12 seconds\n",
"Epoch 368/1000 - loss: 107.6194 - val_loss: 83.9882-> took 0.13 seconds\n",
"Epoch 369/1000 - loss: 107.5575 - val_loss: 83.9706-> took 0.14 seconds\n",
"Epoch 370/1000 - loss: 107.4979 - val_loss: 83.9537-> took 0.15 seconds\n",
"Epoch 371/1000 - loss: 107.4404 - val_loss: 83.9375-> took 0.14 seconds\n",
"Epoch 372/1000 - loss: 107.3848 - val_loss: 83.9220-> took 0.13 seconds\n",
"Epoch 373/1000 - loss: 107.3313 - val_loss: 83.9072-> took 0.12 seconds\n",
"Epoch 374/1000 - loss: 107.2796 - val_loss: 83.8930-> took 0.14 seconds\n",
"Epoch 375/1000 - loss: 107.2297 - val_loss: 83.8795-> took 0.12 seconds\n",
"Epoch 376/1000 - loss: 107.1816 - val_loss: 83.8665-> took 0.12 seconds\n",
"Epoch 377/1000 - loss: 107.1352 - val_loss: 83.8540-> took 0.14 seconds\n",
"Epoch 378/1000 - loss: 107.0905 - val_loss: 83.8421-> took 0.13 seconds\n",
"Epoch 379/1000 - loss: 107.0473 - val_loss: 83.8307-> took 0.12 seconds\n",
"Epoch 380/1000 - loss: 107.0056 - val_loss: 83.8197-> took 0.12 seconds\n",
"Epoch 381/1000 - loss: 106.9654 - val_loss: 83.8093-> took 0.12 seconds\n",
"Epoch 382/1000 - loss: 106.9266 - val_loss: 83.7992-> took 0.13 seconds\n",
"Epoch 383/1000 - loss: 106.8891 - val_loss: 83.7896-> took 0.13 seconds\n",
"Epoch 384/1000 - loss: 106.8530 - val_loss: 83.7804-> took 0.14 seconds\n",
"Epoch 385/1000 - loss: 106.8182 - val_loss: 83.7715-> took 0.13 seconds\n",
"Epoch 386/1000 - loss: 106.7845 - val_loss: 83.7630-> took 0.15 seconds\n",
"Epoch 387/1000 - loss: 106.7521 - val_loss: 83.7549-> took 0.14 seconds\n",
"Epoch 388/1000 - loss: 106.7208 - val_loss: 83.7471-> took 0.12 seconds\n",
"Epoch 389/1000 - loss: 106.6906 - val_loss: 83.7396-> took 0.12 seconds\n",
"Epoch 390/1000 - loss: 106.6614 - val_loss: 83.7324-> took 0.12 seconds\n",
"Epoch 391/1000 - loss: 106.6333 - val_loss: 83.7255-> took 0.13 seconds\n",
"Epoch 392/1000 - loss: 106.6061 - val_loss: 83.7189-> took 0.15 seconds\n",
"Epoch 393/1000 - loss: 106.5799 - val_loss: 83.7125-> took 0.13 seconds\n",
"Epoch 394/1000 - loss: 106.5546 - val_loss: 83.7064-> took 0.14 seconds\n",
"Epoch 395/1000 - loss: 106.5302 - val_loss: 83.7005-> took 0.12 seconds\n",
"Epoch 396/1000 - loss: 106.5067 - val_loss: 83.6950-> took 0.12 seconds\n",
"Epoch 397/1000 - loss: 106.4839 - val_loss: 83.6895-> took 0.13 seconds\n",
"Epoch 398/1000 - loss: 106.4620 - val_loss: 83.6843-> took 0.12 seconds\n",
"Epoch 399/1000 - loss: 106.4408 - val_loss: 83.6793-> took 0.13 seconds\n",
"Epoch 400/1000 - loss: 106.4204 - val_loss: 83.6745-> took 0.16 seconds\n",
"Epoch 401/1000 - loss: 106.4007 - val_loss: 83.6699-> took 0.13 seconds\n",
"Epoch 402/1000 - loss: 106.3816 - val_loss: 83.6655-> took 0.13 seconds\n",
"Epoch 403/1000 - loss: 106.3632 - val_loss: 83.6612-> took 0.13 seconds\n",
"Epoch 404/1000 - loss: 106.3455 - val_loss: 83.6571-> took 0.12 seconds\n",
"Epoch 405/1000 - loss: 106.3283 - val_loss: 83.6532-> took 0.12 seconds\n",
"Epoch 406/1000 - loss: 106.3118 - val_loss: 83.6494-> took 0.13 seconds\n",
"Epoch 407/1000 - loss: 106.2958 - val_loss: 83.6457-> took 0.13 seconds\n",
"Epoch 408/1000 - loss: 106.2804 - val_loss: 83.6422-> took 0.14 seconds\n",
"Epoch 409/1000 - loss: 106.2655 - val_loss: 83.6388-> took 0.13 seconds\n",
"Epoch 410/1000 - loss: 106.2511 - val_loss: 83.6356-> took 0.12 seconds\n",
"Epoch 411/1000 - loss: 106.2372 - val_loss: 83.6325-> took 0.14 seconds\n",
"Epoch 412/1000 - loss: 106.2239 - val_loss: 83.6294-> took 0.16 seconds\n",
"Epoch 413/1000 - loss: 106.2109 - val_loss: 83.6265-> took 0.14 seconds\n",
"Epoch 414/1000 - loss: 106.1984 - val_loss: 83.6237-> took 0.13 seconds\n",
"Epoch 415/1000 - loss: 106.1863 - val_loss: 83.6210-> took 0.14 seconds\n",
"Epoch 416/1000 - loss: 106.1747 - val_loss: 83.6184-> took 0.14 seconds\n",
"Epoch 417/1000 - loss: 106.1634 - val_loss: 83.6160-> took 0.12 seconds\n",
"Epoch 418/1000 - loss: 106.1526 - val_loss: 83.6135-> took 0.12 seconds\n",
"Epoch 419/1000 - loss: 106.1421 - val_loss: 83.6112-> took 0.13 seconds\n",
"Epoch 420/1000 - loss: 106.1319 - val_loss: 83.6090-> took 0.12 seconds\n",
"Epoch 421/1000 - loss: 106.1222 - val_loss: 83.6068-> took 0.12 seconds\n",
"Epoch 422/1000 - loss: 106.1127 - val_loss: 83.6048-> took 0.12 seconds\n",
"Epoch 423/1000 - loss: 106.1036 - val_loss: 83.6028-> took 0.13 seconds\n",
"Epoch 424/1000 - loss: 106.0948 - val_loss: 83.6008-> took 0.12 seconds\n",
"Epoch 425/1000 - loss: 106.0863 - val_loss: 83.5990-> took 0.13 seconds\n",
"Epoch 426/1000 - loss: 106.0781 - val_loss: 83.5972-> took 0.13 seconds\n",
"Epoch 427/1000 - loss: 106.0701 - val_loss: 83.5955-> took 0.12 seconds\n",
"Epoch 428/1000 - loss: 106.0625 - val_loss: 83.5938-> took 0.14 seconds\n",
"Epoch 429/1000 - loss: 106.0551 - val_loss: 83.5922-> took 0.14 seconds\n",
"Epoch 430/1000 - loss: 106.0479 - val_loss: 83.5907-> took 0.13 seconds\n",
"Epoch 431/1000 - loss: 106.0410 - val_loss: 83.5892-> took 0.13 seconds\n",
"Epoch 432/1000 - loss: 106.0343 - val_loss: 83.5878-> took 0.13 seconds\n",
"Epoch 433/1000 - loss: 106.0279 - val_loss: 83.5864-> took 0.13 seconds\n",
"Epoch 434/1000 - loss: 106.0217 - val_loss: 83.5850-> took 0.13 seconds\n",
"Epoch 435/1000 - loss: 106.0156 - val_loss: 83.5838-> took 0.13 seconds\n",
"Epoch 436/1000 - loss: 106.0098 - val_loss: 83.5825-> took 0.14 seconds\n",
"Epoch 437/1000 - loss: 106.0042 - val_loss: 83.5813-> took 0.13 seconds\n",
"Epoch 438/1000 - loss: 105.9988 - val_loss: 83.5802-> took 0.14 seconds\n",
"Epoch 439/1000 - loss: 105.9936 - val_loss: 83.5790-> took 0.14 seconds\n",
"Epoch 440/1000 - loss: 105.9885 - val_loss: 83.5780-> took 0.14 seconds\n",
"Epoch 441/1000 - loss: 105.9836 - val_loss: 83.5770-> took 0.12 seconds\n",
"Epoch 442/1000 - loss: 105.9789 - val_loss: 83.5759-> took 0.13 seconds\n",
"Epoch 443/1000 - loss: 105.9743 - val_loss: 83.5750-> took 0.12 seconds\n",
"Epoch 444/1000 - loss: 105.9700 - val_loss: 83.5740-> took 0.14 seconds\n",
"Epoch 445/1000 - loss: 105.9657 - val_loss: 83.5731-> took 0.13 seconds\n",
"Epoch 446/1000 - loss: 105.9616 - val_loss: 83.5723-> took 0.13 seconds\n",
"Epoch 447/1000 - loss: 105.9576 - val_loss: 83.5714-> took 0.13 seconds\n",
"Epoch 448/1000 - loss: 105.9538 - val_loss: 83.5706-> took 0.13 seconds\n",
"Epoch 449/1000 - loss: 105.9501 - val_loss: 83.5699-> took 0.12 seconds\n",
"Epoch 450/1000 - loss: 105.9465 - val_loss: 83.5691-> took 0.14 seconds\n",
"Epoch 451/1000 - loss: 105.9430 - val_loss: 83.5684-> took 0.13 seconds\n",
"Epoch 452/1000 - loss: 105.9397 - val_loss: 83.5677-> took 0.13 seconds\n",
"Epoch 453/1000 - loss: 105.9365 - val_loss: 83.5670-> took 0.13 seconds\n",
"Epoch 454/1000 - loss: 105.9334 - val_loss: 83.5663-> took 0.13 seconds\n",
"Epoch 455/1000 - loss: 105.9304 - val_loss: 83.5657-> took 0.13 seconds\n",
"Epoch 456/1000 - loss: 105.9274 - val_loss: 83.5651-> took 0.12 seconds\n",
"Epoch 457/1000 - loss: 105.9246 - val_loss: 83.5645-> took 0.13 seconds\n",
"Epoch 458/1000 - loss: 105.9219 - val_loss: 83.5639-> took 0.12 seconds\n",
"Epoch 459/1000 - loss: 105.9193 - val_loss: 83.5634-> took 0.13 seconds\n",
"Epoch 460/1000 - loss: 105.9167 - val_loss: 83.5629-> took 0.13 seconds\n",
"Epoch 461/1000 - loss: 105.9143 - val_loss: 83.5624-> took 0.14 seconds\n",
"Epoch 462/1000 - loss: 105.9119 - val_loss: 83.5619-> took 0.13 seconds\n",
"Epoch 463/1000 - loss: 105.9096 - val_loss: 83.5614-> took 0.14 seconds\n",
"Epoch 464/1000 - loss: 105.9074 - val_loss: 83.5609-> took 0.12 seconds\n",
"Epoch 465/1000 - loss: 105.9053 - val_loss: 83.5605-> took 0.15 seconds\n",
"Epoch 466/1000 - loss: 105.9032 - val_loss: 83.5601-> took 0.14 seconds\n",
"Epoch 467/1000 - loss: 105.9012 - val_loss: 83.5596-> took 0.13 seconds\n",
"Epoch 468/1000 - loss: 105.8993 - val_loss: 83.5593-> took 0.13 seconds\n",
"Epoch 469/1000 - loss: 105.8974 - val_loss: 83.5589-> took 0.14 seconds\n",
"Epoch 470/1000 - loss: 105.8956 - val_loss: 83.5585-> took 0.14 seconds\n",
"Epoch 471/1000 - loss: 105.8939 - val_loss: 83.5581-> took 0.13 seconds\n",
"Epoch 472/1000 - loss: 105.8922 - val_loss: 83.5578-> took 0.13 seconds\n",
"Epoch 473/1000 - loss: 105.8906 - val_loss: 83.5574-> took 0.14 seconds\n",
"Epoch 474/1000 - loss: 105.8890 - val_loss: 83.5571-> took 0.13 seconds\n",
"Epoch 475/1000 - loss: 105.8875 - val_loss: 83.5568-> took 0.14 seconds\n",
"Epoch 476/1000 - loss: 105.8861 - val_loss: 83.5565-> took 0.14 seconds\n",
"Epoch 477/1000 - loss: 105.8846 - val_loss: 83.5562-> took 0.13 seconds\n",
"Epoch 478/1000 - loss: 105.8833 - val_loss: 83.5559-> took 0.13 seconds\n",
"Epoch 479/1000 - loss: 105.8820 - val_loss: 83.5557-> took 0.13 seconds\n",
"Epoch 480/1000 - loss: 105.8807 - val_loss: 83.5554-> took 0.12 seconds\n",
"Epoch 481/1000 - loss: 105.8794 - val_loss: 83.5551-> took 0.12 seconds\n",
"Epoch 482/1000 - loss: 105.8783 - val_loss: 83.5549-> took 0.13 seconds\n",
"Epoch 483/1000 - loss: 105.8771 - val_loss: 83.5547-> took 0.13 seconds\n",
"Epoch 484/1000 - loss: 105.8760 - val_loss: 83.5544-> took 0.13 seconds\n",
"Epoch 485/1000 - loss: 105.8749 - val_loss: 83.5542-> took 0.12 seconds\n",
"Epoch 486/1000 - loss: 105.8739 - val_loss: 83.5540-> took 0.13 seconds\n",
"Epoch 487/1000 - loss: 105.8729 - val_loss: 83.5538-> took 0.13 seconds\n",
"Epoch 488/1000 - loss: 105.8719 - val_loss: 83.5536-> took 0.12 seconds\n",
"Epoch 489/1000 - loss: 105.8710 - val_loss: 83.5534-> took 0.13 seconds\n",
"Epoch 490/1000 - loss: 105.8701 - val_loss: 83.5532-> took 0.14 seconds\n",
"Epoch 491/1000 - loss: 105.8692 - val_loss: 83.5530-> took 0.12 seconds\n",
"Epoch 492/1000 - loss: 105.8684 - val_loss: 83.5528-> took 0.14 seconds\n",
"Epoch 493/1000 - loss: 105.8676 - val_loss: 83.5527-> took 0.14 seconds\n",
"Epoch 494/1000 - loss: 105.8668 - val_loss: 83.5525-> took 0.12 seconds\n",
"Epoch 495/1000 - loss: 105.8660 - val_loss: 83.5524-> took 0.13 seconds\n",
"Epoch 496/1000 - loss: 105.8653 - val_loss: 83.5522-> took 0.13 seconds\n",
"Epoch 497/1000 - loss: 105.8646 - val_loss: 83.5521-> took 0.12 seconds\n",
"Epoch 498/1000 - loss: 105.8639 - val_loss: 83.5519-> took 0.13 seconds\n",
"Epoch 499/1000 - loss: 105.8632 - val_loss: 83.5518-> took 0.13 seconds\n",
"Epoch 500/1000 - loss: 105.8626 - val_loss: 83.5517-> took 0.14 seconds\n",
"Epoch 501/1000 - loss: 105.8619 - val_loss: 83.5515-> took 0.12 seconds\n",
"Epoch 502/1000 - loss: 105.8614 - val_loss: 83.5514-> took 0.12 seconds\n",
"Epoch 503/1000 - loss: 105.8608 - val_loss: 83.5513-> took 0.12 seconds\n",
"Epoch 504/1000 - loss: 105.8602 - val_loss: 83.5512-> took 0.13 seconds\n",
"Epoch 505/1000 - loss: 105.8597 - val_loss: 83.5510-> took 0.15 seconds\n",
"Epoch 506/1000 - loss: 105.8591 - val_loss: 83.5510-> took 0.13 seconds\n",
"Epoch 507/1000 - loss: 105.8587 - val_loss: 83.5508-> took 0.14 seconds\n",
"Epoch 508/1000 - loss: 105.8582 - val_loss: 83.5508-> took 0.12 seconds\n",
"Epoch 509/1000 - loss: 105.8577 - val_loss: 83.5507-> took 0.12 seconds\n",
"Epoch 510/1000 - loss: 105.8572 - val_loss: 83.5506-> took 0.14 seconds\n",
"Epoch 511/1000 - loss: 105.8568 - val_loss: 83.5505-> took 0.14 seconds\n",
"Epoch 512/1000 - loss: 105.8564 - val_loss: 83.5504-> took 0.13 seconds\n",
"Epoch 513/1000 - loss: 105.8560 - val_loss: 83.5503-> took 0.13 seconds\n",
"Epoch 514/1000 - loss: 105.8556 - val_loss: 83.5502-> took 0.13 seconds\n",
"Epoch 515/1000 - loss: 105.8552 - val_loss: 83.5501-> took 0.14 seconds\n",
"Epoch 516/1000 - loss: 105.8548 - val_loss: 83.5501-> took 0.13 seconds\n",
"Epoch 517/1000 - loss: 105.8545 - val_loss: 83.5500-> took 0.13 seconds\n",
"Epoch 518/1000 - loss: 105.8541 - val_loss: 83.5499-> took 0.13 seconds\n",
"Epoch 519/1000 - loss: 105.8538 - val_loss: 83.5499-> took 0.14 seconds\n",
"Epoch 520/1000 - loss: 105.8534 - val_loss: 83.5498-> took 0.14 seconds\n",
"Epoch 521/1000 - loss: 105.8531 - val_loss: 83.5497-> took 0.12 seconds\n",
"Epoch 522/1000 - loss: 105.8528 - val_loss: 83.5497-> took 0.13 seconds\n",
"Epoch 523/1000 - loss: 105.8525 - val_loss: 83.5496-> took 0.13 seconds\n",
"Epoch 524/1000 - loss: 105.8523 - val_loss: 83.5496-> took 0.12 seconds\n",
"Epoch 525/1000 - loss: 105.8520 - val_loss: 83.5495-> took 0.13 seconds\n",
"Epoch 526/1000 - loss: 105.8517 - val_loss: 83.5494-> took 0.13 seconds\n",
"Epoch 527/1000 - loss: 105.8515 - val_loss: 83.5494-> took 0.13 seconds\n",
"Epoch 528/1000 - loss: 105.8512 - val_loss: 83.5493-> took 0.13 seconds\n",
"Epoch 529/1000 - loss: 105.8510 - val_loss: 83.5493-> took 0.13 seconds\n",
"Epoch 530/1000 - loss: 105.8508 - val_loss: 83.5492-> took 0.14 seconds\n",
"Epoch 531/1000 - loss: 105.8505 - val_loss: 83.5492-> took 0.12 seconds\n",
"Epoch 532/1000 - loss: 105.8503 - val_loss: 83.5491-> took 0.13 seconds\n",
"Epoch 533/1000 - loss: 105.8501 - val_loss: 83.5491-> took 0.14 seconds\n",
"Epoch 534/1000 - loss: 105.8499 - val_loss: 83.5491-> took 0.13 seconds\n",
"Epoch 535/1000 - loss: 105.8497 - val_loss: 83.5490-> took 0.13 seconds\n",
"Epoch 536/1000 - loss: 105.8495 - val_loss: 83.5490-> took 0.13 seconds\n",
"Epoch 537/1000 - loss: 105.8494 - val_loss: 83.5489-> took 0.13 seconds\n",
"Epoch 538/1000 - loss: 105.8492 - val_loss: 83.5489-> took 0.13 seconds\n",
"Epoch 539/1000 - loss: 105.8490 - val_loss: 83.5489-> took 0.13 seconds\n",
"Epoch 540/1000 - loss: 105.8489 - val_loss: 83.5489-> took 0.14 seconds\n",
"Epoch 541/1000 - loss: 105.8487 - val_loss: 83.5488-> took 0.14 seconds\n",
"Epoch 542/1000 - loss: 105.8486 - val_loss: 83.5488-> took 0.12 seconds\n",
"Epoch 543/1000 - loss: 105.8484 - val_loss: 83.5488-> took 0.13 seconds\n",
"Epoch 544/1000 - loss: 105.8483 - val_loss: 83.5487-> took 0.12 seconds\n",
"Epoch 545/1000 - loss: 105.8481 - val_loss: 83.5487-> took 0.14 seconds\n",
"Epoch 546/1000 - loss: 105.8480 - val_loss: 83.5487-> took 0.13 seconds\n",
"Epoch 547/1000 - loss: 105.8478 - val_loss: 83.5486-> took 0.13 seconds\n",
"Epoch 548/1000 - loss: 105.8477 - val_loss: 83.5486-> took 0.12 seconds\n",
"Epoch 549/1000 - loss: 105.8475 - val_loss: 83.5486-> took 0.14 seconds\n",
"Epoch 550/1000 - loss: 105.8474 - val_loss: 83.5486-> took 0.13 seconds\n",
"Epoch 551/1000 - loss: 105.8473 - val_loss: 83.5485-> took 0.12 seconds\n",
"Epoch 552/1000 - loss: 105.8472 - val_loss: 83.5485-> took 0.13 seconds\n",
"Epoch 553/1000 - loss: 105.8472 - val_loss: 83.5485-> took 0.13 seconds\n",
"Epoch 554/1000 - loss: 105.8470 - val_loss: 83.5485-> took 0.12 seconds\n",
"Epoch 555/1000 - loss: 105.8469 - val_loss: 83.5485-> took 0.13 seconds\n",
"Epoch 556/1000 - loss: 105.8469 - val_loss: 83.5485-> took 0.13 seconds\n",
"Epoch 557/1000 - loss: 105.8468 - val_loss: 83.5484-> took 0.14 seconds\n",
"Epoch 558/1000 - loss: 105.8467 - val_loss: 83.5484-> took 0.13 seconds\n",
"Epoch 559/1000 - loss: 105.8466 - val_loss: 83.5484-> took 0.13 seconds\n",
"Epoch 560/1000 - loss: 105.8466 - val_loss: 83.5484-> took 0.13 seconds\n",
"Epoch 561/1000 - loss: 105.8465 - val_loss: 83.5484-> took 0.13 seconds\n",
"Epoch 562/1000 - loss: 105.8464 - val_loss: 83.5483-> took 0.13 seconds\n",
"Epoch 563/1000 - loss: 105.8463 - val_loss: 83.5483-> took 0.12 seconds\n",
"Epoch 564/1000 - loss: 105.8462 - val_loss: 83.5483-> took 0.13 seconds\n",
"Epoch 565/1000 - loss: 105.8461 - val_loss: 83.5483-> took 0.12 seconds\n",
"Epoch 566/1000 - loss: 105.8461 - val_loss: 83.5483-> took 0.13 seconds\n",
"Epoch 567/1000 - loss: 105.8460 - val_loss: 83.5483-> took 0.12 seconds\n",
"Epoch 568/1000 - loss: 105.8459 - val_loss: 83.5482-> took 0.13 seconds\n",
"Epoch 569/1000 - loss: 105.8458 - val_loss: 83.5482-> took 0.15 seconds\n",
"Epoch 570/1000 - loss: 105.8458 - val_loss: 83.5482-> took 0.15 seconds\n",
"Epoch 571/1000 - loss: 105.8457 - val_loss: 83.5482-> took 0.13 seconds\n",
"Epoch 572/1000 - loss: 105.8457 - val_loss: 83.5482-> took 0.13 seconds\n",
"Epoch 573/1000 - loss: 105.8456 - val_loss: 83.5482-> took 0.15 seconds\n",
"Epoch 574/1000 - loss: 105.8456 - val_loss: 83.5482-> took 0.13 seconds\n",
"Epoch 575/1000 - loss: 105.8456 - val_loss: 83.5482-> took 0.13 seconds\n",
"Epoch 576/1000 - loss: 105.8455 - val_loss: 83.5481-> took 0.13 seconds\n",
"Epoch 577/1000 - loss: 105.8455 - val_loss: 83.5481-> took 0.13 seconds\n",
"Epoch 578/1000 - loss: 105.8454 - val_loss: 83.5481-> took 0.13 seconds\n",
"Epoch 579/1000 - loss: 105.8454 - val_loss: 83.5481-> took 0.14 seconds\n",
"Epoch 580/1000 - loss: 105.8453 - val_loss: 83.5481-> took 0.13 seconds\n",
"Epoch 581/1000 - loss: 105.8453 - val_loss: 83.5481-> took 0.13 seconds\n",
"Epoch 582/1000 - loss: 105.8452 - val_loss: 83.5481-> took 0.14 seconds\n",
"Epoch 583/1000 - loss: 105.8452 - val_loss: 83.5481-> took 0.14 seconds\n",
"Epoch 584/1000 - loss: 105.8451 - val_loss: 83.5481-> took 0.14 seconds\n",
"Epoch 585/1000 - loss: 105.8451 - val_loss: 83.5481-> took 0.13 seconds\n",
"Epoch 586/1000 - loss: 105.8451 - val_loss: 83.5481-> took 0.14 seconds\n",
"Epoch 587/1000 - loss: 105.8450 - val_loss: 83.5481-> took 0.14 seconds\n",
"Epoch 588/1000 - loss: 105.8450 - val_loss: 83.5481-> took 0.14 seconds\n",
"Epoch 589/1000 - loss: 105.8450 - val_loss: 83.5480-> took 0.13 seconds\n",
"Epoch 590/1000 - loss: 105.8449 - val_loss: 83.5480-> took 0.13 seconds\n",
"Epoch 591/1000 - loss: 105.8449 - val_loss: 83.5480-> took 0.13 seconds\n",
"Epoch 592/1000 - loss: 105.8449 - val_loss: 83.5480-> took 0.14 seconds\n",
"Epoch 593/1000 - loss: 105.8448 - val_loss: 83.5480-> took 0.14 seconds\n",
"Epoch 594/1000 - loss: 105.8448 - val_loss: 83.5480-> took 0.13 seconds\n",
"Epoch 595/1000 - loss: 105.8448 - val_loss: 83.5480-> took 0.13 seconds\n",
"Epoch 596/1000 - loss: 105.8447 - val_loss: 83.5480-> took 0.13 seconds\n",
"Epoch 597/1000 - loss: 105.8447 - val_loss: 83.5480-> took 0.13 seconds\n",
"Epoch 598/1000 - loss: 105.8447 - val_loss: 83.5480-> took 0.13 seconds\n",
"Epoch 599/1000 - loss: 105.8446 - val_loss: 83.5480-> took 0.13 seconds\n",
"Epoch 600/1000 - loss: 105.8446 - val_loss: 83.5480-> took 0.13 seconds\n",
"Epoch 601/1000 - loss: 105.8446 - val_loss: 83.5480-> took 0.12 seconds\n",
"Epoch 602/1000 - loss: 105.8446 - val_loss: 83.5480-> took 0.13 seconds\n",
"Epoch 603/1000 - loss: 105.8446 - val_loss: 83.5480-> took 0.12 seconds\n",
"Epoch 604/1000 - loss: 105.8446 - val_loss: 83.5480-> took 0.13 seconds\n",
"Epoch 605/1000 - loss: 105.8446 - val_loss: 83.5480-> took 0.14 seconds\n",
"Epoch 606/1000 - loss: 105.8446 - val_loss: 83.5480-> took 0.12 seconds\n",
"Epoch 607/1000 - loss: 105.8446 - val_loss: 83.5480-> took 0.14 seconds\n",
"Epoch 608/1000 - loss: 105.8446 - val_loss: 83.5480-> took 0.13 seconds\n",
"Epoch 609/1000 - loss: 105.8446 - val_loss: 83.5480-> took 0.13 seconds\n",
"Epoch 610/1000 - loss: 105.8446 - val_loss: 83.5480-> took 0.13 seconds\n",
"Epoch 611/1000 - loss: 105.8446 - val_loss: 83.5480-> took 0.13 seconds\n",
"Epoch 612/1000 - loss: 105.8446 - val_loss: 83.5480-> took 0.12 seconds\n",
"Epoch 613/1000 - loss: 105.8446 - val_loss: 83.5480-> took 0.13 seconds\n",
"Epoch 614/1000 - loss: 105.8446 - val_loss: 83.5480-> took 0.12 seconds\n",
"Epoch 615/1000 - loss: 105.8446 - val_loss: 83.5480-> took 0.13 seconds\n",
"Epoch 616/1000 - loss: 105.8445 - val_loss: 83.5480-> took 0.14 seconds\n",
"Epoch 617/1000 - loss: 105.8445 - val_loss: 83.5480-> took 0.15 seconds\n",
"Epoch 618/1000 - loss: 105.8445 - val_loss: 83.5480-> took 0.14 seconds\n",
"Epoch 619/1000 - loss: 105.8445 - val_loss: 83.5480-> took 0.13 seconds\n",
"Epoch 620/1000 - loss: 105.8445 - val_loss: 83.5480-> took 0.13 seconds\n",
"Epoch 621/1000 - loss: 105.8445 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 622/1000 - loss: 105.8444 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 623/1000 - loss: 105.8444 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 624/1000 - loss: 105.8444 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 625/1000 - loss: 105.8444 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 626/1000 - loss: 105.8444 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 627/1000 - loss: 105.8444 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 628/1000 - loss: 105.8444 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 629/1000 - loss: 105.8444 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 630/1000 - loss: 105.8444 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 631/1000 - loss: 105.8444 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 632/1000 - loss: 105.8443 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 633/1000 - loss: 105.8443 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 634/1000 - loss: 105.8443 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 635/1000 - loss: 105.8443 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 636/1000 - loss: 105.8443 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 637/1000 - loss: 105.8443 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 638/1000 - loss: 105.8443 - val_loss: 83.5479-> took 0.15 seconds\n",
"Epoch 639/1000 - loss: 105.8443 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 640/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 641/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 642/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 643/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 644/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 645/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 646/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.15 seconds\n",
"Epoch 647/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 648/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 649/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 650/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 651/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 652/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 653/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 654/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 655/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 656/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 657/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 658/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 659/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 660/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 661/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 662/1000 - loss: 105.8443 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 663/1000 - loss: 105.8443 - val_loss: 83.5479-> took 0.15 seconds\n",
"Epoch 664/1000 - loss: 105.8443 - val_loss: 83.5479-> took 0.15 seconds\n",
"Epoch 665/1000 - loss: 105.8443 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 666/1000 - loss: 105.8443 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 667/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 668/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 669/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 670/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 671/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 672/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 673/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 674/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 675/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 676/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 677/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 678/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 679/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 680/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 681/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 682/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 683/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 684/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 685/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 686/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 687/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 688/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 689/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 690/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 691/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 692/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 693/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.15 seconds\n",
"Epoch 694/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 695/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.15 seconds\n",
"Epoch 696/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 697/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 698/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.15 seconds\n",
"Epoch 699/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.15 seconds\n",
"Epoch 700/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 701/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 702/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.16 seconds\n",
"Epoch 703/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.15 seconds\n",
"Epoch 704/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 705/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 706/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 707/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 708/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 709/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 710/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 711/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 712/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 713/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.15 seconds\n",
"Epoch 714/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 715/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 716/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 717/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.15 seconds\n",
"Epoch 718/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 719/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 720/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 721/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 722/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.16 seconds\n",
"Epoch 723/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 724/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 725/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 726/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 727/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 728/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 729/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 730/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 731/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 732/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 733/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 734/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 735/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 736/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 737/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 738/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 739/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 740/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 741/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 742/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 743/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 744/1000 - loss: 105.8443 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 745/1000 - loss: 105.8443 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 746/1000 - loss: 105.8443 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 747/1000 - loss: 105.8443 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 748/1000 - loss: 105.8443 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 749/1000 - loss: 105.8443 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 750/1000 - loss: 105.8443 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 751/1000 - loss: 105.8443 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 752/1000 - loss: 105.8443 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 753/1000 - loss: 105.8443 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 754/1000 - loss: 105.8443 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 755/1000 - loss: 105.8443 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 756/1000 - loss: 105.8443 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 757/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 758/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 759/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 760/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.15 seconds\n",
"Epoch 761/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 762/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 763/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 764/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 765/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 766/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 767/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 768/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.15 seconds\n",
"Epoch 769/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 770/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 771/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 772/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 773/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 774/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 775/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 776/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 777/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 778/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 779/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 780/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 781/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 782/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 783/1000 - loss: 105.8440 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 784/1000 - loss: 105.8440 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 785/1000 - loss: 105.8440 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 786/1000 - loss: 105.8440 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 787/1000 - loss: 105.8440 - val_loss: 83.5478-> took 0.12 seconds\n",
"Epoch 788/1000 - loss: 105.8440 - val_loss: 83.5478-> took 0.14 seconds\n",
"Epoch 789/1000 - loss: 105.8440 - val_loss: 83.5478-> took 0.14 seconds\n",
"Epoch 790/1000 - loss: 105.8440 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 791/1000 - loss: 105.8440 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 792/1000 - loss: 105.8440 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 793/1000 - loss: 105.8440 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 794/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 795/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 796/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 797/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.15 seconds\n",
"Epoch 798/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 799/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.15 seconds\n",
"Epoch 800/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 801/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 802/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 803/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 804/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 805/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 806/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 807/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 808/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 809/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 810/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 811/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 812/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 813/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 814/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 815/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 816/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 817/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 818/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 819/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 820/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 821/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 822/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 823/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 824/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 825/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 826/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 827/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 828/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 829/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 830/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 831/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 832/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 833/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 834/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 835/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 836/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 837/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 838/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 839/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 840/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 841/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 842/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 843/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 844/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 845/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 846/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 847/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 848/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 849/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 850/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 851/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 852/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 853/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 854/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 855/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 856/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 857/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 858/1000 - loss: 105.8440 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 859/1000 - loss: 105.8440 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 860/1000 - loss: 105.8440 - val_loss: 83.5479-> took 0.15 seconds\n",
"Epoch 861/1000 - loss: 105.8440 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 862/1000 - loss: 105.8440 - val_loss: 83.5479-> took 0.16 seconds\n",
"Epoch 863/1000 - loss: 105.8440 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 864/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 865/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 866/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.15 seconds\n",
"Epoch 867/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 868/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 869/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 870/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 871/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 872/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 873/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 874/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 875/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 876/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 877/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 878/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 879/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 880/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 881/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 882/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 883/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 884/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 885/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 886/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 887/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 888/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 889/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 890/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 891/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 892/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 893/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 894/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 895/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 896/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 897/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.15 seconds\n",
"Epoch 898/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 899/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 900/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 901/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 902/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 903/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 904/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 905/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 906/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 907/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 908/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 909/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 910/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 911/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 912/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 913/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 914/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 915/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 916/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.15 seconds\n",
"Epoch 917/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.15 seconds\n",
"Epoch 918/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 919/1000 - loss: 105.8440 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 920/1000 - loss: 105.8440 - val_loss: 83.5479-> took 0.15 seconds\n",
"Epoch 921/1000 - loss: 105.8440 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 922/1000 - loss: 105.8440 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 923/1000 - loss: 105.8440 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 924/1000 - loss: 105.8440 - val_loss: 83.5478-> took 0.15 seconds\n",
"Epoch 925/1000 - loss: 105.8440 - val_loss: 83.5478-> took 0.14 seconds\n",
"Epoch 926/1000 - loss: 105.8440 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 927/1000 - loss: 105.8440 - val_loss: 83.5478-> took 0.13 seconds\n",
"Epoch 928/1000 - loss: 105.8439 - val_loss: 83.5478-> took 0.14 seconds\n",
"Epoch 929/1000 - loss: 105.8439 - val_loss: 83.5478-> took 0.13 seconds\n",
"Epoch 930/1000 - loss: 105.8439 - val_loss: 83.5478-> took 0.13 seconds\n",
"Epoch 931/1000 - loss: 105.8439 - val_loss: 83.5478-> took 0.12 seconds\n",
"Epoch 932/1000 - loss: 105.8440 - val_loss: 83.5478-> took 0.14 seconds\n",
"Epoch 933/1000 - loss: 105.8440 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 934/1000 - loss: 105.8440 - val_loss: 83.5478-> took 0.14 seconds\n",
"Epoch 935/1000 - loss: 105.8440 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 936/1000 - loss: 105.8440 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 937/1000 - loss: 105.8440 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 938/1000 - loss: 105.8440 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 939/1000 - loss: 105.8440 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 940/1000 - loss: 105.8440 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 941/1000 - loss: 105.8440 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 942/1000 - loss: 105.8440 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 943/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 944/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 945/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 946/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 947/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 948/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 949/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.15 seconds\n",
"Epoch 950/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 951/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 952/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 953/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 954/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 955/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 956/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 957/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 958/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 959/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 960/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 961/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 962/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 963/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 964/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 965/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 966/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 967/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 968/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 969/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 970/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 971/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 972/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.15 seconds\n",
"Epoch 973/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 974/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 975/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 976/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 977/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 978/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 979/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 980/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 981/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 982/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 983/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 984/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 985/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 986/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 987/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 988/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 989/1000 - loss: 105.8442 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 990/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 991/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 992/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 993/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 994/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 995/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 996/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.13 seconds\n",
"Epoch 997/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.12 seconds\n",
"Epoch 998/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 999/1000 - loss: 105.8441 - val_loss: 83.5479-> took 0.14 seconds\n",
"Epoch 1000/1000 - loss: 105.8440 - val_loss: 83.5479-> took 0.13 seconds\n"
],
"name": "stdout"
}
]
},
{
"cell_type": "markdown",
"metadata": {
"id": "IVY-4U-KLcNc"
},
"source": [
"# Plot losses"
]
},
{
"cell_type": "code",
"metadata": {
"colab": {
"base_uri": "https://localhost:8080/",
"height": 265
},
"id": "KC_N3bDNLbYe",
"outputId": "7ed27389-584b-4d21-8167-a1732d32cff1"
},
"source": [
"plt.figure()\n",
"plt.plot(losses, label = 'Loss')\n",
"plt.plot(val_losses, label = 'Validation loss')\n",
"plt.grid(True)\n",
"plt.legend()\n",
"plt.show()"
],
"execution_count": 151,
"outputs": [
{
"output_type": "display_data",
"data": {
"image/png": "iVBORw0KGgoAAAANSUhEUgAAAXcAAAD4CAYAAAAXUaZHAAAABHNCSVQICAgIfAhkiAAAAAlwSFlzAAALEgAACxIB0t1+/AAAADh0RVh0U29mdHdhcmUAbWF0cGxvdGxpYiB2ZXJzaW9uMy4yLjIsIGh0dHA6Ly9tYXRwbG90bGliLm9yZy+WH4yJAAAgAElEQVR4nO3deXxU1fn48c8zWclCQhZCIECCbLIEwq7sIFXQgihacQEKlS7Wau3Xil9b0fqzP/yWVqWiX1eE1p+4FhFQqkgUtOwiOxogQDBsIQlZCGQ5vz/mggECmZnMJJk7z/v1mtfce+45N8/h6pOTu5wrxhiUUkrZi6OhA1BKKeV9mtyVUsqGNLkrpZQNaXJXSikb0uSulFI2FNzQAQAkJCSY1NRUj9qWlJQQGRnp3YAaOe1zYNA+B4a69Hnjxo3HjTGJNW1rFMk9NTWVDRs2eNQ2MzOTYcOGeTegRk77HBi0z4GhLn0Wkf2X2qanZZRSyoY0uSullA1pcldKKRtqFOfclVL1r7y8nJycHMrKyho6lHNiYmLYuXNnQ4dRr1zpc3h4OCkpKYSEhLi8X5eSu4jEAq8A3QADTAV2A28BqUA2cKsxJl9EBHgWGAOUAlOMMZtcjkgpVS9ycnKIjo4mNTUV5/+2Da+oqIjo6OiGDqNe1dZnYwx5eXnk5OSQlpbm8n5dPS3zLPCxMaYz0APYCcwAVhhjOgArrHWA0UAH6zMdeMHlaJRS9aasrIz4+PhGk9hVzUSE+Ph4t//CqjW5i0gMMAR4FcAYc8YYUwCMA+Zb1eYDN1rL44AFxmkNECsiyW5FpZSqF5rY/YMnx8mVkXsacAyYJyJfi8grIhIJJBljcq06h4Eka7kVcLBa+xyrzOvWZ5/g7d1n0GmLlVLqfK6ccw8GegH3GmPWisiz/HAKBgBjjBERtzKsiEzHedqGpKQkMjMz3WkOwPLscpbtK2fpJ5lEhQbOCKS4uNijfy9/pn32vpiYGIqKiny2f1ckJyeTm5t7br2ysrLBY6pvrva5rKzMrf8eXEnuOUCOMWattf4uzuR+RESSjTG51mmXo9b2Q0Drau1TrLLzGGNeAl4C6NOnj/HkCa2SLbm8uWsTV3TvzZXJTd1u76/0Kb7A4Os+79y5s1FcvKweg15QvbTw8HAyMjJc3m+tp2WMMYeBgyLSySoaCewAFgOTrbLJwAfW8mJgkjgNAAqrnb7xqhYxYQAcPtl4buVSStXN5s2bGTBgAOnp6YwfP578/HwA5syZQ5cuXUhPT+e2224D4PPPP6dnz5707NmTjIyMgBv1X46r97nfC7whIqHAXuCnOH8xvC0i04D9wK1W3WU4b4PMwnkr5E+9GnE1zaPDAThSqMldqbp4/MPt7Pj+pFf32aVlU2b+uKvb7SZNmsTf//53hg4dyqOPPsrjjz/OM888w6xZs9i3bx9hYWEUFBQAMHv2bObOncvAgQMpLi4mPDzcq33wZy7dCmmM2WyM6WOMSTfG3GiMyTfG5BljRhpjOhhjrjHGnLDqGmPMPcaYK4wx3Y0xns0I5oKkps4DqSN3peyhsLCQgoIChg4dCsDkyZP54osvAEhPT+eOO+7gn//8J8HBznHpwIEDeeCBB5gzZw4FBQXnypWfP6EaGuwgOhSOaHJXqk48GWHXt6VLl/LFF1/w4Ycf8uSTT7J161ZmzJjB9ddfz7Jlyxg4cCDLly+nc+fODR1qo+D3c8vEhzs4VKDJXSk7iImJoVmzZqxatQqAf/zjHwwdOpSqqioOHjzI8OHDeeqppygsLKS4uJg9e/bQvXt3HnroIfr27cuuXbsauAeNh1+P3AESI4SDJ0obOgyllAdKS0tJSUk5t37PPfcwf/58fvGLX1BaWkq7du2YN28elZWV3HnnnRQWFmKM4Te/+Q2xsbH88Y9/ZOXKlTgcDrp27cro0aMbsDeNi/8n9yYONh8opbLKEOQInHvdlbKDqqqq89bP3ha4Zs2ai+quXr36orK///3vPovN3/n9aZnmEUJ5pdGLqkopVY3fJ/fEJs4uHMjTUzNKKXWW/yf3COepGD3vrpRSP/D75B4XLgQ5hAOa3JVS6hy/T+7BDqFlbLgmd6WUqsbvkztA27hITe5KKVWNLZJ7m/gIsvNKGjoMpZQbhg8fzvLly88rmzt3Lr/85S8v2WbYsGFs2OCc0WTMmDHn5pip7rHHHmP27NmX/dmLFi1ix44d59YfffRRPv30U3fCr1FmZiY33HBDnffjDbZI7u0SIikoLedEyZmGDkUp5aKJEyeycOHC88ree+89Jk6c6FL7ZcuWERsb69HPvjC5/+lPf+Kaa67xaF+NlS2S+xWJUQDsPVbcwJEopVw1YcIEli5dypkzzkFZdnY2hw8fZvDgwfzyl7+kT58+dO3alZkzZ9bYPjU1lePHjwPw5JNP0rFjRwYNGsTu3bvP1Xn55Zfp27cvPXr04Oabb6a0tJSvvvqKxYsX8+CDD9KzZ0/27NnDlClTePfddwFYsWIFGRkZdO/enalTp3L69OlzP2/mzJn06tWL7t271zrVwYkTJ7jxxhtJT09nwIABbNmyBah5muLc3FyGDBlCz5496dat27npF+rC759QBWiXGAnA3mMl9EmNa+BolPJDH82Aw1u9u88W3WH0rEtujouLo1+/fnz00UeMGzeOhQsXMn78eESEJ598kri4OCorKxk5ciRbtmwhPT29xv1s3LiRhQsXsnnzZioqKujVqxe9e/cG4KabbuLuu+8G4A9/+AOvvvoq9957L2PHjuWGG25gwoQJ5+2rrKyMKVOmsGLFCjp27MikSZN44YUXuP/++wFISEhg06ZNPP/888yePZtXXnnlkv2bOXMmGRkZLFq0iM8++4xJkyaxefPmi6YpLi8v57XXXuPaa6/lkUceobKyktLSul9DtMXIPaVZBKFBDvYc15G7Uv6k+qmZhQsXnku2b7/9Nr169SIjI4Pt27efdwrlQqtWrWL8+PFERETQtGlTxo4de27btm3bGDx4MN27d+eNN95g+/btl41n9+7dpKWl0bFjR+D8KYfB+csCoHfv3mRnZ192X6tXr+auu+4CYMSIEeTl5XHy5Mkapynu27cv8+bN47HHHmPr1q1eeRuVLUbuQQ6hbXwEe4/pRVWlPHKZEbYvjRs3jt/+9rds2rSJ0tJSMjIy2LdvH7Nnz2b9+vU0a9aMKVOmUFbm2fQiU6ZMYdGiRfTo0YPXX3+9zu+kDQtzvv0tKCiIiooKj/Zx4TTF77//PkOGDOGLL75g6dKlTJkyhQceeIBJkybVKVZbjNzBeWpGz7kr5V+ioqIYPnw4U6dOPXch9eTJk0RGRhITE8ORI0f46KOPLruPIUOGsGjRIk6dOkVRUREffvjhuW1FRUUkJydTXl7OG2+8ca48Ojq6xlfyderUiezsbLKysoAfphz2xODBg8/9zMzMTBISEmjatOlF0xR/++237N+/n6SkJO6++25+9rOfsWnTJo9+ZnW2GLmD86LqZ7uOUl5ZRUiQbX5nKWV7EydOZPz48edOz/To0YOMjAw6d+5M69atGThw4GXb9+rVi5/85Cf06NGD5s2b07dv33PbnnjiCfr3709iYiL9+/c/l9Bvu+027r77bubMmXPuQio4X0I9b948brnlFioqKujbty+/+MUvPOrXY489xtSpU0lPTyciIoL58+cD8Mwzz5w3TfGoUaNYunQpf/nLXwgJCSEqKooFCxZ49DPPY4xp8E/v3r2Np1auXGmMMeadDQdN24eWmD1Hizzel7842+dAon32vh07dvh0/544efJkQ4dQ71ztc03HC9hgLpFXbTPErX7HjFJKBTrbJPcrEqx73fWOGaWUsk9yj4kIIT4yVEfuSrnB+Ze9auw8OU62Se4AVzSP4rujOnJXyhXh4eHk5eVpgm/kjDHk5eURHh7uVjvb3C0D0LlFNO9vOoQxBhF9n6pSl5OSkkJOTg7Hjh1r6FDOKSsrczuJ+TtX+hweHn7ei8RdYbPk3pTi0/vJyT9F67iIhg5HqUYtJCSEtLS0hg7jPJmZmWRkZDR0GPXKV3221WmZTi2cj+zuPnzxwwlKKRVIbJncdx0+2cCRKKVUw7JVco8KC6Z1XBN26chdKRXgbJXcATolNdXkrpQKeLZL7lcmR7PveAll5ZUNHYpSSjUYl5K7iGSLyFYR2SwiG6yyOBH5RES+s76bWeUiInNEJEtEtohIL1924EKdWkRTWWXI0vvdlVIBzJ2R+3BjTE9jTB9rfQawwhjTAVhhrQOMBjpYn+nAC94K1hWdz11U1VMzSqnAVZfTMuOA+dbyfODGauULrEnL1gCxIpJch5/jltT4SMKCHezM1TtmlFKBS1x59FhE9gH5gAFeNMa8JCIFxphYa7sA+caYWBFZAswyxqy2tq0AHjLGbLhgn9NxjuxJSkrqfeFb0F1VXFxMVFTUeWVP/OcUQQ747/5NPNpnY1dTn+1O+xwYtM/uGT58+MZqZ1PO4+oTqoOMMYdEpDnwiYic99pvY4wREbcmqDDGvAS8BNCnTx8zbNgwd5qfk5mZyYVtM09u5631Bxk0eAjBNnxxR019tjvtc2DQPnuPS5nPGHPI+j4K/AvoBxw5e7rF+j5qVT8EtK7WPMUq8749n9Fx91y44K+PHq1jOFVeSZa+dk8pFaBqTe4iEiki0WeXgR8B24DFwGSr2mTgA2t5MTDJumtmAFBojMn1euQAx7+jZe6/oeT4ecXpKbEAbDlY6JMfq5RSjZ0rI/ckYLWIfAOsA5YaYz4GZgGjROQ74BprHWAZsBfIAl4GfuX1qM+KsWZJKzx4XnFafCTR4cFszinw2Y9WSqnGrNZz7saYvUCPGsrzgJE1lBvgHq9EV5uzyf3kIWj1w+30DoeQnhLDFk3uSqkA5d9XG2OsU/uFORdt6pESy67cIn1SVSkVkPw7uTdpRqUjtMbk3rttMyqqDN8c1NG7Uirw+HdyF6EsvDkU7L9oU5+2cYjAun0nGiAwpZRqWP6d3IFTTZIhb+9F5TERIXRKimZdtiZ3pVTgsUFybwkn9kJV1UXb+qfFsXF/PuWVF29TSik78/vkXhrREipOQdH3F23rlxZP6ZlKtn+v88wopQKL3yf3U01aOhfysi7a1jetGQBr9+bVZ0hKKdXgbJ3cm0eHc0ViJF/u0eSulAosfp/cT4fFQUgE5O2pcfvQjs1ZuzdP73dXSgUUv0/uiAPirqhx5A4wtFMipyuqWKOnZpRSAcT/kztAfLtLJvf+aXGEBTvI3H2snoNSSqmGY4/kntAJ8rOhvOyiTeEhQQxoF88X32pyV0oFDnsk9+ZXgqmC49/WuHlox0T2Hi/hQF5pPQemlFINwz7JHeDYrho3D+2UCEDmt0dr3K6UUnZjj+QedwU4guHozho3t0uIpF1iJMu3H67nwJRSqmHYI7kHh0J8+0uO3EWE0d1asGbvCfJLztRzcEopVf/skdwBEjtfcuQOcF3XZCqrDJ/uPFKPQSmlVMOwT3JvfqXzjpkzNV807daqKa1im/DxNj01o5SyP/sk98TOgLnkHTMiwrVdW7Dqu+MUn66o39iUUqqe2Se5n71j5jKnZkZ3b8GZyio+2aGjd6WUvdknucddAcHhcHjrJav0btOMVrFNWPT1xdMDK6WUndgnuQcFQ1I3OLzlklUcDmFsz5aszjrOsaLT9RicUkrVL/skd4DkdMjdAsZcssr4jFZUVhmWbNHRu1LKvmyW3HvA6ULnXTOX0DEpmiuTm7JosyZ3pZR92Su5t0h3fud+c9lq4zNa8s3BAvYdL6mHoJRSqv7ZK7k37+KchuAy590BxvZohUPgvY059RSYUkrVL3sl95Bw5/3utYzcW8SEM7RjIu9sPEhFZVU9BaeUUvXHXskdnOfdc7+57EVVgNv6teHIydP6Eg+llC3ZL7m3SIeSY1B0+QeVRnRuTkJUGAvXH6ynwJRSqv64nNxFJEhEvhaRJdZ6moisFZEsEXlLREKt8jBrPcvanuqb0C+hZU/n96GNl60WEuTglj4prNx9lCMnL36Dk1JK+TN3Ru73AdWf7X8KeNoY0x7IB6ZZ5dOAfKv8aate/UnuCY4QyFlXa9Vb+7Smssrwrl5YVUrZjEvJXURSgOuBV6x1AUYA71pV5gM3WsvjrHWs7SOt+vUjJNw5ej9Ye3JPS4hkQLs4Fq4/QFXV5c/RK6WUPwl2sd4zwO+BaGs9HigwxpydXjEHaGUttwIOAhhjKkSk0Kp/vPoORWQ6MB0gKSmJzMxMjzpQXFx8UdsraEmrg8tY9dknGEfIZdv3jKpgzd7TPP/+CroluPrP0bBq6rPdaZ8Dg/bZe2rNZiJyA3DUGLNRRIZ56wcbY14CXgLo06ePGTbMs11nZmZyUdvmhfD2BwztGAcpvS/b/uqKKt7d+xmbS2L49YS+HsVQ32rss81pnwOD9tl7XDktMxAYKyLZwEKcp2OeBWJF5OwvhxTgkLV8CGgNYG2PAfK8GHPtUvo5vw+urbVqaLCDif1as2LXUQ6eqPlFH0op5W9qTe7GmIeNMSnGmFTgNuAzY8wdwEpgglVtMvCBtbzYWsfa/pkxtdx07m1NkyGmDRxc41L12/u3wSHCP9fs93FgSilVP+pyn/tDwAMikoXznPqrVvmrQLxV/gAwo24heqjt1ZD9JVTV/gRqckwTru2axML1Bzl1prIeglNKKd9yK7kbYzKNMTdYy3uNMf2MMe2NMbcYY05b5WXWentr+15fBF6rdkOh9Dgc3e5S9UlXpVJ4qpwPv9HZIpVS/s9+T6ielTbU+b0306Xq/dPi6JQUzetfZVPfZ5GUUsrb7JvcY1pBQkfY+7lL1UWESVe3ZUfuSTbuz/dxcEop5Vv2Te7gHL3v/xIqzrhUfXxGK5qGBzPvy2zfxqWUUj5m7+TebhiUl7p0SyRARGgwt/dvy0fbcvW2SKWUX7N5ch8KQaHw7ccuN5l8dVscIjp6V0r5NXsn97BoSB0Mu5fVOr/7WckxTfhxj5a8tf4AhafKfRygUkr5hr2TO0Cn0XBiLxz/zuUm0walUXKmkrfWH/BhYEop5TuBkdzBOXp3UbdWMVzVLp55X2ZTrq/hU0r5Ifsn95gU59uZdi11q9nPBqeRW1jGsq25PgpMKaV8x/7JHaDLOOfLOwpcP80yvFNz2iVG8sqqffpQk1LK7wRGcu92s/N723suN3E4hGmD0th6qJC1+074KDCllPKNwEjucWmQ0he2up7cAW7ulUJ8ZCj/+/keHwWmlFK+ERjJHaDbBDiyFY7ucrlJeEgQUwelkbn7GNu/L/RhcEop5V2Bk9y7jgdxwNZ33Gp254C2RIUF83ymjt6VUv4jcJJ7dBK0Gw7fLIQq1+dsj2kSwp0D2vLR1lz2HS/xYYBKKeU9gZPcAXpNgpM5sOczt5pNG5RGSJCDF/Xcu1LKTwRWcu80BiISYOPrbjVLjA7j1j6teW9TDocLy3wTm1JKeVFgJffgUOg50TmRWNERt5pOH9KOKgMvr2qYF0sppZQ7Aiu5A2RMgqoK+OZNt5q1jotgXI+WvLnuAPklrs0Pr5RSDSXwkntiR2hzNWx4za0LqwC/GHYFpWcqef2rbN/EppRSXhJ4yR2g/3Qo2O/WPO8AHZOiGdUlide/yqaoTKcDVko1XoGZ3Dv/GJqmwJoX3G5674j2FJ4qZ76O3pVSjVhgJvegYOh3N2SvgsPb3GqanhLLyM7NeXnVPh29K6UarcBM7uC85z0kAta6P3q//5qOOnpXSjVqgZvcI+Kgx22w5R0oPuZW0+4pMVxzpXP0flJH70qpRihwkzvAgF9B5RlY96LbTe8baY3e9UXaSqlGKLCTe0IHuPIGWPcSlJ10q6lz9J7EK6t19K6UanwCO7kDDPotlBW6PSUBwP3XdKDwVDmv6+hdKdXIaHJv1RvShsJ/5kLFabeadmtljd5X7dXRu1KqUdHkDs7Re/Fht6ckAOfo/WRZBa+t3ueDwJRSyjO1JncRCReRdSLyjYhsF5HHrfI0EVkrIlki8paIhFrlYdZ6lrU91bdd8IJ2w6BlBnz5rNtTEnRrFcN1XVvwyqp9nNA5Z5RSjYQrI/fTwAhjTA+gJ3CdiAwAngKeNsa0B/KBaVb9aUC+Vf60Va9xE3GO3k/she3/crv5f13bkdIzFTy/MssHwSmllPtqTe7GqdhaDbE+BhgBvGuVzwdutJbHWetY20eKiHgtYl/p/GNo3hU+f8rt0Xv75tFM6J3CgjX7OVRwykcBKqWU68QYU3slkSBgI9AemAv8BVhjjc4RkdbAR8aYbiKyDbjOGJNjbdsD9DfGHL9gn9OB6QBJSUm9Fy5c6FEHiouLiYqK8qjthRKOfUW37U+x48rfcjRpmFtt805V8dCqU1yVHMy07mFeiedSvNlnf6F9DgzaZ/cMHz58ozGmT40bjTEuf4BYYCUwCMiqVt4a2GYtbwNSqm3bAyRcbr+9e/c2nlq5cqXHbS9SWWnM81cbM6eXMRXlbjd/4sPtJm3GEvPdkZPei6kGXu2zn9A+Bwbts3uADeYSedWtu2WMMQVWcr8KiBWRYGtTCnDIWj5kJXus7TFAnjs/p8E4HDD0IcjLgm3v1l7/Ar8a3p6I0GBmL//WB8EppZTrXLlbJlFEYq3lJsAoYCfOJD/BqjYZ+MBaXmytY23/zPoN4x863wBJ3Z3n3isr3GoaFxnK3YPb8fH2w2w+WOCjAJVSqnaujNyTgZUisgVYD3xijFkCPAQ8ICJZQDzwqlX/VSDeKn8AmOH9sH3I4YBhM5x3zmx92+3m0wanER8ZylMf7cKffqcppewluLYKxpgtQEYN5XuBfjWUlwG3eCW6htL5emiR7hy9d5vgfLG2i6LCgvn1iPY8/uEOMncfY3jn5j4MVCmlaqZPqNZEBEb8EfKzYdP8Wqtf6I7+bUmNj+DJZTupqKzyfnxKKVULTe6X0mEUtB0In/8PnC6uvX41ocEOHh5zJVlHi3lz/UEfBaiUUpemyf1SROCax6HkKKx53u3mP+qSRP+0OJ7+5FudVEwpVe80uV9O677Ou2e+nAMlx2uvX42I8McbupBfeoa5Oi2BUqqeaXKvzchHobwEVv3V7abdWsVwU0YK81Znc/BEqQ+CU0qpmmlyr01iJ+h5O6x/BQoOuN38wWs7EeQQZn28ywfBKaVUzTS5u2LYw4DAyj+73bRFTDg/H9qOpVty2ZB9wvuxKaVUDTS5uyImBfr/HL5ZCIe3ut18+pB2JMeE8+gH26ms0geblFK+p8ndVYMfgCax8PHD4OaTpxGhwTxy/ZXsyD3JG2v3+yhApZT6gSZ3VzVpBsMfgexVsHuZ282v757M1VfEM3v5bvKK3XtXq1JKuUuTuzt6/xQSOsG//wAV7r1ST0T407iulJ6p5Cm9uKqU8jFN7u4ICoZr/+ycVGzdS243b988mqmD0nh7Qw6bDuT7IECllHLS5O6uDtdA+1HOaQncfLAJ4DcjO5DUNIxHP9imF1eVUj6jyd0T1z4JZ4o9ujUyKiyY/x5zJdsOneT/rXP/vnmllHKFJndPJHaCvtNg4zw4ssPt5mN7tOSqdvH8z8e7OHqyzAcBKqUCnSZ3Tw17GMKawrIH3b41UkT4803dOV1RxczF230UoFIqkGly91REHFwzE/avhq3vuN08LSGS+0Z24KNth/n39sM+CFApFcg0uddFr8nQshcsfwTKCt1uPn1IOzq3iObRD7ZTpNMCK6W8SJN7XTiC4Pq/Qskxjy6uhgQ5mHVzOkeKyvifj3f7IEClVKDS5F5XrXpBn6nO+95zt7jdvGfrWKZcnco/1+5n436dWEwp5R2a3L1h5B+hSRws/R1Uuf/O1P/6USdaxjThofe2UlZe6YMAlVKBRpO7NzRpBqP+BDnrYPMbbjePDAvmyfHdyDpazNOffuuDAJVSgUaTu7f0mAhtrnLOO1N81O3mwzo1Z2K/Nrz0xV49PaOUqjNN7t7icMCP50B5KXz0e4928cj1V9Iqtgm/e/sbSs9UeDlApVQg0eTuTYkdYejvYfu/YNdSt5tHhQUz+5YeZOeV8tRHOnOkUspzmty9beD9kNTNeXHVg3vfB7SLZ+rANOb/Zz9fZrk/MZlSSoEmd+8LCoGxf4fiI/DJox7t4vfXdaJdYiQPvvMNhaX6cJNSyn2a3H2hVS+46h7Y+DrsW+V28/CQIJ75SU+OFp1mxvtbMG7OXaOUUprcfWXYf0OzNFh8L5wpcbt5ekosv7+uEx9tO8yb6w76IECllJ3VmtxFpLWIrBSRHSKyXUTus8rjROQTEfnO+m5mlYuIzBGRLBHZIiK9fN2JRik0AsY9B/nZ8O8/erSLnw1qx+AOCTz+4Xa+PVLk3fiUUrbmysi9AvidMaYLMAC4R0S6ADOAFcaYDsAKax1gNNDB+kwHXvB61P4idZDz9MyGVyHrU7ebOxzCX2/tQXR4ML9582t9elUp5bJak7sxJtcYs8laLgJ2Aq2AccB8q9p84EZreRywwDitAWJFJNnrkfuLEX+ExCth0T1Q6v7DSc2jw5l9Sw92HS7i/yx1/8UgSqnAJO5crBORVOALoBtwwBgTa5ULkG+MiRWRJcAsY8xqa9sK4CFjzIYL9jUd58iepKSk3gsXLvSoA8XFxURFRXnUtr5EFe2l16b/4njCVezo+qBH+1i46wwfZ5czPT2M9KZljb7P3uYPx9nbtM+BoS59Hj58+EZjTJ8aNxpjXPoAUcBG4CZrveCC7fnW9xJgULXyFUCfy+27d+/exlMrV670uG29+vx/jJnZ1Jgt73jUvLyi0tz6v1+ZTn9YZhYsXuHl4Bo/vznOXqR9Dgx16TOwwVwir7p0t4yIhADvAW8YY963io+cPd1ifZ+dUOUQ0Lpa8xSrLLAN/C2k9IWlD0D+frebBwc5eO72XsQ0CeG5r8soPKX3vyulLs2Vu2UEeBXYaYz5W7VNi4HJ1vJk4INq5ZOsu2YGAIXGmFwvxuyfgoLhppecUwK/Nw0q3U/OidFhPH9HL46fMvzu7c1UVen970qpmrkych8I3AWMEJHN1mcMMAsYJSLfAddY6wDLgL1AFvAy8LHUI2QAAA8JSURBVCvvh+2n4trB2GchZz2sfNKjXfRuG8fEzqF8uvMoc1dmeTlApZRdBNdWwTgvjMolNo+sob4B7qljXPbV7WbY+zmsfhpSB0P7i/4JazWyTTAl4Yn89ZNv6ZAUxXXdAvdmJKVUzfQJ1YZw3Szn7ZH/+jkUHXG7uYjwf2/qTkabWO5/azNbc9yfoEwpZW+a3BtCaATc8jqcLob3fwaV7s/dHh4SxEt39SE+Moxp89eTW3jK+3EqpfyWJveG0rwzXP9X2PcFrHjco10kRofx2pS+lJ6pZNrrGyg5rS/4UEo5aXJvSBl3QJ9p8NUc2PZ+7fVr0KlFNH+/PYNdh09y38Kvqah0/wXdSin70eTe0K6bBa37wwe/hiOeTS8wvFNzHh/blU93HuXh97fqFMFKKU3uDS44FG6ZD2FR8NYdcKrAo93cdVUq943swDsbc5j1sb6iT6lAp8m9MWiaDLcugIID8J5nF1gB7r+mA3cNaMuLn+/lpS/2eDlIpZQ/0eTeWLQZAGNmQ9YnsPy/PdqFiPDY2K5cn57Mn5ftYuG6A14OUinlL2p9iEnVoz4/hbws+M9zzqdZB/zC7V0EOYS/3dqDktMVzHh/KyLwk75tfBCsUqox05F7YzPqCeh8Ayx/GHZ/5NEuwoKD+N87ezO0YyIPvbeVt9brCF6pQKPJvbFxOOCmlyG5B7w7Fb7f7NFuwkOCePEuTfBKBSpN7o1RaARMfAsi4uGNWyDPs4ujFyb4V1fv83KgSqnGSpN7YxWdBHe+B1UV8I8b4eT3Hu3mbIIf3a0FTyzZwVMf79L74JUKAJrcG7PETnDnu853r/5jvEfvYAVngn/u9l7c3r8NL2Tu4aH3tuiTrErZnCb3xq5Vb5j4JpzYB/+8GU4XebSbIIfw5I3d+M3IDry9IYe7F2ygqEzf5qSUXWly9wdpQ5yzSOZ+A/+cQFBFqUe7EREeGNWRJ27sxhffHeem57/iQJ5n+1JKNW6a3P1F5zEw4VXIWU/6lseg7KTHu7prQFsWTO3H0aLTjJ27mq/2HPdenEqpRkGTuz/pOh5umUd0URb88yYo8/wlHQPbJ7D41wNJiArjrlfXMXdllr6TVSkb0eTub7qMY0eXB+H7r2HBjR5fZAVoGx/Jv351Ndd1a8Fflu9m8rx1HCs67cVglVINRZO7HzqeeBXc+g84sh1euxYKDnq8r+jwEJ6bmMGfx3dn3b4TjJmzipW7j3oxWqVUQ9Dk7q86j4G7/uV8B+uro5yJ3kMiwu392/DBrwcS2ySEn85bzwNvbSa/5IwXA1ZK1SdN7v4sdSBMteafeW00ZK+u0+46t2jKkt8M4t4R7Vn8zfeMevpzPvzme33oSSk/pMnd3yV1hWmfOJ9oXXAjbJhXp92FBQfxux91YvGvB5Ec04R73/yaW1/8D1tzPL94q5Sqf5rc7SC2tTPBtxsKS+6Hpb+Dyro9oNSlZVMW3TOQ/3tTd/YeK+HHz63m/oVfs+dYsZeCVkr5kiZ3u2gSC7e/DQPvg/WvwIJxUHysTrsMcggT+7Vh5YPD+PnQdny8/TDX/O1zfv3/NrEz1/P77JVSvqfJ3U4cQTDqT3DTK3BoI/zvQNizss67bRoewsOjr2T1QyP4+ZArWLnrKKOfXcXNL3zF+5tyKCuv9ELwSilv0uRuR+m3wN2fQXisc8KxTx+r82kagISoMGaM7syXM0bwyJgrOVFyhgfe/oa+/+dT7lv4NR9vy+XUGU30SjUG+po9u0rqCtMz4eMZsPpp2JsJ4+Y6y+soNiKUu4e042eD0/jPnjwWbT7EJzuO8MHm7wkNdtCzdSwD2sXTN7UZXZKbEh8VVuefqZRyjyZ3OwuNgLFz4IoRzousLw6Fwb9zfoJD67x7EeHq9glc3T6Bisoq1u47wcpdR1m77wTPffYdZ2czSIwOo3OLaNrERdCqWRNaxTo/cZGhxEaEEtMkhCCH1DkepdQPak3uIvIacANw1BjTzSqLA94CUoFs4FZjTL6ICPAsMAYoBaYYYzb5JnTlsq43Qupg5yj+81mw4wMY/ZTz7hovCQ5yMLB9AgPbJwBwsqycLQcL2XX4JLsOF7H7cBHbDuWSX1rz6aGm4cHERIQQERJMeIiD8JAgmoQGER7s/M47dpp/528l2CE4RAh2CEEXfIIdgsPaDnD214UIiLUm1X6HSLV6Z8ulenm1Mmqo62vfHijn0Nr99fPDGolA7DPFvnm3gisj99eB54AF1cpmACuMMbNEZIa1/hAwGuhgffoDL1jfqqFFxsPNL0P3CbDsv2DBWOeLuEf9CeKv8PqPaxoewqAOCQzqkHBeecnpCr4vOMWhglMUlJaTX3qGgtJyCkrPUHiqnFPllZwqr6KsvJITJWcoK6/kVHklRSWV7Cg4TGWVoaLKUHX22zi/bfuc1Y5tDR1B/QuwPk/qUve/omtSa3I3xnwhIqkXFI8DhlnL84FMnMl9HLDAOB9pXCMisSKSbIzJ9VbAqo46XgtpQ2HNXFj1N5jbH/r8FAbeDzGtfP7jI8OC6ZAUTYekaLfaZWZmMmzYsEtur6oyVBpDpZXoDc5s71zGWjbVls8uXLqutflcXUO1CvXgq6++4uqrr66/H9gIBGKfv173H5/s19Nz7knVEvZhIMlabgVUn8UqxyrT5N6YhIQ7z7v3vANW/hk2vAYbX4eMO2HQbyG2TUNH6DaHQ3AghAQ1dCTeExvuoHnT8IYOo14FYp/Dgn1znk9cmTfEGrkvqXbOvcAYE1tte74xppmILAFmGWNWW+UrgIeMMRtq2Od0YDpAUlJS74ULF3rUgeLiYqKiojxq66+83eewsqO03f8uLQ6vQEwVefF9OdRqDPnN0kEax92yepwDg/bZPcOHD99ojOlT0zZPR+5Hzp5uEZFk4OwcsYeA1tXqpVhlFzHGvAS8BNCnTx9zuT+5L6e2P9ftyDd9vhUKc2D9qyRsWkDClpkQdwV0vwW63eR8WXcD0uMcGLTP3uPpsGwxMNlangx8UK18kjgNAAr1fLsfiUmBa2bCAzvgppehaUv4/CmY2w+evxpWPAH7VkGFvtBDqcbOlVsh38R58TRBRHKAmcAs4G0RmQbsB261qi/DeRtkFs5bIX/qg5iVrwWHQfqtzk/RYeetk9v/5XwYatVsCImA1v0guSe07AnJPSC2rXP6A6VUo+DK3TITL7FpZA11DXBPXYNSjUh0C+j/c+enrBCyv4S9K+HgWvjPXKiy7lt3hDgvxMalQUxriEyAyESIiHd+QiOdvxRCmvywHBwOjmBwNI7z+krZiT6hqlwXHuN8A1TnMc71itNwdAcc3gon9sKJfZC/z/l+19ITuH7foDhH/Y5gkKAfEr4j2CpzMOB0GWxqYlWXH9pV+/ph/cLt1dYvt62R6VtaAtsiGzqMehWIfU5s/mN+uLPcezS5K88Fh0HLDOfnQlWVcCofSo7DqRNwphTKS6D8FJwpgfJSqCiDqiqoqnB+TKWzXVXlD2VVFYAhPzeX5BbJnPuFUf1GdZfXa6vbuJQcO0ZkYmJDh1GvArHPFcG+uTtIk7vyDUeQdWomofa6LtidmUlygN1FsSMzk+baZ9vLz8z0yX71ZKdSStmQJnellLIhTe5KKWVDmtyVUsqGNLkrpZQNaXJXSikb0uSulFI2pMldKaVsyKX53H0ehMgxnBOQeSIBOO7FcPyB9jkwaJ8DQ1363NYYU+MjvY0iudeFiGy41GT1dqV9Dgza58Dgqz7raRmllLIhTe5KKWVDdkjuLzV0AA1A+xwYtM+BwSd99vtz7koppS5mh5G7UkqpC2hyV0opG/Lr5C4i14nIbhHJEpEZDR2Pt4hIaxFZKSI7RGS7iNxnlceJyCci8p313cwqFxGZY/07bBGRXg3bA8+ISJCIfC0iS6z1NBFZa/XrLREJtcrDrPUsa3tqQ8btKRGJFZF3RWSXiOwUkasC4Bj/1vpvepuIvCki4XY8ziLymogcFZFt1crcPrYiMtmq/52ITHYnBr9N7iISBMwFRgNdgIki0qVho/KaCuB3xpguwADgHqtvM4AVxpgOwAprHZz/Bh2sz3TghfoP2SvuA3ZWW38KeNoY0x7IB6ZZ5dOAfKv8aaueP3oW+NgY0xnogbPvtj3GItIK+A3QxxjTDQgCbsOex/l14LoLytw6tiISB8wE+gP9gJlnfyG4xBjjlx/gKmB5tfWHgYcbOi4f9fUDYBSwG0i2ypKB3dbyi8DEavXP1fOXD5Bi/Qc/AliC843Vx4HgC483sBy4yloOtupJQ/fBzf7GAPsujNvmx7gVcBCIs47bEuBaux5nIBXY5umxBSYCL1YrP69ebR+/Hbnzw38oZ+VYZbZi/SmaAawFkowxudamw0CStWyHf4tngN8DVdZ6PFBgjKmw1qv36Vx/re2FVn1/kgYcA+ZZp6JeEZFIbHyMjTGHgNnAASAX53HbiL2Pc3XuHts6HXN/Tu62JyJRwHvA/caYk9W3GeevclvcxyoiNwBHjTEbGzqWehQM9AJeMMZkACX88Gc6YK9jDGCdUhiH8xdbSyCSi09dBIT6OLb+nNwPAa2rradYZbYgIiE4E/sbxpj3reIjIpJsbU8Gjlrl/v5vMRAYKyLZwEKcp2aeBWJFJNiqU71P5/prbY8B8uozYC/IAXKMMWut9XdxJnu7HmOAa4B9xphjxphy4H2cx97Ox7k6d49tnY65Pyf39UAH60p7KM4LM4sbOCavEBEBXgV2GmP+Vm3TYuDsFfPJOM/Fny2fZF11HwAUVvvzr9EzxjxsjEkxxqTiPI6fGWPuAFYCE6xqF/b37L/DBKu+X41wjTGHgYMi0skqGgnswKbH2HIAGCAiEdZ/42f7bNvjfAF3j+1y4Eci0sz6q+dHVplrGvqiQx0vWIwBvgX2AI80dDxe7NcgnH+ybQE2W58xOM83rgC+Az4F4qz6gvPOoT3AVpx3IzR4Pzzs+zBgibXcDlgHZAHvAGFWebi1nmVtb9fQcXvY157ABus4LwKa2f0YA48Du4BtwD+AMDseZ+BNnNcVynH+lTbNk2MLTLX6nwX81J0YdPoBpZSyIX8+LaOUUuoSNLkrpZQNaXJXSikb0uSulFI2pMldKaVsSJO7UkrZkCZ3pZSyof8PYuSzu8YHxXEAAAAASUVORK5CYII=\n",
"text/plain": [
"