From 34da38f64b51cac95475393798efe76e9f3c8aec Mon Sep 17 00:00:00 2001 From: Logan Bauman Date: Wed, 4 May 2022 23:51:19 -0400 Subject: hi --- hyperparameters.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) (limited to 'hyperparameters.py') diff --git a/hyperparameters.py b/hyperparameters.py index a0068dd1..71c5325a 100644 --- a/hyperparameters.py +++ b/hyperparameters.py @@ -9,7 +9,7 @@ Number of epochs. If you experiment with more complex networks you might need to increase this. Likewise if you add regularization that slows training. """ -num_epochs = 100 +num_epochs = 2000 """ A critical parameter that can dramatically affect whether training -- cgit v1.2.3-70-g09d2 From 8ff6cd542e4f9594c73812b7db263bec2f2a1df7 Mon Sep 17 00:00:00 2001 From: Logan Bauman Date: Thu, 5 May 2022 00:04:40 -0400 Subject: hi --- hyperparameters.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) (limited to 'hyperparameters.py') diff --git a/hyperparameters.py b/hyperparameters.py index 71c5325a..801d1ce9 100644 --- a/hyperparameters.py +++ b/hyperparameters.py @@ -9,7 +9,7 @@ Number of epochs. If you experiment with more complex networks you might need to increase this. Likewise if you add regularization that slows training. """ -num_epochs = 2000 +num_epochs = 1000 """ A critical parameter that can dramatically affect whether training -- cgit v1.2.3-70-g09d2 From b203623f4723d5c2ccf9146b393a04650cb1c3d0 Mon Sep 17 00:00:00 2001 From: Logan Bauman Date: Thu, 5 May 2022 00:09:36 -0400 Subject: mega --- hyperparameters.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) (limited to 'hyperparameters.py') diff --git a/hyperparameters.py b/hyperparameters.py index 801d1ce9..40120a53 100644 --- a/hyperparameters.py +++ b/hyperparameters.py @@ -9,7 +9,7 @@ Number of epochs. If you experiment with more complex networks you might need to increase this. Likewise if you add regularization that slows training. """ -num_epochs = 1000 +num_epochs = 10000 """ A critical parameter that can dramatically affect whether training -- cgit v1.2.3-70-g09d2 From 6e5f2d1a62f4f3bf0e87829082b2120ca440ddf0 Mon Sep 17 00:00:00 2001 From: Logan Bauman Date: Thu, 5 May 2022 11:12:46 -0400 Subject: hi --- __pycache__/hyperparameters.cpython-38.pyc | Bin 324 -> 341 bytes __pycache__/losses.cpython-38.pyc | Bin 4378 -> 4409 bytes hyperparameters.py | 2 +- losses.py | 4 +++- save.jpg | Bin 20991 -> 39903 bytes 5 files changed, 4 insertions(+), 2 deletions(-) (limited to 'hyperparameters.py') diff --git a/__pycache__/hyperparameters.cpython-38.pyc b/__pycache__/hyperparameters.cpython-38.pyc index 1e5b0ee5..e1a90bd4 100644 Binary files a/__pycache__/hyperparameters.cpython-38.pyc and b/__pycache__/hyperparameters.cpython-38.pyc differ diff --git a/__pycache__/losses.cpython-38.pyc b/__pycache__/losses.cpython-38.pyc index 1fa5462b..71b86245 100644 Binary files a/__pycache__/losses.cpython-38.pyc and b/__pycache__/losses.cpython-38.pyc differ diff --git a/hyperparameters.py b/hyperparameters.py index 40120a53..b03db017 100644 --- a/hyperparameters.py +++ b/hyperparameters.py @@ -9,7 +9,7 @@ Number of epochs. If you experiment with more complex networks you might need to increase this. Likewise if you add regularization that slows training. """ -num_epochs = 10000 +num_epochs = 5000 """ A critical parameter that can dramatically affect whether training diff --git a/losses.py b/losses.py index 7c99d73a..7198ebf4 100644 --- a/losses.py +++ b/losses.py @@ -11,7 +11,7 @@ class YourModel(tf.keras.Model): self.content_image = transform.resize(content_image, tf.shape(style_image), anti_aliasing=True) self.content_image = tf.expand_dims(self.content_image, axis=0) - + #perhaps consider cropping to avoid distortion self.style_image = transform.resize(style_image, tf.shape(style_image), anti_aliasing=True) self.style_image = tf.expand_dims(self.style_image, axis=0) @@ -19,6 +19,8 @@ class YourModel(tf.keras.Model): self.alpha = hp.alpha self.beta = hp.beta + print(self.x.shape) + print(self.content_image.shape, self.style_image.shape) self.optimizer = tf.keras.optimizers.Adam() diff --git a/save.jpg b/save.jpg index c19eff65..86b5f854 100644 Binary files a/save.jpg and b/save.jpg differ -- cgit v1.2.3-70-g09d2 From d802c988a57d6afe4fca979384ba377ecc7edb66 Mon Sep 17 00:00:00 2001 From: Logan Bauman Date: Fri, 6 May 2022 23:32:17 -0400 Subject: hi --- __pycache__/hyperparameters.cpython-38.pyc | Bin 341 -> 332 bytes __pycache__/losses.cpython-38.pyc | Bin 4409 -> 4546 bytes hyperparameters.py | 6 ++--- losses.py | 37 +++++++++++++++++++++-------- main.py | 12 ++++++++-- save.jpg | Bin 39903 -> 46064 bytes 6 files changed, 40 insertions(+), 15 deletions(-) (limited to 'hyperparameters.py') diff --git a/__pycache__/hyperparameters.cpython-38.pyc b/__pycache__/hyperparameters.cpython-38.pyc index e1a90bd4..7d32eefa 100644 Binary files a/__pycache__/hyperparameters.cpython-38.pyc and b/__pycache__/hyperparameters.cpython-38.pyc differ diff --git a/__pycache__/losses.cpython-38.pyc b/__pycache__/losses.cpython-38.pyc index 71b86245..66e565ec 100644 Binary files a/__pycache__/losses.cpython-38.pyc and b/__pycache__/losses.cpython-38.pyc differ diff --git a/hyperparameters.py b/hyperparameters.py index b03db017..460543dc 100644 --- a/hyperparameters.py +++ b/hyperparameters.py @@ -9,17 +9,17 @@ Number of epochs. If you experiment with more complex networks you might need to increase this. Likewise if you add regularization that slows training. """ -num_epochs = 5000 +num_epochs = 1000 """ A critical parameter that can dramatically affect whether training succeeds or fails. The value for this depends significantly on which optimizer is used. Refer to the default learning rate parameter """ -learning_rate = 3e-2 +learning_rate = 1e-2 momentum = 0.01 alpha = 1e-2 -beta = 1e-5 +beta = 1e-4 diff --git a/losses.py b/losses.py index 7198ebf4..5564ca33 100644 --- a/losses.py +++ b/losses.py @@ -1,8 +1,10 @@ import tensorflow as tf +import numpy as np from tensorflow.keras.layers import \ Conv2D, AveragePooling2D from skimage import transform import hyperparameters as hp + class YourModel(tf.keras.Model): """ Your own neural network model. """ @@ -11,17 +13,26 @@ class YourModel(tf.keras.Model): self.content_image = transform.resize(content_image, tf.shape(style_image), anti_aliasing=True) self.content_image = tf.expand_dims(self.content_image, axis=0) + print(self.content_image) #perhaps consider cropping to avoid distortion self.style_image = transform.resize(style_image, tf.shape(style_image), anti_aliasing=True) self.style_image = tf.expand_dims(self.style_image, axis=0) - self.x = tf.Variable(tf.expand_dims(tf.random.uniform(tf.shape(content_image)), axis=0), trainable=True) + #self.x = tf.Variable(initial_value = self.content_image.numpy().astype(np.float32), trainable=True) + self.x = tf.Variable(initial_value = np.random.rand(self.content_image.shape[0], + self.content_image.shape[1], self.content_image.shape[2], self.content_image.shape[3]).astype(np.float32), trainable=True) + self.alpha = hp.alpha self.beta = hp.beta - print(self.x.shape) + self.photo_layers = None + self.art_layers = None + - print(self.content_image.shape, self.style_image.shape) + + #(self.x.shape) + + #print(self.content_image.shape, self.style_image.shape) self.optimizer = tf.keras.optimizers.Adam() @@ -88,14 +99,20 @@ class YourModel(tf.keras.Model): return x, layers def loss_fn(self, p, a, x): - _, photo_layers = self.call(p) - _, art_layers = self.call(a) - _, input_layers = self.call(x) - - content_l = self.content_loss(photo_layers, input_layers) - style_l = self.style_loss(art_layers, input_layers) + # print(p) + if(self.photo_layers == None): + _, self.photo_layers = self.call(p) + # print(a) + if(self.art_layers == None): + _, self.art_layers = self.call(a) + # print(x) + _, input_layers = self.call(x) + + + content_l = self.content_loss(self.photo_layers, input_layers) + style_l = self.style_loss(self.art_layers, input_layers) # Equation 7 - return (self.alpha * content_l) + (self.beta * style_l) + return (self.alpha * content_l) + (self.beta * style_l) def content_loss(self, photo_layers, input_layers): L_content = tf.constant(0.0) diff --git a/main.py b/main.py index 60574e63..a074f71d 100644 --- a/main.py +++ b/main.py @@ -1,7 +1,9 @@ import os import sys import argparse +import cv2 import tensorflow as tf +from skimage import transform import hyperparameters as hp from losses import YourModel @@ -50,9 +52,15 @@ def main(): print('this is',ARGS.content) content_image = imread(ARGS.content) - content_image = np.resize(content_image, (255, 255, 3)) + + style_image = imread(ARGS.style) - style_image = np.resize(style_image, (255, 255, 3)) + cv2.imshow('hi1', style_image) + cv2.waitKey(0) + + style_image = transform.resize(style_image, content_image.shape) + cv2.imshow('hi2', style_image) + cv2.waitKey(0) my_model = YourModel(content_image=content_image, style_image=style_image) my_model.vgg16.build([1, 255, 255, 3]) my_model.vgg16.load_weights('vgg16_imagenet.h5', by_name=True) diff --git a/save.jpg b/save.jpg index 86b5f854..f4f49d6e 100644 Binary files a/save.jpg and b/save.jpg differ -- cgit v1.2.3-70-g09d2 From 1da66023ea97b5c036a97a319843cea7c334af93 Mon Sep 17 00:00:00 2001 From: Logan Bauman Date: Fri, 6 May 2022 23:41:18 -0400 Subject: hi --- hyperparameters.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) (limited to 'hyperparameters.py') diff --git a/hyperparameters.py b/hyperparameters.py index 460543dc..70ad7577 100644 --- a/hyperparameters.py +++ b/hyperparameters.py @@ -9,7 +9,7 @@ Number of epochs. If you experiment with more complex networks you might need to increase this. Likewise if you add regularization that slows training. """ -num_epochs = 1000 +num_epochs = 3000 """ A critical parameter that can dramatically affect whether training -- cgit v1.2.3-70-g09d2 From d758fd450ed2d6ba0086fe3c2cb355829af1b828 Mon Sep 17 00:00:00 2001 From: Logan Bauman Date: Fri, 6 May 2022 23:50:58 -0400 Subject: hi --- hyperparameters.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) (limited to 'hyperparameters.py') diff --git a/hyperparameters.py b/hyperparameters.py index 70ad7577..73f4b497 100644 --- a/hyperparameters.py +++ b/hyperparameters.py @@ -9,7 +9,7 @@ Number of epochs. If you experiment with more complex networks you might need to increase this. Likewise if you add regularization that slows training. """ -num_epochs = 3000 +num_epochs = 5000 """ A critical parameter that can dramatically affect whether training -- cgit v1.2.3-70-g09d2 From 10f2ece4ab4d8df526c5ff77958b25fe6f32344d Mon Sep 17 00:00:00 2001 From: Logan Bauman Date: Sat, 7 May 2022 08:12:20 -0400 Subject: hi --- __pycache__/hyperparameters.cpython-38.pyc | Bin 315 -> 341 bytes __pycache__/losses.cpython-38.pyc | Bin 4529 -> 4610 bytes hyperparameters.py | 6 +++--- losses.py | 6 ++++-- save.jpg | Bin 22251 -> 25059 bytes 5 files changed, 7 insertions(+), 5 deletions(-) (limited to 'hyperparameters.py') diff --git a/__pycache__/hyperparameters.cpython-38.pyc b/__pycache__/hyperparameters.cpython-38.pyc index 99d18197..40cba873 100644 Binary files a/__pycache__/hyperparameters.cpython-38.pyc and b/__pycache__/hyperparameters.cpython-38.pyc differ diff --git a/__pycache__/losses.cpython-38.pyc b/__pycache__/losses.cpython-38.pyc index a642b379..ff01b5a7 100644 Binary files a/__pycache__/losses.cpython-38.pyc and b/__pycache__/losses.cpython-38.pyc differ diff --git a/hyperparameters.py b/hyperparameters.py index 73f4b497..fedd4dd7 100644 --- a/hyperparameters.py +++ b/hyperparameters.py @@ -9,7 +9,7 @@ Number of epochs. If you experiment with more complex networks you might need to increase this. Likewise if you add regularization that slows training. """ -num_epochs = 5000 +num_epochs = 150 """ A critical parameter that can dramatically affect whether training @@ -20,6 +20,6 @@ learning_rate = 1e-2 momentum = 0.01 -alpha = 1e-2 +alpha = 1e1 -beta = 1e-4 +beta = 1e2 diff --git a/losses.py b/losses.py index 5564ca33..4c7b5750 100644 --- a/losses.py +++ b/losses.py @@ -34,7 +34,7 @@ class YourModel(tf.keras.Model): #print(self.content_image.shape, self.style_image.shape) - self.optimizer = tf.keras.optimizers.Adam() + self.optimizer = tf.keras.optimizers.Adam(1e-2) self.vgg16 = [ # Block 1 @@ -112,6 +112,8 @@ class YourModel(tf.keras.Model): content_l = self.content_loss(self.photo_layers, input_layers) style_l = self.style_loss(self.art_layers, input_layers) # Equation 7 + print('style_loss', style_l) + print('content_loss', content_l) return (self.alpha * content_l) + (self.beta * style_l) def content_loss(self, photo_layers, input_layers): @@ -164,7 +166,7 @@ class YourModel(tf.keras.Model): for i in range(len(art_layers)): art_layer = art_layers[i] input_layer = input_layers[i] - L_style = tf.math.add(L_style, self.layer_loss(art_layer, input_layer)) + L_style = tf.math.add(L_style, self.layer_loss(art_layer, input_layer)*(1/5)) #print('style loss', L_style) return L_style diff --git a/save.jpg b/save.jpg index 5a6d6522..a77ce43c 100644 Binary files a/save.jpg and b/save.jpg differ -- cgit v1.2.3-70-g09d2 From e7150a281f23175fd289e3e3050db44717df3a4a Mon Sep 17 00:00:00 2001 From: Logan Bauman Date: Sat, 7 May 2022 08:35:31 -0400 Subject: hi --- __pycache__/hyperparameters.cpython-38.pyc | Bin 341 -> 341 bytes __pycache__/losses.cpython-38.pyc | Bin 4610 -> 4599 bytes hyperparameters.py | 6 +++--- losses.py | 2 +- save.jpg | Bin 25059 -> 46187 bytes 5 files changed, 4 insertions(+), 4 deletions(-) (limited to 'hyperparameters.py') diff --git a/__pycache__/hyperparameters.cpython-38.pyc b/__pycache__/hyperparameters.cpython-38.pyc index 40cba873..d0ec2a3e 100644 Binary files a/__pycache__/hyperparameters.cpython-38.pyc and b/__pycache__/hyperparameters.cpython-38.pyc differ diff --git a/__pycache__/losses.cpython-38.pyc b/__pycache__/losses.cpython-38.pyc index ff01b5a7..9cd2e604 100644 Binary files a/__pycache__/losses.cpython-38.pyc and b/__pycache__/losses.cpython-38.pyc differ diff --git a/hyperparameters.py b/hyperparameters.py index fedd4dd7..75d58118 100644 --- a/hyperparameters.py +++ b/hyperparameters.py @@ -9,7 +9,7 @@ Number of epochs. If you experiment with more complex networks you might need to increase this. Likewise if you add regularization that slows training. """ -num_epochs = 150 +num_epochs = 1000 """ A critical parameter that can dramatically affect whether training @@ -20,6 +20,6 @@ learning_rate = 1e-2 momentum = 0.01 -alpha = 1e1 +alpha = 1e-3 -beta = 1e2 +beta = 1e-1 diff --git a/losses.py b/losses.py index 4c7b5750..eba94c36 100644 --- a/losses.py +++ b/losses.py @@ -34,7 +34,7 @@ class YourModel(tf.keras.Model): #print(self.content_image.shape, self.style_image.shape) - self.optimizer = tf.keras.optimizers.Adam(1e-2) + self.optimizer = tf.keras.optimizers.Adam() self.vgg16 = [ # Block 1 diff --git a/save.jpg b/save.jpg index a77ce43c..a3bc2d80 100644 Binary files a/save.jpg and b/save.jpg differ -- cgit v1.2.3-70-g09d2 From 2b9c0ee6cdf8a2ca5bde047c2192b7a02e3e5a30 Mon Sep 17 00:00:00 2001 From: Logan Bauman Date: Sat, 7 May 2022 08:41:22 -0400 Subject: hi --- hyperparameters.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) (limited to 'hyperparameters.py') diff --git a/hyperparameters.py b/hyperparameters.py index 75d58118..7e4da417 100644 --- a/hyperparameters.py +++ b/hyperparameters.py @@ -20,6 +20,6 @@ learning_rate = 1e-2 momentum = 0.01 -alpha = 1e-3 +alpha = 1e-2 beta = 1e-1 -- cgit v1.2.3-70-g09d2 From 6973d4b62013ef541e674a33fe7bf483f1e6f2ee Mon Sep 17 00:00:00 2001 From: Logan Bauman Date: Sat, 7 May 2022 08:45:29 -0400 Subject: hi --- hyperparameters.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) (limited to 'hyperparameters.py') diff --git a/hyperparameters.py b/hyperparameters.py index 7e4da417..4fa563df 100644 --- a/hyperparameters.py +++ b/hyperparameters.py @@ -9,7 +9,7 @@ Number of epochs. If you experiment with more complex networks you might need to increase this. Likewise if you add regularization that slows training. """ -num_epochs = 1000 +num_epochs = 10000 """ A critical parameter that can dramatically affect whether training -- cgit v1.2.3-70-g09d2 From e1a8c82606afce6d34d7ed858ee625c291dd4685 Mon Sep 17 00:00:00 2001 From: Logan Bauman Date: Sat, 7 May 2022 09:05:55 -0400 Subject: hi --- hyperparameters.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) (limited to 'hyperparameters.py') diff --git a/hyperparameters.py b/hyperparameters.py index 4fa563df..5e98267a 100644 --- a/hyperparameters.py +++ b/hyperparameters.py @@ -9,7 +9,7 @@ Number of epochs. If you experiment with more complex networks you might need to increase this. Likewise if you add regularization that slows training. """ -num_epochs = 10000 +num_epochs = 2500 """ A critical parameter that can dramatically affect whether training -- cgit v1.2.3-70-g09d2 From 590f0c2a57310e52871fe6517275984c466c6c34 Mon Sep 17 00:00:00 2001 From: Logan Bauman Date: Sat, 7 May 2022 09:14:32 -0400 Subject: hi --- hyperparameters.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) (limited to 'hyperparameters.py') diff --git a/hyperparameters.py b/hyperparameters.py index 5e98267a..d09e6623 100644 --- a/hyperparameters.py +++ b/hyperparameters.py @@ -9,7 +9,7 @@ Number of epochs. If you experiment with more complex networks you might need to increase this. Likewise if you add regularization that slows training. """ -num_epochs = 2500 +num_epochs = 5000 """ A critical parameter that can dramatically affect whether training -- cgit v1.2.3-70-g09d2 From 8de640d7e79ed448fe483dd455be897c9bdaa9d7 Mon Sep 17 00:00:00 2001 From: Logan Bauman Date: Sat, 7 May 2022 10:00:28 -0400 Subject: hi --- __pycache__/hyperparameters.cpython-38.pyc | Bin 315 -> 332 bytes __pycache__/losses.cpython-38.pyc | Bin 4582 -> 4599 bytes hyperparameters.py | 2 +- 3 files changed, 1 insertion(+), 1 deletion(-) (limited to 'hyperparameters.py') diff --git a/__pycache__/hyperparameters.cpython-38.pyc b/__pycache__/hyperparameters.cpython-38.pyc index 5e563f46..f1b885e8 100644 Binary files a/__pycache__/hyperparameters.cpython-38.pyc and b/__pycache__/hyperparameters.cpython-38.pyc differ diff --git a/__pycache__/losses.cpython-38.pyc b/__pycache__/losses.cpython-38.pyc index fd07d03c..9cd2e604 100644 Binary files a/__pycache__/losses.cpython-38.pyc and b/__pycache__/losses.cpython-38.pyc differ diff --git a/hyperparameters.py b/hyperparameters.py index d09e6623..337064bd 100644 --- a/hyperparameters.py +++ b/hyperparameters.py @@ -9,7 +9,7 @@ Number of epochs. If you experiment with more complex networks you might need to increase this. Likewise if you add regularization that slows training. """ -num_epochs = 5000 +num_epochs = 3000 """ A critical parameter that can dramatically affect whether training -- cgit v1.2.3-70-g09d2 From 97da9dd646158e6d029833449b7f201486165ce1 Mon Sep 17 00:00:00 2001 From: Logan Bauman Date: Sat, 7 May 2022 10:22:28 -0400 Subject: hi --- hyperparameters.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) (limited to 'hyperparameters.py') diff --git a/hyperparameters.py b/hyperparameters.py index 337064bd..9128dc27 100644 --- a/hyperparameters.py +++ b/hyperparameters.py @@ -22,4 +22,4 @@ momentum = 0.01 alpha = 1e-2 -beta = 1e-1 +beta = 5e-1 -- cgit v1.2.3-70-g09d2 From b674ade7012b0826dfafb7906ab9828527df1fc7 Mon Sep 17 00:00:00 2001 From: Logan Bauman Date: Sat, 7 May 2022 10:32:23 -0400 Subject: hi --- hyperparameters.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) (limited to 'hyperparameters.py') diff --git a/hyperparameters.py b/hyperparameters.py index 9128dc27..308522d7 100644 --- a/hyperparameters.py +++ b/hyperparameters.py @@ -9,7 +9,7 @@ Number of epochs. If you experiment with more complex networks you might need to increase this. Likewise if you add regularization that slows training. """ -num_epochs = 3000 +num_epochs = 4000 """ A critical parameter that can dramatically affect whether training -- cgit v1.2.3-70-g09d2 From 08558c7fd96038cd7209d1562e0b1af50aa34917 Mon Sep 17 00:00:00 2001 From: Logan Bauman Date: Sat, 7 May 2022 10:52:43 -0400 Subject: hi --- hyperparameters.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) (limited to 'hyperparameters.py') diff --git a/hyperparameters.py b/hyperparameters.py index 308522d7..d5eb2f47 100644 --- a/hyperparameters.py +++ b/hyperparameters.py @@ -9,7 +9,7 @@ Number of epochs. If you experiment with more complex networks you might need to increase this. Likewise if you add regularization that slows training. """ -num_epochs = 4000 +num_epochs = 5000 """ A critical parameter that can dramatically affect whether training -- cgit v1.2.3-70-g09d2 From 96b69eac697100499e54c8ac18c10313bd460f3b Mon Sep 17 00:00:00 2001 From: Logan Bauman Date: Sat, 7 May 2022 11:21:54 -0400 Subject: hi --- hyperparameters.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) (limited to 'hyperparameters.py') diff --git a/hyperparameters.py b/hyperparameters.py index d5eb2f47..594368f7 100644 --- a/hyperparameters.py +++ b/hyperparameters.py @@ -9,7 +9,7 @@ Number of epochs. If you experiment with more complex networks you might need to increase this. Likewise if you add regularization that slows training. """ -num_epochs = 5000 +num_epochs = 2000 """ A critical parameter that can dramatically affect whether training @@ -22,4 +22,4 @@ momentum = 0.01 alpha = 1e-2 -beta = 5e-1 +beta = 7e-1 -- cgit v1.2.3-70-g09d2 From 677431e27153473c264c42a99c3374a0d0747c73 Mon Sep 17 00:00:00 2001 From: Logan Bauman Date: Sat, 7 May 2022 11:39:36 -0400 Subject: hi --- hyperparameters.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) (limited to 'hyperparameters.py') diff --git a/hyperparameters.py b/hyperparameters.py index 594368f7..2b1f0d2f 100644 --- a/hyperparameters.py +++ b/hyperparameters.py @@ -9,7 +9,7 @@ Number of epochs. If you experiment with more complex networks you might need to increase this. Likewise if you add regularization that slows training. """ -num_epochs = 2000 +num_epochs = 4000 """ A critical parameter that can dramatically affect whether training -- cgit v1.2.3-70-g09d2 From 761298c2dfba6bfcc44cc7bf3d15b4704feee1d0 Mon Sep 17 00:00:00 2001 From: Logan Bauman Date: Sat, 7 May 2022 12:13:44 -0400 Subject: hi --- hyperparameters.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) (limited to 'hyperparameters.py') diff --git a/hyperparameters.py b/hyperparameters.py index 2b1f0d2f..a691a3ac 100644 --- a/hyperparameters.py +++ b/hyperparameters.py @@ -22,4 +22,4 @@ momentum = 0.01 alpha = 1e-2 -beta = 7e-1 +beta = 1e0 -- cgit v1.2.3-70-g09d2 From ee00602bb385e815b856bb2ef2f85e467d89ad2e Mon Sep 17 00:00:00 2001 From: Logan Bauman Date: Sat, 7 May 2022 12:25:30 -0400 Subject: hi --- hyperparameters.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) (limited to 'hyperparameters.py') diff --git a/hyperparameters.py b/hyperparameters.py index a691a3ac..f87b92c6 100644 --- a/hyperparameters.py +++ b/hyperparameters.py @@ -22,4 +22,4 @@ momentum = 0.01 alpha = 1e-2 -beta = 1e0 +beta = 1e1 -- cgit v1.2.3-70-g09d2 From c073b01a935900abd8357a3005bbfe8bdce266fc Mon Sep 17 00:00:00 2001 From: Logan Bauman Date: Sat, 7 May 2022 12:39:17 -0400 Subject: hi --- hyperparameters.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) (limited to 'hyperparameters.py') diff --git a/hyperparameters.py b/hyperparameters.py index f87b92c6..d68e26d1 100644 --- a/hyperparameters.py +++ b/hyperparameters.py @@ -9,7 +9,7 @@ Number of epochs. If you experiment with more complex networks you might need to increase this. Likewise if you add regularization that slows training. """ -num_epochs = 4000 +num_epochs = 5000 """ A critical parameter that can dramatically affect whether training @@ -22,4 +22,4 @@ momentum = 0.01 alpha = 1e-2 -beta = 1e1 +beta = 5e1 -- cgit v1.2.3-70-g09d2 From 6c68b1b36e0858b96656551695f62d2b556529ca Mon Sep 17 00:00:00 2001 From: Logan Bauman Date: Sat, 7 May 2022 12:53:27 -0400 Subject: hi --- hyperparameters.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) (limited to 'hyperparameters.py') diff --git a/hyperparameters.py b/hyperparameters.py index d68e26d1..e1b8ca75 100644 --- a/hyperparameters.py +++ b/hyperparameters.py @@ -9,7 +9,7 @@ Number of epochs. If you experiment with more complex networks you might need to increase this. Likewise if you add regularization that slows training. """ -num_epochs = 5000 +num_epochs = 10000 """ A critical parameter that can dramatically affect whether training -- cgit v1.2.3-70-g09d2 From 5dc4bbaf666f27f0f19e3cff0991bd66071182f0 Mon Sep 17 00:00:00 2001 From: Logan Bauman Date: Sat, 7 May 2022 13:21:04 -0400 Subject: hi --- hyperparameters.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) (limited to 'hyperparameters.py') diff --git a/hyperparameters.py b/hyperparameters.py index e1b8ca75..80141fcf 100644 --- a/hyperparameters.py +++ b/hyperparameters.py @@ -9,7 +9,7 @@ Number of epochs. If you experiment with more complex networks you might need to increase this. Likewise if you add regularization that slows training. """ -num_epochs = 10000 +num_epochs = 20000 """ A critical parameter that can dramatically affect whether training -- cgit v1.2.3-70-g09d2 From 1b54bfa9de44f29ca2046d1eeb0ab174ae0dadbd Mon Sep 17 00:00:00 2001 From: Logan Bauman Date: Sat, 7 May 2022 14:16:42 -0400 Subject: hi --- hyperparameters.py | 4 ++-- losses.py | 2 +- 2 files changed, 3 insertions(+), 3 deletions(-) (limited to 'hyperparameters.py') diff --git a/hyperparameters.py b/hyperparameters.py index 80141fcf..6c82a745 100644 --- a/hyperparameters.py +++ b/hyperparameters.py @@ -9,14 +9,14 @@ Number of epochs. If you experiment with more complex networks you might need to increase this. Likewise if you add regularization that slows training. """ -num_epochs = 20000 +num_epochs = 7000 """ A critical parameter that can dramatically affect whether training succeeds or fails. The value for this depends significantly on which optimizer is used. Refer to the default learning rate parameter """ -learning_rate = 1e-2 +learning_rate = 2e-3 momentum = 0.01 diff --git a/losses.py b/losses.py index eba94c36..f08a734c 100644 --- a/losses.py +++ b/losses.py @@ -34,7 +34,7 @@ class YourModel(tf.keras.Model): #print(self.content_image.shape, self.style_image.shape) - self.optimizer = tf.keras.optimizers.Adam() + self.optimizer = tf.keras.optimizers.Adam(hp.learning_rate) self.vgg16 = [ # Block 1 -- cgit v1.2.3-70-g09d2 From 26c962e42ca990f741c3667924543265fc38492f Mon Sep 17 00:00:00 2001 From: Logan Bauman Date: Sat, 7 May 2022 14:42:18 -0400 Subject: hi --- hyperparameters.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) (limited to 'hyperparameters.py') diff --git a/hyperparameters.py b/hyperparameters.py index 6c82a745..4760afe4 100644 --- a/hyperparameters.py +++ b/hyperparameters.py @@ -16,10 +16,10 @@ A critical parameter that can dramatically affect whether training succeeds or fails. The value for this depends significantly on which optimizer is used. Refer to the default learning rate parameter """ -learning_rate = 2e-3 +learning_rate = 4e-3 momentum = 0.01 alpha = 1e-2 -beta = 5e1 +beta = 1e2 -- cgit v1.2.3-70-g09d2 From 0832fee98ab118402a52af605bf53d0f5129996e Mon Sep 17 00:00:00 2001 From: Logan Bauman Date: Sat, 7 May 2022 14:45:55 -0400 Subject: hi --- hyperparameters.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) (limited to 'hyperparameters.py') diff --git a/hyperparameters.py b/hyperparameters.py index 4760afe4..ec424dfa 100644 --- a/hyperparameters.py +++ b/hyperparameters.py @@ -9,7 +9,7 @@ Number of epochs. If you experiment with more complex networks you might need to increase this. Likewise if you add regularization that slows training. """ -num_epochs = 7000 +num_epochs = 5000 """ A critical parameter that can dramatically affect whether training -- cgit v1.2.3-70-g09d2 From f46f67c74b19c1db98b30e3f03c166f043079587 Mon Sep 17 00:00:00 2001 From: Logan Bauman Date: Sat, 7 May 2022 15:03:42 -0400 Subject: hi --- hyperparameters.py | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) (limited to 'hyperparameters.py') diff --git a/hyperparameters.py b/hyperparameters.py index ec424dfa..6c82a745 100644 --- a/hyperparameters.py +++ b/hyperparameters.py @@ -9,17 +9,17 @@ Number of epochs. If you experiment with more complex networks you might need to increase this. Likewise if you add regularization that slows training. """ -num_epochs = 5000 +num_epochs = 7000 """ A critical parameter that can dramatically affect whether training succeeds or fails. The value for this depends significantly on which optimizer is used. Refer to the default learning rate parameter """ -learning_rate = 4e-3 +learning_rate = 2e-3 momentum = 0.01 alpha = 1e-2 -beta = 1e2 +beta = 5e1 -- cgit v1.2.3-70-g09d2 From 00991837cc0bbb62b98ab3024ea795a18cf2dde8 Mon Sep 17 00:00:00 2001 From: Logan Bauman Date: Sat, 7 May 2022 15:58:27 -0400 Subject: hi --- data/content.jpeg | Bin 0 -> 68587 bytes data/style.jpeg | Bin 0 -> 65581 bytes hyperparameters.py | 4 ++-- 3 files changed, 2 insertions(+), 2 deletions(-) create mode 100644 data/content.jpeg create mode 100644 data/style.jpeg (limited to 'hyperparameters.py') diff --git a/data/content.jpeg b/data/content.jpeg new file mode 100644 index 00000000..398d20ea Binary files /dev/null and b/data/content.jpeg differ diff --git a/data/style.jpeg b/data/style.jpeg new file mode 100644 index 00000000..0c4015df Binary files /dev/null and b/data/style.jpeg differ diff --git a/hyperparameters.py b/hyperparameters.py index 6c82a745..a15d04ac 100644 --- a/hyperparameters.py +++ b/hyperparameters.py @@ -9,14 +9,14 @@ Number of epochs. If you experiment with more complex networks you might need to increase this. Likewise if you add regularization that slows training. """ -num_epochs = 7000 +num_epochs = 10000 """ A critical parameter that can dramatically affect whether training succeeds or fails. The value for this depends significantly on which optimizer is used. Refer to the default learning rate parameter """ -learning_rate = 2e-3 +learning_rate = .002 momentum = 0.01 -- cgit v1.2.3-70-g09d2