From 160e78fad7d2e75ab8ec52d0fa0d9374979da8ca Mon Sep 17 00:00:00 2001
From: hannandarryl <hannandarryl@gmail.com>
Date: Thu, 21 Apr 2022 13:58:39 +0000
Subject: [PATCH] trying to fix yolov4 issue

---
 sparse_coding_torch/keras_model.py      | 14 +++++++-------
 sparse_coding_torch/train_classifier.py |  2 +-
 sparse_coding_torch/video_loader.py     | 16 ++++++++--------
 yolov4                                  |  1 -
 4 files changed, 16 insertions(+), 17 deletions(-)
 delete mode 160000 yolov4

diff --git a/sparse_coding_torch/keras_model.py b/sparse_coding_torch/keras_model.py
index 4850759..a55afc3 100644
--- a/sparse_coding_torch/keras_model.py
+++ b/sparse_coding_torch/keras_model.py
@@ -258,7 +258,7 @@ class PNBClassifier(keras.layers.Layer):
 
         self.max_pool = keras.layers.MaxPooling2D(pool_size=8, strides=8)
 #         self.conv_1 = keras.layers.Conv2D(32, kernel_size=8, strides=4, activation='relu', padding='valid')
-        self.conv_2 = keras.layers.Conv2D(48, kernel_size=4, strides=2, activation='relu', padding='valid')
+        self.conv_2 = keras.layers.Conv2D(48, kernel_size=(8, 16), strides=4, activation='relu', padding='valid')
 #         self.conv_3 = keras.layers.Conv2D(24, kernel_size=4, strides=2, activation='relu', padding='valid')
 #         self.conv_4 = keras.layers.Conv2D(24, kernel_size=4, strides=2, activation='relu', padding='valid')
 
@@ -266,8 +266,8 @@ class PNBClassifier(keras.layers.Layer):
 
         self.dropout = keras.layers.Dropout(0.5)
 
-#         self.ff_1 = keras.layers.Dense(1000, activation='relu', use_bias=True)
-#         self.ff_2 = keras.layers.Dense(500, activation='relu', use_bias=True)
+        self.ff_1 = keras.layers.Dense(1000, activation='relu', use_bias=True)
+        self.ff_2 = keras.layers.Dense(500, activation='relu', use_bias=True)
         self.ff_3 = keras.layers.Dense(100, activation='relu', use_bias=True)
         self.ff_4 = keras.layers.Dense(1)
 
@@ -280,10 +280,10 @@ class PNBClassifier(keras.layers.Layer):
 #         x = self.conv_3(x)
 #         x = self.conv_4(x)
         x = self.flatten(x)
-#         x = self.ff_1(x)
-#         x = self.dropout(x)
-#         x = self.ff_2(x)
-#         x = self.dropout(x)
+        x = self.ff_1(x)
+        x = self.dropout(x)
+        x = self.ff_2(x)
+        x = self.dropout(x)
         x = self.ff_3(x)
         x = self.dropout(x)
         x = self.ff_4(x)
diff --git a/sparse_coding_torch/train_classifier.py b/sparse_coding_torch/train_classifier.py
index 8f0276c..a188862 100644
--- a/sparse_coding_torch/train_classifier.py
+++ b/sparse_coding_torch/train_classifier.py
@@ -50,7 +50,7 @@ if __name__ == "__main__":
     
     if args.dataset == 'pnb':
         image_height = 285
-        image_width = 235
+        image_width = 470
     elif args.dataset == 'ptx':
         image_height = 100
         image_width = 200
diff --git a/sparse_coding_torch/video_loader.py b/sparse_coding_torch/video_loader.py
index e7060d1..a855e5e 100644
--- a/sparse_coding_torch/video_loader.py
+++ b/sparse_coding_torch/video_loader.py
@@ -98,11 +98,11 @@ def get_yolo_regions(yolo_model, clip, is_right):
         lower_y = upper_y - 285
         
         if is_right:
-            lower_x = center_x - 235
+            lower_x = center_x - 470
             upper_x = center_x
         else:
             lower_x = center_x
-            upper_x = center_x + 235
+            upper_x = center_x + 470
 
         trimmed_clip = clip[:, :, lower_y:upper_y, lower_x:upper_x]
         
@@ -172,8 +172,8 @@ class PNBLoader(Dataset):
         self.augmentation = augmentation
         self.labels = [name for name in listdir(video_path) if isdir(join(video_path, name))]
         
-        clip_cache_file = 'clip_cache_pnb.pt'
-        clip_cache_final_file = 'clip_cache_pnb_final.pt'
+        clip_cache_file = 'clip_cache_pnb_double.pt'
+        clip_cache_final_file = 'clip_cache_pnb_final_double.pt'
         
         region_labels = load_pnb_region_labels(join(video_path, 'sme_region_labels.csv'))
 
@@ -222,15 +222,15 @@ class PNBLoader(Dataset):
 #                                 cv2.imwrite('test.png', vc_sub[0, 0, :, :].unsqueeze(2).numpy())
 
                                 for clip in get_yolo_regions(yolo_model, vc_sub, is_right):
-#                                     if self.transform:
-#                                         clip = self.transform(clip)
+                                    if self.transform:
+                                        clip = self.transform(clip)
                                         
 #                                     print(clip[0, 0, :, :].size())
 #                                     cv2.imwrite('test_yolo.png', clip[0, 0, :, :].unsqueeze(2).numpy())
 #                                     print(clip.shape)
 #                                     tv.io.write_video('test_yolo.mp4', clip.swapaxes(0,1).swapaxes(1,2).swapaxes(2,3).numpy(), fps=20)
-                                    print(path)
-                                    raise Exception
+#                                     print(path)
+#                                     raise Exception
 
                                     self.clips.append(('Negatives', clip, self.videos[vid_idx][2]))
 
diff --git a/yolov4 b/yolov4
deleted file mode 160000
index 9f16748..0000000
--- a/yolov4
+++ /dev/null
@@ -1 +0,0 @@
-Subproject commit 9f16748aa3f45ff240608da4bd9b1216a29127f5
-- 
GitLab