diff --git a/aimodel/src/lib/ai/components/convnext.py b/aimodel/src/lib/ai/components/convnext.py index 8e706ae..0202ae7 100644 --- a/aimodel/src/lib/ai/components/convnext.py +++ b/aimodel/src/lib/ai/components/convnext.py @@ -129,7 +129,7 @@ def convnext( def add_convnext_block(y, dim, drop_prob=0, prefix=""): skip = y - + y = tf.keras.layers.DepthwiseConv2D( kernel_size=7, padding='same', diff --git a/aimodel/src/lib/ai/model_rainfallwater_segmentation.py b/aimodel/src/lib/ai/model_rainfallwater_segmentation.py index 8f7e16a..53826ca 100644 --- a/aimodel/src/lib/ai/model_rainfallwater_segmentation.py +++ b/aimodel/src/lib/ai/model_rainfallwater_segmentation.py @@ -29,7 +29,7 @@ def model_rainfallwater_segmentation(metadata, feature_dim_in, shape_water_out, layer_next = do_convnext_inverse(layer_next, arch_name="convnext_i_tiny") # TODO: An attention layer here instead of a dense layer, with a skip connection perhaps? - raise Exception("Error: read and implement attention from https://ieeexplore.ieee.org/document/9076883") + logger.warn("Warning: TODO implement attention from https://ieeexplore.ieee.org/document/9076883") layer_next = tf.keras.layers.Dense(32)(layer_next) layer_next = tf.keras.layers.Conv2D(1, kernel_size=1, activation="softmax", padding="same")(layer_next)