Commit graph

377 commits

Author SHA1 Message Date
1b489518d0
segmenter: add LayerStack2Image to custom_objects 2022-10-26 17:05:50 +01:00
48ae8a5c20
LossContrastive: normalise features as per the paper 2022-10-26 16:52:56 +01:00
843cc8dc7b
contrastive: rewrite the loss function.
The CLIP paper *does* kinda make sense I think
2022-10-26 16:45:45 +01:00
fad1399c2d
convnext: whitespace 2022-10-26 16:45:20 +01:00
1d872cb962
contrastive: fix initial temperature value
It should be 1/0.07, but we had it set to 0.07......
2022-10-26 16:45:01 +01:00
f994d449f1
Layer2Image: fix 2022-10-25 21:32:17 +01:00
6a29105f56
model_segmentation: stack not reshape 2022-10-25 21:25:15 +01:00
98417a3e06
prepare for NCE loss
.....but Tensorflow's implementation looks to be for supervised models :-(
2022-10-25 21:15:05 +01:00
bb0679a509
model_segmentation: don't softmax twice 2022-10-25 21:11:48 +01:00
f2e2ca1484
model_contrastive: make water encoder significantly shallower 2022-10-24 20:52:31 +01:00
a6b07a49cb
count water/nowater pixels in Jupyter Notebook 2022-10-24 18:05:34 +01:00
a8b101bdae
dataset_predict: add shape_water_desired 2022-10-24 18:05:13 +01:00
587c1dfafa
train_predict: revamp jsonl handling 2022-10-21 16:53:08 +01:00
8195318a42
SparseCategoricalAccuracy: losses → metrics 2022-10-21 16:51:20 +01:00
612735aaae
rename shuffle arg 2022-10-21 16:35:45 +01:00
c98d8d05dd
segmentation: use the right accuracy 2022-10-21 16:17:05 +01:00
bb0258f5cd
flip squeeze operator ordering 2022-10-21 15:38:57 +01:00
af26964c6a
batched_iterator: reset i_item after every time 2022-10-21 15:35:43 +01:00
c5b1501dba
train-predict fixup 2022-10-21 15:27:39 +01:00
42aea7a0cc
plt.close() fixup 2022-10-21 15:23:54 +01:00
12dad3bc87
vis/segmentation: fix titles 2022-10-21 15:22:35 +01:00
0cb2de5d06
train-preedict: close matplotlib after we've finished
they act like file handles
2022-10-21 15:19:31 +01:00
81e53efd9c
PNG: create output dir if doesn't exist 2022-10-21 15:17:39 +01:00
3f7db6fa78
fix embedding confusion 2022-10-21 15:15:59 +01:00
847cd97ec4
fixup 2022-10-21 14:26:58 +01:00
0e814b7e98
Contraster → Segmenter 2022-10-21 14:25:43 +01:00
1b658a1b7c
train-predict: can't destructure array when iterating generator
....it seems to lead to undefined behaviour or something
2022-10-20 19:34:04 +01:00
aed2348a95
train_predict: fixup 2022-10-20 15:42:33 +01:00
cc6679c609
batch data; use generator 2022-10-20 15:22:29 +01:00
d306853c42
use right daataset 2022-10-20 15:16:24 +01:00
59cfa4a89a
basename paths 2022-10-20 15:11:14 +01:00
4d8ae21a45
update cli help text 2022-10-19 17:31:42 +01:00
200076596b
finish train_predict 2022-10-19 17:26:40 +01:00
488f78fca5
pretrain_predict: default to parallel_reads=0 2022-10-19 16:59:45 +01:00
63e909d9fc
datasets: add shuffle=True/False to get_filepaths.
This is important because otherwise it SCAMBLES the filenames, which is a disaster for making predictions in the right order....!
2022-10-19 16:52:07 +01:00
fe43ddfbf9
start implementing driver for train_predict, but not finished yet 2022-10-18 19:37:55 +01:00
b3ea189d37
segmentation: softmax the output 2022-10-13 21:02:57 +01:00
f121bfb981
fixup summaryfile 2022-10-13 17:54:42 +01:00
5c35c0cee4
model_segmentation: document; remove unused args 2022-10-13 17:50:16 +01:00
f12e6ab905
No need for a CLI arg for feature_dim_in - metadata should contain this 2022-10-13 17:37:16 +01:00
e201372252
write quick Jupyter notebook to test data
....I'm paranoid
2022-10-13 17:27:17 +01:00
ae53130e66
layout 2022-10-13 14:54:20 +01:00
7933564c66
typo 2022-10-12 17:33:54 +01:00
dbe4fb0eab
train: add slurm job file 2022-10-12 17:27:10 +01:00
6423bf6702
LayerConvNeXtGamma: avoid adding an EagerTensor to config
Very weird how this is a problem when it wasn't before..
2022-10-12 17:12:07 +01:00
32f5200d3b
pass model_arch properly 2022-10-12 16:50:06 +01:00
5933fb1061
fixup 2022-10-11 19:23:41 +01:00
c45b90764e
segmentation: adds xxtiny, but unsure if it's small enough 2022-10-11 19:22:37 +01:00
f4a2c742d9
typo 2022-10-11 19:19:23 +01:00
11f91a7cf4
train: add --arch; default to convnext_i_xtiny 2022-10-11 19:18:01 +01:00
5666c5a0d9
typo 2022-10-10 18:12:51 +01:00
131c0a0a5b
pretrain-predict: create dir if not exists 2022-10-10 18:00:55 +01:00
deede32241
slurm-pretrain: limit memory usage 2022-10-10 17:45:29 +01:00
13a8f3f511
pretrain-predict: only queue pretrain-plot if I we output jsonl 2022-10-10 17:11:10 +01:00
ffcb2e3735
pretrain-predict: queue for the actual input 2022-10-10 16:53:28 +01:00
f883986eaa
Bugfix: modeset to enable TFRecordWriter instead of bare handle 2022-10-06 20:07:59 +01:00
e9a8e2eb57
fixup 2022-10-06 19:23:31 +01:00
9f3ae96894
finish wiring for --water-size 2022-10-06 19:21:50 +01:00
5dac70aa08
typo 2022-10-06 19:17:03 +01:00
2960d3b645
exception → warning 2022-10-06 18:26:40 +01:00
0ee6703c1e
Add todo and comment 2022-10-03 19:06:56 +01:00
2b182214ea
typo 2022-10-03 17:53:10 +01:00
92c380bff5
fiddle with Conv2DTranspose
you need to set the `stride` argument to actually get it to upscale..... :P
2022-10-03 17:51:41 +01:00
d544553800
fixup 2022-10-03 17:33:06 +01:00
058e3b6248
model_segmentation: cast float → int 2022-10-03 17:31:36 +01:00
04e5ae0c45
model_segmentation: redo reshape
much cheese was applied :P
2022-10-03 17:27:52 +01:00
deffe69202
typo 2022-10-03 16:59:36 +01:00
fc6d2dabc9
Upscale first, THEN convnext... 2022-10-03 16:38:43 +01:00
6a0790ff50
convnext_inverse: add returns; change ordering 2022-10-03 16:32:09 +01:00
fe813cb46d
slurm predict: fix plotting subcall 2022-10-03 16:03:26 +01:00
e51087d0a9
add reshape layer 2022-09-28 18:22:48 +01:00
a336cdee90
and continues 2022-09-28 18:18:10 +01:00
de47a883d9
missing units 2022-09-28 18:17:22 +01:00
b5e08f92fe
the long night continues 2022-09-28 18:14:09 +01:00
dc159ecfdb
and again 2022-09-28 18:11:46 +01:00
4cf0485e32
fixup... again 2022-09-28 18:10:11 +01:00
030d8710b6
fixup 2022-09-28 18:08:31 +01:00
4ee7f2a0d6
add water thresholding 2022-09-28 18:07:26 +01:00
404dc30f08
and again 2022-09-28 17:39:09 +01:00
4cd8fc6ded
segmentation: param name fix 2022-09-28 17:37:42 +01:00
41ba980d69
segmentationP implement dataset parser 2022-09-28 17:19:21 +01:00
d618e6f8d7
pretrain-predict: params.json → metadata.jsonl 2022-09-28 16:35:22 +01:00
e9e6139c7a
typo 2022-09-28 16:28:18 +01:00
3dee3d8908
update cli help 2022-09-28 16:23:47 +01:00
b836f7f70c
again 2022-09-27 19:06:41 +01:00
52dff130dd
slurm pretraian predict: moar logging 2022-09-27 19:05:54 +01:00
2d24174e0a
slurm pretrain predict: add $OUTPUT 2022-09-27 18:58:02 +01:00
7d0e3913ae
fix logging 2022-09-27 18:51:58 +01:00
d765b3b14e
fix crash 2022-09-27 18:43:43 +01:00
2cd59a01a5
slurm-pretrain-plot: add ARGS 2022-09-27 18:41:35 +01:00
f4d1d1d77e
just wh 2022-09-27 18:25:45 +01:00
4c24d69ae6
$d → +d 2022-09-27 18:17:07 +01:00
cdb19b4d9f
fixup 2022-09-27 18:13:21 +01:00
c4d3c16873
add some logging 2022-09-27 18:10:58 +01:00
3772c3227e
fixup 2022-09-27 17:57:21 +01:00
dbfa45a016
write params.json properly 2022-09-27 17:49:54 +01:00
a5455dc22a
fix import 2022-09-27 17:41:24 +01:00
d6ff3fb2ce
pretrain_predict fix write mode 2022-09-27 17:38:12 +01:00
f95fd8f9e4
pretrain-predict: add .tfrecord output function 2022-09-27 16:59:31 +01:00
30b8dd063e
fixup 2022-09-27 15:54:37 +01:00
3cf99587e4
Contraster: add gamma layer to load_model 2022-09-27 15:53:52 +01:00
9e9852d066
UMAP: 100k random
no labels.
2022-09-27 15:52:45 +01:00
58c65bdc86
slurm: allow predictions on gpu 2022-09-23 19:21:57 +01:00
d59de41ebb
embeddings: change title rendering; make even moar widererer
We need to see that parallel coordinates  plot in detail
2022-09-23 18:56:39 +01:00
b4ddb24589
slurm plot: compute → highmem 2022-09-22 18:27:58 +01:00
df12470e78
flip conda and time
hopefully we can capture the exit code this way
2022-09-21 14:40:28 +01:00
5252a81238
vis: don't call add_subplot 2022-09-20 19:06:21 +01:00
32bb55652b
slurm predict: autoqueue UMAP plot 2022-09-16 19:36:57 +01:00
24c5263bf0
slurm predict-plot: fixup 2022-09-16 19:27:26 +01:00
7e9c119b04
slurm: fix job name 2022-09-16 19:20:59 +01:00
1574529704
slurm pretrain-predict: move to compute; make exclusive just in case
also shortent o 3 days
2022-09-16 19:17:42 +01:00
d7f5958af0
slurm: write new job files 2022-09-16 19:00:43 +01:00
a552cc4dad
ai vis: make parallel coordinates wider 2022-09-16 18:51:49 +01:00
a70794e661
umap: no min_dist 2022-09-16 17:09:09 +01:00
5778fc51f7
embeddings: fix title; remove colourmap 2022-09-16 17:08:04 +01:00
4fd852d782
fixup 2022-09-16 16:44:35 +01:00
fcab227f6a
cheese: set label for everything to 1 2022-09-16 16:42:05 +01:00
b31645bd5d
pretrain-plot: fix crash; remove water code
the model doesn't save the water encoder at this time
2022-09-16 16:24:07 +01:00
a5f03390ef
pretrain-plot: handle open w -> r 2022-09-16 16:14:30 +01:00
1103ae5561
ai: tweak progress bar 2022-09-16 16:07:16 +01:00
1e35802d2b
ai: fix embed i/o 2022-09-16 16:02:27 +01:00
ed94da7492
fixup 2022-09-16 15:51:26 +01:00
366db658a8
ds predict: fix filenames in 2022-09-16 15:45:22 +01:00
e333dcba9c
tweak projection head 2022-09-16 15:36:01 +01:00
6defd24000
bugfix: too many values to unpack 2022-09-15 19:56:17 +01:00
e3c8277255
ai: tweak the segmentation model structure 2022-09-15 19:54:50 +01:00
1bc8a5bf13
ai: fix crash 2022-09-15 19:37:06 +01:00
bd64986332
ai: implement batched_iterator to replace .batch()
...apparently .batch() means you get a BatchedDataset or whatever when you iterate it like a tf.function instead of the actual tensor :-/
2022-09-15 19:16:38 +01:00
ccd256c00a
embed rainfall radar, not both 2022-09-15 17:37:04 +01:00
2c74676902
predict → predict_on_batch 2022-09-15 17:31:50 +01:00
f036e79098
fixup 2022-09-15 17:09:26 +01:00
d5f1a26ba3
disable prefetching when predicting a thing 2022-09-15 17:09:09 +01:00
8770638022
ai: call .predict(), not the model itself 2022-09-14 17:41:01 +01:00
a96cefde62
ai: predict oops 2022-09-14 17:37:48 +01:00
fa3165a5b2
dataset: simplify dataset_predict 2022-09-14 17:33:17 +01:00
279e27c898
fixup 2022-09-14 17:16:49 +01:00
fad3313ede
fixup 2022-09-14 17:14:04 +01:00
1e682661db
ai: kwargs in from_checkpoint 2022-09-14 17:11:06 +01:00
6bda24d4da
ai: how did I miss that?!
bugfix ah/c
2022-09-14 16:53:43 +01:00
decdd434d8
ai from_checkpoint: bugfix 2022-09-14 16:49:01 +01:00
c9e00ea485
pretrain_predict: fix import 2022-09-14 16:02:36 +01:00
f568d8d19f
io.open → handle_open
this was we get transparent .gz support
2022-09-14 16:01:22 +01:00
9b25186541
fix --only-gpu 2022-09-14 15:55:21 +01:00
a9c9c70d13
typo 2022-09-14 15:17:59 +01:00
fb8f884487
add umap dependencies 2022-09-14 15:16:45 +01:00
1876a8883c
ai pretrain-predict: fix - → _ in cli parsing 2022-09-14 15:12:07 +01:00
f97b771922
make --help display help 2022-09-14 15:03:07 +01:00
3c1aef5913
update help 2022-09-13 19:36:56 +01:00
206257f9f5
ai pretrain_predict: no need to plot here anymore 2022-09-13 19:35:44 +01:00
7685ec3e8b
implement ability to embed & plot pretrained embeddings 2022-09-13 19:18:59 +01:00
7130c4fdf8
start implementing core image segmentation model 2022-09-07 17:45:38 +01:00
22620a1854
ai: implement saving only the rainfall encoder 2022-09-06 19:48:46 +01:00
4c4358c3e5
whitespace 2022-09-06 16:24:11 +01:00
4202821d98
typo 2022-09-06 15:37:36 +01:00
3e13ad12c8
ai Bugfix LayerContrastiveEncoder: channels → input_channels
for consistency
2022-09-05 23:53:16 +01:00
ead8009425
pretrain: add CLI arg for size of watch prediction width/height 2022-09-05 15:36:40 +01:00
9d39215dd5
dataset: drop incomplete batches 2022-09-05 15:36:10 +01:00
c94f5d042e
ai: slurm fixup again 2022-09-02 19:13:56 +01:00
7917820e59
ai: slurm fixup 2022-09-02 19:11:54 +01:00
cd104190e8
slurm: no need to be exclusive anymore 2022-09-02 19:09:45 +01:00
a9dede7bfe
ai: n=10 slurm 2022-09-02 19:09:07 +01:00
42de502f99
slurm-pretrain: set name 2022-09-02 19:08:16 +01:00
457c2eef0d
ai: fixup 2022-09-02 19:06:54 +01:00
a3f03b6d8d
slurm-pretrain: prepare 2022-09-02 19:05:18 +01:00
1d1533d160
ai: how did things get this confusing 2022-09-02 18:51:46 +01:00
1c5defdcd6
ai: cats 2022-09-02 18:45:23 +01:00
6135bcd0cd
fixup 2022-09-02 18:41:31 +01:00
c33c0a0899
ai: these shapes are so annoying 2022-09-02 18:39:24 +01:00
88acd54a97
ai: off-by-one 2022-09-02 18:13:48 +01:00
23bcd294c4
AI: 0.75 → 0.5? 2022-09-02 18:12:51 +01:00
b0e1aeac35
ai: knew it 2022-09-02 18:10:29 +01:00
ef0f15960d
send correct water shape to 2022-09-02 18:09:36 +01:00
ad156a9a00
ai dataset: centre crop the water data to 75% original size
this should both help the model and reduce memory usage
2022-09-02 18:05:32 +01:00
efe41b96ec
fixup 2022-09-02 17:58:02 +01:00
f8ee0afca1
ai: fix arch_name plumbing 2022-09-02 17:57:07 +01:00
3d44831080
Add NO_PREFETCH env var 2022-09-02 17:55:04 +01:00
3e0ca6a315
ai: fix summary file writing; make water encoder smaller 2022-09-02 17:51:45 +01:00
389216b391
reordering 2022-09-02 17:31:19 +01:00
b9d01ddadc
summary logger → summarywriter 2022-09-02 17:28:00 +01:00
9f7f4af784
dataset: properly resize 2022-09-02 17:09:38 +01:00
c78b6562ef
debug adjustment 2022-09-02 17:00:21 +01:00
c89677abd7
dataset: explicit reshape 2022-09-02 16:57:59 +01:00
c066ea331b
dataset: namespace → dict
Python is so annoying
2022-09-02 16:07:44 +01:00
dbff15d4a9
add bug comment 2022-09-01 19:05:50 +01:00
3e4128c0a8
resize rainfall to be 1/2 size of current 2022-09-01 18:47:07 +01:00
8a86728b54
ai: how did I miss that 2022-09-01 18:07:53 +01:00
b2f1ba29bb
debug a 2022-09-01 18:05:33 +01:00
e2f621e212
even moar debug logging 2022-09-01 17:49:51 +01:00
2663123407
moar logging 2022-09-01 17:40:23 +01:00
c2fcb3b954
ai: channels_first → channels_last 2022-09-01 17:06:18 +01:00
f1d7973f22
ai: add dummy label 2022-09-01 17:01:00 +01:00
17d42fe899
ai: json.dumps 2022-09-01 16:21:52 +01:00
940f7aa1b5
ai: set self.model 2022-09-01 16:20:23 +01:00
f1be5fe2bd
ai: add missing arguments to LossContrastive 2022-09-01 16:14:00 +01:00
ddb375e906
ai: another typo 2022-09-01 16:12:05 +01:00
cb3e1d3a23
ai: fixup 2022-09-01 16:10:50 +01:00
e4c95bc7e3
typo 2022-09-01 16:09:24 +01:00
cfbbe8e8cf
ai: global? really? 2022-09-01 16:06:24 +01:00
4952ead094
ai: try the nonlocal keyword? I'm unsure what's going on here
....clearly I need to read up on scoping in python.
2022-09-01 16:02:37 +01:00
1d9d0fbb73
requirements.txt:a dd tensorflow_addons 2022-09-01 15:56:02 +01:00