-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy pathto-do.txt
65 lines (53 loc) · 3.66 KB
/
to-do.txt
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
create another environment to work with dataset_preprocessing ipynb files
put all "import " statements in dataset_...2.ipynb at the top cell
ChatGPT's opinion on making metadata on its own dataframe vs while generating the images:
Both options have their own advantages and disadvantages, and the choice between them ultimately depends on the specific requirements and constraints of your DL training process.
Option 1: Storing image metadata in a dataframe and passing it as a second input along with the image itself has the advantage of separating the image data and metadata into two distinct inputs, which can simplify the data processing and transformation steps in your DL model. This approach allows you to manipulate the metadata independently of the image data, and can make it easier to apply certain transformations or filters to the metadata.
However, this approach also has some drawbacks. For example, it may require additional storage space and computational resources to manage the metadata dataframe separately from the image data. Additionally, passing the metadata as a separate input can increase the complexity of your DL model, as you need to design a way for the model to integrate the metadata with the image data.
Option 2: Fetching the image metadata while the model is fetching images in a custom data generator function can have the advantage of reducing the storage and computational overhead required to manage the metadata separately. This approach allows you to load the metadata on-the-fly as the model fetches the images, which can be more efficient if you have a large dataset or limited computational resources.
However, this approach also has some drawbacks. For example, it may increase the processing time required to load and integrate the metadata with the image data, which could slow down the overall training process. Additionally, this approach may limit your ability to apply certain transformations or filters to the metadata, as it may be difficult to manipulate the metadata while the model is fetching the images.
Ultimately, the choice between these two options depends on the specific requirements and constraints of your DL training process. If you have limited computational resources and need to optimize the storage and processing overhead, Option 2 may be more efficient. On the other hand, if you need more flexibility and control over the metadata and want to separate it from the image data, Option 1 may be more suitable.
94787 train images
20311 val images
so, #steps per epoch (if sampler is not batch):
115098
2962 training steps
634 validation steps
so #steps per epoch (if sampler is batch):
3596
20312 test images
634 testing steps
loss 5926
val_loss 1270
test_loss 0
step_loss 5926
step_val_loss 1270
step_test_loss 0
loss_output_layer_0 43152
acc_output_layer_0 43176
f1_score_output_layer_0 86328
val_loss_output_layer_0 43152
val_acc_output_layer_0 43176
val_f1_score_output_layer_0 86328
val_y_pred_output_layer_0 1380864
val_y_true_output_layer_0 1380864
test_loss_output_layer_0 43152
test_acc_output_layer_0 43176
test_f1_score_output_layer_0 86328
test_y_pred_output_layer_0 1380864
test_y_true_output_layer_0 1380864
step_loss_output_layer_0 43152
step_acc_output_layer_0 43176
step_f1_score_output_layer_0 86328
step_y_pred_output_layer_0 1380864
step_y_true_output_layer_0 1380864
step_val_loss_output_layer_0 43152
step_val_acc_output_layer_0 43176
step_val_f1_score_output_layer_0 86328
step_val_y_pred_output_layer_0 1380864
step_val_y_true_output_layer_0 1380864
step_test_loss_output_layer_0 43152
step_test_acc_output_layer_0 43176
step_test_f1_score_output_layer_0 86328
step_test_y_pred_output_layer_0 1380864
step_test_y_true_output_layer_0 1380864