You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
3. OpenCLIP instead of huggingface `transformers`[done, testing]
93
93
4. Pre-computing ViT activations
94
94
95
95
I'm going to do each of these independently using a set of runs as references.
96
96
97
97
# 10/22/2024
98
98
99
-
* Removed HookedVisionTransformer (see above)
100
-
* Checkpoint [v6jto37s](https://wandb.ai/samuelstevens/saev/runs/wwb20pa0) worked
99
+
Removed HookedVisionTransformer (see above)
100
+
Checkpoint [v6jto37s](https://wandb.ai/samuelstevens/saev/runs/wwb20pa0) worked for training, analysis, and app data.
101
+
102
+
Testing an implementation using OpenCLIP instead of `transformers`.
103
+
Assuming it works (which seems likely given that the loss curve is identical), then I will pre-compute the activations, save them as a numpy array to disk, and memmap them during training rather than computing them.
104
+
I expect this to take a little bit because I had issues with shuffling and such in the analysis step earlier.
105
+
I think the best strategy is to work backwards.
106
+
The `generate_app_data.py` script doesn't need an activation store at all.
107
+
So I will start with the `analysis.py` script and add a new activations store class that meets the same interface as the original (maybe not for the constructor).
108
+
Then I will verify that the analysis script works correctly.
109
+
110
+
Only after that will I use the new class in training.
111
+
Working with the analysis script is a shorter feedback loop.
0 commit comments