lukahh commited on
Commit
3bd0734
·
verified ·
1 Parent(s): d63e6aa

Training in progress, step 1000

Browse files
Files changed (3) hide show
  1. config.json +22 -0
  2. model.safetensors +3 -0
  3. training_args.bin +3 -0
config.json ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "openai/clip-vit-base-patch32",
3
+ "architectures": [
4
+ "CLIPModel"
5
+ ],
6
+ "initializer_factor": 1.0,
7
+ "logit_scale_init_value": 2.6592,
8
+ "model_type": "clip",
9
+ "projection_dim": 512,
10
+ "text_config": {
11
+ "bos_token_id": 0,
12
+ "dropout": 0.0,
13
+ "eos_token_id": 2,
14
+ "model_type": "clip_text_model"
15
+ },
16
+ "torch_dtype": "float32",
17
+ "transformers_version": "4.48.2",
18
+ "vision_config": {
19
+ "dropout": 0.0,
20
+ "model_type": "clip_vision_model"
21
+ }
22
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:39150a5ead1c0e2dd489e3ef02a5ae728d50312c995dbaa550828b03b445856b
3
+ size 605156676
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dfc0c6ce95ea59128b0acaa98a95f555504c6254f0bd8b143229e699d4f671fc
3
+ size 5368