Skip to content

Commit 2f5e057

Browse files
CLIP : Enable CLIP.from_preset() Kaggle tests (#2357)
* update kaggle handle and weights conversion file * code reformat --------- Co-authored-by: Divyashree Sreepathihalli <divyashreepathihalli>
1 parent cc450d7 commit 2f5e057

File tree

4 files changed

+3233
-355
lines changed

4 files changed

+3233
-355
lines changed

keras_cv/models/feature_extractor/clip/clip_model.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -72,7 +72,7 @@ def __init__(
7272
vision_patch_size=32,
7373
context_length=77,
7474
vocab_size=49408,
75-
transformer_width=768,
75+
transformer_width=512,
7676
transformer_heads=8,
7777
transformer_layers=12,
7878
**kwargs,

keras_cv/models/feature_extractor/clip/clip_model_test.py

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -52,9 +52,9 @@ def test_clip_model_golden_values(self):
5252
processed_image, processed_text, attention_mask
5353
)
5454
print(image_logits)
55-
self.assertAllClose(image_logits, [[2.932678, 2.932678, 2.932675]])
55+
self.assertAllClose(image_logits, [[1.896713, 1.896713, 1.896713]])
5656
self.assertAllClose(
57-
text_logits, ops.transpose([[2.932678, 2.932678, 2.932675]])
57+
text_logits, ops.transpose([[1.896713, 1.896713, 1.896713]])
5858
)
5959

6060
def test_clip_preprocessor(self):
@@ -77,8 +77,8 @@ def test_clip_preprocessor_tf_data(self):
7777

7878
@pytest.mark.large
7979
def test_presets(self):
80-
self.skipTest("TODO: Enable after Kaggle model is public")
81-
model = CLIP.from_preset("clip-vit-base-patch32")
80+
# self.skipTest("TODO: Enable after Kaggle model is public")
81+
model = CLIP.from_preset("clip-vit-base-patch16")
8282
processed_image = np.ones(shape=[1, 224, 224, 3])
8383
processed_text = np.ones(shape=[3, 77])
8484
attention_mask = np.ones(shape=[3, 77])
@@ -109,7 +109,7 @@ def test_text_encoder_golden_values(self):
109109
print(model.text_embeddings)
110110
self.assertAllClose(
111111
model.text_embeddings[0, :3],
112-
[-0.018502, 0.000906, 0.020372],
112+
[0.007531, -0.038361, -0.035686],
113113
)
114114

115115
@pytest.mark.large # Saving is slow, so mark these large.

keras_cv/models/feature_extractor/clip/clip_presets.py

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -28,7 +28,7 @@
2828
"official_name": "CLIP",
2929
"path": "clip",
3030
},
31-
"kaggle_handle": "kaggle://keras/clip/keras/clip-vit-base-patch16/2",
31+
"kaggle_handle": "kaggle://keras/clip/keras/clip-vit-base-patch16/4",
3232
},
3333
"clip-vit-base-patch32": {
3434
"metadata": {
@@ -44,7 +44,7 @@
4444
"official_name": "CLIP",
4545
"path": "clip",
4646
},
47-
"kaggle_handle": "kaggle://keras/clip/keras/clip-vit-base-patch32/2",
47+
"kaggle_handle": "kaggle://keras/clip/keras/clip-vit-base-patch32/4",
4848
},
4949
"clip-vit-large-patch14": {
5050
"metadata": {
@@ -60,7 +60,7 @@
6060
"official_name": "CLIP",
6161
"path": "clip",
6262
},
63-
"kaggle_handle": "kaggle://keras/clip/keras/clip-vit-large-patch14/2",
63+
"kaggle_handle": "kaggle://keras/clip/keras/clip-vit-large-patch14/4",
6464
},
6565
"clip-vit-large-patch14-336": {
6666
"metadata": {
@@ -76,6 +76,6 @@
7676
"official_name": "CLIP",
7777
"path": "clip",
7878
},
79-
"kaggle_handle": "kaggle://keras/clip/keras/clip-vit-large-patch14-336/2", # noqa: E501
79+
"kaggle_handle": "kaggle://keras/clip/keras/clip-vit-large-patch14-336/4", # noqa: E501
8080
},
8181
}

0 commit comments

Comments
 (0)