Add new SentenceTransformer model
Browse files- README.md +4 -5
- config.json +1 -1
- model.safetensors +2 -2
README.md
CHANGED
|
@@ -165,9 +165,9 @@ print(embeddings.shape)
|
|
| 165 |
# Get the similarity scores for the embeddings
|
| 166 |
similarities = model.similarity(embeddings, embeddings)
|
| 167 |
print(similarities)
|
| 168 |
-
# tensor([[1.0000, 1.0000, 0.
|
| 169 |
-
# [1.0000, 1.0000, 0.
|
| 170 |
-
# [0.
|
| 171 |
```
|
| 172 |
|
| 173 |
<!--
|
|
@@ -288,7 +288,6 @@ You can finetune this model on your own dataset.
|
|
| 288 |
- `eval_strategy`: steps
|
| 289 |
- `per_device_train_batch_size`: 1152
|
| 290 |
- `per_device_eval_batch_size`: 1152
|
| 291 |
-
- `gradient_accumulation_steps`: 2
|
| 292 |
- `weight_decay`: 0.001
|
| 293 |
- `adam_beta2`: 0.98
|
| 294 |
- `adam_epsilon`: 1e-06
|
|
@@ -316,7 +315,7 @@ You can finetune this model on your own dataset.
|
|
| 316 |
- `per_device_eval_batch_size`: 1152
|
| 317 |
- `per_gpu_train_batch_size`: None
|
| 318 |
- `per_gpu_eval_batch_size`: None
|
| 319 |
-
- `gradient_accumulation_steps`:
|
| 320 |
- `eval_accumulation_steps`: None
|
| 321 |
- `torch_empty_cache_steps`: None
|
| 322 |
- `learning_rate`: 5e-05
|
|
|
|
| 165 |
# Get the similarity scores for the embeddings
|
| 166 |
similarities = model.similarity(embeddings, embeddings)
|
| 167 |
print(similarities)
|
| 168 |
+
# tensor([[1.0000, 1.0000, 0.3655],
|
| 169 |
+
# [1.0000, 1.0000, 0.3655],
|
| 170 |
+
# [0.3655, 0.3655, 1.0000]])
|
| 171 |
```
|
| 172 |
|
| 173 |
<!--
|
|
|
|
| 288 |
- `eval_strategy`: steps
|
| 289 |
- `per_device_train_batch_size`: 1152
|
| 290 |
- `per_device_eval_batch_size`: 1152
|
|
|
|
| 291 |
- `weight_decay`: 0.001
|
| 292 |
- `adam_beta2`: 0.98
|
| 293 |
- `adam_epsilon`: 1e-06
|
|
|
|
| 315 |
- `per_device_eval_batch_size`: 1152
|
| 316 |
- `per_gpu_train_batch_size`: None
|
| 317 |
- `per_gpu_eval_batch_size`: None
|
| 318 |
+
- `gradient_accumulation_steps`: 1
|
| 319 |
- `eval_accumulation_steps`: None
|
| 320 |
- `torch_empty_cache_steps`: None
|
| 321 |
- `learning_rate`: 5e-05
|
config.json
CHANGED
|
@@ -4,7 +4,7 @@
|
|
| 4 |
],
|
| 5 |
"attention_probs_dropout_prob": 0.1,
|
| 6 |
"classifier_dropout": null,
|
| 7 |
-
"dtype": "
|
| 8 |
"gradient_checkpointing": false,
|
| 9 |
"hidden_act": "gelu",
|
| 10 |
"hidden_dropout_prob": 0.1,
|
|
|
|
| 4 |
],
|
| 5 |
"attention_probs_dropout_prob": 0.1,
|
| 6 |
"classifier_dropout": null,
|
| 7 |
+
"dtype": "float32",
|
| 8 |
"gradient_checkpointing": false,
|
| 9 |
"hidden_act": "gelu",
|
| 10 |
"hidden_dropout_prob": 0.1,
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:41ddaf37362229cd67dce6f7afa240ab7dfba4b1462c9571d8e6e5d6ebed73eb
|
| 3 |
+
size 90864192
|