radoslavralev commited on
Commit
7d10fbe
·
verified ·
1 Parent(s): e1b5203

Add new SentenceTransformer model

Browse files
Files changed (3) hide show
  1. README.md +4 -5
  2. config.json +1 -1
  3. model.safetensors +2 -2
README.md CHANGED
@@ -165,9 +165,9 @@ print(embeddings.shape)
165
  # Get the similarity scores for the embeddings
166
  similarities = model.similarity(embeddings, embeddings)
167
  print(similarities)
168
- # tensor([[1.0000, 1.0000, 0.3433],
169
- # [1.0000, 1.0000, 0.3433],
170
- # [0.3433, 0.3433, 1.0000]])
171
  ```
172
 
173
  <!--
@@ -288,7 +288,6 @@ You can finetune this model on your own dataset.
288
  - `eval_strategy`: steps
289
  - `per_device_train_batch_size`: 1152
290
  - `per_device_eval_batch_size`: 1152
291
- - `gradient_accumulation_steps`: 2
292
  - `weight_decay`: 0.001
293
  - `adam_beta2`: 0.98
294
  - `adam_epsilon`: 1e-06
@@ -316,7 +315,7 @@ You can finetune this model on your own dataset.
316
  - `per_device_eval_batch_size`: 1152
317
  - `per_gpu_train_batch_size`: None
318
  - `per_gpu_eval_batch_size`: None
319
- - `gradient_accumulation_steps`: 2
320
  - `eval_accumulation_steps`: None
321
  - `torch_empty_cache_steps`: None
322
  - `learning_rate`: 5e-05
 
165
  # Get the similarity scores for the embeddings
166
  similarities = model.similarity(embeddings, embeddings)
167
  print(similarities)
168
+ # tensor([[1.0000, 1.0000, 0.3655],
169
+ # [1.0000, 1.0000, 0.3655],
170
+ # [0.3655, 0.3655, 1.0000]])
171
  ```
172
 
173
  <!--
 
288
  - `eval_strategy`: steps
289
  - `per_device_train_batch_size`: 1152
290
  - `per_device_eval_batch_size`: 1152
 
291
  - `weight_decay`: 0.001
292
  - `adam_beta2`: 0.98
293
  - `adam_epsilon`: 1e-06
 
315
  - `per_device_eval_batch_size`: 1152
316
  - `per_gpu_train_batch_size`: None
317
  - `per_gpu_eval_batch_size`: None
318
+ - `gradient_accumulation_steps`: 1
319
  - `eval_accumulation_steps`: None
320
  - `torch_empty_cache_steps`: None
321
  - `learning_rate`: 5e-05
config.json CHANGED
@@ -4,7 +4,7 @@
4
  ],
5
  "attention_probs_dropout_prob": 0.1,
6
  "classifier_dropout": null,
7
- "dtype": "bfloat16",
8
  "gradient_checkpointing": false,
9
  "hidden_act": "gelu",
10
  "hidden_dropout_prob": 0.1,
 
4
  ],
5
  "attention_probs_dropout_prob": 0.1,
6
  "classifier_dropout": null,
7
+ "dtype": "float32",
8
  "gradient_checkpointing": false,
9
  "hidden_act": "gelu",
10
  "hidden_dropout_prob": 0.1,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8aa70c5a8c7f662f3c19176fc258b7ec3880da54fdf54f99cdf12236f5d5bc06
3
- size 45437864
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41ddaf37362229cd67dce6f7afa240ab7dfba4b1462c9571d8e6e5d6ebed73eb
3
+ size 90864192