Spaces:
Runtime error
Runtime error
smhavens
commited on
Update README.md
Browse files
README.md
CHANGED
|
@@ -11,17 +11,45 @@ pinned: false
|
|
| 11 |
|
| 12 |
Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
|
| 13 |
|
| 14 |
-
##
|
| 15 |
-
###
|
| 16 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 17 |
|
| 18 |
This database uses a text with label format, with each label being an integer between 0 and 3, relating to the 4 main categories of the news: World (0), Sports (1), Business (2), Sci/Tech (3).
|
| 19 |
|
| 20 |
I chose this one because of the larger variety of categories compared to sentiment databases, with the themes/categories theoretically being more closely related to analogies. I also chose ag_news because, as a news source, it should avoid slang and other potential hiccups that databases using tweets or general reviews will have.
|
| 21 |
|
| 22 |
-
|
| 23 |
[sentence-transformers/all-MiniLM-L6-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2)
|
| 24 |
|
| 25 |
Because my focus is on using embeddings to evaluate analogies for the AnalogyArcade, I focused my model search for those in the sentence-transformers category, as they are readily made for embedding usage. I chose all-MiniLM-L6-v2 because of its high usage and good reviews: it is a well trained model but smaller and more efficient than its previous version.
|
| 26 |
|
| 27 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 11 |
|
| 12 |
Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
|
| 13 |
|
| 14 |
+
## Model Types
|
| 15 |
+
### Baseline
|
| 16 |
+
For my dataset, I made use of relbert/analogy_questions on huggingface, which has all data in the format of:
|
| 17 |
+
```
|
| 18 |
+
"stem": ["raphael", "painter"],
|
| 19 |
+
"answer": 2,
|
| 20 |
+
"choice": [["andersen", "plato"],
|
| 21 |
+
["reading", "berkshire"],
|
| 22 |
+
["marx", "philosopher"],
|
| 23 |
+
["tolstoi", "edison"]]
|
| 24 |
+
```
|
| 25 |
+
|
| 26 |
+
For a baseline, if I were to do a random selection for answer to train the system on (so the stem analogy is compared to a random choice among the answers), then there would only be a 25% baseline for correct categorization and comparison.
|
| 27 |
+
|
| 28 |
+
### Bag-of-Words Model
|
| 29 |
+
For comparison, I made use of my previously trained bag-of-words model from [our previous project](https://github.com/smhavens/NLPHW03).
|
| 30 |
+
|
| 31 |
+
### Fine-Tuning
|
| 32 |
+
#### Dataset
|
| 33 |
+
[analogy questions dataset](https://huggingface.co/datasets/relbert/analogy_questions)
|
| 34 |
|
| 35 |
This database uses a text with label format, with each label being an integer between 0 and 3, relating to the 4 main categories of the news: World (0), Sports (1), Business (2), Sci/Tech (3).
|
| 36 |
|
| 37 |
I chose this one because of the larger variety of categories compared to sentiment databases, with the themes/categories theoretically being more closely related to analogies. I also chose ag_news because, as a news source, it should avoid slang and other potential hiccups that databases using tweets or general reviews will have.
|
| 38 |
|
| 39 |
+
#### Pre-trained model
|
| 40 |
[sentence-transformers/all-MiniLM-L6-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2)
|
| 41 |
|
| 42 |
Because my focus is on using embeddings to evaluate analogies for the AnalogyArcade, I focused my model search for those in the sentence-transformers category, as they are readily made for embedding usage. I chose all-MiniLM-L6-v2 because of its high usage and good reviews: it is a well trained model but smaller and more efficient than its previous version.
|
| 43 |
|
| 44 |
+
### In-Context
|
| 45 |
+
|
| 46 |
+
## User Guide
|
| 47 |
+
### Introduction
|
| 48 |
+
|
| 49 |
+
### Usage
|
| 50 |
+
|
| 51 |
+
### Documentation
|
| 52 |
+
|
| 53 |
+
### Experiments
|
| 54 |
+
|
| 55 |
+
### Limitations
|