Update README.md
Browse files
README.md
CHANGED
|
@@ -12,7 +12,10 @@ It is trained by fine-tuning the LLaMA2-7B model on the [MALLS-v0.1](https://hug
|
|
| 12 |
**Model type:**
|
| 13 |
This repo contains the LoRA delta weights for direct translation LogicLLaMA, which directly translates the NL statement into a FOL rule in one go.
|
| 14 |
We also provide the delta weights for other modes:
|
| 15 |
-
- [
|
|
|
|
|
|
|
|
|
|
| 16 |
|
| 17 |
**License:**
|
| 18 |
Apache License 2.0
|
|
|
|
| 12 |
**Model type:**
|
| 13 |
This repo contains the LoRA delta weights for direct translation LogicLLaMA, which directly translates the NL statement into a FOL rule in one go.
|
| 14 |
We also provide the delta weights for other modes:
|
| 15 |
+
- [direct translation LogicLLaMA-7B](https://huggingface.co/yuan-yang/LogicLLaMA-7b-direct-translate-delta-v0.1)
|
| 16 |
+
- [naive correction LogicLLaMA-7B](https://huggingface.co/yuan-yang/LogicLLaMA-7b-naive-correction-delta-v0.1)
|
| 17 |
+
- [direct translation LogicLLaMA-13B](https://huggingface.co/yuan-yang/LogicLLaMA-13b-direct-translate-delta-v0.1)
|
| 18 |
+
- [naive correction LogicLLaMA-13B](https://huggingface.co/yuan-yang/LogicLLaMA-13b-naive-correction-delta-v0.1)
|
| 19 |
|
| 20 |
**License:**
|
| 21 |
Apache License 2.0
|