mdeberta-ontonotes5

This is a multilingual DeBERTa model fine-tuned for Named Entity Recognition (NER) task. It is based on the rustemgareev/mdeberta-v3-base-lite model.

Usage

from transformers import pipeline

# Initialize the NER pipeline
ner_pipeline = pipeline(
    "token-classification",
    model="rustemgareev/mdeberta-ontonotes5",
    aggregation_strategy="simple"
)

# Example text
text = "Apple Inc. is looking at buying a U.K. startup for $1 billion in London next week."

# Get predictions
entities = ner_pipeline(text)

# Print the results
for entity in entities:
    print(f"Entity: {entity['word']}, Label: {entity['entity_group']}, Score: {entity['score']:.4f}")

# Expected output:
# Entity: Apple Inc., Label: ORGANIZATION, Score: 0.9975
# Entity: U.K., Label: GPE, Score: 0.9956
# Entity: $1 billion, Label: MONEY, Score: 0.9981
# Entity: London, Label: GPE, Score: 0.9981
# Entity: next week, Label: DATE, Score: 0.9940

License

This model is distributed under the MIT License.

Downloads last month
37
Safetensors
Model size
0.2B params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for rustemgareev/mdeberta-ontonotes5

Finetuned
(1)
this model