Explain Document Pipeline for Danish

Description

The explain_document_md is a pretrained pipeline that we can use to process text with a simple pipeline that performs basic processing steps. It performs most of the common text processing tasks on your dataframe

Predicted Entities

Download Copy S3 URI

How to use

from sparknlp.pretrained import PretrainedPipelinein
pipeline = PretrainedPipeline('explain_document_md', lang = 'da')
annotations =  pipeline.fullAnnotate(""Hej fra John Snow Labs! "")[0]
annotations.keys()
val pipeline = new PretrainedPipeline("explain_document_md", lang = "da")
val result = pipeline.fullAnnotate("Hej fra John Snow Labs! ")(0)
import nlu
text = [""Hej fra John Snow Labs! ""]
result_df = nlu.load('da.explain.md').predict(text)
result_df

Results

Results


|    | document                     | sentence                    | token                                   | lemma                                   | pos                                        | embeddings                   | ner                                   | entities            |
|---:|:-----------------------------|:----------------------------|:----------------------------------------|:----------------------------------------|:-------------------------------------------|:-----------------------------|:--------------------------------------|:--------------------|
|  0 | ['Hej fra John Snow Labs! '] | ['Hej fra John Snow Labs!'] | ['Hej', 'fra', 'John', 'Snow', 'Labs!'] | ['Hej', 'fra', 'John', 'Snow', 'Labs!'] | ['NOUN', 'ADP', 'PROPN', 'PROPN', 'PROPN'] | [[0.4006600081920624,.,...]] | ['O', 'O', 'B-PER', 'I-PER', 'I-PER'] | ['John Snow Labs!'] |


{:.model-param}

Model Information

Model Name: explain_document_md
Type: pipeline
Compatibility: Spark NLP 4.4.2+
License: Open Source
Edition: Official
Language: da
Size: 463.5 MB

Included Models

  • DocumentAssembler
  • SentenceDetector
  • TokenizerModel
  • LemmatizerModel
  • PerceptronModel
  • WordEmbeddingsModel
  • NerDLModel
  • NerConverter