Explain Document pipeline for Finnish (explain_document_lg)

Description

The explain_document_lg is a pretrained pipeline that we can use to process text with a simple pipeline that performs basic processing steps and recognizes entities . It performs most of the common text processing tasks on your dataframe

Predicted Entities

Download Copy S3 URI

How to use

from sparknlp.pretrained import PretrainedPipelinein
pipeline = PretrainedPipeline('explain_document_lg', lang = 'fi')
annotations =  pipeline.fullAnnotate(""Hei John Snow Labs! "")[0]
annotations.keys()
val pipeline = new PretrainedPipeline("explain_document_lg", lang = "fi")
val result = pipeline.fullAnnotate("Hei John Snow Labs! ")(0)
import nlu
text = [""Hei John Snow Labs! ""]
result_df = nlu.load('fi.explain.lg').predict(text)
result_df

Results

Results


|    | document                 | sentence                | token                            | lemma                            | pos                                 | embeddings                   | ner                              | entities            |
|---:|:-------------------------|:------------------------|:---------------------------------|:---------------------------------|:------------------------------------|:-----------------------------|:---------------------------------|:--------------------|
|  0 | ['Hei John Snow Labs! '] | ['Hei John Snow Labs!'] | ['Hei', 'John', 'Snow', 'Labs!'] | ['hei', 'John', 'Snow', 'Labs!'] | ['INTJ', 'PROPN', 'PROPN', 'PROPN'] | [[0.0639619976282119,.,...]] | ['O', 'B-PRO', 'I-PRO', 'I-PRO'] | ['John Snow Labs!'] |


{:.model-param}

Model Information

Model Name: explain_document_lg
Type: pipeline
Compatibility: Spark NLP 4.4.2+
License: Open Source
Edition: Official
Language: fi
Size: 2.5 GB

Included Models

  • DocumentAssembler
  • SentenceDetector
  • TokenizerModel
  • LemmatizerModel
  • PerceptronModel
  • WordEmbeddingsModel
  • NerDLModel
  • NerConverter