- java.lang.Object
-
- com.pervasive.datarush.operators.AbstractLogicalOperator
-
- com.pervasive.datarush.operators.StreamingOperator
-
- com.pervasive.datarush.operators.ExecutableOperator
-
- com.pervasive.datarush.analytics.text.TextTokenizer
-
- All Implemented Interfaces:
LogicalOperator
,PipelineOperator<RecordPort>
,RecordPipelineOperator
public class TextTokenizer extends ExecutableOperator implements RecordPipelineOperator
Tokenizes a string field as a TokenizedText object. This operator can be used to tokenize a String field as an object that can then be used for a variety of other text processing tasks. The TextTokenizer operator has two properties: input field and output field. The input field must be of type string while the output field will be of type object. The contents of the string field will be parsed and tokenized creating a TokenizedText object that will be encoded into the output field. This object can then be used for various other processing tasks.
-
-
Constructor Summary
Constructors Constructor Description TextTokenizer()
Default constructor.TextTokenizer(String textField)
Constructor specifying the string field to tokenize.
-
Method Summary
All Methods Instance Methods Concrete Methods Modifier and Type Method Description void
addWordPattern(String wordPattern)
Add a pattern that should be recognized as a single word.protected void
computeMetadata(StreamingMetadataContext ctx)
Implementations must adhere to the following contractsprotected void
execute(ExecutionContext ctx)
Executes the operator.RecordPort
getInput()
Get the input port of this operator.String
getInputField()
Get the string field to tokenize.RecordPort
getOutput()
Get the output port of this operator.String
getOutputField()
Get the tokenized output field.List<String>
getWordPatterns()
Get the list of custom word patterns that will be recognized.void
setInputField(String textField)
Set the string field to tokenize.void
setOutputField(String tokenField)
Set the tokenized output field.void
setWordPatterns(List<String> wordPatterns)
Set the list of custom word patterns that will be recognized.-
Methods inherited from class com.pervasive.datarush.operators.ExecutableOperator
cloneForExecution, getNumInputCopies, getPortSettings, handleInactiveOutput
-
Methods inherited from class com.pervasive.datarush.operators.AbstractLogicalOperator
disableParallelism, getInputPorts, getOutputPorts, newInput, newInput, newOutput, newRecordInput, newRecordInput, newRecordOutput, notifyError
-
Methods inherited from class java.lang.Object
clone, equals, finalize, getClass, hashCode, notify, notifyAll, toString, wait, wait, wait
-
Methods inherited from interface com.pervasive.datarush.operators.LogicalOperator
disableParallelism, getInputPorts, getOutputPorts
-
-
-
-
Constructor Detail
-
TextTokenizer
public TextTokenizer()
Default constructor. UsesetInputField(String)
andsetOutputField(String)
to set the name of the text field to tokenize and its output field.
-
TextTokenizer
public TextTokenizer(String textField)
Constructor specifying the string field to tokenize.- Parameters:
textField
- The name of the field to tokenize
-
-
Method Detail
-
addWordPattern
public void addWordPattern(String wordPattern)
Add a pattern that should be recognized as a single word.- Parameters:
wordPattern
- A regular expression
-
setWordPatterns
public void setWordPatterns(List<String> wordPatterns)
Set the list of custom word patterns that will be recognized.- Parameters:
wordPatterns
- The list of word patterns
-
getWordPatterns
public List<String> getWordPatterns()
Get the list of custom word patterns that will be recognized.- Returns:
- list of word patterns
-
setInputField
public void setInputField(String textField)
Set the string field to tokenize.If this field does not exist in the input, or is not of type String, an exception will be thrown at composition time.
- Parameters:
textField
- The name of the field to tokenize
-
getInputField
public String getInputField()
Get the string field to tokenize.- Returns:
- The name of the field to tokenize
-
setOutputField
public void setOutputField(String tokenField)
Set the tokenized output field.- Parameters:
tokenField
- The name of the token output field
-
getOutputField
public String getOutputField()
Get the tokenized output field.- Returns:
- The name of the token output field
-
getInput
public RecordPort getInput()
Get the input port of this operator.- Specified by:
getInput
in interfacePipelineOperator<RecordPort>
- Returns:
- input port
-
getOutput
public RecordPort getOutput()
Get the output port of this operator.- Specified by:
getOutput
in interfacePipelineOperator<RecordPort>
- Returns:
- output port
-
computeMetadata
protected void computeMetadata(StreamingMetadataContext ctx)
Description copied from class:StreamingOperator
Implementations must adhere to the following contractsGeneral
Regardless of input ports/output port types, all implementations must do the following:- Validation. Validation of configuration should always be performed first.
- Declare parallelizability.. Implementations must declare parallelizability by calling
StreamingMetadataContext.parallelize(ParallelismStrategy)
.
Input record ports
Implementations with input record ports must declare the following:- Required data ordering: Implementations that have data ordering requirements must declare them by calling
- Required data distribution (only applies to parallelizable operators): Implementations that have data distribution requirements must declare them by calling
RecordPort#setRequiredDataOrdering
, otherwise data may arrive in any order.RecordPort#setRequiredDataDistribution
, otherwise data will arrive in anunspecified partial distribution
.RecordPort#getSourceDataDistribution
andRecordPort#getSourceDataOrdering
. These should be viewed as a hints to help chose a more efficient algorithm. In such cases, though, operators must still declare data ordering and data distribution requirements; otherwise there is no guarantee that data will arrive sorted/distributed as required.Output record ports
Implementations with output record ports must declare the following:- Type: Implementations must declare their output type by calling
RecordPort#setType
.
- Output data ordering: Implementations that can make guarantees as to their output
ordering may do so by calling
RecordPort#setOutputDataOrdering
- Output data distribution (only applies to parallelizable operators): Implementations that can make guarantees as to their output
distribution may do so by calling
RecordPort#setOutputDataDistribution
Input model ports
In general, there is nothing special to declare for input model ports. Models are implicitly duplicated to all partitions when going from non-parallel to parallel operators. The case of a model going from a parallel to a non-parallel node is a special case of a "model reducer" operator. In the case of a model reducer, the downstream operator, must declare the following:- Merge handler: Model reducers must declare a merge handler by
calling
AbstractModelPort#setMergeHandler
.
MergeModel
is a convenient, re-usable model reducer, parameterized with a merge-handler.Output model ports
SimpleModelPort
's have no associated metadata and therefore there is never any output metadata to declare.PMMLPort
's, on the other hand, do have associated metadata. For all PMMLPorts, implementations must declare the following:- pmmlModelSpec: Implementations must declare the PMML model spec
by calling
PMMLPort.setPMMLModelSpec
.
- Specified by:
computeMetadata
in classStreamingOperator
- Parameters:
ctx
- the context
-
execute
protected void execute(ExecutionContext ctx)
Description copied from class:ExecutableOperator
Executes the operator. Implementations should adhere to the following contracts:- Following execution, all input ports must be at end-of-data.
- Following execution, all output ports must be at end-of-data.
- Specified by:
execute
in classExecutableOperator
- Parameters:
ctx
- context in which to lookup physical ports bound to logical ports
-
-