-
-
11 Aug 2011 02:55:56 UTC
- Distribution: Lucy
- Source (raw)
- Browse (raw)
- Changes
- How to Contribute
- Issues (1)
- Testers (3 / 16 / 0)
- Kwalitee
Bus factor: 1- License: apache_2_0
- Perl: v5.8.3
- Activity
24 month- Tools
- Download (1.02MB)
- MetaCPAN Explorer
- Permissions
- Subscribe to distribution
- Permalinks
- This version
- Latest version
and 1 contributors- The Apache Lucy Project <lucy-dev at incubator dot apache dot org>
- Dependencies
- JSON::XS
- and possibly others
- Reverse dependencies
- CPAN Testers List
- Dependency graph
NAME
Lucy::Analysis::RegexTokenizer - Split a string into tokens.
SYNOPSIS
my $whitespace_tokenizer = Lucy::Analysis::RegexTokenizer->new( pattern => '\S+' ); # or... my $word_char_tokenizer = Lucy::Analysis::RegexTokenizer->new( pattern => '\w+' ); # or... my $apostrophising_tokenizer = Lucy::Analysis::RegexTokenizer->new; # Then... once you have a tokenizer, put it into a PolyAnalyzer: my $polyanalyzer = Lucy::Analysis::PolyAnalyzer->new( analyzers => [ $case_folder, $word_char_tokenizer, $stemmer ], );
DESCRIPTION
Generically, "tokenizing" is a process of breaking up a string into an array of "tokens". For instance, the string "three blind mice" might be tokenized into "three", "blind", "mice".
Lucy::Analysis::RegexTokenizer decides where it should break up the text based on a regular expression compiled from a supplied
pattern
matching one token. If our source string is..."Eats, Shoots and Leaves."
... then a "whitespace tokenizer" with a
pattern
of"\\S+"
produces...Eats, Shoots and Leaves.
... while a "word character tokenizer" with a
pattern
of"\\w+"
produces...Eats Shoots and Leaves
... the difference being that the word character tokenizer skips over punctuation as well as whitespace when determining token boundaries.
CONSTRUCTORS
new( [labeled params] )
my $word_char_tokenizer = Lucy::Analysis::RegexTokenizer->new( pattern => '\w+', # required );
pattern - A string specifying a Perl-syntax regular expression which should match one token. The default value is
\w+(?:[\x{2019}']\w+)*
, which matches "it's" as well as "it" and "O'Henry's" as well as "Henry".
INHERITANCE
Lucy::Analysis::RegexTokenizer isa Lucy::Analysis::Analyzer isa Lucy::Object::Obj.
Module Install Instructions
To install Lucy::Simple, copy and paste the appropriate command in to your terminal.
cpanm Lucy::Simple
perl -MCPAN -e shell install Lucy::Simple
For more information on module installation, please visit the detailed CPAN module installation guide.