D
|
JaTinySegmenterTokenizer::segment()
|
0
|
96
|
13
|
13.05
|
0
|
0
|
0
|
D
|
Onoi\Tesa\Sanitizer::sanitizeWith()
|
0
|
16
|
10
|
10.02
|
0
|
0
|
0
|
D
|
Tesa\CharacterExaminer::contains()
|
0
|
14
|
9
|
9
|
0
|
0
|
0
|
C
|
JaCompoundGroupTokenizer::tokenize()
|
0
|
13
|
7
|
7.01
|
0
|
0
|
0
|
C
|
Tokenizer\NGramTokenizer::createNGrams()
|
0
|
14
|
7
|
7
|
0
|
0
|
0
|
B
|
TransliteratorTest::characterProvider()
|
0
|
82
|
1
|
?
|
0
|
0
|
0
|
B
|
JaTinySegmenterTokenizerTest::stringProvider()
|
0
|
119
|
1
|
?
|
0
|
0
|
0
|
B
|
NGramTokenizerTest::stringProvider()
|
0
|
78
|
1
|
?
|
0
|
0
|
0
|
B
|
IcuWordBoundaryTokenizer::createTokens()
|
0
|
13
|
6
|
42
|
0
|
0
|
0
|
B
|
Tesa\Transliterator::transliterate()
|
0
|
16
|
5
|
5
|
0
|
0
|
0
|
B
|
CdbStopwordAnalyzer::createCdbByLanguage()
|
0
|
14
|
4
|
4
|
0
|
0
|
0
|
B
|
IcuWordBoundaryTokenizerTest::stringProvider()
|
0
|
53
|
1
|
?
|
0
|
0
|
0
|
B
|
SanitizerTest::testSanitizeByStopwordsToIncludeExemptionWithMinLengthRestriction()
|
0
|
26
|
1
|
?
|
0
|
0
|
0
|
B
|
CJKSimpleCharacterRegExTokenizerTest::testTokenizeWithEnabledExemptionList()
|
0
|
18
|
1
|
?
|
0
|
0
|
0
|
B
|
JaTokenizerTest::tinyTextProvider()
|
0
|
15
|
1
|
?
|
0
|
0
|
0
|