Conditions | 1 |
Paths | 1 |
Total Lines | 495 |
Code Lines | 336 |
Lines | 0 |
Ratio | 0 % |
Changes | 1 | ||
Bugs | 0 | Features | 0 |
Small methods make your code easier to understand, in particular if combined with a good name. Besides, if your method is small, finding a good name is usually much easier.
For example, if you find yourself adding comments to a method's body, this is usually a good sign to extract the commented part to a new method, and use the comment as a starting point when coming up with a good name for this new method.
Commonly applied refactorings include:
If many parameters/temporary variables are present:
1 | <?php declare(strict_types = 1); |
||
22 | public function providerForTestTokenize(): array |
||
23 | { |
||
24 | return [ |
||
25 | [" \n", [new Token(Tokenizer::TOKEN_WHITESPACE, " \n", 0)]], |
||
26 | ['word', [new WordToken('word', 0, '', 'word')]], |
||
27 | ["word\n", [new WordToken('word', 0, '', 'word'), new Token(Tokenizer::TOKEN_WHITESPACE, "\n", 4)]], |
||
28 | ['word ', [new WordToken('word', 0, '', 'word'), new Token(Tokenizer::TOKEN_WHITESPACE, ' ', 4)]], |
||
29 | ['word(', [new WordToken('word', 0, '', 'word'), new GroupBeginToken('(', 4, '(', null)]], |
||
30 | ['word)', [new WordToken('word', 0, '', 'word'), new Token(Tokenizer::TOKEN_GROUP_END, ')', 4)]], |
||
31 | ['šđčćž', [new WordToken('šđčćž', 0, '', 'šđčćž')]], |
||
32 | [ |
||
33 | $jajeNaOko = mb_convert_encoding('🍳', 'UTF-8', 'HTML-ENTITIES'), |
||
34 | [new WordToken($jajeNaOko, 0, '', $jajeNaOko)], |
||
35 | ], |
||
36 | [ |
||
37 | $blah = mb_convert_encoding( |
||
38 | '👩‍👩‍👧‍👧', |
||
39 | 'UTF-8', |
||
40 | 'HTML-ENTITIES' |
||
41 | ), |
||
42 | [new WordToken($blah, 0, '', $blah)], |
||
43 | ], |
||
44 | ['word-word', [new WordToken('word-word', 0, '', 'word-word')]], |
||
45 | [ |
||
46 | "word\nword", |
||
47 | [ |
||
48 | new WordToken('word', 0, '', 'word'), |
||
49 | new Token(Tokenizer::TOKEN_WHITESPACE, "\n", 4), |
||
50 | new WordToken('word', 5, '', 'word'), |
||
51 | ], |
||
52 | ], |
||
53 | [ |
||
54 | 'word word', |
||
55 | [ |
||
56 | new WordToken('word', 0, '', 'word'), |
||
57 | new Token(Tokenizer::TOKEN_WHITESPACE, ' ', 4), |
||
58 | new WordToken('word', 5, '', 'word'), |
||
59 | ], |
||
60 | ], |
||
61 | ['word\\ word', [new WordToken('word\\ word', 0, '', 'word word')]], |
||
62 | ['[a TO b]', [new RangeToken('[a TO b]', 0, '', 'a', 'b', 'inclusive', 'inclusive')]], |
||
63 | ['[a TO b}', [new RangeToken('[a TO b}', 0, '', 'a', 'b', 'inclusive', 'exclusive')]], |
||
64 | ['{a TO b}', [new RangeToken('{a TO b}', 0, '', 'a', 'b', 'exclusive', 'exclusive')]], |
||
65 | ['{a TO b]', [new RangeToken('{a TO b]', 0, '', 'a', 'b', 'exclusive', 'inclusive')]], |
||
66 | [ |
||
67 | '[2017-01-01 TO 2017-01-05]', |
||
68 | [ |
||
69 | new RangeToken( |
||
70 | '[2017-01-01 TO 2017-01-05]', |
||
71 | 0, |
||
72 | '', |
||
73 | '2017-01-01', |
||
74 | '2017-01-05', |
||
75 | 'inclusive', |
||
76 | 'inclusive' |
||
77 | ), |
||
78 | ], |
||
79 | ], |
||
80 | ['[20 TO *]', [new RangeToken('[20 TO *]', 0, '', '20', '*', 'inclusive', 'inclusive')]], |
||
81 | ['[* TO 20]', [new RangeToken('[* TO 20]', 0, '', '*', '20', 'inclusive', 'inclusive')]], |
||
82 | ['"phrase"', [new PhraseToken('"phrase"', 0, '', '"', 'phrase')]], |
||
83 | [ |
||
84 | '"phrase" "phrase"', |
||
85 | [ |
||
86 | new PhraseToken('"phrase"', 0, '', '"', 'phrase'), |
||
87 | new Token(Tokenizer::TOKEN_WHITESPACE, ' ', 8), |
||
88 | new PhraseToken('"phrase"', 9, '', '"', 'phrase'), |
||
89 | ], |
||
90 | ], |
||
91 | ["\"phrase\nphrase\"", [new PhraseToken("\"phrase\nphrase\"", 0, '', '"', "phrase\nphrase")]], |
||
92 | ["'phrase'", [new WordToken("'phrase'", 0, '', "'phrase'")]], |
||
93 | [ |
||
94 | "'phrase' 'phrase'", |
||
95 | [ |
||
96 | new WordToken("'phrase'", 0, '', "'phrase'"), |
||
97 | new Token(Tokenizer::TOKEN_WHITESPACE, ' ', 8), |
||
98 | new WordToken("'phrase'", 9, '', "'phrase'"), |
||
99 | ], |
||
100 | ], |
||
101 | [ |
||
102 | "'phrase\nphrase'", |
||
103 | [ |
||
104 | new WordToken("'phrase", 0, '', "'phrase"), |
||
105 | new Token(Tokenizer::TOKEN_WHITESPACE, "\n", 7), |
||
106 | new WordToken("phrase'", 8, '', "phrase'"), |
||
107 | ], |
||
108 | ], |
||
109 | ['"phrase\"phrase"', [new PhraseToken('"phrase\"phrase"', 0, '', '"', 'phrase"phrase')]], |
||
110 | ["'phrase\\'phrase'", [new WordToken("'phrase\\'phrase'", 0, '', "'phrase\\'phrase'")]], |
||
111 | ['"phrase\'phrase"', [new PhraseToken('"phrase\'phrase"', 0, '', '"', 'phrase\'phrase')]], |
||
112 | [ |
||
113 | "'phrase\"phrase'", |
||
114 | [ |
||
115 | new WordToken("'phrase", 0, '', "'phrase"), |
||
116 | new Token(Tokenizer::TOKEN_BAILOUT, '"', 7), |
||
117 | new WordToken("phrase'", 8, '', "phrase'"), |
||
118 | ], |
||
119 | ], |
||
120 | ['\"not_phrase\"', [new WordToken('\"not_phrase\"', 0, '', '"not_phrase"')]], |
||
121 | ["\\'not_phrase\\'", [new WordToken("\\'not_phrase\\'", 0, '', "\\'not_phrase\\'")]], |
||
122 | [ |
||
123 | '"phrase + - ! ( ) AND OR NOT \\ phrase"', |
||
124 | [ |
||
125 | new PhraseToken( |
||
126 | '"phrase + - ! ( ) AND OR NOT \\ phrase"', |
||
127 | 0, |
||
128 | '', |
||
129 | '"', |
||
130 | 'phrase + - ! ( ) AND OR NOT \\ phrase' |
||
131 | ), |
||
132 | ], |
||
133 | ], |
||
134 | [ |
||
135 | "'word + - ! ( ) AND OR NOT \\ word'", |
||
136 | [ |
||
137 | new WordToken("'word", 0, '', "'word"), |
||
138 | new Token(Tokenizer::TOKEN_WHITESPACE, ' ', 5), |
||
139 | new Token(Tokenizer::TOKEN_MANDATORY, '+', 6), |
||
140 | new Token(Tokenizer::TOKEN_WHITESPACE, ' ', 7), |
||
141 | new Token(Tokenizer::TOKEN_PROHIBITED, '-', 8), |
||
142 | new Token(Tokenizer::TOKEN_WHITESPACE, ' ', 9), |
||
143 | new Token(Tokenizer::TOKEN_LOGICAL_NOT_2, '!', 10), |
||
144 | new Token(Tokenizer::TOKEN_WHITESPACE, ' ', 11), |
||
145 | new GroupBegin('(', 12, '(', ''), |
||
146 | new Token(Tokenizer::TOKEN_WHITESPACE, ' ', 13), |
||
147 | new Token(Tokenizer::TOKEN_GROUP_END, ')', 14), |
||
148 | new Token(Tokenizer::TOKEN_WHITESPACE, ' ', 15), |
||
149 | new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 16), |
||
150 | new Token(Tokenizer::TOKEN_WHITESPACE, ' ', 19), |
||
151 | new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 20), |
||
152 | new Token(Tokenizer::TOKEN_WHITESPACE, ' ', 22), |
||
153 | new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 23), |
||
154 | new Token(Tokenizer::TOKEN_WHITESPACE, ' ', 26), |
||
155 | new WordToken("\\ word'", 27, '', " word'"), |
||
156 | ], |
||
157 | ], |
||
158 | [ |
||
159 | '"phrase \+ \- \! \( \) \AND \OR \NOT \\\\ phrase"', |
||
160 | [ |
||
161 | new PhraseToken( |
||
162 | '"phrase \+ \- \! \( \) \AND \OR \NOT \\\\ phrase"', |
||
163 | 0, |
||
164 | '', |
||
165 | '"', |
||
166 | 'phrase \+ \- \! \( \) \AND \OR \NOT \\\\ phrase' |
||
167 | ), |
||
168 | ], |
||
169 | ], |
||
170 | [ |
||
171 | "'word \\+ \\- \\! \\( \\) \\AND \\OR \\NOT \\\\ word'", |
||
172 | [ |
||
173 | new WordToken("'word", 0, '', "'word"), |
||
174 | new Token(Tokenizer::TOKEN_WHITESPACE, ' ', 5), |
||
175 | new WordToken('\\+', 6, '', '+'), |
||
176 | new Token(Tokenizer::TOKEN_WHITESPACE, ' ', 8), |
||
177 | new WordToken('\\-', 9, '', '-'), |
||
178 | new Token(Tokenizer::TOKEN_WHITESPACE, ' ', 11), |
||
179 | new WordToken('\\!', 12, '', '!'), |
||
180 | new Token(Tokenizer::TOKEN_WHITESPACE, ' ', 14), |
||
181 | new WordToken('\\(', 15, '', '('), |
||
182 | new Token(Tokenizer::TOKEN_WHITESPACE, ' ', 17), |
||
183 | new WordToken('\\)', 18, '', ')'), |
||
184 | new Token(Tokenizer::TOKEN_WHITESPACE, ' ', 20), |
||
185 | new WordToken('\\AND', 21, '', '\AND'), |
||
186 | new Token(Tokenizer::TOKEN_WHITESPACE, ' ', 25), |
||
187 | new WordToken('\\OR', 26, '', '\OR'), |
||
188 | new Token(Tokenizer::TOKEN_WHITESPACE, ' ', 29), |
||
189 | new WordToken('\\NOT', 30, '', '\NOT'), |
||
190 | new Token(Tokenizer::TOKEN_WHITESPACE, ' ', 34), |
||
191 | new WordToken('\\\\', 35, '', '\\'), |
||
192 | new Token(Tokenizer::TOKEN_WHITESPACE, ' ', 37), |
||
193 | new WordToken("word'", 38, '', "word'"), |
||
194 | ], |
||
195 | ], |
||
196 | ['#tag', [new TagToken('#tag', 0, '#', 'tag')]], |
||
197 | ['\#tag', [new WordToken('\#tag', 0, '', '#tag')]], |
||
198 | ['#tagšđčćž', [new WordToken('#tagšđčćž', 0, '', '#tagšđčćž')]], |
||
199 | ['#_tag-tag', [new TagToken('#_tag-tag', 0, '#', '_tag-tag')]], |
||
200 | ['#-not-tag', [new WordToken('#-not-tag', 0, '', '#-not-tag')]], |
||
201 | ['#tag+', [new TagToken('#tag', 0, '#', 'tag'), new Token(Tokenizer::TOKEN_MANDATORY, '+', 4)]], |
||
202 | ['#tag-', [new TagToken('#tag-', 0, '#', 'tag-')]], |
||
203 | ['#tag!', [new TagToken('#tag', 0, '#', 'tag'), new Token(Tokenizer::TOKEN_LOGICAL_NOT_2, '!', 4)]], |
||
204 | ["#tag\n", [new TagToken('#tag', 0, '#', 'tag'), new Token(Tokenizer::TOKEN_WHITESPACE, "\n", 4)]], |
||
205 | ['#tag ', [new TagToken('#tag', 0, '#', 'tag'), new Token(Tokenizer::TOKEN_WHITESPACE, ' ', 4)]], |
||
206 | ['#tag(', [new TagToken('#tag', 0, '#', 'tag'), new GroupBeginToken('(', 4, '(', null)]], |
||
207 | ['#tag)', [new TagToken('#tag', 0, '#', 'tag'), new Token(Tokenizer::TOKEN_GROUP_END, ')', 4)]], |
||
208 | ['@user', [new UserToken('@user', 0, '@', 'user')]], |
||
209 | ['@user.user', [new UserToken('@user.user', 0, '@', 'user.user')]], |
||
210 | ['\@user', [new WordToken('\@user', 0, '', '@user')]], |
||
211 | ['@useršđčćž', [new WordToken('@useršđčćž', 0, '', '@useršđčćž')]], |
||
212 | ['@_user-user', [new UserToken('@_user-user', 0, '@', '_user-user')]], |
||
213 | ['@-not-user', [new WordToken('@-not-user', 0, '', '@-not-user')]], |
||
214 | ['@user+', [new UserToken('@user', 0, '@', 'user'), new Token(Tokenizer::TOKEN_MANDATORY, '+', 5)]], |
||
215 | ['@user-', [new UserToken('@user-', 0, '@', 'user-')]], |
||
216 | ['@user!', [new UserToken('@user', 0, '@', 'user'), new Token(Tokenizer::TOKEN_LOGICAL_NOT_2, '!', 5)]], |
||
217 | ["@user\n", [new UserToken('@user', 0, '@', 'user'), new Token(Tokenizer::TOKEN_WHITESPACE, "\n", 5)]], |
||
218 | ['@user ', [new UserToken('@user', 0, '@', 'user'), new Token(Tokenizer::TOKEN_WHITESPACE, ' ', 5)]], |
||
219 | ['@user(', [new UserToken('@user', 0, '@', 'user'), new GroupBeginToken('(', 5, '(', null)]], |
||
220 | ['@user)', [new UserToken('@user', 0, '@', 'user'), new Token(Tokenizer::TOKEN_GROUP_END, ')', 5)]], |
||
221 | ['domain:', [new WordToken('domain:', 0, '', 'domain:')]], |
||
222 | ['some.domain:', [new WordToken('some.domain:', 0, '', 'some.domain:')]], |
||
223 | ['domain:domain:', [new WordToken('domain:domain:', 0, 'domain', 'domain:')]], |
||
224 | ['some.domain:some.domain:', [new WordToken('some.domain:some.domain:', 0, 'some.domain', 'some.domain:')]], |
||
225 | [ |
||
226 | 'domain:domain:domain:domain', |
||
227 | [new WordToken('domain:domain:domain:domain', 0, 'domain', 'domain:domain:domain')], |
||
228 | ], |
||
229 | ['domain\:', [new WordToken('domain\:', 0, '', 'domain:')]], |
||
230 | ['domain\::', [new WordToken('domain\::', 0, '', 'domain::')]], |
||
231 | ['domain:word', [new WordToken('domain:word', 0, 'domain', 'word')]], |
||
232 | ['domain\:word', [new WordToken('domain\:word', 0, '', 'domain:word')]], |
||
233 | ['domain:"phrase"', [new PhraseToken('domain:"phrase"', 0, 'domain', '"', 'phrase')]], |
||
234 | ['some.domain:"phrase"', [new PhraseToken('some.domain:"phrase"', 0, 'some.domain', '"', 'phrase')]], |
||
235 | [ |
||
236 | 'domain\:"phrase"', |
||
237 | [new WordToken('domain\:', 0, '', 'domain:'), new PhraseToken('"phrase"', 8, '', '"', 'phrase')], |
||
238 | ], |
||
239 | [ |
||
240 | 'domain:(one)', |
||
241 | [ |
||
242 | new GroupBeginToken('domain:(', 0, '(', 'domain'), |
||
243 | new WordToken('one', 8, '', 'one'), |
||
244 | new Token(Tokenizer::TOKEN_GROUP_END, ')', 11), |
||
245 | ], |
||
246 | ], |
||
247 | [ |
||
248 | 'some.domain:(one)', |
||
249 | [ |
||
250 | new GroupBeginToken('some.domain:(', 0, '(', 'some.domain'), |
||
251 | new WordToken('one', 13, '', 'one'), |
||
252 | new Token(Tokenizer::TOKEN_GROUP_END, ')', 16), |
||
253 | ], |
||
254 | ], |
||
255 | [ |
||
256 | 'one AND two', |
||
257 | [ |
||
258 | new WordToken('one', 0, '', 'one'), |
||
259 | new Token(Tokenizer::TOKEN_WHITESPACE, ' ', 3), |
||
260 | new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 4), |
||
261 | new Token(Tokenizer::TOKEN_WHITESPACE, ' ', 7), |
||
262 | new WordToken('two', 8, '', 'two'), |
||
263 | ], |
||
264 | ], |
||
265 | [ |
||
266 | 'one && two', |
||
267 | [ |
||
268 | new WordToken('one', 0, '', 'one'), |
||
269 | new Token(Tokenizer::TOKEN_WHITESPACE, ' ', 3), |
||
270 | new Token(Tokenizer::TOKEN_LOGICAL_AND, '&&', 4), |
||
271 | new Token(Tokenizer::TOKEN_WHITESPACE, ' ', 6), |
||
272 | new WordToken('two', 7, '', 'two'), |
||
273 | ], |
||
274 | ], |
||
275 | [ |
||
276 | 'one OR two', |
||
277 | [ |
||
278 | new WordToken('one', 0, '', 'one'), |
||
279 | new Token(Tokenizer::TOKEN_WHITESPACE, ' ', 3), |
||
280 | new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 4), |
||
281 | new Token(Tokenizer::TOKEN_WHITESPACE, ' ', 6), |
||
282 | new WordToken('two', 7, '', 'two'), |
||
283 | ], |
||
284 | ], |
||
285 | [ |
||
286 | 'one || two', |
||
287 | [ |
||
288 | new WordToken('one', 0, '', 'one'), |
||
289 | new Token(Tokenizer::TOKEN_WHITESPACE, ' ', 3), |
||
290 | new Token(Tokenizer::TOKEN_LOGICAL_OR, '||', 4), |
||
291 | new Token(Tokenizer::TOKEN_WHITESPACE, ' ', 6), |
||
292 | new WordToken('two', 7, '', 'two'), |
||
293 | ], |
||
294 | ], |
||
295 | [ |
||
296 | 'one NOT two', |
||
297 | [ |
||
298 | new WordToken('one', 0, '', 'one'), |
||
299 | new Token(Tokenizer::TOKEN_WHITESPACE, ' ', 3), |
||
300 | new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 4), |
||
301 | new Token(Tokenizer::TOKEN_WHITESPACE, ' ', 7), |
||
302 | new WordToken('two', 8, '', 'two'), |
||
303 | ], |
||
304 | ], |
||
305 | ['AND', [new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 0)]], |
||
306 | ['ANDword', [new WordToken('ANDword', 0, '', 'ANDword')]], |
||
307 | ['wordAND', [new WordToken('wordAND', 0, '', 'wordAND')]], |
||
308 | [ |
||
309 | 'AND+', |
||
310 | [new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 0), new Token(Tokenizer::TOKEN_MANDATORY, '+', 3)], |
||
311 | ], |
||
312 | ['AND\+', [new WordToken('AND\+', 0, '', 'AND+')]], |
||
313 | [ |
||
314 | '+AND', |
||
315 | [new Token(Tokenizer::TOKEN_MANDATORY, '+', 0), new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 1)], |
||
316 | ], |
||
317 | [ |
||
318 | 'AND-', |
||
319 | [new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 0), new Token(Tokenizer::TOKEN_PROHIBITED, '-', 3)], |
||
320 | ], |
||
321 | ['AND\-', [new WordToken('AND\-', 0, '', 'AND-')]], |
||
322 | [ |
||
323 | '-AND', |
||
324 | [new Token(Tokenizer::TOKEN_PROHIBITED, '-', 0), new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 1)], |
||
325 | ], |
||
326 | [ |
||
327 | 'AND!', |
||
328 | [new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 0), new Token(Tokenizer::TOKEN_LOGICAL_NOT_2, '!', 3)], |
||
329 | ], |
||
330 | ['AND\!', [new WordToken('AND\!', 0, '', 'AND!')]], |
||
331 | [ |
||
332 | '!AND', |
||
333 | [new Token(Tokenizer::TOKEN_LOGICAL_NOT_2, '!', 0), new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 1)], |
||
334 | ], |
||
335 | [ |
||
336 | "AND\n", |
||
337 | [new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 0), new Token(Tokenizer::TOKEN_WHITESPACE, "\n", 3)], |
||
338 | ], |
||
339 | [ |
||
340 | 'AND ', |
||
341 | [new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 0), new Token(Tokenizer::TOKEN_WHITESPACE, ' ', 3)], |
||
342 | ], |
||
343 | ['AND(', [new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 0), new GroupBeginToken('(', 3, '(', null)]], |
||
344 | [ |
||
345 | 'AND)', |
||
346 | [new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 0), new Token(Tokenizer::TOKEN_GROUP_END, ')', 3)], |
||
347 | ], |
||
348 | ['ORword', [new WordToken('ORword', 0, '', 'ORword')]], |
||
349 | ['ORword', [new WordToken('ORword', 0, '', 'ORword')]], |
||
350 | ['OR', [new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 0)]], |
||
351 | ['OR+', [new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 0), new Token(Tokenizer::TOKEN_MANDATORY, '+', 2)]], |
||
352 | ['OR\+', [new WordToken('OR\+', 0, '', 'OR+')]], |
||
353 | ['+OR', [new Token(Tokenizer::TOKEN_MANDATORY, '+', 0), new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 1)]], |
||
354 | ['OR-', [new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 0), new Token(Tokenizer::TOKEN_PROHIBITED, '-', 2)]], |
||
355 | ['OR\+', [new WordToken('OR\+', 0, '', 'OR+')]], |
||
356 | ['-OR', [new Token(Tokenizer::TOKEN_PROHIBITED, '-', 0), new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 1)]], |
||
357 | [ |
||
358 | 'OR!', |
||
359 | [new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 0), new Token(Tokenizer::TOKEN_LOGICAL_NOT_2, '!', 2)], |
||
360 | ], |
||
361 | ['OR\!', [new WordToken('OR\!', 0, '', 'OR!')]], |
||
362 | [ |
||
363 | '!OR', |
||
364 | [new Token(Tokenizer::TOKEN_LOGICAL_NOT_2, '!', 0), new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 1)], |
||
365 | ], |
||
366 | [ |
||
367 | "OR\n", |
||
368 | [new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 0), new Token(Tokenizer::TOKEN_WHITESPACE, "\n", 2)], |
||
369 | ], |
||
370 | ['OR ', [new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 0), new Token(Tokenizer::TOKEN_WHITESPACE, ' ', 2)]], |
||
371 | ['OR(', [new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 0), new GroupBeginToken('(', 2, '(', null)]], |
||
372 | ['OR)', [new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 0), new Token(Tokenizer::TOKEN_GROUP_END, ')', 2)]], |
||
373 | ['NOT', [new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 0)]], |
||
374 | ['NOTword', [new WordToken('NOTword', 0, '', 'NOTword')]], |
||
375 | ['wordNOT', [new WordToken('wordNOT', 0, '', 'wordNOT')]], |
||
376 | [ |
||
377 | 'NOT+', |
||
378 | [new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 0), new Token(Tokenizer::TOKEN_MANDATORY, '+', 3)], |
||
379 | ], |
||
380 | [ |
||
381 | '+NOT', |
||
382 | [new Token(Tokenizer::TOKEN_MANDATORY, '+', 0), new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 1)], |
||
383 | ], |
||
384 | [ |
||
385 | 'NOT-', |
||
386 | [new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 0), new Token(Tokenizer::TOKEN_PROHIBITED, '-', 3)], |
||
387 | ], |
||
388 | [ |
||
389 | '-NOT', |
||
390 | [new Token(Tokenizer::TOKEN_PROHIBITED, '-', 0), new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 1)], |
||
391 | ], |
||
392 | [ |
||
393 | 'NOT!', |
||
394 | [new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 0), new Token(Tokenizer::TOKEN_LOGICAL_NOT_2, '!', 3)], |
||
395 | ], |
||
396 | [ |
||
397 | '!NOT', |
||
398 | [new Token(Tokenizer::TOKEN_LOGICAL_NOT_2, '!', 0), new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 1)], |
||
399 | ], |
||
400 | [ |
||
401 | "NOT\n", |
||
402 | [new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 0), new Token(Tokenizer::TOKEN_WHITESPACE, "\n", 3)], |
||
403 | ], |
||
404 | [ |
||
405 | 'NOT ', |
||
406 | [new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 0), new Token(Tokenizer::TOKEN_WHITESPACE, ' ', 3)], |
||
407 | ], |
||
408 | ['NOT(', [new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 0), new GroupBeginToken('(', 3, '(', null)]], |
||
409 | [ |
||
410 | 'NOT)', |
||
411 | [new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 0), new Token(Tokenizer::TOKEN_GROUP_END, ')', 3)], |
||
412 | ], |
||
413 | ['+', [new Token(Tokenizer::TOKEN_MANDATORY, '+', 0)]], |
||
414 | ['++', [new Token(Tokenizer::TOKEN_MANDATORY, '+', 0), new Token(Tokenizer::TOKEN_MANDATORY, '+', 1)]], |
||
415 | ['-', [new Token(Tokenizer::TOKEN_PROHIBITED, '-', 0)]], |
||
416 | ['--', [new Token(Tokenizer::TOKEN_PROHIBITED, '-', 0), new Token(Tokenizer::TOKEN_PROHIBITED, '-', 1)]], |
||
417 | ['!', [new Token(Tokenizer::TOKEN_LOGICAL_NOT_2, '!', 0)]], |
||
418 | [ |
||
419 | '!!', |
||
420 | [new Token(Tokenizer::TOKEN_LOGICAL_NOT_2, '!', 0), new Token(Tokenizer::TOKEN_LOGICAL_NOT_2, '!', 1)], |
||
421 | ], |
||
422 | ['+word', [new Token(Tokenizer::TOKEN_MANDATORY, '+', 0), new WordToken('word', 1, '', 'word')]], |
||
423 | ['-word', [new Token(Tokenizer::TOKEN_PROHIBITED, '-', 0), new WordToken('word', 1, '', 'word')]], |
||
424 | ['!word', [new Token(Tokenizer::TOKEN_LOGICAL_NOT_2, '!', 0), new WordToken('word', 1, '', 'word')]], |
||
425 | ['(word', [new GroupBeginToken('(', 0, '(', null), new WordToken('word', 1, '', 'word')]], |
||
426 | [')word', [new Token(Tokenizer::TOKEN_GROUP_END, ')', 0), new WordToken('word', 1, '', 'word')]], |
||
427 | ['word+', [new WordToken('word+', 0, '', 'word+')]], |
||
428 | ['word-', [new WordToken('word-', 0, '', 'word-')]], |
||
429 | ['word!', [new WordToken('word!', 0, '', 'word!')]], |
||
430 | ['word(', [new WordToken('word', 0, '', 'word'), new GroupBeginToken('(', 4, '(', null)]], |
||
431 | ['word)', [new WordToken('word', 0, '', 'word'), new Token(Tokenizer::TOKEN_GROUP_END, ')', 4)]], |
||
432 | ['one+two+', [new WordToken('one+two+', 0, '', 'one+two+')]], |
||
433 | ['one-two-', [new WordToken('one-two-', 0, '', 'one-two-')]], |
||
434 | ['one!two!', [new WordToken('one!two!', 0, '', 'one!two!')]], |
||
435 | [ |
||
436 | 'one(two(', |
||
437 | [ |
||
438 | new WordToken('one', 0, '', 'one'), |
||
439 | new GroupBeginToken('(', 3, '(', null), |
||
440 | new WordToken('two', 4, '', 'two'), |
||
441 | new GroupBeginToken('(', 7, '(', null), |
||
442 | ], |
||
443 | ], |
||
444 | [ |
||
445 | 'one)two)', |
||
446 | [ |
||
447 | new WordToken('one', 0, '', 'one'), |
||
448 | new Token(Tokenizer::TOKEN_GROUP_END, ')', 3), |
||
449 | new WordToken('two', 4, '', 'two'), |
||
450 | new Token(Tokenizer::TOKEN_GROUP_END, ')', 7), |
||
451 | ], |
||
452 | ], |
||
453 | ['word\+', [new WordToken('word\+', 0, '', 'word+')]], |
||
454 | ['word\-', [new WordToken('word\-', 0, '', 'word-')]], |
||
455 | ['word\!', [new WordToken('word\!', 0, '', 'word!')]], |
||
456 | ['word\(', [new WordToken('word\(', 0, '', 'word(')]], |
||
457 | ['word\)', [new WordToken('word\)', 0, '', 'word)')]], |
||
458 | ['\+word', [new WordToken('\+word', 0, '', '+word')]], |
||
459 | ['\-word', [new WordToken('\-word', 0, '', '-word')]], |
||
460 | ['\!word', [new WordToken('\!word', 0, '', '!word')]], |
||
461 | ['\(word', [new WordToken('\(word', 0, '', '(word')]], |
||
462 | ['\)word', [new WordToken('\)word', 0, '', ')word')]], |
||
463 | ['one\+two\+', [new WordToken('one\+two\+', 0, '', 'one+two+')]], |
||
464 | ['one\-two\-', [new WordToken('one\-two\-', 0, '', 'one-two-')]], |
||
465 | ['one\!two\!', [new WordToken('one\!two\!', 0, '', 'one!two!')]], |
||
466 | ['one\(two\(', [new WordToken('one\(two\(', 0, '', 'one(two(')]], |
||
467 | ['one\)two\)', [new WordToken('one\)two\)', 0, '', 'one)two)')]], |
||
468 | [ |
||
469 | 'one\\\\\)two\\\\\(one\\\\\+two\\\\\-one\\\\\!two', |
||
470 | [ |
||
471 | new WordToken( |
||
472 | 'one\\\\\)two\\\\\(one\\\\\+two\\\\\-one\\\\\!two', |
||
473 | 0, |
||
474 | '', |
||
475 | 'one\)two\(one\+two\-one\!two' |
||
476 | ), |
||
477 | ], |
||
478 | ], |
||
479 | [ |
||
480 | 'one\\\\)two\\\\(one\\\\+two\\\\-one\\\\!two', |
||
481 | [ |
||
482 | new WordToken('one\\\\', 0, '', 'one\\'), |
||
483 | new Token(Tokenizer::TOKEN_GROUP_END, ')', 5), |
||
484 | new WordToken('two\\\\', 6, '', 'two\\'), |
||
485 | new GroupBeginToken('(', 11, '(', null), |
||
486 | new WordToken('one\\\\+two\\\\-one\\\\!two', 12, '', 'one\+two\-one\!two'), |
||
487 | ], |
||
488 | ], |
||
489 | ['one+two-one!two', [new WordToken('one+two-one!two', 0, '', 'one+two-one!two')]], |
||
490 | ['one\\\'two', [new WordToken('one\\\'two', 0, '', "one\\'two")]], |
||
491 | ['one\\"two', [new WordToken('one\\"two', 0, '', 'one"two')]], |
||
492 | ['\\', [new WordToken('\\', 0, '', '\\')]], |
||
493 | ['one\\two', [new WordToken('one\\two', 0, '', 'one\\two')]], |
||
494 | ['one\\\\+\\-\\!\\(\\)two', [new WordToken('one\\\\+\\-\\!\\(\\)two', 0, '', 'one\\+-!()two')]], |
||
495 | ['\\\\', [new WordToken('\\\\', 0, '', '\\')]], |
||
496 | [ |
||
497 | '(type:)', |
||
498 | [ |
||
499 | new GroupBeginToken('(', 0, '(', null), |
||
500 | new WordToken('type:', 1, '', 'type:'), |
||
501 | new Token(Tokenizer::TOKEN_GROUP_END, ')', 6), |
||
502 | ], |
||
503 | ], |
||
504 | [ |
||
505 | 'type: AND', |
||
506 | [ |
||
507 | new WordToken('type:', 0, '', 'type:'), |
||
508 | new Token(Tokenizer::TOKEN_WHITESPACE, ' ', 5), |
||
509 | new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 6), |
||
510 | ], |
||
511 | ], |
||
512 | ["word'", [new WordToken("word'", 0, '', "word'")]], |
||
513 | ['one\'two', [new WordToken("one'two", 0, '', "one'two")]], |
||
514 | ["AND'", [new WordToken("AND'", 0, '', "AND'")]], |
||
515 | ["OR'", [new WordToken("OR'", 0, '', "OR'")]], |
||
516 | ["NOT'", [new WordToken("NOT'", 0, '', "NOT'")]], |
||
517 | ]; |
||
599 |