1
|
|
|
<?php declare(strict_types = 1); |
2
|
|
|
|
3
|
|
|
namespace Apicart\FQL\Tests; |
4
|
|
|
|
5
|
|
|
use Apicart\FQL\Generator\Common\Aggregate; |
6
|
|
|
use Apicart\FQL\Generator\Native; |
7
|
|
|
use Apicart\FQL\Generator\Native\BinaryOperator; |
8
|
|
|
use Apicart\FQL\Generator\Native\Group as GroupNative; |
9
|
|
|
use Apicart\FQL\Generator\Native\Phrase; |
10
|
|
|
use Apicart\FQL\Generator\Native\Query as QueryNative; |
11
|
|
|
use Apicart\FQL\Generator\Native\Tag; |
12
|
|
|
use Apicart\FQL\Generator\Native\UnaryOperator; |
13
|
|
|
use Apicart\FQL\Generator\Native\User; |
14
|
|
|
use Apicart\FQL\Generator\Native\Word; |
15
|
|
|
use Apicart\FQL\Token\Node\Group; |
16
|
|
|
use Apicart\FQL\Token\Node\LogicalAnd; |
17
|
|
|
use Apicart\FQL\Token\Node\LogicalNot; |
18
|
|
|
use Apicart\FQL\Token\Node\LogicalOr; |
19
|
|
|
use Apicart\FQL\Token\Node\Mandatory; |
20
|
|
|
use Apicart\FQL\Token\Node\Prohibited; |
21
|
|
|
use Apicart\FQL\Token\Node\Query; |
22
|
|
|
use Apicart\FQL\Token\Node\Term; |
23
|
|
|
use Apicart\FQL\Token\Token\GroupBegin as GroupBeginToken; |
24
|
|
|
use Apicart\FQL\Token\Token\Phrase as PhraseToken; |
25
|
|
|
use Apicart\FQL\Token\Token\Tag as TagToken; |
26
|
|
|
use Apicart\FQL\Token\Token\User as UserToken; |
27
|
|
|
use Apicart\FQL\Token\Token\Word as WordToken; |
28
|
|
|
use Apicart\FQL\Tokenizer\Full; |
29
|
|
|
use Apicart\FQL\Tokenizer\Parser; |
30
|
|
|
use Apicart\FQL\Tokenizer\Tokenizer; |
31
|
|
|
use Apicart\FQL\Value\Correction; |
32
|
|
|
use Apicart\FQL\Value\SyntaxTree; |
33
|
|
|
use Apicart\FQL\Value\Token; |
34
|
|
|
use Apicart\FQL\Value\TokenSequence; |
35
|
|
|
use PHPUnit\Framework\TestCase; |
36
|
|
|
|
37
|
|
|
final class IntegrationTest extends TestCase |
38
|
|
|
{ |
39
|
|
|
|
40
|
|
|
public function providerForTestQuery(): array |
41
|
|
|
{ |
42
|
|
|
return [ |
43
|
|
|
['', [], new Query([])], |
44
|
|
|
['one', [$token = new WordToken('one', 0, '', 'one')], new Query([new Term($token)])], |
45
|
|
|
[ |
46
|
|
|
'one two', |
47
|
|
|
[$token1 = new WordToken('one', 0, '', 'one'), $token2 = new WordToken('two', 4, '', 'two')], |
48
|
|
|
new Query([new Term($token1), new Term($token2)]), |
49
|
|
|
], |
50
|
|
|
[ |
51
|
|
|
'one AND two', |
52
|
|
|
[ |
53
|
|
|
$token1 = new WordToken('one', 0, '', 'one'), |
54
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 4), |
55
|
|
|
$token3 = new WordToken('two', 8, '', 'two'), |
56
|
|
|
], |
57
|
|
|
new Query([new LogicalAnd(new Term($token1), new Term($token3), $token2)]), |
58
|
|
|
], |
59
|
|
|
[ |
60
|
|
|
'one OR two', |
61
|
|
|
[ |
62
|
|
|
$token1 = new WordToken('one', 0, '', 'one'), |
63
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 4), |
64
|
|
|
$token3 = new WordToken('two', 7, '', 'two'), |
65
|
|
|
], |
66
|
|
|
new Query([new LogicalOr(new Term($token1), new Term($token3), $token2)]), |
67
|
|
|
], |
68
|
|
|
[ |
69
|
|
|
'one OR two AND three', |
70
|
|
|
[ |
71
|
|
|
$token1 = new WordToken('one', 0, '', 'one'), |
72
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 4), |
73
|
|
|
$token3 = new WordToken('two', 7, '', 'two'), |
74
|
|
|
$token4 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 11), |
75
|
|
|
$token5 = new WordToken('three', 15, '', 'three'), |
76
|
|
|
], |
77
|
|
|
new Query( |
78
|
|
|
[ |
79
|
|
|
new LogicalOr(new Term($token1), new LogicalAnd(new Term($token3), new Term( |
80
|
|
|
$token5 |
81
|
|
|
), $token4), $token2), |
82
|
|
|
] |
83
|
|
|
), |
84
|
|
|
], |
85
|
|
|
[ |
86
|
|
|
'one AND two OR three', |
87
|
|
|
[ |
88
|
|
|
$token1 = new WordToken('one', 0, '', 'one'), |
89
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 4), |
90
|
|
|
$token3 = new WordToken('two', 8, '', 'two'), |
91
|
|
|
$token4 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 12), |
92
|
|
|
$token5 = new WordToken('three', 15, '', 'three'), |
93
|
|
|
], |
94
|
|
|
new Query( |
95
|
|
|
[ |
96
|
|
|
new LogicalOr(new LogicalAnd(new Term($token1), new Term($token3), $token2), new Term( |
97
|
|
|
$token5 |
98
|
|
|
), $token4), |
99
|
|
|
] |
100
|
|
|
), |
101
|
|
|
], |
102
|
|
|
[ |
103
|
|
|
'NOT one', |
104
|
|
|
[ |
105
|
|
|
$token1 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 0), |
106
|
|
|
$token2 = new WordToken('one', 4, '', 'one'), |
107
|
|
|
], |
108
|
|
|
new Query([new LogicalNot(new Term($token2), $token1)]), |
109
|
|
|
], |
110
|
|
|
[ |
111
|
|
|
'one NOT two', |
112
|
|
|
[ |
113
|
|
|
$token1 = new WordToken('one', 0, '', 'one'), |
114
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 4), |
115
|
|
|
$token3 = new WordToken('two', 8, '', 'two'), |
116
|
|
|
], |
117
|
|
|
new Query([new Term($token1), new LogicalNot(new Term($token3), $token2)]), |
118
|
|
|
], |
119
|
|
|
[ |
120
|
|
|
'one AND NOT two', |
121
|
|
|
[ |
122
|
|
|
$token1 = new WordToken('one', 0, '', 'one'), |
123
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 4), |
124
|
|
|
$token3 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 8), |
125
|
|
|
$token4 = new WordToken('two', 12, '', 'two'), |
126
|
|
|
], |
127
|
|
|
new Query([new LogicalAnd(new Term($token1), new LogicalNot(new Term($token4), $token3), $token2)]), |
128
|
|
|
], |
129
|
|
|
[ |
130
|
|
|
'one OR NOT two', |
131
|
|
|
[ |
132
|
|
|
$token1 = new WordToken('one', 0, '', 'one'), |
133
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 4), |
134
|
|
|
$token3 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 7), |
135
|
|
|
$token4 = new WordToken('two', 11, '', 'two'), |
136
|
|
|
], |
137
|
|
|
new Query([new LogicalOr(new Term($token1), new LogicalNot(new Term($token4), $token3), $token2)]), |
138
|
|
|
], |
139
|
|
|
[ |
140
|
|
|
'!one', |
141
|
|
|
[ |
142
|
|
|
$token1 = new Token(Tokenizer::TOKEN_LOGICAL_NOT_2, '!', 0), |
143
|
|
|
$token2 = new WordToken('one', 1, '', 'one'), |
144
|
|
|
], |
145
|
|
|
new Query([new LogicalNot(new Term($token2), $token1)]), |
146
|
|
|
], |
147
|
|
|
[ |
148
|
|
|
'one !two', |
149
|
|
|
[ |
150
|
|
|
$token1 = new WordToken('one', 0, '', 'one'), |
151
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_NOT_2, '!', 4), |
152
|
|
|
$token3 = new WordToken('two', 5, '', 'two'), |
153
|
|
|
], |
154
|
|
|
new Query([new Term($token1), new LogicalNot(new Term($token3), $token2)]), |
155
|
|
|
], |
156
|
|
|
[ |
157
|
|
|
'one AND !two', |
158
|
|
|
[ |
159
|
|
|
$token1 = new WordToken('one', 0, '', 'one'), |
160
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 4), |
161
|
|
|
$token3 = new Token(Tokenizer::TOKEN_LOGICAL_NOT_2, '!', 8), |
162
|
|
|
$token4 = new WordToken('two', 9, '', 'two'), |
163
|
|
|
], |
164
|
|
|
new Query([new LogicalAnd(new Term($token1), new LogicalNot(new Term($token4), $token3), $token2)]), |
165
|
|
|
], |
166
|
|
|
[ |
167
|
|
|
'one OR !two', |
168
|
|
|
[ |
169
|
|
|
$token1 = new WordToken('one', 0, '', 'one'), |
170
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 4), |
171
|
|
|
$token3 = new Token(Tokenizer::TOKEN_LOGICAL_NOT_2, '!', 7), |
172
|
|
|
$token4 = new WordToken('two', 8, '', 'two'), |
173
|
|
|
], |
174
|
|
|
new Query([new LogicalOr(new Term($token1), new LogicalNot(new Term($token4), $token3), $token2)]), |
175
|
|
|
], |
176
|
|
|
[ |
177
|
|
|
'(one two)', |
178
|
|
|
[ |
179
|
|
|
$token1 = new GroupBeginToken('(', 0, '(', null), |
180
|
|
|
$token2 = new WordToken('one', 1, '', 'one'), |
181
|
|
|
$token3 = new WordToken('two', 5, '', 'two'), |
182
|
|
|
$token4 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 8), |
183
|
|
|
], |
184
|
|
|
new Query([new Group([new Term($token2), new Term($token3)], $token1, $token4)]), |
185
|
|
|
], |
186
|
|
|
[ |
187
|
|
|
'(one AND two)', |
188
|
|
|
[ |
189
|
|
|
$token1 = new GroupBeginToken('(', 0, '(', null), |
190
|
|
|
$token2 = new WordToken('one', 1, '', 'one'), |
191
|
|
|
$token3 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 5), |
192
|
|
|
$token4 = new WordToken('two', 9, '', 'two'), |
193
|
|
|
$token5 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 12), |
194
|
|
|
], |
195
|
|
|
new Query([ |
196
|
|
|
new Group([new LogicalAnd(new Term($token2), new Term($token4), $token3)], $token1, $token5), |
197
|
|
|
]), |
198
|
|
|
], |
199
|
|
|
[ |
200
|
|
|
'(NOT one OR two)', |
201
|
|
|
[ |
202
|
|
|
$token1 = new GroupBeginToken('(', 0, '(', null), |
203
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 1), |
204
|
|
|
$token3 = new WordToken('one', 5, '', 'one'), |
205
|
|
|
$token4 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 9), |
206
|
|
|
$token5 = new WordToken('two', 12, '', 'two'), |
207
|
|
|
$token6 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 15), |
208
|
|
|
], |
209
|
|
|
new Query( |
210
|
|
|
[ |
211
|
|
|
new Group( |
212
|
|
|
[new LogicalOr(new LogicalNot(new Term($token3), $token2), new Term($token5), $token4)], |
213
|
|
|
$token1, |
214
|
|
|
$token6 |
215
|
|
|
), |
216
|
|
|
] |
217
|
|
|
), |
218
|
|
|
], |
219
|
|
|
[ |
220
|
|
|
'one AND (two OR three)', |
221
|
|
|
[ |
222
|
|
|
$token1 = new WordToken('one', 0, '', 'one'), |
223
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 4), |
224
|
|
|
$token3 = new GroupBeginToken('(', 8, '(', null), |
225
|
|
|
$token4 = new WordToken('two', 9, '', 'two'), |
226
|
|
|
$token5 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 13), |
227
|
|
|
$token6 = new WordToken('three', 16, '', 'three'), |
228
|
|
|
$token7 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 21), |
229
|
|
|
], |
230
|
|
|
new Query( |
231
|
|
|
[ |
232
|
|
|
new LogicalAnd( |
233
|
|
|
new Term($token1), |
234
|
|
|
new Group([new LogicalOr(new Term($token4), new Term($token6), $token5)], $token3, $token7), |
235
|
|
|
$token2 |
236
|
|
|
), |
237
|
|
|
] |
238
|
|
|
), |
239
|
|
|
], |
240
|
|
|
[ |
241
|
|
|
'((one) AND (two AND (three OR four five)))', |
242
|
|
|
[ |
243
|
|
|
$token1 = new GroupBeginToken('(', 0, '(', null), |
244
|
|
|
$token2 = new GroupBeginToken('(', 1, '(', null), |
245
|
|
|
$token3 = new WordToken('one', 2, '', 'one'), |
246
|
|
|
$token4 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 5), |
247
|
|
|
$token5 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 7), |
248
|
|
|
$token6 = new GroupBeginToken('(', 11, '(', null), |
249
|
|
|
$token7 = new WordToken('two', 12, '', 'two'), |
250
|
|
|
$token8 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 16), |
251
|
|
|
$token9 = new GroupBeginToken('(', 20, '(', null), |
252
|
|
|
$token10 = new WordToken('three', 21, '', 'three'), |
253
|
|
|
$token11 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 27), |
254
|
|
|
$token12 = new WordToken('four', 30, '', 'four'), |
255
|
|
|
$token13 = new WordToken('five', 35, '', 'five'), |
256
|
|
|
$token14 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 39), |
257
|
|
|
$token15 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 40), |
258
|
|
|
$token16 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 41), |
259
|
|
|
], |
260
|
|
|
new Query( |
261
|
|
|
[ |
262
|
|
|
new Group( |
263
|
|
|
[ |
264
|
|
|
new LogicalAnd( |
265
|
|
|
new Group([new Term($token3)], $token2, $token4), |
266
|
|
|
new Group( |
267
|
|
|
[ |
268
|
|
|
new LogicalAnd( |
269
|
|
|
new Term($token7), |
270
|
|
|
new Group( |
271
|
|
|
[ |
272
|
|
|
new LogicalOr(new Term($token10), new Term($token12), $token11), |
273
|
|
|
new Term($token13), |
274
|
|
|
], |
275
|
|
|
$token9, |
276
|
|
|
$token14 |
277
|
|
|
), |
278
|
|
|
$token8 |
279
|
|
|
), |
280
|
|
|
], |
281
|
|
|
$token6, |
282
|
|
|
$token15 |
283
|
|
|
), |
284
|
|
|
$token5 |
285
|
|
|
), |
286
|
|
|
], |
287
|
|
|
$token1, |
288
|
|
|
$token16 |
289
|
|
|
), |
290
|
|
|
] |
291
|
|
|
), |
292
|
|
|
], |
293
|
|
|
[ |
294
|
|
|
'((one) (two OR three))', |
295
|
|
|
[ |
296
|
|
|
$token1 = new GroupBeginToken('(', 0, '(', null), |
297
|
|
|
$token2 = new GroupBeginToken('(', 1, '(', null), |
298
|
|
|
$token3 = new WordToken('one', 2, '', 'one'), |
299
|
|
|
$token4 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 5), |
300
|
|
|
$token5 = new GroupBeginToken('(', 7, '(', null), |
301
|
|
|
$token6 = new WordToken('two', 8, '', 'two'), |
302
|
|
|
$token7 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 12), |
303
|
|
|
$token8 = new WordToken('three', 15, '', 'three'), |
304
|
|
|
$token9 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 20), |
305
|
|
|
$token10 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 21), |
306
|
|
|
], |
307
|
|
|
new Query( |
308
|
|
|
[ |
309
|
|
|
new Group( |
310
|
|
|
[ |
311
|
|
|
new Group([new Term($token3)], $token2, $token4), |
312
|
|
|
new Group([ |
313
|
|
|
new LogicalOr(new Term($token6), new Term($token8), $token7), |
314
|
|
|
], $token5, $token9), |
315
|
|
|
], |
316
|
|
|
$token1, |
317
|
|
|
$token10 |
318
|
|
|
), |
319
|
|
|
] |
320
|
|
|
), |
321
|
|
|
], |
322
|
|
|
[ |
323
|
|
|
'+one', |
324
|
|
|
[$token1 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 0), $token2 = new WordToken('one', 1, '', 'one')], |
325
|
|
|
new Query([new Mandatory(new Term($token2), $token1)]), |
326
|
|
|
], |
327
|
|
|
[ |
328
|
|
|
'+one AND +two', |
329
|
|
|
[ |
330
|
|
|
$token1 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 0), |
331
|
|
|
$token2 = new WordToken('one', 1, '', 'one'), |
332
|
|
|
$token3 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 5), |
333
|
|
|
$token4 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 9), |
334
|
|
|
$token5 = new WordToken('two', 10, '', 'two'), |
335
|
|
|
], |
336
|
|
|
new Query( |
337
|
|
|
[ |
338
|
|
|
new LogicalAnd(new Mandatory(new Term($token2), $token1), new Mandatory(new Term( |
339
|
|
|
$token5 |
340
|
|
|
), $token4), $token3), |
341
|
|
|
] |
342
|
|
|
), |
343
|
|
|
], |
344
|
|
|
[ |
345
|
|
|
'+one OR +two', |
346
|
|
|
[ |
347
|
|
|
$token1 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 0), |
348
|
|
|
$token2 = new WordToken('one', 1, '', 'one'), |
349
|
|
|
$token3 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 5), |
350
|
|
|
$token4 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 8), |
351
|
|
|
$token5 = new WordToken('two', 9, '', 'two'), |
352
|
|
|
], |
353
|
|
|
new Query( |
354
|
|
|
[ |
355
|
|
|
new LogicalOr(new Mandatory(new Term($token2), $token1), new Mandatory(new Term( |
356
|
|
|
$token5 |
357
|
|
|
), $token4), $token3), |
358
|
|
|
] |
359
|
|
|
), |
360
|
|
|
], |
361
|
|
|
[ |
362
|
|
|
'+one OR +two AND +three', |
363
|
|
|
[ |
364
|
|
|
$token1 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 0), |
365
|
|
|
$token2 = new WordToken('one', 1, '', 'one'), |
366
|
|
|
$token3 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 5), |
367
|
|
|
$token4 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 8), |
368
|
|
|
$token5 = new WordToken('two', 9, '', 'two'), |
369
|
|
|
$token6 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 13), |
370
|
|
|
$token7 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 17), |
371
|
|
|
$token8 = new WordToken('three', 18, '', 'three'), |
372
|
|
|
], |
373
|
|
|
new Query( |
374
|
|
|
[ |
375
|
|
|
new LogicalOr( |
376
|
|
|
new Mandatory(new Term($token2), $token1), |
377
|
|
|
new LogicalAnd( |
378
|
|
|
new Mandatory(new Term($token5), $token4), |
379
|
|
|
new Mandatory(new Term($token8), $token7), |
380
|
|
|
$token6 |
381
|
|
|
), |
382
|
|
|
$token3 |
383
|
|
|
), |
384
|
|
|
] |
385
|
|
|
), |
386
|
|
|
], |
387
|
|
|
[ |
388
|
|
|
'+one AND +two OR +three', |
389
|
|
|
[ |
390
|
|
|
$token1 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 0), |
391
|
|
|
$token2 = new WordToken('one', 1, '', 'one'), |
392
|
|
|
$token3 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 5), |
393
|
|
|
$token4 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 9), |
394
|
|
|
$token5 = new WordToken('two', 10, '', 'two'), |
395
|
|
|
$token6 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 14), |
396
|
|
|
$token7 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 17), |
397
|
|
|
$token8 = new WordToken('three', 18, '', 'three'), |
398
|
|
|
], |
399
|
|
|
new Query( |
400
|
|
|
[ |
401
|
|
|
new LogicalOr( |
402
|
|
|
new LogicalAnd( |
403
|
|
|
new Mandatory(new Term($token2), $token1), |
404
|
|
|
new Mandatory(new Term($token5), $token4), |
405
|
|
|
$token3 |
406
|
|
|
), |
407
|
|
|
new Mandatory(new Term($token8), $token7), |
408
|
|
|
$token6 |
409
|
|
|
), |
410
|
|
|
] |
411
|
|
|
), |
412
|
|
|
], |
413
|
|
|
[ |
414
|
|
|
'+(one)', |
415
|
|
|
[ |
416
|
|
|
$token1 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 0), |
417
|
|
|
$token2 = new GroupBeginToken('(', 1, '(', null), |
418
|
|
|
$token3 = new WordToken('one', 2, '', 'one'), |
419
|
|
|
$token4 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 5), |
420
|
|
|
], |
421
|
|
|
new Query([new Mandatory(new Group([new Term($token3)], $token2, $token4), $token1)]), |
422
|
|
|
], |
423
|
|
|
[ |
424
|
|
|
'-one', |
425
|
|
|
[ |
426
|
|
|
$token1 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 0), |
427
|
|
|
$token2 = new WordToken('one', 1, '', 'one'), |
428
|
|
|
], |
429
|
|
|
new Query([new Prohibited(new Term($token2), $token1)]), |
430
|
|
|
], |
431
|
|
|
[ |
432
|
|
|
'-one AND -two', |
433
|
|
|
[ |
434
|
|
|
$token1 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 0), |
435
|
|
|
$token2 = new WordToken('one', 1, '', 'one'), |
436
|
|
|
$token3 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 5), |
437
|
|
|
$token4 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 9), |
438
|
|
|
$token5 = new WordToken('two', 10, '', 'two'), |
439
|
|
|
], |
440
|
|
|
new Query( |
441
|
|
|
[ |
442
|
|
|
new LogicalAnd( |
443
|
|
|
new Prohibited(new Term($token2), $token1), |
444
|
|
|
new Prohibited(new Term($token5), $token4), |
445
|
|
|
$token3 |
446
|
|
|
), |
447
|
|
|
] |
448
|
|
|
), |
449
|
|
|
], |
450
|
|
|
[ |
451
|
|
|
'-one OR -two', |
452
|
|
|
[ |
453
|
|
|
$token1 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 0), |
454
|
|
|
$token2 = new WordToken('one', 1, '', 'one'), |
455
|
|
|
$token3 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 5), |
456
|
|
|
$token4 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 8), |
457
|
|
|
$token5 = new WordToken('two', 9, '', 'two'), |
458
|
|
|
], |
459
|
|
|
new Query( |
460
|
|
|
[ |
461
|
|
|
new LogicalOr( |
462
|
|
|
new Prohibited(new Term($token2), $token1), |
463
|
|
|
new Prohibited(new Term($token5), $token4), |
464
|
|
|
$token3 |
465
|
|
|
), |
466
|
|
|
] |
467
|
|
|
), |
468
|
|
|
], |
469
|
|
|
[ |
470
|
|
|
'-one OR -two AND -three', |
471
|
|
|
[ |
472
|
|
|
$token1 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 0), |
473
|
|
|
$token2 = new WordToken('one', 1, '', 'one'), |
474
|
|
|
$token3 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 5), |
475
|
|
|
$token4 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 8), |
476
|
|
|
$token5 = new WordToken('two', 9, '', 'two'), |
477
|
|
|
$token6 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 13), |
478
|
|
|
$token7 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 17), |
479
|
|
|
$token8 = new WordToken('three', 18, '', 'three'), |
480
|
|
|
], |
481
|
|
|
new Query( |
482
|
|
|
[ |
483
|
|
|
new LogicalOr( |
484
|
|
|
new Prohibited(new Term($token2), $token1), |
485
|
|
|
new LogicalAnd( |
486
|
|
|
new Prohibited(new Term($token5), $token4), |
487
|
|
|
new Prohibited(new Term($token8), $token7), |
488
|
|
|
$token6 |
489
|
|
|
), |
490
|
|
|
$token3 |
491
|
|
|
), |
492
|
|
|
] |
493
|
|
|
), |
494
|
|
|
], |
495
|
|
|
[ |
496
|
|
|
'-one AND -two OR -three', |
497
|
|
|
[ |
498
|
|
|
$token1 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 0), |
499
|
|
|
$token2 = new WordToken('one', 1, '', 'one'), |
500
|
|
|
$token3 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 5), |
501
|
|
|
$token4 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 9), |
502
|
|
|
$token5 = new WordToken('two', 10, '', 'two'), |
503
|
|
|
$token6 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 14), |
504
|
|
|
$token7 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 17), |
505
|
|
|
$token8 = new WordToken('three', 18, '', 'three'), |
506
|
|
|
], |
507
|
|
|
new Query( |
508
|
|
|
[ |
509
|
|
|
new LogicalOr( |
510
|
|
|
new LogicalAnd( |
511
|
|
|
new Prohibited(new Term($token2), $token1), |
512
|
|
|
new Prohibited(new Term($token5), $token4), |
513
|
|
|
$token3 |
514
|
|
|
), |
515
|
|
|
new Prohibited(new Term($token8), $token7), |
516
|
|
|
$token6 |
517
|
|
|
), |
518
|
|
|
] |
519
|
|
|
), |
520
|
|
|
], |
521
|
|
|
[ |
522
|
|
|
'-(one)', |
523
|
|
|
[ |
524
|
|
|
$token1 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 0), |
525
|
|
|
$token2 = new GroupBeginToken('(', 1, '(', null), |
526
|
|
|
$token3 = new WordToken('one', 2, '', 'one'), |
527
|
|
|
$token4 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 5), |
528
|
|
|
], |
529
|
|
|
new Query([new Prohibited(new Group([new Term($token3)], $token2, $token4), $token1)]), |
530
|
|
|
], |
531
|
|
|
[ |
532
|
|
|
'(one OR +two three)', |
533
|
|
|
[ |
534
|
|
|
$token1 = new GroupBeginToken('(', 0, '(', null), |
535
|
|
|
$token2 = new WordToken('one', 1, '', 'one'), |
536
|
|
|
$token3 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 5), |
537
|
|
|
$token4 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 8), |
538
|
|
|
$token5 = new WordToken('two', 9, '', 'two'), |
539
|
|
|
$token6 = new WordToken('three', 13, '', 'three'), |
540
|
|
|
$token7 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 18), |
541
|
|
|
], |
542
|
|
|
new Query( |
543
|
|
|
[ |
544
|
|
|
new Group( |
545
|
|
|
[ |
546
|
|
|
new LogicalOr(new Term($token2), new Mandatory(new Term($token5), $token4), $token3), |
547
|
|
|
new Term($token6), |
548
|
|
|
], |
549
|
|
|
$token1, |
550
|
|
|
$token7 |
551
|
|
|
), |
552
|
|
|
] |
553
|
|
|
), |
554
|
|
|
], |
555
|
|
|
[ |
556
|
|
|
'(one OR -two three)', |
557
|
|
|
[ |
558
|
|
|
$token1 = new GroupBeginToken('(', 0, '(', null), |
559
|
|
|
$token2 = new WordToken('one', 1, '', 'one'), |
560
|
|
|
$token3 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 5), |
561
|
|
|
$token4 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 8), |
562
|
|
|
$token5 = new WordToken('two', 9, '', 'two'), |
563
|
|
|
$token6 = new WordToken('three', 13, '', 'three'), |
564
|
|
|
$token7 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 18), |
565
|
|
|
], |
566
|
|
|
new Query( |
567
|
|
|
[ |
568
|
|
|
new Group( |
569
|
|
|
[ |
570
|
|
|
new LogicalOr(new Term($token2), new Prohibited(new Term($token5), $token4), $token3), |
571
|
|
|
new Term($token6), |
572
|
|
|
], |
573
|
|
|
$token1, |
574
|
|
|
$token7 |
575
|
|
|
), |
576
|
|
|
] |
577
|
|
|
), |
578
|
|
|
], |
579
|
|
|
[ |
580
|
|
|
'((one))', |
581
|
|
|
[ |
582
|
|
|
$token1 = new GroupBeginToken('(', 0, '(', null), |
583
|
|
|
$token2 = new GroupBeginToken('(', 1, '(', null), |
584
|
|
|
$token3 = new WordToken('one', 2, '', 'one'), |
585
|
|
|
$token4 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 5), |
586
|
|
|
$token5 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 6), |
587
|
|
|
], |
588
|
|
|
new Query([new Group([new Group([new Term($token3)], $token2, $token4)], $token1, $token5)]), |
589
|
|
|
], |
590
|
|
|
[ |
591
|
|
|
'NOT NOT one NOT NOT NOT two', |
592
|
|
|
[ |
593
|
|
|
$token1 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 0), |
594
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 4), |
595
|
|
|
$token3 = new WordToken('one', 8, '', 'one'), |
596
|
|
|
$token4 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 12), |
597
|
|
|
$token5 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 16), |
598
|
|
|
$token6 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 20), |
599
|
|
|
$token7 = new WordToken('two', 24, '', 'two'), |
600
|
|
|
], |
601
|
|
|
new Query( |
602
|
|
|
[ |
603
|
|
|
new LogicalNot(new LogicalNot(new Term($token3), $token2), $token1), |
604
|
|
|
new LogicalNot(new LogicalNot(new LogicalNot(new Term($token7), $token6), $token5), $token4), |
605
|
|
|
] |
606
|
|
|
), |
607
|
|
|
], |
608
|
|
|
[ |
609
|
|
|
'NOT !one NOT !!two', |
610
|
|
|
[ |
611
|
|
|
$token1 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 0), |
612
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_NOT_2, '!', 4), |
613
|
|
|
$token3 = new WordToken('one', 5, '', 'one'), |
614
|
|
|
$token4 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 9), |
615
|
|
|
$token5 = new Token(Tokenizer::TOKEN_LOGICAL_NOT_2, '!', 13), |
616
|
|
|
$token6 = new Token(Tokenizer::TOKEN_LOGICAL_NOT_2, '!', 14), |
617
|
|
|
$token7 = new WordToken('two', 15, '', 'two'), |
618
|
|
|
], |
619
|
|
|
new Query( |
620
|
|
|
[ |
621
|
|
|
new LogicalNot(new LogicalNot(new Term($token3), $token2), $token1), |
622
|
|
|
new LogicalNot(new LogicalNot(new LogicalNot(new Term($token7), $token6), $token5), $token4), |
623
|
|
|
] |
624
|
|
|
), |
625
|
|
|
], |
626
|
|
|
[ |
627
|
|
|
'one AND NOT "two"', |
628
|
|
|
[ |
629
|
|
|
$token1 = new WordToken('one', 0, '', 'one'), |
630
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 4), |
631
|
|
|
$token3 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 8), |
632
|
|
|
$token4 = new PhraseToken('"two"', 12, '', '"', 'two'), |
633
|
|
|
], |
634
|
|
|
new Query([new LogicalAnd(new Term($token1), new LogicalNot(new Term($token4), $token3), $token2)]), |
635
|
|
|
], |
636
|
|
|
[ |
637
|
|
|
'one AND NOT @two', |
638
|
|
|
[ |
639
|
|
|
$token1 = new WordToken('one', 0, '', 'one'), |
640
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 4), |
641
|
|
|
$token3 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 8), |
642
|
|
|
$token4 = new UserToken('@two', 12, '@', 'two'), |
643
|
|
|
], |
644
|
|
|
new Query([new LogicalAnd(new Term($token1), new LogicalNot(new Term($token4), $token3), $token2)]), |
645
|
|
|
], |
646
|
|
|
[ |
647
|
|
|
'one AND NOT #two', |
648
|
|
|
[ |
649
|
|
|
$token1 = new WordToken('one', 0, '', 'one'), |
650
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 4), |
651
|
|
|
$token3 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 8), |
652
|
|
|
$token4 = new TagToken('#two', 12, '#', 'two'), |
653
|
|
|
], |
654
|
|
|
new Query([new LogicalAnd(new Term($token1), new LogicalNot(new Term($token4), $token3), $token2)]), |
655
|
|
|
], |
656
|
|
|
]; |
657
|
|
|
} |
658
|
|
|
|
659
|
|
|
|
660
|
|
|
public function providerForTestQueryCorrected(): array |
661
|
|
|
{ |
662
|
|
|
return [ |
663
|
|
|
[ |
664
|
|
|
'one"', |
665
|
|
|
'one', |
666
|
|
|
[$token1 = new WordToken('one', 0, '', 'one'), $token2 = new Token(Tokenizer::TOKEN_BAILOUT, '"', 3)], |
667
|
|
|
new Query([new Term($token1)]), |
668
|
|
|
[new Correction(Parser::CORRECTION_BAILOUT_TOKEN_IGNORED, $token2)], |
669
|
|
|
], |
670
|
|
|
[ |
671
|
|
|
'one AND two AND', |
672
|
|
|
'one AND two', |
673
|
|
|
[ |
674
|
|
|
$token1 = new WordToken('one', 0, '', 'one'), |
675
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 4), |
676
|
|
|
$token3 = new WordToken('two', 8, '', 'two'), |
677
|
|
|
$token4 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 12), |
678
|
|
|
], |
679
|
|
|
new Query([new LogicalAnd(new Term($token1), new Term($token3), $token2)]), |
680
|
|
|
[new Correction(Parser::CORRECTION_BINARY_OPERATOR_MISSING_RIGHT_OPERAND_IGNORED, $token4)], |
681
|
|
|
], |
682
|
|
|
[ |
683
|
|
|
'AND one AND two', |
684
|
|
|
'one AND two', |
685
|
|
|
[ |
686
|
|
|
$token1 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 0), |
687
|
|
|
$token2 = new WordToken('one', 4, '', 'one'), |
688
|
|
|
$token3 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 8), |
689
|
|
|
$token4 = new WordToken('two', 12, '', 'two'), |
690
|
|
|
], |
691
|
|
|
new Query([new LogicalAnd(new Term($token2), new Term($token4), $token3)]), |
692
|
|
|
[new Correction(Parser::CORRECTION_BINARY_OPERATOR_MISSING_LEFT_OPERAND_IGNORED, $token1)], |
693
|
|
|
], |
694
|
|
|
[ |
695
|
|
|
'AND AND one AND AND two', |
696
|
|
|
'one two', |
697
|
|
|
[ |
698
|
|
|
$token1 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 0), |
699
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 4), |
700
|
|
|
$token3 = new WordToken('one', 8, '', 'one'), |
701
|
|
|
$token4 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 12), |
702
|
|
|
$token5 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 16), |
703
|
|
|
$token6 = new WordToken('two', 20, '', 'two'), |
704
|
|
|
], |
705
|
|
|
new Query([new Term($token3), new Term($token6)]), |
706
|
|
|
[ |
707
|
|
|
new Correction(Parser::CORRECTION_BINARY_OPERATOR_MISSING_LEFT_OPERAND_IGNORED, $token1), |
708
|
|
|
new Correction(Parser::CORRECTION_BINARY_OPERATOR_MISSING_LEFT_OPERAND_IGNORED, $token2), |
709
|
|
|
new Correction(Parser::CORRECTION_BINARY_OPERATOR_FOLLOWING_OPERATOR_IGNORED, $token4, $token5), |
710
|
|
|
], |
711
|
|
|
], |
712
|
|
|
[ |
713
|
|
|
'OR one OR two', |
714
|
|
|
'one OR two', |
715
|
|
|
[ |
716
|
|
|
$token1 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 0), |
717
|
|
|
$token2 = new WordToken('one', 3, '', 'one'), |
718
|
|
|
$token3 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 7), |
719
|
|
|
$token4 = new WordToken('two', 10, '', 'two'), |
720
|
|
|
], |
721
|
|
|
new Query([new LogicalOr(new Term($token2), new Term($token4), $token3)]), |
722
|
|
|
[new Correction(Parser::CORRECTION_BINARY_OPERATOR_MISSING_LEFT_OPERAND_IGNORED, $token1)], |
723
|
|
|
], |
724
|
|
|
[ |
725
|
|
|
'OR OR one OR OR two', |
726
|
|
|
'one two', |
727
|
|
|
[ |
728
|
|
|
$token1 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 0), |
729
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 3), |
730
|
|
|
$token3 = new WordToken('one', 6, '', 'one'), |
731
|
|
|
$token4 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 10), |
732
|
|
|
$token5 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 13), |
733
|
|
|
$token6 = new WordToken('two', 16, '', 'two'), |
734
|
|
|
], |
735
|
|
|
new Query([new Term($token3), new Term($token6)]), |
736
|
|
|
[ |
737
|
|
|
new Correction(Parser::CORRECTION_BINARY_OPERATOR_MISSING_LEFT_OPERAND_IGNORED, $token1), |
738
|
|
|
new Correction(Parser::CORRECTION_BINARY_OPERATOR_MISSING_LEFT_OPERAND_IGNORED, $token2), |
739
|
|
|
new Correction(Parser::CORRECTION_BINARY_OPERATOR_FOLLOWING_OPERATOR_IGNORED, $token4, $token5), |
740
|
|
|
], |
741
|
|
|
], |
742
|
|
|
[ |
743
|
|
|
'OR OR one OR OR AND two', |
744
|
|
|
'one two', |
745
|
|
|
[ |
746
|
|
|
$token1 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 0), |
747
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 3), |
748
|
|
|
$token3 = new WordToken('one', 6, '', 'one'), |
749
|
|
|
$token4 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 10), |
750
|
|
|
$token5 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 13), |
751
|
|
|
$token6 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 16), |
752
|
|
|
$token7 = new WordToken('two', 20, '', 'two'), |
753
|
|
|
], |
754
|
|
|
new Query([new Term($token3), new Term($token7)]), |
755
|
|
|
[ |
756
|
|
|
new Correction(Parser::CORRECTION_BINARY_OPERATOR_MISSING_LEFT_OPERAND_IGNORED, $token1), |
757
|
|
|
new Correction(Parser::CORRECTION_BINARY_OPERATOR_MISSING_LEFT_OPERAND_IGNORED, $token2), |
758
|
|
|
new Correction( |
759
|
|
|
Parser::CORRECTION_BINARY_OPERATOR_FOLLOWING_OPERATOR_IGNORED, |
760
|
|
|
$token4, |
761
|
|
|
$token5, |
762
|
|
|
$token6 |
763
|
|
|
), |
764
|
|
|
], |
765
|
|
|
], |
766
|
|
|
[ |
767
|
|
|
'one OR two AND OR NOT', |
768
|
|
|
'one OR two', |
769
|
|
|
[ |
770
|
|
|
$token1 = new WordToken('one', 0, '', 'one'), |
771
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 4), |
772
|
|
|
$token3 = new WordToken('two', 7, '', 'two'), |
773
|
|
|
$token4 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 11), |
774
|
|
|
$token5 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 15), |
775
|
|
|
$token6 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 18), |
776
|
|
|
], |
777
|
|
|
new Query([new LogicalOr(new Term($token1), new Term($token3), $token2)]), |
778
|
|
|
[ |
779
|
|
|
new Correction(Parser::CORRECTION_BINARY_OPERATOR_FOLLOWING_OPERATOR_IGNORED, $token4, $token5), |
780
|
|
|
new Correction(Parser::CORRECTION_UNARY_OPERATOR_MISSING_OPERAND_IGNORED, $token6), |
781
|
|
|
], |
782
|
|
|
], |
783
|
|
|
[ |
784
|
|
|
'AND OR one AND OR two AND OR three', |
785
|
|
|
'one two three', |
786
|
|
|
[ |
787
|
|
|
$token1 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 0), |
788
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 4), |
789
|
|
|
$token3 = new WordToken('one', 7, '', 'one'), |
790
|
|
|
$token4 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 11), |
791
|
|
|
$token5 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 15), |
792
|
|
|
$token6 = new WordToken('two', 18, '', 'two'), |
793
|
|
|
$token7 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 22), |
794
|
|
|
$token8 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 26), |
795
|
|
|
$token9 = new WordToken('three', 29, '', 'three'), |
796
|
|
|
], |
797
|
|
|
new Query([new Term($token3), new Term($token6), new Term($token9)]), |
798
|
|
|
[ |
799
|
|
|
new Correction(Parser::CORRECTION_BINARY_OPERATOR_MISSING_LEFT_OPERAND_IGNORED, $token1), |
800
|
|
|
new Correction(Parser::CORRECTION_BINARY_OPERATOR_MISSING_LEFT_OPERAND_IGNORED, $token2), |
801
|
|
|
new Correction(Parser::CORRECTION_BINARY_OPERATOR_FOLLOWING_OPERATOR_IGNORED, $token4, $token5), |
802
|
|
|
new Correction(Parser::CORRECTION_BINARY_OPERATOR_FOLLOWING_OPERATOR_IGNORED, $token7, $token8), |
803
|
|
|
], |
804
|
|
|
], |
805
|
|
|
[ |
806
|
|
|
'OR AND one OR AND two OR AND three', |
807
|
|
|
'one two three', |
808
|
|
|
[ |
809
|
|
|
$token1 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 0), |
810
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 3), |
811
|
|
|
$token3 = new WordToken('one', 7, '', 'one'), |
812
|
|
|
$token4 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 11), |
813
|
|
|
$token5 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 14), |
814
|
|
|
$token6 = new WordToken('two', 18, '', 'two'), |
815
|
|
|
$token7 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 22), |
816
|
|
|
$token8 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 25), |
817
|
|
|
$token9 = new WordToken('three', 29, '', 'three'), |
818
|
|
|
], |
819
|
|
|
new Query([new Term($token3), new Term($token6), new Term($token9)]), |
820
|
|
|
[ |
821
|
|
|
new Correction(Parser::CORRECTION_BINARY_OPERATOR_MISSING_LEFT_OPERAND_IGNORED, $token1), |
822
|
|
|
new Correction(Parser::CORRECTION_BINARY_OPERATOR_MISSING_LEFT_OPERAND_IGNORED, $token2), |
823
|
|
|
new Correction(Parser::CORRECTION_BINARY_OPERATOR_FOLLOWING_OPERATOR_IGNORED, $token4, $token5), |
824
|
|
|
new Correction(Parser::CORRECTION_BINARY_OPERATOR_FOLLOWING_OPERATOR_IGNORED, $token7, $token8), |
825
|
|
|
], |
826
|
|
|
], |
827
|
|
|
[ |
828
|
|
|
'one AND NOT AND two', |
829
|
|
|
'one two', |
830
|
|
|
[ |
831
|
|
|
$token1 = new WordToken('one', 0, '', 'one'), |
832
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 4), |
833
|
|
|
$token3 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 8), |
834
|
|
|
$token4 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 12), |
835
|
|
|
$token5 = new WordToken('two', 16, '', 'two'), |
836
|
|
|
], |
837
|
|
|
new Query([new Term($token1), new Term($token5)]), |
838
|
|
|
[ |
839
|
|
|
new Correction( |
840
|
|
|
Parser::CORRECTION_BINARY_OPERATOR_FOLLOWING_OPERATOR_IGNORED, |
841
|
|
|
$token2, |
842
|
|
|
$token3, |
843
|
|
|
$token4 |
844
|
|
|
), |
845
|
|
|
], |
846
|
|
|
], |
847
|
|
|
[ |
848
|
|
|
'one NOT AND two', |
849
|
|
|
'one two', |
850
|
|
|
[ |
851
|
|
|
$token1 = new WordToken('one', 0, '', 'one'), |
852
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 4), |
853
|
|
|
$token3 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 8), |
854
|
|
|
$token4 = new WordToken('two', 12, '', 'two'), |
855
|
|
|
], |
856
|
|
|
new Query([new Term($token1), new Term($token4)]), |
857
|
|
|
[new Correction(Parser::CORRECTION_BINARY_OPERATOR_FOLLOWING_OPERATOR_IGNORED, $token2, $token3)], |
858
|
|
|
], |
859
|
|
|
[ |
860
|
|
|
'one NOT AND NOT two', |
861
|
|
|
'one NOT two', |
862
|
|
|
[ |
863
|
|
|
$token1 = new WordToken('one', 0, '', 'one'), |
864
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 4), |
865
|
|
|
$token3 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 8), |
866
|
|
|
$token4 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 12), |
867
|
|
|
$token5 = new WordToken('two', 16, '', 'two'), |
868
|
|
|
], |
869
|
|
|
new Query([new Term($token1), new LogicalNot(new Term($token5), $token4)]), |
870
|
|
|
[new Correction(Parser::CORRECTION_BINARY_OPERATOR_FOLLOWING_OPERATOR_IGNORED, $token2, $token3)], |
871
|
|
|
], |
872
|
|
|
[ |
873
|
|
|
'one OR NOT OR two', |
874
|
|
|
'one two', |
875
|
|
|
[ |
876
|
|
|
$token1 = new WordToken('one', 0, '', 'one'), |
877
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 4), |
878
|
|
|
$token3 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 7), |
879
|
|
|
$token4 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 11), |
880
|
|
|
$token5 = new WordToken('two', 14, '', 'two'), |
881
|
|
|
], |
882
|
|
|
new Query([new Term($token1), new Term($token5)]), |
883
|
|
|
[ |
884
|
|
|
new Correction( |
885
|
|
|
Parser::CORRECTION_BINARY_OPERATOR_FOLLOWING_OPERATOR_IGNORED, |
886
|
|
|
$token2, |
887
|
|
|
$token3, |
888
|
|
|
$token4 |
889
|
|
|
), |
890
|
|
|
], |
891
|
|
|
], |
892
|
|
|
[ |
893
|
|
|
'one NOT OR two', |
894
|
|
|
'one two', |
895
|
|
|
[ |
896
|
|
|
$token1 = new WordToken('one', 0, '', 'one'), |
897
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 4), |
898
|
|
|
$token3 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 8), |
899
|
|
|
$token4 = new WordToken('two', 11, '', 'two'), |
900
|
|
|
], |
901
|
|
|
new Query([new Term($token1), new Term($token4)]), |
902
|
|
|
[new Correction(Parser::CORRECTION_BINARY_OPERATOR_FOLLOWING_OPERATOR_IGNORED, $token2, $token3)], |
903
|
|
|
], |
904
|
|
|
[ |
905
|
|
|
'one NOT OR NOT two', |
906
|
|
|
'one NOT two', |
907
|
|
|
[ |
908
|
|
|
$token1 = new WordToken('one', 0, '', 'one'), |
909
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 4), |
910
|
|
|
$token3 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 8), |
911
|
|
|
$token4 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 11), |
912
|
|
|
$token5 = new WordToken('two', 15, '', 'two'), |
913
|
|
|
], |
914
|
|
|
new Query([new Term($token1), new LogicalNot(new Term($token5), $token4)]), |
915
|
|
|
[new Correction(Parser::CORRECTION_BINARY_OPERATOR_FOLLOWING_OPERATOR_IGNORED, $token2, $token3)], |
916
|
|
|
], |
917
|
|
|
[ |
918
|
|
|
'(one AND two OR NOT)', |
919
|
|
|
'(one AND two)', |
920
|
|
|
[ |
921
|
|
|
$token1 = new GroupBeginToken('(', 0, '(', null), |
922
|
|
|
$token2 = new WordToken('one', 1, '', 'one'), |
923
|
|
|
$token3 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 5), |
924
|
|
|
$token4 = new WordToken('two', 9, '', 'two'), |
925
|
|
|
$token5 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 13), |
926
|
|
|
$token6 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 16), |
927
|
|
|
$token7 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 19), |
928
|
|
|
], |
929
|
|
|
new Query([ |
930
|
|
|
new Group([new LogicalAnd(new Term($token2), new Term($token4), $token3)], $token1, $token7), |
931
|
|
|
]), |
932
|
|
|
[ |
933
|
|
|
new Correction(Parser::CORRECTION_UNARY_OPERATOR_MISSING_OPERAND_IGNORED, $token6), |
934
|
|
|
new Correction(Parser::CORRECTION_BINARY_OPERATOR_MISSING_RIGHT_OPERAND_IGNORED, $token5), |
935
|
|
|
], |
936
|
|
|
], |
937
|
|
|
[ |
938
|
|
|
'(AND one OR two)', |
939
|
|
|
'(one OR two)', |
940
|
|
|
[ |
941
|
|
|
$token1 = new GroupBeginToken('(', 0, '(', null), |
942
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 1), |
943
|
|
|
$token3 = new WordToken('one', 5, '', 'one'), |
944
|
|
|
$token4 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 9), |
945
|
|
|
$token5 = new WordToken('two', 12, '', 'two'), |
946
|
|
|
$token6 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 15), |
947
|
|
|
], |
948
|
|
|
new Query([ |
949
|
|
|
new Group([new LogicalOr(new Term($token3), new Term($token5), $token4)], $token1, $token6), |
950
|
|
|
]), |
951
|
|
|
[new Correction(Parser::CORRECTION_BINARY_OPERATOR_MISSING_LEFT_OPERAND_IGNORED, $token2)], |
952
|
|
|
], |
953
|
|
|
[ |
954
|
|
|
'AND (((OR AND one AND NOT OR))) OR NOT', |
955
|
|
|
'(((one)))', |
956
|
|
|
[ |
957
|
|
|
$token1 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 0), |
958
|
|
|
$token2 = new GroupBeginToken('(', 4, '(', null), |
959
|
|
|
$token3 = new GroupBeginToken('(', 5, '(', null), |
960
|
|
|
$token4 = new GroupBeginToken('(', 6, '(', null), |
961
|
|
|
$token5 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 7), |
962
|
|
|
$token6 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 10), |
963
|
|
|
$token7 = new WordToken('one', 14, '', 'one'), |
964
|
|
|
$token8 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 18), |
965
|
|
|
$token9 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 22), |
966
|
|
|
$token10 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 26), |
967
|
|
|
$token11 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 28), |
968
|
|
|
$token12 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 29), |
969
|
|
|
$token13 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 30), |
970
|
|
|
$token14 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 32), |
971
|
|
|
$token15 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 35), |
972
|
|
|
], |
973
|
|
|
new Query( |
974
|
|
|
[ |
975
|
|
|
new Group( |
976
|
|
|
[new Group([new Group([new Term($token7)], $token4, $token11)], $token3, $token12)], |
977
|
|
|
$token2, |
978
|
|
|
$token13 |
979
|
|
|
), |
980
|
|
|
] |
981
|
|
|
), |
982
|
|
|
[ |
983
|
|
|
new Correction(Parser::CORRECTION_BINARY_OPERATOR_MISSING_LEFT_OPERAND_IGNORED, $token1), |
984
|
|
|
new Correction(Parser::CORRECTION_BINARY_OPERATOR_MISSING_LEFT_OPERAND_IGNORED, $token5), |
985
|
|
|
new Correction(Parser::CORRECTION_BINARY_OPERATOR_MISSING_LEFT_OPERAND_IGNORED, $token6), |
986
|
|
|
new Correction( |
987
|
|
|
Parser::CORRECTION_BINARY_OPERATOR_FOLLOWING_OPERATOR_IGNORED, |
988
|
|
|
$token8, |
989
|
|
|
$token9, |
990
|
|
|
$token10 |
991
|
|
|
), |
992
|
|
|
new Correction(Parser::CORRECTION_UNARY_OPERATOR_MISSING_OPERAND_IGNORED, $token15), |
993
|
|
|
new Correction(Parser::CORRECTION_BINARY_OPERATOR_MISSING_RIGHT_OPERAND_IGNORED, $token14), |
994
|
|
|
], |
995
|
|
|
], |
996
|
|
|
[ |
997
|
|
|
'one ()', |
998
|
|
|
'one', |
999
|
|
|
[ |
1000
|
|
|
$token1 = new WordToken('one', 0, '', 'one'), |
1001
|
|
|
$token2 = new GroupBeginToken('(', 4, '(', null), |
1002
|
|
|
$token3 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 5), |
1003
|
|
|
], |
1004
|
|
|
new Query([new Term($token1)]), |
1005
|
|
|
[new Correction(Parser::CORRECTION_EMPTY_GROUP_IGNORED, $token2, $token3)], |
1006
|
|
|
], |
1007
|
|
|
[ |
1008
|
|
|
'one (())', |
1009
|
|
|
'one', |
1010
|
|
|
[ |
1011
|
|
|
$token1 = new WordToken('one', 0, '', 'one'), |
1012
|
|
|
$token2 = new GroupBeginToken('(', 4, '(', null), |
1013
|
|
|
$token3 = new GroupBeginToken('(', 5, '(', null), |
1014
|
|
|
$token4 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 6), |
1015
|
|
|
$token5 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 7), |
1016
|
|
|
], |
1017
|
|
|
new Query([new Term($token1)]), |
1018
|
|
|
[ |
1019
|
|
|
new Correction(Parser::CORRECTION_EMPTY_GROUP_IGNORED, $token3, $token4), |
1020
|
|
|
new Correction(Parser::CORRECTION_EMPTY_GROUP_IGNORED, $token2, $token5), |
1021
|
|
|
], |
1022
|
|
|
], |
1023
|
|
|
[ |
1024
|
|
|
'one AND (()) OR two', |
1025
|
|
|
'one two', |
1026
|
|
|
[ |
1027
|
|
|
$token1 = new WordToken('one', 0, '', 'one'), |
1028
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 4), |
1029
|
|
|
$token3 = new GroupBeginToken('(', 8, '(', null), |
1030
|
|
|
$token4 = new GroupBeginToken('(', 9, '(', null), |
1031
|
|
|
$token5 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 10), |
1032
|
|
|
$token6 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 11), |
1033
|
|
|
$token7 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 13), |
1034
|
|
|
$token8 = new WordToken('two', 16, '', 'two'), |
1035
|
|
|
], |
1036
|
|
|
new Query([new Term($token1), new Term($token8)]), |
1037
|
|
|
[ |
1038
|
|
|
new Correction(Parser::CORRECTION_EMPTY_GROUP_IGNORED, $token4, $token5), |
1039
|
|
|
new Correction(Parser::CORRECTION_EMPTY_GROUP_IGNORED, $token2, $token3, $token6, $token7), |
1040
|
|
|
], |
1041
|
|
|
], |
1042
|
|
|
[ |
1043
|
|
|
'one (AND OR NOT)', |
1044
|
|
|
'one', |
1045
|
|
|
[ |
1046
|
|
|
$token1 = new WordToken('one', 0, '', 'one'), |
1047
|
|
|
$token2 = new GroupBeginToken('(', 4, '(', null), |
1048
|
|
|
$token3 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 5), |
1049
|
|
|
$token4 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 9), |
1050
|
|
|
$token5 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 12), |
1051
|
|
|
$token6 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 15), |
1052
|
|
|
], |
1053
|
|
|
new Query([new Term($token1)]), |
1054
|
|
|
[ |
1055
|
|
|
new Correction(Parser::CORRECTION_BINARY_OPERATOR_MISSING_LEFT_OPERAND_IGNORED, $token3), |
1056
|
|
|
new Correction(Parser::CORRECTION_BINARY_OPERATOR_MISSING_LEFT_OPERAND_IGNORED, $token4), |
1057
|
|
|
new Correction(Parser::CORRECTION_UNARY_OPERATOR_MISSING_OPERAND_IGNORED, $token5), |
1058
|
|
|
new Correction(Parser::CORRECTION_EMPTY_GROUP_IGNORED, $token2, $token6), |
1059
|
|
|
], |
1060
|
|
|
], |
1061
|
|
|
[ |
1062
|
|
|
'one) (AND)) OR NOT)', |
1063
|
|
|
'one', |
1064
|
|
|
[ |
1065
|
|
|
$token1 = new WordToken('one', 0, '', 'one'), |
1066
|
|
|
$token2 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 3), |
1067
|
|
|
$token3 = new GroupBeginToken('(', 5, '(', null), |
1068
|
|
|
$token4 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 6), |
1069
|
|
|
$token5 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 9), |
1070
|
|
|
$token6 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 10), |
1071
|
|
|
$token7 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 12), |
1072
|
|
|
$token8 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 15), |
1073
|
|
|
$token9 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 18), |
1074
|
|
|
], |
1075
|
|
|
new Query([new Term($token1)]), |
1076
|
|
|
[ |
1077
|
|
|
new Correction(Parser::CORRECTION_UNMATCHED_GROUP_RIGHT_DELIMITER_IGNORED, $token9), |
1078
|
|
|
new Correction(Parser::CORRECTION_UNMATCHED_GROUP_RIGHT_DELIMITER_IGNORED, $token6), |
1079
|
|
|
new Correction(Parser::CORRECTION_UNMATCHED_GROUP_RIGHT_DELIMITER_IGNORED, $token2), |
1080
|
|
|
new Correction(Parser::CORRECTION_BINARY_OPERATOR_MISSING_LEFT_OPERAND_IGNORED, $token4), |
1081
|
|
|
new Correction(Parser::CORRECTION_EMPTY_GROUP_IGNORED, $token3, $token5, $token7), |
1082
|
|
|
new Correction(Parser::CORRECTION_UNARY_OPERATOR_MISSING_OPERAND_IGNORED, $token8), |
1083
|
|
|
], |
1084
|
|
|
], |
1085
|
|
|
[ |
1086
|
|
|
'(one( (AND) OR NOT((', |
1087
|
|
|
'one', |
1088
|
|
|
[ |
1089
|
|
|
$token1 = new GroupBeginToken('(', 0, '(', null), |
1090
|
|
|
$token2 = new WordToken('one', 1, '', 'one'), |
1091
|
|
|
$token3 = new GroupBeginToken('(', 4, '(', null), |
1092
|
|
|
$token4 = new GroupBeginToken('(', 6, '(', null), |
1093
|
|
|
$token5 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 7), |
1094
|
|
|
$token6 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 10), |
1095
|
|
|
$token7 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 12), |
1096
|
|
|
$token8 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 15), |
1097
|
|
|
$token9 = new GroupBeginToken('(', 18, '(', null), |
1098
|
|
|
$token10 = new GroupBeginToken('(', 19, '(', null), |
1099
|
|
|
], |
1100
|
|
|
new Query([new Term($token2)]), |
1101
|
|
|
[ |
1102
|
|
|
new Correction(Parser::CORRECTION_UNMATCHED_GROUP_LEFT_DELIMITER_IGNORED, $token10), |
1103
|
|
|
new Correction(Parser::CORRECTION_UNMATCHED_GROUP_LEFT_DELIMITER_IGNORED, $token9), |
1104
|
|
|
new Correction(Parser::CORRECTION_UNMATCHED_GROUP_LEFT_DELIMITER_IGNORED, $token3), |
1105
|
|
|
new Correction(Parser::CORRECTION_UNMATCHED_GROUP_LEFT_DELIMITER_IGNORED, $token1), |
1106
|
|
|
new Correction(Parser::CORRECTION_BINARY_OPERATOR_MISSING_LEFT_OPERAND_IGNORED, $token5), |
1107
|
|
|
new Correction(Parser::CORRECTION_EMPTY_GROUP_IGNORED, $token4, $token6, $token7), |
1108
|
|
|
new Correction(Parser::CORRECTION_UNARY_OPERATOR_MISSING_OPERAND_IGNORED, $token8), |
1109
|
|
|
], |
1110
|
|
|
], |
1111
|
|
|
[ |
1112
|
|
|
'OR NOT (one OR two AND OR NOT) OR three AND NOT', |
1113
|
|
|
'NOT (one OR two) OR three', |
1114
|
|
|
[ |
1115
|
|
|
$token1 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 0), |
1116
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 3), |
1117
|
|
|
$token3 = new GroupBeginToken('(', 7, '(', null), |
1118
|
|
|
$token4 = new WordToken('one', 8, '', 'one'), |
1119
|
|
|
$token5 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 12), |
1120
|
|
|
$token6 = new WordToken('two', 15, '', 'two'), |
1121
|
|
|
$token7 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 19), |
1122
|
|
|
$token8 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 23), |
1123
|
|
|
$token9 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 26), |
1124
|
|
|
$token10 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 29), |
1125
|
|
|
$token11 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 31), |
1126
|
|
|
$token12 = new WordToken('three', 34, '', 'three'), |
1127
|
|
|
$token13 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 40), |
1128
|
|
|
$token14 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 44), |
1129
|
|
|
], |
1130
|
|
|
new Query( |
1131
|
|
|
[ |
1132
|
|
|
new LogicalOr( |
1133
|
|
|
new LogicalNot( |
1134
|
|
|
new Group([ |
1135
|
|
|
new LogicalOr(new Term($token4), new Term($token6), $token5), |
1136
|
|
|
], $token3, $token10), |
1137
|
|
|
$token2 |
1138
|
|
|
), |
1139
|
|
|
new Term($token12), |
1140
|
|
|
$token11 |
1141
|
|
|
), |
1142
|
|
|
] |
1143
|
|
|
), |
1144
|
|
|
[ |
1145
|
|
|
new Correction(Parser::CORRECTION_BINARY_OPERATOR_MISSING_LEFT_OPERAND_IGNORED, $token1), |
1146
|
|
|
new Correction(Parser::CORRECTION_BINARY_OPERATOR_FOLLOWING_OPERATOR_IGNORED, $token7, $token8), |
1147
|
|
|
new Correction(Parser::CORRECTION_UNARY_OPERATOR_MISSING_OPERAND_IGNORED, $token9), |
1148
|
|
|
new Correction(Parser::CORRECTION_UNARY_OPERATOR_MISSING_OPERAND_IGNORED, $token14), |
1149
|
|
|
new Correction(Parser::CORRECTION_BINARY_OPERATOR_MISSING_RIGHT_OPERAND_IGNORED, $token13), |
1150
|
|
|
], |
1151
|
|
|
], |
1152
|
|
|
[ |
1153
|
|
|
'+ one', |
1154
|
|
|
'one', |
1155
|
|
|
[$token1 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 0), $token2 = new WordToken('one', 2, '', 'one')], |
1156
|
|
|
new Query([new Term($token2)]), |
1157
|
|
|
[new Correction(Parser::CORRECTION_UNARY_OPERATOR_MISSING_OPERAND_IGNORED, $token1)], |
1158
|
|
|
], |
1159
|
|
|
[ |
1160
|
|
|
'! one', |
1161
|
|
|
'one', |
1162
|
|
|
[ |
1163
|
|
|
$token1 = new Token(Tokenizer::TOKEN_LOGICAL_NOT_2, '!', 0), |
1164
|
|
|
$token2 = new WordToken('one', 2, '', 'one'), |
1165
|
|
|
], |
1166
|
|
|
new Query([new Term($token2)]), |
1167
|
|
|
[new Correction(Parser::CORRECTION_UNARY_OPERATOR_MISSING_OPERAND_IGNORED, $token1)], |
1168
|
|
|
], |
1169
|
|
|
[ |
1170
|
|
|
'+++one ++two', |
1171
|
|
|
'+one +two', |
1172
|
|
|
[ |
1173
|
|
|
$token1 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 0), |
1174
|
|
|
$token2 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 1), |
1175
|
|
|
$token3 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 2), |
1176
|
|
|
$token4 = new WordToken('one', 3, '', 'one'), |
1177
|
|
|
$token5 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 7), |
1178
|
|
|
$token6 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 8), |
1179
|
|
|
$token7 = new WordToken('two', 9, '', 'two'), |
1180
|
|
|
], |
1181
|
|
|
new Query([new Mandatory(new Term($token4), $token3), new Mandatory(new Term($token7), $token6)]), |
1182
|
|
|
[ |
1183
|
|
|
new Correction(Parser::CORRECTION_ADJACENT_UNARY_OPERATOR_PRECEDING_OPERATOR_IGNORED, $token1), |
1184
|
|
|
new Correction(Parser::CORRECTION_ADJACENT_UNARY_OPERATOR_PRECEDING_OPERATOR_IGNORED, $token2), |
1185
|
|
|
new Correction(Parser::CORRECTION_ADJACENT_UNARY_OPERATOR_PRECEDING_OPERATOR_IGNORED, $token5), |
1186
|
|
|
], |
1187
|
|
|
], |
1188
|
|
|
[ |
1189
|
|
|
'+one + +AND +++ two', |
1190
|
|
|
'+one AND two', |
1191
|
|
|
[ |
1192
|
|
|
$token1 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 0), |
1193
|
|
|
$token2 = new WordToken('one', 1, '', 'one'), |
1194
|
|
|
$token3 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 5), |
1195
|
|
|
$token4 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 7), |
1196
|
|
|
$token5 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 8), |
1197
|
|
|
$token6 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 12), |
1198
|
|
|
$token7 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 13), |
1199
|
|
|
$token8 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 14), |
1200
|
|
|
$token9 = new WordToken('two', 16, '', 'two'), |
1201
|
|
|
], |
1202
|
|
|
new Query([new LogicalAnd(new Mandatory(new Term($token2), $token1), new Term($token9), $token5)]), |
1203
|
|
|
[ |
1204
|
|
|
new Correction(Parser::CORRECTION_UNARY_OPERATOR_MISSING_OPERAND_IGNORED, $token3), |
1205
|
|
|
new Correction(Parser::CORRECTION_ADJACENT_UNARY_OPERATOR_PRECEDING_OPERATOR_IGNORED, $token4), |
1206
|
|
|
new Correction(Parser::CORRECTION_ADJACENT_UNARY_OPERATOR_PRECEDING_OPERATOR_IGNORED, $token6), |
1207
|
|
|
new Correction(Parser::CORRECTION_ADJACENT_UNARY_OPERATOR_PRECEDING_OPERATOR_IGNORED, $token7), |
1208
|
|
|
new Correction(Parser::CORRECTION_UNARY_OPERATOR_MISSING_OPERAND_IGNORED, $token8), |
1209
|
|
|
], |
1210
|
|
|
], |
1211
|
|
|
[ |
1212
|
|
|
'+one + +OR++ +two ++ +', |
1213
|
|
|
'+one OR +two', |
1214
|
|
|
[ |
1215
|
|
|
$token1 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 0), |
1216
|
|
|
$token2 = new WordToken('one', 1, '', 'one'), |
1217
|
|
|
$token3 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 5), |
1218
|
|
|
$token4 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 7), |
1219
|
|
|
$token5 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 8), |
1220
|
|
|
$token6 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 10), |
1221
|
|
|
$token7 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 11), |
1222
|
|
|
$token8 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 13), |
1223
|
|
|
$token9 = new WordToken('two', 14, '', 'two'), |
1224
|
|
|
$token10 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 18), |
1225
|
|
|
$token11 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 19), |
1226
|
|
|
$token12 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 21), |
1227
|
|
|
], |
1228
|
|
|
new Query( |
1229
|
|
|
[ |
1230
|
|
|
new LogicalOr(new Mandatory(new Term($token2), $token1), new Mandatory(new Term( |
1231
|
|
|
$token9 |
1232
|
|
|
), $token8), $token5), |
1233
|
|
|
] |
1234
|
|
|
), |
1235
|
|
|
[ |
1236
|
|
|
new Correction(Parser::CORRECTION_UNARY_OPERATOR_MISSING_OPERAND_IGNORED, $token3), |
1237
|
|
|
new Correction(Parser::CORRECTION_ADJACENT_UNARY_OPERATOR_PRECEDING_OPERATOR_IGNORED, $token4), |
1238
|
|
|
new Correction(Parser::CORRECTION_ADJACENT_UNARY_OPERATOR_PRECEDING_OPERATOR_IGNORED, $token6), |
1239
|
|
|
new Correction(Parser::CORRECTION_UNARY_OPERATOR_MISSING_OPERAND_IGNORED, $token7), |
1240
|
|
|
new Correction(Parser::CORRECTION_ADJACENT_UNARY_OPERATOR_PRECEDING_OPERATOR_IGNORED, $token10), |
1241
|
|
|
new Correction(Parser::CORRECTION_UNARY_OPERATOR_MISSING_OPERAND_IGNORED, $token11), |
1242
|
|
|
new Correction(Parser::CORRECTION_UNARY_OPERATOR_MISSING_OPERAND_IGNORED, $token12), |
1243
|
|
|
], |
1244
|
|
|
], |
1245
|
|
|
[ |
1246
|
|
|
'NOT +one', |
1247
|
|
|
'+one', |
1248
|
|
|
[ |
1249
|
|
|
$token1 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 0), |
1250
|
|
|
$token2 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 4), |
1251
|
|
|
$token3 = new WordToken('one', 5, '', 'one'), |
1252
|
|
|
], |
1253
|
|
|
new Query([new Mandatory(new Term($token3), $token2)]), |
1254
|
|
|
[new Correction(Parser::CORRECTION_LOGICAL_NOT_OPERATORS_PRECEDING_PREFERENCE_IGNORED, $token1)], |
1255
|
|
|
], |
1256
|
|
|
[ |
1257
|
|
|
'+(+one + +OR++ +two ++ +)', |
1258
|
|
|
'+(+one OR +two)', |
1259
|
|
|
[ |
1260
|
|
|
$token1 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 0), |
1261
|
|
|
$token2 = new GroupBeginToken('(', 1, '(', null), |
1262
|
|
|
$token3 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 2), |
1263
|
|
|
$token4 = new WordToken('one', 3, '', 'one'), |
1264
|
|
|
$token5 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 7), |
1265
|
|
|
$token6 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 9), |
1266
|
|
|
$token7 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 10), |
1267
|
|
|
$token8 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 12), |
1268
|
|
|
$token9 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 13), |
1269
|
|
|
$token10 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 15), |
1270
|
|
|
$token11 = new WordToken('two', 16, '', 'two'), |
1271
|
|
|
$token12 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 20), |
1272
|
|
|
$token13 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 21), |
1273
|
|
|
$token14 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 23), |
1274
|
|
|
$token15 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 24), |
1275
|
|
|
], |
1276
|
|
|
new Query( |
1277
|
|
|
[ |
1278
|
|
|
new Mandatory( |
1279
|
|
|
new Group( |
1280
|
|
|
[ |
1281
|
|
|
new LogicalOr( |
1282
|
|
|
new Mandatory(new Term($token4), $token3), |
1283
|
|
|
new Mandatory(new Term($token11), $token10), |
1284
|
|
|
$token7 |
1285
|
|
|
), |
1286
|
|
|
], |
1287
|
|
|
$token2, |
1288
|
|
|
$token15 |
1289
|
|
|
), |
1290
|
|
|
$token1 |
1291
|
|
|
), |
1292
|
|
|
] |
1293
|
|
|
), |
1294
|
|
|
[ |
1295
|
|
|
new Correction(Parser::CORRECTION_UNARY_OPERATOR_MISSING_OPERAND_IGNORED, $token5), |
1296
|
|
|
new Correction(Parser::CORRECTION_ADJACENT_UNARY_OPERATOR_PRECEDING_OPERATOR_IGNORED, $token6), |
1297
|
|
|
new Correction(Parser::CORRECTION_ADJACENT_UNARY_OPERATOR_PRECEDING_OPERATOR_IGNORED, $token8), |
1298
|
|
|
new Correction(Parser::CORRECTION_UNARY_OPERATOR_MISSING_OPERAND_IGNORED, $token9), |
1299
|
|
|
new Correction(Parser::CORRECTION_ADJACENT_UNARY_OPERATOR_PRECEDING_OPERATOR_IGNORED, $token12), |
1300
|
|
|
new Correction(Parser::CORRECTION_UNARY_OPERATOR_MISSING_OPERAND_IGNORED, $token13), |
1301
|
|
|
new Correction(Parser::CORRECTION_UNARY_OPERATOR_MISSING_OPERAND_IGNORED, $token14), |
1302
|
|
|
], |
1303
|
|
|
], |
1304
|
|
|
[ |
1305
|
|
|
'- one', |
1306
|
|
|
'one', |
1307
|
|
|
[ |
1308
|
|
|
$token1 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 0), |
1309
|
|
|
$token2 = new WordToken('one', 2, '', 'one'), |
1310
|
|
|
], |
1311
|
|
|
new Query([new Term($token2)]), |
1312
|
|
|
[new Correction(Parser::CORRECTION_UNARY_OPERATOR_MISSING_OPERAND_IGNORED, $token1)], |
1313
|
|
|
], |
1314
|
|
|
[ |
1315
|
|
|
'---one --two', |
1316
|
|
|
'-one -two', |
1317
|
|
|
[ |
1318
|
|
|
$token1 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 0), |
1319
|
|
|
$token2 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 1), |
1320
|
|
|
$token3 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 2), |
1321
|
|
|
$token4 = new WordToken('one', 3, '', 'one'), |
1322
|
|
|
$token5 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 7), |
1323
|
|
|
$token6 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 8), |
1324
|
|
|
$token7 = new WordToken('two', 9, '', 'two'), |
1325
|
|
|
], |
1326
|
|
|
new Query([new Prohibited(new Term($token4), $token3), new Prohibited(new Term($token7), $token6)]), |
1327
|
|
|
[ |
1328
|
|
|
new Correction(Parser::CORRECTION_ADJACENT_UNARY_OPERATOR_PRECEDING_OPERATOR_IGNORED, $token1), |
1329
|
|
|
new Correction(Parser::CORRECTION_ADJACENT_UNARY_OPERATOR_PRECEDING_OPERATOR_IGNORED, $token2), |
1330
|
|
|
new Correction(Parser::CORRECTION_ADJACENT_UNARY_OPERATOR_PRECEDING_OPERATOR_IGNORED, $token5), |
1331
|
|
|
], |
1332
|
|
|
], |
1333
|
|
|
[ |
1334
|
|
|
'-one - -AND --- two', |
1335
|
|
|
'-one AND two', |
1336
|
|
|
[ |
1337
|
|
|
$token1 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 0), |
1338
|
|
|
$token2 = new WordToken('one', 1, '', 'one'), |
1339
|
|
|
$token3 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 5), |
1340
|
|
|
$token4 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 7), |
1341
|
|
|
$token5 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 8), |
1342
|
|
|
$token6 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 12), |
1343
|
|
|
$token7 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 13), |
1344
|
|
|
$token8 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 14), |
1345
|
|
|
$token9 = new WordToken('two', 16, '', 'two'), |
1346
|
|
|
], |
1347
|
|
|
new Query([new LogicalAnd(new Prohibited(new Term($token2), $token1), new Term($token9), $token5)]), |
1348
|
|
|
[ |
1349
|
|
|
new Correction(Parser::CORRECTION_UNARY_OPERATOR_MISSING_OPERAND_IGNORED, $token3), |
1350
|
|
|
new Correction(Parser::CORRECTION_ADJACENT_UNARY_OPERATOR_PRECEDING_OPERATOR_IGNORED, $token4), |
1351
|
|
|
new Correction(Parser::CORRECTION_ADJACENT_UNARY_OPERATOR_PRECEDING_OPERATOR_IGNORED, $token6), |
1352
|
|
|
new Correction(Parser::CORRECTION_ADJACENT_UNARY_OPERATOR_PRECEDING_OPERATOR_IGNORED, $token7), |
1353
|
|
|
new Correction(Parser::CORRECTION_UNARY_OPERATOR_MISSING_OPERAND_IGNORED, $token8), |
1354
|
|
|
], |
1355
|
|
|
], |
1356
|
|
|
[ |
1357
|
|
|
'-one - -OR-- -two -- -', |
1358
|
|
|
'-one OR -two', |
1359
|
|
|
[ |
1360
|
|
|
$token1 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 0), |
1361
|
|
|
$token2 = new WordToken('one', 1, '', 'one'), |
1362
|
|
|
$token3 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 5), |
1363
|
|
|
$token4 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 7), |
1364
|
|
|
$token5 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 8), |
1365
|
|
|
$token6 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 10), |
1366
|
|
|
$token7 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 11), |
1367
|
|
|
$token8 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 13), |
1368
|
|
|
$token9 = new WordToken('two', 14, '', 'two'), |
1369
|
|
|
$token10 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 18), |
1370
|
|
|
$token11 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 19), |
1371
|
|
|
$token12 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 21), |
1372
|
|
|
], |
1373
|
|
|
new Query( |
1374
|
|
|
[ |
1375
|
|
|
new LogicalOr( |
1376
|
|
|
new Prohibited(new Term($token2), $token1), |
1377
|
|
|
new Prohibited(new Term($token9), $token8), |
1378
|
|
|
$token5 |
1379
|
|
|
), |
1380
|
|
|
] |
1381
|
|
|
), |
1382
|
|
|
[ |
1383
|
|
|
new Correction(Parser::CORRECTION_UNARY_OPERATOR_MISSING_OPERAND_IGNORED, $token3), |
1384
|
|
|
new Correction(Parser::CORRECTION_ADJACENT_UNARY_OPERATOR_PRECEDING_OPERATOR_IGNORED, $token4), |
1385
|
|
|
new Correction(Parser::CORRECTION_ADJACENT_UNARY_OPERATOR_PRECEDING_OPERATOR_IGNORED, $token6), |
1386
|
|
|
new Correction(Parser::CORRECTION_UNARY_OPERATOR_MISSING_OPERAND_IGNORED, $token7), |
1387
|
|
|
new Correction(Parser::CORRECTION_ADJACENT_UNARY_OPERATOR_PRECEDING_OPERATOR_IGNORED, $token10), |
1388
|
|
|
new Correction(Parser::CORRECTION_UNARY_OPERATOR_MISSING_OPERAND_IGNORED, $token11), |
1389
|
|
|
new Correction(Parser::CORRECTION_UNARY_OPERATOR_MISSING_OPERAND_IGNORED, $token12), |
1390
|
|
|
], |
1391
|
|
|
], |
1392
|
|
|
[ |
1393
|
|
|
'NOT -one', |
1394
|
|
|
'-one', |
1395
|
|
|
[ |
1396
|
|
|
$token1 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 0), |
1397
|
|
|
$token2 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 4), |
1398
|
|
|
$token3 = new WordToken('one', 5, '', 'one'), |
1399
|
|
|
], |
1400
|
|
|
new Query([new Prohibited(new Term($token3), $token2)]), |
1401
|
|
|
[new Correction(Parser::CORRECTION_LOGICAL_NOT_OPERATORS_PRECEDING_PREFERENCE_IGNORED, $token1)], |
1402
|
|
|
], |
1403
|
|
|
[ |
1404
|
|
|
'-(-one - -OR-- -two --)-', |
1405
|
|
|
'-(-one OR -two)', |
1406
|
|
|
[ |
1407
|
|
|
$token1 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 0), |
1408
|
|
|
$token2 = new GroupBeginToken('(', 1, '(', null), |
1409
|
|
|
$token3 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 2), |
1410
|
|
|
$token4 = new WordToken('one', 3, '', 'one'), |
1411
|
|
|
$token5 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 7), |
1412
|
|
|
$token6 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 9), |
1413
|
|
|
$token7 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 10), |
1414
|
|
|
$token8 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 12), |
1415
|
|
|
$token9 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 13), |
1416
|
|
|
$token10 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 15), |
1417
|
|
|
$token11 = new WordToken('two', 16, '', 'two'), |
1418
|
|
|
$token12 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 20), |
1419
|
|
|
$token13 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 21), |
1420
|
|
|
$token15 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 22), |
1421
|
|
|
$token14 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 23), |
1422
|
|
|
], |
1423
|
|
|
new Query( |
1424
|
|
|
[ |
1425
|
|
|
new Prohibited( |
1426
|
|
|
new Group( |
1427
|
|
|
[ |
1428
|
|
|
new LogicalOr( |
1429
|
|
|
new Prohibited(new Term($token4), $token3), |
1430
|
|
|
new Prohibited(new Term($token11), $token10), |
1431
|
|
|
$token7 |
1432
|
|
|
), |
1433
|
|
|
], |
1434
|
|
|
$token2, |
1435
|
|
|
$token15 |
1436
|
|
|
), |
1437
|
|
|
$token1 |
1438
|
|
|
), |
1439
|
|
|
] |
1440
|
|
|
), |
1441
|
|
|
[ |
1442
|
|
|
new Correction(Parser::CORRECTION_UNARY_OPERATOR_MISSING_OPERAND_IGNORED, $token5), |
1443
|
|
|
new Correction(Parser::CORRECTION_ADJACENT_UNARY_OPERATOR_PRECEDING_OPERATOR_IGNORED, $token6), |
1444
|
|
|
new Correction(Parser::CORRECTION_ADJACENT_UNARY_OPERATOR_PRECEDING_OPERATOR_IGNORED, $token8), |
1445
|
|
|
new Correction(Parser::CORRECTION_UNARY_OPERATOR_MISSING_OPERAND_IGNORED, $token9), |
1446
|
|
|
new Correction(Parser::CORRECTION_ADJACENT_UNARY_OPERATOR_PRECEDING_OPERATOR_IGNORED, $token12), |
1447
|
|
|
new Correction(Parser::CORRECTION_UNARY_OPERATOR_MISSING_OPERAND_IGNORED, $token13), |
1448
|
|
|
new Correction(Parser::CORRECTION_UNARY_OPERATOR_MISSING_OPERAND_IGNORED, $token14), |
1449
|
|
|
], |
1450
|
|
|
], |
1451
|
|
|
[ |
1452
|
|
|
'+NOT one', |
1453
|
|
|
'NOT one', |
1454
|
|
|
[ |
1455
|
|
|
$token1 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 0), |
1456
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 1), |
1457
|
|
|
$token3 = new WordToken('one', 5, '', 'one'), |
1458
|
|
|
], |
1459
|
|
|
new Query([new LogicalNot(new Term($token3), $token2)]), |
1460
|
|
|
[new Correction(Parser::CORRECTION_ADJACENT_UNARY_OPERATOR_PRECEDING_OPERATOR_IGNORED, $token1)], |
1461
|
|
|
], |
1462
|
|
|
[ |
1463
|
|
|
'+AND one', |
1464
|
|
|
'one', |
1465
|
|
|
[ |
1466
|
|
|
$token1 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 0), |
1467
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 1), |
1468
|
|
|
$token3 = new WordToken('one', 5, '', 'one'), |
1469
|
|
|
], |
1470
|
|
|
new Query([new Term($token3)]), |
1471
|
|
|
[ |
1472
|
|
|
new Correction(Parser::CORRECTION_ADJACENT_UNARY_OPERATOR_PRECEDING_OPERATOR_IGNORED, $token1), |
1473
|
|
|
new Correction(Parser::CORRECTION_BINARY_OPERATOR_MISSING_LEFT_OPERAND_IGNORED, $token2), |
1474
|
|
|
], |
1475
|
|
|
], |
1476
|
|
|
[ |
1477
|
|
|
'+OR one', |
1478
|
|
|
'one', |
1479
|
|
|
[ |
1480
|
|
|
$token1 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 0), |
1481
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 1), |
1482
|
|
|
$token3 = new WordToken('one', 4, '', 'one'), |
1483
|
|
|
], |
1484
|
|
|
new Query([new Term($token3)]), |
1485
|
|
|
[ |
1486
|
|
|
new Correction(Parser::CORRECTION_ADJACENT_UNARY_OPERATOR_PRECEDING_OPERATOR_IGNORED, $token1), |
1487
|
|
|
new Correction(Parser::CORRECTION_BINARY_OPERATOR_MISSING_LEFT_OPERAND_IGNORED, $token2), |
1488
|
|
|
], |
1489
|
|
|
], |
1490
|
|
|
[ |
1491
|
|
|
'-NOT one', |
1492
|
|
|
'NOT one', |
1493
|
|
|
[ |
1494
|
|
|
$token1 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 0), |
1495
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 1), |
1496
|
|
|
$token3 = new WordToken('one', 5, '', 'one'), |
1497
|
|
|
], |
1498
|
|
|
new Query([new LogicalNot(new Term($token3), $token2)]), |
1499
|
|
|
[new Correction(Parser::CORRECTION_ADJACENT_UNARY_OPERATOR_PRECEDING_OPERATOR_IGNORED, $token1)], |
1500
|
|
|
], |
1501
|
|
|
[ |
1502
|
|
|
'-AND one', |
1503
|
|
|
'one', |
1504
|
|
|
[ |
1505
|
|
|
$token1 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 0), |
1506
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 1), |
1507
|
|
|
$token3 = new WordToken('one', 5, '', 'one'), |
1508
|
|
|
], |
1509
|
|
|
new Query([new Term($token3)]), |
1510
|
|
|
[ |
1511
|
|
|
new Correction(Parser::CORRECTION_ADJACENT_UNARY_OPERATOR_PRECEDING_OPERATOR_IGNORED, $token1), |
1512
|
|
|
new Correction(Parser::CORRECTION_BINARY_OPERATOR_MISSING_LEFT_OPERAND_IGNORED, $token2), |
1513
|
|
|
], |
1514
|
|
|
], |
1515
|
|
|
[ |
1516
|
|
|
'-OR one', |
1517
|
|
|
'one', |
1518
|
|
|
[ |
1519
|
|
|
$token1 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 0), |
1520
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 1), |
1521
|
|
|
$token3 = new WordToken('one', 4, '', 'one'), |
1522
|
|
|
], |
1523
|
|
|
new Query([new Term($token3)]), |
1524
|
|
|
[ |
1525
|
|
|
new Correction(Parser::CORRECTION_ADJACENT_UNARY_OPERATOR_PRECEDING_OPERATOR_IGNORED, $token1), |
1526
|
|
|
new Correction(Parser::CORRECTION_BINARY_OPERATOR_MISSING_LEFT_OPERAND_IGNORED, $token2), |
1527
|
|
|
], |
1528
|
|
|
], |
1529
|
|
|
[ |
1530
|
|
|
'NOT (one', |
1531
|
|
|
'NOT one', |
1532
|
|
|
[ |
1533
|
|
|
$token1 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 0), |
1534
|
|
|
$token2 = new GroupBeginToken('(', 4, '(', null), |
1535
|
|
|
$token3 = new WordToken('one', 5, '', 'one'), |
1536
|
|
|
], |
1537
|
|
|
new Query([new LogicalNot(new Term($token3), $token1)]), |
1538
|
|
|
[new Correction(Parser::CORRECTION_UNMATCHED_GROUP_LEFT_DELIMITER_IGNORED, $token2)], |
1539
|
|
|
], |
1540
|
|
|
[ |
1541
|
|
|
'NOT (one two', |
1542
|
|
|
'NOT one two', |
1543
|
|
|
[ |
1544
|
|
|
$token1 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 0), |
1545
|
|
|
$token2 = new GroupBeginToken('(', 4, '(', null), |
1546
|
|
|
$token3 = new WordToken('one', 5, '', 'one'), |
1547
|
|
|
$token4 = new WordToken('two', 9, '', 'two'), |
1548
|
|
|
], |
1549
|
|
|
new Query([new LogicalNot(new Term($token3), $token1), new Term($token4)]), |
1550
|
|
|
[new Correction(Parser::CORRECTION_UNMATCHED_GROUP_LEFT_DELIMITER_IGNORED, $token2)], |
1551
|
|
|
], |
1552
|
|
|
[ |
1553
|
|
|
'-(one', |
1554
|
|
|
'-one', |
1555
|
|
|
[ |
1556
|
|
|
$token1 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 0), |
1557
|
|
|
$token2 = new GroupBeginToken('(', 1, '(', null), |
1558
|
|
|
$token3 = new WordToken('one', 2, '', 'one'), |
1559
|
|
|
], |
1560
|
|
|
new Query([new Prohibited(new Term($token3), $token1)]), |
1561
|
|
|
[new Correction(Parser::CORRECTION_UNMATCHED_GROUP_LEFT_DELIMITER_IGNORED, $token2)], |
1562
|
|
|
], |
1563
|
|
|
[ |
1564
|
|
|
'-(one two', |
1565
|
|
|
'-one two', |
1566
|
|
|
[ |
1567
|
|
|
$token1 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 0), |
1568
|
|
|
$token2 = new GroupBeginToken('(', 1, '(', null), |
1569
|
|
|
$token3 = new WordToken('one', 2, '', 'one'), |
1570
|
|
|
$token4 = new WordToken('two', 6, '', 'two'), |
1571
|
|
|
], |
1572
|
|
|
new Query([new Prohibited(new Term($token3), $token1), new Term($token4)]), |
1573
|
|
|
[new Correction(Parser::CORRECTION_UNMATCHED_GROUP_LEFT_DELIMITER_IGNORED, $token2)], |
1574
|
|
|
], |
1575
|
|
|
[ |
1576
|
|
|
'+(one', |
1577
|
|
|
'+one', |
1578
|
|
|
[ |
1579
|
|
|
$token1 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 0), |
1580
|
|
|
$token2 = new GroupBeginToken('(', 1, '(', null), |
1581
|
|
|
$token3 = new WordToken('one', 2, '', 'one'), |
1582
|
|
|
], |
1583
|
|
|
new Query([new Mandatory(new Term($token3), $token1)]), |
1584
|
|
|
[new Correction(Parser::CORRECTION_UNMATCHED_GROUP_LEFT_DELIMITER_IGNORED, $token2)], |
1585
|
|
|
], |
1586
|
|
|
[ |
1587
|
|
|
'+(one two', |
1588
|
|
|
'+one two', |
1589
|
|
|
[ |
1590
|
|
|
$token1 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 0), |
1591
|
|
|
$token2 = new GroupBeginToken('(', 1, '(', null), |
1592
|
|
|
$token3 = new WordToken('one', 2, '', 'one'), |
1593
|
|
|
$token4 = new WordToken('two', 6, '', 'two'), |
1594
|
|
|
], |
1595
|
|
|
new Query([new Mandatory(new Term($token3), $token1), new Term($token4)]), |
1596
|
|
|
[new Correction(Parser::CORRECTION_UNMATCHED_GROUP_LEFT_DELIMITER_IGNORED, $token2)], |
1597
|
|
|
], |
1598
|
|
|
[ |
1599
|
|
|
'-(one +(two NOT (three', |
1600
|
|
|
'-one +two NOT three', |
1601
|
|
|
[ |
1602
|
|
|
$token1 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 0), |
1603
|
|
|
$token2 = new GroupBeginToken('(', 1, '(', null), |
1604
|
|
|
$token3 = new WordToken('one', 2, '', 'one'), |
1605
|
|
|
$token4 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 6), |
1606
|
|
|
$token5 = new GroupBeginToken('(', 7, '(', null), |
1607
|
|
|
$token6 = new WordToken('two', 8, '', 'two'), |
1608
|
|
|
$token7 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 12), |
1609
|
|
|
$token8 = new GroupBeginToken('(', 16, '(', null), |
1610
|
|
|
$token9 = new WordToken('three', 17, '', 'three'), |
1611
|
|
|
], |
1612
|
|
|
new Query( |
1613
|
|
|
[ |
1614
|
|
|
new Prohibited(new Term($token3), $token1), |
1615
|
|
|
new Mandatory(new Term($token6), $token4), |
1616
|
|
|
new LogicalNot(new Term($token9), $token7), |
1617
|
|
|
] |
1618
|
|
|
), |
1619
|
|
|
[ |
1620
|
|
|
new Correction(Parser::CORRECTION_UNMATCHED_GROUP_LEFT_DELIMITER_IGNORED, $token8), |
1621
|
|
|
new Correction(Parser::CORRECTION_UNMATCHED_GROUP_LEFT_DELIMITER_IGNORED, $token5), |
1622
|
|
|
new Correction(Parser::CORRECTION_UNMATCHED_GROUP_LEFT_DELIMITER_IGNORED, $token2), |
1623
|
|
|
], |
1624
|
|
|
], |
1625
|
|
|
[ |
1626
|
|
|
'one AND NOT (two', |
1627
|
|
|
'one AND NOT two', |
1628
|
|
|
[ |
1629
|
|
|
$token1 = new WordToken('one', 0, '', 'one'), |
1630
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 4), |
1631
|
|
|
$token3 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 8), |
1632
|
|
|
$token4 = new GroupBeginToken('(', 12, '(', null), |
1633
|
|
|
$token5 = new WordToken('two', 13, '', 'two'), |
1634
|
|
|
], |
1635
|
|
|
new Query([new LogicalAnd(new Term($token1), new LogicalNot(new Term($token5), $token3), $token2)]), |
1636
|
|
|
[new Correction(Parser::CORRECTION_UNMATCHED_GROUP_LEFT_DELIMITER_IGNORED, $token4)], |
1637
|
|
|
], |
1638
|
|
|
[ |
1639
|
|
|
'(one OR two AND) AND', |
1640
|
|
|
'(one OR two)', |
1641
|
|
|
[ |
1642
|
|
|
$token1 = new GroupBeginToken('(', 0, '(', null), |
1643
|
|
|
$token2 = new WordToken('one', 1, '', 'one'), |
1644
|
|
|
$token3 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 5), |
1645
|
|
|
$token4 = new WordToken('two', 8, '', 'two'), |
1646
|
|
|
$token5 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 12), |
1647
|
|
|
$token6 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 15), |
1648
|
|
|
$token7 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 17), |
1649
|
|
|
], |
1650
|
|
|
new Query([ |
1651
|
|
|
new Group([new LogicalOr(new Term($token2), new Term($token4), $token3)], $token1, $token6), |
1652
|
|
|
]), |
1653
|
|
|
[ |
1654
|
|
|
new Correction(Parser::CORRECTION_BINARY_OPERATOR_MISSING_RIGHT_OPERAND_IGNORED, $token5), |
1655
|
|
|
new Correction(Parser::CORRECTION_BINARY_OPERATOR_MISSING_RIGHT_OPERAND_IGNORED, $token7), |
1656
|
|
|
], |
1657
|
|
|
], |
1658
|
|
|
[ |
1659
|
|
|
'(one AND NOT +two)', |
1660
|
|
|
'(one AND +two)', |
1661
|
|
|
[ |
1662
|
|
|
$token1 = new GroupBeginToken('(', 0, '(', null), |
1663
|
|
|
$token2 = new WordToken('one', 1, '', 'one'), |
1664
|
|
|
$token3 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 5), |
1665
|
|
|
$token4 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 9), |
1666
|
|
|
$token5 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 13), |
1667
|
|
|
$token6 = new WordToken('two', 14, '', 'two'), |
1668
|
|
|
$token7 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 17), |
1669
|
|
|
], |
1670
|
|
|
new Query( |
1671
|
|
|
[ |
1672
|
|
|
new Group( |
1673
|
|
|
[new LogicalAnd(new Term($token2), new Mandatory(new Term($token6), $token5), $token3)], |
1674
|
|
|
$token1, |
1675
|
|
|
$token7 |
1676
|
|
|
), |
1677
|
|
|
] |
1678
|
|
|
), |
1679
|
|
|
[new Correction(Parser::CORRECTION_LOGICAL_NOT_OPERATORS_PRECEDING_PREFERENCE_IGNORED, $token4)], |
1680
|
|
|
], |
1681
|
|
|
[ |
1682
|
|
|
'(one AND NOT -two)', |
1683
|
|
|
'(one AND -two)', |
1684
|
|
|
[ |
1685
|
|
|
$token1 = new GroupBeginToken('(', 0, '(', null), |
1686
|
|
|
$token2 = new WordToken('one', 1, '', 'one'), |
1687
|
|
|
$token3 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 5), |
1688
|
|
|
$token4 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 9), |
1689
|
|
|
$token5 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 13), |
1690
|
|
|
$token6 = new WordToken('two', 14, '', 'two'), |
1691
|
|
|
$token7 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 17), |
1692
|
|
|
], |
1693
|
|
|
new Query( |
1694
|
|
|
[ |
1695
|
|
|
new Group( |
1696
|
|
|
[new LogicalAnd(new Term($token2), new Prohibited(new Term($token6), $token5), $token3)], |
1697
|
|
|
$token1, |
1698
|
|
|
$token7 |
1699
|
|
|
), |
1700
|
|
|
] |
1701
|
|
|
), |
1702
|
|
|
[new Correction(Parser::CORRECTION_LOGICAL_NOT_OPERATORS_PRECEDING_PREFERENCE_IGNORED, $token4)], |
1703
|
|
|
], |
1704
|
|
|
[ |
1705
|
|
|
'(one AND NOT -two three)', |
1706
|
|
|
'(one AND -two three)', |
1707
|
|
|
[ |
1708
|
|
|
$token1 = new GroupBeginToken('(', 0, '(', null), |
1709
|
|
|
$token2 = new WordToken('one', 1, '', 'one'), |
1710
|
|
|
$token3 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 5), |
1711
|
|
|
$token4 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 9), |
1712
|
|
|
$token5 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 13), |
1713
|
|
|
$token6 = new WordToken('two', 14, '', 'two'), |
1714
|
|
|
$token7 = new WordToken('three', 18, '', 'three'), |
1715
|
|
|
$token8 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 23), |
1716
|
|
|
], |
1717
|
|
|
new Query( |
1718
|
|
|
[ |
1719
|
|
|
new Group( |
1720
|
|
|
[ |
1721
|
|
|
new LogicalAnd(new Term($token2), new Prohibited(new Term($token6), $token5), $token3), |
1722
|
|
|
new Term($token7), |
1723
|
|
|
], |
1724
|
|
|
$token1, |
1725
|
|
|
$token8 |
1726
|
|
|
), |
1727
|
|
|
] |
1728
|
|
|
), |
1729
|
|
|
[new Correction(Parser::CORRECTION_LOGICAL_NOT_OPERATORS_PRECEDING_PREFERENCE_IGNORED, $token4)], |
1730
|
|
|
], |
1731
|
|
|
[ |
1732
|
|
|
'(one AND NOT +two three)', |
1733
|
|
|
'(one AND +two three)', |
1734
|
|
|
[ |
1735
|
|
|
$token1 = new GroupBeginToken('(', 0, '(', null), |
1736
|
|
|
$token2 = new WordToken('one', 1, '', 'one'), |
1737
|
|
|
$token3 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 5), |
1738
|
|
|
$token4 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 9), |
1739
|
|
|
$token5 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 13), |
1740
|
|
|
$token6 = new WordToken('two', 14, '', 'two'), |
1741
|
|
|
$token7 = new WordToken('three', 18, '', 'three'), |
1742
|
|
|
$token8 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 23), |
1743
|
|
|
], |
1744
|
|
|
new Query( |
1745
|
|
|
[ |
1746
|
|
|
new Group( |
1747
|
|
|
[ |
1748
|
|
|
new LogicalAnd(new Term($token2), new Mandatory(new Term($token6), $token5), $token3), |
1749
|
|
|
new Term($token7), |
1750
|
|
|
], |
1751
|
|
|
$token1, |
1752
|
|
|
$token8 |
1753
|
|
|
), |
1754
|
|
|
] |
1755
|
|
|
), |
1756
|
|
|
[new Correction(Parser::CORRECTION_LOGICAL_NOT_OPERATORS_PRECEDING_PREFERENCE_IGNORED, $token4)], |
1757
|
|
|
], |
1758
|
|
|
[ |
1759
|
|
|
'+()+one', |
1760
|
|
|
'+one', |
1761
|
|
|
[ |
1762
|
|
|
$token1 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 0), |
1763
|
|
|
$token2 = new GroupBeginToken('(', 1, '(', null), |
1764
|
|
|
$token3 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 2), |
1765
|
|
|
$token4 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 3), |
1766
|
|
|
$token5 = new WordToken('one', 4, '', 'one'), |
1767
|
|
|
], |
1768
|
|
|
new Query([new Mandatory(new Term($token5), $token4)]), |
1769
|
|
|
[new Correction(Parser::CORRECTION_EMPTY_GROUP_IGNORED, $token1, $token2, $token3)], |
1770
|
|
|
], |
1771
|
|
|
[ |
1772
|
|
|
'+()!one', |
1773
|
|
|
'!one', |
1774
|
|
|
[ |
1775
|
|
|
$token1 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 0), |
1776
|
|
|
$token2 = new GroupBeginToken('(', 1, '(', null), |
1777
|
|
|
$token3 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 2), |
1778
|
|
|
$token4 = new Token(Tokenizer::TOKEN_LOGICAL_NOT_2, '!', 3), |
1779
|
|
|
$token5 = new WordToken('one', 4, '', 'one'), |
1780
|
|
|
], |
1781
|
|
|
new Query([new LogicalNot(new Term($token5), $token4)]), |
1782
|
|
|
[new Correction(Parser::CORRECTION_EMPTY_GROUP_IGNORED, $token1, $token2, $token3)], |
1783
|
|
|
], |
1784
|
|
|
[ |
1785
|
|
|
'one AND +()!two', |
1786
|
|
|
'one !two', |
1787
|
|
|
[ |
1788
|
|
|
$token1 = new WordToken('one', 0, '', 'one'), |
1789
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 4), |
1790
|
|
|
$token3 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 8), |
1791
|
|
|
$token4 = new GroupBeginToken('(', 9, '(', null), |
1792
|
|
|
$token5 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 10), |
1793
|
|
|
$token6 = new Token(Tokenizer::TOKEN_LOGICAL_NOT_2, '!', 11), |
1794
|
|
|
$token7 = new WordToken('two', 12, '', 'two'), |
1795
|
|
|
], |
1796
|
|
|
new Query([new Term($token1), new LogicalNot(new Term($token7), $token6)]), |
1797
|
|
|
[new Correction(Parser::CORRECTION_EMPTY_GROUP_IGNORED, $token2, $token3, $token4, $token5)], |
1798
|
|
|
], |
1799
|
|
|
[ |
1800
|
|
|
'NOT +()!one', |
1801
|
|
|
'!one', |
1802
|
|
|
[ |
1803
|
|
|
$token1 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 0), |
1804
|
|
|
$token2 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 4), |
1805
|
|
|
$token3 = new GroupBeginToken('(', 5, '(', null), |
1806
|
|
|
$token4 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 6), |
1807
|
|
|
$token5 = new Token(Tokenizer::TOKEN_LOGICAL_NOT_2, '!', 7), |
1808
|
|
|
$token6 = new WordToken('one', 8, '', 'one'), |
1809
|
|
|
], |
1810
|
|
|
new Query([new LogicalNot(new Term($token6), $token5)]), |
1811
|
|
|
[new Correction(Parser::CORRECTION_EMPTY_GROUP_IGNORED, $token1, $token2, $token3, $token4)], |
1812
|
|
|
], |
1813
|
|
|
[ |
1814
|
|
|
'NOT -()!one', |
1815
|
|
|
'!one', |
1816
|
|
|
[ |
1817
|
|
|
$token1 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 0), |
1818
|
|
|
$token2 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 4), |
1819
|
|
|
$token3 = new GroupBeginToken('(', 5, '(', null), |
1820
|
|
|
$token4 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 6), |
1821
|
|
|
$token5 = new Token(Tokenizer::TOKEN_LOGICAL_NOT_2, '!', 7), |
1822
|
|
|
$token6 = new WordToken('one', 8, '', 'one'), |
1823
|
|
|
], |
1824
|
|
|
new Query([new LogicalNot(new Term($token6), $token5)]), |
1825
|
|
|
[new Correction(Parser::CORRECTION_EMPTY_GROUP_IGNORED, $token1, $token2, $token3, $token4)], |
1826
|
|
|
], |
1827
|
|
|
[ |
1828
|
|
|
'NOT ++()!one', |
1829
|
|
|
'!one', |
1830
|
|
|
[ |
1831
|
|
|
$token1 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 0), |
1832
|
|
|
$token2 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 4), |
1833
|
|
|
$token3 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 5), |
1834
|
|
|
$token4 = new GroupBeginToken('(', 6, '(', null), |
1835
|
|
|
$token5 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 7), |
1836
|
|
|
$token6 = new Token(Tokenizer::TOKEN_LOGICAL_NOT_2, '!', 8), |
1837
|
|
|
$token7 = new WordToken('one', 9, '', 'one'), |
1838
|
|
|
], |
1839
|
|
|
new Query([new LogicalNot(new Term($token7), $token6)]), |
1840
|
|
|
[ |
1841
|
|
|
new Correction(Parser::CORRECTION_ADJACENT_UNARY_OPERATOR_PRECEDING_OPERATOR_IGNORED, $token2), |
1842
|
|
|
new Correction(Parser::CORRECTION_EMPTY_GROUP_IGNORED, $token1, $token3, $token4, $token5), |
1843
|
|
|
], |
1844
|
|
|
], |
1845
|
|
|
[ |
1846
|
|
|
'NOT -+()!one', |
1847
|
|
|
'!one', |
1848
|
|
|
[ |
1849
|
|
|
$token1 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 0), |
1850
|
|
|
$token2 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 4), |
1851
|
|
|
$token3 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 5), |
1852
|
|
|
$token4 = new GroupBeginToken('(', 6, '(', null), |
1853
|
|
|
$token5 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 7), |
1854
|
|
|
$token6 = new Token(Tokenizer::TOKEN_LOGICAL_NOT_2, '!', 8), |
1855
|
|
|
$token7 = new WordToken('one', 9, '', 'one'), |
1856
|
|
|
], |
1857
|
|
|
new Query([new LogicalNot(new Term($token7), $token6)]), |
1858
|
|
|
[ |
1859
|
|
|
new Correction(Parser::CORRECTION_ADJACENT_UNARY_OPERATOR_PRECEDING_OPERATOR_IGNORED, $token2), |
1860
|
|
|
new Correction(Parser::CORRECTION_EMPTY_GROUP_IGNORED, $token1, $token3, $token4, $token5), |
1861
|
|
|
], |
1862
|
|
|
], |
1863
|
|
|
[ |
1864
|
|
|
'NOT !()!one', |
1865
|
|
|
'!one', |
1866
|
|
|
[ |
1867
|
|
|
$token1 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 0), |
1868
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_NOT_2, '!', 4), |
1869
|
|
|
$token3 = new GroupBeginToken('(', 5, '(', null), |
1870
|
|
|
$token4 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 6), |
1871
|
|
|
$token5 = new Token(Tokenizer::TOKEN_LOGICAL_NOT_2, '!', 7), |
1872
|
|
|
$token6 = new WordToken('one', 8, '', 'one'), |
1873
|
|
|
], |
1874
|
|
|
new Query([new LogicalNot(new Term($token6), $token5)]), |
1875
|
|
|
[new Correction(Parser::CORRECTION_EMPTY_GROUP_IGNORED, $token1, $token2, $token3, $token4)], |
1876
|
|
|
], |
1877
|
|
|
[ |
1878
|
|
|
'NOT +()+()!one', |
1879
|
|
|
'!one', |
1880
|
|
|
[ |
1881
|
|
|
$token1 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 0), |
1882
|
|
|
$token2 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 4), |
1883
|
|
|
$token3 = new GroupBeginToken('(', 5, '(', null), |
1884
|
|
|
$token4 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 6), |
1885
|
|
|
$token5 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 7), |
1886
|
|
|
$token6 = new GroupBeginToken('(', 8, '(', null), |
1887
|
|
|
$token7 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 9), |
1888
|
|
|
$token8 = new Token(Tokenizer::TOKEN_LOGICAL_NOT_2, '!', 10), |
1889
|
|
|
$token9 = new WordToken('one', 11, '', 'one'), |
1890
|
|
|
], |
1891
|
|
|
new Query([new LogicalNot(new Term($token9), $token8)]), |
1892
|
|
|
[ |
1893
|
|
|
new Correction(Parser::CORRECTION_EMPTY_GROUP_IGNORED, $token1, $token2, $token3, $token4), |
1894
|
|
|
new Correction(Parser::CORRECTION_EMPTY_GROUP_IGNORED, $token5, $token6, $token7), |
1895
|
|
|
], |
1896
|
|
|
], |
1897
|
|
|
[ |
1898
|
|
|
'NOT NOT +()+()!one', |
1899
|
|
|
'!one', |
1900
|
|
|
[ |
1901
|
|
|
$token1 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 0), |
1902
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 4), |
1903
|
|
|
$token3 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 8), |
1904
|
|
|
$token4 = new GroupBeginToken('(', 9, '(', null), |
1905
|
|
|
$token5 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 10), |
1906
|
|
|
$token6 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 11), |
1907
|
|
|
$token7 = new GroupBeginToken('(', 12, '(', null), |
1908
|
|
|
$token8 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 13), |
1909
|
|
|
$token9 = new Token(Tokenizer::TOKEN_LOGICAL_NOT_2, '!', 14), |
1910
|
|
|
$token10 = new WordToken('one', 15, '', 'one'), |
1911
|
|
|
], |
1912
|
|
|
new Query([new LogicalNot(new Term($token10), $token9)]), |
1913
|
|
|
[ |
1914
|
|
|
new Correction(Parser::CORRECTION_EMPTY_GROUP_IGNORED, $token1, $token2, $token3, $token4, $token5), |
1915
|
|
|
new Correction(Parser::CORRECTION_EMPTY_GROUP_IGNORED, $token6, $token7, $token8), |
1916
|
|
|
], |
1917
|
|
|
], |
1918
|
|
|
[ |
1919
|
|
|
'one AND NOT +()+()!two', |
1920
|
|
|
'one !two', |
1921
|
|
|
[ |
1922
|
|
|
$token1 = new WordToken('one', 0, '', 'one'), |
1923
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 4), |
1924
|
|
|
$token3 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 8), |
1925
|
|
|
$token4 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 12), |
1926
|
|
|
$token5 = new GroupBeginToken('(', 13, '(', null), |
1927
|
|
|
$token6 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 14), |
1928
|
|
|
$token7 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 15), |
1929
|
|
|
$token8 = new GroupBeginToken('(', 16, '(', null), |
1930
|
|
|
$token9 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 17), |
1931
|
|
|
$token10 = new Token(Tokenizer::TOKEN_LOGICAL_NOT_2, '!', 18), |
1932
|
|
|
$token11 = new WordToken('two', 19, '', 'two'), |
1933
|
|
|
], |
1934
|
|
|
new Query([new Term($token1), new LogicalNot(new Term($token11), $token10)]), |
1935
|
|
|
[ |
1936
|
|
|
new Correction(Parser::CORRECTION_EMPTY_GROUP_IGNORED, $token2, $token3, $token4, $token5, $token6), |
1937
|
|
|
new Correction(Parser::CORRECTION_EMPTY_GROUP_IGNORED, $token7, $token8, $token9), |
1938
|
|
|
], |
1939
|
|
|
], |
1940
|
|
|
[ |
1941
|
|
|
'one AND NOT NOT +()+()!two', |
1942
|
|
|
'one !two', |
1943
|
|
|
[ |
1944
|
|
|
$token1 = new WordToken('one', 0, '', 'one'), |
1945
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 4), |
1946
|
|
|
$token3 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 8), |
1947
|
|
|
$token4 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 12), |
1948
|
|
|
$token5 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 16), |
1949
|
|
|
$token6 = new GroupBeginToken('(', 17, '(', null), |
1950
|
|
|
$token7 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 18), |
1951
|
|
|
$token8 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 19), |
1952
|
|
|
$token9 = new GroupBeginToken('(', 20, '(', null), |
1953
|
|
|
$token10 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 21), |
1954
|
|
|
$token11 = new Token(Tokenizer::TOKEN_LOGICAL_NOT_2, '!', 22), |
1955
|
|
|
$token12 = new WordToken('two', 23, '', 'two'), |
1956
|
|
|
], |
1957
|
|
|
new Query([new Term($token1), new LogicalNot(new Term($token12), $token11)]), |
1958
|
|
|
[ |
1959
|
|
|
new Correction( |
1960
|
|
|
Parser::CORRECTION_EMPTY_GROUP_IGNORED, |
1961
|
|
|
$token2, |
1962
|
|
|
$token3, |
1963
|
|
|
$token4, |
1964
|
|
|
$token5, |
1965
|
|
|
$token6, |
1966
|
|
|
$token7 |
1967
|
|
|
), |
1968
|
|
|
new Correction(Parser::CORRECTION_EMPTY_GROUP_IGNORED, $token8, $token9, $token10), |
1969
|
|
|
], |
1970
|
|
|
], |
1971
|
|
|
[ |
1972
|
|
|
'one -() +() two', |
1973
|
|
|
'one two', |
1974
|
|
|
[ |
1975
|
|
|
$token1 = new WordToken('one', 0, '', 'one'), |
1976
|
|
|
$token2 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 4), |
1977
|
|
|
$token3 = new GroupBeginToken('(', 5, '(', null), |
1978
|
|
|
$token4 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 6), |
1979
|
|
|
$token5 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 8), |
1980
|
|
|
$token6 = new GroupBeginToken('(', 9, '(', null), |
1981
|
|
|
$token7 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 10), |
1982
|
|
|
$token8 = new WordToken('two', 12, '', 'two'), |
1983
|
|
|
], |
1984
|
|
|
new Query([new Term($token1), new Term($token8)]), |
1985
|
|
|
[ |
1986
|
|
|
new Correction(Parser::CORRECTION_EMPTY_GROUP_IGNORED, $token2, $token3, $token4), |
1987
|
|
|
new Correction(Parser::CORRECTION_EMPTY_GROUP_IGNORED, $token5, $token6, $token7), |
1988
|
|
|
], |
1989
|
|
|
], |
1990
|
|
|
[ |
1991
|
|
|
'one !+ two', |
1992
|
|
|
'one two', |
1993
|
|
|
[ |
1994
|
|
|
$token1 = new WordToken('one', 0, '', 'one'), |
1995
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_NOT_2, '!', 4), |
1996
|
|
|
$token3 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 5), |
1997
|
|
|
$token4 = new WordToken('two', 7, '', 'two'), |
1998
|
|
|
], |
1999
|
|
|
new Query([new Term($token1), new Term($token4)]), |
2000
|
|
|
[ |
2001
|
|
|
new Correction(Parser::CORRECTION_ADJACENT_UNARY_OPERATOR_PRECEDING_OPERATOR_IGNORED, $token2), |
2002
|
|
|
new Correction(Parser::CORRECTION_UNARY_OPERATOR_MISSING_OPERAND_IGNORED, $token3), |
2003
|
|
|
], |
2004
|
|
|
], |
2005
|
|
|
[ |
2006
|
|
|
'one +! two', |
2007
|
|
|
'one two', |
2008
|
|
|
[ |
2009
|
|
|
$token1 = new WordToken('one', 0, '', 'one'), |
2010
|
|
|
$token3 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 4), |
2011
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_NOT_2, '!', 5), |
2012
|
|
|
$token4 = new WordToken('two', 7, '', 'two'), |
2013
|
|
|
], |
2014
|
|
|
new Query([new Term($token1), new Term($token4)]), |
2015
|
|
|
[ |
2016
|
|
|
new Correction(Parser::CORRECTION_ADJACENT_UNARY_OPERATOR_PRECEDING_OPERATOR_IGNORED, $token3), |
2017
|
|
|
new Correction(Parser::CORRECTION_UNARY_OPERATOR_MISSING_OPERAND_IGNORED, $token2), |
2018
|
|
|
], |
2019
|
|
|
], |
2020
|
|
|
[ |
2021
|
|
|
'one !- two', |
2022
|
|
|
'one two', |
2023
|
|
|
[ |
2024
|
|
|
$token1 = new WordToken('one', 0, '', 'one'), |
2025
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_NOT_2, '!', 4), |
2026
|
|
|
$token3 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 5), |
2027
|
|
|
$token4 = new WordToken('two', 7, '', 'two'), |
2028
|
|
|
], |
2029
|
|
|
new Query([new Term($token1), new Term($token4)]), |
2030
|
|
|
[ |
2031
|
|
|
new Correction(Parser::CORRECTION_ADJACENT_UNARY_OPERATOR_PRECEDING_OPERATOR_IGNORED, $token2), |
2032
|
|
|
new Correction(Parser::CORRECTION_UNARY_OPERATOR_MISSING_OPERAND_IGNORED, $token3), |
2033
|
|
|
], |
2034
|
|
|
], |
2035
|
|
|
[ |
2036
|
|
|
'one !AND two', |
2037
|
|
|
'one AND two', |
2038
|
|
|
[ |
2039
|
|
|
$token1 = new WordToken('one', 0, '', 'one'), |
2040
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_NOT_2, '!', 4), |
2041
|
|
|
$token3 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 5), |
2042
|
|
|
$token4 = new WordToken('two', 9, '', 'two'), |
2043
|
|
|
], |
2044
|
|
|
new Query([new LogicalAnd(new Term($token1), new Term($token4), $token3)]), |
2045
|
|
|
[new Correction(Parser::CORRECTION_ADJACENT_UNARY_OPERATOR_PRECEDING_OPERATOR_IGNORED, $token2)], |
2046
|
|
|
], |
2047
|
|
|
[ |
2048
|
|
|
'one !OR two', |
2049
|
|
|
'one OR two', |
2050
|
|
|
[ |
2051
|
|
|
$token1 = new WordToken('one', 0, '', 'one'), |
2052
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_NOT_2, '!', 4), |
2053
|
|
|
$token3 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 5), |
2054
|
|
|
$token4 = new WordToken('two', 8, '', 'two'), |
2055
|
|
|
], |
2056
|
|
|
new Query([new LogicalOr(new Term($token1), new Term($token4), $token3)]), |
2057
|
|
|
[new Correction(Parser::CORRECTION_ADJACENT_UNARY_OPERATOR_PRECEDING_OPERATOR_IGNORED, $token2)], |
2058
|
|
|
], |
2059
|
|
|
[ |
2060
|
|
|
'one +! two', |
2061
|
|
|
'one two', |
2062
|
|
|
[ |
2063
|
|
|
$token1 = new WordToken('one', 0, '', 'one'), |
2064
|
|
|
$token2 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 4), |
2065
|
|
|
$token3 = new Token(Tokenizer::TOKEN_LOGICAL_NOT_2, '!', 5), |
2066
|
|
|
$token4 = new WordToken('two', 7, '', 'two'), |
2067
|
|
|
], |
2068
|
|
|
new Query([new Term($token1), new Term($token4)]), |
2069
|
|
|
[ |
2070
|
|
|
new Correction(Parser::CORRECTION_ADJACENT_UNARY_OPERATOR_PRECEDING_OPERATOR_IGNORED, $token2), |
2071
|
|
|
new Correction(Parser::CORRECTION_UNARY_OPERATOR_MISSING_OPERAND_IGNORED, $token3), |
2072
|
|
|
], |
2073
|
|
|
], |
2074
|
|
|
[ |
2075
|
|
|
'NOT+ one', |
2076
|
|
|
'NOT one', |
2077
|
|
|
[ |
2078
|
|
|
$token1 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 0), |
2079
|
|
|
$token2 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 3), |
2080
|
|
|
$token3 = new WordToken('one', 5, '', 'one'), |
2081
|
|
|
], |
2082
|
|
|
new Query([new LogicalNot(new Term($token3), $token1)]), |
2083
|
|
|
[new Correction(Parser::CORRECTION_UNARY_OPERATOR_MISSING_OPERAND_IGNORED, $token2)], |
2084
|
|
|
], |
2085
|
|
|
[ |
2086
|
|
|
'NOT- one', |
2087
|
|
|
'NOT one', |
2088
|
|
|
[ |
2089
|
|
|
$token1 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 0), |
2090
|
|
|
$token2 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 3), |
2091
|
|
|
$token3 = new WordToken('one', 5, '', 'one'), |
2092
|
|
|
], |
2093
|
|
|
new Query([new LogicalNot(new Term($token3), $token1)]), |
2094
|
|
|
[new Correction(Parser::CORRECTION_UNARY_OPERATOR_MISSING_OPERAND_IGNORED, $token2)], |
2095
|
|
|
], |
2096
|
|
|
[ |
2097
|
|
|
'NOT+one', |
2098
|
|
|
'+one', |
2099
|
|
|
[ |
2100
|
|
|
$token1 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 0), |
2101
|
|
|
$token2 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 3), |
2102
|
|
|
$token3 = new WordToken('one', 4, '', 'one'), |
2103
|
|
|
], |
2104
|
|
|
new Query([new Mandatory(new Term($token3), $token2)]), |
2105
|
|
|
[new Correction(Parser::CORRECTION_LOGICAL_NOT_OPERATORS_PRECEDING_PREFERENCE_IGNORED, $token1)], |
2106
|
|
|
], |
2107
|
|
|
[ |
2108
|
|
|
'+()NOT one', |
2109
|
|
|
'NOT one', |
2110
|
|
|
[ |
2111
|
|
|
$token1 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 0), |
2112
|
|
|
$token2 = new GroupBeginToken('(', 1, '(', null), |
2113
|
|
|
$token3 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 2), |
2114
|
|
|
$token4 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 3), |
2115
|
|
|
$token5 = new WordToken('one', 7, '', 'one'), |
2116
|
|
|
], |
2117
|
|
|
new Query([new LogicalNot(new Term($token5), $token4)]), |
2118
|
|
|
[new Correction(Parser::CORRECTION_EMPTY_GROUP_IGNORED, $token1, $token2, $token3)], |
2119
|
|
|
], |
2120
|
|
|
[ |
2121
|
|
|
'-()NOT one', |
2122
|
|
|
'NOT one', |
2123
|
|
|
[ |
2124
|
|
|
$token1 = new Token(Tokenizer::TOKEN_PROHIBITED, '-', 0), |
2125
|
|
|
$token2 = new GroupBeginToken('(', 1, '(', null), |
2126
|
|
|
$token3 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 2), |
2127
|
|
|
$token4 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 3), |
2128
|
|
|
$token5 = new WordToken('one', 7, '', 'one'), |
2129
|
|
|
], |
2130
|
|
|
new Query([new LogicalNot(new Term($token5), $token4)]), |
2131
|
|
|
[new Correction(Parser::CORRECTION_EMPTY_GROUP_IGNORED, $token1, $token2, $token3)], |
2132
|
|
|
], |
2133
|
|
|
[ |
2134
|
|
|
'+()NOT+()one', |
2135
|
|
|
'one', |
2136
|
|
|
[ |
2137
|
|
|
$token1 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 0), |
2138
|
|
|
$token2 = new GroupBeginToken('(', 1, '(', null), |
2139
|
|
|
$token3 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 2), |
2140
|
|
|
$token4 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 3), |
2141
|
|
|
$token5 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 6), |
2142
|
|
|
$token6 = new GroupBeginToken('(', 7, '(', null), |
2143
|
|
|
$token7 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 8), |
2144
|
|
|
$token8 = new WordToken('one', 9, '', 'one'), |
2145
|
|
|
], |
2146
|
|
|
new Query([new Term($token8)]), |
2147
|
|
|
[ |
2148
|
|
|
new Correction(Parser::CORRECTION_EMPTY_GROUP_IGNORED, $token1, $token2, $token3), |
2149
|
|
|
new Correction(Parser::CORRECTION_EMPTY_GROUP_IGNORED, $token4, $token5, $token6, $token7), |
2150
|
|
|
], |
2151
|
|
|
], |
2152
|
|
|
[ |
2153
|
|
|
'NOT()+one', |
2154
|
|
|
'+one', |
2155
|
|
|
[ |
2156
|
|
|
$token1 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 0), |
2157
|
|
|
$token2 = new GroupBeginToken('(', 3, '(', null), |
2158
|
|
|
$token3 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 4), |
2159
|
|
|
$token4 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 5), |
2160
|
|
|
$token5 = new WordToken('one', 6, '', 'one'), |
2161
|
|
|
], |
2162
|
|
|
new Query([new Mandatory(new Term($token5), $token4)]), |
2163
|
|
|
[new Correction(Parser::CORRECTION_EMPTY_GROUP_IGNORED, $token1, $token2, $token3)], |
2164
|
|
|
], |
2165
|
|
|
[ |
2166
|
|
|
'NOT () NOT one', |
2167
|
|
|
'NOT one', |
2168
|
|
|
[ |
2169
|
|
|
$token1 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 0), |
2170
|
|
|
$token2 = new GroupBeginToken('(', 4, '(', null), |
2171
|
|
|
$token3 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 5), |
2172
|
|
|
$token4 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 7), |
2173
|
|
|
$token5 = new WordToken('one', 11, '', 'one'), |
2174
|
|
|
], |
2175
|
|
|
new Query([new LogicalNot(new Term($token5), $token4)]), |
2176
|
|
|
[new Correction(Parser::CORRECTION_EMPTY_GROUP_IGNORED, $token1, $token2, $token3)], |
2177
|
|
|
], |
2178
|
|
|
[ |
2179
|
|
|
'NOT () +one', |
2180
|
|
|
'+one', |
2181
|
|
|
[ |
2182
|
|
|
$token1 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 0), |
2183
|
|
|
$token2 = new GroupBeginToken('(', 4, '(', null), |
2184
|
|
|
$token3 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 5), |
2185
|
|
|
$token4 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 7), |
2186
|
|
|
$token5 = new WordToken('one', 8, '', 'one'), |
2187
|
|
|
], |
2188
|
|
|
new Query([new Mandatory(new Term($token5), $token4)]), |
2189
|
|
|
[new Correction(Parser::CORRECTION_EMPTY_GROUP_IGNORED, $token1, $token2, $token3)], |
2190
|
|
|
], |
2191
|
|
|
[ |
2192
|
|
|
'NOT +()NOT +one', |
2193
|
|
|
'+one', |
2194
|
|
|
[ |
2195
|
|
|
$token1 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 0), |
2196
|
|
|
$token2 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 4), |
2197
|
|
|
$token3 = new GroupBeginToken('(', 5, '(', null), |
2198
|
|
|
$token4 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 6), |
2199
|
|
|
$token5 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 7), |
2200
|
|
|
$token6 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 11), |
2201
|
|
|
$token7 = new WordToken('one', 12, '', 'one'), |
2202
|
|
|
], |
2203
|
|
|
new Query([new Mandatory(new Term($token7), $token6)]), |
2204
|
|
|
[ |
2205
|
|
|
new Correction(Parser::CORRECTION_EMPTY_GROUP_IGNORED, $token1, $token2, $token3, $token4), |
2206
|
|
|
new Correction(Parser::CORRECTION_LOGICAL_NOT_OPERATORS_PRECEDING_PREFERENCE_IGNORED, $token5), |
2207
|
|
|
], |
2208
|
|
|
], |
2209
|
|
|
[ |
2210
|
|
|
'NOT +() NOT +one', |
2211
|
|
|
'+one', |
2212
|
|
|
[ |
2213
|
|
|
$token1 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 0), |
2214
|
|
|
$token2 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 4), |
2215
|
|
|
$token3 = new GroupBeginToken('(', 5, '(', null), |
2216
|
|
|
$token4 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 6), |
2217
|
|
|
$token5 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 8), |
2218
|
|
|
$token6 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 12), |
2219
|
|
|
$token7 = new WordToken('one', 13, '', 'one'), |
2220
|
|
|
], |
2221
|
|
|
new Query([new Mandatory(new Term($token7), $token6)]), |
2222
|
|
|
[ |
2223
|
|
|
new Correction(Parser::CORRECTION_EMPTY_GROUP_IGNORED, $token1, $token2, $token3, $token4), |
2224
|
|
|
new Correction(Parser::CORRECTION_LOGICAL_NOT_OPERATORS_PRECEDING_PREFERENCE_IGNORED, $token5), |
2225
|
|
|
], |
2226
|
|
|
], |
2227
|
|
|
[ |
2228
|
|
|
'(+()NOT one)AND', |
2229
|
|
|
'(NOT one)', |
2230
|
|
|
[ |
2231
|
|
|
$token1 = new GroupBeginToken('(', 0, '(', null), |
2232
|
|
|
$token2 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 1), |
2233
|
|
|
$token3 = new GroupBeginToken('(', 2, '(', null), |
2234
|
|
|
$token4 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 3), |
2235
|
|
|
$token5 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 4), |
2236
|
|
|
$token6 = new WordToken('one', 8, '', 'one'), |
2237
|
|
|
$token7 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 11), |
2238
|
|
|
$token8 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 12), |
2239
|
|
|
], |
2240
|
|
|
new Query([new Group([new LogicalNot(new Term($token6), $token5)], $token1, $token7)]), |
2241
|
|
|
[ |
2242
|
|
|
new Correction(Parser::CORRECTION_EMPTY_GROUP_IGNORED, $token2, $token3, $token4), |
2243
|
|
|
new Correction(Parser::CORRECTION_BINARY_OPERATOR_MISSING_RIGHT_OPERAND_IGNORED, $token8), |
2244
|
|
|
], |
2245
|
|
|
], |
2246
|
|
|
[ |
2247
|
|
|
'one !NOT two', |
2248
|
|
|
'one NOT two', |
2249
|
|
|
[ |
2250
|
|
|
$token1 = new WordToken('one', 0, '', 'one'), |
2251
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_NOT_2, '!', 4), |
2252
|
|
|
$token3 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 5), |
2253
|
|
|
$token4 = new WordToken('two', 9, '', 'two'), |
2254
|
|
|
], |
2255
|
|
|
new Query([new Term($token1), new LogicalNot(new Term($token4), $token3)]), |
2256
|
|
|
[new Correction(Parser::CORRECTION_ADJACENT_UNARY_OPERATOR_PRECEDING_OPERATOR_IGNORED, $token2)], |
2257
|
|
|
], |
2258
|
|
|
[ |
2259
|
|
|
'NOT NOT +one', |
2260
|
|
|
'+one', |
2261
|
|
|
[ |
2262
|
|
|
$token1 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 0), |
2263
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 4), |
2264
|
|
|
$token3 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 8), |
2265
|
|
|
$token4 = new WordToken('one', 9, '', 'one'), |
2266
|
|
|
], |
2267
|
|
|
new Query([new Mandatory(new Term($token4), $token3)]), |
2268
|
|
|
[ |
2269
|
|
|
new Correction( |
2270
|
|
|
Parser::CORRECTION_LOGICAL_NOT_OPERATORS_PRECEDING_PREFERENCE_IGNORED, |
2271
|
|
|
$token1, |
2272
|
|
|
$token2 |
2273
|
|
|
), |
2274
|
|
|
], |
2275
|
|
|
], |
2276
|
|
|
[ |
2277
|
|
|
'NOT !+one', |
2278
|
|
|
'+one', |
2279
|
|
|
[ |
2280
|
|
|
$token1 = new Token(Tokenizer::TOKEN_LOGICAL_NOT, 'NOT', 0), |
2281
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_NOT_2, '!', 4), |
2282
|
|
|
$token3 = new Token(Tokenizer::TOKEN_MANDATORY, '+', 5), |
2283
|
|
|
$token4 = new WordToken('one', 6, '', 'one'), |
2284
|
|
|
], |
2285
|
|
|
new Query([new Mandatory(new Term($token4), $token3)]), |
2286
|
|
|
[ |
2287
|
|
|
new Correction(Parser::CORRECTION_ADJACENT_UNARY_OPERATOR_PRECEDING_OPERATOR_IGNORED, $token2), |
2288
|
|
|
new Correction(Parser::CORRECTION_LOGICAL_NOT_OPERATORS_PRECEDING_PREFERENCE_IGNORED, $token1), |
2289
|
|
|
], |
2290
|
|
|
], |
2291
|
|
|
[ |
2292
|
|
|
'one OR two AND () three', |
2293
|
|
|
'one OR two three', |
2294
|
|
|
[ |
2295
|
|
|
$token1 = new WordToken('one', 0, '', 'one'), |
2296
|
|
|
$token2 = new Token(Tokenizer::TOKEN_LOGICAL_OR, 'OR', 4), |
2297
|
|
|
$token3 = new WordToken('two', 7, '', 'two'), |
2298
|
|
|
$token4 = new Token(Tokenizer::TOKEN_LOGICAL_AND, 'AND', 11), |
2299
|
|
|
$token5 = new GroupBeginToken('(', 15, '(', null), |
2300
|
|
|
$token6 = new Token(Tokenizer::TOKEN_GROUP_END, ')', 16), |
2301
|
|
|
$token7 = new WordToken('three', 18, '', 'three'), |
2302
|
|
|
], |
2303
|
|
|
new Query([new LogicalOr(new Term($token1), new Term($token3), $token2), new Term($token7)]), |
2304
|
|
|
[new Correction(Parser::CORRECTION_EMPTY_GROUP_IGNORED, $token4, $token5, $token6)], |
2305
|
|
|
], |
2306
|
|
|
]; |
2307
|
|
|
} |
2308
|
|
|
|
2309
|
|
|
|
2310
|
|
|
/** |
2311
|
|
|
* @dataProvider providerForTestQuery |
2312
|
|
|
* |
2313
|
|
|
* @param string $string |
2314
|
|
|
* @param Token[] $expectedTokens |
2315
|
|
|
* @param Query $expectedTree |
2316
|
|
|
*/ |
2317
|
|
|
public function testQuery($string, $expectedTokens, $expectedTree): void |
2318
|
|
|
{ |
2319
|
|
|
$this->doTestQuery($string, $string, $expectedTokens, $expectedTree, []); |
2320
|
|
|
} |
2321
|
|
|
|
2322
|
|
|
|
2323
|
|
|
/** |
2324
|
|
|
* @dataProvider providerForTestQueryCorrected |
2325
|
|
|
* |
2326
|
|
|
* @param string $string |
2327
|
|
|
* @param string $correctedString |
2328
|
|
|
* @param Token[] $expectedTokens |
2329
|
|
|
* @param Query $query |
2330
|
|
|
* @param Correction[] $corrections |
2331
|
|
|
*/ |
2332
|
|
|
public function testQueryCorrected($string, $correctedString, $expectedTokens, $query, $corrections): void |
2333
|
|
|
{ |
2334
|
|
|
$this->doTestQuery($string, $correctedString, $expectedTokens, $query, $corrections); |
2335
|
|
|
} |
2336
|
|
|
|
2337
|
|
|
|
2338
|
|
|
/** |
2339
|
|
|
* @param string $string |
2340
|
|
|
* @param string $expectedCorrectedString |
2341
|
|
|
* @param Token[] $expectedTokens |
2342
|
|
|
* @param Query $query |
2343
|
|
|
* @param Correction[] $corrections |
2344
|
|
|
*/ |
2345
|
|
|
protected function doTestQuery($string, $expectedCorrectedString, $expectedTokens, $query, $corrections): void |
2346
|
|
|
{ |
2347
|
|
|
$tokenExtractor = new Full; |
2348
|
|
|
$tokenizer = new Tokenizer($tokenExtractor); |
2349
|
|
|
$parser = new Parser(); |
2350
|
|
|
$generator = $this->getNativeGenerator(); |
2351
|
|
|
|
2352
|
|
|
$tokenSequence = $tokenizer->tokenize($string); |
2353
|
|
|
self::assertInstanceOf(TokenSequence::class, $tokenSequence); |
2354
|
|
|
|
2355
|
|
|
$syntaxTree = $parser->parse($tokenSequence); |
2356
|
|
|
self::assertInstanceOf(SyntaxTree::class, $syntaxTree); |
2357
|
|
|
|
2358
|
|
|
$correctedString = $generator->generate($syntaxTree); |
2359
|
|
|
|
2360
|
|
|
$tokensWithoutWhitespace = []; |
2361
|
|
|
foreach ($tokenSequence->getTokens() as $token) { |
2362
|
|
|
if ($token->getType() !== Tokenizer::TOKEN_WHITESPACE) { |
2363
|
|
|
$tokensWithoutWhitespace[] = $token; |
2364
|
|
|
} |
2365
|
|
|
} |
2366
|
|
|
|
2367
|
|
|
self::assertEquals($expectedCorrectedString, $correctedString); |
2368
|
|
|
self::assertEquals($expectedTokens, $tokensWithoutWhitespace); |
2369
|
|
|
self::assertEquals($query, $syntaxTree->getRootNode()); |
2370
|
|
|
self::assertEquals($corrections, $syntaxTree->getCorrections()); |
2371
|
|
|
self::assertEquals($tokenSequence, $syntaxTree->getTokenSequence()); |
2372
|
|
|
} |
2373
|
|
|
|
2374
|
|
|
|
2375
|
|
|
protected function getNativeGenerator(): Native |
2376
|
|
|
{ |
2377
|
|
|
$visitors = []; |
2378
|
|
|
|
2379
|
|
|
$visitors[] = new GroupNative; |
2380
|
|
|
$visitors[] = new BinaryOperator(); |
2381
|
|
|
$visitors[] = new Phrase(); |
2382
|
|
|
$visitors[] = new QueryNative(); |
2383
|
|
|
$visitors[] = new Tag(); |
2384
|
|
|
$visitors[] = new UnaryOperator(); |
2385
|
|
|
$visitors[] = new User(); |
2386
|
|
|
$visitors[] = new Word(); |
2387
|
|
|
|
2388
|
|
|
$aggregate = new Aggregate($visitors); |
2389
|
|
|
|
2390
|
|
|
return new Native($aggregate); |
2391
|
|
|
} |
2392
|
|
|
|
2393
|
|
|
} |
2394
|
|
|
|