|
1
|
|
|
package br.ufrj.ppgi.greco.kettle; |
|
2
|
|
|
|
|
3
|
|
|
import java.io.ByteArrayOutputStream; |
|
4
|
|
|
import java.io.IOException; |
|
5
|
|
|
import java.io.PrintWriter; |
|
6
|
|
|
|
|
7
|
|
|
import org.pentaho.di.core.exception.KettleException; |
|
8
|
|
|
import org.pentaho.di.core.exception.KettleStepException; |
|
9
|
|
|
import org.pentaho.di.core.row.RowDataUtil; |
|
10
|
|
|
import org.pentaho.di.core.row.RowMeta; |
|
11
|
|
|
import org.pentaho.di.core.row.RowMetaInterface; |
|
12
|
|
|
import org.pentaho.di.trans.Trans; |
|
13
|
|
|
import org.pentaho.di.trans.TransMeta; |
|
14
|
|
|
import org.pentaho.di.trans.step.BaseStep; |
|
15
|
|
|
import org.pentaho.di.trans.step.StepDataInterface; |
|
16
|
|
|
import org.pentaho.di.trans.step.StepInterface; |
|
17
|
|
|
import org.pentaho.di.trans.step.StepMeta; |
|
18
|
|
|
import org.pentaho.di.trans.step.StepMetaInterface; |
|
19
|
|
|
|
|
20
|
|
|
import org.apache.jena.query.ARQ; |
|
21
|
|
|
import org.apache.jena.query.Query; |
|
22
|
|
|
import org.apache.jena.query.QueryException; |
|
23
|
|
|
import org.apache.jena.query.QueryExecution; |
|
24
|
|
|
import org.apache.jena.query.QueryFactory; |
|
25
|
|
|
import org.apache.jena.query.QuerySolution; |
|
26
|
|
|
import org.apache.jena.query.ResultSet; |
|
27
|
|
|
import org.apache.jena.rdf.model.Literal; |
|
28
|
|
|
import org.apache.jena.rdf.model.Model; |
|
29
|
|
|
import org.apache.jena.rdf.model.RDFNode; |
|
30
|
|
|
import org.apache.jena.rdf.model.Statement; |
|
31
|
|
|
import org.apache.jena.rdf.model.StmtIterator; |
|
32
|
|
|
|
|
33
|
|
|
public class SparqlStep extends BaseStep implements StepInterface { |
|
34
|
|
|
|
|
35
|
|
|
private static int MAX_ATTEMPTS = 4; |
|
36
|
|
|
|
|
37
|
|
|
public SparqlStep(StepMeta stepMeta, StepDataInterface stepDataInterface, int copyNr, TransMeta transMeta, |
|
38
|
|
|
Trans trans) { |
|
39
|
|
|
super(stepMeta, stepDataInterface, copyNr, transMeta, trans); |
|
40
|
|
|
} |
|
41
|
|
|
|
|
42
|
|
View Code Duplication |
public boolean processRow(StepMetaInterface smi, StepDataInterface sdi) throws KettleException { |
|
|
|
|
|
|
43
|
|
|
SparqlStepMeta meta = (SparqlStepMeta) smi; |
|
44
|
|
|
SparqlStepData data = (SparqlStepData) sdi; |
|
45
|
|
|
|
|
46
|
|
|
final Object[] row = getRow(); |
|
47
|
|
|
|
|
48
|
|
|
if (first) { |
|
49
|
|
|
first = false; |
|
50
|
|
|
|
|
51
|
|
|
RowMetaInterface rowMeta = getInputRowMeta(row != null); |
|
52
|
|
|
data.outputRowMeta = rowMeta.clone(); |
|
53
|
|
|
|
|
54
|
|
|
// Adiciona os metadados do step atual |
|
55
|
|
|
meta.getFields(data.outputRowMeta, getStepname(), null, null, this); |
|
56
|
|
|
|
|
57
|
|
|
data.inputRowSize = rowMeta.size(); |
|
58
|
|
|
|
|
59
|
|
|
// Obtem string de consulta e constroi o objeto consulta |
|
60
|
|
|
String queryStr = SparqlStepUtils.toFullQueryString(meta.getPrefixes(), meta.getQueryString()); |
|
61
|
|
|
try { |
|
62
|
|
|
data.originalQuery = QueryFactory.create(queryStr); |
|
63
|
|
|
} catch (QueryException e) { |
|
64
|
|
|
// Se consulta for invalida nao pode continuar |
|
65
|
|
|
throw new KettleException(e); |
|
66
|
|
|
} |
|
67
|
|
|
|
|
68
|
|
|
// Se nao usar SAX o execSelect() nao funciona |
|
69
|
|
|
ARQ.set(ARQ.useSAX, true); |
|
70
|
|
|
|
|
71
|
|
|
// Offset e Limit para Construct/select/describe quando limit nao |
|
72
|
|
|
// especificado |
|
73
|
|
|
if (!data.originalQuery.hasLimit() && (data.originalQuery.getQueryType() != Query.QueryTypeAsk) |
|
74
|
|
|
&& (data.originalQuery.getQueryType() != Query.QueryTypeDescribe)) { |
|
75
|
|
|
// Consulta eh quebrada em varias usando OFFSET e LIMIT |
|
76
|
|
|
data.offset = data.originalQuery.hasOffset() ? data.originalQuery.getOffset() : 0; |
|
77
|
|
|
data.limit = 1000; |
|
78
|
|
|
data.runAtOnce = false; |
|
79
|
|
|
} else { |
|
80
|
|
|
data.runAtOnce = true; |
|
81
|
|
|
} |
|
82
|
|
|
|
|
83
|
|
|
data.remainingTries = MAX_ATTEMPTS; |
|
84
|
|
|
|
|
85
|
|
|
return true; |
|
86
|
|
|
} |
|
87
|
|
|
|
|
88
|
|
|
Query query = null; |
|
89
|
|
|
if (data.runAtOnce) { |
|
90
|
|
|
// Roda consulta num unico HTTP Request |
|
91
|
|
|
query = data.originalQuery; |
|
92
|
|
|
|
|
93
|
|
|
while (data.remainingTries > 0) { |
|
94
|
|
|
// Tenta executar consulta ate MAX_ATTEMPTS vezes |
|
95
|
|
|
try { |
|
96
|
|
|
runQueryAndPutResults(query, meta, data, row); |
|
97
|
|
|
|
|
98
|
|
|
setOutputDone(); |
|
99
|
|
|
return false; // Nao ha mais resultados, ie, processRow() |
|
100
|
|
|
// nao sera' chamado novamente |
|
101
|
|
|
} catch (Throwable e) { |
|
102
|
|
|
handleError(e, MAX_ATTEMPTS - data.remainingTries + 1); |
|
103
|
|
|
} |
|
104
|
|
|
|
|
105
|
|
|
data.remainingTries--; |
|
106
|
|
|
} |
|
107
|
|
|
} else { |
|
108
|
|
|
// Cria consulta que representa o bloco atual |
|
109
|
|
|
query = data.originalQuery.cloneQuery(); |
|
110
|
|
|
query.setOffset(data.offset); |
|
111
|
|
|
query.setLimit(data.limit); |
|
112
|
|
|
|
|
113
|
|
|
while (data.remainingTries > 0) { |
|
114
|
|
|
try { |
|
115
|
|
|
int numRows = runQueryAndPutResults(query, meta, data, row); |
|
116
|
|
|
|
|
117
|
|
|
if (numRows > 0) { |
|
118
|
|
|
data.offset += data.limit; |
|
119
|
|
|
data.remainingTries = MAX_ATTEMPTS; |
|
120
|
|
|
return true; |
|
121
|
|
|
} else { |
|
122
|
|
|
setOutputDone(); |
|
123
|
|
|
return false; |
|
124
|
|
|
} |
|
125
|
|
|
} catch (Throwable e) { |
|
126
|
|
|
handleError(e, MAX_ATTEMPTS - data.remainingTries + 1); |
|
127
|
|
|
} |
|
128
|
|
|
|
|
129
|
|
|
data.remainingTries--; |
|
130
|
|
|
} |
|
131
|
|
|
} |
|
132
|
|
|
|
|
133
|
|
|
StringBuilder sb = new StringBuilder(); |
|
134
|
|
|
sb.append("Todas as tentativas de executar a consulta falharam. "); |
|
135
|
|
|
sb.append("Verifique conexão de rede e o SPARQL Endpoint.\n"); |
|
136
|
|
|
sb.append("Endpoint: "); |
|
137
|
|
|
sb.append(meta.getEndpointUri()); |
|
138
|
|
|
sb.append('\n'); |
|
139
|
|
|
sb.append("Grafo padrão: "); |
|
140
|
|
|
sb.append(meta.getDefaultGraph()); |
|
141
|
|
|
sb.append('\n'); |
|
142
|
|
|
sb.append("Consulta:\n"); |
|
143
|
|
|
sb.append(query.toString()); |
|
144
|
|
|
sb.append('\n'); |
|
145
|
|
|
|
|
146
|
|
|
throw new KettleException(sb.toString()); |
|
147
|
|
|
} |
|
148
|
|
|
|
|
149
|
|
|
private RowMetaInterface getInputRowMeta(boolean hasInputRow) { |
|
150
|
|
|
|
|
151
|
|
|
RowMetaInterface rowMeta = null; |
|
152
|
|
|
if (hasInputRow) |
|
153
|
|
|
rowMeta = getInputRowMeta(); |
|
154
|
|
|
else |
|
155
|
|
|
rowMeta = new RowMeta(); |
|
156
|
|
|
|
|
157
|
|
|
return rowMeta; |
|
158
|
|
|
} |
|
159
|
|
|
|
|
160
|
|
View Code Duplication |
private void handleError(Throwable e, int attempts) { |
|
|
|
|
|
|
161
|
|
|
|
|
162
|
|
|
try { |
|
163
|
|
|
String msg = String.format("Falha ao executar consulta (tentativa %d de %d): ", attempts, MAX_ATTEMPTS); |
|
164
|
|
|
|
|
165
|
|
|
ByteArrayOutputStream baos = new ByteArrayOutputStream(); |
|
166
|
|
|
baos.write(msg.getBytes()); |
|
167
|
|
|
|
|
168
|
|
|
e.printStackTrace(new PrintWriter(baos, true)); |
|
169
|
|
|
|
|
170
|
|
|
long sleepTime = (long) (500 * Math.pow(2, attempts)); |
|
171
|
|
|
msg = String.format("Tentando novamente em %d milissegundos...", sleepTime); |
|
172
|
|
|
baos.write(msg.getBytes()); |
|
173
|
|
|
|
|
174
|
|
|
log.logBasic(baos.toString()); |
|
175
|
|
|
|
|
176
|
|
|
Thread.sleep(sleepTime); |
|
177
|
|
|
|
|
178
|
|
|
} catch (IOException e1) { |
|
179
|
|
|
e1.printStackTrace(); |
|
180
|
|
|
} catch (InterruptedException e2) { |
|
181
|
|
|
e2.printStackTrace(); |
|
182
|
|
|
} |
|
183
|
|
|
} |
|
184
|
|
|
|
|
185
|
|
|
private int tripleWriter(Model model, Object[] row, SparqlStepData data) throws KettleStepException { |
|
186
|
|
|
int numPutRows = 0; |
|
187
|
|
|
StmtIterator it = model.listStatements(); |
|
188
|
|
|
|
|
189
|
|
|
while (it.hasNext()) { |
|
190
|
|
|
Statement stmt = it.next(); |
|
191
|
|
|
|
|
192
|
|
|
incrementLinesInput(); |
|
193
|
|
|
|
|
194
|
|
|
String subject = stmt.getSubject().toString(); |
|
195
|
|
|
String predicate = stmt.getPredicate().toString(); |
|
196
|
|
|
String object = stmt.getObject().toString(); |
|
197
|
|
|
|
|
198
|
|
|
if (subject != null && !subject.isEmpty() && predicate != null && !predicate.isEmpty() && object != null) { |
|
199
|
|
|
// Monta linha com a tripla |
|
200
|
|
|
Object[] outputRow = row; |
|
201
|
|
|
outputRow = RowDataUtil.addValueData(outputRow, data.inputRowSize + 0, subject); |
|
202
|
|
|
outputRow = RowDataUtil.addValueData(outputRow, data.inputRowSize + 1, predicate); |
|
203
|
|
|
outputRow = RowDataUtil.addValueData(outputRow, data.inputRowSize + 2, object); |
|
204
|
|
|
|
|
205
|
|
|
// Joga tripla no fluxo |
|
206
|
|
|
putRow(data.outputRowMeta, outputRow); |
|
207
|
|
|
|
|
208
|
|
|
numPutRows++; |
|
209
|
|
|
} else |
|
210
|
|
|
logBasic("Tripla ignorada: " + stmt.getString()); |
|
211
|
|
|
} |
|
212
|
|
|
|
|
213
|
|
|
return numPutRows; |
|
214
|
|
|
} |
|
215
|
|
|
|
|
216
|
|
|
private int runQueryAndPutResults(Query query, SparqlStepMeta meta, SparqlStepData data, Object[] row) |
|
217
|
|
|
throws KettleStepException { |
|
218
|
|
|
int numPutRows = 0; |
|
219
|
|
|
Model model = null; |
|
220
|
|
|
QueryExecution qexec = SparqlStepUtils.createQueryExecution(query, meta.getEndpointUri(), |
|
221
|
|
|
meta.getDefaultGraph()); |
|
222
|
|
|
|
|
223
|
|
|
try { |
|
224
|
|
|
switch (query.getQueryType()) { |
|
225
|
|
|
case Query.QueryTypeAsk: |
|
226
|
|
|
boolean result = qexec.execAsk(); |
|
227
|
|
|
incrementLinesInput(); |
|
228
|
|
|
putRow(data.outputRowMeta, RowDataUtil.addValueData(row, data.inputRowSize, result)); |
|
229
|
|
|
break; |
|
230
|
|
|
|
|
231
|
|
|
case Query.QueryTypeConstruct: |
|
232
|
|
|
model = qexec.execConstruct(); |
|
233
|
|
|
numPutRows = tripleWriter(model, row, data); |
|
234
|
|
|
break; |
|
235
|
|
|
|
|
236
|
|
|
case Query.QueryTypeDescribe: |
|
237
|
|
|
model = qexec.execDescribe(); |
|
238
|
|
|
numPutRows = tripleWriter(model, row, data); |
|
239
|
|
|
break; |
|
240
|
|
|
|
|
241
|
|
|
case Query.QueryTypeSelect: |
|
242
|
|
|
ResultSet resultSet = qexec.execSelect(); |
|
243
|
|
|
// Gera linhas |
|
244
|
|
|
while (resultSet.hasNext()) { |
|
245
|
|
|
QuerySolution qs = resultSet.next(); |
|
246
|
|
|
|
|
247
|
|
|
// Diz ao Kettle que leu mais uma linha da entrada |
|
248
|
|
|
incrementLinesInput(); |
|
249
|
|
|
|
|
250
|
|
|
// Gera uma linha de saida do fluxo |
|
251
|
|
|
Object[] outputRow = row; |
|
252
|
|
|
String[] fieldNames = data.outputRowMeta.getFieldNames(); |
|
253
|
|
|
int posAddValueData = (outputRow != null) ? outputRow.length : 0; |
|
254
|
|
|
for (int i = 0; i < fieldNames.length; i++) { |
|
255
|
|
|
// Retira o prefixo para obter o nome do campo da |
|
256
|
|
|
// consulta SPARQL |
|
257
|
|
|
String var = fieldNames[i].replaceAll(meta.getVarPrefix(), ""); |
|
258
|
|
|
|
|
259
|
|
|
// Obtem o node RDF |
|
260
|
|
|
RDFNode node = qs.get(var); |
|
261
|
|
|
|
|
262
|
|
|
// Obtem o valor do node RDF |
|
263
|
|
|
String value = null; |
|
264
|
|
|
if (node instanceof Literal) { |
|
265
|
|
|
value = qs.getLiteral(var).getString(); |
|
266
|
|
|
} else { |
|
267
|
|
|
value = qs.getResource(var).getURI(); |
|
268
|
|
|
} |
|
269
|
|
|
|
|
270
|
|
|
// Set output row |
|
271
|
|
|
outputRow = RowDataUtil.addValueData(outputRow, posAddValueData++, value); |
|
272
|
|
|
} |
|
273
|
|
|
|
|
274
|
|
|
putRow(data.outputRowMeta, outputRow); |
|
275
|
|
|
numPutRows++; |
|
276
|
|
|
} |
|
277
|
|
|
break; |
|
278
|
|
|
} |
|
279
|
|
|
} finally { |
|
280
|
|
|
qexec.close(); |
|
281
|
|
|
} |
|
282
|
|
|
|
|
283
|
|
|
return numPutRows; |
|
284
|
|
|
} |
|
285
|
|
|
} |