|
1
|
|
|
# -*- coding: utf-8 -*- |
|
2
|
|
|
# |
|
3
|
|
|
# This file is part of SENAITE.CORE |
|
4
|
|
|
# |
|
5
|
|
|
# Copyright 2018 by it's authors. |
|
6
|
|
|
# Some rights reserved. See LICENSE.rst, CONTRIBUTORS.rst. |
|
7
|
|
|
|
|
8
|
|
|
""" Agilent's 'Masshunter Quant' |
|
9
|
|
|
""" |
|
10
|
|
|
from DateTime import DateTime |
|
11
|
|
|
from Products.Archetypes.event import ObjectInitializedEvent |
|
12
|
|
|
from Products.CMFCore.utils import getToolByName |
|
13
|
|
|
from Products.Five.browser.pagetemplatefile import ViewPageTemplateFile |
|
14
|
|
|
from bika.lims import bikaMessageFactory as _ |
|
15
|
|
|
from bika.lims.utils import t |
|
16
|
|
|
from bika.lims import logger |
|
17
|
|
|
from bika.lims.browser import BrowserView |
|
18
|
|
|
from bika.lims.idserver import renameAfterCreation |
|
19
|
|
|
from bika.lims.utils import changeWorkflowState |
|
20
|
|
|
from bika.lims.utils import tmpID |
|
21
|
|
|
from cStringIO import StringIO |
|
22
|
|
|
from datetime import datetime |
|
23
|
|
|
from operator import itemgetter |
|
24
|
|
|
from plone.i18n.normalizer.interfaces import IIDNormalizer |
|
25
|
|
|
from zope.component import getUtility |
|
26
|
|
|
import csv |
|
27
|
|
|
import json |
|
28
|
|
|
import plone |
|
29
|
|
|
import zope |
|
30
|
|
|
import zope.event |
|
31
|
|
|
from bika.lims.exportimport.instruments.resultsimport import InstrumentCSVResultsFileParser,\ |
|
32
|
|
|
AnalysisResultsImporter |
|
33
|
|
|
import traceback |
|
34
|
|
|
|
|
35
|
|
|
title = "Agilent - Masshunter Quantitative" |
|
36
|
|
|
|
|
37
|
|
|
|
|
38
|
|
|
def Import(context, request): |
|
39
|
|
|
""" Read Agilent's Masshunter Quant analysis results |
|
40
|
|
|
""" |
|
41
|
|
|
infile = request.form['amhq_file'] |
|
42
|
|
|
fileformat = request.form['amhq_format'] |
|
43
|
|
|
artoapply = request.form['amhq_artoapply'] |
|
44
|
|
|
override = request.form['amhq_override'] |
|
45
|
|
|
instrument = request.form.get('instrument', None) |
|
46
|
|
|
errors = [] |
|
47
|
|
|
logs = [] |
|
48
|
|
|
|
|
49
|
|
|
# Load the most suitable parser according to file extension/options/etc... |
|
50
|
|
|
parser = None |
|
51
|
|
|
if not hasattr(infile, 'filename'): |
|
52
|
|
|
errors.append(_("No file selected")) |
|
53
|
|
|
elif fileformat == 'csv': |
|
54
|
|
|
parser = MasshunterQuantCSVParser(infile) |
|
55
|
|
|
else: |
|
56
|
|
|
errors.append(t(_("Unrecognized file format ${fileformat}", |
|
57
|
|
|
mapping={"fileformat": fileformat}))) |
|
58
|
|
|
|
|
59
|
|
|
if parser: |
|
60
|
|
|
# Load the importer |
|
61
|
|
|
status = ['sample_received', 'attachment_due', 'to_be_verified'] |
|
62
|
|
|
if artoapply == 'received': |
|
63
|
|
|
status = ['sample_received'] |
|
64
|
|
|
elif artoapply == 'received_tobeverified': |
|
65
|
|
|
status = ['sample_received', 'attachment_due', 'to_be_verified'] |
|
66
|
|
|
|
|
67
|
|
|
over = [False, False] |
|
68
|
|
|
if override == 'nooverride': |
|
69
|
|
|
over = [False, False] |
|
70
|
|
|
elif override == 'override': |
|
71
|
|
|
over = [True, False] |
|
72
|
|
|
elif override == 'overrideempty': |
|
73
|
|
|
over = [True, True] |
|
74
|
|
|
|
|
75
|
|
|
importer = MasshunterQuantImporter(parser=parser, |
|
76
|
|
|
context=context, |
|
77
|
|
|
allowed_ar_states=status, |
|
78
|
|
|
allowed_analysis_states=None, |
|
79
|
|
|
override=over, |
|
80
|
|
|
instrument_uid=instrument) |
|
81
|
|
|
tbex = '' |
|
82
|
|
|
try: |
|
83
|
|
|
importer.process() |
|
84
|
|
|
except: |
|
85
|
|
|
tbex = traceback.format_exc() |
|
86
|
|
|
errors = importer.errors |
|
87
|
|
|
logs = importer.logs |
|
88
|
|
|
warns = importer.warns |
|
89
|
|
|
if tbex: |
|
90
|
|
|
errors.append(tbex) |
|
91
|
|
|
|
|
92
|
|
|
results = {'errors': errors, 'log': logs, 'warns': warns} |
|
|
|
|
|
|
93
|
|
|
|
|
94
|
|
|
return json.dumps(results) |
|
95
|
|
|
|
|
96
|
|
|
|
|
97
|
|
|
class MasshunterQuantCSVParser(InstrumentCSVResultsFileParser): |
|
98
|
|
|
|
|
99
|
|
|
HEADERKEY_BATCHINFO = 'Batch Info' |
|
100
|
|
|
HEADERKEY_BATCHDATAPATH = 'Batch Data Path' |
|
101
|
|
|
HEADERKEY_ANALYSISTIME = 'Analysis Time' |
|
102
|
|
|
HEADERKEY_ANALYSTNAME = 'Analyst Name' |
|
103
|
|
|
HEADERKEY_REPORTTIME = 'Report Time' |
|
104
|
|
|
HEADERKEY_REPORTERNAME = 'Reporter Name' |
|
105
|
|
|
HEADERKEY_LASTCALIBRATION = 'Last Calib Update' |
|
106
|
|
|
HEADERKEY_BATCHSTATE = 'Batch State' |
|
107
|
|
|
SEQUENCETABLE_KEY = 'Sequence Table' |
|
108
|
|
|
SEQUENCETABLE_HEADER_DATAFILE = 'Data File' |
|
109
|
|
|
SEQUENCETABLE_HEADER_SAMPLENAME = 'Sample Name' |
|
110
|
|
|
SEQUENCETABLE_PRERUN = 'prerunrespchk.d' |
|
111
|
|
|
SEQUENCETABLE_MIDRUN = 'mid_respchk.d' |
|
112
|
|
|
SEQUENCETABLE_POSTRUN = 'post_respchk.d' |
|
113
|
|
|
SEQUENCETABLE_NUMERICHEADERS = ('Inj Vol',) |
|
114
|
|
|
QUANTITATIONRESULTS_KEY = 'Quantification Results' |
|
115
|
|
|
QUANTITATIONRESULTS_TARGETCOMPOUND = 'Target Compound' |
|
116
|
|
|
QUANTITATIONRESULTS_HEADER_DATAFILE = 'Data File' |
|
117
|
|
|
QUANTITATIONRESULTS_PRERUN = 'prerunrespchk.d' |
|
118
|
|
|
QUANTITATIONRESULTS_MIDRUN = 'mid_respchk.d' |
|
119
|
|
|
QUANTITATIONRESULTS_POSTRUN = 'post_respchk.d' |
|
120
|
|
|
QUANTITATIONRESULTS_NUMERICHEADERS = ('Resp', 'ISTD Resp', 'Resp Ratio', |
|
121
|
|
|
'Final Conc', 'Exp Conc', 'Accuracy') |
|
122
|
|
|
QUANTITATIONRESULTS_COMPOUNDCOLUMN = 'Compound' |
|
123
|
|
|
COMMAS = ',' |
|
124
|
|
|
|
|
125
|
|
|
def __init__(self, csv): |
|
126
|
|
|
InstrumentCSVResultsFileParser.__init__(self, csv) |
|
127
|
|
|
self._end_header = False |
|
128
|
|
|
self._end_sequencetable = False |
|
129
|
|
|
self._sequences = [] |
|
130
|
|
|
self._sequencesheader = [] |
|
131
|
|
|
self._quantitationresultsheader = [] |
|
132
|
|
|
self._numline = 0 |
|
133
|
|
|
|
|
134
|
|
|
def getAttachmentFileType(self): |
|
135
|
|
|
return "Agilent's Masshunter Quant CSV" |
|
136
|
|
|
|
|
137
|
|
|
def _parseline(self, line): |
|
138
|
|
|
if self._end_header == False: |
|
139
|
|
|
return self.parse_headerline(line) |
|
140
|
|
|
elif self._end_sequencetable == False: |
|
141
|
|
|
return self.parse_sequencetableline(line) |
|
142
|
|
|
else: |
|
143
|
|
|
return self.parse_quantitationesultsline(line) |
|
144
|
|
|
|
|
145
|
|
|
def parse_headerline(self, line): |
|
146
|
|
|
""" Parses header lines |
|
147
|
|
|
|
|
148
|
|
|
Header example: |
|
149
|
|
|
Batch Info,2013-03-20T07:11:09.9053262-07:00,2013-03-20T07:12:55.5280967-07:00,2013-03-20T07:11:07.1047817-07:00,,,,,,,,,,,,,, |
|
150
|
|
|
Batch Data Path,D:\MassHunter\Data\130129\QuantResults\130129LS.batch.bin,,,,,,,,,,,,,,,, |
|
151
|
|
|
Analysis Time,3/20/2013 7:11 AM,Analyst Name,Administrator,,,,,,,,,,,,,, |
|
152
|
|
|
Report Time,3/20/2013 7:12 AM,Reporter Name,Administrator,,,,,,,,,,,,,, |
|
153
|
|
|
Last Calib Update,3/20/2013 7:11 AM,Batch State,Processed,,,,,,,,,,,,,, |
|
154
|
|
|
,,,,,,,,,,,,,,,,, |
|
155
|
|
|
""" |
|
156
|
|
|
if self._end_header == True: |
|
157
|
|
|
# Header already processed |
|
158
|
|
|
return 0 |
|
159
|
|
|
|
|
160
|
|
|
if line.startswith(self.SEQUENCETABLE_KEY): |
|
161
|
|
|
self._end_header = True |
|
162
|
|
|
if len(self._header) == 0: |
|
163
|
|
|
self.err("No header found", numline=self._numline) |
|
164
|
|
|
return -1 |
|
165
|
|
|
return 0 |
|
166
|
|
|
|
|
167
|
|
|
splitted = [token.strip() for token in line.split(',')] |
|
168
|
|
|
|
|
169
|
|
|
# Batch Info,2013-03-20T07:11:09.9053262-07:00,2013-03-20T07:12:55.5280967-07:00,2013-03-20T07:11:07.1047817-07:00,,,,,,,,,,,,,, |
|
170
|
|
|
if splitted[0] == self.HEADERKEY_BATCHINFO: |
|
171
|
|
|
if self.HEADERKEY_BATCHINFO in self._header: |
|
172
|
|
|
self.warn("Header Batch Info already found. Discarding", |
|
173
|
|
|
numline=self._numline, line=line) |
|
174
|
|
|
return 0 |
|
175
|
|
|
|
|
176
|
|
|
self._header[self.HEADERKEY_BATCHINFO] = [] |
|
177
|
|
|
for i in range(len(splitted) - 1): |
|
178
|
|
|
if splitted[i + 1]: |
|
179
|
|
|
self._header[self.HEADERKEY_BATCHINFO].append(splitted[i + 1]) |
|
180
|
|
|
|
|
181
|
|
|
# Batch Data Path,D:\MassHunter\Data\130129\QuantResults\130129LS.batch.bin,,,,,,,,,,,,,,,, |
|
182
|
|
|
elif splitted[0] == self.HEADERKEY_BATCHDATAPATH: |
|
183
|
|
|
if self.HEADERKEY_BATCHDATAPATH in self._header: |
|
184
|
|
|
self.warn("Header Batch Data Path already found. Discarding", |
|
185
|
|
|
numline=self._numline, line=line) |
|
186
|
|
|
return 0; |
|
187
|
|
|
|
|
188
|
|
|
if splitted[1]: |
|
189
|
|
|
self._header[self.HEADERKEY_BATCHDATAPATH] = splitted[1] |
|
190
|
|
|
else: |
|
191
|
|
|
self.warn("Batch Data Path not found or empty", |
|
192
|
|
|
numline=self._numline, line=line) |
|
193
|
|
|
|
|
194
|
|
|
# Analysis Time,3/20/2013 7:11 AM,Analyst Name,Administrator,,,,,,,,,,,,,, |
|
195
|
|
|
elif splitted[0] == self.HEADERKEY_ANALYSISTIME: |
|
196
|
|
View Code Duplication |
if splitted[1]: |
|
|
|
|
|
|
197
|
|
|
try: |
|
198
|
|
|
d = datetime.strptime(splitted[1], "%m/%d/%Y %I:%M %p") |
|
199
|
|
|
self._header[self.HEADERKEY_ANALYSISTIME] = d |
|
200
|
|
|
except ValueError: |
|
201
|
|
|
self.err("Invalid Analysis Time format", |
|
202
|
|
|
numline=self._numline, line=line) |
|
203
|
|
|
else: |
|
204
|
|
|
self.warn("Analysis Time not found or empty", |
|
205
|
|
|
numline=self._numline, line=line) |
|
206
|
|
|
|
|
207
|
|
|
if splitted[2] and splitted[2] == self.HEADERKEY_ANALYSTNAME: |
|
208
|
|
|
if splitted[3]: |
|
209
|
|
|
self._header[self.HEADERKEY_ANALYSTNAME] = splitted[3] |
|
210
|
|
|
else: |
|
211
|
|
|
self.warn("Analyst Name not found or empty", |
|
212
|
|
|
numline=self._numline, line=line) |
|
213
|
|
|
else: |
|
214
|
|
|
self.err("Analyst Name not found", |
|
215
|
|
|
numline=self._numline, line=line) |
|
216
|
|
|
|
|
217
|
|
|
# Report Time,3/20/2013 7:12 AM,Reporter Name,Administrator,,,,,,,,,,,,,, |
|
218
|
|
|
elif splitted[0] == self.HEADERKEY_REPORTTIME: |
|
219
|
|
|
if splitted[1]: |
|
220
|
|
|
try: |
|
221
|
|
|
d = datetime.strptime(splitted[1], "%m/%d/%Y %I:%M %p") |
|
222
|
|
|
self._header[self.HEADERKEY_REPORTTIME] = d |
|
223
|
|
|
except ValueError: |
|
224
|
|
|
self.err("Invalid Report Time format", |
|
225
|
|
|
numline=self._numline, line=line) |
|
226
|
|
|
else: |
|
227
|
|
|
self.warn("Report time not found or empty", |
|
228
|
|
|
numline=self._numline, line=line) |
|
229
|
|
|
|
|
230
|
|
|
|
|
231
|
|
|
if splitted[2] and splitted[2] == self.HEADERKEY_REPORTERNAME: |
|
232
|
|
|
if splitted[3]: |
|
233
|
|
|
self._header[self.HEADERKEY_REPORTERNAME] = splitted[3] |
|
234
|
|
|
else: |
|
235
|
|
|
self.warn("Reporter Name not found or empty", |
|
236
|
|
|
numline=self._numline, line=line) |
|
237
|
|
|
|
|
238
|
|
|
else: |
|
239
|
|
|
self.err("Reporter Name not found", |
|
240
|
|
|
numline=self._numline, line=line) |
|
241
|
|
|
|
|
242
|
|
|
|
|
243
|
|
|
# Last Calib Update,3/20/2013 7:11 AM,Batch State,Processed,,,,,,,,,,,,,, |
|
244
|
|
|
elif splitted[0] == self.HEADERKEY_LASTCALIBRATION: |
|
245
|
|
|
if splitted[1]: |
|
246
|
|
|
try: |
|
247
|
|
|
d = datetime.strptime(splitted[1], "%m/%d/%Y %I:%M %p") |
|
248
|
|
|
self._header[self.HEADERKEY_LASTCALIBRATION] = d |
|
249
|
|
|
except ValueError: |
|
250
|
|
|
self.err("Invalid Last Calibration time format", |
|
251
|
|
|
numline=self._numline, line=line) |
|
252
|
|
|
|
|
253
|
|
|
else: |
|
254
|
|
|
self.warn("Last Calibration time not found or empty", |
|
255
|
|
|
numline=self._numline, line=line) |
|
256
|
|
|
|
|
257
|
|
|
|
|
258
|
|
|
if splitted[2] and splitted[2] == self.HEADERKEY_BATCHSTATE: |
|
259
|
|
|
if splitted[3]: |
|
260
|
|
|
self._header[self.HEADERKEY_BATCHSTATE] = splitted[3] |
|
261
|
|
|
else: |
|
262
|
|
|
self.warn("Batch state not found or empty", |
|
263
|
|
|
numline=self._numline, line=line) |
|
264
|
|
|
|
|
265
|
|
|
else: |
|
266
|
|
|
self.err("Batch state not found", |
|
267
|
|
|
numline=self._numline, line=line) |
|
268
|
|
|
|
|
269
|
|
|
|
|
270
|
|
|
return 0 |
|
271
|
|
|
|
|
272
|
|
|
def parse_sequencetableline(self, line): |
|
273
|
|
|
""" Parses sequence table lines |
|
274
|
|
|
|
|
275
|
|
|
Sequence Table example: |
|
276
|
|
|
Sequence Table,,,,,,,,,,,,,,,,, |
|
277
|
|
|
Data File,Sample Name,Position,Inj Vol,Level,Sample Type,Acq Method File,,,,,,,,,,, |
|
278
|
|
|
prerunrespchk.d,prerunrespchk,Vial 3,-1.00,,Sample,120824_VitD_MAPTAD_1D_MRM_practice.m,,,,,,,,,,, |
|
279
|
|
|
DSS_Nist_L1.d,DSS_Nist_L1,P1-A2,-1.00,,Sample,120824_VitD_MAPTAD_1D_MRM_practice.m,,,,,,,,,,, |
|
280
|
|
|
DSS_Nist_L2.d,DSS_Nist_L2,P1-B2,-1.00,,Sample,120824_VitD_MAPTAD_1D_MRM_practice.m,,,,,,,,,,, |
|
281
|
|
|
DSS_Nist_L3.d,DSS_Nist_L3,P1-C2,-1.00,,Sample,120824_VitD_MAPTAD_1D_MRM_practice.m,,,,,,,,,,, |
|
282
|
|
|
UTAK_DS_L1.d,UTAK_DS_L1,P1-D2,-1.00,,Sample,120824_VitD_MAPTAD_1D_MRM_practice.m,,,,,,,,,,, |
|
283
|
|
|
UTAK_DS_L2.d,UTAK_DS_L2,P1-E2,-1.00,,Sample,120824_VitD_MAPTAD_1D_MRM_practice.m,,,,,,,,,,, |
|
284
|
|
|
mid_respchk.d,mid_respchk,Vial 3,-1.00,,Sample,120824_VitD_MAPTAD_1D_MRM_practice.m,,,,,,,,,,, |
|
285
|
|
|
UTAK_DS_low.d,UTAK_DS_Low,P1-F2,-1.00,,Sample,120824_VitD_MAPTAD_1D_MRM_practice.m,,,,,,,,,,, |
|
286
|
|
|
FDBS_31.d,FDBS_31,P1-G2,-1.00,,Sample,120824_VitD_MAPTAD_1D_MRM_practice.m,,,,,,,,,,, |
|
287
|
|
|
FDBS_32.d,FDBS_32,P1-H2,-1.00,,Sample,120824_VitD_MAPTAD_1D_MRM_practice.m,,,,,,,,,,, |
|
288
|
|
|
LS_60-r001.d,LS_60,P1-G12,-1.00,,Sample,120824_VitD_MAPTAD_1D_MRM_practice.m,,,,,,,,,,, |
|
289
|
|
|
LS_60-r002.d,LS_60,P1-G12,-1.00,,Sample,120824_VitD_MAPTAD_1D_MRM_practice.m,,,,,,,,,,, |
|
290
|
|
|
LS_61-r001.d,LS_61,P1-H12,-1.00,,Sample,120824_VitD_MAPTAD_1D_MRM_practice.m,,,,,,,,,,, |
|
291
|
|
|
LS_61-r002.d,LS_61,P1-H12,-1.00,,Sample,120824_VitD_MAPTAD_1D_MRM_practice.m,,,,,,,,,,, |
|
292
|
|
|
post_respchk.d,post_respchk,Vial 3,-1.00,,Sample,120824_VitD_MAPTAD_1D_MRM_practice.m,,,,,,,,,,, |
|
293
|
|
|
,,,,,,,,,,,,,,,,, |
|
294
|
|
|
""" |
|
295
|
|
|
|
|
296
|
|
|
# Sequence Table,,,,,,,,,,,,,,,,, |
|
297
|
|
|
# prerunrespchk.d,prerunrespchk,Vial 3,-1.00,,Sample,120824_VitD_MAPTAD_1D_MRM_practice.m,,,,,,,,,,, |
|
298
|
|
|
# mid_respchk.d,mid_respchk,Vial 3,-1.00,,Sample,120824_VitD_MAPTAD_1D_MRM_practice.m,,,,,,,,,,, |
|
299
|
|
|
# ,,,,,,,,,,,,,,,,, |
|
300
|
|
|
if line.startswith(self.SEQUENCETABLE_KEY) \ |
|
301
|
|
|
or line.startswith(self.SEQUENCETABLE_PRERUN) \ |
|
302
|
|
|
or line.startswith(self.SEQUENCETABLE_MIDRUN) \ |
|
303
|
|
|
or self._end_sequencetable == True: |
|
304
|
|
|
|
|
305
|
|
|
# Nothing to do, continue |
|
306
|
|
|
return 0 |
|
307
|
|
|
|
|
308
|
|
|
# Data File,Sample Name,Position,Inj Vol,Level,Sample Type,Acq Method File,,,,,,,,,,, |
|
309
|
|
|
if line.startswith(self.SEQUENCETABLE_HEADER_DATAFILE): |
|
310
|
|
|
self._sequencesheader = [token.strip() for token in line.split(',') if token.strip()] |
|
311
|
|
|
return 0 |
|
312
|
|
|
|
|
313
|
|
|
# post_respchk.d,post_respchk,Vial 3,-1.00,,Sample,120824_VitD_MAPTAD_1D_MRM_practice.m,,,,,,,,,,, |
|
314
|
|
|
# Quantitation Results,,,,,,,,,,,,,,,,, |
|
315
|
|
|
if line.startswith(self.SEQUENCETABLE_POSTRUN) \ |
|
316
|
|
|
or line.startswith(self.QUANTITATIONRESULTS_KEY) \ |
|
317
|
|
|
or line.startswith(self.COMMAS): |
|
318
|
|
|
self._end_sequencetable = True |
|
319
|
|
|
if len(self._sequences) == 0: |
|
320
|
|
|
self.err("No Sequence Table found", linenum=self._numline) |
|
321
|
|
|
return -1 |
|
322
|
|
|
|
|
323
|
|
|
# Jumps 2 lines: |
|
324
|
|
|
# Data File,Sample Name,Position,Inj Vol,Level,Sample Type,Acq Method File,,,,,,,,,,, |
|
325
|
|
|
# prerunrespchk.d,prerunrespchk,Vial 3,-1.00,,Sample,120824_VitD_MAPTAD_1D_MRM_practice.m,,,,,,,,,,, |
|
326
|
|
|
return 2 |
|
327
|
|
|
|
|
328
|
|
|
# DSS_Nist_L1.d,DSS_Nist_L1,P1-A2,-1.00,,Sample,120824_VitD_MAPTAD_1D_MRM_practice.m,,,,,,,,,,, |
|
329
|
|
|
splitted = [token.strip() for token in line.split(',')] |
|
330
|
|
|
sequence = {} |
|
331
|
|
|
for colname in self._sequencesheader: |
|
332
|
|
|
sequence[colname] = '' |
|
333
|
|
|
|
|
334
|
|
|
for i in range(len(splitted)): |
|
335
|
|
|
token = splitted[i] |
|
336
|
|
|
if i < len(self._sequencesheader): |
|
337
|
|
|
colname = self._sequencesheader[i] |
|
338
|
|
|
if token and colname in self.SEQUENCETABLE_NUMERICHEADERS: |
|
339
|
|
|
try: |
|
340
|
|
|
sequence[colname] = float(token) |
|
341
|
|
|
except ValueError: |
|
342
|
|
|
self.warn( |
|
343
|
|
|
"No valid number ${token} in column ${index} (${column_name})", |
|
344
|
|
|
mapping={"token": token, |
|
345
|
|
|
"index": str(i + 1), |
|
346
|
|
|
"column_name": colname}, |
|
347
|
|
|
numline=self._numline, line=line) |
|
348
|
|
|
sequence[colname] = token |
|
349
|
|
|
else: |
|
350
|
|
|
sequence[colname] = token |
|
351
|
|
|
elif token: |
|
352
|
|
|
self.err("Orphan value in column ${index} (${token})", |
|
353
|
|
|
mapping={"index": str(i+1), |
|
354
|
|
|
"token": token}, |
|
355
|
|
|
numline=self._numline, line=line) |
|
356
|
|
|
self._sequences.append(sequence) |
|
357
|
|
|
|
|
358
|
|
|
def parse_quantitationesultsline(self, line): |
|
359
|
|
|
""" Parses quantitation result lines |
|
360
|
|
|
|
|
361
|
|
|
Quantitation results example: |
|
362
|
|
|
Quantitation Results,,,,,,,,,,,,,,,,, |
|
363
|
|
|
Target Compound,25-OH D3+PTAD+MA,,,,,,,,,,,,,,,, |
|
364
|
|
|
Data File,Compound,ISTD,Resp,ISTD Resp,Resp Ratio, Final Conc,Exp Conc,Accuracy,,,,,,,,, |
|
365
|
|
|
prerunrespchk.d,25-OH D3+PTAD+MA,25-OH D3d3+PTAD+MA,5816,274638,0.0212,0.9145,,,,,,,,,,, |
|
366
|
|
|
DSS_Nist_L1.d,25-OH D3+PTAD+MA,25-OH D3d3+PTAD+MA,6103,139562,0.0437,1.6912,,,,,,,,,,, |
|
367
|
|
|
DSS_Nist_L2.d,25-OH D3+PTAD+MA,25-OH D3d3+PTAD+MA,11339,135726,0.0835,3.0510,,,,,,,,,,, |
|
368
|
|
|
DSS_Nist_L3.d,25-OH D3+PTAD+MA,25-OH D3d3+PTAD+MA,15871,141710,0.1120,4.0144,,,,,,,,,,, |
|
369
|
|
|
mid_respchk.d,25-OH D3+PTAD+MA,25-OH D3d3+PTAD+MA,4699,242798,0.0194,0.8514,,,,,,,,,,, |
|
370
|
|
|
DSS_Nist_L3-r002.d,25-OH D3+PTAD+MA,25-OH D3d3+PTAD+MA,15659,129490,0.1209,4.3157,,,,,,,,,,, |
|
371
|
|
|
UTAK_DS_L1-r001.d,25-OH D3+PTAD+MA,25-OH D3d3+PTAD+MA,29846,132264,0.2257,7.7965,,,,,,,,,,, |
|
372
|
|
|
UTAK_DS_L1-r002.d,25-OH D3+PTAD+MA,25-OH D3d3+PTAD+MA,28696,141614,0.2026,7.0387,,,,,,,,,,, |
|
373
|
|
|
post_respchk.d,25-OH D3+PTAD+MA,25-OH D3d3+PTAD+MA,5022,231748,0.0217,0.9315,,,,,,,,,,, |
|
374
|
|
|
,,,,,,,,,,,,,,,,, |
|
375
|
|
|
Target Compound,25-OH D2+PTAD+MA,,,,,,,,,,,,,,,, |
|
376
|
|
|
Data File,Compound,ISTD,Resp,ISTD Resp,Resp Ratio, Final Conc,Exp Conc,Accuracy,,,,,,,,, |
|
377
|
|
|
prerunrespchk.d,25-OH D2+PTAD+MA,25-OH D3d3+PTAD+MA,6222,274638,0.0227,0.8835,,,,,,,,,,, |
|
378
|
|
|
DSS_Nist_L1.d,25-OH D2+PTAD+MA,25-OH D3d3+PTAD+MA,1252,139562,0.0090,0.7909,,,,,,,,,,, |
|
379
|
|
|
DSS_Nist_L2.d,25-OH D2+PTAD+MA,25-OH D3d3+PTAD+MA,3937,135726,0.0290,0.9265,,,,,,,,,,, |
|
380
|
|
|
DSS_Nist_L3.d,25-OH D2+PTAD+MA,25-OH D3d3+PTAD+MA,826,141710,0.0058,0.7697,,,,,,,,,,, |
|
381
|
|
|
mid_respchk.d,25-OH D2+PTAD+MA,25-OH D3d3+PTAD+MA,7864,242798,0.0324,0.9493,,,,,,,,,,, |
|
382
|
|
|
DSS_Nist_L3-r002.d,25-OH D2+PTAD+MA,25-OH D3d3+PTAD+MA,853,129490,0.0066,0.7748,,,,,,,,,,, |
|
383
|
|
|
UTAK_DS_L1-r001.d,25-OH D2+PTAD+MA,25-OH D3d3+PTAD+MA,127496,132264,0.9639,7.1558,,,,,,,,,,, |
|
384
|
|
|
UTAK_DS_L1-r002.d,25-OH D2+PTAD+MA,25-OH D3d3+PTAD+MA,135738,141614,0.9585,7.1201,,,,,,,,,,, |
|
385
|
|
|
post_respchk.d,25-OH D2+PTAD+MA,25-OH D3d3+PTAD+MA,6567,231748,0.0283,0.9219,,,,,,,,,,, |
|
386
|
|
|
,,,,,,,,,,,,,,,,, |
|
387
|
|
|
""" |
|
388
|
|
|
|
|
389
|
|
|
# Quantitation Results,,,,,,,,,,,,,,,,, |
|
390
|
|
|
# prerunrespchk.d,25-OH D3+PTAD+MA,25-OH D3d3+PTAD+MA,5816,274638,0.0212,0.9145,,,,,,,,,,, |
|
391
|
|
|
# mid_respchk.d,25-OH D3+PTAD+MA,25-OH D3d3+PTAD+MA,4699,242798,0.0194,0.8514,,,,,,,,,,, |
|
392
|
|
|
# post_respchk.d,25-OH D2+PTAD+MA,25-OH D3d3+PTAD+MA,6567,231748,0.0283,0.9219,,,,,,,,,,, |
|
393
|
|
|
# ,,,,,,,,,,,,,,,,, |
|
394
|
|
|
if line.startswith(self.QUANTITATIONRESULTS_KEY) \ |
|
395
|
|
|
or line.startswith(self.QUANTITATIONRESULTS_PRERUN) \ |
|
396
|
|
|
or line.startswith(self.QUANTITATIONRESULTS_MIDRUN) \ |
|
397
|
|
|
or line.startswith(self.QUANTITATIONRESULTS_POSTRUN) \ |
|
398
|
|
|
or line.startswith(self.COMMAS): |
|
399
|
|
|
|
|
400
|
|
|
# Nothing to do, continue |
|
401
|
|
|
return 0 |
|
402
|
|
|
|
|
403
|
|
|
# Data File,Compound,ISTD,Resp,ISTD Resp,Resp Ratio, Final Conc,Exp Conc,Accuracy,,,,,,,,, |
|
404
|
|
|
if line.startswith(self.QUANTITATIONRESULTS_HEADER_DATAFILE): |
|
405
|
|
|
self._quantitationresultsheader = [token.strip() for token in line.split(',') if token.strip()] |
|
406
|
|
|
return 0 |
|
407
|
|
|
|
|
408
|
|
|
# Target Compound,25-OH D3+PTAD+MA,,,,,,,,,,,,,,,, |
|
409
|
|
|
if line.startswith(self.QUANTITATIONRESULTS_TARGETCOMPOUND): |
|
410
|
|
|
# New set of Quantitation Results |
|
411
|
|
|
splitted = [token.strip() for token in line.split(',')] |
|
412
|
|
|
if not splitted[1]: |
|
413
|
|
|
self.warn("No Target Compound found", |
|
414
|
|
|
numline=self._numline, line=line) |
|
415
|
|
|
return 0 |
|
416
|
|
|
|
|
417
|
|
|
# DSS_Nist_L1.d,25-OH D2+PTAD+MA,25-OH D3d3+PTAD+MA,1252,139562,0.0090,0.7909,,,,,,,,,,, |
|
418
|
|
|
splitted = [token.strip() for token in line.split(',')] |
|
419
|
|
|
quantitation = {} |
|
420
|
|
|
for colname in self._quantitationresultsheader: |
|
421
|
|
|
quantitation[colname] = '' |
|
422
|
|
|
|
|
423
|
|
|
for i in range(len(splitted)): |
|
424
|
|
|
token = splitted[i] |
|
425
|
|
View Code Duplication |
if i < len(self._quantitationresultsheader): |
|
|
|
|
|
|
426
|
|
|
colname = self._quantitationresultsheader[i] |
|
427
|
|
|
if token and colname in self.QUANTITATIONRESULTS_NUMERICHEADERS: |
|
428
|
|
|
try: |
|
429
|
|
|
quantitation[colname] = float(token) |
|
430
|
|
|
except ValueError: |
|
431
|
|
|
self.warn( |
|
432
|
|
|
"No valid number ${token} in column ${index} (${column_name})", |
|
433
|
|
|
mapping={"token": token, |
|
434
|
|
|
"index": str(i + 1), |
|
435
|
|
|
"column_name": colname}, |
|
436
|
|
|
numline=self._numline, line=line) |
|
437
|
|
|
quantitation[colname] = token |
|
438
|
|
|
else: |
|
439
|
|
|
quantitation[colname] = token |
|
440
|
|
|
elif token: |
|
441
|
|
|
self.err("Orphan value in column ${index} (${token})", |
|
442
|
|
|
mapping={"index": str(i+1), |
|
443
|
|
|
"token": token}, |
|
444
|
|
|
numline=self._numline, line=line) |
|
445
|
|
|
|
|
446
|
|
|
if self.QUANTITATIONRESULTS_COMPOUNDCOLUMN in quantitation: |
|
447
|
|
|
compound = quantitation[self.QUANTITATIONRESULTS_COMPOUNDCOLUMN] |
|
448
|
|
|
|
|
449
|
|
|
# Look for sequence matches and populate rawdata |
|
450
|
|
|
datafile = quantitation.get(self.QUANTITATIONRESULTS_HEADER_DATAFILE, '') |
|
451
|
|
|
if not datafile: |
|
452
|
|
|
self.err("No Data File found for quantitation result", |
|
453
|
|
|
numline=self._numline, line=line) |
|
454
|
|
|
|
|
455
|
|
|
else: |
|
456
|
|
|
seqs = [sequence for sequence in self._sequences \ |
|
457
|
|
|
if sequence.get('Data File', '') == datafile] |
|
458
|
|
|
if len(seqs) == 0: |
|
459
|
|
|
self.err("No sample found for quantitative result ${data_file}", |
|
460
|
|
|
mapping={"data_file": datafile}, |
|
461
|
|
|
numline=self._numline, line=line) |
|
462
|
|
|
elif len(seqs) > 1: |
|
463
|
|
|
self.err("More than one sequence found for quantitative result: ${data_file}", |
|
464
|
|
|
mapping={"data_file": datafile}, |
|
465
|
|
|
numline=self._numline, line=line) |
|
466
|
|
|
else: |
|
467
|
|
|
objid = seqs[0].get(self.SEQUENCETABLE_HEADER_SAMPLENAME, '') |
|
468
|
|
|
if objid: |
|
469
|
|
|
quantitation['DefaultResult'] = 'Final Conc' |
|
470
|
|
|
quantitation['Remarks'] = _("Autoimport") |
|
471
|
|
|
rows = self.getRawResults().get(objid, []) |
|
472
|
|
|
raw = rows[0] if len(rows) > 0 else {} |
|
473
|
|
|
raw[compound] = quantitation |
|
474
|
|
|
self._addRawResult(objid, raw, True) |
|
475
|
|
|
else: |
|
476
|
|
|
self.err("No valid sequence for ${data_file}", |
|
477
|
|
|
mapping={"data_file": datafile}, |
|
478
|
|
|
numline=self._numline, line=line) |
|
479
|
|
|
else: |
|
480
|
|
|
self.err("Value for column '${column}' not found", |
|
481
|
|
|
mapping={"column": self.QUANTITATIONRESULTS_COMPOUNDCOLUMN}, |
|
482
|
|
|
numline=self._numline, line=line) |
|
483
|
|
|
|
|
484
|
|
|
|
|
485
|
|
|
class MasshunterQuantImporter(AnalysisResultsImporter): |
|
486
|
|
|
|
|
487
|
|
|
def __init__(self, parser, context, override, |
|
488
|
|
|
allowed_ar_states=None, allowed_analysis_states=None, |
|
489
|
|
|
instrument_uid=''): |
|
490
|
|
|
AnalysisResultsImporter.__init__(self, parser, context, |
|
491
|
|
|
override, allowed_ar_states, |
|
492
|
|
|
allowed_analysis_states, |
|
493
|
|
|
instrument_uid) |
|
494
|
|
|
|