@@ -49,11 +49,11 @@ |
||
49 | 49 | |
50 | 50 | // Adding options to help archive: |
51 | 51 | /** |
52 | - * We removed the "proc" option as it seemd not to be working any more. But as the complete handling of the crawler has changed regarding the configuration |
|
53 | - * this is completely ok. Since configuration records were introduced to configure "what should be done" additionally to page ts the way to setup jobs |
|
54 | - * has drifted from selecting filtering processing instructions to selecting/filtering configuration keys (you can configure the processing instructions there). |
|
55 | - * This is also reflected in the backend modules and allows you a much clearer and powerful way to work with the crawler extension. |
|
56 | - */ |
|
52 | + * We removed the "proc" option as it seemd not to be working any more. But as the complete handling of the crawler has changed regarding the configuration |
|
53 | + * this is completely ok. Since configuration records were introduced to configure "what should be done" additionally to page ts the way to setup jobs |
|
54 | + * has drifted from selecting filtering processing instructions to selecting/filtering configuration keys (you can configure the processing instructions there). |
|
55 | + * This is also reflected in the backend modules and allows you a much clearer and powerful way to work with the crawler extension. |
|
56 | + */ |
|
57 | 57 | // $this->cli_options[] = array('-proc listOfProcInstr', 'Comma list of processing instructions. These are the "actions" carried out when crawling and you must specify at least one. Depends on third-party extensions. Examples are "tx_cachemgm_recache" from "cachemgm" extension (will recache pages), "tx_staticpub_publish" from "staticpub" (publishing pages to static files) or "tx_indexedsearch_reindex" from "indexed_search" (indexes pages).'); |
58 | 58 | // TODO: cleanup here! |
59 | 59 | $this->cli_options[] = ['-d depth', 'Tree depth, 0-99', "How many levels under the 'page_id' to include."]; |
@@ -59,7 +59,7 @@ discard block |
||
59 | 59 | $this->cli_options[] = ['-d depth', 'Tree depth, 0-99', "How many levels under the 'page_id' to include."]; |
60 | 60 | $this->cli_options[] = ['-o mode', 'Output mode: "url", "exec", "queue"', "Specifies output modes\nurl : Will list URLs which wget could use as input.\nqueue: Will put entries in queue table.\nexec: Will execute all entries right away!"]; |
61 | 61 | $this->cli_options[] = ['-n number', 'Number of items per minute.', 'Specifies how many items are put in the queue per minute. Only valid for output mode "queue"']; |
62 | - $this->cli_options[] = ['-conf configurationkeys','List of Configuration Keys','A commaseperated list of crawler configurations']; |
|
62 | + $this->cli_options[] = ['-conf configurationkeys', 'List of Configuration Keys', 'A commaseperated list of crawler configurations']; |
|
63 | 63 | # $this->cli_options[] = array('-v level', 'Verbosity level 0-3', "The value of level can be:\n 0 = all output\n 1 = info and greater (default)\n 2 = warnings and greater\n 3 = errors"); |
64 | 64 | |
65 | 65 | // Setting help texts: |
@@ -67,7 +67,7 @@ discard block |
||
67 | 67 | $this->cli_help['synopsis'] = 'page_id ###OPTIONS###'; |
68 | 68 | $this->cli_help['description'] = "Works as a CLI interface to some functionality from the Web > Info > Site Crawler module; It can put entries in the queue from command line options, return the list of URLs and even execute all entries right away without having to queue them up - this can be useful for immediate re-cache, re-indexing or static publishing from command line."; |
69 | 69 | $this->cli_help['examples'] = "/.../cli_dispatch.phpsh crawler_im 7 -d=2 -conf=<configurationKey> -o=exec\nWill re-cache pages from page 7 and two levels down, executed immediately.\n"; |
70 | - $this->cli_help['examples'].= "/.../cli_dispatch.phpsh crawler_im 7 -d=0 -conf=<configurationKey> -n=4 -o=queue\nWill put entries for re-caching pages from page 7 into queue, 4 every minute.\n"; |
|
70 | + $this->cli_help['examples'] .= "/.../cli_dispatch.phpsh crawler_im 7 -d=0 -conf=<configurationKey> -n=4 -o=queue\nWill put entries for re-caching pages from page 7 into queue, 4 every minute.\n"; |
|
71 | 71 | $this->cli_help['author'] = 'Kasper Skaarhoj, Daniel Poetzinger, Fabrizio Branca, Tolleiv Nietsch, Timo Schmidt - AOE media 2009'; |
72 | 72 | } |
73 | 73 | } |
@@ -9,5 +9,5 @@ |
||
9 | 9 | try { |
10 | 10 | $processManager->multiProcess($timeout); |
11 | 11 | } catch (Exception $e) { |
12 | - echo PHP_EOL . $e->getMessage(); |
|
12 | + echo PHP_EOL.$e->getMessage(); |
|
13 | 13 | } |
@@ -3,12 +3,12 @@ |
||
3 | 3 | } ?> |
4 | 4 | |
5 | 5 | Page: |
6 | -<?php for ($currentPageOffset = 0; $currentPageOffset < $this->getTotalPagesCount(); $currentPageOffset++) { |
|
6 | +<?php for ($currentPageOffset = 0; $currentPageOffset < $this->getTotalPagesCount(); $currentPageOffset++) { |
|
7 | 7 | ?> |
8 | 8 | <a href="index.php?offset=<?php echo htmlspecialchars($currentPageOffset * $this->getPerPage()); ?>"> |
9 | 9 | <?php echo htmlspecialchars($this->getLabelForPageOffset($currentPageOffset)); ?> |
10 | 10 | </a> |
11 | - <?php if ($currentPageOffset+1 < $this->getTotalPagesCount()) { |
|
11 | + <?php if ($currentPageOffset + 1 < $this->getTotalPagesCount()) { |
|
12 | 12 | ?> |
13 | 13 | | |
14 | 14 | <?php |
@@ -50,8 +50,8 @@ discard block |
||
50 | 50 | */ |
51 | 51 | public function offsetGet($index) |
52 | 52 | { |
53 | - if (! parent::offsetExists($index)) { |
|
54 | - throw new Exception('Index "' . var_export($index, true) . '" for tx_crawler_domain_process are not available'); |
|
53 | + if (!parent::offsetExists($index)) { |
|
54 | + throw new Exception('Index "'.var_export($index, true).'" for tx_crawler_domain_process are not available'); |
|
55 | 55 | } |
56 | 56 | return parent::offsetGet($index); |
57 | 57 | } |
@@ -66,7 +66,7 @@ discard block |
||
66 | 66 | */ |
67 | 67 | public function offsetSet($index, $subject) |
68 | 68 | { |
69 | - if (! $subject instanceof tx_crawler_domain_process) { |
|
69 | + if (!$subject instanceof tx_crawler_domain_process) { |
|
70 | 70 | throw new InvalidArgumentException('Wrong parameter type given, "tx_crawler_domain_process" expected!'); |
71 | 71 | } |
72 | 72 | parent::offsetSet($index, $subject); |
@@ -80,7 +80,7 @@ discard block |
||
80 | 80 | */ |
81 | 81 | public function append($subject) |
82 | 82 | { |
83 | - if (! $subject instanceof tx_crawler_domain_process) { |
|
83 | + if (!$subject instanceof tx_crawler_domain_process) { |
|
84 | 84 | throw new InvalidArgumentException('Wrong parameter type given, "tx_crawler_domain_process" expected!'); |
85 | 85 | } |
86 | 86 | parent::append($subject); |
@@ -92,9 +92,9 @@ discard block |
||
92 | 92 | */ |
93 | 93 | public function getProcessIds() |
94 | 94 | { |
95 | - $result=[]; |
|
95 | + $result = []; |
|
96 | 96 | foreach ($this->getIterator() as $value) { |
97 | - $result[]=$value->getProcess_id(); |
|
97 | + $result[] = $value->getProcess_id(); |
|
98 | 98 | } |
99 | 99 | return $result; |
100 | 100 | } |
@@ -67,8 +67,8 @@ discard block |
||
67 | 67 | */ |
68 | 68 | public function __construct() |
69 | 69 | { |
70 | - $this->processRepository = new tx_crawler_domain_process_repository(); |
|
71 | - $this->queueRepository = new tx_crawler_domain_queue_repository(); |
|
70 | + $this->processRepository = new tx_crawler_domain_process_repository(); |
|
71 | + $this->queueRepository = new tx_crawler_domain_queue_repository(); |
|
72 | 72 | $this->crawlerObj = \TYPO3\CMS\Core\Utility\GeneralUtility::makeInstance('tx_crawler_lib'); |
73 | 73 | $this->timeToLive = intval($this->crawlerObj->extensionSettings['processMaxRunTime']); |
74 | 74 | $this->countInARun = intval($this->crawlerObj->extensionSettings['countInARun']); |
@@ -84,7 +84,7 @@ discard block |
||
84 | 84 | public function multiProcess($timeout) |
85 | 85 | { |
86 | 86 | if ($this->processLimit <= 1) { |
87 | - throw new RuntimeException('To run crawler in multi process mode you have to configure the processLimit > 1.' . PHP_EOL); |
|
87 | + throw new RuntimeException('To run crawler in multi process mode you have to configure the processLimit > 1.'.PHP_EOL); |
|
88 | 88 | } |
89 | 89 | |
90 | 90 | $pendingItemsStart = $this->queueRepository->countAllPendingItems(); |
@@ -102,7 +102,7 @@ discard block |
||
102 | 102 | } |
103 | 103 | if ($currentPendingItems == 0) { |
104 | 104 | if ($this->verbose) { |
105 | - echo 'Finished...' . chr(10); |
|
105 | + echo 'Finished...'.chr(10); |
|
106 | 106 | } |
107 | 107 | break; |
108 | 108 | } |
@@ -114,16 +114,16 @@ discard block |
||
114 | 114 | } |
115 | 115 | sleep(1); |
116 | 116 | if ($nextTimeOut < time()) { |
117 | - $timedOutProcesses = $this->processRepository->findAll('', 'DESC', null, 0, 'ttl >' . $nextTimeOut); |
|
117 | + $timedOutProcesses = $this->processRepository->findAll('', 'DESC', null, 0, 'ttl >'.$nextTimeOut); |
|
118 | 118 | $nextTimeOut = time() + $this->timeToLive; |
119 | 119 | if ($this->verbose) { |
120 | - echo 'Cleanup' . implode(',', $timedOutProcesses->getProcessIds()) . chr(10); |
|
120 | + echo 'Cleanup'.implode(',', $timedOutProcesses->getProcessIds()).chr(10); |
|
121 | 121 | } |
122 | 122 | $this->crawlerObj->CLI_releaseProcesses($timedOutProcesses->getProcessIds(), true); |
123 | 123 | } |
124 | 124 | } |
125 | 125 | if ($currentPendingItems > 0 && $this->verbose) { |
126 | - echo 'Stop with timeout' . chr(10); |
|
126 | + echo 'Stop with timeout'.chr(10); |
|
127 | 127 | } |
128 | 128 | } |
129 | 129 | |
@@ -143,17 +143,17 @@ discard block |
||
143 | 143 | private function startRequiredProcesses() |
144 | 144 | { |
145 | 145 | $ret = false; |
146 | - $currentProcesses= $this->processRepository->countActive(); |
|
147 | - $availableProcessesCount = $this->processLimit-$currentProcesses; |
|
146 | + $currentProcesses = $this->processRepository->countActive(); |
|
147 | + $availableProcessesCount = $this->processLimit - $currentProcesses; |
|
148 | 148 | $requiredProcessesCount = ceil($this->queueRepository->countAllUnassignedPendingItems() / $this->countInARun); |
149 | - $startProcessCount = min([$availableProcessesCount,$requiredProcessesCount]); |
|
149 | + $startProcessCount = min([$availableProcessesCount, $requiredProcessesCount]); |
|
150 | 150 | if ($startProcessCount <= 0) { |
151 | 151 | return $ret; |
152 | 152 | } |
153 | 153 | if ($startProcessCount && $this->verbose) { |
154 | 154 | echo 'Start '.$startProcessCount.' new processes (Running:'.$currentProcesses.')'; |
155 | 155 | } |
156 | - for ($i=0;$i<$startProcessCount;$i++) { |
|
156 | + for ($i = 0; $i < $startProcessCount; $i++) { |
|
157 | 157 | usleep(100); |
158 | 158 | if ($this->startProcess()) { |
159 | 159 | if ($this->verbose) { |
@@ -174,13 +174,13 @@ discard block |
||
174 | 174 | */ |
175 | 175 | public function startProcess() |
176 | 176 | { |
177 | - $ttl = (time() + $this->timeToLive -1); |
|
177 | + $ttl = (time() + $this->timeToLive - 1); |
|
178 | 178 | $current = $this->processRepository->countNotTimeouted($ttl); |
179 | - $completePath = '(' .escapeshellcmd($this->getCrawlerCliPath()) . ' &) > /dev/null'; |
|
179 | + $completePath = '('.escapeshellcmd($this->getCrawlerCliPath()).' &) > /dev/null'; |
|
180 | 180 | if (system($completePath) === false) { |
181 | 181 | throw new Exception('could not start process!'); |
182 | 182 | } else { |
183 | - for ($i=0;$i<10;$i++) { |
|
183 | + for ($i = 0; $i < 10; $i++) { |
|
184 | 184 | if ($this->processRepository->countNotTimeouted($ttl) > $current) { |
185 | 185 | return true; |
186 | 186 | } |
@@ -197,9 +197,9 @@ discard block |
||
197 | 197 | */ |
198 | 198 | public function getCrawlerCliPath() |
199 | 199 | { |
200 | - $phpPath = $this->crawlerObj->extensionSettings['phpPath'] . ' '; |
|
201 | - $pathToTypo3 = rtrim(\TYPO3\CMS\Core\Utility\GeneralUtility::getIndpEnv('TYPO3_DOCUMENT_ROOT'), '/'); |
|
202 | - $pathToTypo3 .= rtrim(\TYPO3\CMS\Core\Utility\GeneralUtility::getIndpEnv('TYPO3_SITE_PATH'), '/'); |
|
200 | + $phpPath = $this->crawlerObj->extensionSettings['phpPath'].' '; |
|
201 | + $pathToTypo3 = rtrim(\TYPO3\CMS\Core\Utility\GeneralUtility::getIndpEnv('TYPO3_DOCUMENT_ROOT'), '/'); |
|
202 | + $pathToTypo3 .= rtrim(\TYPO3\CMS\Core\Utility\GeneralUtility::getIndpEnv('TYPO3_SITE_PATH'), '/'); |
|
203 | 203 | $cliPart = '/typo3/cli_dispatch.phpsh crawler'; |
204 | 204 | return $phpPath.$pathToTypo3.$cliPart; |
205 | 205 | } |
@@ -130,7 +130,7 @@ |
||
130 | 130 | { |
131 | 131 | if (is_array($this->observers[$event])) { |
132 | 132 | foreach ($this->observers[$event] as $eventObserver) { |
133 | - call_user_func([$eventObserver['object'],$eventObserver['method']], $event, $group, $attachedData); |
|
133 | + call_user_func([$eventObserver['object'], $eventObserver['method']], $event, $group, $attachedData); |
|
134 | 134 | } |
135 | 135 | } |
136 | 136 | } |
@@ -45,36 +45,36 @@ |
||
45 | 45 | { |
46 | 46 | $GLOBALS['TYPO3_CONF_VARS']['SC_OPTIONS']['scheduler']['tasks']['AOE\\Crawler\\Task\\CrawlerQueueTask'] = [ |
47 | 47 | 'extension' => $extKey, |
48 | - 'title' => 'LLL:EXT:' . $extKey . '/Resources/Private/Language/Backend.xlf:crawler_im.name', |
|
49 | - 'description' => 'LLL:EXT:' . $extKey . '/Resources/Private/Language/Backend.xlf:crawler_im.description', |
|
48 | + 'title' => 'LLL:EXT:'.$extKey.'/Resources/Private/Language/Backend.xlf:crawler_im.name', |
|
49 | + 'description' => 'LLL:EXT:'.$extKey.'/Resources/Private/Language/Backend.xlf:crawler_im.description', |
|
50 | 50 | 'additionalFields' => CrawlerQueueTaskAdditionalFieldProvider::class |
51 | 51 | ]; |
52 | 52 | |
53 | 53 | $GLOBALS['TYPO3_CONF_VARS']['SC_OPTIONS']['scheduler']['tasks']['AOE\\Crawler\\Task\\CrawlerTask'] = [ |
54 | 54 | 'extension' => $extKey, |
55 | - 'title' => 'LLL:EXT:' . $extKey . '/Resources/Private/Language/Backend.xlf:crawler_crawl.name', |
|
56 | - 'description' => 'LLL:EXT:' . $extKey . '/Resources/Private/Language/Backend.xlf:crawler_crawl.description', |
|
55 | + 'title' => 'LLL:EXT:'.$extKey.'/Resources/Private/Language/Backend.xlf:crawler_crawl.name', |
|
56 | + 'description' => 'LLL:EXT:'.$extKey.'/Resources/Private/Language/Backend.xlf:crawler_crawl.description', |
|
57 | 57 | 'additionalFields' => CrawlerTaskAdditionalFieldProvider::class |
58 | 58 | ]; |
59 | 59 | |
60 | 60 | $GLOBALS['TYPO3_CONF_VARS']['SC_OPTIONS']['scheduler']['tasks']['AOE\\Crawler\\Task\\CrawlMultiProcessTask'] = [ |
61 | 61 | 'extension' => $extKey, |
62 | - 'title' => 'LLL:EXT:' . $extKey . '/Resources/Private/Language/Backend.xlf:crawler_crawlMultiProcess.name', |
|
63 | - 'description' => 'LLL:EXT:' . $extKey . '/Resources/Private/Language/Backend.xlf:crawler_crawl.description', |
|
62 | + 'title' => 'LLL:EXT:'.$extKey.'/Resources/Private/Language/Backend.xlf:crawler_crawlMultiProcess.name', |
|
63 | + 'description' => 'LLL:EXT:'.$extKey.'/Resources/Private/Language/Backend.xlf:crawler_crawl.description', |
|
64 | 64 | 'additionalFields' => CrawlMultiProcessTaskAdditionalFieldProvider::class |
65 | 65 | ]; |
66 | 66 | |
67 | 67 | $GLOBALS['TYPO3_CONF_VARS']['SC_OPTIONS']['scheduler']['tasks']['AOE\\Crawler\\Task\\FlushQueueTask'] = [ |
68 | 68 | 'extension' => $extKey, |
69 | - 'title' => 'LLL:EXT:' . $extKey . '/Resources/Private/Language/Backend.xlf:crawler_flush.name', |
|
70 | - 'description' => 'LLL:EXT:' . $extKey . '/Resources/Private/Language/Backend.xlf:crawler_flush.description', |
|
69 | + 'title' => 'LLL:EXT:'.$extKey.'/Resources/Private/Language/Backend.xlf:crawler_flush.name', |
|
70 | + 'description' => 'LLL:EXT:'.$extKey.'/Resources/Private/Language/Backend.xlf:crawler_flush.description', |
|
71 | 71 | 'additionalFields' => FlushQueueTaskAdditionalFieldProvider::class |
72 | 72 | ]; |
73 | 73 | |
74 | 74 | $GLOBALS['TYPO3_CONF_VARS']['SC_OPTIONS']['scheduler']['tasks']['AOE\\Crawler\\Task\\ProcessCleanupTask'] = [ |
75 | 75 | 'extension' => $extKey, |
76 | - 'title' => 'LLL:EXT:' . $extKey . '/Resources/Private/Language/Backend.xlf:crawler_processCleanup.name', |
|
77 | - 'description' => 'LLL:EXT:' . $extKey . '/Resources/Private/Language/Backend.xlf:crawler_processCleanup.description', |
|
76 | + 'title' => 'LLL:EXT:'.$extKey.'/Resources/Private/Language/Backend.xlf:crawler_processCleanup.name', |
|
77 | + 'description' => 'LLL:EXT:'.$extKey.'/Resources/Private/Language/Backend.xlf:crawler_processCleanup.description', |
|
78 | 78 | ]; |
79 | 79 | } |
80 | 80 | } |
@@ -45,17 +45,17 @@ |
||
45 | 45 | public static function registerHooks($extKey) |
46 | 46 | { |
47 | 47 | $GLOBALS['TYPO3_CONF_VARS']['SC_OPTIONS']['tslib/class.tslib_fe.php']['connectToDB']['tx_crawler'] = |
48 | - TsfeHook::class . '->fe_init'; |
|
48 | + TsfeHook::class.'->fe_init'; |
|
49 | 49 | $GLOBALS['TYPO3_CONF_VARS']['SC_OPTIONS']['tslib/class.tslib_fe.php']['initFEuser']['tx_crawler'] = |
50 | - TsfeHook::class . '->fe_feuserInit'; |
|
50 | + TsfeHook::class.'->fe_feuserInit'; |
|
51 | 51 | $GLOBALS['TYPO3_CONF_VARS']['SC_OPTIONS']['tslib/class.tslib_fe.php']['isOutputting']['tx_crawler'] = |
52 | - TsfeHook::class . '->fe_isOutputting'; |
|
52 | + TsfeHook::class.'->fe_isOutputting'; |
|
53 | 53 | $GLOBALS['TYPO3_CONF_VARS']['SC_OPTIONS']['tslib/class.tslib_fe.php']['hook_eofe']['tx_crawler'] = |
54 | - TsfeHook::class . '->fe_eofe'; |
|
54 | + TsfeHook::class.'->fe_eofe'; |
|
55 | 55 | |
56 | 56 | // Activating NC Static File Cache hook |
57 | 57 | $GLOBALS['TYPO3_CONF_VARS']['SC_OPTIONS']['nc_staticfilecache/class.tx_ncstaticfilecache.php']['createFile_initializeVariables']['tx_crawler'] = |
58 | - StaticFileCacheCreateUriHook::class . '->initialize'; |
|
58 | + StaticFileCacheCreateUriHook::class.'->initialize'; |
|
59 | 59 | |
60 | 60 | // Activating Crawler cli_hooks |
61 | 61 | $GLOBALS['TYPO3_CONF_VARS']['EXTCONF'][$extKey]['cli_hooks'][] = |
@@ -70,14 +70,14 @@ discard block |
||
70 | 70 | $results = $this->getDatabaseConnection()->exec_SELECTgetRows( |
71 | 71 | 'process_id, system_process_id', |
72 | 72 | 'tx_crawler_process', |
73 | - 'ttl <= ' . intval(time() - $this->extensionSettings['processMaxRunTime'] - 3600) . ' AND active = 1' |
|
73 | + 'ttl <= '.intval(time() - $this->extensionSettings['processMaxRunTime'] - 3600).' AND active = 1' |
|
74 | 74 | ); |
75 | 75 | |
76 | 76 | if (!is_array($results)) { |
77 | 77 | return; |
78 | 78 | } |
79 | 79 | foreach ($results as $result) { |
80 | - $systemProcessId = (int)$result['system_process_id']; |
|
80 | + $systemProcessId = (int) $result['system_process_id']; |
|
81 | 81 | $processId = $result['process_id']; |
82 | 82 | if ($systemProcessId > 1) { |
83 | 83 | if ($this->doProcessStillExists($systemProcessId)) { |
@@ -98,7 +98,7 @@ discard block |
||
98 | 98 | $results = $this->getDatabaseConnection()->exec_SELECTgetRows( |
99 | 99 | 'process_id, system_process_id', |
100 | 100 | 'tx_crawler_process', |
101 | - 'ttl <= ' . intval(time() - $this->extensionSettings['processMaxRunTime']) . ' AND active = 1' |
|
101 | + 'ttl <= '.intval(time() - $this->extensionSettings['processMaxRunTime']).' AND active = 1' |
|
102 | 102 | ); |
103 | 103 | |
104 | 104 | if (!is_array($results)) { |
@@ -106,7 +106,7 @@ discard block |
||
106 | 106 | } |
107 | 107 | foreach ($results as $result) { |
108 | 108 | $processExists = false; |
109 | - $systemProcessId = (int)$result['system_process_id']; |
|
109 | + $systemProcessId = (int) $result['system_process_id']; |
|
110 | 110 | $processId = $result['process_id']; |
111 | 111 | if ($systemProcessId > 1) { |
112 | 112 | $dispatcherProcesses = $this->findDispatcherProcesses(); |
@@ -116,7 +116,7 @@ discard block |
||
116 | 116 | } |
117 | 117 | foreach ($dispatcherProcesses as $process) { |
118 | 118 | $responseArray = $this->createResponseArray($process); |
119 | - if ($systemProcessId === (int)$responseArray[1]) { |
|
119 | + if ($systemProcessId === (int) $responseArray[1]) { |
|
120 | 120 | $processExists = true; |
121 | 121 | }; |
122 | 122 | } |
@@ -138,12 +138,12 @@ discard block |
||
138 | 138 | { |
139 | 139 | $this->getDatabaseConnection()->exec_DELETEquery( |
140 | 140 | 'tx_crawler_process', |
141 | - 'process_id = ' . $this->getDatabaseConnection()->fullQuoteStr($processId, 'tx_crawler_process') |
|
141 | + 'process_id = '.$this->getDatabaseConnection()->fullQuoteStr($processId, 'tx_crawler_process') |
|
142 | 142 | ); |
143 | 143 | |
144 | 144 | $this->getDatabaseConnection()->exec_UPDATEquery( |
145 | 145 | 'tx_crawler_queue', |
146 | - 'process_id = ' . $this->getDatabaseConnection()->fullQuoteStr($processId, 'tx_crawler_queue'), |
|
146 | + 'process_id = '.$this->getDatabaseConnection()->fullQuoteStr($processId, 'tx_crawler_queue'), |
|
147 | 147 | ['process_id' => ''] |
148 | 148 | ); |
149 | 149 | } |
@@ -177,12 +177,12 @@ discard block |
||
177 | 177 | $doProcessStillExists = false; |
178 | 178 | if (!$this->isOsWindows()) { |
179 | 179 | // Not windows |
180 | - if (file_exists('/proc/' . $pid)) { |
|
180 | + if (file_exists('/proc/'.$pid)) { |
|
181 | 181 | $doProcessStillExists = true; |
182 | 182 | } |
183 | 183 | } else { |
184 | 184 | // Windows |
185 | - exec('tasklist | find "' . $pid . '"', $returnArray, $returnValue); |
|
185 | + exec('tasklist | find "'.$pid.'"', $returnArray, $returnValue); |
|
186 | 186 | if (count($returnArray) > 0 && preg_match('/php/i', $returnValue[0])) { |
187 | 187 | $doProcessStillExists = true; |
188 | 188 | } |
@@ -204,7 +204,7 @@ discard block |
||
204 | 204 | posix_kill($pid, 9); |
205 | 205 | } else { |
206 | 206 | // Windows |
207 | - exec('taskkill /PID ' . $pid); |
|
207 | + exec('taskkill /PID '.$pid); |
|
208 | 208 | } |
209 | 209 | } |
210 | 210 |