Completed
Push — 5-2 ( 9a4d58...1354ac )
by Tomas Norre
03:30
created
domain/queue/class.tx_crawler_domain_queue_repository.php 1 patch
Indentation   +1 added lines, -1 removed lines patch added patch discarded remove patch
@@ -66,7 +66,7 @@
 block discarded – undo
66 66
     {
67 67
         $db = $this->getDB();
68 68
         $where = 'process_id_completed='.$db->fullQuoteStr($process->getProcess_id(), $this->tableName).
69
-                 ' AND exec_time > 0 ';
69
+                    ' AND exec_time > 0 ';
70 70
         $limit = 1;
71 71
         $groupby = '';
72 72
 
Please login to merge, or discard this patch.
cli/class.tx_crawler_cli_im.php 1 patch
Indentation   +5 added lines, -5 removed lines patch added patch discarded remove patch
@@ -49,11 +49,11 @@
 block discarded – undo
49 49
 
50 50
         // Adding options to help archive:
51 51
         /**
52
-        * We removed the "proc" option as it seemd not to be working any more. But as the complete handling of the crawler has changed regarding the configuration
53
-        * this is completely ok. Since configuration records were introduced to configure "what should be done" additionally to page ts the way to setup jobs
54
-        * has drifted from selecting filtering processing instructions to selecting/filtering configuration keys (you can configure the processing instructions there).
55
-        * This is also reflected in the backend modules and allows you a much clearer and powerful way to work with the crawler extension.
56
-        */
52
+         * We removed the "proc" option as it seemd not to be working any more. But as the complete handling of the crawler has changed regarding the configuration
53
+         * this is completely ok. Since configuration records were introduced to configure "what should be done" additionally to page ts the way to setup jobs
54
+         * has drifted from selecting filtering processing instructions to selecting/filtering configuration keys (you can configure the processing instructions there).
55
+         * This is also reflected in the backend modules and allows you a much clearer and powerful way to work with the crawler extension.
56
+         */
57 57
         // $this->cli_options[] = array('-proc listOfProcInstr', 'Comma list of processing instructions. These are the "actions" carried out when crawling and you must specify at least one. Depends on third-party extensions. Examples are "tx_cachemgm_recache" from "cachemgm" extension (will recache pages), "tx_staticpub_publish" from "staticpub" (publishing pages to static files) or "tx_indexedsearch_reindex" from "indexed_search" (indexes pages).');
58 58
         // TODO: cleanup here!
59 59
         $this->cli_options[] = ['-d depth', 'Tree depth, 0-99', "How many levels under the 'page_id' to include."];
Please login to merge, or discard this patch.