diff --git a/app/Jobs/Search.php b/app/Jobs/Search.php new file mode 100644 index 0000000000000000000000000000000000000000..1e02491d59a52d52a7b167bf5b9e6f4a1a8e8e93 --- /dev/null +++ b/app/Jobs/Search.php @@ -0,0 +1,341 @@ +<?php + +namespace App\Jobs; + +use App\Jobs\Job; +use Illuminate\Queue\SerializesModels; +use Illuminate\Queue\InteractsWithQueue; +use Illuminate\Contracts\Queue\ShouldQueue; +use Illuminate\Http\Request; +use Redis; +use Log; + +class Search extends Job implements ShouldQueue +{ + use InteractsWithQueue, SerializesModels; + + protected $hash, $host, $port, $name, $getString, $useragent, $fp, $sumaFile; + protected $buffer_length = 8192; + + /** + * Create a new job instance. + * + * @return void + */ + public function __construct($hash, $host, $port, $name, $getString, $useragent, $sumaFile) + { + $this->hash = $hash; + $this->host = $host; + $this->port = $port; + $this->name = $name; + $this->getString = $getString; + $this->useragent = $useragent; + $this->sumaFile = $sumaFile; + } + + /** + * Execute the job. + * + * @return void + */ + public function handle(Request $request) + { + $this->fp = $this->getFreeSocket(); + + if(!$this->fp) + { + $this->disable($this->sumaFile, "Die Suchmaschine " . $this->name . " wurde für 1h deaktiviert, weil keine Verbindung aufgebaut werden konnte"); + }else + { + if($this->writeRequest()) + { + $this->readAnswer(); + } + } + } + + public function disable($sumaFile, $message) + { + $xml = simplexml_load_file($sumaFile); + $xml->xpath("//sumas/suma[@name='" . $this->name . "']")['0']['disabled'] = date(DATE_RFC822, mktime(date("H")+1,date("i"), date("s"), date("m"), date("d"), date("Y"))); + $xml->saveXML($sumaFile); + } + + private function readAnswer () + { + $time = microtime(true); + $headers = ''; + $body = ''; + $length = 0; + + if(!$this->fp) + { + return; + } + + // get headers FIRST + $c = 0; + stream_set_blocking($this->fp, 1); + do + { + // use fgets() not fread(), fgets stops reading at first newline + // or buffer which ever one is reached first + $data = fgets($this->fp, 8192); + // a sincle CRLF indicates end of headers + if ($data === false || $data == "\r\n" || feof($this->fp) ) { + // break BEFORE OUTPUT + break; + } + if( sizeof(($tmp = explode(": ", $data))) === 2 ) + $headers[trim($tmp[0])] = trim($tmp[1]); + $c++; + } + while (true); + + // end of headers + if(sizeof($headers) > 1){ + $bodySize = 0; + if( isset($headers["Transfer-Encoding"]) && $headers["Transfer-Encoding"] === "chunked" ) + { + $body = $this->readChunked(); + + }elseif( isset($headers['Content-Length']) ) + { + $length = trim($headers['Content-Length']); + if(is_numeric($length) && $length >= 1) + $body = $this->readBody($length); + $bodySize = strlen($body); + }else + { + Log::error("Konnte nicht herausfinden, wie ich die Serverantwort von: " . $this->name . " auslesen soll. Header war: " . print_r($headers)); + exit; + } + }else + { + return; + } + + Redis::del($this->host . "." . $this->socketNumber); + if( isset($headers["Content-Encoding"]) && $headers['Content-Encoding'] === "gzip") + { + $body = $this->gunzip($body); + } + Redis::hset('search.' . $this->hash, $this->name, $body); + Redis::expire('search.' . $this->hash, 5); + } + + private function readBody($length) + { + $theData = ''; + $done = false; + stream_set_blocking($this->fp, 0); + $startTime = time(); + $lastTime = $startTime; + while (!feof($this->fp) && !$done && (($startTime + 1) > time()) && $length !== 0) + { + usleep(100); + $theNewData = fgets($this->fp, 8192); + $theData .= $theNewData; + $length -= strlen($theNewData); + $done = (trim($theNewData) === '0'); + + } + return $theData; + } + + private function readChunked() + { + $body = ''; + // read from chunked stream + // loop though the stream + do + { + // NOTE: for chunked encoding to work properly make sure + // there is NOTHING (besides newlines) before the first hexlength + + // get the line which has the length of this chunk (use fgets here) + $line = fgets($this->fp, 8192); + + // if it's only a newline this normally means it's read + // the total amount of data requested minus the newline + // continue to next loop to make sure we're done + if ($line == "\r\n") { + continue; + } + + // the length of the block is sent in hex decode it then loop through + // that much data get the length + // NOTE: hexdec() ignores all non hexadecimal chars it finds + $length = hexdec($line); + + if (!is_int($length)) { + trigger_error('Most likely not chunked encoding', E_USER_ERROR); + } + + // zero is sent when at the end of the chunks + // or the end of the stream or error + if ($line === false || $length < 1 || feof($this->fp)) { + if($length <= 0) + fgets($this->fp, 8192); + // break out of the streams loop + break; + } + + // loop though the chunk + do + { + // read $length amount of data + // (use fread here) + $data = fread($this->fp, $length); + + // remove the amount received from the total length on the next loop + // it'll attempt to read that much less data + $length -= strlen($data); + + // PRINT out directly + // you could also save it directly to a file here + + // store in string for later use + $body .= $data; + + // zero or less or end of connection break + if ($length <= 0 || feof($this->fp)) + { + // break out of the chunk loop + if($length <= 0) + fgets($this->fp, 8192); + break; + } + } + while (true); + // end of chunk loop + } + while (true); + // end of stream loop + return $body; + } + + private function gunzip($zipped) { + $offset = 0; + if (substr($zipped,0,2) == "\x1f\x8b") + $offset = 2; + if (substr($zipped,$offset,1) == "\x08") + { + try + { + return gzinflate(substr($zipped, $offset + 8)); + } catch (\Exception $e) + { + abort(500, "Fehler beim unzip des Ergebnisses von folgendem Anbieter: " . $this->name); + } + } + return "Unknown Format"; + } + + private function writeRequest () + { + + $out = "GET " . $this->getString . " HTTP/1.1\r\n"; + $out .= "Host: " . $this->host . "\r\n"; + $out .= "User-Agent: " . $this->useragent . "\r\n"; + $out .= "Accept: text/html,application/xhtml+xml,application/xml;q=0.9,*/*;q=0.8\r\n"; + $out .= "Accept-Language: de,en-US;q=0.7,en;q=0.3\r\n"; + $out .= "Accept-Encoding: gzip, deflate, br\r\n"; + $out .= "Connection: keep-alive\r\n\r\n"; + # Anfrage senden: + $sent = 0; $string = $out; $time = microtime(true); + while(true) + { + try{ + $tmp = fwrite($this->fp, $string); + }catch(\ErrorException $e) + { + # Irgendwas ist mit unserem Socket passiert. Wir brauchen einen neuen: + fclose($this->fp); + Redis::del($this->name . "." . $this->socketNumber); + $this->fp = $this->getFreeSocket(); + $sent = 0; + $string = $out; + continue; + } + if($tmp){ + $sent += $tmp; + $string = substr($string, $tmp); + }else + Log::error("Fehler beim schreiben."); + + if(((microtime(true) - $time) / 1000000) >= 500) + { + Log::error("Konnte die Request Daten nicht an: " . $this->name . " senden"); + } + + if($sent >= strlen($out)) + break; + } + if( $sent === strlen($out) ) + { + return true; + } + return false; + } + + private function getFreeSocket() + { + # Je nach Auslastung des Servers ( gleichzeitige Abfragen ), kann es sein, dass wir mehrere Sockets benötigen um die Abfragen ohne Wartezeit beantworten zu können. + # pfsockopen öffnet dabei einen persistenten Socket, der also auch zwischen den verschiedenen php Prozessen geteilt werden kann. + # Wenn der Hostname mit einem bereits erstellten Socket übereinstimmt, wird die Verbindung also aufgegriffen und fortgeführt. + # Allerdings dürfen wir diesen nur verwenden, wenn er nicht bereits von einem anderen Prozess zur Kommunikation verwendet wird. + # Wenn dem so ist, probieren wir den nächsten Socket zu verwenden. + # Dies festzustellen ist komplizierter, als man sich das vorstellt. Folgendes System sollte funktionieren: + # 1. Stelle fest, ob dieser Socket neu erstellt wurde, oder ob ein existierender geöffnet wurde. + $counter = 0; $fp = null; + do + { + + if( intval(Redis::exists($this->host . ".$counter")) === 0 ) + { + Redis::set($this->host . ".$counter", 1); + Redis::expire($this->host . ".$counter", 5); + $this->socketNumber = $counter; + + try + { + $fp = pfsockopen($this->getHost() . ":" . $this->port . "/$counter", $this->port, $errstr, $errno, 1); + if(!$fp) + Log::error("lkajng"); + }catch(\ErrorException $e) + { + Log::error('Fehler beim erstellen des Sockets zu: ' . $this->getHost()); + break; + } + # Wir gucken, ob der Lesepuffer leer ist: + stream_set_blocking($fp, 0); + $string = fgets($fp, 8192); + if( $string !== false || feof($fp) ) + { + Log::error("Der Lesepuffer von: " . $this->name . " war nach dem Erstellen nicht leer. Musste den Socket neu starten."); + fclose($fp); + continue; + } + Log::info($this->getHost() . ":" . $this->port . "/$counter"); + break; + } + $counter++; + }while(true); + return $fp; + } + + private function getHost() + { + $return = ""; + if( $this->port === "443" ) + { + $return .= "tls://"; + }else + { + $return .= "tcp://"; + } + $return .= $this->host; + return $return; + } +} diff --git a/app/MetaGer.php b/app/MetaGer.php index 37c4b5cca3183ea8c46befcf0f78bcf1efdafd8e..2d8dffb3e83bfe29f5366967ea9fda61aa5fdfe6 100644 --- a/app/MetaGer.php +++ b/app/MetaGer.php @@ -52,8 +52,6 @@ class MetaGer function __construct() { $this->starttime = microtime(true); - define('CRLF', "\r\n"); - define('BUFFER_LENGTH', 8192); if( file_exists(config_path() . "/blacklistDomains.txt") && file_exists(config_path() . "/blacklistUrl.txt") ) { # Blacklists einlesen: @@ -70,6 +68,12 @@ class MetaGer $this->languageDetect->setNameMode("2"); } + public function getHashCode () + { + $string = url()->full(); + return md5($string); + } + public function rankAll () { foreach( $this->engines as $engine ) @@ -278,7 +282,6 @@ class MetaGer public function createSearchEngines (Request $request) { - #die(SocketRocket::get("tls", "dominik-pfennig.de", "", 443)); # Überprüfe, welche Sumas eingeschaltet sind @@ -312,7 +315,7 @@ class MetaGer if(!(isset($suma['disabled']) && $suma['disabled']->__toString() === "1")) { - if($suma["name"]->__toString() === "overture") + if($suma["name"]->__toString() === "overture" || $suma["name"]->__toString() === "overtureAds") { $overtureEnabled = TRUE; } @@ -336,7 +339,7 @@ class MetaGer ){ if(!(isset($suma['disabled']) && $suma['disabled']->__toString() === "1")) { - if($suma["name"]->__toString() === "overture") + if($suma["name"]->__toString() === "overture" || $suma["name"]->__toString() === "overtureAds") { $overtureEnabled = TRUE; } @@ -353,6 +356,8 @@ class MetaGer $this->errors[] = "Achtung: Sie haben in ihren Einstellungen keine Suchmaschine ausgewählt."; } + + $engines = []; foreach($enabledSearchengines as $engine){ @@ -360,8 +365,8 @@ class MetaGer { continue; } - # Wenn diese Suchmaschine gar nicht eingeschaltet sein soll + # Wenn diese Suchmaschine gar nicht eingeschaltet sein soll $path = "App\Models\parserSkripte\\" . ucfirst($engine["package"]->__toString()); $time = microtime(); @@ -378,43 +383,51 @@ class MetaGer $this->sockets[$tmp->name] = $tmp->fp; } } + # Nun passiert ein elementarer Schritt. # Wir warten auf die Antwort der Suchmaschinen, da wir vorher nicht weiter machen können. # aber natürlich nicht ewig. # Die Verbindung steht zu diesem Zeitpunkt und auch unsere Request wurde schon gesendet. # Wir geben der Suchmaschine nun bis zu 500ms Zeit zu antworten. - - # Jetzt lesen wir alles aus, was da ist und verwerfen den Rest: $enginesToLoad = count($engines); $loadedEngines = 0; - $time = 0; + $timeStart = microtime(true); + while( true ) { + $time = (microtime(true) - $timeStart) * 1000; + $loadedEngines = intval(Redis::hlen('search.' . $this->getHashCode())); + $canBreak = true; + if( $overtureEnabled && !Redis::hexists('search.' . $this->getHashCode(), 'overture') && !Redis::hexists('search.' . $this->getHashCode(), 'overtureAds')) + $canBreak = false; + + # Abbruchbedingung if($time < 500) { - if($loadedEngines >= $enginesToLoad) + if($loadedEngines >= $enginesToLoad && $canBreak) break; }elseif( $time >= 500 && $time < $this->time) { - if( ($loadedEngines / ($enginesToLoad * 1.0)) >= 0.8 ) + if( ($loadedEngines / ($enginesToLoad * 1.0)) >= 0.8 && $canBreak ) break; }else { break; } - foreach($engines as $engine) + usleep(50000); + } + + + foreach($engines as $engine) + { + if(!$engine->loaded) { - if(!$engine->loaded) - { - $success = $engine->retrieveResults(); - if($engine->loaded) - $loadedEngines += 1; - } + $engine->retrieveResults(); } - usleep(50000); - $time += 50; } + + # und verwerfen den Rest: foreach( $engines as $engine ) { if( !$engine->loaded ) diff --git a/app/Models/Searchengine.php b/app/Models/Searchengine.php index 977d3f27fc0e894f34746f738e9885bc3f7765ad..cb508ff46edcb506b6b43c9e91885a039a261e73 100644 --- a/app/Models/Searchengine.php +++ b/app/Models/Searchengine.php @@ -4,9 +4,13 @@ namespace App\Models; use App\MetaGer; use Log; use Redis; +use App\Jobs\Search; +use Illuminate\Foundation\Bus\DispatchesJobs; -abstract class Searchengine + +abstract class Searchengine { + use DispatchesJobs; protected $ch; # Curl Handle zum erhalten der Ergebnisse public $fp; @@ -58,67 +62,20 @@ abstract class Searchengine $this->startTime = microtime(); $this->getString = $this->generateGetString($metager->getQ(), $metager->getUrl(), $metager->getLanguage(), $metager->getCategory()); - $counter = 0; - # Wir benötigen einen verfügbaren Socket, über den wir kommunizieren können: - $time = microtime(true); - $this->fp = $this->getFreeSocket(); - - $this->setStatistic("connection_time", ((microtime(true)-$time) / 1000000)); - if(!$this->fp) - { - $this->disable($metager->getSumaFile(), "Die Suchmaschine " . $this->name . " wurde für 1h deaktiviert, weil keine Verbindung aufgebaut werden konnte"); - }else - { - $time = microtime(true); - $this->writeRequest(); - $this->setStatistic("write_time", ((microtime(true)-$time) / 1000000)); - } + $this->hash = $metager->getHashCode(); + + # Die Anfragen an die Suchmaschinen werden nun von der Laravel-Queue bearbeitet: + # Hinweis: solange in der .env der QUEUE_DRIVER auf "sync" gestellt ist, werden die Abfragen + # nacheinander abgeschickt. + # Sollen diese Parallel verarbeitet werden, muss ein anderer QUEUE_DRIVER verwendet werden. + # siehe auch: https://laravel.com/docs/5.2/queues + $this->dispatch(new Search($this->hash, $this->host, $this->port, $this->name, $this->getString, $this->useragent, $metager->getSumaFile())); } public abstract function loadResults($result); - private function writeRequest () - { - $out = "GET " . $this->getString . " HTTP/1.1\r\n"; - $out .= "Host: " . $this->host . "\r\n"; - $out .= "User-Agent: " . $this->useragent . "\r\n"; - $out .= "Accept: text/html,application/xhtml+xml,application/xml;q=0.9,*/*;q=0.8\r\n"; - $out .= "Accept-Language: de,en-US;q=0.7,en;q=0.3\r\n"; - $out .= "Accept-Encoding: gzip, deflate, br\r\n"; - $out .= "Connection: keep-alive\r\n\r\n"; - - # Anfrage senden: - $sent = 0; $string = $out; $time = microtime(true); - while(true) - { - try{ - $tmp = fwrite($this->fp, $string); - }catch(\ErrorException $e) - { - # Irgendwas ist mit unserem Socket passiert. Wir brauchen einen neuen: - fclose($this->fp); - Redis::del($this->name . "." . $this->socketNumber); - $this->fp = $this->getFreeSocket(); - $sent = 0; - $string = $out; - continue; - } - if($tmp){ - $sent += $tmp; - $string = substr($string, $tmp); - }else - abort(500, "Fehler beim schreiben."); - - if(((microtime(true) - $time) / 1000000) >= 500) - { - abort(500, "Konnte die Request Daten nicht an: " . $this->name . " senden"); - } - - if($sent >= strlen($out)) - break; - } - } + public function rank (\App\MetaGer $metager) { @@ -128,47 +85,7 @@ abstract class Searchengine } } - private function getFreeSocket() - { - # Je nach Auslastung des Servers ( gleichzeitige Abfragen ), kann es sein, dass wir mehrere Sockets benötigen um die Abfragen ohne Wartezeit beantworten zu können. - # pfsockopen öffnet dabei einen persistenten Socket, der also auch zwischen den verschiedenen php Prozessen geteilt werden kann. - # Wenn der Hostname mit einem bereits erstellten Socket übereinstimmt, wird die Verbindung also aufgegriffen und fortgeführt. - # Allerdings dürfen wir diesen nur verwenden, wenn er nicht bereits von einem anderen Prozess zur Kommunikation verwendet wird. - # Wenn dem so ist, probieren wir den nächsten Socket zu verwenden. - # Dies festzustellen ist komplizierter, als man sich das vorstellt. Folgendes System sollte funktionieren: - # 1. Stelle fest, ob dieser Socket neu erstellt wurde, oder ob ein existierender geöffnet wurde. - $counter = 0; $fp = null; - do - { - - if( intval(Redis::exists($this->host . ".$counter")) === 0 ) - { - Redis::set($this->host . ".$counter", 1); - Redis::expire($this->host . ".$counter", 5); - $this->socketNumber = $counter; - - try - { - $fp = pfsockopen($this->getHost() . ":" . $this->port . "/$counter", $this->port, $errstr, $errno, 1); - }catch(\ErrorException $e) - { - break; - } - # Wir gucken, ob der Lesepuffer leer ist: - stream_set_blocking($fp, 0); - if(fgets($fp, BUFFER_LENGTH) !== false) - { - Log::error("Der Lesepuffer von: " . $this->name . " war nach dem Erstellen nicht leer. Musste den Socket neu starten."); - fclose($fp); - $fp = pfsockopen($this->getHost() . ":" . $this->port . "/$counter", $this->port, $errstr, $errno, 1); - } - break; - } - $counter++; - }while(true); - - return $fp; - } + private function setStatistic($key, $val) { @@ -179,13 +96,7 @@ abstract class Searchengine $this->$key = $newVal; } - public function disable($sumaFile, $message) - { - Log::info($message); - $xml = simplexml_load_file($sumaFile); - $xml->xpath("//sumas/suma[@name='" . $this->name . "']")['0']['disabled'] = date(DATE_RFC822, mktime(date("H")+1,date("i"), date("s"), date("m"), date("d"), date("Y"))); - $xml->saveXML($sumaFile); - } + public function enable($sumaFile, $message) { @@ -200,192 +111,38 @@ abstract class Searchengine fclose($this->fp); } - public function retrieveResults() + public function getSocket() { - $time = microtime(true); - $headers = ''; - $body = ''; - $length = 0; - if(!$this->fp) - { - return; - } - // get headers FIRST - $c = 0; - stream_set_blocking($this->fp, 0); - do + $number = Redis::hget('search.' . $this->hash, $this->name); + if( $number === null ) { - // use fgets() not fread(), fgets stops reading at first newline - // or buffer which ever one is reached first - $data = fgets($this->fp, BUFFER_LENGTH); - // a sincle CRLF indicates end of headers - if ($data === false || $data == CRLF || feof($this->fp) || ((microtime()-$time)/1000000) > 100 ) { - // break BEFORE OUTPUT - break; - } - if( sizeof(($tmp = explode(": ", $data))) === 2 ) - $headers[trim($tmp[0])] = trim($tmp[1]); - $c++; - } - while (true); - // end of headers - if(sizeof($headers) > 1){ - $bodySize = 0; - stream_set_blocking($this->fp, 1); - if( isset($headers["Transfer-Encoding"]) && $headers["Transfer-Encoding"] === "chunked" ) - { - $body = $this->readChunked(); - - }elseif( isset($headers['Content-Length']) ) - { - $length = trim($headers['Content-Length']); - if(is_numeric($length) && $length >= 1) - $body = $this->readBody($length); - $bodySize = strlen($body); - }else - { - die("Konnte nicht herausfinden, wie ich die Serverantwort von: " . $this->name . " auslesen soll. Header war: " . print_r($headers)); - } - $this->loaded = true; + die("test"); + return null; }else { - return; + return pfsockopen($this->getHost() . ":" . $this->port . "/$number", $this->port, $errstr, $errno, 1); } + } - Redis::del($this->host . "." . $this->socketNumber); - $this->setStatistic("read_time", ((microtime(true)-$time) / 1000000)); - if( isset($headers["Content-Encoding"]) && $headers['Content-Encoding'] === "gzip") + public function retrieveResults() + { + if( Redis::hexists('search.' . $this->hash, $this->name)) { - $body = $this->gunzip($body); + $body = Redis::hget('search.' . $this->hash, $this->name); + $this->loadResults($body); + $this->loaded = true; + Redis::hdel('search.' . $this->hash, $this->name); + return true; } - #print_r($headers); - #print($body); - #print("\r\n". $bodySize); - #exit; - #die(print_r($headers)); - // $body and $headers should contain your stream data - $this->loadResults($body); - #print(print_r($headers, TRUE) . $body); - #exit; + return false; + } public function shutdown() { - if( $this->fp ) - fclose($this->fp); Redis::del($this->host . "." . $this->socketNumber); } - private function readBody($length) - { - $theData = ''; - $done = false; - stream_set_blocking($this->fp, 0); - $startTime = time(); - $lastTime = $startTime; - while (!feof($this->fp) && !$done && (($startTime + 1) > time()) && $length !== 0) - { - usleep(100); - $theNewData = fgets($this->fp, BUFFER_LENGTH); - $theData .= $theNewData; - $length -= strlen($theNewData); - $done = (trim($theNewData) === '0'); - - } - return $theData; - } - - private function readChunked() - { - $body = ''; - // read from chunked stream - // loop though the stream - do - { - // NOTE: for chunked encoding to work properly make sure - // there is NOTHING (besides newlines) before the first hexlength - - // get the line which has the length of this chunk (use fgets here) - $line = fgets($this->fp, BUFFER_LENGTH); - - // if it's only a newline this normally means it's read - // the total amount of data requested minus the newline - // continue to next loop to make sure we're done - if ($line == CRLF) { - continue; - } - - // the length of the block is sent in hex decode it then loop through - // that much data get the length - // NOTE: hexdec() ignores all non hexadecimal chars it finds - $length = hexdec($line); - - if (!is_int($length)) { - trigger_error('Most likely not chunked encoding', E_USER_ERROR); - } - - // zero is sent when at the end of the chunks - // or the end of the stream or error - if ($line === false || $length < 1 || feof($this->fp)) { - if($length <= 0) - fgets($this->fp, BUFFER_LENGTH); - // break out of the streams loop - break; - } - - // loop though the chunk - do - { - // read $length amount of data - // (use fread here) - $data = fread($this->fp, $length); - - // remove the amount received from the total length on the next loop - // it'll attempt to read that much less data - $length -= strlen($data); - - // PRINT out directly - #print $data; - #flush(); - // you could also save it directly to a file here - - // store in string for later use - $body .= $data; - - // zero or less or end of connection break - if ($length <= 0 || feof($this->fp)) - { - // break out of the chunk loop - if($length <= 0) - fgets($this->fp, BUFFER_LENGTH); - break; - } - } - while (true); - // end of chunk loop - } - while (true); - // end of stream loop - return $body; - } - - private function gunzip($zipped) { - $offset = 0; - if (substr($zipped,0,2) == "\x1f\x8b") - $offset = 2; - if (substr($zipped,$offset,1) == "\x08") - { - try - { - return gzinflate(substr($zipped, $offset + 8)); - } catch (\Exception $e) - { - abort(500, "Fehler beim unzip des Ergebnisses von folgendem Anbieter: " . $this->name); - } - } - return "Unknown Format"; - } - protected function getHost() { $return = ""; diff --git a/config/database.php b/config/database.php index 076fc54e5d82e13b4cf00bdf7499b5090c647789..57cc7e5c76b968776b5e2d60571a7ada84e7e565 100755 --- a/config/database.php +++ b/config/database.php @@ -48,7 +48,7 @@ return [ 'sqlite' => [ 'driver' => 'sqlite', - 'database' => env('DB_DATABASE', database_path('database.sqlite')), + 'database' => database_path(env('DB_DATABASE', 'database.sqlite')), 'prefix' => '', ], diff --git a/database/migrations/2016_06_14_132524_create_jobs_table.php b/database/migrations/2016_06_14_132524_create_jobs_table.php new file mode 100644 index 0000000000000000000000000000000000000000..81b2d29f2b9eb6f7c3f245c517b2293485c8dc44 --- /dev/null +++ b/database/migrations/2016_06_14_132524_create_jobs_table.php @@ -0,0 +1,37 @@ +<?php + +use Illuminate\Database\Schema\Blueprint; +use Illuminate\Database\Migrations\Migration; + +class CreateJobsTable extends Migration +{ + /** + * Run the migrations. + * + * @return void + */ + public function up() + { + Schema::create('jobs', function (Blueprint $table) { + $table->bigIncrements('id'); + $table->string('queue'); + $table->longText('payload'); + $table->tinyInteger('attempts')->unsigned(); + $table->tinyInteger('reserved')->unsigned(); + $table->unsignedInteger('reserved_at')->nullable(); + $table->unsignedInteger('available_at'); + $table->unsignedInteger('created_at'); + $table->index(['queue', 'reserved', 'reserved_at']); + }); + } + + /** + * Reverse the migrations. + * + * @return void + */ + public function down() + { + Schema::drop('jobs'); + } +} diff --git a/database/migrations/2016_06_14_132551_create_failed_jobs_table.php b/database/migrations/2016_06_14_132551_create_failed_jobs_table.php new file mode 100644 index 0000000000000000000000000000000000000000..3d733b6bf029a112fc5885f57501c6dc85963afe --- /dev/null +++ b/database/migrations/2016_06_14_132551_create_failed_jobs_table.php @@ -0,0 +1,33 @@ +<?php + +use Illuminate\Database\Schema\Blueprint; +use Illuminate\Database\Migrations\Migration; + +class CreateFailedJobsTable extends Migration +{ + /** + * Run the migrations. + * + * @return void + */ + public function up() + { + Schema::create('failed_jobs', function (Blueprint $table) { + $table->increments('id'); + $table->text('connection'); + $table->text('queue'); + $table->longText('payload'); + $table->timestamp('failed_at')->useCurrent(); + }); + } + + /** + * Reverse the migrations. + * + * @return void + */ + public function down() + { + Schema::drop('failed_jobs'); + } +} diff --git a/database/migrations/2016_06_14_132601_create_failed_jobs_table.php b/database/migrations/2016_06_14_132601_create_failed_jobs_table.php new file mode 100644 index 0000000000000000000000000000000000000000..3d733b6bf029a112fc5885f57501c6dc85963afe --- /dev/null +++ b/database/migrations/2016_06_14_132601_create_failed_jobs_table.php @@ -0,0 +1,33 @@ +<?php + +use Illuminate\Database\Schema\Blueprint; +use Illuminate\Database\Migrations\Migration; + +class CreateFailedJobsTable extends Migration +{ + /** + * Run the migrations. + * + * @return void + */ + public function up() + { + Schema::create('failed_jobs', function (Blueprint $table) { + $table->increments('id'); + $table->text('connection'); + $table->text('queue'); + $table->longText('payload'); + $table->timestamp('failed_at')->useCurrent(); + }); + } + + /** + * Reverse the migrations. + * + * @return void + */ + public function down() + { + Schema::drop('failed_jobs'); + } +} diff --git a/public/index.php b/public/index.php index 79a6cbfc5b5b79827ca11c38bbac1d0aabb64378..76ed52b9c96dfdf120c03a00f9698c40ca756d04 100644 --- a/public/index.php +++ b/public/index.php @@ -9,21 +9,24 @@ # Unser erster Schritt wird sein, IP-Adresse und USER-Agent zu anonymisieren, damit # nicht einmal wir selbst noch Zugriff auf die Daten haben: -if( !isset($_SERVER['HTTP_X_FORWARDED_FOR'] )) +if( !isset($_SERVER['HTTP_X_FORWARDED_FOR']) && isset($_SERVER['REMOTE_ADDR']) ) { $_SERVER['REMOTE_ADDR'] = preg_replace("/(\d+)\.(\d+)\.\d+.\d+/s", "$1.$2.0.0", $_SERVER['REMOTE_ADDR']); -}else +}elseif( isset($_SERVER['HTTP_X_FORWARDED_FOR'])) { $_SERVER['HTTP_X_FORWARDED_FOR'] = preg_replace("/(\d+)\.(\d+)\.\d+.\d+/s", "$1.$2.0.0", $_SERVER['HTTP_X_FORWARDED_FOR']); } -$agentPieces = explode(" ", $_SERVER['HTTP_USER_AGENT']); - -for($i = 0; $i < count($agentPieces); $i++) +if( isset($_SERVER['HTTP_USER_AGENT']) ) { - $agentPieces[$i] = preg_replace("/(\d+\.\d+)/s", "0.0", $agentPieces[$i]); - $agentPieces[$i] = preg_replace("/([^\/]*)\/\w+/s", "$1/0.0", $agentPieces[$i]); + $agentPieces = explode(" ", $_SERVER['HTTP_USER_AGENT']); + + for($i = 0; $i < count($agentPieces); $i++) + { + $agentPieces[$i] = preg_replace("/(\d+\.\d+)/s", "0.0", $agentPieces[$i]); + $agentPieces[$i] = preg_replace("/([^\/]*)\/\w+/s", "$1/0.0", $agentPieces[$i]); + } + $_SERVER['HTTP_USER_AGENT'] = implode(" ", $agentPieces); } -$_SERVER['HTTP_USER_AGENT'] = implode(" ", $agentPieces); /* |-------------------------------------------------------------------------- diff --git a/readme.md b/readme.md index 57c1d177d77cbaff62017800a8f8741e5c19537e..6ae2f1b9bcdba81e09250b3218fe3a84a868a0f7 100644 --- a/readme.md +++ b/readme.md @@ -8,6 +8,13 @@ Während wir daran arbeiten sie soweit fertig zu stellen, dass alles exakt so fu ein Mechanismus für den automatischen Deploy dafür, dass zu jedem beliebigen Zeitpunkt der Master-Branch in seiner aktuellen Version über die Adresse https://metager3.de ( Entwicklungsserver ) erreichbar ist. +## MetaGer zu langsam? +Damit MetaGer so schnell wird, wie auf unserem Live-Server, erfordert es ein wenig Konfigurationsarbeit. Der Grund, warum die Version nach dem Checkout langsamer als normal ist, ist der, dass die eingestellten Suchmaschinen im Standard synchron abgefragt werden. +Das heißt, dass bei einer Suche mit 20 Suchmaschinen eine Suchmaschine nach der anderen abgefragt wird. +Die parallele abarbeitung kann mit Hilfe von Laravels Queue-System ( https://laravel.com/docs/5.2/queues ) hergestellt werden. +Im Standard, ist in der Datei ".env" QUEUE_DRIVER=sync gesetzt. +Wir verwenden auf unseren Servern den QUEUE_DRIVER=redis und haben mit Hilfe von Supervisor ( https://laravel.com/docs/5.2/queues#supervisor-configuration ) eine Menge queue:worker Prozesse am laufen, die für eine parallele bearbeitung sorgen. + ## Offizielle Dokumentation Die Dokumentation ist im Wiki des Gitlab-Projektes zu finden.