From 6230c45fcd3974c6a294a36d71084965e9ea94f7 Mon Sep 17 00:00:00 2001 From: freekmurze Date: Wed, 31 Jul 2024 10:44:32 +0000 Subject: [PATCH] Fix styling --- src/Crawler.php | 8 ++++---- src/Handlers/CrawlRequestFulfilled.php | 7 ++++--- tests/ArrayCrawlQueueTest.php | 2 +- tests/CrawlObserverCollectionTest.php | 4 ++-- tests/CrawlerTest.php | 18 +++++++++--------- tests/Pest.php | 2 +- 6 files changed, 21 insertions(+), 20 deletions(-) diff --git a/src/Crawler.php b/src/Crawler.php index 3d54d46..141ab02 100644 --- a/src/Crawler.php +++ b/src/Crawler.php @@ -96,11 +96,11 @@ public function __construct( protected Client $client, protected int $concurrency = 10, ) { - $this->crawlProfile = new CrawlAllUrls(); + $this->crawlProfile = new CrawlAllUrls; - $this->crawlQueue = new ArrayCrawlQueue(); + $this->crawlQueue = new ArrayCrawlQueue; - $this->crawlObservers = new CrawlObserverCollection(); + $this->crawlObservers = new CrawlObserverCollection; $this->crawlRequestFulfilledClass = CrawlRequestFulfilled::class; @@ -399,7 +399,7 @@ public function getUserAgent(): string public function getBrowsershot(): Browsershot { if (! $this->browsershot) { - $this->browsershot = new Browsershot(); + $this->browsershot = new Browsershot; } return $this->browsershot; diff --git a/src/Handlers/CrawlRequestFulfilled.php b/src/Handlers/CrawlRequestFulfilled.php index 1f9a0e7..6b16e16 100644 --- a/src/Handlers/CrawlRequestFulfilled.php +++ b/src/Handlers/CrawlRequestFulfilled.php @@ -3,6 +3,8 @@ namespace Spatie\Crawler\Handlers; use Exception; +use GuzzleHttp\Exception\RequestException; +use GuzzleHttp\Psr7\Request; use GuzzleHttp\Psr7\Uri; use GuzzleHttp\Psr7\Utils; use GuzzleHttp\RedirectMiddleware; @@ -16,8 +18,6 @@ use Spatie\Crawler\ResponseWithCachedBody; use Spatie\Crawler\UrlParsers\UrlParser; use Symfony\Component\Process\Exception\ProcessFailedException; -use GuzzleHttp\Psr7\Request; -use GuzzleHttp\Exception\RequestException; class CrawlRequestFulfilled { @@ -45,13 +45,14 @@ public function __invoke(ResponseInterface $response, $index) try { $body = $this->getBodyAfterExecutingJavaScript($crawlUrl->url); } catch (ProcessFailedException $exception) { - $request = new Request("GET", $crawlUrl->url); + $request = new Request('GET', $crawlUrl->url); $exception = new RequestException($exception->getMessage(), $request); $crawlUrl = $this->crawler->getCrawlQueue()->getUrlById($index); $this->crawler->getCrawlObservers()->crawlFailed($crawlUrl, $exception); usleep($this->crawler->getDelayBetweenRequests()); + return; } diff --git a/tests/ArrayCrawlQueueTest.php b/tests/ArrayCrawlQueueTest.php index 728f8f4..2aec812 100644 --- a/tests/ArrayCrawlQueueTest.php +++ b/tests/ArrayCrawlQueueTest.php @@ -5,7 +5,7 @@ use Spatie\Crawler\CrawlUrl; beforeEach(function () { - $this->crawlQueue = new ArrayCrawlQueue(); + $this->crawlQueue = new ArrayCrawlQueue; }); test('a url can be added to crawl queue', function () { diff --git a/tests/CrawlObserverCollectionTest.php b/tests/CrawlObserverCollectionTest.php index 60170b8..c0e95d2 100644 --- a/tests/CrawlObserverCollectionTest.php +++ b/tests/CrawlObserverCollectionTest.php @@ -11,7 +11,7 @@ use Spatie\Crawler\CrawlUrl; beforeEach(function () { - $this->crawlObserver = new class() extends CrawlObserver + $this->crawlObserver = new class extends CrawlObserver { public $crawled = false; @@ -44,7 +44,7 @@ public function crawlFailed( $observers->crawled( CrawlUrl::create(new Uri('')), - new Response() + new Response ); expect($this->crawlObserver) diff --git a/tests/CrawlerTest.php b/tests/CrawlerTest.php index e0ea670..2ecbed9 100644 --- a/tests/CrawlerTest.php +++ b/tests/CrawlerTest.php @@ -71,7 +71,7 @@ $crawler = Crawler::create(); if (getenv('TRAVIS')) { - $browsershot = new Browsershot(); + $browsershot = new Browsershot; $browsershot->noSandbox(); @@ -80,7 +80,7 @@ $crawler ->executeJavaScript() - ->setCrawlObserver(new CrawlLogger()) + ->setCrawlObserver(new CrawlLogger) ->startCrawling('http://localhost:8080'); expect(regularUrls())->each->toBeCrawledOnce(); @@ -89,7 +89,7 @@ }); it('allows for a browsershot instance to be set', function () { - $browsershot = new Browsershot(); + $browsershot = new Browsershot; if (getenv('TRAVIS')) { $browsershot->noSandbox(); @@ -98,7 +98,7 @@ Crawler::create() ->setBrowsershot($browsershot) ->executeJavaScript() - ->setCrawlObserver(new CrawlLogger()) + ->setCrawlObserver(new CrawlLogger) ->startCrawling('http://localhost:8080'); expect(regularUrls())->each->toBeCrawledOnce(); @@ -110,7 +110,7 @@ Crawler::create() ->executeJavaScript() ->doNotExecuteJavaScript() - ->setCrawlObserver(new CrawlLogger()) + ->setCrawlObserver(new CrawlLogger) ->startCrawling('http://localhost:8080'); expect(regularUrls())->each->toBeCrawledOnce(); @@ -129,7 +129,7 @@ ]) ->setBrowsershot($browsershot) ->executeJavaScript() - ->setCrawlObserver(new CrawlLogger()) + ->setCrawlObserver(new CrawlLogger) ->startCrawling('http://localhost:8080/simulate-activity'); })->not->toThrow(ProcessFailedException::class); @@ -137,7 +137,7 @@ }); it('uses a crawl profile to determine what should be crawled', function () { - $crawlProfile = new class() extends CrawlProfile + $crawlProfile = new class extends CrawlProfile { public function shouldCrawl(UriInterface $url): bool { @@ -146,7 +146,7 @@ public function shouldCrawl(UriInterface $url): bool }; createCrawler() - ->setCrawlProfile(new $crawlProfile()) + ->setCrawlProfile(new $crawlProfile) ->startCrawling('http://localhost:8080'); expect([ @@ -200,7 +200,7 @@ function ($url) { }); it('can handle pages with invalid urls', function () { - $crawlProfile = new class() extends CrawlProfile + $crawlProfile = new class extends CrawlProfile { public function shouldCrawl(UriInterface $url): bool { diff --git a/tests/Pest.php b/tests/Pest.php index 1e69644..fc2b2cd 100644 --- a/tests/Pest.php +++ b/tests/Pest.php @@ -82,5 +82,5 @@ function createCrawler($options = []): Crawler { return Crawler::create($options) ->setMaximumDepth(3) - ->setCrawlObserver(new CrawlLogger()); + ->setCrawlObserver(new CrawlLogger); }