mirror of
https://github.com/shlinkio/shlink.git
synced 2024-12-14 23:22:21 +03:00
76 lines
2.2 KiB
PHP
76 lines
2.2 KiB
PHP
|
<?php
|
||
|
|
||
|
declare(strict_types=1);
|
||
|
|
||
|
namespace ShlinkioTest\Shlink\Core\Action;
|
||
|
|
||
|
use Laminas\Diactoros\ServerRequestFactory;
|
||
|
use PHPUnit\Framework\TestCase;
|
||
|
use Prophecy\PhpUnit\ProphecyTrait;
|
||
|
use Prophecy\Prophecy\ObjectProphecy;
|
||
|
use Shlinkio\Shlink\Core\Action\RobotsAction;
|
||
|
use Shlinkio\Shlink\Core\Crawling\CrawlingHelperInterface;
|
||
|
|
||
|
class RobotsActionTest extends TestCase
|
||
|
{
|
||
|
use ProphecyTrait;
|
||
|
|
||
|
private RobotsAction $action;
|
||
|
private ObjectProphecy $helper;
|
||
|
|
||
|
protected function setUp(): void
|
||
|
{
|
||
|
$this->helper = $this->prophesize(CrawlingHelperInterface::class);
|
||
|
$this->action = new RobotsAction($this->helper->reveal());
|
||
|
}
|
||
|
|
||
|
/**
|
||
|
* @test
|
||
|
* @dataProvider provideShortCodes
|
||
|
*/
|
||
|
public function buildsRobotsLinesFromCrawlableShortCodes(array $shortCodes, string $expected): void
|
||
|
{
|
||
|
$getShortCodes = $this->helper->listCrawlableShortCodes()->willReturn($shortCodes);
|
||
|
|
||
|
$response = $this->action->handle(ServerRequestFactory::fromGlobals());
|
||
|
|
||
|
self::assertEquals(200, $response->getStatusCode());
|
||
|
self::assertEquals($expected, $response->getBody()->__toString());
|
||
|
self::assertEquals('text/plain', $response->getHeaderLine('Content-Type'));
|
||
|
$getShortCodes->shouldHaveBeenCalledOnce();
|
||
|
}
|
||
|
|
||
|
public function provideShortCodes(): iterable
|
||
|
{
|
||
|
yield 'three short codes' => [['foo', 'bar', 'baz'], <<<ROBOTS
|
||
|
# For more information about the robots.txt standard, see:
|
||
|
# https://www.robotstxt.org/orig.html
|
||
|
|
||
|
User-agent: *
|
||
|
Allow: /foo
|
||
|
Allow: /bar
|
||
|
Allow: /baz
|
||
|
Disallow: /
|
||
|
ROBOTS];
|
||
|
yield 'five short codes' => [['foo', 'bar', 'some', 'thing', 'baz'], <<<ROBOTS
|
||
|
# For more information about the robots.txt standard, see:
|
||
|
# https://www.robotstxt.org/orig.html
|
||
|
|
||
|
User-agent: *
|
||
|
Allow: /foo
|
||
|
Allow: /bar
|
||
|
Allow: /some
|
||
|
Allow: /thing
|
||
|
Allow: /baz
|
||
|
Disallow: /
|
||
|
ROBOTS];
|
||
|
yield 'no short codes' => [[], <<<ROBOTS
|
||
|
# For more information about the robots.txt standard, see:
|
||
|
# https://www.robotstxt.org/orig.html
|
||
|
|
||
|
User-agent: *
|
||
|
Disallow: /
|
||
|
ROBOTS];
|
||
|
}
|
||
|
}
|