mirror of
https://github.com/Llewellynvdm/php-ml.git
synced 2024-11-25 22:37:34 +00:00
db82afa263
* Update to phpunit 8 * Require at least PHP 7.2
60 lines
1.4 KiB
PHP
60 lines
1.4 KiB
PHP
<?php
|
|
|
|
declare(strict_types=1);
|
|
|
|
namespace Phpml\Tests\NeuralNetwork\ActivationFunction;
|
|
|
|
use Phpml\NeuralNetwork\ActivationFunction\PReLU;
|
|
use PHPUnit\Framework\TestCase;
|
|
|
|
class PReLUTest extends TestCase
|
|
{
|
|
/**
|
|
* @dataProvider preluProvider
|
|
*
|
|
* @param float|int $value
|
|
*/
|
|
public function testPReLUActivationFunction(float $beta, float $expected, $value): void
|
|
{
|
|
$prelu = new PReLU($beta);
|
|
|
|
self::assertEqualsWithDelta($expected, $prelu->compute($value), 0.001);
|
|
}
|
|
|
|
public function preluProvider(): array
|
|
{
|
|
return [
|
|
[0.01, 0.367, 0.367],
|
|
[0.0, 1, 1],
|
|
[0.3, -0.3, -1],
|
|
[0.9, 3, 3],
|
|
[0.02, -0.06, -3],
|
|
];
|
|
}
|
|
|
|
/**
|
|
* @dataProvider preluDerivativeProvider
|
|
*
|
|
* @param float|int $value
|
|
*/
|
|
public function testPReLUDerivative(float $beta, float $expected, $value): void
|
|
{
|
|
$prelu = new PReLU($beta);
|
|
$activatedValue = $prelu->compute($value);
|
|
self::assertEquals($expected, $prelu->differentiate($value, $activatedValue));
|
|
}
|
|
|
|
public function preluDerivativeProvider(): array
|
|
{
|
|
return [
|
|
[0.5, 0.5, -3],
|
|
[0.5, 1, 0],
|
|
[0.5, 1, 1],
|
|
[0.01, 1, 1],
|
|
[1, 1, 1],
|
|
[0.3, 1, 0.1],
|
|
[0.1, 0.1, -0.1],
|
|
];
|
|
}
|
|
}
|