mirror of
https://github.com/Llewellynvdm/php-ml.git
synced 2024-12-03 18:38:25 +00:00
db82afa263
* Update to phpunit 8 * Require at least PHP 7.2
223 lines
8.1 KiB
PHP
223 lines
8.1 KiB
PHP
<?php
|
|
|
|
declare(strict_types=1);
|
|
|
|
namespace Phpml\Tests\Classification\Linear;
|
|
|
|
use Phpml\Classification\Linear\LogisticRegression;
|
|
use Phpml\Exception\InvalidArgumentException;
|
|
use PHPUnit\Framework\TestCase;
|
|
use ReflectionMethod;
|
|
use ReflectionProperty;
|
|
|
|
class LogisticRegressionTest extends TestCase
|
|
{
|
|
public function testConstructorThrowWhenInvalidTrainingType(): void
|
|
{
|
|
$this->expectException(InvalidArgumentException::class);
|
|
$this->expectExceptionMessage('Logistic regression can only be trained with '.
|
|
'batch (gradient descent), online (stochastic gradient descent) '.
|
|
'or conjugate batch (conjugate gradients) algorithms');
|
|
|
|
new LogisticRegression(
|
|
500,
|
|
true,
|
|
-1,
|
|
'log',
|
|
'L2'
|
|
);
|
|
}
|
|
|
|
public function testConstructorThrowWhenInvalidCost(): void
|
|
{
|
|
$this->expectException(InvalidArgumentException::class);
|
|
$this->expectExceptionMessage("Logistic regression cost function can be one of the following: \n".
|
|
"'log' for log-likelihood and 'sse' for sum of squared errors");
|
|
|
|
new LogisticRegression(
|
|
500,
|
|
true,
|
|
LogisticRegression::CONJUGATE_GRAD_TRAINING,
|
|
'invalid',
|
|
'L2'
|
|
);
|
|
}
|
|
|
|
public function testConstructorThrowWhenInvalidPenalty(): void
|
|
{
|
|
$this->expectException(InvalidArgumentException::class);
|
|
$this->expectExceptionMessage('Logistic regression supports only \'L2\' regularization');
|
|
|
|
new LogisticRegression(
|
|
500,
|
|
true,
|
|
LogisticRegression::CONJUGATE_GRAD_TRAINING,
|
|
'log',
|
|
'invalid'
|
|
);
|
|
}
|
|
|
|
public function testPredictSingleSample(): void
|
|
{
|
|
// AND problem
|
|
$samples = [[0, 0], [1, 0], [0, 1], [1, 1], [0.4, 0.4], [0.6, 0.6]];
|
|
$targets = [0, 0, 0, 1, 0, 1];
|
|
$classifier = new LogisticRegression();
|
|
$classifier->train($samples, $targets);
|
|
self::assertEquals(0, $classifier->predict([0.1, 0.1]));
|
|
self::assertEquals(1, $classifier->predict([0.9, 0.9]));
|
|
}
|
|
|
|
public function testPredictSingleSampleWithBatchTraining(): void
|
|
{
|
|
$samples = [[0, 0], [1, 0], [0, 1], [1, 1], [0.4, 0.4], [0.6, 0.6]];
|
|
$targets = [0, 0, 0, 1, 0, 1];
|
|
|
|
// $maxIterations is set to 10000 as batch training needs more
|
|
// iteration to converge than CG method in general.
|
|
$classifier = new LogisticRegression(
|
|
10000,
|
|
true,
|
|
LogisticRegression::BATCH_TRAINING,
|
|
'log',
|
|
'L2'
|
|
);
|
|
$classifier->train($samples, $targets);
|
|
self::assertEquals(0, $classifier->predict([0.1, 0.1]));
|
|
self::assertEquals(1, $classifier->predict([0.9, 0.9]));
|
|
}
|
|
|
|
public function testPredictSingleSampleWithOnlineTraining(): void
|
|
{
|
|
$samples = [[0, 0], [1, 0], [0, 1], [1, 1], [0.4, 0.4], [0.6, 0.6]];
|
|
$targets = [0, 0, 0, 1, 0, 1];
|
|
|
|
// $penalty is set to empty (no penalty) because L2 penalty seems to
|
|
// prevent convergence in online training for this dataset.
|
|
$classifier = new LogisticRegression(
|
|
10000,
|
|
true,
|
|
LogisticRegression::ONLINE_TRAINING,
|
|
'log',
|
|
''
|
|
);
|
|
$classifier->train($samples, $targets);
|
|
self::assertEquals(0, $classifier->predict([0.1, 0.1]));
|
|
self::assertEquals(1, $classifier->predict([0.9, 0.9]));
|
|
}
|
|
|
|
public function testPredictSingleSampleWithSSECost(): void
|
|
{
|
|
$samples = [[0, 0], [1, 0], [0, 1], [1, 1], [0.4, 0.4], [0.6, 0.6]];
|
|
$targets = [0, 0, 0, 1, 0, 1];
|
|
$classifier = new LogisticRegression(
|
|
500,
|
|
true,
|
|
LogisticRegression::CONJUGATE_GRAD_TRAINING,
|
|
'sse',
|
|
'L2'
|
|
);
|
|
$classifier->train($samples, $targets);
|
|
self::assertEquals(0, $classifier->predict([0.1, 0.1]));
|
|
self::assertEquals(1, $classifier->predict([0.9, 0.9]));
|
|
}
|
|
|
|
public function testPredictSingleSampleWithoutPenalty(): void
|
|
{
|
|
$samples = [[0, 0], [1, 0], [0, 1], [1, 1], [0.4, 0.4], [0.6, 0.6]];
|
|
$targets = [0, 0, 0, 1, 0, 1];
|
|
$classifier = new LogisticRegression(
|
|
500,
|
|
true,
|
|
LogisticRegression::CONJUGATE_GRAD_TRAINING,
|
|
'log',
|
|
''
|
|
);
|
|
$classifier->train($samples, $targets);
|
|
self::assertEquals(0, $classifier->predict([0.1, 0.1]));
|
|
self::assertEquals(1, $classifier->predict([0.9, 0.9]));
|
|
}
|
|
|
|
public function testPredictMultiClassSample(): void
|
|
{
|
|
// By use of One-v-Rest, Perceptron can perform multi-class classification
|
|
// The samples should be separable by lines perpendicular to the dimensions
|
|
$samples = [
|
|
[0, 0], [0, 1], [1, 0], [1, 1], // First group : a cluster at bottom-left corner in 2D
|
|
[5, 5], [6, 5], [5, 6], [7, 5], // Second group: another cluster at the middle-right
|
|
[3, 10], [3, 10], [3, 8], [3, 9], // Third group : cluster at the top-middle
|
|
];
|
|
$targets = [0, 0, 0, 0, 1, 1, 1, 1, 2, 2, 2, 2];
|
|
|
|
$classifier = new LogisticRegression();
|
|
$classifier->train($samples, $targets);
|
|
self::assertEquals(0, $classifier->predict([0.5, 0.5]));
|
|
self::assertEquals(1, $classifier->predict([6.0, 5.0]));
|
|
self::assertEquals(2, $classifier->predict([3.0, 9.5]));
|
|
}
|
|
|
|
public function testPredictProbabilitySingleSample(): void
|
|
{
|
|
$samples = [[0, 0], [1, 0], [0, 1], [1, 1], [0.4, 0.4], [0.6, 0.6]];
|
|
$targets = [0, 0, 0, 1, 0, 1];
|
|
$classifier = new LogisticRegression();
|
|
$classifier->train($samples, $targets);
|
|
|
|
$property = new ReflectionProperty($classifier, 'classifiers');
|
|
$property->setAccessible(true);
|
|
$predictor = $property->getValue($classifier)[0];
|
|
$method = new ReflectionMethod($predictor, 'predictProbability');
|
|
$method->setAccessible(true);
|
|
|
|
$zero = $method->invoke($predictor, [0.1, 0.1], 0);
|
|
$one = $method->invoke($predictor, [0.1, 0.1], 1);
|
|
self::assertEqualsWithDelta(1, $zero + $one, 1e-6);
|
|
self::assertTrue($zero > $one);
|
|
|
|
$zero = $method->invoke($predictor, [0.9, 0.9], 0);
|
|
$one = $method->invoke($predictor, [0.9, 0.9], 1);
|
|
self::assertEqualsWithDelta(1, $zero + $one, 1e-6);
|
|
self::assertTrue($zero < $one);
|
|
}
|
|
|
|
public function testPredictProbabilityMultiClassSample(): void
|
|
{
|
|
$samples = [
|
|
[0, 0], [0, 1], [1, 0], [1, 1],
|
|
[5, 5], [6, 5], [5, 6], [6, 6],
|
|
[3, 10], [3, 10], [3, 8], [3, 9],
|
|
];
|
|
$targets = [0, 0, 0, 0, 1, 1, 1, 1, 2, 2, 2, 2];
|
|
|
|
$classifier = new LogisticRegression();
|
|
$classifier->train($samples, $targets);
|
|
|
|
$property = new ReflectionProperty($classifier, 'classifiers');
|
|
$property->setAccessible(true);
|
|
|
|
$predictor = $property->getValue($classifier)[0];
|
|
$method = new ReflectionMethod($predictor, 'predictProbability');
|
|
$method->setAccessible(true);
|
|
$zero = $method->invoke($predictor, [3.0, 9.5], 0);
|
|
$not_zero = $method->invoke($predictor, [3.0, 9.5], 'not_0');
|
|
|
|
$predictor = $property->getValue($classifier)[1];
|
|
$method = new ReflectionMethod($predictor, 'predictProbability');
|
|
$method->setAccessible(true);
|
|
$one = $method->invoke($predictor, [3.0, 9.5], 1);
|
|
$not_one = $method->invoke($predictor, [3.0, 9.5], 'not_1');
|
|
|
|
$predictor = $property->getValue($classifier)[2];
|
|
$method = new ReflectionMethod($predictor, 'predictProbability');
|
|
$method->setAccessible(true);
|
|
$two = $method->invoke($predictor, [3.0, 9.5], 2);
|
|
$not_two = $method->invoke($predictor, [3.0, 9.5], 'not_2');
|
|
|
|
self::assertEqualsWithDelta(1, $zero + $not_zero, 1e-6);
|
|
self::assertEqualsWithDelta(1, $one + $not_one, 1e-6);
|
|
self::assertEqualsWithDelta(1, $two + $not_two, 1e-6);
|
|
self::assertTrue($zero < $two);
|
|
self::assertTrue($one < $two);
|
|
}
|
|
}
|