This project is a subproject from a bigger and older project called CAI and is sister to Keras based K-CAI NEURAL API.
- Compiled pascal code is super fast! This API can outperform some major APIs in some architectures.
- Pascal is easy to learn and easy to make a readable and understandable source code. You'll be able to make super fast native code and at the same time have a readable code.
You'll need Lazarus development environment. If you have an OpenCL capable device, you'll need its OpenCL drivers.
This project is Lazarus based. That said, as of release v0.98, a number of units do compile with Delphi and you can create and run neural networks with Delphi. You'll be able to compile these units with Delphi: neuralvolume, neuralnetwork, neuralab, neuralabfun, neuralbit, neuralbyteprediction, neuralcache, neuraldatasets, neuralgeneric, neuralplanbuilder and neuralfit. At this moment, Neural OpenCL and Neural Threading for Delphi are experimental.
Clone this project, add the neural folder to your Lazarus unit search path and you'll be ready to go!
This is an example for image classification:
NN := TNNet.Create();
NN.AddLayer([
TNNetInput.Create(32, 32, 3), //32x32x3 Input Image
TNNetConvolutionReLU.Create({Features=}16, {FeatureSize=}5, {Padding=}0, {Stride=}1, {SuppressBias=}0),
TNNetMaxPool.Create({Size=}2),
TNNetConvolutionReLU.Create({Features=}32, {FeatureSize=}5, {Padding=}0, {Stride=}1, {SuppressBias=}0),
TNNetMaxPool.Create({Size=}2),
TNNetConvolutionReLU.Create({Features=}32, {FeatureSize=}5, {Padding=}0, {Stride=}1, {SuppressBias=}0),
TNNetFullConnectReLU.Create({Neurons=}32),
TNNetFullConnectLinear.Create(NumClasses),
TNNetSoftMax.Create()
]);
CreateCifar10Volumes(ImgTrainingVolumes, ImgValidationVolumes, ImgTestVolumes);
WriteLn('Neural Network will minimize error with:');
WriteLn(' Layers: ', NN.CountLayers());
WriteLn(' Neurons:', NN.CountNeurons());
WriteLn(' Weights:', NN.CountWeights());
NeuralFit := TNeuralImageFit.Create;
NeuralFit.InitialLearningRate := fLearningRate;
NeuralFit.Inertia := fInertia;
NeuralFit.Fit(NN, ImgTrainingVolumes, ImgValidationVolumes, ImgTestVolumes, NumClasses, {batchsize}128, {epochs}100);
The documentation is under construction and is currently composed by:
- Introductory Examples.
- Youtube Videos.
- Advanced Examples.
Some recommended introductory source code examples are:
- Training a neural network to learn the hypotenuse function
- Training a neural network to learn the hypotenuse function with FitLoading
- Training a neural network to learn boolean functions AND, OR and XOR with neuralfit unit
- Training a neural network to learn boolean functions AND, OR and XOR without neuralfit unit
- Simple CIFAR-10 Image Classifier
- Simple CIFAR-10 Image Classifier with OpenCL
- Many neural network architectures for CIFAR-10 image classification
- MNIST, Fashion MNIST and CIFAR-100
There are some available videos:
- Increasing Image Resolution with Neural Networks
- Ultra Fast Single Precision Floating Point Computing
- AVX and AVX2 Code Optimization
Some videos make referrence to uvolume unit. The current neuralvolume unit used to be called uvolume. This is why it's mentioned.
Although these examples require deeper understanding about neural networks, they are very interesting:
- DenseNetBC L40
- Separable Convolutions - MobileNet building block
- Identity Shortcut Connection - ResNet building block
- Gradient Ascent - Visualizing patterns from inner neurons in image classification
- Artificial Art - Let a neural network produce art via a generative adversarial network
- Super Resolution - A neural network learns how to increase image resolution
- Autoencoder - Shows an autoencoder built with hyperbolic tangents and trained with Tiny ImageNet 200.
There are also some older code examples that you can look at.
This API heavily relies on volumes. TNNetVolume
class allows you to create volumes that can be accessed as 1D, 2D or 3D arrays and be operated with AVX SIMD instruction set. This is the most common way to create a volume:
constructor Create(pSizeX, pSizeY, pDepth: integer; c: T = 0);
You can access the data as 1D or 3D with:
property Raw[x: integer]: T read GetRaw write SetRaw;
property Data[x, y, d: integer]: T read Get write Store; default;
Your code will look like this:
// Usage Examples
vInput := TNNetVolume.Create(32, 32, 3);
vInput[1, 1, 1] := 1;
vInput[2, 2, 2] := vInput[1, 1, 1] + 1;
vInput.Raw[10] := 5;
vInput.RandomizeGaussian();
WriteLn('Avg: ', vInput.GetAvg());
WriteLn('Variance: ', vInput.GetVariance());
WriteLn('Std Dev: ', vInput.GetStdDeviation());
WriteLn('Multiplying by 10');
vInput.Mul(10);
WriteLn('Avg: ', vInput.GetAvg());
WriteLn('Variance: ', vInput.GetVariance());
WriteLn('Std Dev: ', vInput.GetStdDeviation());
As examples, you can add, subtract, multiply and calculate dot products with:
procedure Add(Original: TNNetVolume); overload;
procedure Sub(Original: TNNetVolume); overload;
procedure Mul(Value: Single); overload;
function DotProduct(Original: TNNetVolume): TNeuralFloat; overload;
In the case that you need the raw position or raw pointer to an element of the volume, you can get with:
function GetRawPos(x, y, d: integer): integer; overload;
function GetRawPos(x, y: integer): integer; overload;
function GetRawPtr(x, y, d: integer): pointer; overload;
function GetRawPtr(x, y: integer): pointer; overload;
function GetRawPtr(x: integer): pointer; overload;
You can easily operate volumes with OpenCL via TEasyOpenCLV
:
TEasyOpenCLV = class (TEasyOpenCL)
public
function CreateBuffer(flags: cl_mem_flags; V: TNNetVolume): cl_mem; overload;
function CreateInputBuffer(V: TNNetVolume): cl_mem; overload;
function CreateHostInputBuffer(V: TNNetVolume): cl_mem; overload;
function CreateOutputBuffer(V: TNNetVolume): cl_mem; overload;
function CreateBuffer(V: TNNetVolume): cl_mem; overload;
function WriteBuffer(buffer: cl_mem; V: TNNetVolume; blocking: cl_bool = CL_FALSE): integer;
function ReadBuffer(buffer: cl_mem; V: TNNetVolume; blocking: cl_bool = CL_TRUE): integer;
function CreateAndWriteBuffer(V: TNNetVolume; var buffer: cl_mem): integer; overload;
function CreateAndWriteBuffer(V: TNNetVolume): cl_mem; overload;
function CreateWriteSetArgument(V: TNNetVolume; kernel:cl_kernel; arg_index: cl_uint): cl_mem;
function CreateOutputSetArgument(V: TNNetVolume; kernel:cl_kernel; arg_index: cl_uint): cl_mem;
end;
Volumes can be organized in pairs:
/// Implements a pair of volumes
TNNetVolumePair = class(TObject)
protected
FA: TNNetVolume;
FB: TNNetVolume;
public
constructor Create(); overload;
constructor Create(pA, pB: TNNetVolume); overload;
constructor CreateCopying(pA, pB: TNNetVolume); overload;
destructor Destroy(); override;
property A:TNNetVolume read FA;
property B:TNNetVolume read FB;
property I:TNNetVolume read FA;
property O:TNNetVolume read FB;
end;
Depending on the problem that you are trying to solve, modelling the training with pairs or pair lists might be helpful. Typically, a pair will be (input, desired output). This is how volume lists and volume pair lists have been implemented:
TNNetVolumeList = class (specialize TFPGObjectList<TNNetVolume>
TNNetVolumePairList = class (specialize TFPGObjectList<TNNetVolumePair>)
This API is really big. The following list gives a general idea about this API but it doesn't contain everything.
TNNetInput
(input/output: 1D, 2D or 3D).
TNNetConvolution
(input/output: 1D, 2D or 3D - feature size: 1D or 2D).TNNetConvolutionReLU
(input/output: 1D, 2D or 3D - feature size: 1D or 2D).TNNetConvolutionLinear
(input/output: 1D, 2D or 3D - feature size: 1D or 2D).TNNetPointwiseConvReLU
(input/output: 1D, 2D or 3D).TNNetPointwiseConvLinear
(input/output: 1D, 2D or 3D).TNNetDepthwiseConv
(input/output: 1D, 2D or 3D).TNNetDepthwiseConvReLU
(input/output: 1D, 2D or 3D).TNNetDepthwiseConvLinear
(input/output: 1D, 2D or 3D).TNNet.AddSeparableConvReLU
(input/output: 1D, 2D or 3D). Adds a separable convolution.TNNet.AddSeparableConvLinear
(input/output: 1D, 2D or 3D). Adds a separable convolution.TNNet.AddConvOrSeparableConv
(input/output: 1D, 2D or 3D). Adds a convolution or a separable convolutions with/without ReLU and normalization.
TNNetFullConnect
(input/output: 1D, 2D or 3D).TNNetFullConnectReLU
(input/output: 1D, 2D or 3D).TNNetFullConnectLinear
(input/output: 1D, 2D or 3D).TNNetFullConnectSigmoid
(input/output: 1D, 2D or 3D).
TNNetLocalConnect
(input/output: 1D, 2D or 3D - feature size: 1D or 2D).TNNetLocalConnectReLU
(input/output: 1D, 2D or 3D - feature size: 1D or 2D).
TNNetAvgPool
(input/output: 1D, 2D or 3D).TNNetMaxPool
(input/output: 1D, 2D or 3D).TNNetMinPool
(input/output: 1D, 2D or 3D).TNNet.AddMinMaxPool
(input/output: 1D, 2D or 3D). Does both min and max pools and then concatenates results.TNNet.AddAvgMaxPool
(input/output: 1D, 2D or 3D ). Does both average and max pools and then concatenates results.
TNNetAvgChannel
(input: 2D or 3D - output: 1D). Calculates the channel average.TNNetMaxChannel
(input: 2D or 3D - output: 1D). Calculates the channel max.TNNetMinChannel
(input: 2D or 3D - output: 1D). Calculates the channel min.TNNet.AddMinMaxChannel
(input/output: 1D, 2D or 3D). Does both min and max channel and then concatenates results.TNNet.AddAvgMaxChannel
(input/output: 1D, 2D or 3D). Does both average and max channel and then concatenates results.
TNNetChannelZeroCenter
(input/output: 1D, 2D or 3D). Trainable zero centering.TNNetMovingStdNormalization
(input/output: 1D, 2D or 3D). Trainable std. normalization.TNNetChannelStdNormalization
(input/output: 1D, 2D or 3D). Trainable per channel std. normalization.TNNet.AddMovingNorm
(input/output: 1D, 2D or 3D). Possible replacement for batch normalization.TNNet.AddChannelMovingNorm
(input/output: 1D, 2D or 3D). Possible replacement for per batch normalization.
TNNetLayerMaxNormalization
(input/output: 1D, 2D or 3D).TNNetLayerStdNormalization
(input/output: 1D, 2D or 3D).TNNetLocalResponseNorm2D
(input/output: 2D or 3D).TNNetLocalResponseNormDepth
(input/output: 2D or 3D).TNNetRandomMulAdd
(input/output: 1D, 2D or 3D). Adds a random multiplication (scale) and a random bias (shift).TNNetChannelRandomMulAdd
(input/output: 1D, 2D or 3D). Adds a random multiplication (scale) and random bias (shift) per channel.
TNNetConcat
(input/output: 1D, 2D or 3D). Allows concatenating previous layers into a single layer.TNNetDeepConcat
(input/output: 1D, 2D or 3D). Concatenates previous layers into the depth axis. This is useful with DenseNet like architectures. UseTNNetDeepConcat
instead ofTNNetConcat
if you need to add convolutions after concating layers.TNNetIdentity
(input/output: 1D, 2D or 3D).TNNetIdentityWithoutBackprop
(input/output: 1D, 2D or 3D). Allows the forward pass to proceed but prevents backpropagation.TNNetReshape
(input/output: 1D, 2D or 3D).TNNetSplitChannels
(input: 1D, 2D or 3D / output: 1D, 2D or 3D). Splits layers/channels from the input.TNNetSum
(input/output: 1D, 2D or 3D). Sums outputs from previous layers allowing ResNet style networks.
TNNetReLU
(input/output: 1D, 2D or 3D).TNNetLeakyReLU
(input/output: 1D, 2D or 3D).TNNetVeryLeakyReLU
(input/output: 1D, 2D or 3D).TNNetReLUSqrt
(input/output: 1D, 2D or 3D).TNNetSELU
(input/output: 1D, 2D or 3D).TNNetSigmoid
(input/output: 1D, 2D or 3D).TNNetSoftMax
(input/output: 1D, 2D or 3D).TNNetHyperbolicTangent
(input/output: 1D, 2D or 3D).TNNetPower
(input/output: 1D, 2D or 3D).
Trainable Bias (Shift) and Multiplication (Scaling) per Cell or Channel Allowing Faster Learning and Convergence
TNNetCellBias
(input/output: 1D, 2D or 3D).TNNetCellMul
(input/output: 1D, 2D or 3D).TNNetChannelBias
(input/output: 1D, 2D or 3D).TNNetChannelMul
(input/output: 1D, 2D or 3D).
TNNetDeLocalConnect
(input/output: 1D, 2D or 3D - feature size: 1D or 2D).TNNetDeLocalConnectReLU
(input/output: 1D, 2D or 3D - feature size: 1D or 2D).TNNetDeconvolution
(input/output: 1D, 2D or 3D - feature size: 1D or 2D).TNNetDeconvolutionReLU
(input/output: 1D, 2D or 3D - feature size: 1D or 2D).TNNetDeMaxPool
(input/output: 1D, 2D or 3D - max is done on a single layer).
InitUniform(Value: TNeuralFloat = 1)
.InitLeCunUniform(Value: TNeuralFloat = 1)
.InitHeUniform(Value: TNeuralFloat = 1)
.InitHeUniformDepthwise(Value: TNeuralFloat = 1)
.InitHeGaussian(Value: TNeuralFloat = 0.5)
.InitHeGaussianDepthwise(Value: TNeuralFloat = 0.5)
.InitGlorotBengioUniform(Value: TNeuralFloat = 1)
.
procedure FlipX();
procedure FlipY();
procedure CopyCropping(Original: TVolume; StartX, StartY, pSizeX, pSizeY: integer);
procedure CopyResizing(Original: TVolume; NewSizeX, NewSizeY: integer);
procedure AddGaussianNoise(pMul: TNeuralFloat);
procedure AddSaltAndPepper(pNum: integer; pSalt: integer = 2; pPepper: integer = -2);
You can add layers one by one or you can add an array of layers in one go. Follows an example adding layers one by one:
NN.AddLayer(TNNetConvolutionReLU.Create({Features=}64, {FeatureSize=}5, {Padding=}2, {Stride=}1));
NN.AddLayer(TNNetMaxPool.Create(2));
The next example shows how to add an array of layers that is equivalent to the above example:
NN.AddLayer([
TNNetConvolutionReLU.Create({Features=}64, {FeatureSize=}5, {Padding=}2, {Stride=}1),
TNNetMaxPool.Create(2)
]);
Since 2017, this API supports multi-paths architectures. You can create multi-paths with AddLayerAfter
method. For concatenating (merging) paths, you can call either TNNetConcat
or TNNetDeepConcat
. Follows an example:
// Creates The Neural Network
NN := TNNet.Create();
// This network splits into 2 paths and then is later concatenated
InputLayer := NN.AddLayer(TNNetInput.Create(32, 32, 3));
// First branch starting from InputLayer (5x5 features)
NN.AddLayerAfter(TNNetConvolutionReLU.Create({Features=}16, {FeatureSize=}5, {Padding=}2, {Stride=}1), InputLayer);
NN.AddLayer(TNNetMaxPool.Create(2));
NN.AddLayer(TNNetConvolutionReLU.Create({Features=}64, {FeatureSize=}5, {Padding=}2, {Stride=}1));
NN.AddLayer(TNNetMaxPool.Create(2));
EndOfFirstPath := NN.AddLayer(TNNetConvolutionReLU.Create({Features=}64, {FeatureSize=}5, {Padding=}2, {Stride=}1));
// Another branch starting from InputLayer (3x3 features)
NN.AddLayerAfter(TNNetConvolutionReLU.Create({Features=}16, {FeatureSize=}3, {Padding=}1, {Stride=}1), InputLayer);
NN.AddLayer(TNNetMaxPool.Create(2));
NN.AddLayer(TNNetConvolutionReLU.Create({Features=}64, {FeatureSize=}3, {Padding=}1, {Stride=}1));
NN.AddLayer(TNNetMaxPool.Create(2));
EndOfSecondPath := NN.AddLayer(TNNetConvolutionReLU.Create({Features=}64, {FeatureSize=}3, {Padding=}1, {Stride=}1));
// Concats both branches into one branch.
NN.AddLayer(TNNetDeepConcat.Create([EndOfFirstPath, EndOfSecondPath]));
NN.AddLayer(TNNetConvolutionReLU.Create({Features=}64, {FeatureSize=}3, {Padding=}1, {Stride=}1));
NN.AddLayer(TNNetLayerFullConnectReLU.Create(64));
NN.AddLayer(TNNetLayerFullConnectReLU.Create(NumClasses));
These source code examples show AddLayerAfter
:
- DenseNetBC L40
- Identity Shortcut Connection - ResNet building block
You can find more about multi-path architectures at:
These datasets can be easily loaded:
procedure CreateCifar10Volumes(out ImgTrainingVolumes, ImgValidationVolumes, ImgTestVolumes: TNNetVolumeList);
Source code example: Simple CIFAR-10 Image Classifier
procedure CreateCifar100Volumes(out ImgTrainingVolumes, ImgValidationVolumes, ImgTestVolumes: TNNetVolumeList);
Source code example: CAI Optimized DenseNet CIFAR-100 Image Classifier
procedure CreateMNISTVolumes(out ImgTrainingVolumes, ImgValidationVolumes,
ImgTestVolumes: TNNetVolumeList;
TrainFileName, TestFileName: string;
Verbose:boolean = true;
IsFashion:boolean = false);
Source code examples:
In the case that your dataset has one class per folder, you can call CreateVolumesFromImagesFromFolder for loading your data into RAM:
// change ProportionToLoad to a smaller number if you don't have enough RAM.
ProportionToLoad := 1;
WriteLn('Loading ', Round(ProportionToLoad*100), '% of the Plant leave disease dataset into memory.');
CreateVolumesFromImagesFromFolder
(
ImgTrainingVolumes, ImgValidationVolumes, ImgTestVolumes,
{FolderName=}'plant', {pImageSubFolder=}'',
{color_encoding=}csRGB{RGB},
{TrainingProp=}0.9*ProportionToLoad,
{ValidationProp=}0.05*ProportionToLoad,
{TestProp=}0.05*ProportionToLoad,
{NewSizeX=}128, {NewSizeY=}128
);
The example above shows how to load the dataset with 90% loaded into training and 5% loaded for each validation and testing. Images are being resized to 128x128.
Source code examples:
When loading an image from a file, the easiest and fastest method is calling LoadImageFromFileIntoVolume(ImageFileName:string; V:TNNetVolume)
. When loading from an TFPMemoryImage, you can load with LoadImageIntoVolume(M: TFPMemoryImage; Vol:TNNetVolume)
. For saving an image, the fastest method is SaveImageFromVolumeIntoFile(V: TNNetVolume; ImageFileName: string)
.
The easiest way to train your neural network is utilizing unit neuralfit.pas
. Inside this unit, you’ll find the class TNeuralImageFit
that is used by many examples.
TNeuralImageFit
has been designed for image classification tasks and can be called as follows:
procedure Fit(pNN: TNNet;
pImgVolumes, pImgValidationVolumes, pImgTestVolumes: TNNetVolumeList;
pNumClasses, pBatchSize, Epochs: integer);
Each volume should be provided with property tag
that contains the corresponding class. TNeuralImageFit
internally implements data augmentation techniques: flipping, making gray, cropping and resizing. These techniques can be controlled with:
property HasImgCrop: boolean read FHasImgCrop write FHasImgCrop;
property HasMakeGray: boolean read FHasMakeGray write FHasMakeGray;
property HasFlipX: boolean read FHasFlipX write FHasFlipX;
property HasFlipY: boolean read FHasFlipY write FHasFlipY;
property MaxCropSize: integer read FMaxCropSize write FMaxCropSize;
Once you have a trained neural network, you can use an advanced classification procedure that will average the classification probability of the input image with its flipped and cropped versions. This process frequently gives a higher classification accuracy at the expense of internally running the very same neural network a number of times. This is how you can classify images:
procedure ClassifyImage(pNN: TNNet; pImgInput, pOutput: TNNetVolume);
In the case that you would like to look into TNeuralImageFit
in more detail, the Simple CIFAR-10 Image Classifier example is a good starting point.
In the case that your training, validation and testing data can be defined as volume pairs from input volume to output volume, the easiest way to train your neural network will be calling TNeuralFit
. This class has the following fitting method:
procedure Fit(pNN: TNNet;
pTrainingVolumes, pValidationVolumes, pTestVolumes: TNNetVolumePairList;
pBatchSize, Epochs: integer);
Both AND, OR and XOR with neuralfit unit and hypotenuse function examples load volume pair lists for training.
The TNeuralFit
implementation has a limitation: your dataset needs to be placed into RAM. In the case that your dataset is too large for RAM, you can call TNeuralDataLoadingFit
:
TNNetGetPairFn = function(Idx: integer; ThreadId: integer): TNNetVolumePair of object;
TNNetGet2VolumesProc = procedure(Idx: integer; ThreadId: integer; pInput, pOutput: TNNetVolume) of object;
TNeuralDataLoadingFit = class(TNeuralFitBase)
...
procedure FitLoading(pNN: TNNet;
TrainingCnt, ValidationCnt, TestCnt, pBatchSize, Epochs: integer;
pGetTrainingPair, pGetValidationPair, pGetTestPair: TNNetGetPairFn); overload;
procedure FitLoading(pNN: TNNet;
TrainingCnt, ValidationCnt, TestCnt, pBatchSize, Epochs: integer;
pGetTrainingProc, pGetValidationProc, pGetTestProc: TNNetGet2VolumesProc); overload;
The Hypotenuse with FitLoading example uses TNeuralDataLoadingFit
so it creates training pairs on the fly.
TNeuralImageFit
and TNeuralDataLoadingFit
both descend from TNeuralFitBase
. From TNeuralFitBase
, you can define training properties:
property Inertia: single read FInertia write FInertia;
property InitialEpoch: integer read FInitialEpoch write FInitialEpoch;
property InitialLearningRate: single read FInitialLearningRate write FInitialLearningRate;
property LearningRateDecay: single read FLearningRateDecay write FLearningRateDecay;
property CyclicalLearningRateLen: integer read FCyclicalLearningRateLen write FCyclicalLearningRateLen;
property Momentum: single read FInertia write FInertia;
property L2Decay: single read FL2Decay write FL2Decay;
property FileNameBase: string read FFileNameBase write FFileNameBase;
You can also collect current statistics:
property CurrentEpoch: integer read FCurrentEpoch;
property CurrentStep: integer read FCurrentStep;
property CurrentLearningRate: single read FCurrentLearningRate;
property TestAccuracy: TNeuralFloat read FTestAccuracy;
property TrainingAccuracy: TNeuralFloat read FTrainingAccuracy;
property Running: boolean read FRunning;
Some events are available:
property OnStart: TNotifyEvent read FOnStart write FOnStart;
property OnAfterStep: TNotifyEvent read FOnAfterStep write FOnAfterStep;
property OnAfterEpoch: TNotifyEvent read FOnAfterEpoch write FOnAfterEpoch;
You can define your own learning rate schedule:
property CustomLearningRateScheduleFn: TCustomLearningRateScheduleFn read FCustomLearningRateScheduleFn write FCustomLearningRateScheduleFn;
property CustomLearningRateScheduleObjFn: TCustomLearningRateScheduleObjFn read FCustomLearningRateScheduleObjFn write FCustomLearningRateScheduleObjFn;
TNeuralFitBase
descends from TMObject
that allows you to code your own message treatment:
property MessageProc: TGetStrProc read FMessageProc write FMessageProc;
property ErrorProc: TGetStrProc read FErrorProc write FErrorProc;
On your own code, you could something is:
MyFit.MessageProc := {$IFDEF FPC}@{$ENDIF}Self.MessageProc;
MyFit.ErrorProc := {$IFDEF FPC}@{$ENDIF}Self.ErrorProc;
If you don’t need any message at all, you can hide messages by calling:
procedure HideMessages();
You can also disable fitting verbosity with:
property Verbose: boolean read FVerbose write FVerbose;
Your code will look like this:
NeuralFit := TNeuralImageFit.Create;
...
NeuralFit.Verbose := false;
NeuralFit.HideMessages();
In the case that you need help with your own A.I. project (Pascal, Python, PHP or Java), please feel free to contact me.
Pull requests are welcome. Having requests accepted might be hard.