<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:dc="http://purl.org/dc/elements/1.1/" version="2.0">
<channel>
<title>OVER-FITTING AMELIORATION IN BAYESIAN NEURAL NETWORK MODEL ESTIMATION USING HETEROGENEOUS ACTIVATION FUNCTIONS</title>
<link>http://hdl.handle.net/123456789/2150</link>
<description/>
<pubDate>Sat, 04 Apr 2026 22:51:55 GMT</pubDate>
<dc:date>2026-04-04T22:51:55Z</dc:date>
<item>
<title>OVER-FITTING AMELIORATION IN BAYESIAN NEURAL NETWORK MODEL ESTIMATION USING HETEROGENEOUS ACTIVATION FUNCTIONS</title>
<link>http://hdl.handle.net/123456789/2151</link>
<description>OVER-FITTING AMELIORATION IN BAYESIAN NEURAL NETWORK MODEL ESTIMATION USING HETEROGENEOUS ACTIVATION FUNCTIONS
OGUNDUNMADE, Tayo Peter
Neural Network (NN) allows complex nonlinear relationships between the response&#13;
variables and its predictors. The Deep NN have made notable contributions across&#13;
computer vision, reinforcement learning, speech recognition and natural language&#13;
processing. Previous studies have obtained the parameters of NN through the classical approach using Homogeneous Activation Functions (HOMAFs). However, a&#13;
major setback of NN using the classical approach is its tendency to over-fit. Therefore, this study was aimed at developing a Bayesian NN (BNN) model to ameliorate&#13;
over-fitting using Heterogeneous Activation Functions (HETAFs).&#13;
A BNN model was developed with Gaussian error distribution for the likelihood&#13;
function; inverse gamma and inverse Wishart priors for the parameters, to obtain&#13;
the BNN estimators. The HOMAFs (Rectified Linear Unit (ReLU), Sigmoid and&#13;
Hyperbolic Tangent Sigmoid (TANSIG)) and HETAFs (Symmetric Saturated Linear Hyperbolic Tangent (SSLHT) and Symmetric Saturated Linear Hyperbolic Tangent Sigmoid (SSLHTS)) were used to activate the model parameters.The Bayesian&#13;
approach was used to ameliorate the problem of over-fitting, while the Posterior&#13;
Mean (PM), Posterior Standard Deviation (PSD) and Numerical Standard Error&#13;
(NSE) were used to determine the estimators’ sensitivity. The performance of the&#13;
Bayesian estimators from each of the activation functions was evaluated in the&#13;
Monte Carlo experiment using the Mean Square Error (MSE), Mean Absolute Error (MAE) and training error as metrics. The proximity of MSE and training error&#13;
values were used to generalise on the problem of over-fitting.&#13;
The derived Bayesian estimators were β ∼ N(Kβ, Hβ) and γ ∼ exp (−1 2{Fγ +Mγ);&#13;
where Kβ is derived mean of β, Hβ is derived standard deviation of β; Fγ and&#13;
Mγ&#13;
are the derived posteriors of γ. For ReLU, the PM, PSD and NSE values for&#13;
β and γ were 0.4755, 0.0646, 0.0020; and 0.2370, 0.0642, 0.0020, respectively; for&#13;
Sigmoid: 0.4476, 0.2734, 0.0087; and 1.0269, 0.2732, 0.0086, respectively; for TANSIG: 0.4718, 0.0826, 0.0026, and 1.0239, 0.0822, 0.0026, respectively. For SSLHT,&#13;
the PM, PSD and NSE values for β and γ were 0.8344, 0.0567, 0.0018; and 1.0242,&#13;
0.0566, 0.0016, respectively; and for SSLHTS: 0.89825, 0.01278, 0.0004; and 1.0236,&#13;
v0.0127, 0.0003, respectively. The MSE, MAE and training error values for the performance of the activation functions were ReLU: 0.1631, 0.2465, 0.1522; Sigmoid:&#13;
0.1834, 0.2074, 0.1862; TANSIG: 0.1943, 0.269, 0.1813; SSLHT: 0.0714, 0.0131,&#13;
0.0667; and SSLHTS: 0.0322, 0.0339, 0.0328, respectively. The HETAFs showed&#13;
closer proximity between MSE and training error implying amelioration of overfitting and minimum error values compared to HOMAFS.&#13;
The derived Bayesian neural network estimators ameliorated the problem of overfitting with close values of Mean Square Error and training error, thus making&#13;
them more appropriate in handling Neural Network models. They could be used&#13;
in solving problems in machine learning.
</description>
<pubDate>Wed, 16 Aug 2023 00:00:00 GMT</pubDate>
<guid isPermaLink="false">http://hdl.handle.net/123456789/2151</guid>
<dc:date>2023-08-16T00:00:00Z</dc:date>
</item>
</channel>
</rss>
