Skip to content

kumbasart/Learning-with-type-2-fuzzy-activation-functions-to-improve-the-performance-of-deep-neural-networks

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

15 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Learning-with-type-2-fuzzy-activation-functions-to-improve-the-performance-of-deep-neural-networks

Matlab implementation of the composite learning algorithm in the following paper:

A. Beke and T. Kumbasar, "Learning with type-2 fuzzy activation functions to improve the performance of deep neural networks," Engineering Applications of Artificial Intelligence, vol. 85, pp. 372-384, 2019. 

We kindly ask that to cite the above mentioned paper if you use type-2 fuzzy activation functions and you publish papers about work that was performed using these codes.

How to use type-2 fuzzy activation functions in your model

  1. Download or clone the repository into a convenient directory. Open MATLAB 2019a or a newer version

  2. Add default layer folder to your path

  3. While constructing your model, just add 'SIT2FRU' or 'SIT2FMLayerOpt' class as an activation function and insert the appropriate layer size as an input argument. In the following, we give an example:

     FF_NetworkLayers = [sequenceInputLayer(size(inputRegressor{1}',1),'Name','FF-Giris')
     fullyConnectedLayer(60,'Name','RegressionLayer')
     sequenceFoldingLayer
     fullyConnectedLayer(6,'Name','FF-Output2')
     SIT2FRU(6, 'SIT2')
     fullyConnectedLayer(1,'Name','FF-Output')
     SIT2FRU(1, 'SIT22')
     regressionLayer];
    
  4. 'SIT2FRU' or 'SIT2FMLayerOpt' classes baically do the some job, the only difference 'SIT2FMLayerOpt' is the time optimized version. The functions takes 2 arguments layer size and customizable name.

  5. We also provide two dataset examples that are ICEngine and MNIST. ICEngine is a time series while MNIST is an image dataset. Besides, we provide classification and regression examples on MNIST dataset.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • MATLAB 98.9%
  • M 1.1%