Artificial Intelligence TechC++Learn C++

AI Techs :: A Simple Artificial Neuron Model in C++

A Simple Artificial Neuron

Do you want to develop your of artificial intelligence application from the scratch ? Want to learn how you can develop a simple artificial neuron model in C++? In this post we will explain with a very simple artificial neuron example.

A Minimum Artificial Neuron has an activation value (a), an activation function ( phi() ) and weighted (w) input net links. So it has one activation value, one activation function and one or more weights depends on number of it’s input nets.

Activation value (a) is a result of activation function also called as transfer function or threshold.

Activation Function ( phi() ) also called as transfer function, or threshold function that determines the activation value ( a = phi(sum) ) from a given value (sum). Here sum is a sum of signals in their weights. In another terms. Activation function is a way to transfer the sum of all weighted signals to a new activation value of that signal. There are different activation functions, mostly Linear (Identity) , bipolar and logistic (sigmoid) function are used. Activation function and it’s types are explained well here.

Weights are weight of a input signal for that neuron. Every net link has a link affecting to other neurons sum.

It is analogous to the axon of a biological neuron, and its value propagates to the input of the next layer, through a synapse. In Artificial Intelligence Theory the main and the core formula that simulates all neural activity is described with one signal formula. Activation function (transfer function), transfers sum input to a activation value as given below,

In some old articles here sum showed as a transfer function or result of a transfer function, now most articles uses activations function as transfer function, both same. It transfers given sum to activation value. So sum can be a result of a sum() function but not a a result of a transfer function, it will be used in a transfer function (activation function) to define activation output. This formula calculates the current activation value of an artificial neuron with activation function corresponding with the sum of activations in their weights. In this formula,,

From the this very brief theory, we can create a simple neuron model as below used in a simple ANN example,

Very Simple Artificial Neural Network Example in C++

Here below we prepared a a very simple Artificial Neural Network example with an Activation Function given,

In this ANN example we have 2 input neurons and one output neuron. Neuron 0 and Neuron 1 are connected with Neuron 2. Each neuron will have their activation value (a) and each of the links between neurons will have weights. So output neuron will be the result of activation of the sum of 2 input neuron activations multiplied in their weights.

Output of the neuron may have a value in this current state. We should calculate it’s new activation value by the activation function phi(). Let’s define a simple linear activation function

Here we assume that our network was trained before, that means weights of net links are known. Let’s define and initialize our all neural activation and weight values,

Note that, the number of neural net links are not same with number of neurons. That means, links and weights can be handled in another array. We defined a2 before and we set it to 0 which is not correct in this network, our ANN should be stable, correct ! So we need to calculate the activation value of the output with these inputs. We should use our activation function with the sum of incoming signals in their weights.

As you see, this simple equation shows that activation value of an artificial neuron depends on activation values of input neurons and weights of input links.

We think that, this example above was the simplest artificial neuron and ANN example in C++ on the internet. It is good to understand how a simple neuron simulation works. Let’s sum all in this full example below,

Get started building powerful apps with C++Builder!

Oh hi there 👋
It’s nice to meet you.

Sign up to receive awesome C++ content in your inbox, every day.

We don’t spam! Read our privacy policy for more info.

About author

Dr. Yilmaz Yoru has 35+ years of coding with more than 30+ programming languages, mostly C++ on Windows, Android, Mac-OS, iOS, Linux, and some other operating systems. He graduated and received his MSc and PhD degrees from the Department of Mechanical Engineering of Eskisehir Osmangazi University. He is the founder and CEO of ESENJA LLC Company. His interests are Programming, Thermodynamics, Fluid Mechanics, Artificial Intelligence, 2D & 3D Designs, and high-end innovations.
Related posts
C++C++11C++14C++17C++20

What Is The Stack (std::stack) In Modern C++?

C++C++11C++14C++17C++20Learn C++

What Is The Queue (std::queue) In Modern C++?

C++C++11C++14C++17Learn C++SyntaxTemplates

What Are The Logical Operation Metafunctions In Modern C++?

C++C++14C++17C++20Learn C++

What Are The Deprecated C++14 Features In C++17?

Worth reading...
AI Techs :: Introduction to Artificial Intelligence in C++