Tuesday, June 4, 2019
The Science Of Artificial Neural Networks Psychology Essay
The Science Of ersatz Neural Networks Psychology EssayThe science of Artificial Neural Networks (ANNs), ordinarily referred as Neural Networks, stills a new and promising bea of research. The concept of creation of neural networks exists for legion(predicate) decades. Nevertheless neural networks have become known and have been developed in international levels only in the recent years. It is noteworthy, scientist showing interest in neural networks, come from various scientific areas such as chemistry, medicine, physics, mathematics, engineering and the disceptation goes on. That shows Neural Networks is a new ch eachenge in science. No other science today combines and needs direct knowledge from such diverse areas. nonpareil of the main differences of the Artificial Neural Networks from the biological is that while ANNs learn through schooling and experience just like the biological ones but they follow different rules from regular computers. A Neural Network is a parallel data processing system consisted by multitude of artificial neurons, organized in social organisations standardized with the ones in human beings brain. They influence as parallel computing devices made by umteen highly interconnected fair processors. Artificial Neurons are generally organized in layers. The first of those layers called the stimulus layer and is used to insert the data. Input layers are unable to proceed to any sort of computation as its elements do not comprise input predict weights or bias (threshold).The axon mean of transfer of neural signals from the neuron. Its length can be tens of thousands of times the diameter of its body and it is characterized by high electrical resistance and very large capacitance. Every neuron has only one axon, however it can branch and hence enabling communication with many target cells or other neurons.The dendrite short highly branched cell projections (filaments). Most neurons have many dendrites, attached on the som a and increase the surface area. thither are approximately 103 to 104 dendrites per neuron, to receive information from other neurons through synapses they are covered with and transmit electrochemical stimulation to the soma.The axon terminal located in the end of the axon and is responsible for transmitting signals on to other neurons. On axon terminals are attached the terminal buttons, that store the information in synaptic vesicles and secreting them in neurotransmitters.As mentioned above, the connection between neurons happens through the synapses. Neural synapses are a silent exchange of information. The electrical nerve impulses travel along neurons and transmitted by chemical transmitters (neurotransmitters) in the next neuron across a tiny gap, the synapses and are located between the neuron and the neighboring cell (target cell). Therefore dendrites are very close to each other but neer in touch. It is estimated that there are approximately 10 billion neurons in the hum an cortex, and 60 trillion synapses or connections (Shepherd and Koch, 1990).A number of neurons and their connections form a neural network. The entire system of neural networks in the human body forms the Central Nervous System. This system goes through the whole human body with profound points the brain and the spine. During lifetime, synapses are in constant dynamic chemical equilibrium, new are created and old are destroyed. The creation of new synapse happens when the brain acquires more experiences from the surrounding environment, learns, recognizes and understands. On the other side, diseases cause the devastation of neurons and therefore the devastation of synapses.In comparison to other cells, neurons might not replaced by new ones, if destroyed. That means after the lineage of a new individual, its neural system is fully developed within the first few months of its life.A neuron can be either restless or inactive. When it is activated, it produces an electric signal. This signal has intensity of only a few mVolt.The way those electric signals are produced is pretty similar with the way a capacitor works. Between the external and internal surface of the cell of the neuron there is a dynamic difference.Although the mass of the human brain is only the 2% of human body mass, con spousal relationshipes more than the 20% of the oxygen that goes in the organism. The power con mettleption in the brain is some 20 Watt in comparison to a computer that needs a lot more.The computational power of brain is measured by three possible approachesThe number of synapses (Kandel, 1985), the computational power of the retina and cypher it by their brain-to-retina ratio (Moravec, 1998b), and the total useful energy used by the brain per second by the amount of energy used for each canonic operation to give the maximum operation per second (Merkle, 1989)From the three approaches above, is concluded that the estimated computational power of human brain is about 1014 operations per second (Ng, 2009).It is interesting to mention how the electric pulses are created to stimulate neurons. On the membrane of the cell it is appeared to be an electric potential difference between its external and internal surface just like a capacitor. Most of the times the negative charges found in the internal surface as they cant broadcast the membrane and leave the cell.The membrane has many openings that allow ions and atoms to go through each element from its own channel. The endings of the channels are secured by gates which direct the flow of those elements. Proteins that act like pumps force the elements to travel in the opposite direction from their natural and thus neurons consume larger amounts of energy. Eventually the balanced drive of the elements along the surface of the membrane produces an electric current which is the corresponding electrical pulse that stimulates the neuron.Once the neuron has fired it returns to a state of potential equili brium and in this state it cant be fired again until it recovers.Each neuron has a specific threshold or weight. When electric signals reach that point, sum up and if their weight value is same or larger than the one of the threshold the neuron stimulates. If the sum of the signals is smaller than the required value of the threshold, then the neuron stays inactive.Add images.Models of artificial neuronsAs mentioned earlier, ANNs are parallel data processing systems, consisting out of large numbers of artificial neurons, inspired by the biological neurons.A neuron is an information-processing unit that is fundamental to the operation of a neural network (Haykin, 1999, pg-10).A neuron may have many inputs, an internal structure consisting out of multiple layers but it always has a single output.Every single neuron accepts variable input signals x0, x1, x2 xn. This corresponds to the electric pulses of the biological brain.Every input signal is multiplied by the synaptic weights of the neuron, wi, where i=1,2,3..n, the input nodes. The weights represent the biological synapses and indicate the strength of the bond (the connection) between them.The range of value of a weight can be positive or negative depending on if the function of synapse suspend or propagate (transmit) the stimuli from other neurons, unlike the biological synapses that do not take negative values. This is because external bias, b, are applied when the weights added.Bias or threshold, is the standard value of the internal potential energy of the neuron that the sum of the combined output must be reached in order the activation (or squashing) function to be activated.An important element of the neuronal body is the adder .At the adder, all the input signals, influenced by weight vectors are summing up together and produce a resultant combined output u. When the sum of weight is big (0Therefore, the harvest-festival u is given by the relationshipThe result of combined output u, pass through the activation function, denoted with the letter ( ).The activation function is a non linear function where the resultant combined output u takes its final value y.The calculated activation output signal of the neuron is shown asand whereTherefore,Activation functionsThere are several activation functions, however three of the most basic types are the following they slightly vary from book to book)The threshold activation function, which gives as an output 1, if the adder produce a value greater than the one of the threshold. This is expressed asThe Piecewise-Linear function, where the unity is assumed to be the amplification factor inside(a) the linear region of operation (Haykin, 1999, pg14)The Sigmoid function, which is expressed asWhere is is the slope parameter of the sigmoid function. This function is one of the most important and most commonly used as it provides non-linearity to the neuron.Some other activation functions are, the rump function, the bipolar sigmoid function, and the signum function.The signum function gives a positive or negative output, with values unremarkably ranging from 1 to -1 depending on the value of the summation of the weights on the threshold. This can be applied to the activation functions mentioned above and more specifically to the threshold faction freehandedAdd images and graphsA simple neural networkIn this paragraph, neural networks will be introduced, starting from their simplest form. Every neural network consists out of hundreds or thousands of tiny units, the neurons. Each neuron has an input where the electric signals are received. A neuron may have more than one input but no matter how many layers of neurons and synaptic connections are in between (the body), there is always one output value. The neurons of a layer between each input and output are not connected to each other however each layer is interconnected with the layer of the next and the previous level. In its simplest form, a neuron has no layers but is limited only to an input and an output. Every signal that leaves an output and enters an input has a value, the weights. The weights represent the importance of each signal reaching the threshold of an input. Depending on the value of weight (wn), the contribution of the electric signal can be great or small for the function of the system.Artificial intelligence and neural networksHistorical background(The study of the brain and the biological neurons has started thousands of years ago.) However, as artificial neural networks started to be developing the past century, the historic background still not as broad as in other sciences.The first union of mathematical logic and neuropsychology, commenced in 1943 by warren S. McCulloch and Walter Pitts.McCulloch was a pioneer neuroanatomist and psychiatrist. Pitts was a young mathematical prodigy, who joined McCulloch in 1942. (Haykin, 1999, pg 38).Together they created the first model a neural network that was represented by a great n umber of interconnected neurons. In their well-known paper, A logical calculus of the ideas immanent in nervous activity, (1943), came up with theorems that describe the function of neurons and the neural networks. As a result of those theorems, neural networks and artificial intelligence ideas established a new era of research began.The paper of McCulloch and Pitts, triggered the interest of many scientists like von Neumann, Wiener and Uttley in their effort to extract information of the function of biological neurons and create corresponding artificial ones.In 1949 another idea appeared by D. Hebb who published the book The Organisation of Behavior. Although his book had greater influence on the psychological rather than the engineering community, he introduced the concept of postulate and learning and the synaptic modification rule, which suggests that the connectivity of the brain changes continually thorough its entire life in the process of learning new tasks.From 1950 to 1979 , a number of remarkable books were written about neural networks developing the ideas of neurons abilities, such as learning and memorising.Some of these books are the Design for a Brain The origin of Adaptive Behaviors, (1952) by Ashby, that still fire to read nowadays, and the Learning Machines, (1965) by Nilsson, one of the best-written expositions about linearly separable patterns in hypersurfaces. (Haykin, 1999, pg 40).A novel model, the perceptron, introduced in 1958 by F. Rosenblatt. The perceptron is a very simple model of supervised learning, which has only one input and one output built around a nonlinear neuron (Haykin pg 135). Although this model appeared to have many limitations the idea of training the neurons encouraged many scientists for building larger neural networks.In 1969, Minsky and Papert in their book Perceptron they make a gross(a) evaluation of the features and uses of the perceptrons. It be with mathematics that there were fundamental limitations on t he computational ability of single-layered perceptrons and therefore those limitations assumed to carry on in the multilayered levels of perceptrons.A period followed were scientists start losing commit about neural networks and turned to other knowledge based systems.In 1982, neural networks make an interesting come back when John Hopfield proved in a strict mathematical way that by time a neural network can be adjusted to use the minimum energy to function just like human brain does. In addition, Hopfield proved that a simple neural network can be used as storage devise. Such networks are called the Hopfield networks.A very important work was published in 1986 by Rumelhart and McClelland. The two-volume book, Parallel Distributed Processing Explorations in the Microstructures of Cognitions, shows new methods of training neural networks and introduces the idea of parallel data processor. This theory had a great influence in the use of back-propagation learning as and allowed the d evelopment of multilayered networks (perceptrons).The books published by McCulloch- Pitts (1943), Hopfield (1982) and Rumelhart-McClelland (1986), are the most influential in the revolution of neural networks.Since 1980 to nowadays, Neural Networks have been established as a new independent science branch. Conferences and magazines appeared with complete interest on artificial neural networks while the first commercial companies dedicated to the improvement of them, created, supported by thousands of members worldwide especially in America, atomic number 63 and Japan.Learning processes/ trainingFundamental ideasThe present, looking to futureAnn applications areasAnns in civil engineeringCan it be applied in?Benefits/disadvantagesProgramObservationscommentssummaryreferences
Subscribe to:
Post Comments (Atom)
No comments:
Post a Comment