280 likes | 298 Views
Gain insights into the foundational concepts and models of neural networks, including neuron modeling, activation functions, network architectures, and feedback mechanisms. Dive into knowledge representation and challenges faced by neural networks in learning from and adapting to real-world environments.
E N D
KULIAH II JST:BASIC CONCEPTS Amer Sharif, S.Si, M.Kom
INTRODUCTION REVIEW • Neural Network definition: • A massively parallel distributed processor of simple processing units (neuron) • Store experiential knowledge and make it available for use • Knowledge is acquired from the environment through learning process • Knowledge is stored as internerneuron connection strengths (synaptic weights)
INTRODUCTION REVIEW • Benefits: • Nonlinearity • Input Output Mapping • Adaptivity • Evidential Response • Contextual Information • Fault Tolerance/Graceful Degrading • VLSI Implementability • Uniform Analysis & Design
NEURON MODELLING Basic elements of neuron: • A set of synapses or connecting links • Each synapse is characterized by its weight • Signal xjat synapse j connected to neuron kis multiplied by synaptic weight wkj • Biasis bk • An adder for summing the input signals • An activation function for limiting the output amplitude of the neuron
NEURON MODELLING • Block diagram of a nonlinier neuron
NEURON MODELLING Note • x1, x2,…, xmare input signals • wk1, wk2,…, wkmare synaptic weights of neuron k • uk is the linier combiner output • bk is bias • is the activation function • yk is the output signal of the neuron
NEURON MODELLING • If and bias is substituted for a synapse where x0= + 1 with weight wk0 = bk then and
NEURON MODELLING • Modified block diagram of a nonlinier neuron
1.2 1 0.8 0.6 0.4 0.2 0 -2 -1 0 1 2 v ACTIVATION FUNCTIONS Activation Function types: • Threshold Function and • also known as the McCulloch-Pitts model
ACTIVATION FUNCTIONS • Piecewise-Linear Function
ACTIVATION FUNCTIONS • Sigmoid Function • S-shaped • Sample logistic function: • ais the slope parameter: the larger athe steeper the function • Differentiable everywhere increasing a
wkj yk= wkj xj xj xj yi yk=yi + yj yj xj xj xj NEURAL NETWORKS AS DIRECTED GRAPHS • Neural networks maybe represented as directed graphs: • Synaptic links (linier I/O) • Activation links (nonlinier I/O) • Synaptic convergence • Synaptic divergence
x0 =+1 x1 Output yk x2 . . . xm NEURAL NETWORKS AS DIRECTED GRAPHS • Architectural graph: partially complete directed graph
FEEDBACK • Output of a system influences some of the input applied to the system • One or more closed paths of signal transmission around the system • Feedback plays an important role in recurrent networks
w x’j (n) yk(n) xj(n) z-1 FEEDBACK • Sample single-loop feedback system • Output signal yk(n) is an infinite weighted summation of present and past samples of input signal xj(n) wis fixed weight z-1is unit-delay operator is sample of input signal delayed by l time units
yk(n) wxj(0) 0 1 2 3 4 n FEEDBACK • Dynamic system behavior is determined by weight w w < 1 • w < 1 • System isexponentially convergent/stable • System posses infinite memory: Output depends on input samples extending into the infinite past • Memory is fading: influence of past samples is reduced exponentially with time n
yk(n) wxj(0) 0 1 2 3 4 n yk(n) wxj(0) 0 1 2 3 4 n FEEDBACK • w = 1 • System is linearly divergent • w > 1 • System is exponentially divergent w = 1 w > 1
input layer of source nodes output layer of neurons NETWORK ARCHITECTURES • Single-Layered Feedforward Networks • Neurons are organized in layers • “Single-layer” refers to output neurons • Source nodes supply to output neurons but not vice versa • Network is feedforward or acyclic
Layer of output neurons Layer of hidden neurons Input layer of source nodes NETWORK ARCHITECTURES • One or more hidden layers • Hidden neurons enable extractions of higher-order statistic • Network acquires global perspective due to extra set of synaptic connections and neural interactions • Multilayer Feedforward Networks • 7-4-2 fully connected network: • 7 source nodes • 4 hidden neurons • 2 output neurons
z-1 z-1 Outputs z-1 z-1 Unit-delay operators Inputs NETWORK ARCHITECTURES • Recurrent Networks • At least one feedback loop • Feedback loops affect learning capability and performance of the network
KNOWLEDGE REPRESENTATION • Definition of Knowledge: • Knowledge refers to stored information or models used by a person or a machine to interpret, predict, and appropriately respond to the outside world • Issues: • What information is actually made explicit • How information is physically encoded for subsequent use • Knowledge representation is goal-directed • Good solution depends on good representation of knowledge
KNOWLEDGE REPRESENTATION • Challenges faced by Neural Networks: • Learn the model of the world/environment • Maintain the model to be consistent with the real world to achieve the goals desired • Neural Networks may learn from a set of observations data in form of input-output pairs (training data/training sample) • Input is input signal and output is the corresponding desired response
KNOWLEDGE REPRESENTATION • Handwritten digit recognition problem • Input signal: one of 10 images of digits • Goal: to identify image presented to the network as input • Design steps: • Select the appropriate architecture • Train the network with subset of examples (learning phase) • Test the network with presentation of data/digit image not seen before, then compare response of network with actual identity of the digit image presented (generalization phase)
KNOWLEDGE REPRESENTATION • Difference with classical pattern-classifier: • Classical pattern-classifier design steps: • Formulate mathematical model of the problem • Validate model with real data • Build based on model • Neural Network design is: • Based on real life data • Data may “speak for itself” • Neural network not only provides model of the environment but also process the information
ARTIFICIAL INTELLIGENCE AND NEURAL NETWORKS • AI systems must be able to: • Store knowledge • Use stored knowledge to solve problem • Acquire new knowledge through experience • AI components: • Representation • Knowledge is presented in a language of symbolic structures • Symbolic representation makes it relatively easy for human users
ARTIFICIAL INTELLIGENCE AND NEURAL NETWORKS • Reasoning • Able to express and solve broad range of problems • Able to make explicit and implicit information known to it • Have a control mechanism to determine which operation for a particular problem, when a solution is obtained, or when further work on the problem must be terminated • Rules, Data, and Control: • Rules operate on Data • Control operate on Rules • The Travelling Salesman Problem: • Data: possible tours and cost • Rules: ways to go from city to city • Control: which Rules to apply and when
ARTIFICIAL INTELLIGENCE AND NEURAL NETWORKS • Learning • Inductive learning: determine rules from raw data and experience • Deductive learning:use rules to determine specific facts Performance element Knowlegdge Base Learning element Environment