SlideShare uma empresa Scribd logo
1 de 8
Baixar para ler offline
Getting Started with Neural Network Toolbox 4:20
Use graphical tools to apply neural networks to data fitting, pattern recognition, clustering,
and time series problems.
Neural Network Toolbox
Create, train, and simulate neural networks
Neural Network Toolbox™ provides functions and apps for modeling complex nonlinear systems that are not
easily modeled with a closed-form equation. Neural Network Toolbox supports supervised learning with
feedforward, radial basis, and dynamic networks. It also supports unsupervised learning with self-organizing maps
and competitive layers. With the toolbox you can design, train, visualize, and simulate neural networks. You can
use Neural Network Toolbox for applications such as data fitting, pattern recognition, clustering, time-series
prediction, and dynamic system modeling and control.
To speed up training and handle large data sets, you can distribute computations and data across multicore
processors, GPUs, and computer clusters using Parallel Computing Toolbox™.
Key Features
▪ Supervised networks, including multilayer, radial basis, learning vector quantization (LVQ), time-delay,
nonlinear autoregressive (NARX), and layer-recurrent
▪ Unsupervised networks, including self-organizing maps and competitive layers
▪ Apps for data-fitting, pattern recognition, and clustering
▪ Parallel computing and GPU support for accelerating training (using Parallel Computing Toolbox)
▪ Preprocessing and postprocessing for improving the efficiency of network training and assessing network
performance
▪ Modular network representation for managing and visualizing networks of arbitrary size
▪ Simulink®
blocks for building and evaluating neural networks and for control systems applications
Data Fitting, Clustering, and Pattern Recognition
Like its counterpart in the biological nervous system, a neural network can learn and therefore can be trained to
find solutions, recognize patterns, classify data, and forecast future events. The behavior of a neural network is
defined by the way its individual computing elements are connected and by the strengths of those connections, or
weights. The weights are automatically adjusted by training the network according to a specified learning rule
until it performs the desired task correctly.
Neural Network Toolbox includes command-line functions and apps for creating, training, and simulating neural
networks. The apps make it easy to develop neural networks for tasks such as data-fitting (including time-series
data), pattern recognition, and clustering. After creating your networks in these tools, you can automatically
generate MATLAB®
code to capture your work and automate tasks.
1
House Pricing Estimation with Neural Net Fitting App 4:21
Estimate median house prices for neighborhoods based on various neighborhood attributes.
Iris Flower Clustering with Neural Net Clustering App 3:48
Cluster iris flowers based on petal and sepal size.
Wine Classification with Neural Net Pattern Recognition App 3:36
Identify the winery that particular wines came from based on chemical attributes of the wine.
Network Architectures
Neural Network Toolbox supports a variety of supervised and unsupervised network architectures. With the
toolbox’s modular approach to building networks, you can develop custom network architectures for your specific
problem. You can view the network architecture including all inputs, layers, outputs, and interconnections.
Supervised Networks
Supervised neural networks are trained to produce desired outputs in response to sample inputs, making them
particularly well-suited to modeling and controlling dynamic systems, classifying noisy data, and predicting
future events.
Neural Network Toolbox includes four types of supervised networks: feedforward, radial basis, dynamic, and
learning vector quantization.
Feedforward networks have one-way connections from input to output layers. They are most commonly used for
prediction, pattern recognition, and nonlinear function fitting. Supported feedforward networks include
feedforward backpropagation, cascade-forward backpropagation, feedforward input-delay backpropagation,
linear, and perceptron networks.
A two-layer feedforward network with sigmoid hidden neurons and linear output neurons. This type of network can fit
multidimensional mapping problems arbitrarily well, given consistent data and enough neurons in its hidden layer.
Radial basis networks provide an alternative, fast method for designing nonlinear feedforward networks.
Supported variations include generalized regression and probabilistic neural networks.
2
Maglev Modeling with Neural Time Series App 5:28
Model the position of a levitated magnet as current passes through an electromagnet beneath
it.
Dynamic networks use memory and recurrent feedback connections to recognize spatial and temporal patterns in
data. They are commonly used for time-series prediction, nonlinear dynamic system modeling, and control
systems applications. Prebuilt dynamic networks in the toolbox include focused and distributed time-delay,
nonlinear autoregressive (NARX), layer-recurrent, Elman, and Hopfield networks. The toolbox also supports
dynamic training of custom networks with arbitrary connections.
Learning vector quantization (LVQ) networks use a method for classifying patterns that are not linearly
separable. LVQ lets you specify class boundaries and the granularity of classification.
Unsupervised Networks
Unsupervised neural networks are trained by letting the network continually adjust itself to new inputs. They find
relationships within data and can automatically define classification schemes.
Neural Network Toolbox includes two types of self-organizing, unsupervised networks: competitive layers and
self-organizing maps.
Competitive layers recognize and group similar input vectors, enabling them to automatically sort inputs into
categories. Competitive layers are commonly used for classification and pattern recognition.
Self-organizing maps learn to classify input vectors according to similarity. Like competitive layers, they are
used for classification and pattern recognition tasks; however, they differ from competitive layers because they are
able to preserve the topology of the input vectors, assigning nearby inputs to nearby categories.
A self-organizing map consisting of a competitive layer that can classify a data set of vectors with any number of dimensions
into as many classes as the layer has neurons.
Training Algorithms
Training and learning functions are mathematical procedures used to automatically adjust the network’s weights
and biases. The training function dictates a global algorithm that affects all the weights and biases of a given
network. The learning function can be applied to individual weights and biases within a network.
Neural Network Toolbox supports a variety of training algorithms, including several gradient descent methods,
conjugate gradient methods, the Levenberg-Marquardt algorithm (LM), and the resilient backpropagation
algorithm (Rprop). The toolbox’s modular framework lets you quickly develop custom training algorithms that
can be integrated with built-in algorithms. While training your neural network, you can use error weights to define
the relative importance of desired outputs, which can be prioritized in terms of sample, time step (for time-series
problems), output element, or any combination of these. You can access training algorithms from the command
3
line or via apps that show diagrams of the network being trained and provide network performance plots and
status information to help you monitor the training process.
A suite of learning functions, including gradient descent, Hebbian learning, LVQ, Widrow-Hoff, and Kohonen, is
also provided.
Neural network apps that automate training your neural network to fit input and target data (left), monitor training progress
(right), and calculate statistical results and plots to assess training quality.
Preprocessing and Postprocessing
Preprocessing the network inputs and targets improves the efficiency of neural network training. Postprocessing
enables detailed analysis of network performance. Neural Network Toolbox provides preprocessing and
postprocessing functions and Simulink blocks that enable you to:
▪ Reduce the dimensions of the input vectors using principal component analysis
▪ Perform regression analysis between the network response and the corresponding targets
▪ Scale inputs and targets so they fall in the range [-1,1]
▪ Normalize the mean and standard deviation of the training set
▪ Use automated data preprocessing and data division when creating your networks
4
Postprocessing plots to analyze network performance, including mean squared error validation performance for successive
training epochs (top left), error histogram (top right), and confusion matrices (bottom) for training, validation, and test phases.
Improved Generalization
Improving the network’s ability to generalize helps prevent overfitting, a common problem in neural network
design. Overfitting occurs when a network has memorized the training set but has not learned to generalize to new
inputs. Overfitting produces a relatively small error on the training set but a much larger error when new data is
presented to the network.
Neural Network Toolbox provides two solutions to improve generalization:
▪ Regularization modifies the network’s performance function (the measure of error that the training process
minimizes). By including the sizes of the weights and biases, regularization produces a network that performs
well with the training data and exhibits smoother behavior when presented with new data.
▪ Early stopping uses two different data sets: the training set, to update the weights and biases, and the
validation set, to stop training when the network begins to overfit the data.
Simulink Blocks and Control Systems Applications
Simulink Support
Neural Network Toolbox provides a set of blocks for building neural networks in Simulink. All blocks are
compatible with Simulink Coder™. These blocks are divided into four libraries:
▪ Transfer function blocks, which take a net input vector and generate a corresponding output vector
5
▪ Net input function blocks, which take any number of weighted input vectors, weight-layer output vectors,
and bias vectors, and return a net input vector
▪ Weight function blocks, which apply a neuron’s weight vector to an input vector (or a layer output vector) to
get a weighted input value for a neuron
▪ Data preprocessing blocks, which map input and output data into the ranges best suited for the neural
network to handle directly
Alternatively, you can create and train your networks in the MATLAB environment and automatically generate
network simulation blocks for use with Simulink. This approach also enables you to view your networks
graphically.
Control Systems Applications
You can apply neural networks to the identification and control of nonlinear systems. The toolbox includes
descriptions, examples, and Simulink blocks for three popular control applications:
▪ Model predictive control, which uses a neural network model to predict future plant responses to potential
control signals. An optimization algorithm then computes the control signals that optimize future plant
performance. The neural network plant model is trained offline and in batch form.
▪ Feedback linearization, which uses a rearrangement of the neural network plant model and is trained offline.
This controller requires the least computation of these three architectures; however, the plant must either be in
companion form or be capable of approximation by a companion form model.
▪ Model reference adaptive control, which requires that a separate neural network controller be trained
offline, in addition to the neural network plant model. While the controller training is computationally
expensive, the model reference control applies to a larger class of plant than feedback linearization.
You can incorporate neural network predictive control blocks included in the toolbox into your Simulink models.
By changing the parameters of these blocks, you can tailor the network’s performance to your application.
6
A Simulink model that uses the Neural Network Predictive Controller block with a tank reactor plant model (top). You can
visualize the dynamic response (bottom left) and manage the controller block (bottom center) and your plant identification
(bottom right).
Accelerated Training and Large Data Sets
You can speed up neural network training and simulation of large data sets by using Neural Network Toolbox with
Parallel Computing Toolbox. Training and simulation involve many parallel computations, which can be
accelerated with multicore processors, CUDA-enabled NVIDIA GPUs, and computer clusters with multiple
processors and GPUs.
Distributed Computing
Parallel Computing Toolbox lets neural network training and simulation run across multiple processor cores on a
single PC, or across multiple processors on multiple computers on a network using MATLAB Distributed
Computing Server™. Using multiple cores can speed up calculations. Using multiple computers lets you solve
problems using data sets too big to fit within the system memory of any single computer. The only limit to
problem size is the total system memory available across all computers.
GPU Computing
Parallel Computing Toolbox enables Neural Network Toolbox simulation and training to be parallelized across the
multiprocessors and cores of a general-purpose graphics processing unit (GPU). GPUs are highly efficient on
parallel algorithms such as neural networks. You can achieve higher levels of parallelism by using multiple GPUs
or by using GPUs and processors together. With MATLAB Distributed Computing Server, you can harness all the
processors and GPUs on a network cluster of computers for neural network training and simulation.
Learn more about GPU computing with MATLAB.
7
Product Details, Examples, and System Requirements
www.mathworks.com/products/neural-network
Trial Software
www.mathworks.com/trialrequest
Sales
www.mathworks.com/contactsales
Technical Support
www.mathworks.com/support
Resources
Online User Community
www.mathworks.com/matlabcentral
Training Services
www.mathworks.com/training
Third-Party Products and Services
www.mathworks.com/connections
Worldwide Contacts
www.mathworks.com/contact
© 2012 The MathWorks, Inc. MATLAB and Simulink are registered trademarks of The MathWorks, Inc. See www.mathworks.com/trademarks for a list of
additional trademarks. Other product or brand names may be trademarks or registered trademarks of their respective holders. 8

Mais conteúdo relacionado

Mais procurados

artificial neural network
artificial neural networkartificial neural network
artificial neural networkPallavi Yadav
 
Activation functions and Training Algorithms for Deep Neural network
Activation functions and Training Algorithms for Deep Neural networkActivation functions and Training Algorithms for Deep Neural network
Activation functions and Training Algorithms for Deep Neural networkGayatri Khanvilkar
 
Machine Learning (ML) in Wireless Sensor Networks (WSNs)
Machine Learning (ML) in Wireless Sensor Networks (WSNs)Machine Learning (ML) in Wireless Sensor Networks (WSNs)
Machine Learning (ML) in Wireless Sensor Networks (WSNs)mabualsh
 
Artificial neural network
Artificial neural networkArtificial neural network
Artificial neural networknainabhatt2
 
Artificial Neural Network(Artificial intelligence)
Artificial Neural Network(Artificial intelligence)Artificial Neural Network(Artificial intelligence)
Artificial Neural Network(Artificial intelligence)spartacus131211
 
Artificial neural networks in hydrology
Artificial neural networks in hydrology Artificial neural networks in hydrology
Artificial neural networks in hydrology Jonathan D'Cruz
 
Advance deep learning
Advance deep learningAdvance deep learning
Advance deep learningaliaKhan71
 
Introduction Of Artificial neural network
Introduction Of Artificial neural networkIntroduction Of Artificial neural network
Introduction Of Artificial neural networkNagarajan
 
Neural network image recognition
Neural network image recognitionNeural network image recognition
Neural network image recognitionOleksii Sekundant
 
UNET: Massive Scale DNN on Spark
UNET: Massive Scale DNN on SparkUNET: Massive Scale DNN on Spark
UNET: Massive Scale DNN on SparkZhan Zhang
 
A Survey of Convolutional Neural Networks
A Survey of Convolutional Neural NetworksA Survey of Convolutional Neural Networks
A Survey of Convolutional Neural NetworksRimzim Thube
 
Adaptive Resonance Theory
Adaptive Resonance TheoryAdaptive Resonance Theory
Adaptive Resonance Theorysurat murthy
 
Neural Networks Ver1
Neural  Networks  Ver1Neural  Networks  Ver1
Neural Networks Ver1ncct
 
Neural Network Research Projects Topics
Neural Network Research Projects TopicsNeural Network Research Projects Topics
Neural Network Research Projects TopicsMatlab Simulation
 
Cloud data management
Cloud data managementCloud data management
Cloud data managementambitlick
 

Mais procurados (20)

artificial neural network
artificial neural networkartificial neural network
artificial neural network
 
Artificial Neural Network.pptx
Artificial Neural Network.pptxArtificial Neural Network.pptx
Artificial Neural Network.pptx
 
Activation functions and Training Algorithms for Deep Neural network
Activation functions and Training Algorithms for Deep Neural networkActivation functions and Training Algorithms for Deep Neural network
Activation functions and Training Algorithms for Deep Neural network
 
Machine Learning (ML) in Wireless Sensor Networks (WSNs)
Machine Learning (ML) in Wireless Sensor Networks (WSNs)Machine Learning (ML) in Wireless Sensor Networks (WSNs)
Machine Learning (ML) in Wireless Sensor Networks (WSNs)
 
Artificial neural network
Artificial neural networkArtificial neural network
Artificial neural network
 
Artificial Neural Network(Artificial intelligence)
Artificial Neural Network(Artificial intelligence)Artificial Neural Network(Artificial intelligence)
Artificial Neural Network(Artificial intelligence)
 
Neural networks
Neural networksNeural networks
Neural networks
 
Artificial neural networks in hydrology
Artificial neural networks in hydrology Artificial neural networks in hydrology
Artificial neural networks in hydrology
 
Neural network
Neural networkNeural network
Neural network
 
Multi tasking learning
Multi tasking learningMulti tasking learning
Multi tasking learning
 
Advance deep learning
Advance deep learningAdvance deep learning
Advance deep learning
 
Introduction Of Artificial neural network
Introduction Of Artificial neural networkIntroduction Of Artificial neural network
Introduction Of Artificial neural network
 
Neural network image recognition
Neural network image recognitionNeural network image recognition
Neural network image recognition
 
UNET: Massive Scale DNN on Spark
UNET: Massive Scale DNN on SparkUNET: Massive Scale DNN on Spark
UNET: Massive Scale DNN on Spark
 
A Survey of Convolutional Neural Networks
A Survey of Convolutional Neural NetworksA Survey of Convolutional Neural Networks
A Survey of Convolutional Neural Networks
 
Adaptive Resonance Theory
Adaptive Resonance TheoryAdaptive Resonance Theory
Adaptive Resonance Theory
 
Neural Networks Ver1
Neural  Networks  Ver1Neural  Networks  Ver1
Neural Networks Ver1
 
Neural Network Research Projects Topics
Neural Network Research Projects TopicsNeural Network Research Projects Topics
Neural Network Research Projects Topics
 
Ann
Ann Ann
Ann
 
Cloud data management
Cloud data managementCloud data management
Cloud data management
 

Destaque

Intro matlab-nn
Intro matlab-nnIntro matlab-nn
Intro matlab-nnsoldier78
 
learning Vector Quantization LVQ2 LVQ3
learning Vector Quantization LVQ2 LVQ3learning Vector Quantization LVQ2 LVQ3
learning Vector Quantization LVQ2 LVQ3ESCOM
 
Matlab Neural Network Toolbox
Matlab Neural Network ToolboxMatlab Neural Network Toolbox
Matlab Neural Network ToolboxAliMETN
 
Kohonen SOM dan Learning Vector Quantization (LVQ)
Kohonen SOM dan Learning Vector Quantization (LVQ)Kohonen SOM dan Learning Vector Quantization (LVQ)
Kohonen SOM dan Learning Vector Quantization (LVQ)petrus fendiyanto
 
Neural Network Toolbox MATLAB
Neural Network Toolbox MATLABNeural Network Toolbox MATLAB
Neural Network Toolbox MATLABESCOM
 
Matlab Neural Network Toolbox MATLAB
Matlab Neural Network Toolbox MATLABMatlab Neural Network Toolbox MATLAB
Matlab Neural Network Toolbox MATLABESCOM
 
Learning Vector Quantization LVQ
Learning Vector Quantization LVQLearning Vector Quantization LVQ
Learning Vector Quantization LVQESCOM
 
Learning Vector Quantization LVQ
Learning Vector Quantization LVQLearning Vector Quantization LVQ
Learning Vector Quantization LVQESCOM
 
Neural networks Self Organizing Map by Engr. Edgar Carrillo II
Neural networks Self Organizing Map by Engr. Edgar Carrillo IINeural networks Self Organizing Map by Engr. Edgar Carrillo II
Neural networks Self Organizing Map by Engr. Edgar Carrillo IIEdgar Carrillo
 
Neural network in matlab
Neural network in matlab Neural network in matlab
Neural network in matlab Fahim Khan
 
Self-Organising Maps for Customer Segmentation using R - Shane Lynn - Dublin R
Self-Organising Maps for Customer Segmentation using R - Shane Lynn - Dublin RSelf-Organising Maps for Customer Segmentation using R - Shane Lynn - Dublin R
Self-Organising Maps for Customer Segmentation using R - Shane Lynn - Dublin Rshanelynn
 
Self-organizing maps - Tutorial
Self-organizing maps -  TutorialSelf-organizing maps -  Tutorial
Self-organizing maps - Tutorialaskroll
 
Artificial neural networks
Artificial neural networksArtificial neural networks
Artificial neural networksstellajoseph
 

Destaque (14)

Intro matlab-nn
Intro matlab-nnIntro matlab-nn
Intro matlab-nn
 
learning Vector Quantization LVQ2 LVQ3
learning Vector Quantization LVQ2 LVQ3learning Vector Quantization LVQ2 LVQ3
learning Vector Quantization LVQ2 LVQ3
 
Matlab Neural Network Toolbox
Matlab Neural Network ToolboxMatlab Neural Network Toolbox
Matlab Neural Network Toolbox
 
Kohonen SOM dan Learning Vector Quantization (LVQ)
Kohonen SOM dan Learning Vector Quantization (LVQ)Kohonen SOM dan Learning Vector Quantization (LVQ)
Kohonen SOM dan Learning Vector Quantization (LVQ)
 
Neural Network Toolbox MATLAB
Neural Network Toolbox MATLABNeural Network Toolbox MATLAB
Neural Network Toolbox MATLAB
 
Matlab Neural Network Toolbox MATLAB
Matlab Neural Network Toolbox MATLABMatlab Neural Network Toolbox MATLAB
Matlab Neural Network Toolbox MATLAB
 
Learning Vector Quantization LVQ
Learning Vector Quantization LVQLearning Vector Quantization LVQ
Learning Vector Quantization LVQ
 
Learning Vector Quantization LVQ
Learning Vector Quantization LVQLearning Vector Quantization LVQ
Learning Vector Quantization LVQ
 
Sefl Organizing Map
Sefl Organizing MapSefl Organizing Map
Sefl Organizing Map
 
Neural networks Self Organizing Map by Engr. Edgar Carrillo II
Neural networks Self Organizing Map by Engr. Edgar Carrillo IINeural networks Self Organizing Map by Engr. Edgar Carrillo II
Neural networks Self Organizing Map by Engr. Edgar Carrillo II
 
Neural network in matlab
Neural network in matlab Neural network in matlab
Neural network in matlab
 
Self-Organising Maps for Customer Segmentation using R - Shane Lynn - Dublin R
Self-Organising Maps for Customer Segmentation using R - Shane Lynn - Dublin RSelf-Organising Maps for Customer Segmentation using R - Shane Lynn - Dublin R
Self-Organising Maps for Customer Segmentation using R - Shane Lynn - Dublin R
 
Self-organizing maps - Tutorial
Self-organizing maps -  TutorialSelf-organizing maps -  Tutorial
Self-organizing maps - Tutorial
 
Artificial neural networks
Artificial neural networksArtificial neural networks
Artificial neural networks
 

Semelhante a Neural network-toolbox

Handwritten Digit Recognition using Convolutional Neural Networks
Handwritten Digit Recognition using Convolutional Neural  NetworksHandwritten Digit Recognition using Convolutional Neural  Networks
Handwritten Digit Recognition using Convolutional Neural NetworksIRJET Journal
 
Neural network based numerical digits recognization using nnt in matlab
Neural network based numerical digits recognization using nnt in matlabNeural network based numerical digits recognization using nnt in matlab
Neural network based numerical digits recognization using nnt in matlabijcses
 
Artificial Neural Network ANN
Artificial Neural Network ANNArtificial Neural Network ANN
Artificial Neural Network ANNAbdullah al Mamun
 
# Neural network toolbox
# Neural network toolbox # Neural network toolbox
# Neural network toolbox VineetKumar508
 
friction factor modelling.pptx
friction factor modelling.pptxfriction factor modelling.pptx
friction factor modelling.pptxOKORIE1
 
Artificial Neural Network Seminar Report
Artificial Neural Network Seminar ReportArtificial Neural Network Seminar Report
Artificial Neural Network Seminar ReportTodd Turner
 
IRJET-AI Neural Network Disaster Recovery Cloud Operations Systems
IRJET-AI Neural Network Disaster Recovery Cloud Operations SystemsIRJET-AI Neural Network Disaster Recovery Cloud Operations Systems
IRJET-AI Neural Network Disaster Recovery Cloud Operations SystemsIRJET Journal
 
Neural Network Based Individual Classification System
Neural Network Based Individual Classification SystemNeural Network Based Individual Classification System
Neural Network Based Individual Classification SystemIRJET Journal
 
Table of Contents
Table of ContentsTable of Contents
Table of Contentsbutest
 
The Use of Wireless networks for Control Applications
The Use of Wireless networks for Control ApplicationsThe Use of Wireless networks for Control Applications
The Use of Wireless networks for Control ApplicationsNativ Levy
 
report
reportreport
reportbutest
 
House price prediction
House price predictionHouse price prediction
House price predictionSabahBegum
 
Lecture 1 - Introduction.pptx
Lecture 1 - Introduction.pptxLecture 1 - Introduction.pptx
Lecture 1 - Introduction.pptxaida alsamawi
 
Artificial neural network for machine learning
Artificial neural network for machine learningArtificial neural network for machine learning
Artificial neural network for machine learninggrinu
 

Semelhante a Neural network-toolbox (20)

N ns 1
N ns 1N ns 1
N ns 1
 
Handwritten Digit Recognition using Convolutional Neural Networks
Handwritten Digit Recognition using Convolutional Neural  NetworksHandwritten Digit Recognition using Convolutional Neural  Networks
Handwritten Digit Recognition using Convolutional Neural Networks
 
Neural network based numerical digits recognization using nnt in matlab
Neural network based numerical digits recognization using nnt in matlabNeural network based numerical digits recognization using nnt in matlab
Neural network based numerical digits recognization using nnt in matlab
 
Artificial Neural Network ANN
Artificial Neural Network ANNArtificial Neural Network ANN
Artificial Neural Network ANN
 
# Neural network toolbox
# Neural network toolbox # Neural network toolbox
# Neural network toolbox
 
friction factor modelling.pptx
friction factor modelling.pptxfriction factor modelling.pptx
friction factor modelling.pptx
 
Artificial Neural Network Seminar Report
Artificial Neural Network Seminar ReportArtificial Neural Network Seminar Report
Artificial Neural Network Seminar Report
 
IRJET-AI Neural Network Disaster Recovery Cloud Operations Systems
IRJET-AI Neural Network Disaster Recovery Cloud Operations SystemsIRJET-AI Neural Network Disaster Recovery Cloud Operations Systems
IRJET-AI Neural Network Disaster Recovery Cloud Operations Systems
 
Neural Network Based Individual Classification System
Neural Network Based Individual Classification SystemNeural Network Based Individual Classification System
Neural Network Based Individual Classification System
 
Project Report -Vaibhav
Project Report -VaibhavProject Report -Vaibhav
Project Report -Vaibhav
 
Table of Contents
Table of ContentsTable of Contents
Table of Contents
 
Deep Learning
Deep LearningDeep Learning
Deep Learning
 
The Use of Wireless networks for Control Applications
The Use of Wireless networks for Control ApplicationsThe Use of Wireless networks for Control Applications
The Use of Wireless networks for Control Applications
 
++Matlab 14 sesiones
++Matlab 14 sesiones++Matlab 14 sesiones
++Matlab 14 sesiones
 
report
reportreport
report
 
House price prediction
House price predictionHouse price prediction
House price prediction
 
Lecture 1 - Introduction.pptx
Lecture 1 - Introduction.pptxLecture 1 - Introduction.pptx
Lecture 1 - Introduction.pptx
 
Artificial neural network for machine learning
Artificial neural network for machine learningArtificial neural network for machine learning
Artificial neural network for machine learning
 
Stock Market Prediction Using ANN
Stock Market Prediction Using ANNStock Market Prediction Using ANN
Stock Market Prediction Using ANN
 
Cc module 3.pptx
Cc module 3.pptxCc module 3.pptx
Cc module 3.pptx
 

Último

SIP trunking in Janus @ Kamailio World 2024
SIP trunking in Janus @ Kamailio World 2024SIP trunking in Janus @ Kamailio World 2024
SIP trunking in Janus @ Kamailio World 2024Lorenzo Miniero
 
Developer Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQLDeveloper Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQLScyllaDB
 
H2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo Day
H2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo DayH2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo Day
H2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo DaySri Ambati
 
Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024Scott Keck-Warren
 
Commit 2024 - Secret Management made easy
Commit 2024 - Secret Management made easyCommit 2024 - Secret Management made easy
Commit 2024 - Secret Management made easyAlfredo García Lavilla
 
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)Mark Simos
 
What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024Stephanie Beckett
 
Nell’iperspazio con Rocket: il Framework Web di Rust!
Nell’iperspazio con Rocket: il Framework Web di Rust!Nell’iperspazio con Rocket: il Framework Web di Rust!
Nell’iperspazio con Rocket: il Framework Web di Rust!Commit University
 
Artificial intelligence in cctv survelliance.pptx
Artificial intelligence in cctv survelliance.pptxArtificial intelligence in cctv survelliance.pptx
Artificial intelligence in cctv survelliance.pptxhariprasad279825
 
Anypoint Exchange: It’s Not Just a Repo!
Anypoint Exchange: It’s Not Just a Repo!Anypoint Exchange: It’s Not Just a Repo!
Anypoint Exchange: It’s Not Just a Repo!Manik S Magar
 
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek SchlawackFwdays
 
WordPress Websites for Engineers: Elevate Your Brand
WordPress Websites for Engineers: Elevate Your BrandWordPress Websites for Engineers: Elevate Your Brand
WordPress Websites for Engineers: Elevate Your Brandgvaughan
 
From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .Alan Dix
 
Streamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project SetupStreamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project SetupFlorian Wilhelm
 
Scanning the Internet for External Cloud Exposures via SSL Certs
Scanning the Internet for External Cloud Exposures via SSL CertsScanning the Internet for External Cloud Exposures via SSL Certs
Scanning the Internet for External Cloud Exposures via SSL CertsRizwan Syed
 
Connect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck PresentationConnect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck PresentationSlibray Presentation
 
Unleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubUnleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubKalema Edgar
 
CloudStudio User manual (basic edition):
CloudStudio User manual (basic edition):CloudStudio User manual (basic edition):
CloudStudio User manual (basic edition):comworks
 
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024BookNet Canada
 

Último (20)

SIP trunking in Janus @ Kamailio World 2024
SIP trunking in Janus @ Kamailio World 2024SIP trunking in Janus @ Kamailio World 2024
SIP trunking in Janus @ Kamailio World 2024
 
Developer Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQLDeveloper Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQL
 
H2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo Day
H2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo DayH2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo Day
H2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo Day
 
Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024
 
Commit 2024 - Secret Management made easy
Commit 2024 - Secret Management made easyCommit 2024 - Secret Management made easy
Commit 2024 - Secret Management made easy
 
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
 
What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024
 
Nell’iperspazio con Rocket: il Framework Web di Rust!
Nell’iperspazio con Rocket: il Framework Web di Rust!Nell’iperspazio con Rocket: il Framework Web di Rust!
Nell’iperspazio con Rocket: il Framework Web di Rust!
 
Artificial intelligence in cctv survelliance.pptx
Artificial intelligence in cctv survelliance.pptxArtificial intelligence in cctv survelliance.pptx
Artificial intelligence in cctv survelliance.pptx
 
Anypoint Exchange: It’s Not Just a Repo!
Anypoint Exchange: It’s Not Just a Repo!Anypoint Exchange: It’s Not Just a Repo!
Anypoint Exchange: It’s Not Just a Repo!
 
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
 
WordPress Websites for Engineers: Elevate Your Brand
WordPress Websites for Engineers: Elevate Your BrandWordPress Websites for Engineers: Elevate Your Brand
WordPress Websites for Engineers: Elevate Your Brand
 
From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .
 
Streamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project SetupStreamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project Setup
 
E-Vehicle_Hacking_by_Parul Sharma_null_owasp.pptx
E-Vehicle_Hacking_by_Parul Sharma_null_owasp.pptxE-Vehicle_Hacking_by_Parul Sharma_null_owasp.pptx
E-Vehicle_Hacking_by_Parul Sharma_null_owasp.pptx
 
Scanning the Internet for External Cloud Exposures via SSL Certs
Scanning the Internet for External Cloud Exposures via SSL CertsScanning the Internet for External Cloud Exposures via SSL Certs
Scanning the Internet for External Cloud Exposures via SSL Certs
 
Connect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck PresentationConnect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck Presentation
 
Unleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubUnleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding Club
 
CloudStudio User manual (basic edition):
CloudStudio User manual (basic edition):CloudStudio User manual (basic edition):
CloudStudio User manual (basic edition):
 
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
 

Neural network-toolbox

  • 1. Getting Started with Neural Network Toolbox 4:20 Use graphical tools to apply neural networks to data fitting, pattern recognition, clustering, and time series problems. Neural Network Toolbox Create, train, and simulate neural networks Neural Network Toolbox™ provides functions and apps for modeling complex nonlinear systems that are not easily modeled with a closed-form equation. Neural Network Toolbox supports supervised learning with feedforward, radial basis, and dynamic networks. It also supports unsupervised learning with self-organizing maps and competitive layers. With the toolbox you can design, train, visualize, and simulate neural networks. You can use Neural Network Toolbox for applications such as data fitting, pattern recognition, clustering, time-series prediction, and dynamic system modeling and control. To speed up training and handle large data sets, you can distribute computations and data across multicore processors, GPUs, and computer clusters using Parallel Computing Toolbox™. Key Features ▪ Supervised networks, including multilayer, radial basis, learning vector quantization (LVQ), time-delay, nonlinear autoregressive (NARX), and layer-recurrent ▪ Unsupervised networks, including self-organizing maps and competitive layers ▪ Apps for data-fitting, pattern recognition, and clustering ▪ Parallel computing and GPU support for accelerating training (using Parallel Computing Toolbox) ▪ Preprocessing and postprocessing for improving the efficiency of network training and assessing network performance ▪ Modular network representation for managing and visualizing networks of arbitrary size ▪ Simulink® blocks for building and evaluating neural networks and for control systems applications Data Fitting, Clustering, and Pattern Recognition Like its counterpart in the biological nervous system, a neural network can learn and therefore can be trained to find solutions, recognize patterns, classify data, and forecast future events. The behavior of a neural network is defined by the way its individual computing elements are connected and by the strengths of those connections, or weights. The weights are automatically adjusted by training the network according to a specified learning rule until it performs the desired task correctly. Neural Network Toolbox includes command-line functions and apps for creating, training, and simulating neural networks. The apps make it easy to develop neural networks for tasks such as data-fitting (including time-series data), pattern recognition, and clustering. After creating your networks in these tools, you can automatically generate MATLAB® code to capture your work and automate tasks. 1
  • 2. House Pricing Estimation with Neural Net Fitting App 4:21 Estimate median house prices for neighborhoods based on various neighborhood attributes. Iris Flower Clustering with Neural Net Clustering App 3:48 Cluster iris flowers based on petal and sepal size. Wine Classification with Neural Net Pattern Recognition App 3:36 Identify the winery that particular wines came from based on chemical attributes of the wine. Network Architectures Neural Network Toolbox supports a variety of supervised and unsupervised network architectures. With the toolbox’s modular approach to building networks, you can develop custom network architectures for your specific problem. You can view the network architecture including all inputs, layers, outputs, and interconnections. Supervised Networks Supervised neural networks are trained to produce desired outputs in response to sample inputs, making them particularly well-suited to modeling and controlling dynamic systems, classifying noisy data, and predicting future events. Neural Network Toolbox includes four types of supervised networks: feedforward, radial basis, dynamic, and learning vector quantization. Feedforward networks have one-way connections from input to output layers. They are most commonly used for prediction, pattern recognition, and nonlinear function fitting. Supported feedforward networks include feedforward backpropagation, cascade-forward backpropagation, feedforward input-delay backpropagation, linear, and perceptron networks. A two-layer feedforward network with sigmoid hidden neurons and linear output neurons. This type of network can fit multidimensional mapping problems arbitrarily well, given consistent data and enough neurons in its hidden layer. Radial basis networks provide an alternative, fast method for designing nonlinear feedforward networks. Supported variations include generalized regression and probabilistic neural networks. 2
  • 3. Maglev Modeling with Neural Time Series App 5:28 Model the position of a levitated magnet as current passes through an electromagnet beneath it. Dynamic networks use memory and recurrent feedback connections to recognize spatial and temporal patterns in data. They are commonly used for time-series prediction, nonlinear dynamic system modeling, and control systems applications. Prebuilt dynamic networks in the toolbox include focused and distributed time-delay, nonlinear autoregressive (NARX), layer-recurrent, Elman, and Hopfield networks. The toolbox also supports dynamic training of custom networks with arbitrary connections. Learning vector quantization (LVQ) networks use a method for classifying patterns that are not linearly separable. LVQ lets you specify class boundaries and the granularity of classification. Unsupervised Networks Unsupervised neural networks are trained by letting the network continually adjust itself to new inputs. They find relationships within data and can automatically define classification schemes. Neural Network Toolbox includes two types of self-organizing, unsupervised networks: competitive layers and self-organizing maps. Competitive layers recognize and group similar input vectors, enabling them to automatically sort inputs into categories. Competitive layers are commonly used for classification and pattern recognition. Self-organizing maps learn to classify input vectors according to similarity. Like competitive layers, they are used for classification and pattern recognition tasks; however, they differ from competitive layers because they are able to preserve the topology of the input vectors, assigning nearby inputs to nearby categories. A self-organizing map consisting of a competitive layer that can classify a data set of vectors with any number of dimensions into as many classes as the layer has neurons. Training Algorithms Training and learning functions are mathematical procedures used to automatically adjust the network’s weights and biases. The training function dictates a global algorithm that affects all the weights and biases of a given network. The learning function can be applied to individual weights and biases within a network. Neural Network Toolbox supports a variety of training algorithms, including several gradient descent methods, conjugate gradient methods, the Levenberg-Marquardt algorithm (LM), and the resilient backpropagation algorithm (Rprop). The toolbox’s modular framework lets you quickly develop custom training algorithms that can be integrated with built-in algorithms. While training your neural network, you can use error weights to define the relative importance of desired outputs, which can be prioritized in terms of sample, time step (for time-series problems), output element, or any combination of these. You can access training algorithms from the command 3
  • 4. line or via apps that show diagrams of the network being trained and provide network performance plots and status information to help you monitor the training process. A suite of learning functions, including gradient descent, Hebbian learning, LVQ, Widrow-Hoff, and Kohonen, is also provided. Neural network apps that automate training your neural network to fit input and target data (left), monitor training progress (right), and calculate statistical results and plots to assess training quality. Preprocessing and Postprocessing Preprocessing the network inputs and targets improves the efficiency of neural network training. Postprocessing enables detailed analysis of network performance. Neural Network Toolbox provides preprocessing and postprocessing functions and Simulink blocks that enable you to: ▪ Reduce the dimensions of the input vectors using principal component analysis ▪ Perform regression analysis between the network response and the corresponding targets ▪ Scale inputs and targets so they fall in the range [-1,1] ▪ Normalize the mean and standard deviation of the training set ▪ Use automated data preprocessing and data division when creating your networks 4
  • 5. Postprocessing plots to analyze network performance, including mean squared error validation performance for successive training epochs (top left), error histogram (top right), and confusion matrices (bottom) for training, validation, and test phases. Improved Generalization Improving the network’s ability to generalize helps prevent overfitting, a common problem in neural network design. Overfitting occurs when a network has memorized the training set but has not learned to generalize to new inputs. Overfitting produces a relatively small error on the training set but a much larger error when new data is presented to the network. Neural Network Toolbox provides two solutions to improve generalization: ▪ Regularization modifies the network’s performance function (the measure of error that the training process minimizes). By including the sizes of the weights and biases, regularization produces a network that performs well with the training data and exhibits smoother behavior when presented with new data. ▪ Early stopping uses two different data sets: the training set, to update the weights and biases, and the validation set, to stop training when the network begins to overfit the data. Simulink Blocks and Control Systems Applications Simulink Support Neural Network Toolbox provides a set of blocks for building neural networks in Simulink. All blocks are compatible with Simulink Coder™. These blocks are divided into four libraries: ▪ Transfer function blocks, which take a net input vector and generate a corresponding output vector 5
  • 6. ▪ Net input function blocks, which take any number of weighted input vectors, weight-layer output vectors, and bias vectors, and return a net input vector ▪ Weight function blocks, which apply a neuron’s weight vector to an input vector (or a layer output vector) to get a weighted input value for a neuron ▪ Data preprocessing blocks, which map input and output data into the ranges best suited for the neural network to handle directly Alternatively, you can create and train your networks in the MATLAB environment and automatically generate network simulation blocks for use with Simulink. This approach also enables you to view your networks graphically. Control Systems Applications You can apply neural networks to the identification and control of nonlinear systems. The toolbox includes descriptions, examples, and Simulink blocks for three popular control applications: ▪ Model predictive control, which uses a neural network model to predict future plant responses to potential control signals. An optimization algorithm then computes the control signals that optimize future plant performance. The neural network plant model is trained offline and in batch form. ▪ Feedback linearization, which uses a rearrangement of the neural network plant model and is trained offline. This controller requires the least computation of these three architectures; however, the plant must either be in companion form or be capable of approximation by a companion form model. ▪ Model reference adaptive control, which requires that a separate neural network controller be trained offline, in addition to the neural network plant model. While the controller training is computationally expensive, the model reference control applies to a larger class of plant than feedback linearization. You can incorporate neural network predictive control blocks included in the toolbox into your Simulink models. By changing the parameters of these blocks, you can tailor the network’s performance to your application. 6
  • 7. A Simulink model that uses the Neural Network Predictive Controller block with a tank reactor plant model (top). You can visualize the dynamic response (bottom left) and manage the controller block (bottom center) and your plant identification (bottom right). Accelerated Training and Large Data Sets You can speed up neural network training and simulation of large data sets by using Neural Network Toolbox with Parallel Computing Toolbox. Training and simulation involve many parallel computations, which can be accelerated with multicore processors, CUDA-enabled NVIDIA GPUs, and computer clusters with multiple processors and GPUs. Distributed Computing Parallel Computing Toolbox lets neural network training and simulation run across multiple processor cores on a single PC, or across multiple processors on multiple computers on a network using MATLAB Distributed Computing Server™. Using multiple cores can speed up calculations. Using multiple computers lets you solve problems using data sets too big to fit within the system memory of any single computer. The only limit to problem size is the total system memory available across all computers. GPU Computing Parallel Computing Toolbox enables Neural Network Toolbox simulation and training to be parallelized across the multiprocessors and cores of a general-purpose graphics processing unit (GPU). GPUs are highly efficient on parallel algorithms such as neural networks. You can achieve higher levels of parallelism by using multiple GPUs or by using GPUs and processors together. With MATLAB Distributed Computing Server, you can harness all the processors and GPUs on a network cluster of computers for neural network training and simulation. Learn more about GPU computing with MATLAB. 7
  • 8. Product Details, Examples, and System Requirements www.mathworks.com/products/neural-network Trial Software www.mathworks.com/trialrequest Sales www.mathworks.com/contactsales Technical Support www.mathworks.com/support Resources Online User Community www.mathworks.com/matlabcentral Training Services www.mathworks.com/training Third-Party Products and Services www.mathworks.com/connections Worldwide Contacts www.mathworks.com/contact © 2012 The MathWorks, Inc. MATLAB and Simulink are registered trademarks of The MathWorks, Inc. See www.mathworks.com/trademarks for a list of additional trademarks. Other product or brand names may be trademarks or registered trademarks of their respective holders. 8