当前位置:高等教育资讯网  >  中国高校课件下载中心  >  大学文库  >  浏览文档

《机器学习 Machine Learning》课程教学资源(书籍文献)An introduction to neural networks for beginners

资源类别:文库,文档格式:PDF,文档页数:40,文件大小:1.33MB,团购合买
点击下载完整版文档(PDF)

accuracy 0.980 0.975 0.970 0.965 0.960 0.955 0.950 0.945 0.940 0.835 0.930 0.000 1.0002.0003.0004.0005.0006.0007.0008.0009.000 An introduction to neural networks for beginners By Dr Andy Thomas Adyentures in Machine Learning

An introduction to neural networks for beginners By Dr Andy Thomas Adventures in Machine Learning

Table of Contents Introduction....... Who I am and my approach 0…2 The code,pre-requisites and installation. …3 Part 1-Introduction to neural networks 3 1.1 What are artificial neural networks?. 1.2 The structure of an ANN.................. 4 1.2.I The artificial neuron............ 4 1.2.2N0des. …5 1.2.3 The bias.… .6 1.2.4 Putting together the structure .8 1.2.5 The notation........ 9 1.3 The feed-forward pass..... .10 1.3.1A feed-forward example............ 1.3.2 Our first attempt at a feed-forward function. 1.3.3 A more efficient implementation ..13 1.3.4 Vectorisation in neural networks. 13 1.3.5 Matrix multiplication......... 1.4 Gradient descent and optimisation..... .16 1.4.1 A simple example in code...... .18 14.2The cost function.. .19 1.4.3 Gradient descent in neural networks..... 。.。220 1.4.4 A two dimensional gradient descent example .20 1.4.5 Backpropagation in depth............... 21 1.4.6 Propagating into the hidden layers.......... .24 1.4.7 Vectorisation of backpropagation..... .26 1.4.8 Implementing the gradient descent step. …27 1.4.9 The final gradient descent algorithm. ..28 1.5 Implementing the neural network in Python ...29 1.5.1 Scaling data....... .30 PAGE1

PAGE 1 Table of Contents Introduction.................................................................................................................................... 2 Who I am and my approach ..................................................................................................... 2 The code, pre-requisites and installation................................................................................ 3 Part 1 – Introduction to neural networks .................................................................................... 3 1.1 What are artificial neural networks?................................................................................... 3 1.2 The structure of an ANN......................................................................................................4 1.2.1 The artificial neuron.......................................................................................................4 1.2.2 Nodes............................................................................................................................... 5 1.2.3 The bias ...........................................................................................................................6 1.2.4 Putting together the structure .....................................................................................8 1.2.5 The notation...................................................................................................................9 1.3 The feed-forward pass .........................................................................................................10 1.3.1 A feed-forward example.................................................................................................11 1.3.2 Our first attempt at a feed-forward function..............................................................11 1.3.3 A more efficient implementation................................................................................13 1.3.4 Vectorisation in neural networks................................................................................13 1.3.5 Matrix multiplication....................................................................................................14 1.4 Gradient descent and optimisation ...................................................................................16 1.4.1 A simple example in code.............................................................................................18 1.4.2 The cost function..........................................................................................................19 1.4.3 Gradient descent in neural networks ........................................................................ 20 1.4.4 A two dimensional gradient descent example ......................................................... 20 1.4.5 Backpropagation in depth ...........................................................................................21 1.4.6 Propagating into the hidden layers ........................................................................... 24 1.4.7 Vectorisation of backpropagation.............................................................................. 26 1.4.8 Implementing the gradient descent step.................................................................. 27 1.4.9 The final gradient descent algorithm........................................................................ 28 1.5 Implementing the neural network in Python.................................................................. 29 1.5.1 Scaling data ................................................................................................................... 30

1.5.2 Creating test and training datasets............. 1.5.3 Setting up the output layer.... .32 1.5.4 Creating the neural network .................... 32 1.5.5 Assessing the accuracy of the trained model 38 Introduction Welcome to the"An introduction to neural networks for beginners"book.The aim of this much larger book is to get you up to speed with all you need to start on the deep learning journey using TensorFlow.Once you're finished,you may like to check out my follow-up book entitled Coding the Deep Learning Revolution -a step by step introduction using Python,Keras and TensorFlow.What is deep learning,and what is TensorFlow?Deep learning is the field of machine learning that is making many state-of-the-art advancements,from beating players at Go and Poker,to speeding up drug discovery and assisting self-driving cars.If these types of cutting edge applications excite you like they excite me,then you will be interesting in learning as much as you can about deep learning.However,that requires you to know quite a bit about how neural networks work.This will be what this book covers-getting you up to speed on the basic concepts of neural networks and how to create them in Python. WHO I AM AND MY APPROACH I am an engineer who works in the energy utility business who uses machine learning almost daily to excel in my duties.I believe that knowledge of machine learning,and its associated concepts,gives you a significant edge in many different industries,and allows you to approach a multitude of problems in novel and interesting ways.I also maintain an avid interest in machine and deep learning in my spare time,and wish to leverage my previous experience as a university lecturer and academic to educate others in the coming Al and machine learning revolution.My main base for doing this is my website- Adventures in Machine Learning. Some educators in this area tend to focus solely on the code,with neglect of the theory. Others focus more on the theory,with neglect of the code.There are problems with both these types of approaches.The first leads to a stunted understanding of what one is doing -you get quite good at implementing frameworks but when something goes awry or not quite to plan,you have no idea how to fix it.The second often leads to people getting swamped in theory and mathematics and losing interest before implementing anything in code. PAGE 2

PAGE 2 1.5.2 Creating test and training datasets ............................................................................. 31 1.5.3 Setting up the output layer......................................................................................... 32 1.5.4 Creating the neural network ...................................................................................... 32 1.5.5 Assessing the accuracy of the trained model............................................................ 38 Introduction Welcome to the “An introduction to neural networks for beginners” book. The aim of this much larger book is to get you up to speed with all you need to start on the deep learning journey using TensorFlow. Once you’re finished, you may like to check out my follow-up book entitled Coding the Deep Learning Revolution – A step by step introduction using Python, Keras and TensorFlow. What is deep learning, and what is TensorFlow? Deep learning is the field of machine learning that is making many state-of-the-art advancements, from beating players at Go and Poker, to speeding up drug discovery and assisting self-driving cars. If these types of cutting edge applications excite you like they excite me, then you will be interesting in learning as much as you can about deep learning. However, that requires you to know quite a bit about how neural networks work. This will be what this book covers – getting you up to speed on the basic concepts of neural networks and how to create them in Python. WHO I AM AND MY APPROACH I am an engineer who works in the energy / utility business who uses machine learning almost daily to excel in my duties. I believe that knowledge of machine learning, and its associated concepts, gives you a significant edge in many different industries, and allows you to approach a multitude of problems in novel and interesting ways. I also maintain an avid interest in machine and deep learning in my spare time, and wish to leverage my previous experience as a university lecturer and academic to educate others in the coming AI and machine learning revolution. My main base for doing this is my website – Adventures in Machine Learning. Some educators in this area tend to focus solely on the code, with neglect of the theory. Others focus more on the theory, with neglect of the code. There are problems with both these types of approaches. The first leads to a stunted understanding of what one is doing – you get quite good at implementing frameworks but when something goes awry or not quite to plan, you have no idea how to fix it. The second often leads to people getting swamped in theory and mathematics and losing interest before implementing anything in code

My approach is to try to walk a middle path-with some focus on theory but only as much as is necessary before trying it out in code.I also take things slowly,in a step-by-step fashion as much as possible.I get frustrated when educators take multiple steps at once and perform large leaps in logic,which makes things difficult to follow,so I assume my readers are likewise annoyed at such leaps and therefore I try not to assume too much. THE CODE,PRE-REQUISITES AND INSTALLATION This book will feature snippets of code as we go through the explanations,however the full set of code can be found for download at my github repository.This book does require some loose pre-requisites of the reader-these are as follows: A basic understanding of Python variables,arrays,functions,loops and control statements A basic understanding of the numpy library,and multi-dimensional indexing Basic matrix multiplication concepts and differentiation While I list these points as pre-requisites,I expect that you will still be able to follow along reasonably well if you are lacking in some of these areas.I expect you'll be able to pick up these ideas as you go along-I'll provide links and go slowly to ensure that is the case. To install the required software,consult the following links: Python 3.6(this version is required for TensorFlow): https://www.python.org/downloads/ Numpy:https://www.scipy org/install html Sci-kit learn:http://scikit-learn org/stable/install.html It may be easier for you to install Anaconda,which comes with most of these packages ready to go and allows easy installation of virtual environments. Part 1-Introduction to neural networks 1.1 WHAT ARE ARTIFICIAL NEURAL NETWORKS? Artificial neural networks(ANNs)are software implementations of the neuronal structure of our brains.We don't need to talk about the complex biology of our brain structures,but suffice to say,the brain contains neurons which are kind of like organic switches.These can change their output state depending on the strength of their electrical or chemical input.The neural network in a person's brain is a hugely interconnected network of neurons,where the output of any given neuron may be the input to thousands of other neurons.Learning occurs by repeatedly activating certain neural connections over others,and this reinforces those connections.This makes them more likely to produce a desired outcome given a specified input.This PAGE 3

PAGE 3 My approach is to try to walk a middle path – with some focus on theory but only as much as is necessary before trying it out in code. I also take things slowly, in a step-by-step fashion as much as possible. I get frustrated when educators take multiple steps at once and perform large leaps in logic, which makes things difficult to follow, so I assume my readers are likewise annoyed at such leaps and therefore I try not to assume too much. THE CODE, PRE-REQUISITES AND INSTALLATION This book will feature snippets of code as we go through the explanations, however the full set of code can be found for download at my github repository. This book does require some loose pre-requisites of the reader – these are as follows: - A basic understanding of Python variables, arrays, functions, loops and control statements - A basic understanding of the numpy library, and multi-dimensional indexing - Basic matrix multiplication concepts and differentiation While I list these points as pre-requisites, I expect that you will still be able to follow along reasonably well if you are lacking in some of these areas. I expect you’ll be able to pick up these ideas as you go along – I’ll provide links and go slowly to ensure that is the case. To install the required software, consult the following links: - Python 3.6 (this version is required for TensorFlow): https://www.python.org/downloads/ - Numpy: https://www.scipy.org/install.html - Sci-kit learn: http://scikit-learn.org/stable/install.html It may be easier for you to install Anaconda, which comes with most of these packages ready to go and allows easy installation of virtual environments. Part 1 – Introduction to neural networks 1.1 WHAT ARE ARTIFICIAL NEURAL NETWORKS? Artificial neural networks (ANNs) are software implementations of the neuronal structure of our brains. We don’t need to talk about the complex biology of our brain structures, but suffice to say, the brain contains neurons which are kind of like organic switches. These can change their output state depending on the strength of their electrical or chemical input. The neural network in a person’s brain is a hugely interconnected network of neurons, where the output of any given neuron may be the input to thousands of other neurons. Learning occurs by repeatedly activating certain neural connections over others, and this reinforces those connections. This makes them more likely to produce a desired outcome given a specified input. This

learning involves feedback-when the desired outcome occurs,the neural connections causing that outcome becomes strengthened. Artificial neural networks attempt to simplify and mimic this brain behavior.They can be trained in a supervised or unsupervised manner.In a supervised ANN,the network is trained by providing matched input and output data samples,with the intention of getting the ANN to provide a desired output for a given input.An example is an e-mail spam filter-the input training data could be the count of various words in the body of the e- mail,and the output training data would be a classification of whether the e-mail was truly spam or not.If many examples of e-mails are passed through the neural network this allows the network to learn what input data makes it likely that an e-mail is spam or not.This learning takes place be adjusting the weights of the ANN connections,but this will be discussed further in the next section. Unsupervised learning in an ANN is an attempt to get the ANN to"understand"the structure of the provided input data"on its own".This type of ANN will not be discussed in this book 1.2 THE STRUCTURE OF AN ANN 1.2.1 The artificial neuron The biological neuron is simulated in an ANN by an activation function.In classification tasks (e.g.identifying spam e-mails)this activation function must have a"switch on" characteristic-in other words,once the input is greater than a certain value,the output should change state i.e.from o to 1,from-1 to 1 or from o to >o.This simulates the "turning on"of a biological neuron.A common activation function that is used is the sigmoid function: 1 f(2)=1+exp(-2 Which looks like this: import matplotlib.pylab as plt import numpy as np x np.arange(-8,8,0.1) f 1 (1 np.exp(-x)) plt.plot(x,f) plt.xlabel('x') plt.ylabel('f(x)') plt.show() PAGE 4

PAGE 4 learning involves feedback – when the desired outcome occurs, the neural connections causing that outcome becomes strengthened. Artificial neural networks attempt to simplify and mimic this brain behavior. They can be trained in a supervised or unsupervised manner. In a supervised ANN, the network is trained by providing matched input and output data samples, with the intention of getting the ANN to provide a desired output for a given input. An example is an e-mail spam filter – the input training data could be the count of various words in the body of the e￾mail, and the output training data would be a classification of whether the e-mail was truly spam or not. If many examples of e-mails are passed through the neural network this allows the network to learn what input data makes it likely that an e-mail is spam or not. This learning takes place be adjusting the weights of the ANN connections, but this will be discussed further in the next section. Unsupervised learning in an ANN is an attempt to get the ANN to “understand” the structure of the provided input data “on its own”. This type of ANN will not be discussed in this book. 1.2 THE STRUCTURE OF AN ANN 1.2.1 The artificial neuron The biological neuron is simulated in an ANN by an activation function. In classification tasks (e.g. identifying spam e-mails) this activation function must have a “switch on” characteristic – in other words, once the input is greater than a certain value, the output should change state i.e. from 0 to 1, from -1 to 1 or from 0 to >0. This simulates the “turning on” of a biological neuron. A common activation function that is used is the sigmoid function: 𝑓(𝑧) = 1 1 + 𝑒𝑥𝑝(−𝑧) Which looks like this:

1.0 0.8 0.6 0.4 0.2 0.0 8 -6 一4 -20 Figure 1 The sigmoid function As can be seen in the figure above,the function is"activated"i.e.it moves from o to1 when the input x is greater than a certain value.The sigmoid function isn't a step function however,the edge is"soft",and the output doesn't change instantaneously.This means that there is a derivative of the function and this is important for the training algorithm which is discussed more in Section 1.4.5 Backpropagation in depth. 1.2.2 Nodes As mentioned previously,biological neurons are connected hierarchical networks,with the outputs of some neurons being the inputs to others.We can represent these networks as connected layers of nodes.Each node takes multiple weighted inputs,applies the activation function to the summation of these inputs,and in doing so generates an output. I'll break this down further,but to help things along,consider the diagram below: 1 X2 ÷hw,b(X) X3 +1 Figure 2 Node with inputs The circle in the image above represents the node.The node is the "seat"of the activation function,and takes the weighted inputs,sums them,then inputs them to the activation function.The output of the activation function is shown as h in the above diagram.Note:a node as I have shown above is also called a perceptron in some literature. What about this"weight"idea that has been mentioned?The weights are real valued numbers(i.e.not binary is or os),which are multiplied by the inputs and then summed up in the node.So,in other words,the weighted input to the node above would be: PAGE 5

PAGE 5 Figure 1 The sigmoid function As can be seen in the figure above, the function is “activated” i.e. it moves from 0 to 1 when the input x is greater than a certain value. The sigmoid function isn’t a step function however, the edge is “soft”, and the output doesn’t change instantaneously. This means that there is a derivative of the function and this is important for the training algorithm which is discussed more in Section 1.4.5 Backpropagation in depth. 1.2.2 Nodes As mentioned previously, biological neurons are connected hierarchical networks, with the outputs of some neurons being the inputs to others. We can represent these networks as connected layers of nodes. Each node takes multiple weighted inputs, applies the activation function to the summation of these inputs, and in doing so generates an output. I’ll break this down further, but to help things along, consider the diagram below: Figure 2 Node with inputs The circle in the image above represents the node. The node is the “seat” of the activation function, and takes the weighted inputs, sums them, then inputs them to the activation function. The output of the activation function is shown as h in the above diagram. Note: a node as I have shown above is also called a perceptron in some literature. What about this “weight” idea that has been mentioned? The weights are real valued numbers (i.e. not binary 1s or 0s), which are multiplied by the inputs and then summed up in the node. So, in other words, the weighted input to the node above would be:

x1W1+x2W2+x3W3+b Here the wi values are weights(ignore the b for the moment).What are these weights all about?Well,they are the variables that are changed during the learning process,and, along with the input,determine the output of the node.The b is the weight of the +1 bias element-the inclusion of this bias enhances the flexibility of the node,which is best demonstrated in an example. 1.2.3 The bias Let's take an extremely simple node,with only one input and one output: X1 hw(x) Figure 3 Simple node The input to the activation function of the node in this case is simply xw.What does changing wi do in this simple network? w1=8.5 w2=1.8 w3=2.0 11='w=0.5 12='w=1.0 13='w=2.8 forw,1in[(w1,11),(w2,12),(w3,13)]: f=1/(1+np.exp(-x*w)) plt.plot(x,f,label=1) plt.xlabel('x') plt.ylabel('h_w(x)') plt.legend(loc=2) plt.show() PAGE6

PAGE 6 𝑥1𝑤1 + 𝑥2𝑤2 + 𝑥3𝑤3 + 𝑏 Here the 𝑤𝑖 values are weights (ignore the b for the moment). What are these weights all about? Well, they are the variables that are changed during the learning process, and, along with the input, determine the output of the node. The b is the weight of the +1 bias element – the inclusion of this bias enhances the flexibility of the node, which is best demonstrated in an example. 1.2.3 The bias Let’s take an extremely simple node, with only one input and one output: Figure 3 Simple node The input to the activation function of the node in this case is simply 𝑥1𝑤1.What does changing 𝑤1do in this simple network?

1.0 w=0.5 w=1.0 0.8 w=2.0 0.6 0.4 0.2 0.0 -6 -4 -2 0 2 6 Figure 4 Effect of adjusting weights Here we can see that changing the weight changes the slope of the output of the sigmoid activation function,which is obviously useful if we want to model different strengths of relationships between the input and output variables.However,what if we only want the output to change when x is greater than 1?This is where the bias comes in-let's consider the same network with a bias input: X1 W1 hw.b(x) 1 Figure 5 Node with bias PAGE 7

PAGE 7 Figure 4 Effect of adjusting weights Here we can see that changing the weight changes the slope of the output of the sigmoid activation function, which is obviously useful if we want to model different strengths of relationships between the input and output variables. However, what if we only want the output to change when x is greater than 1? This is where the bias comes in – let’s consider the same network with a bias input: Figure 5 Node with bias

w=5.8 b1=-8.0 b2=8.a b3=8.0 11=b=-8.8 12='b=6.8 13=b=8.0' forb,1in[(b1,11),(b2,12),(b3,13)]: f=1/(1+np.exp(-(x*w+b)) plt.plot(x,f,label=1) plt.xlabel('x') plt.ylabel('h_wb(x)') plt.legend(loc=2) plt.show() 1.0 b=-8.0 0.8 b=0.0 b=8.0 0.6 0.2 0.0 8 -6 -4 -20 2 6 8 Figure 6 Effect of bias adjustments In this case,the wi has been increased to simulate a more defined "turn on"function.As you can see,by varying the bias "weight"b,you can change when the node activates.Therefore,by adding a bias term,you can make the node simulate a generic if function,i.e.if(x>z)then 1 else o.Without a bias term,you are unable to vary the z in that if statement,it will be always stuck around o.This is obviously very useful if you are trying to simulate conditional relationships. 1.2.4 Putting together the structure Hopefully the previous explanations have given you a good overview of how a given node/neuron/perceptron in a neural network operates.However,as you are probably aware,there are many such interconnected nodes in a fully fledged neural network.These structures can come in a myriad of different forms,but the most common simple neural PAGE 8

PAGE 8 Figure 6 Effect of bias adjustments In this case, the 𝑤1 has been increased to simulate a more defined “turn on” function. As you can see, by varying the bias “weight” b, you can change when the node activates. Therefore, by adding a bias term, you can make the node simulate a generic if function, i.e. if (x > z) then 1 else 0. Without a bias term, you are unable to vary the z in that if statement, it will be always stuck around 0. This is obviously very useful if you are trying to simulate conditional relationships. 1.2.4 Putting together the structure Hopefully the previous explanations have given you a good overview of how a given node/neuron/perceptron in a neural network operates. However, as you are probably aware, there are many such interconnected nodes in a fully fledged neural network. These structures can come in a myriad of different forms, but the most common simple neural

network structure consists of an input layer,a hidden layer and an output layer.An example of such a structure can be seen below: h,②) X X2 h22 h,3) hw.b(x) ha) Layer 1 Layer 2 Layer3 Figure 7 Three layer neural network The three layers of the network can be seen in the above figure-Layer 1 represents the input layer,where the external input data enters the network.Layer 2 is called the hidden layer as this layer is not part of the input or output.Note:neural networks can have many hidden layers,but in this case for simplicity I have just included one.Finally, Layer 3 is the output layer.You can observe the many connections between the layers,in particular between Layer 1(Li)and Layer 2(L2).As can be seen,each node in Li has a connection to all the nodes in L2.Likewise for the nodes in L2 to the single output node L3.Each of these connections will have an associated weight. 1.2.5 The notation The maths below requires some fairly precise notation so that we know what we are talking about.The notation I am using here is similar to that used in the Stanford deep learning tutorial.In the upcoming equations,each of these weights are identified with the following notation:wij.irefers to the node number of the connection in layer+1and j refers to the node number of the connection in layer I.Take special note of this order.So, for the connection between node 1 in layer 1 and node 2 in layer 2,the weight notation would be w21(1).This notation may seem a bit odd,as you would expect the i and jto refer the node numbers in layers I and l+1 respectively(i.e.in the direction of input to output),rather than the opposite.However,this notation makes more sense when you add the bias. As you can observe in the figure above-the (+1)bias is connected to each of the nodes in the subsequent layer.The bias in layer 1 is connected to the all the nodes in layer two. Because the bias is not a true node with an activation function,it has no inputs(it always outputs the value+).The notation of the bias weight is b,whereiis the node number in the layerl+1-the same as used for the normal weight notation w2().So,the PAGE 9

PAGE 9 network structure consists of an input layer, a hidden layer and an output layer. An example of such a structure can be seen below: Figure 7 Three layer neural network The three layers of the network can be seen in the above figure – Layer 1 represents the input layer, where the external input data enters the network. Layer 2 is called the hidden layer as this layer is not part of the input or output. Note: neural networks can have many hidden layers, but in this case for simplicity I have just included one. Finally, Layer 3 is the output layer. You can observe the many connections between the layers, in particular between Layer 1 (L1) and Layer 2 (L2). As can be seen, each node in L1 has a connection to all the nodes in L2. Likewise for the nodes in L2 to the single output node L3. Each of these connections will have an associated weight. 1.2.5 The notation The maths below requires some fairly precise notation so that we know what we are talking about. The notation I am using here is similar to that used in the Stanford deep learning tutorial. In the upcoming equations, each of these weights are identified with the following notation: 𝑤𝑖𝑗 (𝑙) . i refers to the node number of the connection in layer 𝑙 + 1 and j refers to the node number of the connection in layer l. Take special note of this order. So, for the connection between node 1 in layer 1 and node 2 in layer 2, the weight notation would be 𝑤21 (1) . This notation may seem a bit odd, as you would expect the i and j to refer the node numbers in layers l and 𝑙 + 1 respectively (i.e. in the direction of input to output), rather than the opposite. However, this notation makes more sense when you add the bias. As you can observe in the figure above – the (+1) bias is connected to each of the nodes in the subsequent layer. The bias in layer 1 is connected to the all the nodes in layer two. Because the bias is not a true node with an activation function, it has no inputs (it always outputs the value +1). The notation of the bias weight is 𝑏𝑖 (𝑙) , where i is the node number in the layer 𝑙 + 1 – the same as used for the normal weight notation 𝑤21 (1) . So, the

点击下载完整版文档(PDF)VIP每日下载上限内不扣除下载券和下载次数;
按次数下载不扣除下载券;
24小时内重复下载只扣除一次;
顺序:VIP每日次数-->可用次数-->下载券;
共40页,可试读14页,点击继续阅读 ↓↓
相关文档

关于我们|帮助中心|下载说明|相关软件|意见反馈|联系我们

Copyright © 2008-现在 cucdc.com 高等教育资讯网 版权所有