{"id":208309,"date":"2024-06-24T11:15:08","date_gmt":"2024-06-24T15:15:08","guid":{"rendered":"https:\/\/ibkrcampus.com\/?p=208309"},"modified":"2024-06-24T11:15:32","modified_gmt":"2024-06-24T15:15:32","slug":"forward-propagation-in-neural-networks-components-and-applications-part-i","status":"publish","type":"post","link":"https:\/\/www.interactivebrokers.com\/campus\/ibkr-quant-news\/forward-propagation-in-neural-networks-components-and-applications-part-i\/","title":{"rendered":"Forward Propagation In Neural Networks: Components and Applications &#8211; Part I"},"content":{"rendered":"\n<p>What exactly is forward propagation in neural networks? Well, if we break down the words, &#8220;forward&#8221; implies moving ahead, and &#8220;propagation&#8221; refers to the spreading of something. In neural networks, forward propagation means moving in only one direction: from input to output. Think of it as moving forward in time, where we have no option but to keep moving ahead!<\/p>\n\n\n\n<p>In this blog, we will delve into the intricacies of forward propagation, its calculation process, and its significance in different types of neural networks, including feedforward propagation, CNNs, and ANNs.<\/p>\n\n\n\n<p>We will also explore the components involved, such as activation functions, weights, and biases, and discuss its applications across various domains, including trading. Additionally, we will discuss the examples of forward propagation implemented using Python, along with potential future developments and FAQs.<\/p>\n\n\n\n<p>This blog covers:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>What are neural networks?<\/li>\n\n\n\n<li>What is forward propagation?<\/li>\n\n\n\n<li>Forward propagation algorithm<\/li>\n\n\n\n<li>Forward vs backward propagation in neural network<\/li>\n\n\n\n<li>Forward propagation in different types of neural networks<\/li>\n\n\n\n<li>Components of forward propagation<\/li>\n\n\n\n<li>Applications of forward propagation<\/li>\n\n\n\n<li>Process of forward propagation in trading<\/li>\n\n\n\n<li>Forward propagation in neural networks for trading using Python<\/li>\n\n\n\n<li>Challenges with forward propagation in trading<\/li>\n\n\n\n<li>FAQs while using forward propagation in neural networks for trading<\/li>\n<\/ul>\n\n\n\n<hr class=\"wp-block-separator has-alpha-channel-opacity\" \/>\n\n\n\n<h2 class=\"wp-block-heading\" id=\"what-are-neural-networks\">What are neural networks?<\/h2>\n\n\n\n<p>For centuries, we&#8217;ve been fascinated by how the human mind works. Philosophers have long grappled with understanding human thought processes. However, it&#8217;s only in recent years that we&#8217;ve started making real progress in deciphering how our brains operate. This is where conventional computers diverge from humans.<\/p>\n\n\n\n<p>You see, while we can create algorithms to solve problems, we have to consider all sorts of probabilities. Humans, on the other hand, can start with limited information and still learn and solve problems quickly and accurately. Hence, we began researching and developing artificial brains, now known as&nbsp;<a href=\"https:\/\/blog.quantinsti.com\/neural-network-python\/\">neural networks<\/a>.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"definition-of-a-neural-network\">Definition of a neural network<\/h3>\n\n\n\n<p>A neural network is a computational model inspired by the human brain&#8217;s neural structure, consisting of interconnected layers of artificial neurons. These networks process input data, adjust through learning, and produce outputs, making them effective for tasks like pattern recognition, classification, and predictive modelling.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"what-does-a-neural-network-look-like\">What does a neural network look like?<\/h3>\n\n\n\n<p>A neural network could be simply described as follows:<\/p>\n\n\n\n<figure class=\"wp-block-image size-full\"><img decoding=\"async\" width=\"720\" height=\"271\" data-src=\"\/campus\/wp-content\/uploads\/sites\/2\/2024\/06\/Neuron-1-quantinsti.png\" alt=\"\" class=\"wp-image-208313 lazyload\" data-srcset=\"https:\/\/ibkrcampus.com\/campus\/wp-content\/uploads\/sites\/2\/2024\/06\/Neuron-1-quantinsti.png 720w, https:\/\/ibkrcampus.com\/campus\/wp-content\/uploads\/sites\/2\/2024\/06\/Neuron-1-quantinsti-700x263.png 700w, https:\/\/ibkrcampus.com\/campus\/wp-content\/uploads\/sites\/2\/2024\/06\/Neuron-1-quantinsti-300x113.png 300w\" data-sizes=\"(max-width: 720px) 100vw, 720px\" src=\"data:image\/svg+xml;base64,PHN2ZyB3aWR0aD0iMSIgaGVpZ2h0PSIxIiB4bWxucz0iaHR0cDovL3d3dy53My5vcmcvMjAwMC9zdmciPjwvc3ZnPg==\" style=\"--smush-placeholder-width: 720px; aspect-ratio: 720\/271;\" \/><\/figure>\n\n\n\n<ul class=\"wp-block-list\">\n<li>The basic structure of a neural network is the perceptron, inspired by the neurons in our brains.<\/li>\n\n\n\n<li>In a neural network, there are inputs to the neuron, marked with yellow circles, and then it emits an output signal after processing these inputs.<\/li>\n\n\n\n<li>The input layer resembles the dendrites of a neuron, while the output signal is comparable to the axon. Each input signal is assigned a weight (wi), which is multiplied by the input value. Then the weighted sum of all input variables is stored.<\/li>\n\n\n\n<li>Following this an activation function is applied to the weighted sum, resulting in the output signal.<\/li>\n<\/ul>\n\n\n\n<p>One popular application of neural networks is image recognition software, capable of identifying faces and tagging the same person in different lighting conditions.<\/p>\n\n\n\n<p>Now, let&#8217;s delve into the details of forward propagation beginning with its definition.<\/p>\n\n\n\n<hr class=\"wp-block-separator has-alpha-channel-opacity\" \/>\n\n\n\n<h2 class=\"wp-block-heading\" id=\"what-is-forward-propagation\">What is forward propagation?<\/h2>\n\n\n\n<p>Forward propagation is a fundamental process in neural networks that involves moving input data through the network to produce an output. It&#8217;s essentially the process of feeding input data into the network and computing an output value through the layers of the network.<\/p>\n\n\n\n<p>During forward propagation, each neuron in the network receives input from the previous layer, performs a computation using weights and biases, applies an activation function, and passes the result to the next layer. This process continues until the output is generated. In simple terms, forward propagation is like passing a message through a series of people, with each person adding some information before passing it to the next person until it reaches its destination.<\/p>\n\n\n\n<p>Next, we will see the forward propagation algorithm in detail.<\/p>\n\n\n\n<hr class=\"wp-block-separator has-alpha-channel-opacity\" \/>\n\n\n\n<h2 class=\"wp-block-heading\" id=\"forward-propagation-algorithm\">Forward propagation algorithm<\/h2>\n\n\n\n<p>Here&#8217;s a simplified explanation of the forward propagation algorithm:<\/p>\n\n\n\n<ol class=\"wp-block-list\">\n<li><strong>Input Layer:\u00a0<\/strong>The process begins with the input layer, where the input data is fed into the network.<\/li>\n\n\n\n<li><strong>Hidden Layers:\u00a0<\/strong>The input data is passed through one or more hidden layers. Each neuron in these hidden layers receives input from the previous layer, computes a weighted sum of these inputs, adds a bias term, and applies an activation function.<\/li>\n\n\n\n<li><strong>Output Layer:\u00a0<\/strong>Finally, the processed data moves to the output layer, where the network produces its output.<\/li>\n\n\n\n<li><strong>Error Calculation:\u00a0<\/strong>Once the output is generated, it is compared to the actual output (in the case of supervised learning). The error, also known as the loss, is calculated using a predefined loss function, such as mean squared error or cross-entropy loss.<\/li>\n<\/ol>\n\n\n\n<p>The output of the neural network is then compared to the actual output (in the case of supervised learning) to calculate the error. This error is then used to adjust the weights and biases of the network during the backpropagation phase, which is crucial for training the neural network.<\/p>\n\n\n\n<p>I will explain forward propagation with the help of a simple equation of a line next.<\/p>\n\n\n\n<p>We all know that a line can be represented with the help of the equation:<\/p>\n\n\n\n<pre class=\"EnlighterJSRAW\" data-enlighter-language=\"generic\" data-enlighter-theme=\"\" data-enlighter-highlight=\"\" data-enlighter-linenumbers=\"\" data-enlighter-lineoffset=\"\" data-enlighter-title=\"\" data-enlighter-group=\"\">y = mx + b<\/pre>\n\n\n\n<p>Where,<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>y is the y coordinate of the point<\/li>\n\n\n\n<li>m is the slope<\/li>\n\n\n\n<li>x is the x coordinate<\/li>\n\n\n\n<li>b is the y-intercept i.e. the point at which the line crosses the y-axis<\/li>\n<\/ul>\n\n\n\n<p><em>But why are we jotting the line equation here?<br><\/em>This will help us later on when we understand the components of a neural network in detail.<\/p>\n\n\n\n<p><em>Remember how we said neural networks are supposed to mimic the thinking process of humans?<br><\/em>Well, let us just assume that we do not know the equation of a line, but we do have graph paper and draw a line randomly on it.<\/p>\n\n\n\n<p>For the sake of this example, you drew a line through the origin and when you saw the x and y coordinates, they looked like this:<\/p>\n\n\n\n<figure class=\"wp-block-image size-full\"><img decoding=\"async\" width=\"342\" height=\"314\" data-src=\"\/campus\/wp-content\/uploads\/sites\/2\/2024\/06\/X-cord-Y-cord-quantinsti.png\" alt=\"\" class=\"wp-image-208316 lazyload\" data-srcset=\"https:\/\/ibkrcampus.com\/campus\/wp-content\/uploads\/sites\/2\/2024\/06\/X-cord-Y-cord-quantinsti.png 342w, https:\/\/ibkrcampus.com\/campus\/wp-content\/uploads\/sites\/2\/2024\/06\/X-cord-Y-cord-quantinsti-300x275.png 300w\" data-sizes=\"(max-width: 342px) 100vw, 342px\" src=\"data:image\/svg+xml;base64,PHN2ZyB3aWR0aD0iMSIgaGVpZ2h0PSIxIiB4bWxucz0iaHR0cDovL3d3dy53My5vcmcvMjAwMC9zdmciPjwvc3ZnPg==\" style=\"--smush-placeholder-width: 342px; aspect-ratio: 342\/314;\" \/><\/figure>\n\n\n\n<p>This looks familiar. If I asked you to find the relation between x and y, you would directly say it is y = 3x. But let us go through the process of how forward propagation works. We will assume here that x is the input and y is the output.<\/p>\n\n\n\n<p>The first step here is the initialisation of the parameters. We will guess that y must be a multiplication factor of x. So we will assume that y = 5x and see the results then. Let us add this to the table and see how far we are from the answer.<\/p>\n\n\n\n<figure class=\"wp-block-image size-full\"><img decoding=\"async\" width=\"504\" height=\"368\" data-src=\"\/campus\/wp-content\/uploads\/sites\/2\/2024\/06\/input-output-guess-quantinsti.png\" alt=\"\" class=\"wp-image-208317 lazyload\" data-srcset=\"https:\/\/ibkrcampus.com\/campus\/wp-content\/uploads\/sites\/2\/2024\/06\/input-output-guess-quantinsti.png 504w, https:\/\/ibkrcampus.com\/campus\/wp-content\/uploads\/sites\/2\/2024\/06\/input-output-guess-quantinsti-300x219.png 300w\" data-sizes=\"(max-width: 504px) 100vw, 504px\" src=\"data:image\/svg+xml;base64,PHN2ZyB3aWR0aD0iMSIgaGVpZ2h0PSIxIiB4bWxucz0iaHR0cDovL3d3dy53My5vcmcvMjAwMC9zdmciPjwvc3ZnPg==\" style=\"--smush-placeholder-width: 504px; aspect-ratio: 504\/368;\" \/><\/figure>\n\n\n\n<p>Note that taking the number 5 is just a random guess and nothing else. We could have taken any other number here. I should point out that here we can term 5 as the weight of the model.<\/p>\n\n\n\n<p>All right, this was our first attempt, now we will see how close (or far) we are from the actual output. One way to do that is to use the difference between the actual output and the output we calculated. We will call this the error. Here, we aren\u2019t concerned with the positive or negative sign and hence we take the absolute difference of the error.<\/p>\n\n\n\n<p>Thus, we will update the table now with the error.<\/p>\n\n\n\n<figure class=\"wp-block-image size-full\"><img decoding=\"async\" width=\"720\" height=\"342\" data-src=\"\/campus\/wp-content\/uploads\/sites\/2\/2024\/06\/Error-inclusion-quantinsti.png\" alt=\"\" class=\"wp-image-208324 lazyload\" data-srcset=\"https:\/\/ibkrcampus.com\/campus\/wp-content\/uploads\/sites\/2\/2024\/06\/Error-inclusion-quantinsti.png 720w, https:\/\/ibkrcampus.com\/campus\/wp-content\/uploads\/sites\/2\/2024\/06\/Error-inclusion-quantinsti-700x333.png 700w, https:\/\/ibkrcampus.com\/campus\/wp-content\/uploads\/sites\/2\/2024\/06\/Error-inclusion-quantinsti-300x143.png 300w\" data-sizes=\"(max-width: 720px) 100vw, 720px\" src=\"data:image\/svg+xml;base64,PHN2ZyB3aWR0aD0iMSIgaGVpZ2h0PSIxIiB4bWxucz0iaHR0cDovL3d3dy53My5vcmcvMjAwMC9zdmciPjwvc3ZnPg==\" style=\"--smush-placeholder-width: 720px; aspect-ratio: 720\/342;\" \/><\/figure>\n\n\n\n<p>If we take the sum of this error, we get the value 30. But why did we total the error? Since we are going to try multiple guesses to come to the closest answer, we need to know how close or how far we were from the previous answers. This helps us refine our guesses and calculate the correct answer.<\/p>\n\n\n\n<p>Wait. But if we just add up all the error values, it feels like we are giving equal weightage to all the answers. Shouldn\u2019t we penalise the values which are way off the mark? For example, 10 here is much higher than 2. It is here that we introduce the somewhat famous \u201cSum of squared Errors\u201d or SSE for short. In SSE, we square all the error values and then add them. Thus, the error values which are very high get exaggerated and thus, help us in knowing how to proceed further.<\/p>\n\n\n\n<p>Let\u2019s put these values in the table below.<\/p>\n\n\n\n<figure class=\"wp-block-image size-full\"><img decoding=\"async\" width=\"720\" height=\"345\" data-src=\"\/campus\/wp-content\/uploads\/sites\/2\/2024\/06\/Square-of-errors-quantinsti.png\" alt=\"\" class=\"wp-image-208326 lazyload\" data-srcset=\"https:\/\/ibkrcampus.com\/campus\/wp-content\/uploads\/sites\/2\/2024\/06\/Square-of-errors-quantinsti.png 720w, https:\/\/ibkrcampus.com\/campus\/wp-content\/uploads\/sites\/2\/2024\/06\/Square-of-errors-quantinsti-700x335.png 700w, https:\/\/ibkrcampus.com\/campus\/wp-content\/uploads\/sites\/2\/2024\/06\/Square-of-errors-quantinsti-300x144.png 300w\" data-sizes=\"(max-width: 720px) 100vw, 720px\" src=\"data:image\/svg+xml;base64,PHN2ZyB3aWR0aD0iMSIgaGVpZ2h0PSIxIiB4bWxucz0iaHR0cDovL3d3dy53My5vcmcvMjAwMC9zdmciPjwvc3ZnPg==\" style=\"--smush-placeholder-width: 720px; aspect-ratio: 720\/345;\" \/><\/figure>\n\n\n\n<p>Now the SSE for the weight 5 (Recall that we assumed y = 5x), is 145. We call this the loss function. The loss function is important to understand the efficiency of the neural network and also helps us when we incorporate&nbsp;<a href=\"https:\/\/blog.quantinsti.com\/backpropagation\/\">backpropagation<\/a>&nbsp;in the neural network.<\/p>\n\n\n\n<p>All right, so far we understood the principle of how the neural network tries to learn. We have also seen the basic principle of the neuron. Next, we will see the forward vs backward propagation in the neural network.<\/p>\n\n\n\n<p>Author: Chainika Thakar (Originally written by Varun Divakar and Rekhit Pachanekar)<\/p>\n\n\n\n<p><em>Originally posted on <a href=\"https:\/\/blog.quantinsti.com\/forward-propagation-neural-networks\/\">QuantInsti<\/a>.<\/em><\/p>\n\n\n\n<p><em>Stay tuned for Part II to read about forward vs backward propagation in neural network<\/em>.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>In this blog, we will delve into the intricacies of forward propagation, its calculation process, and its significance in different types of neural networks, including feedforward propagation, CNNs, and ANNs.<\/p>\n","protected":false},"author":186,"featured_media":112415,"comment_status":"open","ping_status":"closed","sticky":true,"template":"","format":"standard","meta":{"_acf_changed":false,"footnotes":""},"categories":[339,343,349,338,341],"tags":[4087,806,4363,852,15593,595],"contributors-categories":[13654],"class_list":{"0":"post-208309","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-data-science","8":"category-programing-languages","9":"category-python-development","10":"category-ibkr-quant-news","11":"category-quant-development","12":"tag-backpropagation","13":"tag-data-science","14":"tag-forward-propagation","15":"tag-machine-learning","16":"tag-neural-networks","17":"tag-python","18":"contributors-categories-quantinsti"},"pp_statuses_selecting_workflow":false,"pp_workflow_action":"current","pp_status_selection":"publish","acf":[],"yoast_head":"<!-- This site is optimized with the Yoast SEO Premium plugin v26.9 (Yoast SEO v27.3) - https:\/\/yoast.com\/product\/yoast-seo-premium-wordpress\/ -->\n<title>Forward Propagation In Neural Networks: Components and Applications &#8211; Part I<\/title>\n<meta name=\"description\" content=\"In this blog, we will delve into the intricacies of forward propagation, its calculation process, and its significance in different types of neural...\" \/>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/www.interactivebrokers.com\/campus\/wp-json\/wp\/v2\/posts\/208309\/\" \/>\n<meta property=\"og:locale\" content=\"en_US\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"Forward Propagation In Neural Networks: Components and Applications \u2013 Part I\" \/>\n<meta property=\"og:description\" content=\"In this blog, we will delve into the intricacies of forward propagation, its calculation process, and its significance in different types of neural networks, including feedforward propagation, CNNs, and ANNs.\" \/>\n<meta property=\"og:url\" content=\"https:\/\/www.interactivebrokers.com\/campus\/ibkr-quant-news\/forward-propagation-in-neural-networks-components-and-applications-part-i\/\" \/>\n<meta property=\"og:site_name\" content=\"IBKR Campus US\" \/>\n<meta property=\"article:published_time\" content=\"2024-06-24T15:15:08+00:00\" \/>\n<meta property=\"article:modified_time\" content=\"2024-06-24T15:15:32+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/www.interactivebrokers.com\/campus\/wp-content\/uploads\/sites\/2\/2021\/11\/globe-candles.jpg\" \/>\n\t<meta property=\"og:image:width\" content=\"1000\" \/>\n\t<meta property=\"og:image:height\" content=\"563\" \/>\n\t<meta property=\"og:image:type\" content=\"image\/jpeg\" \/>\n<meta name=\"author\" content=\"Contributor Author\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:label1\" content=\"Written by\" \/>\n\t<meta name=\"twitter:data1\" content=\"Contributor Author\" \/>\n\t<meta name=\"twitter:label2\" content=\"Est. reading time\" \/>\n\t<meta name=\"twitter:data2\" content=\"8 minutes\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\n\t    \"@context\": \"https:\\\/\\\/schema.org\",\n\t    \"@graph\": [\n\t        {\n\t            \"@type\": \"NewsArticle\",\n\t            \"@id\": \"https:\\\/\\\/ibkrcampus.com\\\/campus\\\/ibkr-quant-news\\\/forward-propagation-in-neural-networks-components-and-applications-part-i\\\/#article\",\n\t            \"isPartOf\": {\n\t                \"@id\": \"https:\\\/\\\/ibkrcampus.com\\\/campus\\\/ibkr-quant-news\\\/forward-propagation-in-neural-networks-components-and-applications-part-i\\\/\"\n\t            },\n\t            \"author\": {\n\t                \"name\": \"Contributor Author\",\n\t                \"@id\": \"https:\\\/\\\/ibkrcampus.com\\\/campus\\\/#\\\/schema\\\/person\\\/e823e46b42ca381080387e794318a485\"\n\t            },\n\t            \"headline\": \"Forward Propagation In Neural Networks: Components and Applications &#8211; Part I\",\n\t            \"datePublished\": \"2024-06-24T15:15:08+00:00\",\n\t            \"dateModified\": \"2024-06-24T15:15:32+00:00\",\n\t            \"mainEntityOfPage\": {\n\t                \"@id\": \"https:\\\/\\\/ibkrcampus.com\\\/campus\\\/ibkr-quant-news\\\/forward-propagation-in-neural-networks-components-and-applications-part-i\\\/\"\n\t            },\n\t            \"wordCount\": 1476,\n\t            \"commentCount\": 0,\n\t            \"publisher\": {\n\t                \"@id\": \"https:\\\/\\\/ibkrcampus.com\\\/campus\\\/#organization\"\n\t            },\n\t            \"image\": {\n\t                \"@id\": \"https:\\\/\\\/ibkrcampus.com\\\/campus\\\/ibkr-quant-news\\\/forward-propagation-in-neural-networks-components-and-applications-part-i\\\/#primaryimage\"\n\t            },\n\t            \"thumbnailUrl\": \"https:\\\/\\\/www.interactivebrokers.com\\\/campus\\\/wp-content\\\/uploads\\\/sites\\\/2\\\/2021\\\/11\\\/globe-candles.jpg\",\n\t            \"keywords\": [\n\t                \"Backpropagation\",\n\t                \"Data Science\",\n\t                \"Forward propagation\",\n\t                \"Machine Learning\",\n\t                \"Neural Networks\",\n\t                \"Python\"\n\t            ],\n\t            \"articleSection\": [\n\t                \"Data Science\",\n\t                \"Programming Languages\",\n\t                \"Python Development\",\n\t                \"Quant\",\n\t                \"Quant Development\"\n\t            ],\n\t            \"inLanguage\": \"en-US\",\n\t            \"potentialAction\": [\n\t                {\n\t                    \"@type\": \"CommentAction\",\n\t                    \"name\": \"Comment\",\n\t                    \"target\": [\n\t                        \"https:\\\/\\\/ibkrcampus.com\\\/campus\\\/ibkr-quant-news\\\/forward-propagation-in-neural-networks-components-and-applications-part-i\\\/#respond\"\n\t                    ]\n\t                }\n\t            ]\n\t        },\n\t        {\n\t            \"@type\": \"WebPage\",\n\t            \"@id\": \"https:\\\/\\\/ibkrcampus.com\\\/campus\\\/ibkr-quant-news\\\/forward-propagation-in-neural-networks-components-and-applications-part-i\\\/\",\n\t            \"url\": \"https:\\\/\\\/ibkrcampus.com\\\/campus\\\/ibkr-quant-news\\\/forward-propagation-in-neural-networks-components-and-applications-part-i\\\/\",\n\t            \"name\": \"Forward Propagation In Neural Networks: Components and Applications - Part I | IBKR Campus US\",\n\t            \"isPartOf\": {\n\t                \"@id\": \"https:\\\/\\\/ibkrcampus.com\\\/campus\\\/#website\"\n\t            },\n\t            \"primaryImageOfPage\": {\n\t                \"@id\": \"https:\\\/\\\/ibkrcampus.com\\\/campus\\\/ibkr-quant-news\\\/forward-propagation-in-neural-networks-components-and-applications-part-i\\\/#primaryimage\"\n\t            },\n\t            \"image\": {\n\t                \"@id\": \"https:\\\/\\\/ibkrcampus.com\\\/campus\\\/ibkr-quant-news\\\/forward-propagation-in-neural-networks-components-and-applications-part-i\\\/#primaryimage\"\n\t            },\n\t            \"thumbnailUrl\": \"https:\\\/\\\/www.interactivebrokers.com\\\/campus\\\/wp-content\\\/uploads\\\/sites\\\/2\\\/2021\\\/11\\\/globe-candles.jpg\",\n\t            \"datePublished\": \"2024-06-24T15:15:08+00:00\",\n\t            \"dateModified\": \"2024-06-24T15:15:32+00:00\",\n\t            \"description\": \"In this blog, we will delve into the intricacies of forward propagation, its calculation process, and its significance in different types of neural networks, including feedforward propagation, CNNs, and ANNs.\",\n\t            \"inLanguage\": \"en-US\",\n\t            \"potentialAction\": [\n\t                {\n\t                    \"@type\": \"ReadAction\",\n\t                    \"target\": [\n\t                        \"https:\\\/\\\/ibkrcampus.com\\\/campus\\\/ibkr-quant-news\\\/forward-propagation-in-neural-networks-components-and-applications-part-i\\\/\"\n\t                    ]\n\t                }\n\t            ]\n\t        },\n\t        {\n\t            \"@type\": \"ImageObject\",\n\t            \"inLanguage\": \"en-US\",\n\t            \"@id\": \"https:\\\/\\\/ibkrcampus.com\\\/campus\\\/ibkr-quant-news\\\/forward-propagation-in-neural-networks-components-and-applications-part-i\\\/#primaryimage\",\n\t            \"url\": \"https:\\\/\\\/www.interactivebrokers.com\\\/campus\\\/wp-content\\\/uploads\\\/sites\\\/2\\\/2021\\\/11\\\/globe-candles.jpg\",\n\t            \"contentUrl\": \"https:\\\/\\\/www.interactivebrokers.com\\\/campus\\\/wp-content\\\/uploads\\\/sites\\\/2\\\/2021\\\/11\\\/globe-candles.jpg\",\n\t            \"width\": 1000,\n\t            \"height\": 563,\n\t            \"caption\": \"Quant\"\n\t        },\n\t        {\n\t            \"@type\": \"WebSite\",\n\t            \"@id\": \"https:\\\/\\\/ibkrcampus.com\\\/campus\\\/#website\",\n\t            \"url\": \"https:\\\/\\\/ibkrcampus.com\\\/campus\\\/\",\n\t            \"name\": \"IBKR Campus US\",\n\t            \"description\": \"Financial Education from Interactive Brokers\",\n\t            \"publisher\": {\n\t                \"@id\": \"https:\\\/\\\/ibkrcampus.com\\\/campus\\\/#organization\"\n\t            },\n\t            \"potentialAction\": [\n\t                {\n\t                    \"@type\": \"SearchAction\",\n\t                    \"target\": {\n\t                        \"@type\": \"EntryPoint\",\n\t                        \"urlTemplate\": \"https:\\\/\\\/ibkrcampus.com\\\/campus\\\/?s={search_term_string}\"\n\t                    },\n\t                    \"query-input\": {\n\t                        \"@type\": \"PropertyValueSpecification\",\n\t                        \"valueRequired\": true,\n\t                        \"valueName\": \"search_term_string\"\n\t                    }\n\t                }\n\t            ],\n\t            \"inLanguage\": \"en-US\"\n\t        },\n\t        {\n\t            \"@type\": \"Organization\",\n\t            \"@id\": \"https:\\\/\\\/ibkrcampus.com\\\/campus\\\/#organization\",\n\t            \"name\": \"Interactive Brokers\",\n\t            \"alternateName\": \"IBKR\",\n\t            \"url\": \"https:\\\/\\\/ibkrcampus.com\\\/campus\\\/\",\n\t            \"logo\": {\n\t                \"@type\": \"ImageObject\",\n\t                \"inLanguage\": \"en-US\",\n\t                \"@id\": \"https:\\\/\\\/ibkrcampus.com\\\/campus\\\/#\\\/schema\\\/logo\\\/image\\\/\",\n\t                \"url\": \"https:\\\/\\\/www.interactivebrokers.com\\\/campus\\\/wp-content\\\/uploads\\\/sites\\\/2\\\/2024\\\/05\\\/ibkr-campus-logo.jpg\",\n\t                \"contentUrl\": \"https:\\\/\\\/www.interactivebrokers.com\\\/campus\\\/wp-content\\\/uploads\\\/sites\\\/2\\\/2024\\\/05\\\/ibkr-campus-logo.jpg\",\n\t                \"width\": 669,\n\t                \"height\": 669,\n\t                \"caption\": \"Interactive Brokers\"\n\t            },\n\t            \"image\": {\n\t                \"@id\": \"https:\\\/\\\/ibkrcampus.com\\\/campus\\\/#\\\/schema\\\/logo\\\/image\\\/\"\n\t            },\n\t            \"publishingPrinciples\": \"https:\\\/\\\/www.interactivebrokers.com\\\/campus\\\/about-ibkr-campus\\\/\",\n\t            \"ethicsPolicy\": \"https:\\\/\\\/www.interactivebrokers.com\\\/campus\\\/cyber-security-notice\\\/\"\n\t        },\n\t        {\n\t            \"@type\": \"Person\",\n\t            \"@id\": \"https:\\\/\\\/ibkrcampus.com\\\/campus\\\/#\\\/schema\\\/person\\\/e823e46b42ca381080387e794318a485\",\n\t            \"name\": \"Contributor Author\",\n\t            \"url\": \"https:\\\/\\\/www.interactivebrokers.com\\\/campus\\\/author\\\/contributor-author\\\/\"\n\t        }\n\t    ]\n\t}<\/script>\n<!-- \/ Yoast SEO Premium plugin. -->","yoast_head_json":{"title":"Forward Propagation In Neural Networks: Components and Applications &#8211; Part I","description":"In this blog, we will delve into the intricacies of forward propagation, its calculation process, and its significance in different types of neural...","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/www.interactivebrokers.com\/campus\/wp-json\/wp\/v2\/posts\/208309\/","og_locale":"en_US","og_type":"article","og_title":"Forward Propagation In Neural Networks: Components and Applications \u2013 Part I","og_description":"In this blog, we will delve into the intricacies of forward propagation, its calculation process, and its significance in different types of neural networks, including feedforward propagation, CNNs, and ANNs.","og_url":"https:\/\/www.interactivebrokers.com\/campus\/ibkr-quant-news\/forward-propagation-in-neural-networks-components-and-applications-part-i\/","og_site_name":"IBKR Campus US","article_published_time":"2024-06-24T15:15:08+00:00","article_modified_time":"2024-06-24T15:15:32+00:00","og_image":[{"width":1000,"height":563,"url":"https:\/\/www.interactivebrokers.com\/campus\/wp-content\/uploads\/sites\/2\/2021\/11\/globe-candles.jpg","type":"image\/jpeg"}],"author":"Contributor Author","twitter_card":"summary_large_image","twitter_misc":{"Written by":"Contributor Author","Est. reading time":"8 minutes"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"NewsArticle","@id":"https:\/\/ibkrcampus.com\/campus\/ibkr-quant-news\/forward-propagation-in-neural-networks-components-and-applications-part-i\/#article","isPartOf":{"@id":"https:\/\/ibkrcampus.com\/campus\/ibkr-quant-news\/forward-propagation-in-neural-networks-components-and-applications-part-i\/"},"author":{"name":"Contributor Author","@id":"https:\/\/ibkrcampus.com\/campus\/#\/schema\/person\/e823e46b42ca381080387e794318a485"},"headline":"Forward Propagation In Neural Networks: Components and Applications &#8211; Part I","datePublished":"2024-06-24T15:15:08+00:00","dateModified":"2024-06-24T15:15:32+00:00","mainEntityOfPage":{"@id":"https:\/\/ibkrcampus.com\/campus\/ibkr-quant-news\/forward-propagation-in-neural-networks-components-and-applications-part-i\/"},"wordCount":1476,"commentCount":0,"publisher":{"@id":"https:\/\/ibkrcampus.com\/campus\/#organization"},"image":{"@id":"https:\/\/ibkrcampus.com\/campus\/ibkr-quant-news\/forward-propagation-in-neural-networks-components-and-applications-part-i\/#primaryimage"},"thumbnailUrl":"https:\/\/www.interactivebrokers.com\/campus\/wp-content\/uploads\/sites\/2\/2021\/11\/globe-candles.jpg","keywords":["Backpropagation","Data Science","Forward propagation","Machine Learning","Neural Networks","Python"],"articleSection":["Data Science","Programming Languages","Python Development","Quant","Quant Development"],"inLanguage":"en-US","potentialAction":[{"@type":"CommentAction","name":"Comment","target":["https:\/\/ibkrcampus.com\/campus\/ibkr-quant-news\/forward-propagation-in-neural-networks-components-and-applications-part-i\/#respond"]}]},{"@type":"WebPage","@id":"https:\/\/ibkrcampus.com\/campus\/ibkr-quant-news\/forward-propagation-in-neural-networks-components-and-applications-part-i\/","url":"https:\/\/ibkrcampus.com\/campus\/ibkr-quant-news\/forward-propagation-in-neural-networks-components-and-applications-part-i\/","name":"Forward Propagation In Neural Networks: Components and Applications - Part I | IBKR Campus US","isPartOf":{"@id":"https:\/\/ibkrcampus.com\/campus\/#website"},"primaryImageOfPage":{"@id":"https:\/\/ibkrcampus.com\/campus\/ibkr-quant-news\/forward-propagation-in-neural-networks-components-and-applications-part-i\/#primaryimage"},"image":{"@id":"https:\/\/ibkrcampus.com\/campus\/ibkr-quant-news\/forward-propagation-in-neural-networks-components-and-applications-part-i\/#primaryimage"},"thumbnailUrl":"https:\/\/www.interactivebrokers.com\/campus\/wp-content\/uploads\/sites\/2\/2021\/11\/globe-candles.jpg","datePublished":"2024-06-24T15:15:08+00:00","dateModified":"2024-06-24T15:15:32+00:00","description":"In this blog, we will delve into the intricacies of forward propagation, its calculation process, and its significance in different types of neural networks, including feedforward propagation, CNNs, and ANNs.","inLanguage":"en-US","potentialAction":[{"@type":"ReadAction","target":["https:\/\/ibkrcampus.com\/campus\/ibkr-quant-news\/forward-propagation-in-neural-networks-components-and-applications-part-i\/"]}]},{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/ibkrcampus.com\/campus\/ibkr-quant-news\/forward-propagation-in-neural-networks-components-and-applications-part-i\/#primaryimage","url":"https:\/\/www.interactivebrokers.com\/campus\/wp-content\/uploads\/sites\/2\/2021\/11\/globe-candles.jpg","contentUrl":"https:\/\/www.interactivebrokers.com\/campus\/wp-content\/uploads\/sites\/2\/2021\/11\/globe-candles.jpg","width":1000,"height":563,"caption":"Quant"},{"@type":"WebSite","@id":"https:\/\/ibkrcampus.com\/campus\/#website","url":"https:\/\/ibkrcampus.com\/campus\/","name":"IBKR Campus US","description":"Financial Education from Interactive Brokers","publisher":{"@id":"https:\/\/ibkrcampus.com\/campus\/#organization"},"potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/ibkrcampus.com\/campus\/?s={search_term_string}"},"query-input":{"@type":"PropertyValueSpecification","valueRequired":true,"valueName":"search_term_string"}}],"inLanguage":"en-US"},{"@type":"Organization","@id":"https:\/\/ibkrcampus.com\/campus\/#organization","name":"Interactive Brokers","alternateName":"IBKR","url":"https:\/\/ibkrcampus.com\/campus\/","logo":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/ibkrcampus.com\/campus\/#\/schema\/logo\/image\/","url":"https:\/\/www.interactivebrokers.com\/campus\/wp-content\/uploads\/sites\/2\/2024\/05\/ibkr-campus-logo.jpg","contentUrl":"https:\/\/www.interactivebrokers.com\/campus\/wp-content\/uploads\/sites\/2\/2024\/05\/ibkr-campus-logo.jpg","width":669,"height":669,"caption":"Interactive Brokers"},"image":{"@id":"https:\/\/ibkrcampus.com\/campus\/#\/schema\/logo\/image\/"},"publishingPrinciples":"https:\/\/www.interactivebrokers.com\/campus\/about-ibkr-campus\/","ethicsPolicy":"https:\/\/www.interactivebrokers.com\/campus\/cyber-security-notice\/"},{"@type":"Person","@id":"https:\/\/ibkrcampus.com\/campus\/#\/schema\/person\/e823e46b42ca381080387e794318a485","name":"Contributor Author","url":"https:\/\/www.interactivebrokers.com\/campus\/author\/contributor-author\/"}]}},"jetpack_featured_media_url":"https:\/\/www.interactivebrokers.com\/campus\/wp-content\/uploads\/sites\/2\/2021\/11\/globe-candles.jpg","_links":{"self":[{"href":"https:\/\/ibkrcampus.com\/campus\/wp-json\/wp\/v2\/posts\/208309","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/ibkrcampus.com\/campus\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/ibkrcampus.com\/campus\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/ibkrcampus.com\/campus\/wp-json\/wp\/v2\/users\/186"}],"replies":[{"embeddable":true,"href":"https:\/\/ibkrcampus.com\/campus\/wp-json\/wp\/v2\/comments?post=208309"}],"version-history":[{"count":0,"href":"https:\/\/ibkrcampus.com\/campus\/wp-json\/wp\/v2\/posts\/208309\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/ibkrcampus.com\/campus\/wp-json\/wp\/v2\/media\/112415"}],"wp:attachment":[{"href":"https:\/\/ibkrcampus.com\/campus\/wp-json\/wp\/v2\/media?parent=208309"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/ibkrcampus.com\/campus\/wp-json\/wp\/v2\/categories?post=208309"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/ibkrcampus.com\/campus\/wp-json\/wp\/v2\/tags?post=208309"},{"taxonomy":"contributors-categories","embeddable":true,"href":"https:\/\/ibkrcampus.com\/campus\/wp-json\/wp\/v2\/contributors-categories?post=208309"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}