Conventional applications of neural networks usually predict a single value as a function of given inputs. In forecasting, for example, a standard objective is to predict the future value of some entity of interest on the basis of a time series of past measurements or observations. Typical training schemes aim to minimise the sum of squared deviations between predicted and actual values (the 'targets'), by which, ideally, the network learns the conditional mean of the target given the input. If the underlying conditional distribution is Gaus sian or at least unimodal, this may be a satisfactory approach. However, for a multimodal distribution, the conditional mean does not capture the relevant features of the system, and the prediction performance will, in general, be very poor. This calls for a more powerful and sophisticated model, which can learn the whole conditional probability distribution. Chapter 1 demonstrates that even for a deterministic system and 'be nign' Gaussian observational noise, the conditional distribution of a future observation, conditional on a set of past observations, can become strongly skewed and multimodal. In Chapter 2, a general neural network structure for modelling conditional probability densities is derived, and it is shown that a universal approximator for this extended task requires at least two hidden layers. A training scheme is developed from a maximum likelihood approach in Chapter 3, and the performance ofthis method is demonstrated on three stochastic time series in chapters 4 and 5.
"synopsis" may belong to another edition of this title.
Using two alternative approaches, the GM network and the GM-RVFL model, this study of the applications of neural networks includes a case study of the prediction of housing prices in the Boston metropolitan area, and examines the required modifications to standard approaches.
"About this title" may belong to another edition of this title.
£ 5.94 shipping within U.S.A.
Destination, rates & speedsSeller: Solr Books, Lincolnwood, IL, U.S.A.
Condition: very_good. This books is in Very good condition. There may be a few flaws like shelf wear and some light wear. Seller Inventory # 5D4000008WHW_ns
Quantity: 1 available
Seller: Phatpocket Limited, Waltham Abbey, HERTS, United Kingdom
Condition: Good. Your purchase helps support Sri Lankan Children's Charity 'The Rainbow Centre'. Ex-library, so some stamps and wear, but in good overall condition. Our donations to The Rainbow Centre have helped provide an education and a safe haven to hundreds of children who live in appalling conditions. Seller Inventory # Z1-B-017-02028
Quantity: 1 available
Seller: Best Price, Torrance, CA, U.S.A.
Condition: New. SUPER FAST SHIPPING. Seller Inventory # 9781852330958
Quantity: 2 available
Seller: GreatBookPrices, Columbia, MD, U.S.A.
Condition: New. Seller Inventory # 890682-n
Quantity: 15 available
Seller: Grand Eagle Retail, Mason, OH, U.S.A.
Paperback. Condition: new. Paperback. Conventional applications of neural networks usually predict a single value as a function of given inputs. In forecasting, for example, a standard objective is to predict the future value of some entity of interest on the basis of a time series of past measurements or observations. Typical training schemes aim to minimise the sum of squared deviations between predicted and actual values (the 'targets'), by which, ideally, the network learns the conditional mean of the target given the input. If the underlying conditional distribution is Gaus sian or at least unimodal, this may be a satisfactory approach. However, for a multimodal distribution, the conditional mean does not capture the relevant features of the system, and the prediction performance will, in general, be very poor. This calls for a more powerful and sophisticated model, which can learn the whole conditional probability distribution. Chapter 1 demonstrates that even for a deterministic system and 'be nign' Gaussian observational noise, the conditional distribution of a future observation, conditional on a set of past observations, can become strongly skewed and multimodal. In Chapter 2, a general neural network structure for modelling conditional probability densities is derived, and it is shown that a universal approximator for this extended task requires at least two hidden layers. A training scheme is developed from a maximum likelihood approach in Chapter 3, and the performance ofthis method is demonstrated on three stochastic time series in chapters 4 and 5. Chapter 1 demonstrates that even for a deterministic system and 'be nign' Gaussian observational noise, the conditional distribution of a future observation, conditional on a set of past observations, can become strongly skewed and multimodal. Shipping may be from multiple locations in the US or from the UK, depending on stock availability. Seller Inventory # 9781852330958
Quantity: 1 available
Seller: GreatBookPrices, Columbia, MD, U.S.A.
Condition: As New. Unread book in perfect condition. Seller Inventory # 890682
Quantity: 15 available
Seller: California Books, Miami, FL, U.S.A.
Condition: New. Seller Inventory # I-9781852330958
Quantity: Over 20 available
Seller: Ria Christie Collections, Uxbridge, United Kingdom
Condition: New. In. Seller Inventory # ria9781852330958_new
Quantity: Over 20 available
Seller: Chiron Media, Wallingford, United Kingdom
Paperback. Condition: New. Seller Inventory # 6666-IUK-9781852330958
Quantity: 10 available
Seller: Roland Antiquariat UG haftungsbeschränkt, Weinheim, Germany
Softcover. XXIII, 275 S. : graph. Darst. ; 24 cm Like new. Unread book. --- Neuwertiger Zustand. Ungelesenes Buch. 9781852330958 Sprache: Deutsch Gewicht in Gramm: 467 Softcover reprint of the original 1st ed. 1999. Seller Inventory # 200027
Quantity: 1 available