Neural Network Learning: Theoretical Foundations

by
Edition: 1st
Format: Paperback
Pub. Date: 2009-08-20
Publisher(s): Cambridge University Press
  • Free Shipping Icon

    Receive Free Shipping To The More Store!*

    *Marketplace items do not qualify for the free shipping promotion.

  • eCampus.com Device Compatibility Matrix

    Click the device icon to install or view instructions

    Apple iOS | iPad, iPhone, iPod
    Apple iOS | iPad, iPhone, iPod
    Android Devices | Android Tables & Phones OS 2.2 or higher | *Kindle Fire
    Android Devices | Android Tables & Phones OS 2.2 or higher | *Kindle Fire
    Windows 10 / 8 / 7 / Vista / XP
    Windows 10 / 8 / 7 / Vista / XP
    Mac OS X | **iMac / Macbook
    Mac OS X | **iMac / Macbook
    Enjoy offline reading with these devices
    Apple Devices
    Android Devices
    Windows Devices
    Mac Devices
    iPad, iPhone, iPod
    Our reader is compatible
     
     
     
    Android 2.2 +
     
    Our reader is compatible
     
     
    Kindle Fire
     
    Our reader is compatible
     
     
    Windows
    10 / 8 / 7 / Vista / XP
     
     
    Our reader is compatible
     
    Mac
     
     
     
    Our reader is compatible
List Price: $62.00

Buy New

Usually Ships in 8 - 10 Business Days.
$61.94

Rent Textbook

Select for Price
There was a problem. Please try again later.

Rent Digital

Rent Digital Options
Online:1825 Days access
Downloadable:Lifetime Access
$73.19
*To support the delivery of the digital material to you, a digital delivery fee of $3.99 will be charged on each digital item.
$73.19*

Used Textbook

We're Sorry
Sold Out

How Marketplace Works:

  • This item is offered by an independent seller and not shipped from our warehouse
  • Item details like edition and cover design may differ from our description; see seller's comments before ordering.
  • Sellers much confirm and ship within two business days; otherwise, the order will be cancelled and refunded.
  • Marketplace purchases cannot be returned to eCampus.com. Contact the seller directly for inquiries; if no response within two days, contact customer service.
  • Additional shipping costs apply to Marketplace purchases. Review shipping costs at checkout.

Summary

This important work describes recent theoretical advances in the study of artificial neural networks. It explores probabilistic models of supervised learning problems, and addresses the key statistical and computational questions. Chapters survey research on pattern classification with binary-output networks, including a discussion of the relevance of the Vapnik Chervonenkis dimension, and of estimates of the dimension for several neural network models. In addition, Anthony and Bartlett develop a model of classification by real-output networks, and demonstrate the usefulness of classification with a "large margin." The authors explain the role of scale-sensitive versions of the Vapnik Chervonenkis dimension in large margin classification, and in real prediction. Key chapters also discuss the computational complexity of neural network learning, describing a variety of hardness results, and outlining two efficient, constructive learning algorithms. The book is self-contained and accessible to researchers and graduate students in computer science, engineering, and mathematics.

Table of Contents

Introduction
Pattern Recognition with Binary-output Neural Networks:
The pattern recognition problem
The growth function and VC-dimension
General upper bounds on sample complexity
General lower bounds
The VC-dimension of linear threshold networks
Bounding the VC-dimension using geometric techniques
VC-dimension bounds for neural networks
Pattern Recognition with Real-output Neural Networks:
Classification with real values
Covering numbers and uniform convergence
The pseudo-dimension and fat-shattering dimension
Bounding covering numbers with dimensions
The sample complexity of classification learning
The dimensions of neural networks
Model selection
Learning Real-Valued Functions:
Learning classes of real functions
Uniform convergence results for real function classes
Bounding covering numbers
The sample complexity of learning function classes
Convex classes
Other learning problems
Algorithmics:
Efficient learning
Learning as optimisation
The Boolean perceptron
Hardness results for feed-forward networks
Constructive learning algorithms for two-layered networks
Table of Contents provided by Publisher. All Rights Reserved.

An electronic version of this book is available through VitalSource.

This book is viewable on PC, Mac, iPhone, iPad, iPod Touch, and most smartphones.

By purchasing, you will be able to view this book online, as well as download it, for the chosen number of days.

Digital License

You are licensing a digital product for a set duration. Durations are set forth in the product description, with "Lifetime" typically meaning five (5) years of online access and permanent download to a supported device. All licenses are non-transferable.

More details can be found here.

A downloadable version of this book is available through the eCampus Reader or compatible Adobe readers.

Applications are available on iOS, Android, PC, Mac, and Windows Mobile platforms.

Please view the compatibility matrix prior to purchase.