人工神经网络原理 英文版_第1页
人工神经网络原理 英文版_第2页
人工神经网络原理 英文版_第3页
人工神经网络原理 英文版_第4页
人工神经网络原理 英文版_第5页
已阅读5页,还剩370页未读 继续免费阅读

下载本文档

版权说明:本文档由用户提供并上传,收益归属内容提供方,若内容存在侵权,请进行举报或认领

文档简介

1、FundamentalsofArtificialNeuralNetworksbyMohamadH.Hassoun(MITPress,1995) HYPERLINK / /PrefaceMypurposeinwritingthisbookhasbeentogiveasystematicaccountofmajorconceptsandmethodologiesofartificialneuralnetworksandtopresentaunifiedframeworkthatmakesthesubjectmoreaccessibletostudentsandpractitioners.Thisb

2、ookemphasizesfundamentaltheoreticalaspectsofthecomputationalcapabilitiesandlearningabilitiesofartificialneuralnetworks.Itintegratesimportanttheoreticalresultsonartificialneuralnetworksandusesthemtoexplainawiderangeofexistingempiricalobservationsandcommonlyusedheuristics.Themainaudienceisfirst-yeargr

3、aduatestudentsinelectricalengineering,computerengineering,andcomputerscience.Thisbookmaybeadaptedforuseasaseniorundergraduatetextbookbyselectivechoiceoftopics.Alternatively,itmayalsobeusedasavaluableresourceforpracticingengineers,computerscientists,andothersinvolvedinresearchinartificialneuralnetwor

4、ks.Thisbookhasevolvedfromlecturenotesoftwocoursesonartificialneuralnetworks,asenior-levelcourseandagraduate-levelcourse,whichIhavetaughtduringthelast6yearsintheDepartmentofElectricalandComputerEngineeringatWayneStateUniversity.Thebackgroundmaterialneededtounderstandthisbookisgeneralknowledgeofsomeba

5、sictopicsinmathematics,suchasprobabilityandstatistics,differentialequationsandlinearalgebra,andsomethingaboutmultivariatecalculus.Thereaderisalsoassumedtohaveenoughfamiliaritywiththeconceptofasystemandthenotionofstate,aswellaswiththebasicelementsofBooleanalgebraandswitchingtheory.Therequiredtechnica

6、lmaturityisthatofaseniorundergraduateinelectricalengineering,computerengineering,orcomputerscience.Artificialneuralnetworksareviewedhereasparallelcomputationalmodels,withvaryingdegreesofcomplexity,comprisedofdenselyinterconnectedadaptiveprocessingunits.Thesenetworksarefine-grainedparallelimplementat

7、ionsofnonlinearstaticordynamicsystems.Averyimportantfeatureofthesenetworksistheiradaptivenature,wherelearningbyexamplereplacestraditionalprogramminginsolvingproblems.Thisfeaturemakessuchcomputationalmodelsveryappealinginapplicationdomainswhereonehaslittleorincompleteunderstandingoftheproblemtobesolv

8、edbutwheretrainingdataisreadilyavailable.Anotherkeyfeatureistheintrinsicparallelismthatallowsforfastcomputationsofsolutionswhenthesenetworksareimplementedonparalleldigitalcomputersor,ultimately,whenimplementedincustomizedhardware.Artificialneuralnetworksareviablecomputationalmodelsforawidevarieyofpr

9、oblems,includingpatternclassification,speechsynthesisandrecognition,adaptiveinterfacesbetweenhumansandcomplexphysicalsystems,functionapproximation,imagedatacompression,associativememory,clustering,forecastingandprediction,combinatorialoptimization,nonlinearsystemmodeling,andcontrol.Thesenetworksaren

10、euralinthesensethattheymayhavebeeninspiredbyneuroscience,butnotbecausetheyarefaithfulmodelsofbiologicneuralorcognitivephenomena.Infact,themajorityofthenetworkmodelscoveredinthisbookaremorecloselyrelatedtotraditionalmathematicaland/orstatisticalmodelssuchasoptimizationalgorithms,nonparametricpatternc

11、lassifiers,clusteringalgorithms,linearandnonlinearfilters,andstatisticalregressionmodelsthantheyaretoneurobiologicmodels.Thetheoriesandtechniquesofartificialneuralnetworksoutlinedherearefairlymathematical,althoughthelevelofmathematicalrigorisrelativelylow.InmyexpositionIhaveusedmathematicstoprovidei

12、nsightandunderstandingratherthantoestablishrigorousmathematicalfoundations.Theselectionandtreatmentofmaterialreflectmybackgroundasanelectricalandcomputerengineer.Theoperationofartificialneuralnetworksisviewedasthatofnonlinearsystems:Staticnetworksareviewedasmappingorstaticinput/outputsystems,andrecu

13、rrentnetworksareviewedasdynamicalsystemswithevolvingstate.Thesystemsapproachisalsoevidentwhenitcomestodiscussingthestabilityoflearningalgorithmsandrecurrentnetworkretrievaldynamics,aswellasintheadoptedclassificationsofneuralnetworksasdiscrete-stateorcontinuous-stateanddiscrete-timeorcontinuous-time.

14、Theneuralnetworkparadigms(architecturesandtheirassociatedlearningrules)treatedherewereselectedbecauseoftheirrelevence,mathematicaltractability,and/orpracticality.Omissionshavebeenmadeforanumberofreasons,includingcomplexity,obscurity,andspace.Thisbookisorganizedintoeightchapters.Chapter1introducesthe

15、readertothemostbasicartificialneuralnet,consistingofasinglelinearthresholdgate(LTG).Thecomputationalcapabilitiesoflinearandpolynomialthresholdgatesarederived.Afundamentaltheorem,thefunctioncountingtheorem,isprovedandisappliedtostudythecapacityandthegeneralizationcapabilityofthresholdgates.Theconcept

16、scoveredinthischapterarecrucialbecausetheylaythetheoreticalfoundationsforjustifyingandexploringthemoregeneralartificialneuralnetworkarchitecturestreatedinlaterchapters.Chapter2mainlydealswiththeoreticalfoundationsofmultivariatefunctionapproximationusingneuralnetworks.Thefunctioncountingtheoremofchap

17、ter1isemployedtoderiveupperboundsonthecapacityofvariousfeedforwardnetsofLTGs.ThenecessaryboundsonthesizeofLTG-basedmultilayerclassifiersforthecasesoftrainingdataingeneralpositionandinarbitrarypositionarederived.Theoreticalresultsoncontinuousfunctionapproximationcapabilitiesoffeedforwardnets,withunit

18、semployingvariousnonlinearities,aresummarized.Thechapterconcludeswithadiscussionofthecomputationaleffectivenessofneuralnetarchitecturesandtheefficiencyoftheirhardwareimplementations.Learningrulesforsingle-unitandsingle-ayernetsarecoveredinChapter3.Morethan20basicdiscrete-timelearningrulesarepresente

19、d.Supervisedrulesareconsideredfirst,followedbyreinforcement,Hebbian,competitive,andfeaturemappingrules.Thepresentationoftheselearningrulesisunifiedinthesensethattheymayallbeviewedasrealizingincrementalsteepest-gradient-descentsearchonasuitablecriterionfunction.Examplesofsingle-layerarchitecturesareg

20、iventoillustratetheapplicationofunsupervisedlearningrules(e.g.,principalcomponentanalysis,clustering,vectorquantization,andself-organizingfeaturemaps).Chapter4isconcernedwiththetheoreticalaspectsofsupervised,unsupervised,andreinforcementlearningrules.Thechapterstartsbydevelopingaunifyingframeworkfor

21、thecharaterizationofvariouslearningrules(supervisedandunsupervised).Underthisframework,acontinuous-timelearningruleisviewedasafirst-orderstochasticdifferentialequation/dynamicalsystemwherebythestateofthesystemevolvessoas丫彳rptominimizeanassociatedinstantaneouscriterionfunction.Statisticalapproximatio

22、ntechniquesareemployedtostudythedynamicsandstability,inanaveragesense,ofthestochasticsystem.Thisapproximationleadstoanaveragelearningequationthat,inmostcases,canbecastasaglobally,asymptoticallystablegradientsystemwhosestableequilibriaareminimizersofawell-definedcriterionfunction.Formalanalysisisprov

23、idedforsupervised,reinforcement,Hebbian,competitive,andtopologypreservinglearning.Also,thegeneralizationpropertiesofdeterministicandstochasticneuralnetsareanalyzed.Thechapterconcludeswithaninvestigatinonofthecomplexityoflearninginmultilayerneuralnets.Chapter5dealswithlearninginmultilayerartificialne

24、uralnets.Itextendsthegradientdescent-basedlearningtomultilayerfeedforwardnets,whichresultsinthebackerrorpropagationlearningrule(orbackprop).Anextensivenumberofmethodsandheuristicsforimprovingbackpropsconvergencespeedandsolutionqualityarepresented,andanattemptismadetogiveatheoreticalbasisforsuchmetho

25、dsandheuristics.Severalsignificantapplicationsofbackprop-trainedmultilayernetsaredescribed.TheseapplicationsincludeconversionofEnglishtexttospeech,mappingofhandgesturestospeech,recognitionofhandwrittenZIPcodes,continuousvehiclenavigation,andmedicaldiagnosis.Thechapteralsoextendsbackproptorecurrentne

26、tworkscapableoftemporalassociation,nonlineardynamicalsystemmodeling,andcontrol.Chapter6isconcernedwithotherimportantadaptivemultilayernetarchitectures,suchastheradialbasisfunction(RBF)netandthecerebellermodelarticulationcontroller(CMAC)net,andtheirassociatedlearningrules.Thesenetworksoftenhavesimila

27、rcomputationalcapabilitiestofeedforwardmultilayernetsofsigmoidalunits,butwiththepotentialforfasterlearning.Adaptivemulilayerunit-allocatingnetssuchashypersphericalclassifiers,restrictedCoulombenergy(RCE)net,andcascadecorrelationnetarediscussed.Thechapteralsoaddressestheissueofunsupervisedlearninginm

28、ultilayernets,anditdescribestwospecificnetworksadaptiveresonancetheory(ART)netandtheautoassociativeclusteringnetsuitableforadaptivedataclustering.Theclusteringcapabilitiesofthesenetsaredemonstratedthroughexamples,includingthedecompositionofcomplexelectromyogramsignals.Chapter7discussesassociativeneu

29、ralmemories.Variousmodelsofassociativelearningandretrievalarepresentedandanalyzed,withemphasisonrecurrentmodels.Thestability,capacity,anderror-correctioncapabilitiesofthesemodelsareanalyzed.Thechapterconcludesbydescribingtheuseofoneparticularrecurrentmodel(theHopfieldcontinuousmodel)forsolvingcombin

30、atorialoptimizationproblems.GlobalsearchmethodsforoptimallearningandretrievalinmultilayerneuralnetworksisthetopicofChapter8.Itcoverstheuseofsimulatedannealing,meanfieldannealing,andgeneticalgorithmsforoptimallearning.Simulatedannealingisalsodiscussedinthecontextoflocal-minima-freeretrievalsinrecurre

31、ntneuralnetworks(Boltzmannmachines).Finally,ahybridgeneticalgorithm/gradient-descentsearchmethodthatcombinesoptimalandfastlearningisdescribed.Eachchapterconcludeswithasetofproblemsdesignedtoallowthereadertofurtherexploretheconceptsdiscussed.Morethan200problemsofvaryingdegreesofdifficultyareprovided.

32、Theproblemscanbedividedroughlyintothreecategories.Thefirstcategoryconsistsofproblemsthatarerelativelyeasytosolve.Theseproblemsaredesignedtodirectlyreinforcethetopicsdiscussedinthebook.Thesecondcategoryofproblems,markedwithanasterisk(*),isrelativelymoredifficult.Theseproblemsnormallyinvolvemathematic

33、alderivationsandproofsandareintendedtobethoughtprovoking.Manyoftheseproblemsincludereferencetotechnicalpapersintheliteraturethatmaygivecompleteorpartialsolutions.Thissecondcategoryofproblemsisintendedmainlyforreadersinterestedinexploringadvancedtopicsforthepurposeofstimulatingoriginalresearchideas.P

34、roblemsmarkedwithadagger()representathirdcategoryofproblemsthatarenumericalinnatureandrequiretheuseofacomputer.Someoftheseproblemsareminiprogrammingprojects,whichshouldbeespeciallyusefulforstudents.Thisbookcontainsenoughmaterialforafullsemestercourseonartificialneuralnetworksatthefirst-yeargraduatel

35、evel.Ihavealsousedthismaterialselectivelytoteachanupper-levelundergraduateintroductorycourse.Fortheundergraduatecourse,onemaychoosetoskipallorasubsetofthefollowingmaterial:Sections1.4-1.6,2.12.2,4.3-4.8,5.1.2,5.4.3-5.4.5,6.1.2,6.2-6.4,6.4.2,7.2.2,7.4.1-7.4.4,8.3.2,8.4.2,and8.6.hopethatthisbookwillpr

36、oveusefultothosestudentsandpracticingprofessionalswhoareinterestednotonlyinunderstandingtheunderlyingtheoryofartificialneuralnetworksbutalsoinpursuingresearchinthisarea.Alistofabout700releventreferencesisincludedwiththeaimofprovidingguidanceanddirectionforthereadersownsearchoftheresearchliterature.E

37、venthoughthisreferencelistmayseemcomprehensive,thepublishedliteratureistooextensivetoallowsuchalisttobecomplete.AcknowledgmentsFirstandforemost,Iacknowledgethecontributionsofthemanyresearchersintheareaofartificialneuralnetworksonwhichmostofthematerialinthistextisbased.Itwouldhavebeenextremelydifficu

38、lt(ifnotimpossible)towritethisbookwithoutthesupportandassistanceofanumberoforganizationsandindividuals.IwouldfirstliketothanktheNationalScienceFoundation(throughaPYIAward),ElectricPowerResearchInstitute(EPRI),FordMotorCompany,MentorGraphics,SunMicroSystems,UnisisCorporation,WhitakerFoundation,andZen

39、ithDataSystemsforsupportingmyresearch.IamalsogratefulforthesupportIhavereceivedforthisprojectfromWayneStateUniversitythroughaCareerDevelopmentChairAward.thankmystudents,whohavemadeclassroomuseofpreliminaryversionsofthisbookandwhosequestionsandcommentshavedefinitelyenhancedit.Inparticular,Iwouldliket

40、othankRaedAbuZitar,DavidClark,MikeFinta,JingSong,AgusSudjianto,Chuanming(Chuck)Wang,HuiWang,PaulWatta,andAbbasYoussef.IalsowouldliketothankmymanycolleaguesintheartificialneuralnetworkscommunityandatWayneStateUniversity,especiallyDr.A.RobertSpitzer,formanyenjoyableandproductiveconversationsandcollabo

41、rations.amindebttoMikeFinta,whoverycapablyandenthusiasticallytypedthecompletemanuscriptandhelpedwithmostoftheartwork,andtoDr.PaulWattaoftheComputationandNeuralNetworksLaboratory,WayneStateUniversity,forhiscriticalreadingofthemanuscriptandassistancewiththesimulationsthatledtoFigures5.3.8and5.3.9.Myde

42、epgratitudegoestothereviewersfortheircriticalandconstructivesuggestions.TheyareProfessorsShun-IchiAmarioftheUniversityofTokyo,JamesAndersonofBrownUniversity,ThomasCoverofStanfordUniversity,RichardGoldenoftheUniversityofTexas-Dallas,LaveenKanaloftheUniversityofMaryland,JohnTaylorofKingsCollegeLondon,

43、FrancisT.S.YuoftheUniversityofPennsylvania,Dr.GraninoKornofG.A.andT.M.KornIndustrialConsultants,andotheranonymousreviewers.Finally,letmethankmywifeAmal,daughterLamees,andsonTarekfortheirquietpatiencethroughthemanylonelyhoursduringthepreparationofthemanuscript.MohamadH.HassounDetroit,1994ProblemsProb

44、lemsTableofContentsFundamentalsofArtificialNeuralNetworksbyMohamadH.Hassoun(MITPress,1995)Chapter1ThresholdGates1.0IntroductionThresholdGatesLinearThresholdGatesQuadraticThresholdGatesPolynomialThresholdGatesComputationalCapabilitiesofPolynomialThresholdGatesGeneralPositionandtheFunctionCountingTheo

45、remWeierstrasssApproximationTheoremPointsinGeneralPositionFunctionCountingTheoremSeparabilityinf-SpaceMinimalPTGRealizationofArbitrarySwitchingFunctionsAmbiguityandGeneralizationExtremePointsSummaryProblemsChapter2ComputationalCapabilitiesofArtificialNeuralNetworks2.0IntroductionSomePreliminaryResul

46、tsonNeuralNetworkMappingCapabilitiesNetworkRealizationofBooleanFunctionsBoundsontheNumberofFunctionsRealizablebyaFeedforwardNetworkofLTGsNecessaryLowerBoundsontheSizeofLTGNetworksTwoLayerFeedforwardNetworksThreeLayerFeedforwardNetworksGenerallyInterconnectedNetworkswithnoFeedbackApproximationCapabil

47、itiesofFeedforwardNeuralNetworksforContinuousFunctionsKolmogorovsTheoremSingleHiddenLayerNeuralNetworksareUniversalApproximatorsSingleHiddenLayerNeuralNetworksareUniversalClassifiersComputationalEffectivenessofNeuralNetworksAlgorithmicComplexityComputationalEnergySummaryChapter3LearningRules3.0Intro

48、ductionSupervisedLearninginaSingleUnitSettingErrorCorrectionRulesPerceptronLearningRuleGeneralizationsofthePerceptronLearningRuleThePerceptronCriterionFunctionMaysLearningRuleWidrow-Hoff(alpha-LMS)LearningRuleOtherGradientDescent-BasedLearningRulesmu-LMSLearningRuleThemu-LMSasaStochasticProcessCorre

49、lationLearningRuleExtensionofthemu-LMSRuletoUnitswithDifferentiableActivationFunctions:DeltaRuleAdaptiveHo-Kashyap(AHK)LearningRulesOtherCriterionFunctionsExtensionofGradientDescent-BasedLearningtoStochasticUnitsReinforcementLearningAssociativeReward-PenaltyReinforcementLearningRuleUnsupervisedLearn

50、ingHebbianLearningOjasRuleYuilleetal.RuleLinskersRuleHebbianLearninginaNetworkSetting:PrincipalComponentAnalysis(PCA)PCAinaNetworkofInteractingUnitsPCAinaSingleLayerNetworkwithAdaptiveLateralConnectionsNonlinearPCACompetitivelearningSimpleCompetitiveLearningVectorQuantizationSelf-OrganizingFeatureMa

51、ps:TopologyPreservingCompetitiveLearningKohonensSOFMExamplesofSOFMsSummaryProblemsChapter4MathematicalTheoryofNeuralLearning4.0IntroductionLearningasaSearchMechanismMathematicalTheoryofLearninginaSingleUnitSettingGeneralLearningEquationAnalysisoftheLearningEquationAnalysisofsomeBasicLearningRulesCha

52、racterizationofAdditionalLearningRulesSimpleHebbianLearningProblemsProblemsImprovedHebbianLearningOjasRuleYuilleetal.RuleHassounsRulePrincipalComponentAnalysis(PCA)TheoryofReinforcementLearningTheoryofSimpleCompetitiveLearningDeterministicAnalysisStochasticAnalysisTheoryofFeatureMappingCharacterizat

53、ionofKohonensFeatureMapSelf-OrganizingNeuralFieldsGeneralizationGeneralizationCapabilitiesofDeterministicNetworksGeneralizationinStochasticNetworksComplexityofLearningSummaryProblemsChapter5AdaptiveMultilayerNeuralNetworksI5.0IntroductionLearningRuleforMultilayerFeedforwardNeuralNetworksErrorBackpro

54、pagationLearningRuleGlobalDescent-BasedErrorBackpropagationBackpropEnhancementsandVariationsWeightsInitializationLearningRateMomentumActivationFunctionWeightDecay,WeightElimination,andUnitEliminationCross-ValidationCriterionFunctionsApplicationsNetTalkGlove-TalkHandwrittenZIPCodeRecognitionALVINN:AT

55、rainableAutonomousLandVehicleMedicalDiagnosisExpertNetImageCompressionandDimensionalityReductionExtensionsofBackpropforTemporalLearningTime-DelayNeuralNetworksBackpropagationThroughTimeRecurrentBack-PropagationTime-DependentRecurentBack-Propagation545Real-TimeRecurrentLearningSummaryChapter6Adaptive

56、MultilayerNeuralNetworksII6.0IntroductionRadialBasisFunction(RBF)NetworksRBFNetworksversusBackpropNetworksRBFNetworkVariationsCerebellerModelArticulationController(CMAC)CMACRelationtoRosenblattsPerceptronandOtherModelsUnit-AllocatingAdaptiveNetworksHypersphericalClassifiersRestrictedCoulombEnergy(RC

57、E)ClassifierReal-TimeTrainedHypersphericalClassifierCascade-CorrelationNetworkClusteringNetworks641AdaptiveResonanceTheory(ART)NetworksAutoassociativeClusteringNetworkSummaryProblemsChapter7AssociativeNeuralMemories7.0IntroductionBasicAssociativeNeuralMemoryModelsSimpleAssociativeMemoriesandtheirAss

58、ociatedRecordingRecipesCorrelationRecordingRecipeASimpleNonlinearAssociativeMemoryModelOptimalLinearAssociativeMemory(OLAM)OLAMErrorCorrectionCapabilitiesStrategiesforImprovingMemoryRecordingDynamicAssociativeMemories(DAM)Continuous-TimeContinuous-StateModelDiscrete-TimeContinuous-StateModelDiscrete

59、-TimeDiscrete-StateModelDAMCapacityandRetrievalDyanamicsCorrelationDAMsProjectionDAMsCharacteristicsofHigh-PerformanceDAMsOtherDAMModelsBrain-State-in-a-Box(BSB)DAMNon-MonotonicActivationsDAMDiscreteModelContinuousModelHystereticActivationsDAMExponentialCapacityDAMSequenceGeneratorDAMHeteroassociati

60、veDAMTheDAMasaGradientNetanditsApplicationtoCombinatorialOptimizationSummaryChapter8GlobalSearchMethodsforNeuralNetworks1.8.0IntroductionLocalversusGlobalSearchAGradientDescent/AscentSearchStrategyStochasticGradientSearch:GlobalSearchviaDiffusionSimulatedAnnealing-BasedGlobalSearchSimulatedAnnealing

温馨提示

  • 1. 本站所有资源如无特殊说明,都需要本地电脑安装OFFICE2007和PDF阅读器。图纸软件为CAD,CAXA,PROE,UG,SolidWorks等.压缩文件请下载最新的WinRAR软件解压。
  • 2. 本站的文档不包含任何第三方提供的附件图纸等,如果需要附件,请联系上传者。文件的所有权益归上传用户所有。
  • 3. 本站RAR压缩包中若带图纸,网页内容里面会有图纸预览,若没有图纸预览就没有图纸。
  • 4. 未经权益所有人同意不得将文件中的内容挪作商业或盈利用途。
  • 5. 人人文库网仅提供信息存储空间,仅对用户上传内容的表现方式做保护处理,对用户上传分享的文档内容本身不做任何修改或编辑,并不能对任何下载内容负责。
  • 6. 下载文件中如有侵权或不适当内容,请与我们联系,我们立即纠正。
  • 7. 本站不保证下载资源的准确性、安全性和完整性, 同时也不承担用户因使用这些下载资源对自己和他人造成任何形式的伤害或损失。

最新文档

评论

0/150

提交评论