中文翻译.docx

用更快的R-CNN进行人脸表情识别【中文3606字】

收藏

压缩包内文档预览:
预览图 预览图 预览图 预览图 预览图 预览图 预览图 预览图
编号:9931492    类型:共享资源    大小:1.86MB    格式:ZIP    上传时间:2018-04-05 上传人:闰*** IP属地:河南
15
积分
关 键 词:
用更快 cnn 进行 表情 识别 辨认 中文
资源描述:
用更快的R-CNN进行人脸表情识别【中文3606字】,用更快,cnn,进行,表情,识别,辨认,中文
内容简介:
AVAILABLEONLINEATWWWSCIENCEDIRECTCOMSCIENCEDIRECTPROCEDIACOMPUTERSCIENCE1072017135140INTERNATIONALCONGRESSOFINFORMATIONANDCOMMUNICATIONTECHNOLOGYICICT2017FACIALEXPRESSIONRECOGNITIONWITHFASTERRCNNJIAXINGLIA,DEXIANGZHANGA,JINGJINGZHANGA,JUNZHANGA,TENGLIA,YIXIAA,QINGYANA,ANDLINAXUNAATHESCHOOLOFELECTRICALENGINEERINGANDAUTOMATIONANHUIUNIVERSITY,HEFEI230601,CHINACORRESPONDINGAUTHORZDXDZXY126COMTEL086055163861094ABSTRACTINORDERTOAVOIDTHECOMPLEXEXPLICITFEATUREEXTRACTIONPROCESSANDTHEPROBLEMOFLOWLEVELDATAOPERATIONINVOLVEDINTRADITIONALFACIALEXPRESSIONRECOGNITION,WEPROPOSEDAMETHODOFFASTERRCNNFASTERREGIONSWITHCONVOLUTIONALNEURALNETWORKFEATURESFORFACIALEXPRESSIONRECOGNITIONINTHISPAPERFIRSTLY,THEFACIALEXPRESSIONIMAGEISNORMALIZEDANDTHEIMPLICITFEATURESAREEXTRACTEDBYUSINGTHETRAINABLECONVOLUTIONKERNELTHEN,THEMAXIMUMPOOLINGISUSEDTOREDUCETHEDIMENSIONSOFTHEEXTRACTEDIMPLICITFEATURESAFTERTHAT,RPNSREGIONPROPOSALNETWORKSISUSEDTOGENERATEHIGHQUALITYREGIONPROPOSALS,WHICHAREUSEDBYFASTERRCNNFORDETECTIONFINALLY,THESOFTMAXCLASSIFIERANDREGRESSIONLAYERISUSEDTOCLASSIFYTHEFACIALEXPRESSIONSANDPREDICTBOUNDARYBOXOFTHETESTSAMPLE,RESPECTIVELYTHEDATASETISPROVIDEDBYCHINESELINGUISTICDATACONSORTIUMCLDC,WHICHISCOMPOSEDOFMULTIMODALEMOTIONALAUDIOANDVIDEODATAEXPERIMENTALRESULTSSHOWTHEPERFORMANCEANDTHEGENERALIZATIONABILITYOFTHEFASTERRCNNFORFACIALEXPRESSIONRECOGNITIONTHEVALUEOFTHEMAPISAROUND082KEYWORDSFACIALEXPRESSIONRECOGNITIONFASTERRCNNDEEPLEARNINGGRAPHICSPROCESSINGUNIT1INTRODUCTIONFACIALEXPRESSIONISAKINDOFEFFECTIVEWAYOFHUMANCOMMUNICATIONFACIALEXPRESSIONRECOGNITIONISTHEKEYTECHNOLOGYFORREALIZINGHUMANCOMPUTERINTERACTIONTOBETHEEMOTIONALCOMPUTINGSYSTEMFACIALEXPRESSIONHASABROADAPPLICATIONPROSPECTINMANYRESEARCHFIELDS,SUCHASVIRTUALREALITY,VIDEOCONFERENCE,CUSTOMERSATISFACTIONSURVEYANDOTHERFIELDSDESPITEOFGREATENCOURAGINGPROGRESSHAVEBEENMADEINTHISRESEARCHFIELD,THEREARESTILLMANYPROBLEMSEXISTINGONONEHAND,THETRADITIONALFEATUREEXTRACTIONMETHODSARECOMPLETELYRELIEDONHUMANEXPERIENCE,WHICHISSTILLTOOCOMPLICATEDFORREALAPPLICATIONTHEREFORE,THETRADITIONALMETHODSAREVERYDIFFICULTTOEXTRACTUSEFULFEATURESCOMPREHENSIVELYANDEFFECTIVELYONTHEOTHERHAND,THETRADITIONALMETHODSCANNOTDISPOSETHEBIGDATAANDACHIEVEBETTERPERFORMANCESOITISNOTEASYTOMEETTHEREALAPPLICATIONREQUIREMENTINMOSTSITUATIONS,THISKINDOFMETHODCANNOTBEEMPLOYEDEFFECTIVELY187705092017THEAUTHORSPUBLISHEDBYELSEVIERBVTHISISANOPENACCESSARTICLEUNDERTHECCBYNCNDLICENSEHTTP/CREATIVECOMMONSORG/LICENSES/BYNCND/40/PEERREVIEWUNDERRESPONSIBILITYOFTHESCIENTIFICCOMMITTEEOFTHE7THINTERNATIONALCONGRESSOFINFORMATIONANDCOMMUNICATIONTECHNOLOGYDOI101016/JPROCS201703069136JIAXINGLIETAL/PROCEDIACOMPUTERSCIENCE1072017135140TOADDRESSTHEABOVEISSUES,WEPROPOSEANENDTOENDRECOGNITIONMETHODBASEDONFASTERRCNN1THEPROPOSEDMETHODCANBEUSEDTOSOLVETHEEXISTINGPROBLEMSFIRST,THEREGIONPROPOSALNETWORKSRPNS1AREUSEDTOPREDICTEFFICIENTANDACCURATEREGIONPROPOSAL4THEPIPELINEOFTHEPROPOSEDMETHODJUSTUSEONECONVOLUTIONALNEURALNETWORKCNNFORALLPURPOSESO,THEREGIONPROPOSALISNEARLYCOSTFREEBYSHARINGCONVOLUTIONALFEATURESOFTHEDOWNFLOWDETECTIONNETWORKSECOND,THERPNALSOIMPROVEDTHEREGIONALPROPOSAL4QUALITYANDTHEACCURACYOFTHEOVERALLTARGETDETECTION2FASTERRCNNALGORITHMFASTERRCNN1CANBESIMPLYREGARDEDASTHESYSTEMCONSISTINGOFREGIONALPROPOSALNETWORKANDFASTREGIONSWITHCONVOLUTIONALNEURALNETWORKFEATURESFASTRCNNTHEREGIONALPROPOSALNETWORKISUSEDTOINSTEADSELECTIVESEARCHALGORITHM4OFFASTRCNNTHEPROPOSEDMETHODFOCUSESONSOLVINGTHREEPROBLEMS1HOWTODESIGNAREGIONALPROPOSALNETWORK2HOWTOMAKEPROPOSALNETWORKREGION3HOWTOSHAREFEATUREEXTRACTIONNETWORK21CANDIDATEREGIONANCHORSCHARACTERISTICSCANBESEENASA256CHANNELIMAGEWITHASCALEOF5139,FOREACHPOSITIONOFTHEIMAGETHEMETHODCONSIDERSTHENINEPOSSIBLECANDIDATEWINDOWS,WHICHARETHREEAREASOF128,256,512MULTIPLIEDBYTHREERATIOSOF11,12,21THESECANDIDATEWINDOWSARESAIDTOBEAS“ANCHORS“FIG1SHOWSTHEANCHOR5139CENTER,ASWELLAS9ANCHOREXAMPLESFIG1THE5139ANCHORCENTERSASWELLAS9ANCHOREXAMPLESEACHPOSITIONOFTHEOUTPUTOFCLASSIFICATIONLAYER6CLS_SCORESHOWSTHEPROBABILITYTHATTHE9ANCHORSBELONGTOTHEFOREGROUNDANDBACKGROUNDANDEACHPOSITIONOFTHEOUTPUTOFREGRESSIONLAYER6BBOX_PREDSHOWSTHATTHECORRESPONDINGWINDOWOFTHE9ANCHORSSHOULDBETRANSLATEDTOSCALEPARAMETERSFOREACHLOCATION,THECLASSIFICATIONLAYEROUTPUTSTHEPROBABILITYOFTHEFOREGROUNDANDTHEBACKGROUNDFROMTHE256DIMENSION,WHILETHEREGRESSIONLAYEROUTPUTS4TRANSLATIONSCALINGPARAMETERSFROMALOCALRESPECTIVE,THETWOLAYERSARETHEWHOLECONNECTIONNETWORK,WHILEFROMAGLOBALRESPECTIVE,SINCETHENETWORKSHARESAMEPARAMETERSINALLPOSITIONS5139,THENETWORKUSEDINTHEPRESENTSTUDYISACTUALLYWITHASIZEOF1122SHARINGFEATUREREGIONPROPOSALNETWORKRPNANDFASTRCNNREQUIREANORIGINALFEATUREEXTRACTIONNETWORKTHEIMAGENETCLASSIFICATIONLIBRARYISUSEDTOTRAINTHEINITIALPARAMETER0OFTHENETWORKTHENTHENETWORKISFINETUNEDBYTHESPECIFIEDDATASETTHEPROPOSEDMETHODPROVIDESTHREEMETHODS1TRAINRPNTOEXTRACTTHEANCHORSONTHETRAININGSETFROM02TRAINFASTRCNNBYUSINGTHEANCHORSFROM0,ANDTHEPARAMETERISDENOTEDAS13TRAINRPNFROM1FIG2SHOWSTHEDETAILSTEPSOFTHESHAREDFEATUREJIAXINGLIETAL/PROCEDIACOMPUTERSCIENCE1072017135140137INPUTTRAININGIMAGECONVFEATUREMAPROIPIFIG2THESTEPSOFTHESHAREDFEATUREWITHTHESEDEFINITIONS,WEMINIMIZEANOBJECTIVEFUNCTIONFOLLOWINGTHEMULTITASKLOSSINFASTRCNN2FORANANCHORBOXI,ITSLOSSFUNCTIONISDEFINEDASLP,TLP,PPLT,T1IICLSIIIREGIIHEREPIISTHEANCHORIBEINGANOBJECT1OFTHEPROBABILITYOFPREDICTIONIFTHEANCHORLABELISPOSITIVE,IS1,IFITISNEGATIVE,ITIS0TITX,TY,T,THIREPRESENTSTHEPREDICTEDBOUNDINGBOXOF4PARAMETERIZEDCOORDINATES6,ANDTT,T,T,TINDICATESTHATTHEGROUNDTRUTHBOX6ISASSOCIATEDWITHAPOSITIVEIXYHIANCHORTHECLASSIFICATIONLOSSLREGISTHESOFTMAXLOSSOFTWOCLASSES5THEDETAILSOFOURFACIALEXPRESSIONRECOGNITIONMETHODISILLUSTRATEDINFIG3DEEPROIPOOLINGLAYERRPNSFULLCONNECTIONNETWORKSOFTMAXREGRESSOROUTPUTSFIG3FLOWCHARTOFOURFACIALEXPRESSIONRECOGNITIONMETHOD3EXPERIMENTS31DATASETANDFEATURESTHEDATASETISPROVIDEDBYCHINESELINGUISTICDATACONSORTIUMCLDC9,WHICHISCOMPOSEDOFMULTIMODALEMOTIONALAUDIOANDVIDEODATATOTALEIGHTEXPRESSIONARECOLLECTEDFROMTVSERIESORMOVIESSOMEEXAMPLESARESHOWNINFIG4138JIAXINGLIETAL/PROCEDIACOMPUTERSCIENCE1072017135140FIG4THEPICTUREOFTHEDATAFROMCLDCFROMLEFTTORIGHTWORRIED,ANGRY,DISGUST,SURPRISE,ANXIOUS,HAPPY,SADANDNEURALINTHEEXPERIMENT,THEDATASETCONSISTSOF66486PICTURESWITHEIGHTCATEGORIESAMONGTHEMARE6174WORRIEDPICTURES,10862ANGRYPICTURES,1687DISGUSTPICTURES,2574SURPRISEPICTURES,12019ANXIOUSPICTURES,9867HAPPYPICTURES,18326SADPICTURESAND4977NEURALPICTURESTHERATIOOFTRAINING,VALIDATIONANDTESTINGDATAIS811SINCEFASTERRCNNWASUSEDTODETECTFACEDIRECTLY,THEBACKGROUNDWASCONSIDEREDASONECLASSSO,TOTAL9CATEGORIESWEREUSEDINTHISRESEARCH32DATALABELMAKINGSINCEFASTERRCNNISUSED,THEREGIONOFINTERESTROIOFEACHIMAGEMUSTBEMARKEDFIRSTTHESOFTWARECANBEEMPLOYEDTOACHIEVETHECOORDINATESOFROITHENTHECOORDINATESARETRANSFERINTOXMLFORMATTHEROIOFSOMEIMAGESISSHOWNINFIG5FIG5DATALABELMAKING33MODELDESIGNWEUSETHETHREEPASCAL_VOCMODELPROVIDEDBYTHEPYFASTERRCNN,RESPECTIVELY,VGG_CNN_M_10248,ZF7ANDVGG163,WHICHARECALLEDTHEFASTER_RCNN_ALT_OPTNETWORKTOFINETURNIMAGENETMODELTHEDEPTHOFTHETHREENETWORKSISINCREASINGSOWECANCOMPARETHEEXPERIMENTALRESULTSTOFURTHERANALYZETHEACCURACYOFTHEEXPERIMENTTHROUGHTHETHREENETWORKS34PARAMETERSETTINGTHESOLVERPARAMETERSOFTHENETWORKARESETASFOLLOWSSTAGE1_FAST_RCNN_TRAINPTBASE_LRFASTER0001,LR_POLICYFASTER“STEP“,STEPSIZEFASTER30000,DISPLAYFASTER20,AVERAGE_LOSSFASTER100,MOMENTUMJIAXINGLIETAL/PROCEDIACOMPUTERSCIENCE1072017135140139FASTER09,WEIGHT_DECAYFASTER00005STAGE1_RPN_TRAINPTBASE_LRFASTER0001,LR_POLICYFASTER“STEP“,STEPSIZEFASTER60000,DISPLAYFASTER20,AVERAGE_LOSSFASTER100,MOMENTUMFASTER09,WEIGHT_DECAYFASTER00005STAGE2_FAST_RCNN_TRAINPTBASE_LRFASTER0001,LR_POLICYFASTER“STEP“,STEPSIZEFASTER30000,DISPLAYFASTER20,AVERAGE_LOSSFASTER100,MOMENTUMFASTER09,WEIGHT_DECAYFASTER00005STAGE2_RPN_TRAINPTBASE_LRFASTER0001,LR_POLICYFASTER“STEP“,STEPSIZEFASTER60000,DISPLAYFASTER20,AVERAGE_LOSSFASTER100,MOMENTUMFASTER09,WEIGHT_DECAYFASTER00005THEPARTIALPARAMETERSOFTHENETWORKARESETASFOLLOWSDATA_PARAM_STR_NUM_CLASSES9,CLS_SCORE_NUM_OUTPUT9,BBOX_PRED_NUM_OUTPUT3635TRAININGANDTESTINGWETREATOVER08IOUOVERLAPFORALLREGIONPROPOSALSWITHAGROUNDTRUTHBOXASPOSITIVESANDTHERESTASNEGATIVESFORBOXSCLASSWEBIASTHESAMPLINGTOWARDSPOSITIVEWINDOWSBECAUSETHEYAREEXTREMELYRARECOMPAREDTOBACKGROUND4EXPERIMENTALRESULTINORDERTOIMPROVETHERELIABILITYOFTHEEXPERIMENTALRESULTS,THREEDIFFERENTDEPTHOFTHENETWORKISUSEDTOTRAINANDTESTDATA,THATIS,VGG_CNN_M_10248,ZF7ANDVGG163THETRAININGRESULTSOFTHETHREEKINDSOFNETWORKSARESHOWNINTABLE1ASWECANSEEFROMTABLE1,WITHTHEINCREASEOFTHEDEPTHOFTHENETWORK,THEMAPVALUEOFTHETRAININGDETECTIONISIMPROVEDITCANBECLEARTOFINDTHATSOMETYPESWITHLARGEAMOUNTDATA,SUCHASTHETYPEOFANXIOUS,THERECOGNITIONRATEISNOTVERYHIGHONTHECONTRARY,SOMETYPESWITHSMALLTRAININGDATA,SUCHASTHETYPEOFDISGUST,BUTTHERECOGNITIONRATEISRELATIVELYHIGHTHATSBECAUSEITISNOTBALANCEDFORTHETRAININGDATAOFEACHTYPE,ITLEADSTOTHEMUTUALINTERFERENCEOFSOMEKINDSOFDATAINTRAININGBUTIT,THISPROBLEM,HASNOTMUCHIMPACTONTHERESULTSOFTHETRAININGDATA,WECANSTILLOBTAINTHEFINALRECOGNITIONRATEANDRECOGNITIONRESULTSINTHEFUTURE,WEWILLMAKETHEAPPROPRIATEJUDGMENTANDCORRECTIONFORTHEIMBALANCEOFTHEDATA,SOTHATTHEPROBLEMWILLNOTAFFECTTHERESULTSOFTHEEXPERIMENT,SOASTOIMPROVETHEREADABILITYANDACCURACYOFTHEDATAFROMTHETABLE1,THERECOGNITIONRATEOFTHENEURALTYPEISVERYLOWMAKETHEMAPVALUEISNOTVERYHIGHFORALLTESTDATATHENEURALEXPRESSIONISVERYHARDTOIDENTIFYBECAUSETHEREARETOOMANYHUMANFACTORSTODETERMINENEURALTYPEWHENLABELIMAGES,SOMEOTHERCLASSEXPRESSIONAREVERYEASILYDETERMINEDTOBEANEURALTYPEINTHEFUTURE,WECANIMPROVETHERECOGNITIONRATEBYINCREASINGTHETRAININGDATAOFTHENEURALTYPETABLE1THETRAININGRESULTSOFTHETHREEKINDSOFNETWORKSPASCAL_VOCMODELWORRIEDANGRYDISGUSTSURPRISEANXIOUSHAPPYSADNEURALMAPVGG_CNN_MSOMEEXAMPLEDETECTIONSUSINGFASTERRCNNONCLDCARESHOWNINFIG6WECANSEEFROMTHEFIG5,ITISVERYGOODFORRECOGNITIONEFFECTOF6KINDSOFTYPESWORRIED,ANGRY,DISGUST,ANXIOUS,HAPPYANDSAD,ANDSOMERECOGNITIONRATEEVENREACHED100,SUCHASTHETYPESOFDISGUSTANDANGRYITCANBESEENTHATTHEEXPERIMENTALRESULTSHAVEREACHEDTHEEXPECTEDGOALHOWEVER,THEREARENOTVERYGOODIDENTIFICATIONFORTHETYPESOFSURPRISEANDNEURALFIRST,THETYPEOFSURPRISEHAVEMANYSIMILARITIESWITHTHETYPESOFANGRYANDHAPPY,WHICHDOESNOTHAVETHEVERYGOODRECOGNITIONFEATURESOTHETYPEOFSURPRISEAREIDENTIFIEDASANGRYORHAPPYINTESTINGDATATHETYPEOFNEURALISALSODUETOTHEHUMANELEMENTWHENITISUSEDASALABELINLATEREXPERIMENTS,WECANINCREASETHEWEIGHTOFTHEEYESANDMOUTHFROMTHEIMAGETORECOGNIZETHEFEATUREOFIMAGESACCURATELY,BECAUSEITISMAINLYTHROUGHTHEBEHAVIORCHARACTERISTICSOFTHE_1024079220899809798088790880608513090780360408200ZF080150899809739089790879908666090830334408203VGG16083150901609782089730885708812090910364608312140JIAXINGLIETAL/PROCEDIACOMPUTERSCIENCE1072017135140EYESANDMOUTHFORTHEFACIALEXPRESSIONRECOGNITIONATTHESAMETIME,ITCANINCREASETHETRAININGDATAOFEACHTYPEANDENHANCETHERECOGNITIONOFTHEDATATOINCREASETHERECOGNITIONRATEOFTHETYPEFIG6EXAMPLEDETECTIONSUSINGFASTERRCNNONCLDC5CONCLUSIONINTHISPAPER,FASTERRCNNWASUSEDTOIDENTIFYFACIALEXPRESSIONTHEREARETHEFOLLOWINGADVANTAGESUSEDINFACIALEXPRESSIONRECOGNITIONTHEORIGINALIMAGEWASUSEDASTHEWHOLENETWORKINPUTTHEPROCESSOFFEATUREEXTRACTIONINTHETRADITIONALFACIALEXPRESSIONRECOGNITIONISAVOIDEDTHEFEATURESAREEXTRACTEDBYNETWORKFROMTRAININGDATASETAUTOMATICALLYTHEREGIONPROPOSALNETWORKSRPNSWASUSEDTOGENERATEAEFFICIENTANDAACCURATEREGIONPROPOSALINEACHIMAGE,THEPROPOSEDMETHODLOCATETHEFACEREGIONANDRECOGNIZETHEEXPRESSIONDIRECTLYTHEEXPERIMENTALRESULTSSHOWTHATTHEPROPOSEDMETHODACHIEVEDBETTERRECOGNITIONPERFORMANCEACKNOWLEDGEMENTSTHISWORKWASFINANCIALLYSUPPORTEDBYTHECHINESENATIONALSCIENCEFOUNDATIONGRANTNO61272025,NO61402004,NO61602002,NO61300056,NO61572029ANDNO61271098,ANDISSUPPORTEDBYANHUIPROVINCIALNATURALSCIENCEFOUNDATIONNO1608085MF136ANDNO1408085QF118THISPAPERISPARTIALLYSUPPORTEDBYSCIENCEANDTECHNOLOGYPROJECTOFANHUIPROVINCENO1501B042207ANDNO1604D0802019REFERENCES1RENS,HEK,GIRSHICKR,ETALFASTERRCNNTOWARDSREALTIMEOBJECTDETECTIONWITHREGIONPROPOSALNETWORKSJCOMPUTERSCIENCE,20152RGIRSHICKFASTRCNNARXIV150408083,20153KSIMONYANANDAZISSERMANVERYDEEPCONVOLUTIONALNETWORKSFORLARGESCALEIMAGERECOGNITIONINICLR,20154JRUIJLINGS,KEVANDESANDE,TGEVERS,ANDAWSMEULDERSSELECTIVESEARCHFOROBJECTRECOGNITIONIJCV,20135RGIRSHICK,JDONAHUE,TDARRELL,ANDJMALIKRICHFEATUREHIERARCHIESFORACCURATEOBJECTDETECTIONANDSEMANTICSEGMENTATIONINCVPR,20146CLZITNICKANDPDOLLREDGEBOXESLOCATINGOBJECTPROPOSALSFROMEDGESINECCV,20147MDZEILERANDRFERGUSVISUALIZINGANDUNDERSTANDINGCONVOLUTIONALNEURALNETWORKSINECCV,20148KSIMONYANANDAZISSERMANVERYDEEPCONVOLUTIONALNETWORKSFORLARGESCALEIMAGERECOGNITIONINICLR,20159CHINESELINGUISTICDATACONSORTIUM,CHINESECONFERENCEONPATTERNRECOGNITIONCCPR,2016【中文3606字】国际信息和通信技术大会(ICICT2017)用更快的RCNN进行人脸表情识别JIAXINGLIA,DEXIANGZHANGA,JINGJINGZHANGA,JUNZHANGA,TENGLIA,YIXIAA,QINGYANA,ANDLINAXUNAA安徽大学电气与自动化工程学院,安徽合肥230601通讯作者ZDXDZXY126COM电话086055163861094摘要为了避免复杂的显式特征提取过程和传统面部表情识别所涉及的低层次数据操作问题,提出了一种快速RCNN(具有卷积神经网络特征的快速区域)的人脸表情识别方法纸。首先对表情图像进行归一化处理,利用可训练卷积核提取隐含特征。然后,使用最大池来减少提取的隐式特征的维度。之后,RPN(地区提案网络)用于生成高质量地区提案,RASTERCN用于检测。最后,SOFTMAX分类器和回归图层分别用于对面部表情进行分类并预测测试样本的边界框。该数据集由中国语言数据联盟(CLDC)提供,由多模式情感音频和视频数据组成。实验结果证明了RCNN用于人脸表情识别的性能和泛化能力。MAP的价值在082附近。关键词面部表情识别更快的RCNN深入学习图形处理单元1简介面部表情是一种有效的人际沟通方式。面部表情识别是实现人机交互成为情感计算系统的关键技术。面部表情在虚拟现实,视频会议,客户满意度调查等多个研究领域具有广阔的应用前景。尽管该研究领域取得了令人鼓舞的进展,但仍存在很多问题。一方面,传统的特征提取方法完全依赖于人类的经验,对于实际应用来说仍然过于复杂。因此,传统方法难以全面,有效地提取有用的特征。另一方面,传统的方法不能处理大数据,取得更好的性能。所以要满足实际的应用要求并不容易。在大多数情况下,这种方法不能有效地使用。为解决上述问题,我们提出了一种基于更快的RCNN的端到端识别方法1。所提出的方法可以用来解决存在的问题。首先,区域提议网络(RPN)1用于预测有效和准确的区域提案4。该方法的流水线只用一个卷积神经网络(CNN)用于各种目的。因此,通过共享下行流检测网络的卷积特性,该区域提案几乎无成本。其次,RPN还改善了区域提案4质量和整体目标检测的准确性。2更快的RCNN算法更快的RCNN1可以简单地认为是由区域提议网络和带有卷积神经网络特征的快速区域(FASTRCNN)组成的系统。区域提案网络被用来代替FASTRCNN的选择性搜索算法4。所提出的方法着重解决三个问题1)如何设计区域提案网络2)如何提出建议网络区域3)如何共享特征提取网络。21候选地区(锚)。对于图像的每个位置,可以将特征视为具有5139比例的256通道图像。该方法考虑了九个可能的候选窗口,这三个窗口是128,256,512的三个区域乘以11,12,21三个比率。这些候选窗口被称为“锚”。图1显示了锚5139中心以及9个锚示例。图15139个锚中心以及9个锚示例分类层6(CLS_SCORE)输出的每个位置显示9个锚点属于前景和背景的概率。并且回归层6(BBOX_PRED)输出的每个位置都表明9个锚点的相应窗口应该被转换为尺度参数。对于每个位置,分类层输出来自256维的前景和背景的概率,而回归层输出4个翻译缩放参数。从本地各自来看,这两层是整个连接网络,而从全球各自来看,由于网络在所有位置共享相同的参数(5139),本研究中使用的网络实际上大小为11。22共享功能区域提案网络(RPN)和快速RCNN需要原始特征提取网络。IMAGENET分类库用于训练网络的初始参数。然后通过指定的数据集对网络进行微调。所0提出的方法提供了三种方法)训练RPN以从提取训练集上的锚点。10使用的锚点训练FASTRCNN,参数记为3)从开始训练RPN。图2显2011示了共享功能的详细步骤。图2共享功能的步骤通过这些定义,我们将FASTRCNN中的多任务丢失后的目标函数最小化2。对于锚箱I,其损失函数定义为这里,我是作为预测概率的对象1的锚点。如果锚标签为正,则为1,IP如果为负,则为0,I表示4个参数化坐标6的预测包围盒,I表示地面实况框6与正锚点相关联。分类损失是两类的SOFTMAX损失5。我们的面部表情识别方法的细节如图3所示。REGL图3我们的面部表情识别方法的流程图3实验31数据集和功能图4来自CLDC的数据图(从左至右担心,愤怒,厌恶,惊奇,焦虑,快乐,悲伤和神经)该数据集由中国语言数据联盟(CLDC)提供9,由多模式情感音频和视频数据组成。总共八个表情是从电视剧或电影中收集的。一些例子如图1所示。4。在实验中,数据集由66486个图片组成。其中有6174张担忧图片,10862张生气图片,1687张反感图片,2574张惊喜图片,12019张急性图片,9867张幸福图片,18326张悲伤图片和4977张神经图片。培训,验证和测试数据的比例是811。由于更快的RCNN被用来直接检测脸部,因此背景被认为是一类。因此,本研究共使用了9个类别。32数据标签制作由于使用更快的RCNN,因此必须首先标记每个图像的感兴趣区域(ROI)。该软件可用于实现ROI的坐标。然后坐标转换成XML格式。一些图像的ROI如图5所示。图5数据标签制作33模型设计我们分别使用由PYFASTERRCNN提供的三个PASCAL_VOCMODEL,VGG_CNN_M_10248,ZF7和VGG163,它们分别称为FASTER_RCNN_ALT_OPT网络,用于细化IMAGENET模型。三个网络的深度在增加。因此,我们可以通过三个网络比较实验结果来进一步分析实验的准确性。34参数设置网络的解算参数设置如下STAGE1_FAST_RCNN_TRAINPTBASE_LR(更快)0001,LR_POLICY(更快)“STEP”,STEPSIZE(更快)30000,DISPLAY(更快)20,AVERAGE_LOSS(更快)100,动量(更快)09,WEIGHT_DECAY(更快)00005。STAGE1_RPN_TRAINPTBASE_LR(更快)0001,LR_POLICY(更快)“STEP”,STEPSIZE(更快)60000,显示(更快)20,AVERAGE_LOSS(更快)100,动量(更快)09,WEIGHT_DECAY更快)00005。STAGE2_FAST_RCNN_TRAINPTBASE_LR(更快)0001,LR_POLICY(更快)“STEP”,STEPSIZE(更快)30000,显示(更快)20,AVERAGE_LOSS(更快)100,动量(更快)09,WEIGHT_DECAY更快)00005。STAGE2_RPN_TRAINPTBASE_LR(更快)0001,LR_POLICY(更快)“STEP”,STEPSIZE(更快)60000,显示(更快)20,AVERAGE_LOSS(更快)100,动量(更快)09,WEIGHT_DECAY更快)00005。网络的部分参数设置如下DATA_PARAM_STR_NUM_CLASSES9,CLS_SCORE_NUM_OUTPUT9,BBOX_PRED_NUM_OUTPUT36。35培训和测试我们将所有区域提案的08IOU重叠视为正片,其余视为片盒类的负片。我们偏向正窗口采样,因为它们与背景相比极其罕见。4实验结果为了提高实验结果的可靠性,使用三种不同深度的网络来训练和测试数据,即VGG_CNN_M_10248,
温馨提示:
1: 本站所有资源如无特殊说明,都需要本地电脑安装OFFICE2007和PDF阅读器。图纸软件为CAD,CAXA,PROE,UG,SolidWorks等.压缩文件请下载最新的WinRAR软件解压。
2: 本站的文档不包含任何第三方提供的附件图纸等,如果需要附件,请联系上传者。文件的所有权益归上传用户所有。
3.本站RAR压缩包中若带图纸,网页内容里面会有图纸预览,若没有图纸预览就没有图纸。
4. 未经权益所有人同意不得将文件中的内容挪作商业或盈利用途。
5. 人人文库网仅提供信息存储空间,仅对用户上传内容的表现方式做保护处理,对用户上传分享的文档内容本身不做任何修改或编辑,并不能对任何下载内容负责。
6. 下载文件中如有侵权或不适当内容,请与我们联系,我们立即纠正。
7. 本站不保证下载资源的准确性、安全性和完整性, 同时也不承担用户因使用这些下载资源对自己和他人造成任何形式的伤害或损失。
提示  人人文库网所有资源均是用户自行上传分享,仅供网友学习交流,未经上传用户书面授权,请勿作他用。
关于本文
本文标题:用更快的R-CNN进行人脸表情识别【中文3606字】
链接地址:https://www.renrendoc.com/p-9931492.html

官方联系方式

2:不支持迅雷下载,请使用浏览器下载   
3:不支持QQ浏览器下载,请用其他浏览器   
4:下载后的文档和图纸-无水印   
5:文档经过压缩,下载后原文更清晰   
关于我们 - 网站声明 - 网站地图 - 资源地图 - 友情链接 - 网站客服 - 联系我们

网站客服QQ:2881952447     

copyright@ 2020-2024  renrendoc.com 人人文库版权所有   联系电话:400-852-1180

备案号:蜀ICP备2022000484号-2       经营许可证: 川B2-20220663       公网安备川公网安备: 51019002004831号

本站为文档C2C交易模式,即用户上传的文档直接被用户下载,本站只是中间服务平台,本站所有文档下载所得的收益归上传人(含作者)所有。人人文库网仅提供信息存储空间,仅对用户上传内容的表现方式做保护处理,对上载内容本身不做任何修改或编辑。若文档所含内容侵犯了您的版权或隐私,请立即通知人人文库网,我们立即给予删除!