combininggoogle

google中国地图  时间:2021-02-05  阅读:()
Zhuetal.
BMCBioinformatics2019,20(Suppl18):575https://doi.
org/10.
1186/s12859-019-3131-8RESEARCHOpenAccessAttention-basedrecurrentneuralnetworkforinfluenzaepidemicpredictionXiangleiZhu1,BofengFu1,YaodongYang1,YuMa3,JianyeHao1*,SiqiChen1,ShuangLiu1,TiegangLi3,SenLiu2,WeimingGuo2andZhenyuLiao4,5FromBiologicalOntologiesandKnowledgebasesworkshopatIEEEBIBM2018Madrid,Spain.
3-6December2018AbstractBackground:Influenzaisaninfectiousrespiratorydiseasethatcancauseseriouspublichealthhazard.
Duetoitshugethreattothesociety,precisereal-timeforecastingofinfluenzaoutbreaksisofgreatvaluetoourpublic.
Results:Inthispaper,weproposeanewdeepneuralnetworkstructurethatforecastsareal-timeinfluenza-likeillnessrate(ILI%)inGuangzhou,China.
Longshort-termmemory(LSTM)neuralnetworksisappliedtopreciselyforecastaccuratenessduetothelong-termattributeanddiversityofinfluenzaepidemicdata.
Wedeviseamulti-channelLSTMneuralnetworkthatcandrawmultipleinformationfromdifferenttypesofinputs.
Wealsoaddattentionmechanismtoimproveforecastingaccuracy.
Byusingthisstructure,weareabletodealwithrelationshipsbetweenmultipleinputsmoreappropriately.
Ourmodelfullyconsidertheinformationinthedataset,targetedlysolvingpracticalproblemsoftheGuangzhouinfluenzaepidemicforecasting.
Conclusion:Weassesstheperformanceofourmodelbycomparingitwithdifferentneuralnetworkstructuresandotherstate-of-the-artmethods.
Theexperimentalresultsindicatethatourmodelhasstrongcompetitivenessandcanprovideeffectivereal-timeinfluenzaepidemicforecasting.
Keywords:Influenzaepidemicprediction,Attentionmechanism,Multi-channelLSTMneuralnetworkBackgroundInfluenzaisaninfectiousrespiratorydiseasethatcancauseseriouspublichealthhazard.
Itcanaggravatetheoriginalunderlyingdiseaseafterinfection,causingsec-ondarybacterialpneumoniaandacuteexacerbationofchronicheartandlungdisease.
Furthermore,the2009H1N1pandemiccausedbetween151,700and575,400deathsinworldwideduringthefirstyeartheviruscir-culated[1].
Therefore,preciseon-linemonitoringandforecastingofinfluenzaepidemicoutbreakshasagreatvaluetopublichealthdepartments.
Influenzadetectionandsurveillancesystemsprovideepidemiologicinforma-tionthatcanhelppublichealthsectorsdeveloppreventive*Correspondence:haojianye@gmail.
comXiangleiZhuandBofengFucontributedequallytothiswork.
1CollegeofIntelligenceandComputing,TianjinUniversity,PeiyangParkCampus:No.
135YaguanRoad,HaiheEducationPark,300350Tianjin,ChinaFulllistofauthorinformationisavailableattheendofthearticlemeasuresandassistlocalmedicalinstitutionsindeploy-mentplanning[2].
Influenza-like-illness(ILI)isaninfectiousrespiratoryinfectionmeasurementdefinedbytheWorldHealthOrganization(WHO).
ILIwithameasuredfeverhigherthan38°C,andcough,withonsetwithintheprevious10days[3].
Ourpredictiontarget,ILI%,isequaltotheratiooftheinfluenza-likecasesnumbertothevisitingpatients'number.
Inthefieldofinfluenzasurveillance,ILI%isoftenusedasanindicatortohelpdetermineifthereisapossibleinfluenzaepidemic.
WhentheILI%baselineisexceeded,theinfluenzaseasonhasarrived,remindingthehealthadministrationstotaketimelypreventivemeasures.
Inrecentyears,moreandmoreresearchershavecon-centratedonpreciseon-linemonitoring,earlydetec-tionandinfluenzaepidemicoutbreaksforecasting.
Thus,influenzaepidemicoutbreaksforecastinghasbecomethemostactiveresearchdirection.
TheinformationfromTheAuthor(s).
2019OpenAccessThisarticleisdistributedunderthetermsoftheCreativeCommonsAttribution4.
0InternationalLicense(http://creativecommons.
org/licenses/by/4.
0/),whichpermitsunrestricteduse,distribution,andreproductioninanymedium,providedyougiveappropriatecredittotheoriginalauthor(s)andthesource,providealinktotheCreativeCommonslicense,andindicateifchangesweremade.
TheCreativeCommonsPublicDomainDedicationwaiver(http://creativecommons.
org/publicdomain/zero/1.
0/)appliestothedatamadeavailableinthisarticle,unlessotherwisestated.
Zhuetal.
BMCBioinformatics2019,20(Suppl18):575Page2of10websitesearchorsocialnetworkapplications,suchasTwitterandGoogleCorrelate[4–6],providessufficientdatasupportforthisresearcharea.
Previousmethodsarecommonlybuiltonlinearmodels,suchasleastabsoluteshrinkageandselectionoperator(LASSO)orpenalizedregression[4,6,7].
Somepeoplealsoimplementdeeplearningmodelswhensolvinginfluenzaepidemicfore-castingproblems[8,9].
However,thesemethodscan'teffi-cientlyprovidethepreciseforecastingofILI%oneweekinadvance.
First,theonlinedataisnotaccurateenoughandlacksnecessaryfeatures,whichcannotfullyreflectthetrendoftheinfluenzaepidemic.
Second,influenzaepi-demicdataisusuallyverycomplex,non-stationary,andverynoisy.
Traditionallinearmodelscannothandlemulti-variableinputsappropriately.
Third,previouslyproposeddeeplearningmethodsdidn'tconsiderthetime-sequencepropertyofinfluenzaepidemicdata.
Inthispaper,weuseinfluenzasurveillancedataasourdataset,whichisprovidedbytheGuangzhouCenterforDiseaseControlandPrevention.
Thisdatasetincludesmultiplefeaturesandiscountseparatelyofeachdis-trictinGuangzhou.
Ourapproachtakesadvantageofthesetwocharacteristics.
Meanwhile,weconsiderthetime-sequenceproperty,makingourapproachsolvetheinfluenzaepidemicforecastingprobleminGuangzhouwithpertinence.
Duetotherelevantspecificationsofdatacollection,ourmethodisalsoapplicableinotherregions.
Weconcentrateonimplementingdeeplearningmodelstoaddresstheinfluenzaoutbreaksforecastingprob-lem.
Recently,deeplearningmethodshaveobtainedremarkableperformancesinvariousresearchareasfromcomputervision,speechrecognitiontoclimatefore-casting[10–12].
Weimplementlong-shorttermmemory(LSTM)neuralnetworks[13]asafundamentalmethodforforecasting,becausetheinfluenzaepidemicdatanat-urallyhastimeseriesattribute.
Consideringthatdifferenttypesofinputdatacorrespondtodifferentcharacter-istics,onesingleLSTMwithaspecificfiltermaynotcapturethetimeseriesinformationcomprehensively.
Byusingamulti-channelarchitecture,wecanbettercap-turethetimeseriesattributesfromthedata.
Notonlyensurestheintegrationofvariousrelevantdescriptorsinthehigh-levelnetwork,butalsoensuresthattheinputdatawillnotinterferewitheachotherintheunderlyingnetwork.
ThestructuredLSTMcanproviderobustfittingabilitythathasbeenprovidedinseveralpapers[14,15].
Wefurtherenhanceourmethodusingattentionmecha-nism.
Inattentionlayer,theprobabilityofoccurrenceofeachvalueintheoutputsequencedependsontheval-uesintheinputsequence.
Bydesigningthisarchitecture,wecanbetterdealwithinputstreamrelationshipsamongmultipleregionsmoreappropriately.
WenamedourmodelasAtt-MCLSTM,whichstandsforattention-basedmulti-channelLSTM.
Ourmaincontributionscanbesummarizedasfollows:(1)WetestourmodelonGuangzhouinfluenzasurveil-lancedataset,whichisauthenticandreliable.
Itcontainsmultipleattributesandtimeseriesfeatures.
(2)Wepro-poseanattention-basedmulti-channelLSTMstructurethatassociatesdifferentwell-behavedapproaches.
Thestructuretakestheforecastingproblemandtheinfluenzaepidemicdataattributesintoaccount.
Theproposedmodelcanbeseenasanalternativetoforecastinfluenzaepidemicoutbreaksinotherdistricts.
(3)Theproposedmodelmakesfulluseofinformationinthedataset,solv-ingtheactualproblemofinfluenzaepidemicforecastinginGuangzhouwithpertinence.
Theexperimentalresultsdemonstratethevalidityofourmethod.
Tothebestofourknowledge,thisisthefirststudythatappliesLSTMneuralnetworkstotheinfluenzaoutbreaksforecastingproblem.
Therestofthispaperisorganizedasfollows.
Inthesecondsection,weillustratedetailsofourmethod.
Inthethirdsection,weevaluateperformancesofourmethodbycomparingitwithdifferentneuralnetworkstructuresandotherpriorartmethods.
Inthefourthsection,wediscussconclusionsandprospectsforfutureworks.
MethodsTheaccuratenessoftheforecastingproblemscanbeenhancedbycombiningmultiplemodels[16–26].
Inthispaper,wedeviseannovelLSTMneuralnetworkstructuretosettletheinfluenzaepidemicforecastingprobleminGuangzhou,China.
Ourmodelcanextractcharacteristicsmoreeffectivelyfromtimeseriesdata,andtakediffer-entimpactsofdifferentpartsofdataintoconsideration.
Inordertoillustrateourmodelclearly,weillustrateourdatasetfirst.
Thefollowingsectionswillgivefurtherillus-trationsonthedataset,theoverallideaofourmodel,detailsofLSTMneuralnetworks,attentionmechanism,attention-basedmulti-channelLSTM,datanormalization,andevaluationmethod.
DatasetdescriptionTheinfluenzasurveillancedataweusedincludes9yearsdata.
Statisticsoninfluenzaepidemicdatain9regionsarecountedeachyear.
Thedatasetincludes6modules,andeachofthesemoduleshasmultiplefeatures.
Thedatasethasonerecordeachweek,anddatafor52weeksiscountedeachyear.
DesignoftheproposedmodelInFig.
1,wedemonstratetheflowdiagramofourmethod.
Theintegratedflowdiagramhastwoparts,trainingpartandtestpart.
Inthetrainingpart,first,weselect19relevantfeaturesafterdatacleaningandnormalizationprocesses.
WefurtherillustratethechosenmodulesandZhuetal.
BMCBioinformatics2019,20(Suppl18):575Page3of10Fig.
1TheflowchartofAttention-basedmulti-channelLSTMfeaturesinTable1.
Table1doesn'tincludebasicinforma-tionmodule,whichincludestimeinformation,districts,andpopulation.
Weusemodel-basedrankingmethodasourfeatureselectionmethod.
Inordertoimplementmodel-basedrankingmethod,wedeleteonefeatureatatime,andinputtherestoffeaturesintothesamefore-castingmodeleverytime.
Iftheforecastingaccuracyislow,thismeansthatthefeatureweremovedisrelevanttoourforecastingobjective.
Afterrankingallthefore-castingaccuracy,weselect19featuresthatarerelevanttotheforecastingobjective.
Thenweseparatethedatasetintotrainingdatasetandtestdataset.
Thetrain-ingdatasetcontains80percentofdatatoextractannualtrendandseasonalperiodicity.
Inthetestpart,wetestourmodelonthetestdataset.
Then,wepreformdenor-malizationprocesstoreconstructtheoriginalvalues.
Finally,weassessourmodelandcompareitwithothermodels.
DatanormalizationMin-Maxnormalizationisalineartransformationstrategy[27].
Thismethodmaintainstherelationshipamongalltheoriginaldata.
Min-Maxnormalizationtransformsavaluextoy,yisdefinedasEq.
1.
y=(xmin)(maxmin)(1)Whereministhesmallestvalueinthedata,maxisthebiggestvalueinthedata.
Afterdatanormalization,thefeaturesofdatawillbescaledbetween0and1.
Wepreformdenormalizationprocesstoreconstructtheoriginaldata.
Givenanormalizedvaluey,itsoriginalvaluexisdefinedasEq.
2.
x=(maxmin)y+min(2)Zhuetal.
BMCBioinformatics2019,20(Suppl18):575Page4of10Table1ModulesandfeaturesdescriptionforSection2.
1ModulenameFeaturenameDescriptionLegalinfluenzacasesreportmoduleLegalinfluenzacasesnumbersThenumberofinfluenzacasesinthenationalinfectiousdiseasereportingsystem.
EpidemicmonitoringmoduleInfluenzaoutbreaksnumbersMorethan10influenza-likecasesoccurredwithinoneweekinthesameunit.
AffectedcasesnumbersThetotalnumberofpeopleaffectedbytheepidemic.
SymptommonitoringmoduleInfluenza-likecasesnumbers(0-5age)Thenumberofinfluenza-likecases(0-5age).
Influenza-likecasesnumbers(5-15age)Thenumberofinfluenza-likecases(5-15age).
Influenza-likecasesnumbers(15-25age)Thenumberofinfluenza-likecases(15-25age).
Influenza-likecasesnumbers(25-60age)Thenumberofinfluenza-likecases(25-60age).
Influenza-likecasesnumbers(>60age)Thenumberofinfluenza-likecases(over60age).
Totalinfluenza-likecasesnumbersThetotalnumberofinfluenza-likecases.
TotalvisitingpatientsnumbersThetotalnumberofvisitingpatients.
UpperrespiratorytractinfectionsnumbersThenumberofupperrespiratorytractinfections.
PharmacymonitoringmoduleChinesepatentcoldmedicinesSalesofChinesepatentcoldmedicines.
OthercoldmedicinesSalesofothercoldmedicines.
ClimatedatamoduleAveragetemperature(°C)Averagetemperature.
Maximumtemperature(°C)Maximumtemperature.
Minimumtemperature(°C)Minimumtemperature.
Rainfall(mm)Rainfall.
Airpressure(hPa)Airpressure.
Relativehumidity(%)Relativehumidity.
Long-shorttermmemoryneuralnetworkRecurrentneuralnetworkshavetheabilitytodynam-icallycombineexperiencesbecauseoftheirinternalrecurrence[28].
DifferentfromothertraditionalRNNs,LSTMcandealwiththegradientvanishingprob-lem[29].
ThememoryunitsofLSTMcellsretaintimeseriesattributesofgivencontext[29].
SomeresearcheshaveproventhatLSTMneuralnetworkscanyieldabetterperformancecomparedwithothertradi-tionalRNNswhendealingwithlong-termtimeseriesdata[30].
ThestructureofasingleLSTMcellillustrateinFig.
2.
Thegatescontroltheflowofinformation,thatis,inter-actionsbetweendifferentcellsandcellitself.
Inputgatecontrolsthememorystateupdatingprocess.
Outputgatecontrolswhethertheoutputflowcanalterothercells'memorystate.
Forgetgatecanchoosetorememberorfor-getitspreviousstate.
LSTMisimplementedbyfollowingcompositefunctions:it=σ(Wxixt+Whiht1+Wcict1+bi)(3)ft=σ(Wxfxt+Whfht1+Wcfct1+bf)(4)ct=ft1+ittanh(Wxcxt+Whcht1+bc)(5)ot=σ(Wxoxt+Whoht1+Wcoct+bo)(6)ht=ottanh(ct)(7)Whereσrepresentthelogisticsigmoidfunction.
i,f,o,andcrepresenttheinputgate,forgetgate,outputgate,Fig.
2ThestructureofsingleLSTMcellZhuetal.
BMCBioinformatics2019,20(Suppl18):575Page5of10cellinputactivationvectorsrespectively.
hrepresentsthehiddenvector.
Theweightmatrixsubscriptshavetheintu-itivemeaning.
Like,Whirepresentsthehidden-inputgatematrixetc.
AttentionmechanismTraditionalEncode-Decodestructurestypicallyencodeaninputsequenceintoafixed-lengthvectorrepresenta-tion.
However,thismodelhasdrawbacks.
Whentheinputsequenceisverylong,itisdifficulttolearnafeasiblevectorrepresentation.
Onefundamentaltheoryofattentionmechanism[31]istoabandontheconventionalEncoder-Decoderstruc-ture.
Attentionmechanismtrainsamodelthatselectivelylearnstheinputstreamsbyconservingtheintermedi-ateoutputsofLSTM.
Inattentionstructure,theoutputsequencesareaffiliatedwiththeinputsequences.
Inotherwords,theprobabilityofoccurrenceofeachvalueintheoutputsequencedependsonthevalueintheinputsequence.
Figure3illustratestheattentionmechanism.
AttentionlayercalculatestheweighteddistributionofX1,.
.
.
,XT.
TheinputofStcontainstheoutputoftheattentionlayer.
Theprobabilityofoccurrenceoftheout-putsequence.
.
.
,yt1,yt,.
.
.
dependsoninputsequenceX1,X2,.
.
.
,XT.
hirepresentsthehiddenvector.
At,irepre-sentstheweightofithinputattimestept.
Attentionlayerinputsnparametersy1,.
.
.
,yn,contextsequencec,andoutputsvectorz,zistheweighteddistributionofyiforagivencontextc.
Attentionmechanismisimplementedbyfollowingcompositefunction:mi=tanh(Wcmc+Wymyi)(8)si∝exp(wm,mi)(9)isi=1(10)z=isiyi(11)Wheremiiscalculatedbytanhlayer,siisthesoftmaxofthemiprojectedonalearneddirection.
TheoutputzisFig.
3Thediagramofattentionmechanism.
AttentionlayercalculatestheweighteddistributionofX1,.
.
.
,XT.
TheinputofStcontainstheoutputoftheattentionlayer.
Theprobabilityofoccurrenceoftheoutputsequence.
.
.
,yt1,yt,.
.
.
dependsoninputsequenceX1,X2,.
.
.
,XT.
hirepresentsthehiddenvector.
At,irepresentstheweightofithinputattimesteptZhuetal.
BMCBioinformatics2019,20(Suppl18):575Page6of10theweightedarithmeticmeanofallyi,Wrepresentstherelevanceforeachvariableaccordingtothecontextc.
Attention-basedmulti-channelLSTMInFig.
4,weillustratetheoverallarchitectureofourmodel.
Weseparateourdatasetintotwocategories.
First,weclassifyaveragetemperature,maximumtemperature,minimumtemperature,rainfall,airpressureandrelativehumiditytogetherasclimate-relateddatacategory.
Then,therestoffeaturesareclassifiedtogetherasinfluenza-relateddatacategory.
Inourdataset,eachregionhasitsowninfluenza-relateddata,andtheysharethesameclimate-relateddataeveryweek.
Becauseourdatasethastheabovecharacteristics,theinputsofAtt-MCLSTMcontainstwoparts.
First,theinfluenza-relateddataisinputintoaseriesofLSTMneu-ralnetworks(LSTM1,.
.
.
,LSTM9)tocapturecorrelativefeatures.
Second,theclimate-relateddataisinputintoasingleLSTMneuralnetwork(LSTM10)tocapturethelong-termtimeseriesattributeofinfluenzaepidemicdata.
Forthefirstpart,eachLSTMneuralnetworkacquirestheinfluenza-relateddatafromonedistinctregion.
Inordertomakefulluseofthecomplementarityamongeveryregions,theoutputsofLSTMneuralnetworks(LSTM1,.
.
.
,LSTM9)areconcatenatedinahigherlayer(Merge1).
Thishigherlayercanobtainthefuseddescriptorsofunderlyingneuralnetworks.
Afterwecapturethefea-turesofeveryregions,westillwanttoweightintermediatesequences.
Thereasonisthatthedataofeachregionhasdifferentinfluencesonthefinalforecastingresult.
Therefore,theintermediatesequencespassthroughanattentionlayer(Attention)andafullyconnectedlayer(Dense1)inturn.
Thereafter,weconcatenatetheoutputsofthesetwopartstogether(Merge2).
Finally,theinterme-diatesequencesarepassedthroughtwofullyconnectedlayers(Dense2,Dense3).
Sofar,weacquirethehigh-levelfeaturesoftheinputdata,andtheyareusedtosolvetheinfluenzaepidemicforecasting.
Bydesigningamulti-channelstructure,wecanbet-terextractthetime-sequencepropertyofeachtypeofdata.
Notonlyensurestheintegrationofvariousrelevantdescriptorsinthehigh-levelnetwork,butalsoensuresthatinputdatawillnotinterferewitheachotherintheunderlyingnetwork.
Intheattentionlayer,theprobabilityofoccurrenceofeachvalueintheoutputsequencedependsonthevalueintheinputsequence.
Thisstructureallowsustohandletherelationshipofinputdatabetweendifferentdistrictsmoreappropriately.
EvaluationmethodToevaluateourmethod,weusethemeanabsoluteper-centageerror(MAPE)asthecriteriastandard.
ItsformulaisexpressasEq.
12.
MAPE=1nni=1|yixiyi|*100(12)Fig.
4ThestructureofAttention-basedmulti-channelLSTMZhuetal.
BMCBioinformatics2019,20(Suppl18):575Page7of10Whereyidenotestheithactualvalue,andxidenotestheithpredictedvalue.
IfthevalueofMAPEislow,theaccuracyofthemethodishigh.
ExperimentsInthissection,wedidtwoexperimentstoverifytheAtt-MCLSTMmodel.
Inthefirstexperiment,weevaluatethenumbersofconsecutiveweeksofdatathatweneedtoforecastILI%forthenextweek.
Inthesecondexperiment,wecompareourmodelwithdifferentneuralnetworkstructuresandothermethods.
Eachexperimentresultistheaverageof10repeatedtrials.
SelectionofconsecutiveweeksInthisexperiment,wesetthenumbersofconsecu-tiveweeksas6,8,10,12,14respectively.
Thehyper-parametersofeachlayerarelistedinTable2.
Theacti-vationfunctionsweusedarelinearactivationfunction.
Thelossfunctionandoptimizeraremapeandadamrespectively.
Weusethefirst370consecutiveweeks'dataintrainingphaseandtheremainingdatainthetestphase.
Eachdatasampleincludes6featuresinclimate-relateddatacate-goryand9differentdistricts'influenza-relateddata.
Eachinfluenza-relateddatacontains13features.
Theclimate-relateddataandeachdistrict'sinfluenza-relateddataareinputintotheclimate-relatedchannelandtheinfluenza-relatedchannelrespectively.
TheforecastingresultsareshowninTable3.
PerformancevalidationInthisexperiment,weverifythevalidityofourmodel.
First,wecompareAtt-MCLSTMwithMCLSTMbycomparingtheirforecastingaccuracy.
Thepurposeofdoingthisistoverifytheeffectoftheattentionmechanism.
Forbothmodels,weusethesamemulti-channelarchitecture(asshowninFig.
4).
TheonlydifferencebetweenthesetwomodelsisthatwedeletetheattentionlayerinMCLSTM.
Theparameterssettingsanddatainputsmethodareasdescribedinthefirstexperiment.
Table2ThesizeofeveryunitinAtt-MCLSTMneuralnetworkforSection3.
1LayernameUnitsnumberLSTM1,.
.
.
,LSTM932LSTM1032Dense116Dense210Dense31Table3TheMAPEofthepredictionresultsforSection3.
1NumberofweeksMAPE60.
10780.
092100.
086120.
106140.
109Second,wecompareMCLSTMwithLSTMbycomparingtheirforecastingaccuracy.
Thepurposeofdoingthisistoverifytheeffectofthemulti-channelstructure.
ForMCLSTM,parameterssettingsanddatainputsmethodareasdescribedinthefirstexperiment.
ForLSTM,weinputentirefeaturesintooneLSTMlayertocapturethefuseddescriptors.
Insteadofseparatingdatasetaccordingtodifferentregions,wesumcorrespondinginfluenza-relatedfeaturesineachweekfromeveryregionstogether.
Therefore,eachdatarecordincludes19selectedfeatures.
Thedatathatcontainsthese19featuresarepassedthroughafullyconnectedlayertoacquirehigh-levelfeatures.
Theunits'numberofLSTMlayerandfullyconnectedlayerare32and1respectively.
Third,wedemonstratethatLSTMscanyieldbetterperformancethanRNNswhendealingwithtimeseriesdata.
Results(1)AscanbeseenfromTable3,10consecutiveweeks'datayieldsthebestperformance.
(2)Table4showsthatAtt-MCLSTMhasstrongcompetitivenessandcanpro-videeffectivereal-timeinfluenzaepidemicforecasting.
DiscussionTheresultsofthefirstexperimentindicatethat10consec-utiveweeksdatacanappropriatelyreflectthetimeseriesattributeofinfluenzadata.
Ifthelengthofinputdataisshorterthan10,theinputdatadoesn'tcontainenoughtimeseriesinformation.
Onthecontrary,ifthelengthofinputdataislongerthan10,thenoiseinsidetheinputdataincreased,leadingtoadecreaseinforecastingaccuracy.
Therefore,inourexperiments,eachdatarecordincludes10consecutiveweeks'data.
Table4TheMAPEofthepredictionresultsforSection3.
2SchemesMAPEAtt-MCLSTM0.
086MCLSTM0.
105LSTM0.
118RNN0.
132Zhuetal.
BMCBioinformatics2019,20(Suppl18):575Page8of10TheresultsofthesecondexperimentshowthatAtt-MCLSTMcanyieldthebestperformance.
InTable4,fromthefirsttworows,wecanconcludethatusingatten-tionmechanismcanimprovetheMAPEfrom0.
105to0.
086.
Thereasonisthattheattentionlayercanbet-terdealwiththerelationshipsofinputstreamsamongeveryregionsmoreappropriately.
Fromthesecondrowandthethirdrow,wecanconcludethatusingmulti-channelstructurecanimprovetheMAPEfrom0.
118to0.
105.
Thereasonisthatthemulti-channelstructurecanbettercapturethetimeseriesattributesfromdif-ferentinputstreams.
Fromthelasttworows,wecanconcludethatusingLSTMcanimprovetheMAPEfrom0.
132to0.
118.
ThereasonisthatLSTMneuralnet-workcanbetterdealwithtimeseriesdata.
Thisresultalsodemonstratesthetimeseriesattributeofinfluenzaepidemicdata.
Figure5showstheactualvaluesandpredictedval-uesoffourmodels.
WecanseethattheresultofAtt-MCLSTMisclosetotheactualoutput.
Therearemoreobviousdifferencesbetweenthepre-dictedresultsandtheactualvaluebyusingtheotherthreemodels.
So,thiscanverifythatadoptingAtt-MCLSTMtoanalyzethesequentialinformationcanhelptoextracttime-sequencecharacteristicmoreaccuratelyandcomprehensively.
ConclusionandfutureworkInthispaper,weproposeanewdeepneuralnet-workstructure(Att-MCLSTM)toforecasttheILI%inGuangzhou,China.
First,weimplementthemulti-channelarchitecturetocapturetimeseriesattributesfromdifferentinputstreams.
Then,theattentionmechanismisappliedtoweightthefusedfeaturesequences,whichallowsustodealwithrelationshipsbetweendifferentinputstreamsmoreappropriately.
Ourmodelfullycon-sidertheinformationinthedataset,targetedlysolvingthepracticalproblemofinfluenzaepidemicforecastinginGuangzhou.
Weassesstheperformanceofourmodelbycomparingitwithdifferentneuralnetworkstructuresandotherstate-of-the-artmodels.
Theexperimentalresultsindicatethatourmodelhasstrongcompetitivenessandcanprovideeffectivereal-timeinfluenzaepidemicfore-casting.
Tothebestofourknowledge,thisisthefirststudythatappliesLSTMneuralnetworkstotheinfluenzaout-breaksforecasting.
Continuingworkwillfurtherimprovetheexpansionabilityofourmodelbyintroducingtransferlearning.
Fig.
5Theresultsofone-weekaheadpredictionbyusingfourindividualmodels.
ashowsthecomparisonofAtt-MCLSTMandrealdata;bshowsthecomparisonofMCLSTMandrealdata;cshowsthecomparisonofLSTMandrealdata;dshowsthecomparisonoftraditionalRNNandrealdata.
Ineachfigure,thebluelinedenotestheactualvalues,andtheorangelinedenotesthepredictedvaluesZhuetal.
BMCBioinformatics2019,20(Suppl18):575Page9of10AbbreviationsILI:Influenza-likeillness;LSTM:Longshort-termmemory;LASSO:Leastabsoluteshrinkageandselectionoperator;MAPE:MeanabsolutepercentageerrorAcknowledgementsWethankthereviewers'valuablecommentsforimprovingthequalityofthiswork.
AboutthissupplementThisarticlehasbeenpublishedaspartofBMCBioinformaticsVolume20Supplement18,2019:SelectedarticlesfromtheBiologicalOntologiesandKnowledgebasesworkshop2018.
Thefullcontentsofthesupplementareavailableonlineathttps://bmcbioinformatics.
biomedcentral.
com/articles/supplements/volume-20-supplement-18.
Authors'contributionsXZandBFcontributedequallytothealgorithmdesignandtheoreticalanalysis.
YY,YM,JH,SC,SL,TL,SL,WG,andZLcontributedequallytothethequalitycontrolanddocumentreviewing.
Allauthorsreadandapprovedthefinalmanuscript.
FundingPublicationcostsarefundedbyTheNationalNaturalScienceFoundationofChina(GrantNos.
:U1836214),TianjinDevelopmentProgramforInnovationandEntrepreneurshipandSpecialProgramofArtificialIntelligenceofTianjinMunicipalScienceandTechnologyCommission(NO.
:17ZXRGGX00150).
AvailabilityofdataandmaterialsAlldatainformationoranalyzedduringthisstudyareincludedinthisarticle.
EthicsapprovalandconsenttoparticipateNotapplicable.
ConsentforpublicationNotapplicable.
CompetinginterestsTheauthorsdeclarethattheyhavenocompetinginterests.
Authordetails1CollegeofIntelligenceandComputing,TianjinUniversity,PeiyangParkCampus:No.
135YaguanRoad,HaiheEducationPark,300350Tianjin,China.
2AutomotiveDataCenter,ChinaAutomotiveTechnology&Research,300300Tianjin,China.
3GuangzhouCenterforDiseaseControlandPrevention,510440Guangzhou,China.
4PonyTestingInternationalGroup,300051Tianjin,China.
5TianjinFoodSafetyInspectionTechnologyInstitute,300300Tianjin,China.
Published:25November2019References1.
YangS,SantillanaM,KouSC.
Accurateestimationofinfluenzaepidemicsusinggooglesearchdataviaargo.
ProcNatlAcadSci.
2015;112(47):14473–8.
2.
BrownsteinJS,MandlKD.
Reengineeringrealtimeoutbreakdetectionsystemsforinfluenzaepidemicmonitoring.
In:AMIAAnnualSymposiumProceedings,vol.
2006.
AmericanMedicalInformaticsAssociation;2006.
p.
866.
3.
OrganizationWH,etal.
Whointerimglobalepidemiologicalsurveillancestandardsforinfluenza.
20121–61.
4.
SantillanaM,ZhangDW,AlthouseBM,AyersJW.
Whatcandigitaldiseasedetectionlearnfrom(anexternalrevisionto)googleflutrendsAmJPrevMed.
2014;47(3):341–7.
5.
AchrekarH,GandheA,LazarusR,YuS.
-H.
,LiuB.
Predictingflutrendsusingtwitterdata.
In:ComputerCommunicationsWorkshops(INFOCOMWKSHPS),2011IEEEConferenceOn.
IEEE;2011.
p.
702–7.
https://doi.
org/10.
1109/infcomw.
2011.
5928903.
6.
BroniatowskiDA,PaulMJ,DredzeM.
Nationalandlocalinfluenzasurveillancethroughtwitter:ananalysisofthe2012-2013influenzaepidemic.
PLoSONE.
2013;8(12):83672.
7.
SantillanaM,NsoesieEO,MekaruSR,ScalesD,BrownsteinJS.
Usingclinicians'searchquerydatatomonitorinfluenzaepidemics.
ClinInfectDisOffPublInfectDisSocAm.
2014;59(10):1446.
8.
XuQ,GelYR,RamirezLLR,NezafatiK,ZhangQ,TsuiK.
-L.
Forecastinginfluenzainhongkongwithgooglesearchqueriesandstatisticalmodelfusion.
PLoSONE.
2017;12(5):0176690.
9.
HuH,WangH,WangF,LangleyD,AvramA,LiuM.
Predictionofinfluenza-likeillnessbasedontheimprovedartificialtreealgorithmandartificialneuralnetwork.
SciRep.
2018;8(1):4895.
10.
HintonGE,SalakhutdinovRR.
Reducingthedimensionalityofdatawithneuralnetworks.
science.
2006;313(5786):504–7.
11.
ZouB,LamposV,GortonR,CoxIJ.
Oninfectiousintestinaldiseasesurveillanceusingsocialmediacontent.
In:Proceedingsofthe6thInternationalConferenceonDigitalHealthConference.
ACM;2016.
p.
157–61.
https://doi.
org/10.
1145/2896338.
2896372.
12.
HuangW,SongG,HongH,XieK.
Deeparchitecturefortrafficflowprediction:Deepbeliefnetworkswithmultitasklearning.
IEEETransIntellTranspSyst.
2014;15(5):2191–201.
13.
HowDNT,LooCK,SahariKSM.
Behaviorrecognitionforhumanoidrobotsusinglongshort-termmemory.
IntJAdvRobotSyst.
2016;13(6):1729881416663369.
14.
YangY,HaoJ,SunM,WangZ,FanC,StrbacG.
Recurrentdeepmultiagentq-learningforautonomousbrokersinsmartgrid.
In:IJCAI,vol.
18;2018.
p.
569–75.
https://doi.
org/10.
24963/ijcai.
2018/79.
15.
YangY,HaoJ,WangZ,SunM,StrbacG.
Recurrentdeepmultiagentq-learningforautonomousagentsinfuturesmartgrid.
In:Proceedingsofthe17thInternationalConferenceonAutonomousAgentsandMultiAgentSystems.
InternationalFoundationforAutonomousAgentsandMultiagentSystems;2018.
p.
2136–8.
https://doi.
org/10.
24963/ijcai.
2018/79.
16.
Shafie-KhahM,MoghaddamMP,Sheikh-El-EslamiM.
Priceforecastingofday-aheadelectricitymarketsusingahybridforecastmethod.
EnergyConversManag.
2011;52(5):2165–9.
17.
XiaotianH,WeixunW,JianyeH,YaodongY.
Independentgenerativeadversarialself-imitationlearningincooperativemultiagentsystems.
In:Proceedingsofthe18thInternationalConferenceonAutonomousAgentsandMultiAgentSystems;2019.
p.
1315–1323.
InternationalFoundationforAutonomousAgentsandMultiagentSystems.
18.
YaodongY,JianyeH,YanZ,XiaotianH,BofengF.
Large-scalehomeenergymanagementusingentropy-basedcollectivemultiagentreinforcementlearningframework.
In:Proceedingsofthe18thInternationalConferenceonAutonomousAgentsandMultiAgentSystems;2019.
https://doi.
org/10.
24963/ijcai.
2019/89.
19.
HongyaoT,JianyeH,TangjieLv,YingfengC,ZongzhangZ,HangtianJ,ChunxuR,YanZ,ChangjieF,LiW.
HierarchicaldeepmultiagentreinforcementlearningwithTemporalAbstraction.
In:arXivpreprintarXiv:1809.
09332;2018.
20.
PengJ,GuanJ,ShangX.
Predictingparkinson'sdiseasegenesbasedonnode2vecandautoencoder.
FrontGenet.
2019;10:226.
21.
PengJ,ZhuL,WangY,ChenJ.
Miningrelationshipsamongmultipleentitiesinbiologicalnetworks.
IEEE/ACMTransComputBiolBioinforma.
2019.
https://doi.
org/10.
1109/tcbb.
2019.
2904965.
22.
PengJ,XueH,ShaoY,ShangX,WangY,ChenJ.
Anovelmethodtomeasurethesemanticsimilarityofhpoterms.
IJDMB.
2017;17(2):173–88.
23.
ChengL,HuY,SunJ,ZhouM,JiangQ.
Dincrna:acomprehensiveweb-basedbioinformaticstoolkitforexploringdiseaseassociationsandncrnafunction.
Bioinformatics.
2018;34(11):1953–6.
24.
ChengL,WangP,TianR,WangS,GuoQ,LuoM,ZhouW,LiuG,JiangH,JiangQ.
Lncrna2targetv2.
0:acomprehensivedatabasefortargetgenesoflncrnasinhumanandmouse.
NucleicAcidsRes.
2018;47(D1):140–4.
25.
HuY,ZhaoT,ZangT,ZhangY,ChengL.
Identificationofalzheimer'sdisease-relatedgenesbasedondataintegrationmethod.
FrontGenet.
2018;9:.
https://doi.
org/10.
3389/fgene.
2018.
00703.
26.
PengJ,HuiW,LiQ,ChenB,JiangQ,WeiZ,ShangX.
Alearning-basedframeworkformirna-diseaseassociationpredictionusingneuralnetworks.
bioRxiv.
2018276048.
https://doi.
org/10.
1101/276048.
27.
PandaSK,JanaPK.
Efficienttaskschedulingalgorithmsforheterogeneousmulti-cloudenvironment.
JSupercomput.
2015;71(4):1505–33.
28.
MurtaghF,StarckJ-L,RenaudO.
Onneuro-waveletmodeling.
DecSupportSyst.
2004;37(4):475–84.
29.
HochreiterS,SchmidhuberJ.
Longshort-termmemory.
NeuralComput.
1997;9(8):1735–80.
Zhuetal.
BMCBioinformatics2019,20(Suppl18):575Page10of1030.
PalangiH,DengL,ShenY,GaoJ,HeX,ChenJ,SongX,WardR.
Deepsentenceembeddingusinglongshort-termmemorynetworks:Analysisandapplicationtoinformationretrieval.
IEEE/ACMTransAudioSpeechLangProcess(TASLP).
2016;24(4):694–707.
31.
VaswaniA,ShazeerN,ParmarN,UszkoreitJ,JonesL,GomezAN,Kaiser,PolosukhinI.
Attentionisallyouneed.
In:AdvancesinNeuralInformationProcessingSystems;2017.
p.
5998–6008.
Publisher'sNoteSpringerNatureremainsneutralwithregardtojurisdictionalclaimsinpublishedmapsandinstitutionalaffiliations.

EtherNetservers年付仅10美元,美国洛杉矶VPS/1核512M内存10GB硬盘1Gpbs端口月流量500GB/2个IP

EtherNetservers是一家成立于2013年的英国主机商,提供基于OpenVZ和KVM架构的VPS,数据中心包括美国洛杉矶、新泽西和杰克逊维尔,商家支持使用PayPal、支付宝等付款方式,提供 60 天退款保证,这在IDC行业来说很少见,也可见商家对自家产品很有信心。有需要便宜VPS、多IP VPS的朋友可以关注一下。优惠码SUMMER-VPS-15 (终身 15% 的折扣)SUMMER-...

iON Cloud七月促销适合稳定不折腾的用户,云服务器新购半年付8.5折,洛杉矶/圣何塞CN2 GT线路,可选Windows系统

iON Cloud怎么样?iON Cloud今天发布了7月份优惠,使用优惠码:VC4VF8RHFL,新购指定型号VPS半年付或以上可享八五折!iON的云服务器包括美国洛杉矶、美国圣何塞(包含了优化线路、CN2 GIA线路)、新加坡(CN2 GIA线路、PCCW线路、移动CMI线路)这几个机房或者线路可供选择,有Linux和Windows系统之分,整体来说针对中国的优化是非常明显的,机器稳定可靠,比...

Hosteons - 限时洛杉矶/达拉斯/纽约 免费升级至10G带宽 低至年$21

Hosteons,一家海外主机商成立于2018年,在之前还没有介绍和接触这个主机商,今天是有在LEB上看到有官方发送的活动主要是针对LEB的用户提供的洛杉矶、达拉斯和纽约三个机房的方案,最低年付21美元,其特点主要在于可以从1G带宽升级至10G,而且是免费的,是不是很吸引人?本来这次活动是仅仅在LEB留言提交账单ID才可以,这个感觉有点麻烦。不过看到老龚同学有拿到识别优惠码,于是就一并来分享给有需...

google中国地图为你推荐
SinglesbDescriptionios5ipad如何上网IPAD4怎样上网?勒索病毒win7补丁win7有针对勒索病毒的补丁吗win10445端口win的22端口和23端口作用分别是什么 ?tcpip上的netbios怎么启用TCP/IP上的NetBIOSicloudiphone苹果6显示已停用请连接itunes什么意思micromediaMacromedia翻译成中文是什么?googleadsensegoogle adsense 和google adwords有什么区别?适合什么样的人群?迅雷下载速度为什么现在迅雷下载的速度比原来慢得多?
免费虚拟主机申请 bluevm 免费主机 mach5 10t等于多少g 域名优惠码 godaddy 好看的桌面背景图片 godaddy域名证书 ftp教程 me空间社区 中国电信测速网 新睿云 跟踪路由命令 四川电信商城 网站加速软件 cxz smtp服务器地址 主机管理系统 国外在线代理服务器 更多