Please use this identifier to cite or link to this item: http://dx.doi.org/10.25673/60208
Full metadata record
DC FieldValueLanguage
dc.contributor.authorBeyer, Christian-
dc.contributor.authorUnnikrishnan, Vishnu-
dc.contributor.authorBrüggemann, Robert-
dc.contributor.authorToulouse, Vincent-
dc.contributor.authorOmar, Hafez Kader-
dc.contributor.authorNtoutsi, Eirini-
dc.contributor.authorSpiliopoulou, Myra-
dc.date.accessioned2022-01-26T10:31:40Z-
dc.date.available2022-01-26T10:31:40Z-
dc.date.issued2020-
dc.date.submitted2020-
dc.identifier.urihttps://opendata.uni-halle.de//handle/1981185920/62159-
dc.identifier.urihttp://dx.doi.org/10.25673/60208-
dc.description.abstractMany current and future applications plan to provide entity-specific predictions. These range from individualized healthcare applications to user-specific purchase recommendations. In our previous stream-based work on Amazon review data, we could show that error-weighted ensembles that combine entity-centric classifiers, which are only trained on reviews of one particular product (entity), and entity-ignorant classifiers, which are trained on all reviews irrespective of the product, can improve prediction quality. This came at the cost of storing multiple entity-centric models in primary memory, many of which would never be used again as their entities would not receive future instances in the stream. To overcome this drawback and make entity-centric learning viable in these scenarios, we investigated two different methods of reducing the primary memory requirement of our entity-centric approach. Our first method uses the lossy counting algorithm for data streams to identify entities whose instances make up a certain percentage of the total data stream within an error-margin. We then store all models which do not fulfil this requirement in secondary memory, from which they can be retrieved in case future instances belonging to them should arrive later in the stream. The second method replaces entity-centric models with a much more naive model which only stores the past labels and predicts the majority label seen so far. We applied our methods on the previously used Amazon data sets which contained up to 1.4M reviews and added two subsets of the Yelp data set which contain up to 4.2M reviews. Both methods were successful in reducing the primary memory requirements while still outperforming an entity-ignorant model.eng
dc.description.sponsorshipProjekt DEAL 2020-
dc.language.isoeng-
dc.relation.ispartofhttp://link.springer.com/journal/12243-
dc.rights.urihttps://creativecommons.org/licenses/by/4.0/-
dc.subjectEntity-centric learningeng
dc.subjectStream classificationeng
dc.subjectDocument predictioneng
dc.subjectMemory reductioneng
dc.subjectText ignorant modelseng
dc.subject.ddc000-
dc.titleResource management for model learning at entity leveleng
dc.typeArticle-
dc.identifier.urnurn:nbn:de:gbv:ma9:1-1981185920-621591-
local.versionTypepublishedVersion-
local.bibliographicCitation.journaltitleAnnals of telecommunications-
local.bibliographicCitation.volume75-
local.bibliographicCitation.pagestart549-
local.bibliographicCitation.pageend561-
local.bibliographicCitation.publishernameLavoisier-
local.bibliographicCitation.publisherplaceParis-
local.bibliographicCitation.doi10.1007/s12243-020-00800-4-
local.openaccesstrue-
dc.identifier.ppn1735990388-
local.bibliographicCitation.year2020-
cbs.sru.importDate2022-01-26T10:23:06Z-
local.bibliographicCitationEnthalten in Annals of telecommunications - Paris : Lavoisier, 1946-
local.accessrights.dnbfree-
Appears in Collections:Fakultät für Informatik (OA)

Files in This Item:
File Description SizeFormat 
Beyer et al._Resource_2020.pdfZweitveröffentlichung1.84 MBAdobe PDFThumbnail
View/Open