System disruptions
We are currently experiencing disruptions on the search portals due to high traffic. We are working to resolve the issue, you may temporarily encounter an error message.
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • apa.csl
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Modeling Batch Tasks Using Recurrent Neural Networks in Co-Located Alibaba Workloads
Tufts University, United States.
Karlstad University, Faculty of Health, Science and Technology (starting 2013), Department of Mathematics and Computer Science (from 2013).ORCID iD: 0000-0001-7547-8111
Karlstad University, Faculty of Health, Science and Technology (starting 2013), Department of Mathematics and Computer Science (from 2013). Red Hat.ORCID iD: 0000-0002-0722-2656
Tufts University, United States.
2024 (English)In: / [ed] Modesto Castrillon-Santana; Maria De Marsico; Ana Fred, SciTePress, 2024, Vol. 1, p. 558-569Conference paper, Published paper (Refereed)
Abstract [en]

Accurate predictive models for cloud workloads can be helpful in improving task scheduling, capacity planning and preemptive resource conflict resolution, especially in the setting of co-located jobs. Alibaba, one of the leading cloud providers co-locates transient batch tasks and high priority latency sensitive online jobs on the same cluster. In this paper, we consider the problem of using a publicly released dataset by Alibaba to model the batch tasks that are often overlooked compared to online services. The dataset contains the arrivals and resource requirements (CPU, memory, etc.) for both batch and online tasks. Our trained model predicts, with high accuracy, the number of batch tasks that arrive in any 30 minute window, their associated CPU and memory requirements, and their lifetimes. It captures over 94% of arrivals in each 30 minute window within a 95% prediction interval. The F1 scores for the most frequent CPU classes exceed 75%, and our memory and lifetime predictions incur less than 1% test data loss. The prediction accuracy of the lifetime of a batch-task drops when the model uses both CPU and memory information, as opposed to only using memory information. 

Place, publisher, year, edition, pages
SciTePress, 2024. Vol. 1, p. 558-569
Keywords [en]
Cloud Workload Modeling, Co-Located Workloads, Time Series Forecasting, Recurrent Neural Networks.
National Category
Computer Systems
Research subject
Computer Science
Identifiers
URN: urn:nbn:se:kau:diva-99726DOI: 10.5220/0012392700003654Scopus ID: 2-s2.0-85190672119ISBN: 978-989-758-684-2 (electronic)OAI: oai:DiVA.org:kau-99726DiVA, id: diva2:1859591
Conference
13th International Conference on Pattern Recognition Applications and Methods, ICPRAM, Rome, Italy, February 24-26, 2024.
Available from: 2024-05-22 Created: 2024-05-22 Last updated: 2024-05-22Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Ramaswamy, ArunselvanFerlin, Simone

Search in DiVA

By author/editor
Ramaswamy, ArunselvanFerlin, Simone
By organisation
Department of Mathematics and Computer Science (from 2013)
Computer Systems

Search outside of DiVA

GoogleGoogle Scholar

doi
isbn
urn-nbn

Altmetric score

doi
isbn
urn-nbn
Total: 81 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • apa.csl
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf