[2509.03394] CloudFormer: An Attention-based Performance Prediction for Public Clouds with Unknown Workload
Nlp

[2509.03394] CloudFormer: An Attention-based Performance Prediction for Public Clouds with Unknown Workload

arXiv - Machine Learning 4 min read

About this article

Abstract page for arXiv paper 2509.03394: CloudFormer: An Attention-based Performance Prediction for Public Clouds with Unknown Workload

Computer Science > Distributed, Parallel, and Cluster Computing arXiv:2509.03394 (cs) [Submitted on 3 Sep 2025 (v1), last revised 25 Mar 2026 (this version, v2)] Title:CloudFormer: An Attention-based Performance Prediction for Public Clouds with Unknown Workload Authors:Amirhossein Shahbazinia, Darong Huang, Luis Costero, David Atienza View a PDF of the paper titled CloudFormer: An Attention-based Performance Prediction for Public Clouds with Unknown Workload, by Amirhossein Shahbazinia and 3 other authors View PDF HTML (experimental) Abstract:Cloud platforms are increasingly relied upon to host diverse, resource-intensive workloads due to their scalability, flexibility, and cost-efficiency. In multi-tenant cloud environments, virtual machines are consolidated on shared physical servers to improve resource utilization. While virtualization guarantees resource partitioning for CPU, memory, and storage, it cannot ensure performance isolation. Competition for shared resources such as last-level cache, memory bandwidth, and network interfaces often leads to severe performance degradation. Existing management techniques, including VM scheduling and resource provisioning, require accurate performance prediction to mitigate interference. However, this remains challenging in public clouds due to the black-box nature of VMs and the highly dynamic nature of workloads. To address these limitations, we propose CloudFormer, a dual-branch Transformer-based model designed to predict VM p...

Originally published on March 26, 2026. Curated by AI News.

Related Articles

Machine Learning

[D] Looking for definition of open-world ish learning problem

Hello! Recently I did a project where I initially had around 30 target classes. But at inference, the model had to be able to handle a lo...

Reddit - Machine Learning · 1 min ·
[2603.11687] SemBench: A Universal Semantic Framework for LLM Evaluation
Llms

[2603.11687] SemBench: A Universal Semantic Framework for LLM Evaluation

Abstract page for arXiv paper 2603.11687: SemBench: A Universal Semantic Framework for LLM Evaluation

arXiv - AI · 4 min ·
[2603.11583] UtilityMax Prompting: A Formal Framework for Multi-Objective Large Language Model Optimization
Llms

[2603.11583] UtilityMax Prompting: A Formal Framework for Multi-Objective Large Language Model Optimization

Abstract page for arXiv paper 2603.11583: UtilityMax Prompting: A Formal Framework for Multi-Objective Large Language Model Optimization

arXiv - AI · 3 min ·
[2512.05245] STAR-GO: Improving Protein Function Prediction by Learning to Hierarchically Integrate Ontology-Informed Semantic Embeddings
Machine Learning

[2512.05245] STAR-GO: Improving Protein Function Prediction by Learning to Hierarchically Integrate Ontology-Informed Semantic Embeddings

Abstract page for arXiv paper 2512.05245: STAR-GO: Improving Protein Function Prediction by Learning to Hierarchically Integrate Ontology...

arXiv - Machine Learning · 4 min ·
More in Nlp: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime