CLC number: TP39
On-line Access: 2023-01-21
Received: 2021-06-24
Revision Accepted: 2023-01-21
Crosschecked: 2021-11-02
Cited: 0
Clicked: 3160
Citations: Bibtex RefMan EndNote GB/T7714
Yuzhao WANG, Junqing YU, Zhibin YU. Resource scheduling techniques in cloud from a view of coordination: a holistic survey[J]. Frontiers of Information Technology & Electronic Engineering, 2023, 24(1): 1-40.
@article{title="Resource scheduling techniques in cloud from a view of coordination: a holistic survey",
author="Yuzhao WANG, Junqing YU, Zhibin YU",
journal="Frontiers of Information Technology & Electronic Engineering",
volume="24",
number="1",
pages="1-40",
year="2023",
publisher="Zhejiang University Press & Springer",
doi="10.1631/FITEE.2100298"
}
%0 Journal Article
%T Resource scheduling techniques in cloud from a view of coordination: a holistic survey
%A Yuzhao WANG
%A Junqing YU
%A Zhibin YU
%J Frontiers of Information Technology & Electronic Engineering
%V 24
%N 1
%P 1-40
%@ 2095-9184
%D 2023
%I Zhejiang University Press & Springer
%DOI 10.1631/FITEE.2100298
TY - JOUR
T1 - Resource scheduling techniques in cloud from a view of coordination: a holistic survey
A1 - Yuzhao WANG
A1 - Junqing YU
A1 - Zhibin YU
J0 - Frontiers of Information Technology & Electronic Engineering
VL - 24
IS - 1
SP - 1
EP - 40
%@ 2095-9184
Y1 - 2023
PB - Zhejiang University Press & Springer
ER -
DOI - 10.1631/FITEE.2100298
Abstract: Nowadays, the management of resource contention in shared cloud remains a pending problem. The evolution and deployment of new application paradigms (e.g., deep learning training and microservices) and custom hardware (e.g., graphics processing unit (GPU) and tensor processing unit (TPU)) have posed new challenges in resource management system design. Current solutions tend to trade cluster efficiency for guaranteed application performance, e.g., resource over-allocation, leaving a lot of resources underutilized. Overcoming this dilemma is not easy, because different components across the software stack are involved. Nevertheless, massive efforts have been devoted to seeking effective performance isolation and highly efficient resource scheduling. The goal of this paper is to systematically cover related aspects to deliver the techniques from the coordination perspective, and to identify the corresponding trends they indicate. Briefly, four topics are involved. First, isolation mechanisms deployed at different levels (micro-architecture, system, and virtualization levels) are reviewed, including GPU multitasking methods. Second, resource scheduling techniques within an individual machine and at the cluster level are investigated, respectively. Particularly, GPU scheduling for deep learning applications is described in detail. Third, adaptive resource management including the latest microservice-related research is thoroughly explored. Finally, future research directions are discussed in the light of advanced work. We hope that this review paper will help researchers establish a global view of the landscape of resource management techniques in shared cloud, and see technology trends more clearly.
[1]Achermann R, Panwar A, Bhattacharjee A, et al., 2020. Mitosis: transparently self-replicating page-tables for large-memory machines. Proc 25th Int Conf on Architectural Support for Programming Languages and Operating Systems, p.283-300.
[2]Akkus IE, Chen RC, Rimac I, et al., 2018. SAND: towards high-performance serverless computing. Proc USENIX Annual Technical Conf, p.923-935.
[3]Alibaba, 2020. Fuxi 2.0—The Core Dispatching System of Ali Economy Towards the Big Data and Cloud Computing Scheduling Challenge (in Chinese). https://developer.aliyun.com/article/760083 [Accessed on July 1, 2021].
[4]Ananthanarayanan G, Douglas C, Ramakrishnan R, et al., 2012. True elasticity in multi-tenant data-intensive compute clusters. Proc 3rd ACM Symp on Cloud Computing, p.1-7.
[5]Asmussen N, Völp M, Nöthen B, et al., 2016. M3: a hardware/operating-system co-design to tame heterogeneous manycores. Proc 21st Int Conf on Architectural Support for Programming Languages and Operating Systems, p.189-203.
[6]Ausavarungnirun R, Miller V, Landgraf J, et al., 2018. MASK: redesigning the GPU memory hierarchy to support multi-application concurrency. Proc 23rd Int Conf on Architectural Support for Programming Languages and Operating Systems, p.503-518.
[7]Bao YX, Peng YH, Wu C, 2019. Deep learning-based job placement in distributed machine learning clusters. Proc IEEE Conf on Computer Communications, p.505-513.
[8]Bauman E, Ayoade G, Lin ZQ, 2015. A survey on hypervisor-based monitoring: approaches, applications, and evolutions. ACM Comput Surv, 48(1):10.
[9]Baumann A, Barham P, Dagand PE, et al., 2009. The multi-kernel: a new OS architecture for scalable multicore systems. Proc ACM SIGOPS 22nd Symp on Operating Systems Principles, p.29-44.
[10]Berger DS, Berg B, Zhu T, et al., 2018. RobinHood: tail latency-aware caching—dynamically reallocating from cache-rich to cache-poor. Proc 13th USENIX Conf on Operating Systems Design and Implementation, p.195-212.
[11]Bhadauria M, McKee SA, 2010. An approach to resource-aware co-scheduling for CMPs. Proc 24th ACM Int Conf on Supercomputing, p.189-199.
[12]Bitirgen R, Ipek E, Martinez JF, 2008. Coordinated management of multiple interacting resources in chip multi-processors: a machine learning approach. Proc 41st IEEE/ACM Int Symp on Microarchitecture, p.318-329.
[13]Blagodurov S, Zhuravlev S, Fedorova A, et al., 2010. A case for NUMA-aware contention management on multicore systems. Proc 19th Int Conf on Parallel Architectures and Compilation Techniques, p.557-558.
[14]Boucher S, Kalia A, Andersen DG, et al., 2018. Putting the “micro” back in microservice. Proc USENIX Annual Technical Conf, p.645-650.
[15]Boutin E, Ekanayake J, Lin W, et al., 2014. Apollo: scalable and coordinated scheduling for cloud-scale computing. Proc 11th USENIX Symp on Operating Systems Design and Implementation, p.285-300.
[16]Cadden J, Unger T, Awad Y, et al., 2020. SEUSS: skip redundant paths to make serverless fast. Proc 15th European Conf on Computer Systems, p.1-15.
[17]Carastan-Santos D, de Camargo RY, 2017. Obtaining dynamic scheduling policies with simulation and machine learning. Proc Int Conf for High Performance Computing, Networking, Storage and Analysis, p.1-13.
[18]Carvalho M, Cirne W, Brasileiro F, et al., 2014. Long-term SLOs for reclaimed cloud computing resources. Proc ACM Symp on Cloud Computing, p.1-13.
[19]Castelló A, Peña AJ, Mayo R, 2018. Exploring the interoperability of remote GPGPU virtualization using rCUDA and directive-based programming models. J Supercomput, 74(11):5628-5642.
[20]Chandra D, Guo F, Kim S, et al., 2005. Predicting inter-thread cache contention on a chip multi-processor architecture. Proc 11th Int Symp on High-Performance Computer Architecture, p.340-351.
[21]Chaudhary S, Ramjee R, Sivathanu M, et al., 2020. Balancing efficiency and fairness in heterogeneous GPU clusters for deep learning. Proc 15th European Conf on Computer Systems, p.1-16.
[22]Chen L, Lingys J, Chen K, et al., 2018. AuTO: scaling deep reinforcement learning for datacenter-scale automatic traffic optimization. Proc Conf of the ACM Special Interest Group on Data Communication, p.191-205.
[23]Chen Q, Yang HL, Mars J, et al., 2016. Baymax: QoS awareness and increased utilization for non-preemptive accelerators in warehouse scale computers. Proc 21st Int Conf on Architectural Support for Programming Languages and Operating Systems, p.681-696.
[24]Chen Q, Yang HL, Guo MY, et al., 2017. Prophet: precise QoS prediction on non-preemptive accelerators to improve utilization in warehouse-scale computers. Proc 22nd Int Conf on Architectural Support for Programming Languages and Operating Systems, p.17-32.
[25]Chen Q, Wang ZN, Leng JW, et al., 2019. Avalon: towards QoS awareness and improved utilization through multi-resource management in datacenters. Proc ACM Int Conf on Supercomputing, p.272-283.
[26]Chen W, Rao J, Zhou XB, 2017. Preemptive, low latency datacenter scheduling via lightweight virtualization. Proc USENIX Annual Technical Conf, p.251-263.
[27]Cherkasova L, Gupta D, Vahdat A, 2007. Comparison of the three CPU schedulers in Xen. ACM SIGMETRICS Perform Eval Rev, 35(2):42-51.
[28]Cho S, Jin L, 2006. Managing distributed, shared L2 caches through OS-level page allocation. Proc 39th Annual IEEE/ACM Int Symp on Microarchitecture, p.455-468.
[29]Curino C, Difallah DE, Douglas C, et al., 2014. Reservation-based scheduling: if you’re late don’t blame us! Proc ACM Symp on Cloud Computing, p.1-14.
[30]Dai GH, Huang TH, Chi YZ, et al., 2017. ForeGraph: exploring large-scale graph processing on multi-FPGA architecture. Proc ACM/SIGDA Int Symp on Field-Programmable Gate Arrays, p.217-226.
[31]Dean J, Barroso LA, 2013. The tail at scale. Commun ACM, 56(2):74-80.
[32]Delgado P, Dinu F, Kermarrec AM, et al., 2015. Hawk: hybrid datacenter scheduling. Proc USENIX Annual Technical Conf, p.499-510.
[33]Delgado P, Didona D, Dinu F, et al., 2016. Job-aware scheduling in Eagle: divide and stick to your probes. Proc 7th ACM Symp on Cloud Computing, p.497-509.
[34]Delimitrou C, Kozyrakis C, 2014. Quasar: resource-efficient and QoS-aware cluster management. Proc 19th Int Conf on Architectural Support for Programming Languages and Operating Systems, p.127-144.
[35]Delimitrou C, Kozyrakis C, 2016. HCloud: resource-efficient provisioning in shared cloud systems. Proc 21st Int Conf on Architectural Support for Programming Languages and Operating Systems, p.473-488.
[36]Delimitrou C, Sanchez D, Kozyrakis C, 2015. Tarcil: reconciling scheduling speed and quality in large shared clusters. Proc 6th ACM Symp on Cloud Computing, p.97-110.
[37]Dhakal A, Kulkarni SG, Ramakrishnan KK, 2020. GSLICE: controlled spatial sharing of GPUs for a scalable inference platform. Proc 11th ACM Symp on Cloud Computing, p.492-506.
[38]Ebrahimi E, Lee CJ, Mutlu O, et al., 2010. Fairness via source throttling: a configurable and high-performance fairness substrate for multi-core memory systems. Proc 15th Int Conf on Architectural Support for Programming Languages and Operating Systems, p.335-346.
[39]Engler DR, Kaashoek MF, O’Toole J, 1995. Exokernel: an operating system architecture for application-level resource management. Proc 15th ACM Symp on Operating Systems Principles, p.251-266.
[40]Eyerman S, Eeckhout L, 2010. Probabilistic job symbiosis modeling for SMT processor scheduling. Proc 15th Int Conf on Architectural Support for Programming Languages and Operating Systems, p.91-102.
[41]Facebook, 2015. Facebook Disaggregated Rack. http://goo.gl/6h2Ut [Accessed on July 1, 2021].
[42]Feliu J, Sahuquillo J, Petit S, et al., 2013. L1-bandwidth aware thread allocation in multicore SMT processors. Proc 22nd Int Conf on Parallel Architectures and Compilation Techniques, p.123-132.
[43]Feliu J, Eyerman S, Sahuquillo J, et al., 2016. Symbiotic job scheduling on the IBM POWER8. Proc IEEE Int Symp on High Performance Computer Architecture, p.669-680.
[44]Firestone D, Putnam A, Mundkur S, et al., 2018. Azure accelerated networking: smartnics in the public cloud. Proc 15th USENIX Symp on Networked Systems Design and Implementation, p.51-66.
[45]Fowers J, Ovtcharov K, Papamichael M, et al., 2018. A configurable cloud-scale DNN processor for real-time AI. Proc ACM/IEEE 45th Annual Int Symp on Computer Architecture, p.1-14.
[46]Gan Y, Zhang YQ, Cheng DL, et al., 2019a. An open-source benchmark suite for microservices and their hardware-software implications for cloud & edge systems. Proc 24th Int Conf on Architectural Support for Programming Languages and Operating Systems, p.3-18.
[47]Gan Y, Zhang YQ, Hu K, et al., 2019b. Seer: leveraging big data to navigate the complexity of performance debugging in cloud microservices. Proc 24th Int Conf on Architectural Support for Programming Languages and Operating Systems, p.19-33.
[48]Gan Y, Liang MY, Dev S, et al., 2021. Sage: practical and scalable ML-driven performance debugging in microservices. Proc 26th ACM Int Conf on Architectural Support for Programming Languages and Operating Systems, p.135-151.
[49]Giceva J, 2016. Database/Operating System Co-design. PhD Thesis, ETH Zurich, Switzerland.
[50]Goder A, Spiridonov A, Wang Y, 2015. Bistro: scheduling data-parallel jobs against live production systems. Proc USENIX Annual Technical Conf, p.459-471.
[51]Gog I, Schwarzkopf M, Gleave A, et al., 2016. Firmament: fast, centralized cluster scheduling at scale. Proc 12th USENIX Conf on Operating Systems Design and Implementation, p.99-115.
[52]Goglin B, Furmento N, 2009. Enabling high-performance memory migration for multithreaded applications on LINUX. Proc IEEE Int Symp on Parallel & Distributed Processing, p.1-9.
[53]Grandl R, Ananthanarayanan G, Kandula S, et al., 2014. Multi-resource packing for cluster schedulers. Proc ACM Conf on SIGCOMM, p.455-466.
[54]Grandl R, Chowdhury M, Akella A, et al., 2016a. Altruistic scheduling in multi-resource clusters. Proc 12th USENIX Conf on Operating Systems Design and Implementation, p.65-80.
[55]Grandl R, Kandula S, Rao S, et al., 2016b. Graphene: packing and dependency-aware scheduling for data-parallel clusters. Proc 12th USENIX Conf on Operating Systems Design and Implementation, p.81-97.
[56]Grulich PM, Nawab F, 2018. Collaborative edge and cloud neural networks for real-time video processing. Proc VLDB Endow, 11(12):2046-2049.
[57]Gu JC, Chowdhury M, Shin KG, et al., 2019. Tiresias: a GPU cluster manager for distributed deep learning. Proc 16th USENIX Symp on Networked Systems Design and Implementation, p.485-500.
[58]Guo F, Li YK, Lui JCS, et al., 2019. DCUDA: dynamic GPU scheduling with live migration support. Proc ACM Symp on Cloud Computing, p.114-125.
[59]Gysi T, Bär J, Hoefler T, 2016. dCUDA: hardware supported overlap of computation and communication. Proc Int Conf for High Performance Computing, Networking, Storage and Analysis, p.609-620.
[60]Han J, Jeon S, Choi YR, et al., 2016. Interference management for distributed parallel applications in consolidated clusters. Proc 21st Int Conf on Architectural Support for Programming Languages and Operating Systems, p.443-456.
[61]Haque E, Eom YH, He YX, et al., 2015. Few-to-Many: incremental parallelism for reducing tail latency in interactive services. Proc 20th Int Conf on Architectural Support for Programming Languages and Operating Systems, p.161-175.
[62]Herbst NR, Kounev S, Reussner R, 2013. Elasticity in cloud computing: what it is, and what it is not. Proc 10th Int Conf on Autonomic Computing, p.23-27.
[63]Hindman B, Konwinski A, Zaharia M, et al., 2011. Mesos: a platform for fine-grained resource sharing in the data center. Proc 8th USENIX Conf on Networked Systems Design and Implementation, p.295-308.
[64]Hong CH, Spence I, Nikolopoulos DS, 2017. GPU virtualization and scheduling methods: a comprehensive survey. ACM Comput Surv, 50(3):35.
[65]Hou XF, Li C, Liu JC, et al., 2020. ANT-Man: towards agile power management in the microservice era. Proc Int Conf for High Performance Computing, Networking, Storage and Analysis, Article 78.
[66]Hsu CH, Zhang YQ, Laurenzano MA, et al., 2015. Adrenaline: pinpointing and reining in tail queries with quick voltage boosting. Proc IEEE 21st Int Symp on High Performance Computer Architecture, p.271-282.
[67]Hu ZM, Tu J, Li BC, 2019. Spear: optimized dependency-aware task scheduling with deep reinforcement learning. Proc IEEE 39th Int Conf on Distributed Computing Systems, p.2037-2046.
[68]Ibanez S, Shahbaz M, McKeown N, 2019. The case for a network fast path to the CPU. Proc 18th ACM Workshop on Hot Topics in Networks, p.52-59.
[69]Intel, 2016. Intel Cache Allocation Technique. https://software.intel.com/en-us/articles/introduction-to-cache-allocation-technology [Accessed on July 1, 2021].
[70]Isard M, Prabhakaran V, Currey J, et al., 2009. Quincy: fair scheduling for distributed computing clusters. Proc ACM SIGOPS 22nd Symp on Operating Systems Principles, p.261-276.
[71]Islam S, Venugopal S, Liu AN, 2015. Evaluating the impact of fine-scale burstiness on cloud elasticity. Proc 6th ACM Symp on Cloud Computing, p.250-261.
[72]Jeon M, He YX, Kim H, et al., 2016. TPC: target-driven parallelism combining prediction and correction to reduce tail latency in interactive services. Proc 21st Int Conf on Architectural Support for Programming Languages and Operating Systems, p.129-141.
[73]Jeon M, Venkataraman S, Phanishayee A, et al., 2018. Multi-Tenant GPU Clusters for Deep Learning Workloads: Analysis and Implications. Technical Report No. MSR-TR-2018-13, Microsoft Research, USA.
[74]Jeon M, Venkataraman S, Phanishayee A, et al., 2019. Analysis of large-scale multi-tenant GPU clusters for DNN training workloads. Proc USENIX Annual Technical Conf, p.947-960.
[75]Jeong EY, Woo S, Jamshed M, et al., 2014. mTCP: a highly scalable user-level TCP stack for multicore systems. Proc 11th USENIX Conf on Networked Systems Design and Implementation, p.489-502.
[76]Jeyakumar V, Alizadeh M, Mazières D, et al., 2013. EyeQ: practical network performance isolation at the edge. Proc 10th USENIX Symp on Networked Systems Design and Implementation, p.297-311.
[77]Jia ZP, Witchel E, 2021. Nightcore: efficient and scalable serverless computing for latency-sensitive, interactive microservices. Proc 26th ACM Int Conf on Architectural Support for Programming Languages and Operating Systems, p.152-166.
[78]Jyothi SA, Curino C, Menache I, et al., 2016. Morpheus: towards automated SLOs for enterprise clusters. Proc 12th USENIX Conf on Operating Systems Design and Implementation, p.117-134.
[79]Kakivaya G, Xun L, Hasha R, et al., 2018. Service fabric: a distributed platform for building microservices in the cloud. Proc 13th EuroSys Conf, Article 33.
[80]Kalia A, Kaminsky M, Andersen DG, 2016. FaSST: fast, scalable and simple distributed transactions with two-sided (RDMA) datagram RPCs. Proc 12th USENIX Conf on Operating Systems Design and Implementation, p.185-201.
[81]Kalia A, Kaminsky M, Andersen D, 2019. Datacenter RPCs can be general and fast. Proc 16th USENIX Symp on Networked Systems Design and Implementation, p.1-16.
[82]Kang YP, Hauswald J, Gao C, et al., 2017. Neurosurgeon: collaborative intelligence between the cloud and mobile edge. Proc 22nd Int Conf on Architectural Support for Programming Languages and Operating Systems, p.615-629.
[83]Kannan RS, Subramanian L, Raju A, et al., 2019. GrandSLAm: guaranteeing SLAs for jobs in microservices execution frameworks. Proc 14th EuroSys Conf, Article 34.
[84]Kannan S, Gavrilovska A, Gupta V, et al., 2017. HeteroOS: OS design for heterogeneous memory management in datacenter. Proc 44th Annual Int Symp on Computer Architecture, p.521-534.
[85]Kannan S, Ren YJ, Bhattacharjee A, 2021. KLOCs: kernel-level object contexts for heterogeneous memory systems. Proc 26th ACM Int Conf on Architectural Support for Programming Languages and Operating Systems, p.65-78.
[86]Kapoor R, Porter G, Tewari M, et al., 2012. Chronos: predictable low latency for data center applications. Proc 3rd ACM Symp on Cloud Computing, Article 9.
[87]Karanasos K, Rao S, Curino C, et al., 2015. Mercury: hybrid centralized and distributed scheduling in large shared clusters. Proc USENIX Annual Technical Conf, p.485-497.
[88]Kasture H, Sanchez D, 2014. Ubik: efficient cache sharing with strict QoS for latency-critical workloads. Proc 19th Int Conf on Architectural Support for Programming Languages and Operating Systems, p.729-742.
[89]Khawaja A, Landgraf J, Prakash R, et al., 2018. Sharing, protection, and compatibility for reconfigurable fabric with AMORPHOS. Proc 13th USENIX Conf on Operating Systems Design and Implementation, p.107-127.
[90]Khorasani F, Esfeden HA, Farmahini-Farahani A, et al., 2018. RegMutex: inter-warp GPU register time-sharing. Proc ACM/IEEE 45th Annual Int Symp on Computer Architecture, p.816-828.
[91]Klimovic A, Kozyrakis C, Thereska E, et al., 2016. Flash storage disaggregation. Proc 11th European Conf on Computer Systems, Article 29.
[92]Knauerhase R, Brett P, Hohlt B, et al., 2008. Using OS observations to improve performance in multicore systems. IEEE Micro, 28(3):54-66.
[93]Korolija D, Roscoe T, Alonso G, 2020. Do OS abstractions make sense on FPGAs? Proc 14th USENIX Symp on Operating Systems Design and Implementation, p.991-1010.
[94]Kotra JB, Zhang HB, Alameldeen AR, et al., 2018. CHAMELEON: a dynamically reconfigurable heterogeneous memory system. Proc 51st Annual IEEE/ACM Int Symp on Microarchitecture, p.533-545.
[95]Lazarev N, Xiang SJ, Adit N, et al., 2021. Dagger: efficient and fast RPCs in cloud microservices with near-memory reconfigurable NICs. Proc 26th ACM Int Conf on Architectural Support for Programming Languages and Operating Systems, p.36-51.
[96]Le TN, Sun X, Chowdhury M, et al., 2020. AlloX: compute allocation in hybrid clusters. Proc 15th European Conf on Computer Systems, Article 31.
[97]Le YF, Chang H, Mukherjee S, et al., 2017. UNO: uniflying host and smart NIC offload for flexible packet processing. Proc Symp on Cloud Computing, p.506-519.
[98]Li CL, Andersen DG, Fu Q, et al., 2017. Workload analysis and caching strategies for search advertising systems. Proc Symp on Cloud Computing, p.170-180.
[99]Li J, Agrawal K, Elnikety S, et al., 2016. Work stealing for interactive services to meet target latency. Proc 21st ACM SIGPLAN Symp on Principles and Practice of Parallel Programming, Article 14.
[100]Li JL, Sharma NK, Ports DRK, et al., 2014. Tales of the tail: hardware, OS, and application-level sources of tail latency. Proc ACM Symp on Cloud Computing, p.1-14.
[101]Lim K, Chang JC, Mudge T, et al., 2009. Disaggregated memory for expansion and sharing in blade servers. Proc 36th Annual Int Symp on Computer Architecture, p.267-278.
[102]Linux Community, 2016. Linux Kernel Namespace. https://en.wikipedia.org/wiki/Linux_namespaces [Accessed on Feb. 23, 2021].
[103]Liu M, Peter S, Krishnamurthy A, et al., 2019. E3: energy-efficient microservices on SmartNIC-accelerated servers. Proc USENIX Annual Technical Conf, p.363-378.
[104]Lo D, Cheng LQ, Govindaraju R, et al., 2015. Heracles: improving resource efficiency at scale. Proc 42nd Annual Int Symp on Computer Architecture, p.450-462.
[105]Luo QY, Lin JK, Zhuo YW, et al., 2019. Hop: heterogeneity-aware decentralized training. Proc 24th Int Conf on Architectural Support for Programming Languages and Operating Systems, p.893-907.
[106]Ma JC, Zuo GF, Loughlin K, et al., 2020. A hypervisor for shared-memory FPGA platforms. Proc 25th Int Conf on Architectural Support for Programming Languages and Operating Systems, p.827-844.
[107]Madhavapeddy A, Scott DJ, 2014. Unikernels: the rise of the virtual library operating system. Commun ACM, 57(1):61-69.
[108]Mahajan K, Balasubramanian A, Singhvi A, et al., 2020. Themis: fair and efficient GPU cluster scheduling. Proc 17th USENIX Symp on Networked Systems Design and Implementation, p.289-304.
[109]Manco F, Lupu C, Schmidt F, et al., 2017. My VM is lighter (and safer) than your container. Proc 26th Symp on Operating Systems Principles, p.218-233.
[110]Mao HZ, Alizadeh M, Menache I, et al., 2016. Resource management with deep reinforcement learning. Proc 15th ACM Workshop on Hot Topics in Networks, p.50-56.
[111]Mao HZ, Schwarzkopf M, Venkatakrishnan SB, et al., 2019. Learning scheduling algorithms for data processing clusters. Proc Special Interest Group on Data Communication, p.270-288.
[112]Mars J, Tang LJ, 2013. Whare-Map: heterogeneity in “homogeneous” warehouse-scale computers. Proc 40th Annual Int Symp on Computer Architecture, p.619-630.
[113]Min C, Kang W, Kumar M, et al., 2018. Solros: a data-centric operating system architecture for heterogeneous computing. Proc 13th EuroSys Conf, Article 36.
[114]Moon Y, Lee S, Jamshed MA, et al., 2020. AccelTCP: accelerating network applications with stateful TCP offloading. Proc 17th USENIX Symp on Networked Systems Design and Implementation, p.77-92.
[115]Moritz P, Nishihara R, Wang S, et al., 2018. Ray: a distributed framework for emerging AI applications. Proc 13th USENIX Conf on Operating Systems Design and Implementation, p.561-577.
[116]Multicluster Special Interest Group, 2020. Kubernetes Multicluster. https://github.com/kubernetes/community/tree/master/sigmulticluster [Accessed on July 1, 2021].
[117]Mutlu O, Moscibroda T, 2008. Parallelism-aware batch scheduling: enhancing both performance and fairness of shared DRAM systems. Proc Int Symp on Computer Architecture, p.63-74.
[118]Nagaraj K, Bharadia D, Mao HZ, et al., 2016. NUMFabric: fast and flexible bandwidth allocation in datacenters. Proc ACM SIGCOMM Conf, p.188-201.
[119]Narayanan D, Santhanam K, Kazhamiaka F, et al., 2020. Heterogeneity-aware cluster scheduling policies for deep learning workloads. Proc 14th USENIX Symp on Operating Systems Design and Implementation, p.481-498.
[120]Nightingale EB, Hodson O, McIlroy R, et al., 2009. Helios: heterogeneous multiprocessing with satellite kernels. Proc ACM SIGOPS 22nd Symp on Operating Systems Principles, p.221-234.
[121]Novaković D, Vasić N, Novaković S, et al., 2013. DeepDive: transparently identifying and managing performance interference in virtualized environments. Proc USENIX Annual Technical Conf, p.219-230.
[122]Novaković S, Daglis A, Bugnion E, et al., 2014. Scale-out NUMA. Proc 19th Int Conf on Architectural Support for Programming Languages and Operating Systems, p.3-18.
[123]Ousterhout K, Wendell P, Zaharia M, et al., 2013. Sparrow: distributed, low latency scheduling. Proc 24th ACM Symp on Operating Systems Principles, p.69-84.
[124]Ousterhout K, Canel C, Ratnasamy S, et al., 2017. Monotasks: architecting for performance clarity in data analytics frameworks. Proc 26th Symp on Operating Systems Principles, p.184-200.
[125]Panda A, Zheng WT, Hu XH, et al., 2017. SCL: simplifying distributed SDN control planes. Proc 14th USENIX Symp on Networked Systems Design and Implementation, p.329-345.
[126]Park JJK, Park Y, Mahlke S, 2015. Chimera: collaborative preemption for multitasking on a shared GPU. Proc 20th Int Conf on Architectural Support for Programming Languages and Operating Systems, p.593-606.
[127]Peng X, Shi XH, Dai HL, et al., 2020. Capuchin: tensor-based GPU memory management for deep learning. Proc 25th Int Conf on Architectural Support for Programming Languages and Operating Systems, p.891-905.
[128]Peng YH, Bao YX, Chen YR, et al., 2018. Optimus: an efficient dynamic resource scheduler for deep learning clusters. Proc 20th EuroSys Conf, Article 3.
[129]Popov M, Jimborean A, Black-Schaffer D, 2019. Efficient thread/page/parallelism autotuning for NUMA systems. Proc ACM Int Conf on Supercomputing, p.342-353.
[130]Pothukuchi RP, Greathouse JL, Rao K, et al., 2019. Tangram: integrated control of heterogeneous computers. Proc 52nd Annual IEEE/ACM Int Symp on Microarchitecture, p.384-398.
[131]Pratheek B, Jawalkar N, Basu A, 2021. Improving GPU multi-tenancy with page walk stealing. Proc IEEE Int Symp on High-Performance Computer Architecture, p.626-639.
[132]Qiu HR, Banerjee SS, Jha S, et al., 2020. FIRM: an intelligent fine-grained resource management framework for SLO-oriented microservices. Proc 14th USENIX Symp on Operating Systems Design and Implementation, p.805-825.
[133]Qureshi MK, Patt YN, 2006. Utility-based cache partitioning: a low-overhead, high-performance, runtime mechanism to partition shared caches. Proc 39th Annual IEEE/ACM Int Symp on Microarchitecture, p.423-432.
[134]Rao J, Wang K, Zhou XB, et al., 2013. Optimizing virtual machine scheduling in NUMA multicore systems. Proc IEEE 19th Int Symp on High Performance Computer Architecture, p.306-317.
[135]Reiss C, Tumanov A, Ganger GR, et al., 2012. Heterogeneity and dynamicity of clouds at scale: Google trace analysis. Proc 3rd ACM Symp on Cloud Computing, Article 7.
[136]Rhu M, Gimelshein N, Clemons J, et al., 2016. vDNN: virtualized deep neural networks for scalable, memory-efficient neural network design. Proc 49th Annual IEEE/ACM Int Symp on Microarchitecture, p.1-13.
[137]Rossbach CJ, Currey J, Silberstein M, et al., 2011. PTask: operating system abstractions to manage GPUs as compute devices. Proc 23rd ACM Symp on Operating Systems Principles, p.233-248.
[138]Sanchez D, Kozyrakis C, 2011. Vantage: scalable and efficient fine-grain cache partitioning. Proc 38th Annual Int Symp on Computer Architecture, p.57-68.
[139]Schwarzkopf M, Konwinski A, Abd-El-Malek M, et al., 2013. Omega: flexible, scalable schedulers for large compute clusters. Proc 8th ACM European Conf on Computer Systems, p.351-364.
[140]Sengupta D, Belapure R, Schwan K, 2013. Multi-tenancy on GPGPU-based servers. Proc 7th Int Workshop on Virtualization Technologies in Distributed Computing, p.3-10.
[141]Sengupta D, Goswami A, Schwan K, et al., 2014. Scheduling multi-tenant cloud workloads on accelerator-based systems. Proc Int Conf for High Performance Computing, Networking, Storage and Analysis, p.513-524.
[142]Shan YZ, Huang YT, Chen YL, et al., 2018. LegoOS: a disseminated, distributed OS for hardware resource disaggregation. Proc 13th USENIX Conf on Operating Systems Design and Implementation, p.69-87.
[143]Sharma NK, Zhao CXY, Liu M, et al., 2020. Programmable calendar queues for high-speed packet scheduling. Proc 17th USENIX Symp on Networked Systems Design and Implementation, p.685-699.
[144]Sharma P, Guo T, He X, et al., 2016. Flint: batch-interactive data-intensive processing on transient servers. Proc 11th European Conf on Computer Systems, Article 6.
[145]Shen ZM, Sun Z, Sela GE, et al., 2019. X-Containers: breaking down barriers to improve performance and isolation of cloud-native containers. Proc 24th Int Conf on Architectural Support for Programming Languages and Operating Systems, p.121-135.
[146]Shillaker S, Pietzuch P, 2020. Faasm: lightweight isolation for efficient stateful serverless computing. Proc USENIX Annual Technical Conf, p.419-433.
[147]Sigelman BH, Barroso LA, Burrows M, et al., 2010. Dapper, a Large-Scale Distributed Systems Tracing Infrastructure. https://storage.googleapis.com/pub-tools-public-publication-data/pdf/36356.pdf [Accessed on July 1, 2021].
[148]Singh S, Chana I, 2016. A survey on resource scheduling in cloud computing: issues and challenges. J Grid Comput, 14(2):217-264.
[149]Snavely A, Tullsen DM, 2000. Symbiotic jobscheduling for a simultaneous multithreaded processor. ACM SIGOPS Oper Syst Rev, 34(5):234-244.
[150]Song X, Shi JC, Chen HB, et al., 2013. Schedule processes, not VCPUs. Proc 4th Asia-Pacific Workshop on Systems, p.1-7.
[151]Sriraman A, Dhanotia A, 2020. Accelerometer: understanding acceleration opportunities for data center overheads at hyperscale. Proc 25th Int Conf on Architectural Support for Programming Languages and Operating Systems, p.733-750.
[152]Sriraman A, Wenisch TF, 2018. μTune: auto-tuned threading for OLDI microservices. Proc 13th USENIX Conf on Operating Systems Design and Implementation, p.177-194.
[153]Sriraman A, Dhanotia A, Wenisch TF, 2019. SoftSKU: optimizing server architectures for microservice diversity @scale. Proc 46th Int Symp on Computer Architecture, p.513-526.
[154]Staples G, 2006. TORQUE resource manager. Proc ACM/IEEE Conf on Supercomputing.
[155]Subramanian L, Seshadri V, Ghosh A, et al., 2015. The application slowdown model: quantifying and controlling the impact of inter-application interference at shared caches and main memory. Proc 48th Int Symp on Microarchitecture, p.62-75.
[156]Tanasic I, Gelado I, Cabezas J, et al., 2014. Enabling preemptive multiprogramming on GPUs. Proc ACM/IEEE 41st Int Symp on Computer Architecture, p.193-204.
[157]Tang CQ, Yu K, Veeraraghavan K, et al., 2020. Twine: a unified cluster management system for shared infrastructure. Proc 14th USENIX Symp on Operating Systems Design and Implementation, p.787-803.
[158]Tang LJ, Mars J, Vachharajani N, et al., 2011. The impact of memory subsystem resource sharing on datacenter applications. Proc 38th Annual Int Symp on Computer Architecture, p.283-294.
[159]Tembey P, Gavrilovska A, Schwan K, 2014. Merlin: application- and platform-aware resource allocation in consolidated server systems. Proc ACM Symp on Cloud Computing, p.1-14.
[160]Thinakaran P, Gunasekaran JR, Sharma B, et al., 2017. Phoenix: a constraint-aware scheduler for heterogeneous datacenters. Proc IEEE 37th Int Conf on Distributed Computing Systems, p.977-987.
[161]Tirmazi M, Barker A, Deng N, et al., 2020. Borg: the next generation. Proc 15th European Conf on Computer Systems, Article 30.
[162]Tumanov A, Zhu T, Park JW, et al., 2016. TetriSched: global rescheduling with adaptive plan-ahead in dynamic heterogeneous clusters. Proc 11th European Conf on Computer Systems, Article 35.
[163]Vanga M, Gujarati A, Brandenburg BB, 2018. Tableau: a high-throughput and predictable VM scheduler for high-density workloads. Proc 13th EuroSys Conf, Article 28.
[164]Vasić N, Novaković D, Miučin S, et al., 2012. DejaVu: accelerating resource allocation in virtualized environments. Proc 17th Int Conf on Architectural Support for Programming Languages and Operating Systems, p.423-436.
[165]Vavilapalli VK, Murthy AC, Douglas C, et al., 2013. Apache Hadoop YARN: yet another resource negotiator. Proc 4th Annual Symp on Cloud Computing, Article 5.
[166]Verma A, Pedrosa L, Korupolu M, et al., 2015. Large-scale cluster management at Google with Borg. Proc 10th European Conf on Computer Systems, Article 18.
[167]Vulimiri A, Curino C, Godfrey PB, et al., 2015. Wanaly-tics: geo-distributed analytics for a data intensive world. Proc ACM SIGMOD Int Conf on Management of Data, p.1087-1092.
[168]Wang JJ, Balazinska M, 2017. Elastic memory management for cloud data analytics. Proc USENIX Annual Technical Conf, p.745-758.
[169]Wang JY, Pan JL, Esposito F, et al., 2019. Edge cloud offloading algorithms: issues, methods, and perspectives. ACM Comput Surv, 52(1):2.
[170]Wang LN, Ye JM, Zhao YM, et al., 2018. SuperNeurons: dynamic GPU memory management for training deep neural networks. Proc 23rd ACM SIGPLAN Symp on Principles and Practice of Parallel Programming, p.41-53.
[171]Wang LP, Weng QZ, Wang W, et al., 2020. Metis: learning to schedule long-running applications in shared container clusters at scale. Proc Int Conf for High Performance Computing, Networking, Storage and Analysis, Article 68.
[172]Wang SQ, Gonzalez OJ, Zhou XB, et al., 2020. An efficient and non-intrusive GPU scheduling framework for deep learning training systems. Proc Int Conf for High Performance Computing, Networking, Storage and Analysis, Article 90.
[173]Wang ZN, Yang J, Melhem R, et al., 2016. Simultaneous multikernel GPU: multi-tasking throughput processors via fine-grained sharing. Proc IEEE Int Symp on High Performance Computer Architecture, p.358-369.
[174]Weerasiri D, Barukh MC, Benatallah B, et al., 2017. A taxonomy and survey of cloud resource orchestration techniques. ACM Comput Surv, 50(2):26.
[175]Williams D, Koller R, 2016. Unikernel monitors: extending minimalism outside of the box. Proc 8th USENIX Workshop on Hot Topics in Cloud Computing, p.1-6.
[176]Xiao WC, Bhardwaj R, Ramjee R, et al., 2018. Gandiva: introspective cluster scheduling for deep learning. Proc 13th USENIX Conf on Operating Systems Design and Implementation, p.595-610.
[177]Xiao WX, Ren SR, Li Y, et al., 2020. AntMan: dynamic scaling on GPU clusters for deep learning. Proc 14th USENIX Symp on Operating Systems Design and Implementation, p.533-548.
[178]Xu QM, Jeon H, Kim K, et al., 2016. Warped-Slicer: efficient intra-SM slicing through dynamic resource partitioning for GPU multiprogramming. Proc ACM/IEEE 43rd Annual Int Symp on Computer Architecture, p.230-242.
[179]Xu YJ, Musgrave Z, Noble B, et al., 2013. Bobtail: avoiding long tails in the cloud. Proc 10th USENIX Symp on Networked Systems Design and Implementation, p.329-341.
[180]Yan Y, Gao YJ, Chen Y, et al., 2016. TR-Spark: transient computing for big data analytics. Proc 7th ACM Symp on Cloud Computing, p.484-496.
[181]Yang HL, Breslow A, Mars J, et al., 2013. Bubble-Flux: precise online QoS management for increased utilization in warehouse scale computers. Proc 40th Annual Int Symp on Computer Architecture, p.607-618.
[182]Yang X, Blackburn SM, McKinley KS, 2016. Elfen scheduling: fine-grain principled borrowing from latency-critical workloads using simultaneous multithreading. Proc USENIX Annual Technical Conf, p.309-322.
[183]Yang Y, Kim GW, Song WW, et al., 2017. Pado: a data processing engine for harnessing transient resources in datacenters. Proc 12th European Conf on Computer Systems, p.575-588.
[184]Yeh TT, Sabne A, Sakdhnagool P, et al., 2017. Pagoda: fine-grained GPU resource virtualization for narrow tasks. Proc 22nd ACM SIGPLAN Symp on Principles and Practice of Parallel Programming, p.221-234.
[185]Yeh TT, Sinclair MD, Beckmann BM, et al., 2021. Deadline-aware offloading for high-throughput accelerators. Proc IEEE Int Symp on High-Performance Computer Architecture, p.479-492.
[186]Zellweger G, Gerber S, Kourtis K, et al., 2014. Decoupling cores, kernels, and operating systems. Proc 11th USENIX Symp on Operating Systems Design and Implementation, p.17-31.
[187]Zha Y, Li J, 2020. Virtualizing FPGAs in the cloud. Proc 25th Int Conf on Architectural Support for Programming Languages and Operating Systems, p.845-858.
[188]Zha Y, Li J, 2021. When application-specific ISA meets FPGAs: a multi-layer virtualization framework for heterogeneous cloud FPGAs. Proc 26th ACM Int Conf on Architectural Support for Programming Languages and Operating Systems, p.123-134.
[189]Zhang D, Dai D, He YB, et al., 2020. RLScheduler: an automated HPC batch job scheduler using reinforcement learning. Proc Int Conf for High Performance Computing, Networking, Storage and Analysis, p.1-15.
[190]Zhang JS, Xiong YQ, Xu NY, et al., 2017. The Feniks FPGA operating system for cloud computing. Proc 8th Asia-Pacific Workshop on Systems, Article 22.
[191]Zhang X, Dwarkadas S, Shen K, 2009. Towards practical page coloring-based multicore cache management. Proc 4th ACM European Conf on Computer Systems, p.89-102.
[192]Zhang X, Tune E, Hagmann R, et al., 2013. CPI2: CPU performance isolation for shared compute clusters. Proc 8th ACM European Conf on Computer Systems, p.379-391.
[193]Zhang XT, Zheng X, Wang Z, et al., 2019. Fast and scalable VMM live upgrade in large cloud infrastructure. Proc 24th Int Conf on Architectural Support for Programming Languages and Operating Systems, p.93-105.
[194]Zhang YQ, Laurenzano MA, Mars J, et al., 2014. SMiTe: precise QoS prediction on real-system SMT processors to improve utilization in warehouse scale computers. Proc 47th Annual IEEE/ACM Int Symp on Microarchitecture, p.406-418.
[195]Zhang YQ, Prekas G, Fumarola GM, et al., 2016. History-based harvesting of spare cycles and storage in large-scale datacenters. Proc 12th USENIX Conf on Operating Systems Design and Implementation, p.755-770.
[196]Zhang YQ, Hua WZ, Zhou ZZ, et al., 2021. Sinan: ML-based and QoS-aware resource management for cloud microservices. Proc 26th ACM Int Conf on Architectural Support for Programming Languages and Operating Systems, p.167-181.
[197]Zhao HY, Han ZH, Yang Z, et al., 2020. HiveD: sharing a GPU cluster for deep learning with guarantees. Proc 14th USENIX Symp on Operating Systems Design and Implementation, p.515-532.
[198]Zhao M, Cabrera J, 2018. RTVirt: enabling time-sensitive computing on virtualized systems through cross-layer CPU scheduling. Proc 13th EuroSys Conf, Article 27.
[199]Zheng L, Li XL, Zheng YH, et al., 2020. Scaph: scalable GPU-accelerated graph processing with value-driven differential scheduling. Proc USENIX Annual Technical Conf, p.573-588.
[200]Zhou H, Chen M, Lin Q, et al., 2018. Overload control for scaling WeChat microservices. Proc ACM Symp on Cloud Computing, p.149-161.
[201]Zhou ZY, Benson TA, 2019. Composing SDN controller enhancements with Mozart. Proc ACM Symp on Cloud Computing, p.351-363.
[202]Zhu H, Kaffes K, Chen ZX, et al., 2020. RackSched: a microsecond-scale scheduler for rack-scale computers. Proc 14th USENIX Symp on Operating Systems Design and Implementation, p.1225-1240.
[203]Zhu HS, Erez M, 2016. Dirigent: enforcing QoS for latency-critical tasks on shared multicore systems. Proc 21st Int Conf on Architectural Support for Programming Languages and Operating Systems, p.33-47.
[204]Zhu T, Kozuch MA, Harchol-Balter M, 2017. WorkloadCompactor: reducing datacenter cost while providing tail latency SLO guarantees. Proc Symp on Cloud Computing, p.598-610.
[205]Zhuravlev S, Blagodurov S, Fedorova A, 2010. Addressing shared resource contention in multicore processors via scheduling. Proc 15th Int Conf on Architectural Support for Programming Languages and Operating Systems, p.129-142.
Open peer comments: Debate/Discuss/Question/Opinion
<1>