Computation offloading explained

Computation offloading is the transfer of resource intensive computational tasks to a separate processor, such as a hardware accelerator, or an external platform, such as a cluster, grid, or a cloud. Offloading to a coprocessor can be used to accelerate applications including: image rendering and mathematical calculations. Offloading computing to an external platform over a network can provide computing power and overcome hardware limitations of a device, such as limited computational power, storage, and energy.

History

The first concepts of stored-program computers were developed in the design of the ENIAC, the first general-purpose digital computer. The ENIAC was limited in performance to single tasks which led to the development of the EDVAC which would become the first computer designed to perform instructions of various types. Developing computing technologies facilitated the increase in performance of computers, and subsequently has led to a variety of configurations and architecture.

The first instances of computation offloading were the use of simple sub-processors to handle Input/output processing through a separate system called Channel I/O. This concept improved overall system performance as the mainframe only needed to set parameters for the operations while the channel processors carried out the I/O formatting and processing. During the 1970s, coprocessors began being used to accelerate floating-point arithmetic faster than earlier 8-bit and 16-bit processors which used software. As a result, math coprocessors became common for scientific and engineering calculations. Another form of coprocessor was the graphics coprocessor. As image processing became more popular, specialized graphics chips began being used to offload the creation of images from the CPU. Coprocessors were common in most computers, however, declined in usage due to the development in microprocessor technologies which integrated many coprocessor functions. Dedicated graphics processing units, however, are still widely used for their effectiveness in many tasks including; image processing, machine learning, parallel computing, computer vision, and physical simulation.

The concept of time-sharing, the sharing of computing resources, was first implemented by John McCarthy. At the time, mainframe computing was not practical due to the costs associated with purchasing and maintaining mainframe computers. Time-sharing was a viable solution for this problem as computing time could be available to smaller companies. In the 1990s telecommunications companies started offering virtual private network (VPN) services. This allowed companies to balance traffic on servers which resulted in an effective use of bandwidth. The cloud symbol became synonymous with the interaction between providers and users. This computing extended past network servers and allowed computing power to be available to users through time-sharing. The availability of virtual computers allowed users to offload tasks from a local processor.[1]

In 1997 distributed.net sought to obtain volunteer computing to solve computational intensive tasks by using the performance of networked PCs. This concept known as grid computing was associated with cloud computing systems.

The first concept of linking large mainframes to provide an effective form of parallelism was developed in the 1960s by IBM. Cluster computing was used by IBM was to increase hardware, operating system, and software performance while allowing users to run existing applications. This concept gained momentum during the 1980s as high-performance microprocessors and high-speed networks, tools for high performance distributed computing, emerged. Clusters could efficiently split and offload computation to individual nodes for increased performance while also gaining scalability.[2]

Concept

Computational tasks are handled by a central processor which executes instructions by carrying out rudimentary arithmetic, control logic, and input/output operations. The efficiency of computational tasks is dependent on the instructions per seconds that a CPU can perform which vary with different types of processors.[3] Certain application processes can be accelerated by offloading tasks from the main processor to a coprocessor while other processes might require an external processing platform.

Hardware Acceleration

Hardware offers greater possible performance for certain tasks when compared to software. The use of specialized hardware can perform functions faster than software processed on a CPU. Hardware has the benefit of customization which allows for dedicated technologies to be used for distinct functions. For example, a graphics processing unit (GPU), which consists of numerous low-performance cores, is more efficient at graphical computation than a CPU which features fewer, high power, cores.[4] Hardware accelerators, however, are less versatile when compared to a CPU.

Cloud Computing

Cloud Computing refers to both the applications transported over the Internet and the hardware and software in the data centers that provide services; which include data storage and computing.[5] This form of computing is reliant on high-speed internet access and infrastructure investment.[6] Through network access, a computer can migrate part of its computing to the cloud. This process involves sending data to a network of data centers that have access to computing power needed for computation.

Cluster Computing

Cluster computing is a type of parallel processing system which combines interconnected stand-alone computers to work as a single computing resource.[7] Clusters employ a parallel programming model which requires fast interconnection technologies to support high-bandwidth and low latency for communication between nodes. In a shared memory model, parallel processes have access to all memory as a global address space. Multiple processors have the ability to operate independently but they share the same memory, therefore, changes in memory by one processor are reflected on all other processors.

Grid Computing

Grid computing is a group of networked computers that work together as a virtual supercomputer to perform intensive computational tasks, such as analyzing huge sets of data. Through the cloud, it is possible to create and use computer grids for purposes and specific periods. Splitting computational tasks over multiple machines significantly reduces processing time to increase efficiency and minimize wasted resources. Unlike with parallel computing, grid computing tasks usually have no time dependency associated with them, but instead, they use computers that are part of the grid only when idle, and users can perform tasks unrelated to the grid at any time.[8]

Advantages

There are several advantages in offloading computation to an external processor:

Limitations

There are several limitations in offloading computation to an external processor:

Application

Cloud Services

Cloud services can be described by 3 main cloud service models: SaaS, PaaS, and IaaS. Software as a service (SaaS) is an externally hosted service which can be easily accessed by a consumer through a web browser. Platform as a service (PaaS) is a development environment where software can be built, tested, and deployed without the user having to focus on building and maintaining computational infrastructure. Infrastructure as a service (IaaS) is access to an infrastructure's resources, network technology, and security compliance which can be used by enterprises to build software. Cloud computing services provide users with access to large amounts of computational power and storage not viable on local computers without having significant expenditure.

Mobile Cloud Computing

Mobile devices, such as smartphones and wearable devices, are limited in terms of computational power, storage, and energy. Despite the constant development in key components including; CPU, GPU, memory, and wireless access technologies; mobile devices need to be portable and energy efficient. Mobile cloud computing is the combination of cloud computing and mobile computing, in which mobile devices perform computation offloading to balance the power of the cloud for accelerating application execution and saving energy consumption. In this computation offloading, a mobile device migrates part of its computing to the cloud. This process involves application partitioning, offloading decision, and distributed task execution.[11] [12]

Video Gaming

Video games are electronic games that involve input, an interaction with a user interface, and generate output, usually visual feedback on a video display device. These input/output operations rely on a computer and its components, including the CPU, GPU, RAM, and storage. Game files are stored in a form of secondary memory which is then loaded into the main memory when executed. The CPU is responsible for processing input from the user and passing information to the GPU. The GPU does not have access to a computer's main memory so instead, graphical assets have to be loaded into VRAM which is the GPU's memory. The CPU is responsible for instructing the GPU while the GPU uses the information to render an image on to an output device. CPU's are able to run games without a GPU through software rendering, however, offloading rendering to a GPU which has specialized hardware results in improved performance.[13]

Notes and References

  1. Web site: A Brief History of Cloud Computing. Foote. Keith D.. 2017-06-22. DATAVERSITY. 2019-10-17.
  2. Book: Cluster Computing: High-Performance, High-Availability, and High-Throughput Processing on a Network of Computers. Yeo. Chin. Buyya. Rajkumar. Pourreza. Hossien. Eskicioglu. Rasit. Peter. Graham. Sommers. Frank. Springer. 978-0-387-40532-2. Boston, MA. 2006-01-10.
  3. Web site: CPU Frequency. www.cpu-world.com. 2019-10-16.
  4. Book: Embedded Computing for High Performance. Cardoso. Joao. Gabriel. Jose. Pedro. Diniz. Morgan Kaufmann. June 15, 2017. 9780128041994.
  5. Book: Above the Clouds: A Berkeley View of Cloud Computing. Armbrust. Michael. Fox. Armondo. Griffith. Rean. Joseph. Anthony. Katz. Randy. Konwinski. Andrew. Lee. Gunho. Patterson. David. Rabkin. Ariel. February 10, 2009.
  6. Web site: How Does Cloud Computing Work? Cloud Academy Blog. 2019-03-25. Cloud Academy. 2019-10-22.
  7. Web site: Introduction to Cluster Computing — Distributed Computing Fundamentals. selkie.macalester.edu. 2019-10-22.
  8. Web site: What is Grid Computing - Definition Microsoft Azure. azure.microsoft.com. 2019-10-22.
  9. Akherfi. Khadija. Gerndt. Micheal. Harroud. Hamid. 2018-01-01. Mobile cloud computing for computation offloading: Issues and challenges. Applied Computing and Informatics. 14. 1. 1–16. 10.1016/j.aci.2016.11.002. 2210-8327. free.
  10. Ananthanarayanan. Ganesh. Bahl. Paramvir. Bodik. Peter. Chintalapudi. Krishna. Philipose. Matthai. Ravindranath. Lenin. Sinha. Sudipta. 2017. Real-Time Video Analytics: The Killer App for Edge Computing. Computer. 50. 10. 58–67. 10.1109/mc.2017.3641638. 206449115 . 0018-9162.
  11. Lin. L.. Liao. X.. Jin. H.. Li. P.. August 2019. Computation Offloading Toward Edge Computing. Proceedings of the IEEE. 107. 8. 1584–1607. 10.1109/JPROC.2019.2922285. 199017142 .
  12. Ma. X.. Zhao. Y.. Zhang. L.. Wang. H.. Peng. L.. September 2013. When mobile terminals meet the cloud: computation offloading as the bridge. IEEE Network. 27. 5. 28–33. 10.1109/MNET.2013.6616112. 16674645 .
  13. Web site: How Graphics Cards Work - ExtremeTech. www.extremetech.com. 2019-11-11.