diff --git a/docs/contents/ai/socratiq.html b/docs/contents/ai/socratiq.html index c20e2327..73b009a7 100644 --- a/docs/contents/ai/socratiq.html +++ b/docs/contents/ai/socratiq.html @@ -674,9 +674,9 @@
The top nav bar provices quick access to the following features:
Choosing the right machine learning framework for a given application requires carefully evaluating models, hardware, and software considerations. Figure 6.13 provides a comparison of different TensorFlow frameworks, which we’ll discuss in more detail:
-As shown in Figure 6.14, TensorFlow Lite Micro does not have OS support, while TensorFlow and TensorFlow Lite do. This design choice for TensorFlow Lite Micro helps reduce memory overhead, make startup times faster, and consume less energy. Instead, TensorFlow Lite Micro can be used in conjunction with real-time operating systems (RTOS) like FreeRTOS, Zephyr, and Mbed OS.
The figure also highlights an important memory management feature: TensorFlow Lite and TensorFlow Lite Micro support model memory mapping, allowing models to be directly accessed from flash storage rather than loaded into RAM. In contrast, TensorFlow does not offer this capability.
-TensorFlow Lite and TensorFlow Lite Micro have significantly smaller base binary sizes and memory footprints than TensorFlow (see Figure 6.15). For example, a typical TensorFlow Lite Micro binary is less than 200KB, whereas TensorFlow is much larger. This is due to the resource-constrained environments of embedded systems. TensorFlow supports x86, TPUs, and GPUs like NVIDIA, AMD, and Intel.
-Currently, the ML system stack consists of four abstractions as shown in Figure 6.16, namely (1) computational graphs, (2) tensor programs, (3) libraries and runtimes, and (4) hardware primitives.
-Understand the key characteristics and differences between Cloud ML, Edge ML, and TinyML systems.
Understand the key characteristics and differences between Cloud ML, Edge ML, Mobile ML, and TinyML systems.
Analyze the benefits and challenges associated with each ML paradigm.
Explore real-world applications and use cases for Cloud ML, Edge ML, and TinyML.
Explore real-world applications and use cases for Cloud ML, Edge ML, Mobile ML, and TinyML.
Compare the performance aspects of each ML approach, including latency, privacy, and resource utilization.
Examine the evolving landscape of ML systems and potential future developments.
ML is rapidly evolving, with new paradigms reshaping how models are developed, trained, and deployed. The field is experiencing significant innovation driven by advancements in hardware, software, and algorithmic techniques. These developments are enabling machine learning to be applied in diverse settings, from large-scale cloud infrastructures to edge devices and even tiny, resource-constrained environments.
Modern machine learning systems span a spectrum of deployment options, each with its own set of characteristics and use cases. At one end, we have cloud-based ML, which leverages powerful centralized computing resources for complex, data-intensive tasks. Moving along the spectrum, we encounter edge ML, which brings computation closer to the data source for reduced latency and improved privacy. At the far end, we find TinyML, which enables machine learning on extremely low-power devices with severe memory and processing constraints.
-This chapter explores the landscape of contemporary machine learning systems, covering three key approaches: Cloud ML, Edge ML, and TinyML. Figure 2.1 illustrates the spectrum of distributed intelligence across these approaches, providing a visual comparison of their characteristics. We will examine the unique characteristics, advantages, and challenges of each approach, as depicted in the figure. Additionally, we will discuss the emerging trends and technologies that are shaping the future of machine learning deployment, considering how they might influence the balance between these three paradigms.
+To better understand the dramatic differences between these ML deployment options, Table 2.1 provides examples of representative hardware platforms for each category. These examples illustrate the vast range of computational resources, power requirements, and cost considerations across the ML systems spectrum. As we explore each paradigm in detail, you can refer back to these concrete examples to better understand the practical implications of each approach.
+Category | +Example Device | +Processor | +Memory | +Storage | +Power | +Price Range | +Example Models/Tasks | +
---|---|---|---|---|---|---|---|
Cloud ML | +NVIDIA DGX A100 | +8x NVIDIA A100 GPUs (40GB/80GB) | +1TB System RAM | +15TB NVMe SSD | +6.5kW | +$200K+ | +Large language models (GPT-3), real-time video processing | +
+ | Google TPU v4 Pod | +4096 TPU v4 chips | +128TB+ | +Networked storage | +~MW | +Pay-per-use | +Training foundation models, large-scale ML research | +
Edge ML | +NVIDIA Jetson AGX Orin | +12-core Arm® Cortex®-A78AE, NVIDIA Ampere GPU | +32GB LPDDR5 | +64GB eMMC | +15-60W | +$899 | +Computer vision, robotics, autonomous systems | +
+ | Intel NUC 12 Pro | +Intel Core i7-1260P, Intel Iris Xe | +32GB DDR4 | +1TB SSD | +28W | +$750 | +Edge AI servers, industrial automation | +
Mobile ML | +iPhone 15 Pro | +A17 Pro (6-core CPU, 6-core GPU) | +8GB RAM | +128GB-1TB | +3-5W | +$999+ | +Face ID, computational photography, voice recognition | +
TinyML | +Arduino Nano 33 BLE Sense | +Arm Cortex-M4 @ 64MHz | +256KB RAM | +1MB Flash | +0.02-0.04W | +$35 | +Gesture recognition, voice detection | +
+ | ESP32-CAM | +Dual-core @ 240MHz | +520KB RAM | +4MB Flash | +0.05-0.25W | +$10 | +Image classification, motion detection | +
This chapter explores the landscape of contemporary machine learning systems, covering four key approaches: Cloud ML, Edge ML, and TinyML. Figure 2.1 illustrates the spectrum of distributed intelligence across these approaches, providing a visual comparison of their characteristics. We will examine the unique characteristics, advantages, and challenges of each approach, as depicted in the figure. Additionally, we will discuss the emerging trends and technologies that are shaping the future of machine learning deployment, considering how they might influence the balance between these three paradigms.