Intel ARCHITECTURE IA-32 User Manual Page 292

  • Download
  • Add to my manuals
  • Print
  • Page
    / 568
  • Table of contents
  • BOOKMARKS
  • Rated. / 5. Based on customer reviews
Page view 291
IA-32 Intel® Architecture Optimization
6-2
Memory Optimization Using Hardware Prefetching, Software
Prefetch and Cacheability Instructions: discusses techniques for
implementing memory optimizations using the above instructions.
Using deterministic cache parameters to manage cache hierarchy.
General Prefetch Coding Guidelines
The following guidelines will help you to reduce memory traffic and
utilize peak memory system bandwidth more effectively when large
amounts of data movement must originate from the memory system:
Take advantage of the hardware prefetchers ability to prefetch data
that are accessed in linear patterns, either forward or backward
direction.
Take advantage of the hardware prefetchers ability to prefetch data
that are accessed in a regular pattern with access stride that are
substantially smaller than half of the trigger distance of the
hardware prefetch (see Table 1-2).
Use a current-generation compiler, such as the Intel
®
C++ Compiler
that supports C++ language-level features for Streaming SIMD
Extensions. Streaming SIMD Extensions and MMX technology
instructions provide intrinsics that allow you to optimize cache
utilization. The examples of such Intel
®
compiler intrinsics are
_mm_prefetch, _mm_stream and _mm_load, _mm_sfence. For more
details on these intrinsics, refer to the Intel® C++ Compiler Users
Guide, doc. number 718195.
NOTE. In a number of cases presented in this chapter,
the prefetching and cache utilization are specific to the
current implementation of Intel NetBurst
microarchitecture but are largely applicable for the
future processors.
Page view 291
1 2 ... 287 288 289 290 291 292 293 294 295 296 297 ... 567 568

Comments to this Manuals

No comments