This articlerelies excessively onreferences toprimary sources. Please improve this article by addingsecondary or tertiary sources. Find sources: "Intel C++ Compiler" – news ·newspapers ·books ·scholar ·JSTOR(March 2023) (Learn how and when to remove this message) |
Other names |
|
---|---|
Developer(s) | Intel |
Stable release | |
Repository | |
Operating system | Windows,Linux |
Type | Compiler |
License | Freeware,proprietary |
Website | software |
Other names |
|
---|---|
Developer(s) | Intel |
Stable release | |
Repository | |
Operating system | Windows,macOS,Linux |
Type | Compiler |
License | Freeware,proprietary |
Website | software |
Intel oneAPI DPC++/C++ Compiler andIntel C++ Compiler Classic (deprecated icc and icl[3] is in Intel OneAPI HPC toolkit) areIntel’sC,C++,SYCL, andData Parallel C++ (DPC++) compilers for Intel processor-based systems, available forWindows,Linux, andmacOS operating systems.[4]
Intel oneAPI DPC++/C++ Compiler is available for Windows and Linux and supports compiling C, C++, SYCL, and Data Parallel C++ (DPC++)source, targeting IntelIA-32, Intel 64 (akax86-64),Core,Xeon, andXeon Scalable processors, as well as GPUs including Intel Processor Graphics Gen9 and above, IntelXe architecture, and Intel Programmable Acceleration Card with Intel Arria 10 GX FPGA.[5] Like Intel C++ Compiler Classic, it also supports theMicrosoft Visual Studio andEclipse IDE development environments, and supports threading via Intel oneAPIThreading Building Blocks,OpenMP, and native threads.
DPC++[6][7] builds on theSYCL specification from The Khronos Group. It is designed to allow developers to reuse code across hardware targets (CPUs and accelerators such as GPUs and FPGAs) and perform custom tuning for a specific accelerator. DPC++ comprisesC++17 and SYCL language features and incorporates open-source community extensions that make SYCL easier to use. Many of these extensions were adopted by the SYCL 2020 provisional specification[8] including unified shared memory, group algorithms, and sub-groups.
Intel announced in August 2021 the complete adoption ofLLVM for faster build times and benefits from supporting the latest C++ standards.[9]
Intel C++ Compiler Classic is available for Windows, Linux, and macOS and supports compiling C and C++ source, targeting Intel IA-32, Intel 64 (x86-64), Core, Xeon, and Xeon Scalable processors.[5] It supports the Microsoft Visual Studio and Eclipse IDE development environments. Intel C++ Compiler Classic supports threading via Intel oneAPI Threading Building Blocks, OpenMP, and native threads.
According to Intel,[10] starting with the 2023.0 release, Intel oneAPI DPC++/C++ Compiler supports all current Intel general-purposex86-64 CPUs andGPUs including:
Intel C++ Compiler Classic targets general-purpose Intelx86-64 architecture CPUs including:[5]
The Intel oneAPI DPC++/C++ Compiler is available either as a standalone component[11] or as part of the Intel oneAPI Base Toolkit, Intel oneAPI HPC Toolkit, and Intel oneAPI IoT Toolkit.[5]
The Intel C++ Compiler Classic is available either as a standalone component[12] or as part of the Intel oneAPI Base Toolkit.[5]
The Intel compiler provides debugging information that is standard for the common debuggers (DWARF 2 on Linux, similar togdb, andCOFF for Windows). The flags to compile with debugging information are/Zi
on Windows and-g
on Linux. Debugging is done on Windows using the Visual Studio debugger and, on Linux, using gdb.
While the Intel compiler can generate a gprof compatibleprofiling output, Intel also provides a kernel level, system-wide statistical profiler calledIntel VTune Profiler. VTune can be used from a command line or through an included GUI on Linux or Windows. It can also be integrated intoVisual Studio on Windows, orEclipse on Linux). In addition to the VTune profiler, there isIntel Advisor that specializes in vectorization optimization, offload modeling, flow graph design and tools for threading design and prototyping.
Intel also offers a tool for memory and threading error detection calledIntel Inspector XE. Regarding memory errors, it helps detectmemory leaks, memory corruption, allocation/de-allocation of API mismatches and inconsistent memory API usage. Regarding threading errors, it helps detect data races (both heap and stack), deadlocks and thread and synch API errors.
Previous versions of Intel’s C/C++ compilers have been criticized for optimizing less aggressively for non-Intel processors; for example, Steve Westfield wrote in a 2005 article at the AMD website:[13]
Intel 8.1 C/C++ compiler uses the flag -xN (for Linux) or -QxN (for Windows) to take advantage of the SSE2 extensions. For SSE3, the compiler switch is -xP (for Linux) and -QxP (for Windows). ... With the -xN/-QxN and -xP/-QxP flags set, it checks the processor vendor string—and if it's not "GenuineIntel", it stops execution without even checking the feature flags. Ouch!
The Danish developer and scholarAgner Fog wrote in 2009:[14]
The Intel compiler and several different Intel function libraries have suboptimal performance onAMD andVIA processors. The reason is that the compiler or library can make multiple versions of a piece of code, each optimized for a certain processor andinstruction set, for exampleSSE2,SSE3, etc. The system includes a function that detects which type of CPU it is running on and chooses the optimal code path for that CPU. This is called a CPU dispatcher. However, the Intel CPU dispatcher does not only check which instruction set is supported by the CPU, it also checks the vendor ID string. If the vendor string is "GenuineIntel" then it uses the optimal code path. If the CPU is not from Intel then, in most cases, it will run the slowest possible version of the code, even if the CPU is fully compatible with a better version.
This vendor-specific CPU dispatching may potentially impact the performance of software built with an Intel compiler or an Intel function library on non-Intel processors, possibly without the programmer’s knowledge. This has allegedly led to misleadingbenchmarks,[14] including one incident when changing theCPUID of aVIA Nano significantly improved results.[15] In November 2009, AMD and Intel reached a legal settlement over this and related issues,[16] and in late 2010, AMD settled aUS Federal Trade Commissionantitrust investigation against Intel.[17]
The FTC settlement included a disclosure provision where Intel must:[18]
publish clearly that its compiler discriminates against non-Intel processors (such as AMD's designs), not fully utilizing their features and producing inferior code.
In compliance with this ruling, Intel added disclaimers to its compiler documentation:[19]
Intel's compilers may or may not optimize to the same degree for non-Intel microprocessors for optimizations that are not unique to Intel microprocessors. These optimizations include SSE2, SSE3, and SSSE3 instruction sets and other optimizations. Intel does not guarantee the availability, functionality, or effectiveness of any optimization on microprocessors not manufactured by Intel. Microprocessor-dependent optimizations in this product are intended for use with Intel microprocessors. Certain optimizations not specific to Intel microarchitecture are reserved for Intel microprocessors. Please refer to the applicable product User and Reference Guides for more information regarding the specific instruction sets covered by this notice.
As late as 2013, an article inThe Register alleged that the object code produced by the Intel compiler for theAnTuTu Mobile Benchmark omitted portions of the benchmark which showed increased performance compared to ARM platforms.[20]
The following lists versions of the Intel C++ Compiler since 1999:[21]
Compiler version | Release date | Major new features |
---|---|---|
Intel C/C++ Compiler 4.0 | 1999 | Support for Streaming SIMD, enhanced profile-guided optimization, vectorization[22] |
Intel C/C++ Compiler 4.5 | 2000 | Support for OpenMP, Itanium[23] |
Intel C/C++ Compiler 5.0 | 2001 | [24] |
Intel C/C++ Compiler 6.0 | 2002 | Increased Linux support, Auto-parallelization[25] |
Intel C++ Compiler 7.0 | 2003 | [26] |
Intel C++ Compiler 8.0 | December 15, 2003 | Precompiled headers, code-coverage tools. |
Intel C++ Compiler 8.1 | September 2004 | AMD64 architecture (for Linux). |
Intel C++ Compiler 9.0 | June 14, 2005 | AMD64 architecture (for Windows), software-based speculative pre-computation (SSP) optimization, improved loop optimization reports. |
Intel C++ Compiler 10.0 | June 5, 2007 | Improved parallelizer and vectorizer, Streaming SIMD Extensions 4 (SSE4), new and enhanced optimization reports for advanced loop transformations, new optimized exception handling implementation. |
Intel C++ Compiler 10.1 | November 7, 2007 | New OpenMP* compatibility runtime library: if you use the new OpenMP RTL, you can mix and match with libraries and objects built by Visual C++. To use the new libraries, you need to use the new option "-Qopenmp /Qopenmp-lib:compat" on Windows, and "-openmp -openmp-lib:compat" on Linux. This version of the Intel compiler supports more intrinsics fromVisual Studio 2005. VS2008 support: command line only in this release. The IDE integration was not supported yet. |
Intel C++ Compiler 11.0 | November 2008 | InitialC++11 support. VS2008 IDE integration on Windows. OpenMP 3.0. Source Checker for static memory/parallel diagnostics. |
Intel C++ Compiler 11.1 | June 23, 2009 | Support for latest Intel SSESSE4.2,AVX andAES instructions. Parallel Debugger Extension. Improved integration into Microsoft Visual Studio, Eclipse CDT 5.0 and Mac Xcode IDE. |
Intel C++ Composer XE 2011 up to Update 5 (compiler 12.0) | November 7, 2010 | Cilk Plus language extensions, Guided Auto-Parallelism, ImprovedC++11 support.[27] |
Intel C++ Composer XE 2011 Update 6 and above (compiler 12.1) | September 8, 2011 | Cilk Plus language extensions updated to support specification version 1.1 and available onMac OS X in addition to Windows and Linux,Threading Building Blocks updated to support version 4.0, Apple blocks supported on Mac OS X, improvedC++11 support including support for Variadic templates,OpenMP 3.1 support. |
Intel C++ Composer XE 2013 (compiler 13.0) | September 5, 2012 | Linux-based support for Intel Xeon Phi coprocessors, support for Microsoft Visual Studio 12 (Desktop), support for gcc 4.7, support for Intel AVX 2 instructions, updates to existing functionality focused on improved application performance.[28] |
Intel C++ Composer XE 2013 SP1 (compiler 14.0) | September 4, 2013 | Online installer; support for Intel Xeon Phi coprocessors; preview Win32 only support for Intel graphics; improved C++11 support |
Intel C++ Composer XE 2013 SP1 Update 1 (compiler 14.0.1) | October 18, 2013 | Japanese localization of 14.0; Windows 8.1 and Xcode 5.0 support |
Intel C++ Compiler for Android (compiler 14.0.1) | November 12, 2013 | Hosted on Windows, Linux, or OS X, compatible with Android NDK tools including the gcc compiler and Eclipse |
Intel C++ Composer XE 2015 (compiler 15.0) | July 25, 2014 | Full C++11 language support; Additional OpenMP 4.0 and Cilk Plus enhancements |
Intel C++ Composer XE 2015 Update 1 (compiler 15.0.1) | October 30, 2014 | AVX-512 support; Japanese localization |
Intel C++ 16.0 | August 25, 2015 | Suite-based availability (Intel Parallel Studio XE, Intel System Studio) |
Intel C++ 17.0 | September 15, 2016 | Suite-based availability (Intel Parallel Studio XE, Intel System Studio) |
Intel C++ 18.0 | January 26, 2017 | Suite-based availability (Intel Parallel Studio XE, Intel System Studio) |
Intel C++ 19.0 | April 3, 2018 | Suite-based availability (Intel Parallel Studio XE, Intel System Studio) |
Intel C++ Compiler Classic 19.1 | October 22, 2020 | Initial Open MP 5.1 CPU only |
Intel oneAPI DPC++ / C++ Compiler 2021 | December 8, 2020 | SYCL, DPC++, initial Open MP 5.1 |
Intel C++ Compiler Classic 2021.1.2 Intel oneAPI DPC++/C++ Compiler 2021.1.2 | December 16, 2020 | oneAPI DPC++/C++ introduces support forGPUoffloading |
Intel C++ Compiler Classic 2022.2.1 Intel oneAPI DPC++/C++ Compiler 2022.2.1 | November 2, 2022 | Support for latest Intel CPUs, GPUs, and FPGAs; support for upcoming ISO/IEC 9899:2023 (C23) and ISO/IEC 14882:2023 (C++23) language standards |
Intel C++ Compiler Classic 2023.0 Intel oneAPI DPC++/C++ Compiler 2023.0 | Q1 2023[10] | Support for Intel Advanced Matrix Extensions (Intel AMX), Quick Assist Technology (QAT), IntelAVX-512 withVector Neural Network Instructions (VNNI),bfloat16 , GPU datatype flexibility, Intel Xe matrix extensions (Intel XMX), Intel GPU vector engine, XE-Link |
{{cite web}}
: CS1 maint: bot: original URL status unknown (link){{cite web}}
: CS1 maint: bot: original URL status unknown (link){{cite web}}
: CS1 maint: bot: original URL status unknown (link){{cite web}}
: CS1 maint: bot: original URL status unknown (link){{cite web}}
: CS1 maint: bot: original URL status unknown (link)