diff --git a/clang/lib/Basic/Targets/X86.cpp b/clang/lib/Basic/Targets/X86.cpp --- a/clang/lib/Basic/Targets/X86.cpp +++ b/clang/lib/Basic/Targets/X86.cpp @@ -1170,6 +1170,7 @@ bool X86TargetInfo::validateCpuSupports(StringRef FeatureStr) const { return llvm::StringSwitch(FeatureStr) #define X86_FEATURE_COMPAT(ENUM, STR, PRIORITY) .Case(STR, true) +#define X86_MICROARCH_LEVEL(ENUM, STR, PRIORITY) .Case(STR, true) #include "llvm/TargetParser/X86TargetParser.def" .Default(false); } diff --git a/clang/lib/CodeGen/CGBuiltin.cpp b/clang/lib/CodeGen/CGBuiltin.cpp --- a/clang/lib/CodeGen/CGBuiltin.cpp +++ b/clang/lib/CodeGen/CGBuiltin.cpp @@ -13325,9 +13325,7 @@ } Value *CodeGenFunction::EmitX86CpuSupports(ArrayRef FeatureStrs) { - uint64_t Mask = llvm::X86::getCpuSupportsMask(FeatureStrs); - std::array FeatureMask{Lo_32(Mask), Hi_32(Mask), 0, 0}; - return EmitX86CpuSupports(FeatureMask); + return EmitX86CpuSupports(llvm::X86::getCpuSupportsMask(FeatureStrs)); } llvm::Value * diff --git a/clang/lib/CodeGen/CodeGenFunction.cpp b/clang/lib/CodeGen/CodeGenFunction.cpp --- a/clang/lib/CodeGen/CodeGenFunction.cpp +++ b/clang/lib/CodeGen/CodeGenFunction.cpp @@ -2683,24 +2683,12 @@ if (!RO.Conditions.Architecture.empty()) { StringRef Arch = RO.Conditions.Architecture; - std::array Mask{}; - // If arch= specifies an x86-64 micro-architecture level, test a special - // feature named FEATURE_X86_64_*, otherwise we use __builtin_cpu_is. - if (Arch.consume_front("x86-64")) { - if (Arch.empty()) // FEATURE_X86_64_BASELINE 95=2*32+31 - Mask[2] = 1u << 31; - else if (Arch == "-v2") // FEATURE_X86_64_V2 96==3*32+0 - Mask[3] = 1u << 0; - else if (Arch == "-v3") // FEATURE_X86_64_V3 97==3*32+1 - Mask[3] = 1u << 1; - else if (Arch == "-v4") // FEATURE_X86_64_V3 98==3*32+2 - Mask[3] = 1u << 2; - else - llvm_unreachable("invalid x86-64 micro-architecture level"); - Condition = EmitX86CpuSupports(Mask); - } else { + // If arch= specifies an x86-64 micro-architecture level, test the feature + // with __builtin_cpu_supports, otherwise use __builtin_cpu_is. + if (Arch.starts_with("x86-64")) + Condition = EmitX86CpuSupports({Arch}); + else Condition = EmitX86CpuIs(Arch); - } } if (!RO.Conditions.Features.empty()) { diff --git a/clang/lib/CodeGen/CodeGenModule.cpp b/clang/lib/CodeGen/CodeGenModule.cpp --- a/clang/lib/CodeGen/CodeGenModule.cpp +++ b/clang/lib/CodeGen/CodeGenModule.cpp @@ -4168,8 +4168,9 @@ // always run on at least a 'pentium'). We do this by deleting the 'least // advanced' (read, lowest mangling letter). while (Options.size() > 1 && - llvm::X86::getCpuSupportsMask( - (Options.end() - 2)->Conditions.Features) == 0) { + llvm::all_of(llvm::X86::getCpuSupportsMask( + (Options.end() - 2)->Conditions.Features), + [](auto X) { return X == 0; })) { StringRef LHSName = (Options.end() - 2)->Function->getName(); StringRef RHSName = (Options.end() - 1)->Function->getName(); if (LHSName.compare(RHSName) < 0) diff --git a/clang/test/CodeGen/builtin-cpu-supports.c b/clang/test/CodeGen/builtin-cpu-supports.c --- a/clang/test/CodeGen/builtin-cpu-supports.c +++ b/clang/test/CodeGen/builtin-cpu-supports.c @@ -30,3 +30,23 @@ } // CHECK: declare dso_local void @__cpu_indicator_init() + +// CHECK-LABEL: define{{.*}} @baseline( +// CHECK: [[LOAD:%.*]] = load i32, ptr getelementptr inbounds ([[[#]] x i32], ptr @__cpu_features2, i32 0, i32 1) +// CHECK-NEXT: and i32 [[LOAD]], -2147483648 +int baseline() { return __builtin_cpu_supports("x86-64"); } + +// CHECK-LABEL: define{{.*}} @v2( +// CHECK: [[LOAD:%.*]] = load i32, ptr getelementptr inbounds ([[[#]] x i32], ptr @__cpu_features2, i32 0, i32 2) +// CHECK-NEXT: and i32 [[LOAD]], 1 +int v2() { return __builtin_cpu_supports("x86-64-v2"); } + +// CHECK-LABEL: define{{.*}} @v3( +// CHECK: [[LOAD:%.*]] = load i32, ptr getelementptr inbounds ([[[#]] x i32], ptr @__cpu_features2, i32 0, i32 2) +// CHECK-NEXT: and i32 [[LOAD]], 2 +int v3() { return __builtin_cpu_supports("x86-64-v3"); } + +// CHECK-LABEL: define{{.*}} @v4( +// CHECK: [[LOAD:%.*]] = load i32, ptr getelementptr inbounds ([[[#]] x i32], ptr @__cpu_features2, i32 0, i32 2) +// CHECK-NEXT: and i32 [[LOAD]], 4 +int v4() { return __builtin_cpu_supports("x86-64-v4"); } diff --git a/clang/test/Sema/attr-cpuspecific.c b/clang/test/Sema/attr-cpuspecific.c --- a/clang/test/Sema/attr-cpuspecific.c +++ b/clang/test/Sema/attr-cpuspecific.c @@ -181,3 +181,8 @@ // expected-error@+1 {{attribute 'cpu_specific' multiversioning cannot be combined with attribute 'overloadable'}} int __attribute__((__overloadable__)) __attribute__((cpu_specific(atom))) good_overload7(void); int __attribute__((cpu_specific(ivybridge))) good_overload7(int); + +// expected-error@+1 {{invalid option 'x86_64'}} +int __attribute__((cpu_specific(x86_64))) baseline(void); +// expected-error@+1 {{invalid option 'x86_64_v2'}} +int __attribute__((cpu_specific(x86_64_v2))) baseline(void); diff --git a/clang/test/Sema/attr-target.c b/clang/test/Sema/attr-target.c --- a/clang/test/Sema/attr-target.c +++ b/clang/test/Sema/attr-target.c @@ -26,6 +26,11 @@ //expected-warning@+1 {{unknown tune CPU 'hiss' in the 'target' attribute string; 'target' attribute ignored}} int __attribute__((target("tune=hiss,tune=woof"))) apple_tree(void) { return 4; } +//expected-warning@+1 {{unsupported 'x86-64' in the 'target' attribute string}} +void __attribute__((target("x86-64"))) baseline(void) {} +//expected-warning@+1 {{unsupported 'x86-64-v2' in the 'target' attribute string}} +void __attribute__((target("x86-64-v2"))) v2(void) {} + #elifdef __aarch64__ int __attribute__((target("sve,arch=armv8-a"))) foo(void) { return 4; } diff --git a/llvm/include/llvm/TargetParser/X86TargetParser.h b/llvm/include/llvm/TargetParser/X86TargetParser.h --- a/llvm/include/llvm/TargetParser/X86TargetParser.h +++ b/llvm/include/llvm/TargetParser/X86TargetParser.h @@ -15,6 +15,7 @@ #include "llvm/ADT/ArrayRef.h" #include "llvm/ADT/StringMap.h" +#include namespace llvm { template class SmallVectorImpl; @@ -57,7 +58,10 @@ enum ProcessorFeatures { #define X86_FEATURE(ENUM, STRING) FEATURE_##ENUM, #include "llvm/TargetParser/X86TargetParser.def" - CPU_FEATURE_MAX + CPU_FEATURE_MAX, + +#define X86_MICROARCH_LEVEL(ENUM, STRING, PRIORITY) FEATURE_##ENUM = PRIORITY, +#include "llvm/TargetParser/X86TargetParser.def" }; enum CPUKind { @@ -171,7 +175,7 @@ char getCPUDispatchMangling(StringRef Name); bool validateCPUSpecificCPUDispatch(StringRef Name); -uint64_t getCpuSupportsMask(ArrayRef FeatureStrs); +std::array getCpuSupportsMask(ArrayRef FeatureStrs); unsigned getFeaturePriority(ProcessorFeatures Feat); } // namespace X86 diff --git a/llvm/include/llvm/TargetParser/X86TargetParser.def b/llvm/include/llvm/TargetParser/X86TargetParser.def --- a/llvm/include/llvm/TargetParser/X86TargetParser.def +++ b/llvm/include/llvm/TargetParser/X86TargetParser.def @@ -128,6 +128,10 @@ #define X86_FEATURE(ENUM, STR) #endif +#ifndef X86_MICROARCH_LEVEL +#define X86_MICROARCH_LEVEL(ENUM, STR, PRIORITY) +#endif + X86_FEATURE_COMPAT(CMOV, "cmov", 0) X86_FEATURE_COMPAT(MMX, "mmx", 1) X86_FEATURE_COMPAT(POPCNT, "popcnt", 9) @@ -242,5 +246,11 @@ X86_FEATURE (RETPOLINE_INDIRECT_CALLS, "retpoline-indirect-calls") X86_FEATURE (LVI_CFI, "lvi-cfi") X86_FEATURE (LVI_LOAD_HARDENING, "lvi-load-hardening") + +X86_MICROARCH_LEVEL(X86_64_BASELINE,"x86-64", 95) +X86_MICROARCH_LEVEL(X86_64_V2, "x86-64-v2", 96) +X86_MICROARCH_LEVEL(X86_64_V3, "x86-64-v3", 97) +X86_MICROARCH_LEVEL(X86_64_V4, "x86-64-v4", 98) #undef X86_FEATURE_COMPAT #undef X86_FEATURE +#undef X86_MICROARCH_LEVEL diff --git a/llvm/lib/TargetParser/X86TargetParser.cpp b/llvm/lib/TargetParser/X86TargetParser.cpp --- a/llvm/lib/TargetParser/X86TargetParser.cpp +++ b/llvm/lib/TargetParser/X86TargetParser.cpp @@ -703,18 +703,22 @@ return I != std::end(Processors); } -uint64_t llvm::X86::getCpuSupportsMask(ArrayRef FeatureStrs) { +std::array +llvm::X86::getCpuSupportsMask(ArrayRef FeatureStrs) { // Processor features and mapping to processor feature value. - uint64_t FeaturesMask = 0; - for (const StringRef &FeatureStr : FeatureStrs) { + std::array FeatureMask{}; + for (StringRef FeatureStr : FeatureStrs) { unsigned Feature = StringSwitch(FeatureStr) #define X86_FEATURE_COMPAT(ENUM, STR, PRIORITY) \ .Case(STR, llvm::X86::FEATURE_##ENUM) +#define X86_MICROARCH_LEVEL(ENUM, STR, PRIORITY) \ + .Case(STR, llvm::X86::FEATURE_##ENUM) #include "llvm/TargetParser/X86TargetParser.def" ; - FeaturesMask |= (1ULL << Feature); + assert(Feature / 32 < FeatureMask.size()); + FeatureMask[Feature / 32] |= 1U << (Feature % 32); } - return FeaturesMask; + return FeatureMask; } unsigned llvm::X86::getFeaturePriority(ProcessorFeatures Feat) {