From owner-svn-src-projects@freebsd.org Tue Sep 22 20:42:15 2015 Return-Path: Delivered-To: svn-src-projects@mailman.ysv.freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:1900:2254:206a::19:1]) by mailman.ysv.freebsd.org (Postfix) with ESMTP id 351A0A02C0C for ; Tue, 22 Sep 2015 20:42:15 +0000 (UTC) (envelope-from dim@FreeBSD.org) Received: from repo.freebsd.org (repo.freebsd.org [IPv6:2001:1900:2254:2068::e6a:0]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (Client did not present a certificate) by mx1.freebsd.org (Postfix) with ESMTPS id 171C21CA0; Tue, 22 Sep 2015 20:42:15 +0000 (UTC) (envelope-from dim@FreeBSD.org) Received: from repo.freebsd.org ([127.0.1.70]) by repo.freebsd.org (8.15.2/8.15.2) with ESMTP id t8MKgEpO033325; Tue, 22 Sep 2015 20:42:14 GMT (envelope-from dim@FreeBSD.org) Received: (from dim@localhost) by repo.freebsd.org (8.15.2/8.15.2/Submit) id t8MKgEDK033324; Tue, 22 Sep 2015 20:42:14 GMT (envelope-from dim@FreeBSD.org) Message-Id: <201509222042.t8MKgEDK033324@repo.freebsd.org> X-Authentication-Warning: repo.freebsd.org: dim set sender to dim@FreeBSD.org using -f From: Dimitry Andric Date: Tue, 22 Sep 2015 20:42:14 +0000 (UTC) To: src-committers@freebsd.org, svn-src-projects@freebsd.org Subject: svn commit: r288128 - projects/clang370-import/contrib/llvm/patches X-SVN-Group: projects MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit X-BeenThere: svn-src-projects@freebsd.org X-Mailman-Version: 2.1.20 Precedence: list List-Id: "SVN commit messages for the src " projects" tree" List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Tue, 22 Sep 2015 20:42:15 -0000 Author: dim Date: Tue Sep 22 20:42:14 2015 New Revision: 288128 URL: https://svnweb.freebsd.org/changeset/base/288128 Log: Add clang patch corresponding to r288127. Added: projects/clang370-import/contrib/llvm/patches/patch-05-clang-r244063-missing-atomic-libcall.diff Added: projects/clang370-import/contrib/llvm/patches/patch-05-clang-r244063-missing-atomic-libcall.diff ============================================================================== --- /dev/null 00:00:00 1970 (empty, because file is newly added) +++ projects/clang370-import/contrib/llvm/patches/patch-05-clang-r244063-missing-atomic-libcall.diff Tue Sep 22 20:42:14 2015 (r288128) @@ -0,0 +1,229 @@ +Pull in r244063 from upstream clang trunk (by James Y Knight): + + Add missing atomic libcall support. + + Support for emitting libcalls for __atomic_fetch_nand and + __atomic_{add,sub,and,or,xor,nand}_fetch was missing; add it, and some + test cases. + + Differential Revision: http://reviews.llvm.org/D10847 + +This fixes "cannot compile this atomic library call yet" errors when +compiling code which calls the above builtins, on arm < v6. + +Introduced here: http://svnweb.freebsd.org/changeset/base/288127 + +Index: tools/clang/docs/LanguageExtensions.rst +=================================================================== +--- tools/clang/docs/LanguageExtensions.rst ++++ tools/clang/docs/LanguageExtensions.rst +@@ -1715,6 +1715,9 @@ The macros ``__ATOMIC_RELAXED``, ``__ATOMIC_CONSUM + provided, with values corresponding to the enumerators of C11's + ``memory_order`` enumeration. + ++(Note that Clang additionally provides GCC-compatible ``__atomic_*`` ++builtins) ++ + Low-level ARM exclusive memory builtins + --------------------------------------- + +Index: tools/clang/lib/CodeGen/CGAtomic.cpp +=================================================================== +--- tools/clang/lib/CodeGen/CGAtomic.cpp ++++ tools/clang/lib/CodeGen/CGAtomic.cpp +@@ -699,7 +699,7 @@ RValue CodeGenFunction::EmitAtomicExpr(AtomicExpr + + switch (E->getOp()) { + case AtomicExpr::AO__c11_atomic_init: +- llvm_unreachable("Already handled!"); ++ llvm_unreachable("Already handled above with EmitAtomicInit!"); + + case AtomicExpr::AO__c11_atomic_load: + case AtomicExpr::AO__atomic_load_n: +@@ -785,6 +785,9 @@ RValue CodeGenFunction::EmitAtomicExpr(AtomicExpr + if (UseLibcall) { + bool UseOptimizedLibcall = false; + switch (E->getOp()) { ++ case AtomicExpr::AO__c11_atomic_init: ++ llvm_unreachable("Already handled above with EmitAtomicInit!"); ++ + case AtomicExpr::AO__c11_atomic_fetch_add: + case AtomicExpr::AO__atomic_fetch_add: + case AtomicExpr::AO__c11_atomic_fetch_and: +@@ -791,14 +794,34 @@ RValue CodeGenFunction::EmitAtomicExpr(AtomicExpr + case AtomicExpr::AO__atomic_fetch_and: + case AtomicExpr::AO__c11_atomic_fetch_or: + case AtomicExpr::AO__atomic_fetch_or: ++ case AtomicExpr::AO__atomic_fetch_nand: + case AtomicExpr::AO__c11_atomic_fetch_sub: + case AtomicExpr::AO__atomic_fetch_sub: + case AtomicExpr::AO__c11_atomic_fetch_xor: + case AtomicExpr::AO__atomic_fetch_xor: ++ case AtomicExpr::AO__atomic_add_fetch: ++ case AtomicExpr::AO__atomic_and_fetch: ++ case AtomicExpr::AO__atomic_nand_fetch: ++ case AtomicExpr::AO__atomic_or_fetch: ++ case AtomicExpr::AO__atomic_sub_fetch: ++ case AtomicExpr::AO__atomic_xor_fetch: + // For these, only library calls for certain sizes exist. + UseOptimizedLibcall = true; + break; +- default: ++ ++ case AtomicExpr::AO__c11_atomic_load: ++ case AtomicExpr::AO__c11_atomic_store: ++ case AtomicExpr::AO__c11_atomic_exchange: ++ case AtomicExpr::AO__c11_atomic_compare_exchange_weak: ++ case AtomicExpr::AO__c11_atomic_compare_exchange_strong: ++ case AtomicExpr::AO__atomic_load_n: ++ case AtomicExpr::AO__atomic_load: ++ case AtomicExpr::AO__atomic_store_n: ++ case AtomicExpr::AO__atomic_store: ++ case AtomicExpr::AO__atomic_exchange_n: ++ case AtomicExpr::AO__atomic_exchange: ++ case AtomicExpr::AO__atomic_compare_exchange_n: ++ case AtomicExpr::AO__atomic_compare_exchange: + // Only use optimized library calls for sizes for which they exist. + if (Size == 1 || Size == 2 || Size == 4 || Size == 8) + UseOptimizedLibcall = true; +@@ -820,6 +843,9 @@ RValue CodeGenFunction::EmitAtomicExpr(AtomicExpr + QualType RetTy; + bool HaveRetTy = false; + switch (E->getOp()) { ++ case AtomicExpr::AO__c11_atomic_init: ++ llvm_unreachable("Already handled!"); ++ + // There is only one libcall for compare an exchange, because there is no + // optimisation benefit possible from a libcall version of a weak compare + // and exchange. +@@ -903,7 +929,49 @@ RValue CodeGenFunction::EmitAtomicExpr(AtomicExpr + AddDirectArgument(*this, Args, UseOptimizedLibcall, Val1, MemTy, + E->getExprLoc(), sizeChars); + break; +- default: return EmitUnsupportedRValue(E, "atomic library call"); ++ // T __atomic_fetch_nand_N(T *mem, T val, int order) ++ case AtomicExpr::AO__atomic_fetch_nand: ++ LibCallName = "__atomic_fetch_nand"; ++ AddDirectArgument(*this, Args, UseOptimizedLibcall, Val1, MemTy, ++ E->getExprLoc(), sizeChars); ++ break; ++ ++ // T __atomic_add_fetch_N(T *mem, T val, int order) ++ case AtomicExpr::AO__atomic_add_fetch: ++ LibCallName = "__atomic_add_fetch"; ++ AddDirectArgument(*this, Args, UseOptimizedLibcall, Val1, LoweredMemTy, ++ E->getExprLoc(), sizeChars); ++ break; ++ // T __atomic_and_fetch_N(T *mem, T val, int order) ++ case AtomicExpr::AO__atomic_and_fetch: ++ LibCallName = "__atomic_and_fetch"; ++ AddDirectArgument(*this, Args, UseOptimizedLibcall, Val1, MemTy, ++ E->getExprLoc(), sizeChars); ++ break; ++ // T __atomic_or_fetch_N(T *mem, T val, int order) ++ case AtomicExpr::AO__atomic_or_fetch: ++ LibCallName = "__atomic_or_fetch"; ++ AddDirectArgument(*this, Args, UseOptimizedLibcall, Val1, MemTy, ++ E->getExprLoc(), sizeChars); ++ break; ++ // T __atomic_sub_fetch_N(T *mem, T val, int order) ++ case AtomicExpr::AO__atomic_sub_fetch: ++ LibCallName = "__atomic_sub_fetch"; ++ AddDirectArgument(*this, Args, UseOptimizedLibcall, Val1, LoweredMemTy, ++ E->getExprLoc(), sizeChars); ++ break; ++ // T __atomic_xor_fetch_N(T *mem, T val, int order) ++ case AtomicExpr::AO__atomic_xor_fetch: ++ LibCallName = "__atomic_xor_fetch"; ++ AddDirectArgument(*this, Args, UseOptimizedLibcall, Val1, MemTy, ++ E->getExprLoc(), sizeChars); ++ break; ++ // T __atomic_nand_fetch_N(T *mem, T val, int order) ++ case AtomicExpr::AO__atomic_nand_fetch: ++ LibCallName = "__atomic_nand_fetch"; ++ AddDirectArgument(*this, Args, UseOptimizedLibcall, Val1, MemTy, ++ E->getExprLoc(), sizeChars); ++ break; + } + + // Optimized functions have the size in their name. +Index: tools/clang/test/CodeGen/atomic-ops-libcall.c +=================================================================== +--- tools/clang/test/CodeGen/atomic-ops-libcall.c ++++ tools/clang/test/CodeGen/atomic-ops-libcall.c +@@ -35,3 +35,75 @@ int *fp2a(int **p) { + // Note, the GNU builtins do not multiply by sizeof(T)! + return __atomic_fetch_sub(p, 4, memory_order_relaxed); + } ++ ++int test_atomic_fetch_add(int *p) { ++ // CHECK: test_atomic_fetch_add ++ // CHECK: {{%[^ ]*}} = tail call i32 @__atomic_fetch_add_4(i8* {{%[0-9]+}}, i32 55, i32 5) ++ return __atomic_fetch_add(p, 55, memory_order_seq_cst); ++} ++ ++int test_atomic_fetch_sub(int *p) { ++ // CHECK: test_atomic_fetch_sub ++ // CHECK: {{%[^ ]*}} = tail call i32 @__atomic_fetch_sub_4(i8* {{%[0-9]+}}, i32 55, i32 5) ++ return __atomic_fetch_sub(p, 55, memory_order_seq_cst); ++} ++ ++int test_atomic_fetch_and(int *p) { ++ // CHECK: test_atomic_fetch_and ++ // CHECK: {{%[^ ]*}} = tail call i32 @__atomic_fetch_and_4(i8* {{%[0-9]+}}, i32 55, i32 5) ++ return __atomic_fetch_and(p, 55, memory_order_seq_cst); ++} ++ ++int test_atomic_fetch_or(int *p) { ++ // CHECK: test_atomic_fetch_or ++ // CHECK: {{%[^ ]*}} = tail call i32 @__atomic_fetch_or_4(i8* {{%[0-9]+}}, i32 55, i32 5) ++ return __atomic_fetch_or(p, 55, memory_order_seq_cst); ++} ++ ++int test_atomic_fetch_xor(int *p) { ++ // CHECK: test_atomic_fetch_xor ++ // CHECK: {{%[^ ]*}} = tail call i32 @__atomic_fetch_xor_4(i8* {{%[0-9]+}}, i32 55, i32 5) ++ return __atomic_fetch_xor(p, 55, memory_order_seq_cst); ++} ++ ++int test_atomic_fetch_nand(int *p) { ++ // CHECK: test_atomic_fetch_nand ++ // CHECK: {{%[^ ]*}} = tail call i32 @__atomic_fetch_nand_4(i8* {{%[0-9]+}}, i32 55, i32 5) ++ return __atomic_fetch_nand(p, 55, memory_order_seq_cst); ++} ++ ++int test_atomic_add_fetch(int *p) { ++ // CHECK: test_atomic_add_fetch ++ // CHECK: {{%[^ ]*}} = tail call i32 @__atomic_add_fetch_4(i8* {{%[0-9]+}}, i32 55, i32 5) ++ return __atomic_add_fetch(p, 55, memory_order_seq_cst); ++} ++ ++int test_atomic_sub_fetch(int *p) { ++ // CHECK: test_atomic_sub_fetch ++ // CHECK: {{%[^ ]*}} = tail call i32 @__atomic_sub_fetch_4(i8* {{%[0-9]+}}, i32 55, i32 5) ++ return __atomic_sub_fetch(p, 55, memory_order_seq_cst); ++} ++ ++int test_atomic_and_fetch(int *p) { ++ // CHECK: test_atomic_and_fetch ++ // CHECK: {{%[^ ]*}} = tail call i32 @__atomic_and_fetch_4(i8* {{%[0-9]+}}, i32 55, i32 5) ++ return __atomic_and_fetch(p, 55, memory_order_seq_cst); ++} ++ ++int test_atomic_or_fetch(int *p) { ++ // CHECK: test_atomic_or_fetch ++ // CHECK: {{%[^ ]*}} = tail call i32 @__atomic_or_fetch_4(i8* {{%[0-9]+}}, i32 55, i32 5) ++ return __atomic_or_fetch(p, 55, memory_order_seq_cst); ++} ++ ++int test_atomic_xor_fetch(int *p) { ++ // CHECK: test_atomic_xor_fetch ++ // CHECK: {{%[^ ]*}} = tail call i32 @__atomic_xor_fetch_4(i8* {{%[0-9]+}}, i32 55, i32 5) ++ return __atomic_xor_fetch(p, 55, memory_order_seq_cst); ++} ++ ++int test_atomic_nand_fetch(int *p) { ++ // CHECK: test_atomic_nand_fetch ++ // CHECK: {{%[^ ]*}} = tail call i32 @__atomic_nand_fetch_4(i8* {{%[0-9]+}}, i32 55, i32 5) ++ return __atomic_nand_fetch(p, 55, memory_order_seq_cst); ++}