From 53becd87b874cda5f3ae29e00538fb41d336dbb9 Mon Sep 17 00:00:00 2001 From: Timur Iskhodzhanov Date: Mon, 28 Jan 2013 17:29:50 +0000 Subject: [PATCH] Use the correct order of NOINLINE vs ret type to fix Windows build git-svn-id: https://llvm.org/svn/llvm-project/compiler-rt/trunk@173694 91177308-0d34-0410-b5e6-96231b3b80d8 --- lib/sanitizer_common/sanitizer_allocator.h | 14 +++++++------- 1 file changed, 7 insertions(+), 7 deletions(-) diff --git a/lib/sanitizer_common/sanitizer_allocator.h b/lib/sanitizer_common/sanitizer_allocator.h index 39c20b962..9d8812106 100644 --- a/lib/sanitizer_common/sanitizer_allocator.h +++ b/lib/sanitizer_common/sanitizer_allocator.h @@ -313,7 +313,7 @@ class SizeClassAllocator64 { alignment <= SizeClassMap::kMaxSize; } - Batch *NOINLINE AllocateBatch(AllocatorStats *stat, AllocatorCache *c, + NOINLINE Batch* AllocateBatch(AllocatorStats *stat, AllocatorCache *c, uptr class_id) { CHECK_LT(class_id, kNumClasses); RegionInfo *region = GetRegionInfo(class_id); @@ -324,7 +324,7 @@ class SizeClassAllocator64 { return b; } - void NOINLINE DeallocateBatch(AllocatorStats *stat, uptr class_id, Batch *b) { + NOINLINE void DeallocateBatch(AllocatorStats *stat, uptr class_id, Batch *b) { RegionInfo *region = GetRegionInfo(class_id); region->free_list.Push(b); region->n_freed += b->count; @@ -451,7 +451,7 @@ class SizeClassAllocator64 { return offset / (u32)size; } - Batch *NOINLINE PopulateFreeList(AllocatorStats *stat, AllocatorCache *c, + NOINLINE Batch* PopulateFreeList(AllocatorStats *stat, AllocatorCache *c, uptr class_id, RegionInfo *region) { BlockingMutexLock l(®ion->mutex); Batch *b = region->free_list.Pop(); @@ -573,7 +573,7 @@ class SizeClassAllocator32 { return reinterpret_cast(meta); } - Batch *NOINLINE AllocateBatch(AllocatorStats *stat, AllocatorCache *c, + NOINLINE Batch* AllocateBatch(AllocatorStats *stat, AllocatorCache *c, uptr class_id) { CHECK_LT(class_id, kNumClasses); SizeClassInfo *sci = GetSizeClassInfo(class_id); @@ -586,7 +586,7 @@ class SizeClassAllocator32 { return b; } - void NOINLINE DeallocateBatch(AllocatorStats *stat, uptr class_id, Batch *b) { + NOINLINE void DeallocateBatch(AllocatorStats *stat, uptr class_id, Batch *b) { CHECK_LT(class_id, kNumClasses); SizeClassInfo *sci = GetSizeClassInfo(class_id); SpinMutexLock l(&sci->mutex); @@ -782,7 +782,7 @@ struct SizeClassAllocatorLocalCache { } } - void NOINLINE Refill(SizeClassAllocator *allocator, uptr class_id) { + NOINLINE void Refill(SizeClassAllocator *allocator, uptr class_id) { InitCache(); PerClass *c = &per_class_[class_id]; Batch *b = allocator->AllocateBatch(&stats_, this, class_id); @@ -794,7 +794,7 @@ struct SizeClassAllocatorLocalCache { Deallocate(allocator, SizeClassMap::ClassID(sizeof(Batch)), b); } - void NOINLINE Drain(SizeClassAllocator *allocator, uptr class_id) { + NOINLINE void Drain(SizeClassAllocator *allocator, uptr class_id) { InitCache(); PerClass *c = &per_class_[class_id]; Batch *b; -- 2.11.4.GIT