[PATCH] Add ForEachChunk() to sanitizer allocators.
David Blaikie
dblaikie at gmail.com
Wed Mar 13 08:28:12 PDT 2013
On Mar 13, 2013 8:20 AM, "Sergey Matveev" <earthdok at google.com> wrote:
>
> Hi kcc, glider, samsonov,
>
> ForEachChunk() iterates over known chunks, passing each of them to the
> callback.
>
> http://llvm-reviews.chandlerc.com/D539
>
> Files:
> lib/sanitizer_common/sanitizer_allocator.h
> lib/sanitizer_common/tests/sanitizer_allocator_test.cc
>
> Index: lib/sanitizer_common/sanitizer_allocator.h
> ===================================================================
> --- lib/sanitizer_common/sanitizer_allocator.h
> +++ lib/sanitizer_common/sanitizer_allocator.h
> @@ -433,6 +433,24 @@
> }
> }
>
> + // Iterate over existing chunks. May include chunks that are not
currently
> + // allocated to the user (e.g. freed).
> + // The caller is expected to do a ForceLock() before calling this
function.
> + void ForEachChunk(void (*callback)(void *p, void *arg), void
*argument) {
This is a rather C-ish API. Any reason it's done this way rather than more
idiomatically C++ with a template and a Callable? (The immediately obvious
benefit being type safety, though beyond that this sort of API would be
even more convenient from C++11 (lambdas, std::function, etc))
> + for (uptr class_id = 1; class_id < kNumClasses; class_id++) {
> + RegionInfo *region = GetRegionInfo(class_id);
> + uptr chunk_size = SizeClassMap::Size(class_id);
> + uptr region_beg = kSpaceBeg + class_id * kRegionSize;
> + for (uptr p = region_beg;
> + p < region_beg + region->allocated_user;
> + p += chunk_size)
> + {
> + // Too slow: CHECK_EQ((void *)p, GetBlockBegin((void *)p));
> + callback((void *)p, argument);
> + }
> + }
> + }
> +
> typedef SizeClassMap SizeClassMapT;
> static const uptr kNumClasses = SizeClassMap::kNumClasses;
> static const uptr kNumClassesRounded =
SizeClassMap::kNumClassesRounded;
> @@ -681,6 +699,25 @@
> }
> }
>
> + // Iterate over existing chunks. May include chunks that are not
currently
> + // allocated to the user (e.g. freed).
> + // The caller is expected to do a ForceLock() before calling this
function.
> + void ForEachChunk(void (*callback)(void *p, void *arg), void
*argument) {
> + for (uptr region = 0; region < kNumPossibleRegions; region++)
> + if (state_->possible_regions[region]) {
> + uptr chunk_size =
SizeClassMap::Size(state_->possible_regions[region]);
> + uptr max_chunks_in_region = kRegionSize / (chunk_size +
kMetadataSize);
> + uptr region_beg = region * kRegionSize;
> + for (uptr p = region_beg;
> + p < region_beg + max_chunks_in_region * chunk_size;
> + p += chunk_size)
> + {
> + // Too slow: CHECK_EQ((void *)p, GetBlockBegin((void *)p));
> + callback((void *)p, argument);
> + }
> + }
> + }
> +
> void PrintStats() {
> }
>
> @@ -1005,6 +1042,14 @@
> mutex_.Unlock();
> }
>
> + // Iterate over existing chunks. May include chunks that are not
currently
> + // allocated to the user (e.g. freed).
> + // The caller is expected to do a ForceLock() before calling this
function.
> + void ForEachChunk(void (*callback)(void *p, void *arg), void
*argument) {
> + for (uptr i = 0; i < n_chunks_; i++)
> + callback(GetUser(chunks_[i]), argument);
> + }
> +
> private:
> static const int kMaxNumChunks = 1 << FIRST_32_SECOND_64(15, 18);
> struct Header {
> @@ -1168,6 +1213,14 @@
> primary_.ForceUnlock();
> }
>
> + // Iterate over existing chunks. May include chunks that are not
currently
> + // allocated to the user (e.g. freed).
> + // The caller is expected to do a ForceLock() before calling this
function.
> + void ForEachChunk(void (*callback)(void *p, void *arg), void
*argument) {
> + primary_.ForEachChunk(callback, argument);
> + secondary_.ForEachChunk(callback, argument);
> + }
> +
> private:
> PrimaryAllocator primary_;
> SecondaryAllocator secondary_;
> Index: lib/sanitizer_common/tests/sanitizer_allocator_test.cc
> ===================================================================
> --- lib/sanitizer_common/tests/sanitizer_allocator_test.cc
> +++ lib/sanitizer_common/tests/sanitizer_allocator_test.cc
> @@ -22,6 +22,7 @@
> #include <pthread.h>
> #include <algorithm>
> #include <vector>
> +#include <set>
>
> // Too slow for debug build
> #if TSAN_DEBUG == 0
> @@ -565,4 +566,88 @@
> }
> }
>
> +void IterationTestCallback(void *chunk, void *argument) {
> + std::set<void *> *chunks = reinterpret_cast<std::set<void *>
*>(argument);
> + chunks->insert(chunk);
> +}
> +
> +template <class Allocator>
> +void TestSizeClassAllocatorIteration() {
> + Allocator *a = new Allocator;
> + a->Init();
> + SizeClassAllocatorLocalCache<Allocator> cache;
> + memset(&cache, 0, sizeof(cache));
> + cache.Init(0);
> +
> + static const uptr sizes[] = {1, 16, 30, 40, 100, 1000, 10000,
> + 50000, 60000, 100000, 120000, 300000, 500000, 1000000, 2000000};
> +
> + std::vector<void *> allocated;
> +
> + // Allocate a bunch of chunks.
> + for (uptr s = 0; s < ARRAY_SIZE(sizes); s++) {
> + uptr size = sizes[s];
> + if (!a->CanAllocate(size, 1)) continue;
> + // printf("s = %ld\n", size);
> + uptr n_iter = std::max((uptr)6, 80000 / size);
> + // fprintf(stderr, "size: %ld iter: %ld\n", size, n_iter);
> + for (uptr j = 0; j < n_iter; j++) {
> + uptr class_id0 = Allocator::SizeClassMapT::ClassID(size);
> + void *x = cache.Allocate(a, class_id0);
> + allocated.push_back(x);
> + }
> + }
> +
> + std::set<void *> reported_chunks;
> + a->ForceLock();
> + a->ForEachChunk(IterationTestCallback,
> + reinterpret_cast<void *>(&reported_chunks));
> + a->ForceUnlock();
> +
> + for (uptr i = 0; i < allocated.size(); i++) {
> + // Don't use EXPECT_NE. Reporting the first mismatch is enough.
> + ASSERT_NE(reported_chunks.find(allocated[i]), reported_chunks.end());
> + }
> +
> + a->TestOnlyUnmap();
> + delete a;
> +}
> +
> +#if SANITIZER_WORDSIZE == 64
> +TEST(SanitizerCommon, SizeClassAllocator64Iteration) {
> + TestSizeClassAllocatorIteration<Allocator64>();
> +}
> +#endif
> +
> +TEST(SanitizerCommon, SizeClassAllocator32Iteration) {
> + TestSizeClassAllocatorIteration<Allocator32Compact>();
> +}
> +
> +
> +TEST(SanitizerCommon, LargeMmapAllocatorIteration) {
> + LargeMmapAllocator<> a;
> + a.Init();
> + AllocatorStats stats;
> + stats.Init();
> +
> + static const int kNumAllocs = 1000;
> + char *allocated[kNumAllocs];
> + static const uptr size = 40;
> + // Allocate some.
> + for (int i = 0; i < kNumAllocs; i++) {
> + allocated[i] = (char *)a.Allocate(&stats, size, 1);
> + }
> +
> + std::set<void *> reported_chunks;
> + a.ForceLock();
> + a.ForEachChunk(IterationTestCallback,
> + reinterpret_cast<void *>(&reported_chunks));
> + a.ForceUnlock();
> +
> + for (uptr i = 0; i < kNumAllocs; i++) {
> + // Don't use EXPECT_NE. Reporting the first mismatch is enough.
> + ASSERT_NE(reported_chunks.find(allocated[i]), reported_chunks.end());
> + }
> +}
> +
> #endif // #if TSAN_DEBUG==0
>
> _______________________________________________
> llvm-commits mailing list
> llvm-commits at cs.uiuc.edu
> http://lists.cs.uiuc.edu/mailman/listinfo/llvm-commits
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.llvm.org/pipermail/llvm-commits/attachments/20130313/c456890d/attachment.html>
More information about the llvm-commits
mailing list