• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 //===-- stack_depot.h -------------------------------------------*- C++ -*-===//
2 //
3 // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4 // See https://llvm.org/LICENSE.txt for license information.
5 // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6 //
7 //===----------------------------------------------------------------------===//
8 
9 #ifndef SCUDO_STACK_DEPOT_H_
10 #define SCUDO_STACK_DEPOT_H_
11 
12 #include "atomic_helpers.h"
13 #include "mutex.h"
14 
15 namespace scudo {
16 
17 class MurMur2HashBuilder {
18   static const u32 M = 0x5bd1e995;
19   static const u32 Seed = 0x9747b28c;
20   static const u32 R = 24;
21   u32 H;
22 
23 public:
24   explicit MurMur2HashBuilder(u32 Init = 0) { H = Seed ^ Init; }
add(u32 K)25   void add(u32 K) {
26     K *= M;
27     K ^= K >> R;
28     K *= M;
29     H *= M;
30     H ^= K;
31   }
get()32   u32 get() {
33     u32 X = H;
34     X ^= X >> 13;
35     X *= M;
36     X ^= X >> 15;
37     return X;
38   }
39 };
40 
41 class StackDepot {
42   HybridMutex RingEndMu;
43   u32 RingEnd = 0;
44 
45   // This data structure stores a stack trace for each allocation and
46   // deallocation when stack trace recording is enabled, that may be looked up
47   // using a hash of the stack trace. The lower bits of the hash are an index
48   // into the Tab array, which stores an index into the Ring array where the
49   // stack traces are stored. As the name implies, Ring is a ring buffer, so a
50   // stack trace may wrap around to the start of the array.
51   //
52   // Each stack trace in Ring is prefixed by a stack trace marker consisting of
53   // a fixed 1 bit in bit 0 (this allows disambiguation between stack frames
54   // and stack trace markers in the case where instruction pointers are 4-byte
55   // aligned, as they are on arm64), the stack trace hash in bits 1-32, and the
56   // size of the stack trace in bits 33-63.
57   //
58   // The insert() function is potentially racy in its accesses to the Tab and
59   // Ring arrays, but find() is resilient to races in the sense that, barring
60   // hash collisions, it will either return the correct stack trace or no stack
61   // trace at all, even if two instances of insert() raced with one another.
62   // This is achieved by re-checking the hash of the stack trace before
63   // returning the trace.
64 
65 #ifdef SCUDO_FUZZ
66   // Use smaller table sizes for fuzzing in order to reduce input size.
67   static const uptr TabBits = 4;
68 #else
69   static const uptr TabBits = 16;
70 #endif
71   static const uptr TabSize = 1 << TabBits;
72   static const uptr TabMask = TabSize - 1;
73   atomic_u32 Tab[TabSize] = {};
74 
75 #ifdef SCUDO_FUZZ
76   static const uptr RingBits = 4;
77 #else
78   static const uptr RingBits = 19;
79 #endif
80   static const uptr RingSize = 1 << RingBits;
81   static const uptr RingMask = RingSize - 1;
82   atomic_u64 Ring[RingSize] = {};
83 
84 public:
85   // Insert hash of the stack trace [Begin, End) into the stack depot, and
86   // return the hash.
insert(uptr * Begin,uptr * End)87   u32 insert(uptr *Begin, uptr *End) {
88     MurMur2HashBuilder B;
89     for (uptr *I = Begin; I != End; ++I)
90       B.add(u32(*I) >> 2);
91     u32 Hash = B.get();
92 
93     u32 Pos = Hash & TabMask;
94     u32 RingPos = atomic_load_relaxed(&Tab[Pos]);
95     u64 Entry = atomic_load_relaxed(&Ring[RingPos]);
96     u64 Id = (u64(End - Begin) << 33) | (u64(Hash) << 1) | 1;
97     if (Entry == Id)
98       return Hash;
99 
100     ScopedLock Lock(RingEndMu);
101     RingPos = RingEnd;
102     atomic_store_relaxed(&Tab[Pos], RingPos);
103     atomic_store_relaxed(&Ring[RingPos], Id);
104     for (uptr *I = Begin; I != End; ++I) {
105       RingPos = (RingPos + 1) & RingMask;
106       atomic_store_relaxed(&Ring[RingPos], *I);
107     }
108     RingEnd = (RingPos + 1) & RingMask;
109     return Hash;
110   }
111 
112   // Look up a stack trace by hash. Returns true if successful. The trace may be
113   // accessed via operator[] passing indexes between *RingPosPtr and
114   // *RingPosPtr + *SizePtr.
find(u32 Hash,uptr * RingPosPtr,uptr * SizePtr)115   bool find(u32 Hash, uptr *RingPosPtr, uptr *SizePtr) const {
116     u32 Pos = Hash & TabMask;
117     u32 RingPos = atomic_load_relaxed(&Tab[Pos]);
118     if (RingPos >= RingSize)
119       return false;
120     u64 Entry = atomic_load_relaxed(&Ring[RingPos]);
121     u64 HashWithTagBit = (u64(Hash) << 1) | 1;
122     if ((Entry & 0x1ffffffff) != HashWithTagBit)
123       return false;
124     u32 Size = u32(Entry >> 33);
125     if (Size >= RingSize)
126       return false;
127     *RingPosPtr = (RingPos + 1) & RingMask;
128     *SizePtr = Size;
129     MurMur2HashBuilder B;
130     for (uptr I = 0; I != Size; ++I) {
131       RingPos = (RingPos + 1) & RingMask;
132       B.add(u32(atomic_load_relaxed(&Ring[RingPos])) >> 2);
133     }
134     return B.get() == Hash;
135   }
136 
137   u64 operator[](uptr RingPos) const {
138     return atomic_load_relaxed(&Ring[RingPos & RingMask]);
139   }
140 };
141 
142 } // namespace scudo
143 
144 #endif // SCUDO_STACK_DEPOT_H_
145