1 // Copyright (c) 2009-2010 Satoshi Nakamoto
2 // Copyright (c) 2009-2013 The Bitcoin developers
3 // Distributed under the MIT/X11 software license, see the accompanying
4 // file COPYING or http://www.opensource.org/licenses/mit-license.php.
6 #ifndef BITCOIN_ALLOCATORS_H
7 #define BITCOIN_ALLOCATORS_H
14 #include <boost/thread/mutex.hpp>
15 #include <boost/thread/once.hpp>
17 #include <openssl/crypto.h> // for OPENSSL_cleanse()
20 * Thread-safe class to keep track of locked (ie, non-swappable) memory pages.
22 * Memory locks do not stack, that is, pages which have been locked several times by calls to mlock()
23 * will be unlocked by a single call to munlock(). This can result in keying material ending up in swap when
24 * those functions are used naively. This class simulates stacking memory locks by keeping a counter per page.
26 * @note By using a map from each page base address to lock count, this class is optimized for
27 * small objects that span up to a few pages, mostly smaller than a page. To support large allocations,
28 * something like an interval tree would be the preferred data structure.
30 template <class Locker>
31 class LockedPageManagerBase
34 LockedPageManagerBase(size_t page_size) : page_size(page_size)
36 // Determine bitmask for extracting page from address
37 assert(!(page_size & (page_size - 1))); // size must be power of two
38 page_mask = ~(page_size - 1);
41 ~LockedPageManagerBase()
43 assert(this->GetLockedPageCount() == 0);
47 // For all pages in affected range, increase lock count
48 void LockRange(void* p, size_t size)
50 boost::mutex::scoped_lock lock(mutex);
53 const size_t base_addr = reinterpret_cast<size_t>(p);
54 const size_t start_page = base_addr & page_mask;
55 const size_t end_page = (base_addr + size - 1) & page_mask;
56 for (size_t page = start_page; page <= end_page; page += page_size) {
57 Histogram::iterator it = histogram.find(page);
58 if (it == histogram.end()) // Newly locked page
60 locker.Lock(reinterpret_cast<void*>(page), page_size);
61 histogram.insert(std::make_pair(page, 1));
62 } else // Page was already locked; increase counter
69 // For all pages in affected range, decrease lock count
70 void UnlockRange(void* p, size_t size)
72 boost::mutex::scoped_lock lock(mutex);
75 const size_t base_addr = reinterpret_cast<size_t>(p);
76 const size_t start_page = base_addr & page_mask;
77 const size_t end_page = (base_addr + size - 1) & page_mask;
78 for (size_t page = start_page; page <= end_page; page += page_size) {
79 Histogram::iterator it = histogram.find(page);
80 assert(it != histogram.end()); // Cannot unlock an area that was not locked
81 // Decrease counter for page, when it is zero, the page will be unlocked
83 if (it->second == 0) // Nothing on the page anymore that keeps it locked
85 // Unlock page and remove the count from histogram
86 locker.Unlock(reinterpret_cast<void*>(page), page_size);
92 // Get number of locked pages for diagnostics
93 int GetLockedPageCount()
95 boost::mutex::scoped_lock lock(mutex);
96 return histogram.size();
102 size_t page_size, page_mask;
103 // map of page base address to lock count
104 typedef std::map<size_t, int> Histogram;
110 * OS-dependent memory page locking/unlocking.
111 * Defined as policy class to make stubbing for test possible.
113 class MemoryPageLocker
116 /** Lock memory pages.
117 * addr and len must be a multiple of the system page size
119 bool Lock(const void* addr, size_t len);
120 /** Unlock memory pages.
121 * addr and len must be a multiple of the system page size
123 bool Unlock(const void* addr, size_t len);
127 * Singleton class to keep track of locked (ie, non-swappable) memory pages, for use in
128 * std::allocator templates.
130 * Some implementations of the STL allocate memory in some constructors (i.e., see
131 * MSVC's vector<T> implementation where it allocates 1 byte of memory in the allocator.)
132 * Due to the unpredictable order of static initializers, we have to make sure the
133 * LockedPageManager instance exists before any other STL-based objects that use
134 * secure_allocator are created. So instead of having LockedPageManager also be
135 * static-initialized, it is created on demand.
137 class LockedPageManager : public LockedPageManagerBase<MemoryPageLocker>
140 static LockedPageManager& Instance()
142 boost::call_once(LockedPageManager::CreateInstance, LockedPageManager::init_flag);
143 return *LockedPageManager::_instance;
149 static void CreateInstance()
151 // Using a local static instance guarantees that the object is initialized
152 // when it's first needed and also deinitialized after all objects that use
153 // it are done with it. I can think of one unlikely scenario where we may
154 // have a static deinitialization order/problem, but the check in
155 // LockedPageManagerBase's destructor helps us detect if that ever happens.
156 static LockedPageManager instance;
157 LockedPageManager::_instance = &instance;
160 static LockedPageManager* _instance;
161 static boost::once_flag init_flag;
165 // Functions for directly locking/unlocking memory objects.
166 // Intended for non-dynamically allocated structures.
168 template <typename T>
169 void LockObject(const T& t)
171 LockedPageManager::Instance().LockRange((void*)(&t), sizeof(T));
174 template <typename T>
175 void UnlockObject(const T& t)
177 OPENSSL_cleanse((void*)(&t), sizeof(T));
178 LockedPageManager::Instance().UnlockRange((void*)(&t), sizeof(T));
182 // Allocator that locks its contents from being paged
183 // out of memory and clears its contents before deletion.
185 template <typename T>
186 struct secure_allocator : public std::allocator<T> {
187 // MSVC8 default copy constructor is broken
188 typedef std::allocator<T> base;
189 typedef typename base::size_type size_type;
190 typedef typename base::difference_type difference_type;
191 typedef typename base::pointer pointer;
192 typedef typename base::const_pointer const_pointer;
193 typedef typename base::reference reference;
194 typedef typename base::const_reference const_reference;
195 typedef typename base::value_type value_type;
196 secure_allocator() throw() {}
197 secure_allocator(const secure_allocator& a) throw() : base(a) {}
198 template <typename U>
199 secure_allocator(const secure_allocator<U>& a) throw() : base(a)
202 ~secure_allocator() throw() {}
203 template <typename _Other>
205 typedef secure_allocator<_Other> other;
208 T* allocate(std::size_t n, const void* hint = 0)
211 p = std::allocator<T>::allocate(n, hint);
213 LockedPageManager::Instance().LockRange(p, sizeof(T) * n);
217 void deallocate(T* p, std::size_t n)
220 OPENSSL_cleanse(p, sizeof(T) * n);
221 LockedPageManager::Instance().UnlockRange(p, sizeof(T) * n);
223 std::allocator<T>::deallocate(p, n);
229 // Allocator that clears its contents before deletion.
231 template <typename T>
232 struct zero_after_free_allocator : public std::allocator<T> {
233 // MSVC8 default copy constructor is broken
234 typedef std::allocator<T> base;
235 typedef typename base::size_type size_type;
236 typedef typename base::difference_type difference_type;
237 typedef typename base::pointer pointer;
238 typedef typename base::const_pointer const_pointer;
239 typedef typename base::reference reference;
240 typedef typename base::const_reference const_reference;
241 typedef typename base::value_type value_type;
242 zero_after_free_allocator() throw() {}
243 zero_after_free_allocator(const zero_after_free_allocator& a) throw() : base(a) {}
244 template <typename U>
245 zero_after_free_allocator(const zero_after_free_allocator<U>& a) throw() : base(a)
248 ~zero_after_free_allocator() throw() {}
249 template <typename _Other>
251 typedef zero_after_free_allocator<_Other> other;
254 void deallocate(T* p, std::size_t n)
257 OPENSSL_cleanse(p, sizeof(T) * n);
258 std::allocator<T>::deallocate(p, n);
262 // This is exactly like std::string, but with a custom allocator.
263 typedef std::basic_string<char, std::char_traits<char>, secure_allocator<char> > SecureString;
265 // Byte-vector that clears its contents before deletion.
266 typedef std::vector<char, zero_after_free_allocator<char> > CSerializeData;
268 #endif // BITCOIN_ALLOCATORS_H