alloc_slab.hpp 29 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781
  1. /*************************************************************************
  2. *
  3. * Copyright 2016 Realm Inc.
  4. *
  5. * Licensed under the Apache License, Version 2.0 (the "License");
  6. * you may not use this file except in compliance with the License.
  7. * You may obtain a copy of the License at
  8. *
  9. * http://www.apache.org/licenses/LICENSE-2.0
  10. *
  11. * Unless required by applicable law or agreed to in writing, software
  12. * distributed under the License is distributed on an "AS IS" BASIS,
  13. * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  14. * See the License for the specific language governing permissions and
  15. * limitations under the License.
  16. *
  17. **************************************************************************/
  18. #ifndef REALM_ALLOC_SLAB_HPP
  19. #define REALM_ALLOC_SLAB_HPP
  20. #include <cstdint> // unint8_t etc
  21. #include <vector>
  22. #include <map>
  23. #include <string>
  24. #include <atomic>
  25. #include <mutex>
  26. #include <realm/util/features.h>
  27. #include <realm/util/file.hpp>
  28. #include <realm/util/thread.hpp>
  29. #include <realm/alloc.hpp>
  30. #include <realm/disable_sync_to_disk.hpp>
  31. namespace realm {
  32. // Pre-declarations
  33. class Group;
  34. class GroupWriter;
  35. namespace util {
  36. struct SharedFileInfo;
  37. }
  38. /// Thrown by Group and DB constructors if the specified file
  39. /// (or memory buffer) does not appear to contain a valid Realm
  40. /// database.
  41. struct InvalidDatabase;
  42. /// The allocator that is used to manage the memory of a Realm
  43. /// group, i.e., a Realm database.
  44. ///
  45. /// Optionally, it can be attached to an pre-existing database (file
  46. /// or memory buffer) which then becomes an immuatble part of the
  47. /// managed memory.
  48. ///
  49. /// To attach a slab allocator to a pre-existing database, call
  50. /// attach_file() or attach_buffer(). To create a new database
  51. /// in-memory, call attach_empty().
  52. ///
  53. /// For efficiency, this allocator manages its mutable memory as a set
  54. /// of slabs.
  55. class SlabAlloc : public Allocator {
  56. public:
  57. ~SlabAlloc() noexcept override;
  58. SlabAlloc();
  59. // Disable copying. Copying an allocator can produce double frees.
  60. SlabAlloc(const SlabAlloc&) = delete;
  61. SlabAlloc& operator=(const SlabAlloc&) = delete;
  62. /// \struct Config
  63. /// \brief Storage for combining setup flags for initialization to
  64. /// the SlabAlloc.
  65. ///
  66. /// \var Config::is_shared
  67. /// Must be true if, and only if we are called on behalf of DB.
  68. ///
  69. /// \var Config::read_only
  70. /// Open the file in read-only mode. This implies \a Config::no_create.
  71. ///
  72. /// \var Config::no_create
  73. /// Fail if the file does not already exist.
  74. ///
  75. /// \var Config::skip_validate
  76. /// Skip validation of file header. In a
  77. /// set of overlapping DBs, only the first one (the one
  78. /// that creates/initlializes the coordination file) may validate
  79. /// the header, otherwise it will result in a race condition.
  80. ///
  81. /// \var Config::encryption_key
  82. /// 32-byte key to use to encrypt and decrypt the backing storage,
  83. /// or nullptr to disable encryption.
  84. ///
  85. /// \var Config::session_initiator
  86. /// If set, the caller is the session initiator and
  87. /// guarantees exclusive access to the file. If attaching in
  88. /// read/write mode, the file is modified: files on streaming form
  89. /// is changed to non-streaming form, and if needed the file size
  90. /// is adjusted to match mmap boundaries.
  91. /// Must be set to false if is_shared is false.
  92. ///
  93. /// \var Config::clear_file
  94. /// Always initialize the file as if it was a newly
  95. /// created file and ignore any pre-existing contents. Requires that
  96. /// Config::session_initiator be true as well.
  97. struct Config {
  98. bool is_shared = false;
  99. bool read_only = false;
  100. bool no_create = false;
  101. bool skip_validate = false;
  102. bool session_initiator = false;
  103. bool clear_file = false;
  104. bool disable_sync = false;
  105. const char* encryption_key = nullptr;
  106. };
  107. struct Retry {
  108. };
  109. /// \brief Attach this allocator to the specified file.
  110. ///
  111. /// It is an error if this function is called at a time where the specified
  112. /// Realm file (file system inode) is modified asynchronously.
  113. ///
  114. /// In non-shared mode (when this function is called on behalf of a
  115. /// free-standing Group instance), it is the responsibility of the
  116. /// application to ensure that the Realm file is not modified concurrently
  117. /// from any other thread or process.
  118. ///
  119. /// In shared mode (when this function is called on behalf of a DB
  120. /// instance), the caller (DB::do_open()) must take steps to ensure
  121. /// cross-process mutual exclusion.
  122. ///
  123. /// Except for \a file_path, the parameters are passed in through a
  124. /// configuration object.
  125. ///
  126. /// \return The `ref` of the root node, or zero if there is none.
  127. ///
  128. /// Please note that attach_file can fail to attach to a file due to a
  129. /// collision with a writer extending the file. This can only happen if the
  130. /// caller is *not* the session initiator. When this happens, attach_file()
  131. /// throws SlabAlloc::Retry, and the caller must retry the call. The caller
  132. /// should check if it has become the session initiator before retrying.
  133. /// This can happen if the conflicting thread (or process) terminates or
  134. /// crashes before the next retry.
  135. ///
  136. /// \throw util::File::AccessError
  137. /// \throw SlabAlloc::Retry
  138. ref_type attach_file(const std::string& file_path, Config& cfg);
  139. /// Get the attached file. Only valid when called on an allocator with
  140. /// an attached file.
  141. util::File& get_file();
  142. /// Attach this allocator to the specified memory buffer.
  143. ///
  144. /// It is an error to call this function on an attached
  145. /// allocator. Doing so will result in undefined behavor.
  146. ///
  147. /// \return The `ref` of the root node, or zero if there is none.
  148. ///
  149. /// \sa own_buffer()
  150. ///
  151. /// \throw InvalidDatabase
  152. ref_type attach_buffer(const char* data, size_t size);
  153. /// Reads file format from file header. Must be called from within a write
  154. /// transaction.
  155. int get_committed_file_format_version() const noexcept;
  156. bool is_file_on_streaming_form() const
  157. {
  158. const Header& header = *reinterpret_cast<const Header*>(m_data);
  159. return is_file_on_streaming_form(header);
  160. }
  161. /// Attach this allocator to an empty buffer.
  162. ///
  163. /// It is an error to call this function on an attached
  164. /// allocator. Doing so will result in undefined behavor.
  165. void attach_empty();
  166. /// Detach from a previously attached file or buffer.
  167. ///
  168. /// This function does not reset free space tracking. To
  169. /// completely reset the allocator, you must also call
  170. /// reset_free_space_tracking().
  171. ///
  172. /// This function has no effect if the allocator is already in the
  173. /// detached state (idempotency).
  174. void detach() noexcept;
  175. class DetachGuard;
  176. /// If a memory buffer has been attached using attach_buffer(),
  177. /// mark it as owned by this slab allocator. Behaviour is
  178. /// undefined if this function is called on a detached allocator,
  179. /// one that is not attached using attach_buffer(), or one for
  180. /// which this function has already been called during the latest
  181. /// attachment.
  182. void own_buffer() noexcept;
  183. /// Returns true if, and only if this allocator is currently
  184. /// in the attached state.
  185. bool is_attached() const noexcept;
  186. /// Returns true if, and only if this allocator is currently in
  187. /// the attached state and attachment was not established using
  188. /// attach_empty().
  189. bool nonempty_attachment() const noexcept;
  190. /// Reserve disk space now to avoid allocation errors at a later
  191. /// point in time, and to minimize on-disk fragmentation. In some
  192. /// cases, less fragmentation translates into improved
  193. /// performance. On flash or SSD-drives this is likely a waste.
  194. ///
  195. /// Note: File::prealloc() may misbehave under race conditions (see
  196. /// documentation of File::prealloc()). For that reason, to avoid race
  197. /// conditions, when this allocator is used in a transactional mode, this
  198. /// function may be called only when the caller has exclusive write
  199. /// access. In non-transactional mode it is the responsibility of the user
  200. /// to ensure non-concurrent file mutation.
  201. ///
  202. /// This function will call File::sync().
  203. ///
  204. /// It is an error to call this function on an allocator that is not
  205. /// attached to a file. Doing so will result in undefined behavior.
  206. void resize_file(size_t new_file_size);
  207. #ifdef REALM_DEBUG
  208. /// Deprecated method, only called from a unit test
  209. ///
  210. /// WARNING: This method is NOT thread safe on multiple platforms; see
  211. /// File::prealloc().
  212. ///
  213. /// Reserve disk space now to avoid allocation errors at a later point in
  214. /// time, and to minimize on-disk fragmentation. In some cases, less
  215. /// fragmentation translates into improved performance. On SSD-drives
  216. /// preallocation is likely a waste.
  217. ///
  218. /// When supported by the system, a call to this function will make the
  219. /// database file at least as big as the specified size, and cause space on
  220. /// the target device to be allocated (note that on many systems on-disk
  221. /// allocation is done lazily by default). If the file is already bigger
  222. /// than the specified size, the size will be unchanged, and on-disk
  223. /// allocation will occur only for the initial section that corresponds to
  224. /// the specified size.
  225. ///
  226. /// This function will call File::sync() if it changes the size of the file.
  227. ///
  228. /// It is an error to call this function on an allocator that is not
  229. /// attached to a file. Doing so will result in undefined behavior.
  230. void reserve_disk_space(size_t size_in_bytes);
  231. #endif
  232. /// Get the size of the attached database file or buffer in number
  233. /// of bytes. This size is not affected by new allocations. After
  234. /// attachment, it can only be modified by a call to update_reader_view().
  235. ///
  236. /// It is an error to call this function on a detached allocator,
  237. /// or one that was attached using attach_empty(). Doing so will
  238. /// result in undefined behavior.
  239. size_t get_baseline() const noexcept;
  240. /// Get the total amount of managed memory. This is the baseline plus the
  241. /// sum of the sizes of the allocated slabs. It includes any free space.
  242. ///
  243. /// It is an error to call this function on a detached
  244. /// allocator. Doing so will result in undefined behavior.
  245. size_t get_total_size() const noexcept;
  246. /// Mark all mutable memory (ref-space outside the attached file) as free
  247. /// space.
  248. void reset_free_space_tracking();
  249. /// Update the readers view of the file:
  250. ///
  251. /// Remap the attached file such that a prefix of the specified
  252. /// size becomes available in memory. If sucessfull,
  253. /// get_baseline() will return the specified new file size.
  254. ///
  255. /// It is an error to call this function on a detached allocator,
  256. /// or one that was not attached using attach_file(). Doing so
  257. /// will result in undefined behavior.
  258. ///
  259. /// Updates the memory mappings to reflect a new size for the file.
  260. /// Stale mappings are retained so that they remain valid for other threads,
  261. /// which haven't yet seen the file size change. The stale mappings are
  262. /// associated with a version count if one is provided.
  263. /// They are later purged by calls to purge_old_mappings().
  264. /// The version parameter is subtly different from the mapping_version obtained
  265. /// by get_mapping_version() below. The mapping version changes whenever a
  266. /// ref->ptr translation changes, and is used by Group to enforce re-translation.
  267. void update_reader_view(size_t file_size);
  268. void purge_old_mappings(uint64_t oldest_live_version, uint64_t youngest_live_version);
  269. void init_mapping_management(uint64_t currently_live_version);
  270. /// Get an ID for the current mapping version. This ID changes whenever any part
  271. /// of an existing mapping is changed. Such a change requires all refs to be
  272. /// retranslated to new pointers. This will happen whenever the reader view
  273. /// is extended unless the old size was aligned to a section boundary.
  274. uint64_t get_mapping_version()
  275. {
  276. return m_mapping_version;
  277. }
  278. /// Returns true initially, and after a call to reset_free_space_tracking()
  279. /// up until the point of the first call to SlabAlloc::alloc(). Note that a
  280. /// call to SlabAlloc::alloc() corresponds to a mutation event.
  281. bool is_free_space_clean() const noexcept;
  282. /// Returns the amount of memory requested by calls to SlabAlloc::alloc().
  283. size_t get_commit_size() const
  284. {
  285. return m_commit_size;
  286. }
  287. /// Returns the total amount of memory currently allocated in slab area
  288. size_t get_allocated_size() const noexcept;
  289. /// Returns total amount of slab for all slab allocators
  290. static size_t get_total_slab_size() noexcept;
  291. /// Hooks used to keep the encryption layer informed of the start and stop
  292. /// of transactions.
  293. void note_reader_start(const void* reader_id);
  294. void note_reader_end(const void* reader_id) noexcept;
  295. void verify() const override;
  296. #ifdef REALM_DEBUG
  297. void enable_debug(bool enable)
  298. {
  299. m_debug_out = enable;
  300. }
  301. bool is_all_free() const;
  302. void print() const;
  303. #endif
  304. protected:
  305. MemRef do_alloc(const size_t size) override;
  306. MemRef do_realloc(ref_type, char*, size_t old_size, size_t new_size) override;
  307. void do_free(ref_type, char*) override;
  308. char* do_translate(ref_type) const noexcept override;
  309. /// Returns the first section boundary *above* the given position.
  310. size_t get_upper_section_boundary(size_t start_pos) const noexcept;
  311. /// Returns the section boundary at or above the given size
  312. size_t align_size_to_section_boundary(size_t size) const noexcept;
  313. /// Returns the first section boundary *at or below* the given position.
  314. size_t get_lower_section_boundary(size_t start_pos) const noexcept;
  315. /// Returns true if the given position is at a section boundary
  316. bool matches_section_boundary(size_t pos) const noexcept;
  317. /// Actually compute the starting offset of a section. Only used to initialize
  318. /// a table of predefined results, which are then used by get_section_base().
  319. size_t compute_section_base(size_t index) const noexcept;
  320. /// Find a possible allocation of 'request_size' that will fit into a section
  321. /// which is inside the range from 'start_pos' to 'start_pos'+'free_chunk_size'
  322. /// If found return the position, if not return 0.
  323. size_t find_section_in_range(size_t start_pos, size_t free_chunk_size, size_t request_size) const noexcept;
  324. private:
  325. enum AttachMode {
  326. attach_None, // Nothing is attached
  327. attach_OwnedBuffer, // We own the buffer (m_data = nullptr for empty buffer)
  328. attach_UsersBuffer, // We do not own the buffer
  329. attach_SharedFile, // On behalf of DB
  330. attach_UnsharedFile // Not on behalf of DB
  331. };
  332. // A slab is a dynamically allocated contiguous chunk of memory used to
  333. // extend the amount of space available for database node
  334. // storage. Inter-node references are represented as file offsets
  335. // (a.k.a. "refs"), and each slab creates an apparently seamless extension
  336. // of this file offset addressable space. Slabs are stored as rows in the
  337. // Slabs table in order of ascending file offsets.
  338. struct Slab {
  339. ref_type ref_end;
  340. char* addr;
  341. size_t size;
  342. Slab(ref_type r, size_t s);
  343. ~Slab();
  344. Slab(const Slab&) = delete;
  345. Slab(Slab&& other) noexcept
  346. : ref_end(other.ref_end)
  347. , addr(other.addr)
  348. , size(other.size)
  349. {
  350. other.addr = nullptr;
  351. other.size = 0;
  352. other.ref_end = 0;
  353. }
  354. Slab& operator=(const Slab&) = delete;
  355. Slab& operator=(Slab&&) = delete;
  356. };
  357. // free blocks that are in the slab area are managed using the following structures:
  358. // - FreeBlock: Placed at the start of any free space. Holds the 'ref' corresponding to
  359. // the start of the space, and prev/next links used to place it in a size-specific
  360. // freelist.
  361. // - BetweenBlocks: Structure sitting between any two free OR allocated spaces.
  362. // describes the size of the space before and after.
  363. // Each slab (area obtained from the underlying system) has a terminating BetweenBlocks
  364. // at the beginning and at the end of the Slab.
  365. struct FreeBlock {
  366. ref_type ref; // ref for this entry. Saves a reverse translate / representing links as refs
  367. FreeBlock* prev; // circular doubly linked list
  368. FreeBlock* next;
  369. void clear_links()
  370. {
  371. prev = next = nullptr;
  372. }
  373. void unlink();
  374. };
  375. struct BetweenBlocks { // stores sizes and used/free status of blocks before and after.
  376. int32_t block_before_size; // negated if block is in use,
  377. int32_t block_after_size; // positive if block is free - and zero at end
  378. };
  379. Config m_cfg;
  380. using FreeListMap = std::map<int, FreeBlock*>; // log(N) addressing for larger blocks
  381. FreeListMap m_block_map;
  382. // abstract notion of a freelist - used to hide whether a freelist
  383. // is residing in the small blocks or the large blocks structures.
  384. struct FreeList {
  385. int size = 0; // size of every element in the list, 0 if not found
  386. FreeListMap::iterator it;
  387. bool found_something()
  388. {
  389. return size != 0;
  390. }
  391. bool found_exact(int sz)
  392. {
  393. return size == sz;
  394. }
  395. };
  396. // simple helper functions for accessing/navigating blocks and betweenblocks (TM)
  397. BetweenBlocks* bb_before(FreeBlock* entry) const
  398. {
  399. return reinterpret_cast<BetweenBlocks*>(entry) - 1;
  400. }
  401. BetweenBlocks* bb_after(FreeBlock* entry) const
  402. {
  403. auto bb = bb_before(entry);
  404. size_t sz = bb->block_after_size;
  405. char* addr = reinterpret_cast<char*>(entry) + sz;
  406. return reinterpret_cast<BetweenBlocks*>(addr);
  407. }
  408. FreeBlock* block_before(BetweenBlocks* bb) const
  409. {
  410. size_t sz = bb->block_before_size;
  411. if (sz <= 0)
  412. return nullptr; // only blocks that are not in use
  413. char* addr = reinterpret_cast<char*>(bb) - sz;
  414. return reinterpret_cast<FreeBlock*>(addr);
  415. }
  416. FreeBlock* block_after(BetweenBlocks* bb) const
  417. {
  418. if (bb->block_after_size <= 0)
  419. return nullptr;
  420. return reinterpret_cast<FreeBlock*>(bb + 1);
  421. }
  422. int size_from_block(FreeBlock* entry) const
  423. {
  424. return bb_before(entry)->block_after_size;
  425. }
  426. void mark_allocated(FreeBlock* entry);
  427. // mark the entry freed in bordering BetweenBlocks. Also validate size.
  428. void mark_freed(FreeBlock* entry, int size);
  429. // hook for the memory verifier in Group.
  430. template <typename Func>
  431. void for_all_free_entries(Func f) const;
  432. // Main entry points for alloc/free:
  433. FreeBlock* allocate_block(int size);
  434. void free_block(ref_type ref, FreeBlock* addr);
  435. // Searching/manipulating freelists
  436. FreeList find(int size);
  437. FreeList find_larger(FreeList hint, int size);
  438. FreeBlock* pop_freelist_entry(FreeList list);
  439. void push_freelist_entry(FreeBlock* entry);
  440. void remove_freelist_entry(FreeBlock* element);
  441. void rebuild_freelists_from_slab();
  442. void clear_freelists();
  443. // grow the slab area.
  444. // returns a free block large enough to handle the request.
  445. FreeBlock* grow_slab(int size);
  446. // create a single free chunk with "BetweenBlocks" at both ends and a
  447. // single free chunk between them. This free chunk will be of size:
  448. // slab_size - 2 * sizeof(BetweenBlocks)
  449. FreeBlock* slab_to_entry(const Slab& slab, ref_type ref_start);
  450. // breaking/merging of blocks
  451. FreeBlock* get_prev_block_if_mergeable(FreeBlock* block);
  452. FreeBlock* get_next_block_if_mergeable(FreeBlock* block);
  453. // break 'block' to give it 'new_size'. Return remaining block.
  454. // If the block is too small to split, return nullptr.
  455. FreeBlock* break_block(FreeBlock* block, int new_size);
  456. FreeBlock* merge_blocks(FreeBlock* first, FreeBlock* second);
  457. // Values of each used bit in m_flags
  458. enum {
  459. flags_SelectBit = 1,
  460. };
  461. // 24 bytes
  462. struct Header {
  463. uint64_t m_top_ref[2]; // 2 * 8 bytes
  464. // Info-block 8-bytes
  465. uint8_t m_mnemonic[4]; // "T-DB"
  466. uint8_t m_file_format[2]; // See `library_file_format`
  467. uint8_t m_reserved;
  468. // bit 0 of m_flags is used to select between the two top refs.
  469. uint8_t m_flags;
  470. };
  471. // 16 bytes
  472. struct StreamingFooter {
  473. uint64_t m_top_ref;
  474. uint64_t m_magic_cookie;
  475. };
  476. // Description of to-be-deleted memory mapping
  477. struct OldMapping {
  478. uint64_t replaced_at_version;
  479. util::File::Map<char> mapping;
  480. };
  481. struct OldRefTranslation {
  482. OldRefTranslation(uint64_t v, size_t c, RefTranslation* m) noexcept
  483. : replaced_at_version(v)
  484. , translation_count(c)
  485. , translations(m)
  486. {
  487. }
  488. uint64_t replaced_at_version;
  489. size_t translation_count;
  490. std::unique_ptr<RefTranslation[]> translations;
  491. };
  492. static_assert(sizeof(Header) == 24, "Bad header size");
  493. static_assert(sizeof(StreamingFooter) == 16, "Bad footer size");
  494. static const Header empty_file_header;
  495. static void init_streaming_header(Header*, int file_format_version);
  496. static const uint_fast64_t footer_magic_cookie = 0x3034125237E526C8ULL;
  497. util::RaceDetector changes;
  498. void verify_old_translations(uint64_t verify_old_translations);
  499. // mappings used by newest transactions - additional mappings may be open
  500. // and in use by older transactions. These translations are in m_old_mappings.
  501. struct MapEntry {
  502. util::File::Map<char> primary_mapping;
  503. size_t lowest_possible_xover_offset = 0;
  504. util::File::Map<char> xover_mapping;
  505. };
  506. std::vector<MapEntry> m_mappings;
  507. size_t m_translation_table_size = 0;
  508. std::atomic<uint64_t> m_mapping_version = 1;
  509. uint64_t m_youngest_live_version = 1;
  510. std::mutex m_mapping_mutex;
  511. util::File m_file;
  512. util::SharedFileInfo* m_realm_file_info = nullptr;
  513. // vectors where old mappings, are held from deletion to ensure translations are
  514. // kept open and ref->ptr translations work for other threads..
  515. std::vector<OldMapping> m_old_mappings;
  516. std::vector<OldRefTranslation> m_old_translations;
  517. // Rebuild the ref translations in a thread-safe manner. Save the old one along with it's
  518. // versioning information for later deletion - 'requires_new_fast_mapping' must be
  519. // true if there are changes to entries among the existing translations. Must be called
  520. // with m_mapping_mutex locked.
  521. void rebuild_translations(bool requires_new_fast_mapping, size_t old_num_sections);
  522. // Add a translation covering a new section in the slab area. The translation is always
  523. // added at the end.
  524. void extend_fast_mapping_with_slab(char* address);
  525. void get_or_add_xover_mapping(RefTranslation& txl, size_t index, size_t offset, size_t size) override;
  526. const char* m_data = nullptr;
  527. size_t m_initial_section_size = 0;
  528. int m_section_shifts = 0;
  529. AttachMode m_attach_mode = attach_None;
  530. enum FeeeSpaceState {
  531. free_space_Clean,
  532. free_space_Dirty,
  533. free_space_Invalid,
  534. };
  535. constexpr static int minimal_alloc = 128 * 1024;
  536. constexpr static int maximal_alloc = 1 << section_shift;
  537. /// When set to free_space_Invalid, the free lists are no longer
  538. /// up-to-date. This happens if do_free() or
  539. /// reset_free_space_tracking() fails, presumably due to
  540. /// std::bad_alloc being thrown during updating of the free space
  541. /// list. In this this case, alloc(), realloc_(), and
  542. /// get_free_read_only() must throw. This member is deliberately
  543. /// placed here (after m_attach_mode) in the hope that it leads to
  544. /// less padding between members due to alignment requirements.
  545. FeeeSpaceState m_free_space_state = free_space_Clean;
  546. typedef std::vector<Slab> Slabs;
  547. using Chunks = std::map<ref_type, size_t>;
  548. Slabs m_slabs;
  549. Chunks m_free_read_only;
  550. size_t m_commit_size = 0;
  551. bool m_debug_out = false;
  552. /// Throws if free-lists are no longer valid.
  553. size_t consolidate_free_read_only();
  554. /// Throws if free-lists are no longer valid.
  555. const Chunks& get_free_read_only() const;
  556. /// Throws InvalidDatabase if the file is not a Realm file, if the file is
  557. /// corrupted, or if the specified encryption key is incorrect. This
  558. /// function will not detect all forms of corruption, though.
  559. /// Returns the top_ref for the latest commit.
  560. ref_type validate_header(const char* data, size_t len, const std::string& path);
  561. ref_type validate_header(const Header* header, const StreamingFooter* footer, size_t size,
  562. const std::string& path);
  563. void throw_header_exception(std::string msg, const Header& header, const std::string& path);
  564. static bool is_file_on_streaming_form(const Header& header);
  565. /// Read the top_ref from the given buffer and set m_file_on_streaming_form
  566. /// if the buffer contains a file in streaming form
  567. static ref_type get_top_ref(const char* data, size_t len);
  568. // Gets the path of the attached file, or other relevant debugging info.
  569. std::string get_file_path_for_assertions() const;
  570. static bool ref_less_than_slab_ref_end(ref_type, const Slab&) noexcept;
  571. friend class Group;
  572. friend class DB;
  573. friend class GroupWriter;
  574. };
  575. class SlabAlloc::DetachGuard {
  576. public:
  577. DetachGuard(SlabAlloc& alloc) noexcept
  578. : m_alloc(&alloc)
  579. {
  580. }
  581. ~DetachGuard() noexcept;
  582. SlabAlloc* release() noexcept;
  583. private:
  584. SlabAlloc* m_alloc;
  585. };
  586. // Implementation:
  587. struct InvalidDatabase : util::File::AccessError {
  588. InvalidDatabase(const std::string& msg, const std::string& path)
  589. : util::File::AccessError(msg, path)
  590. {
  591. }
  592. };
  593. inline void SlabAlloc::own_buffer() noexcept
  594. {
  595. REALM_ASSERT_3(m_attach_mode, ==, attach_UsersBuffer);
  596. REALM_ASSERT(m_data);
  597. m_attach_mode = attach_OwnedBuffer;
  598. }
  599. inline bool SlabAlloc::is_attached() const noexcept
  600. {
  601. return m_attach_mode != attach_None;
  602. }
  603. inline bool SlabAlloc::nonempty_attachment() const noexcept
  604. {
  605. return is_attached() && m_data;
  606. }
  607. inline size_t SlabAlloc::get_baseline() const noexcept
  608. {
  609. REALM_ASSERT_DEBUG(is_attached());
  610. return m_baseline.load(std::memory_order_relaxed);
  611. }
  612. inline bool SlabAlloc::is_free_space_clean() const noexcept
  613. {
  614. return m_free_space_state == free_space_Clean;
  615. }
  616. inline SlabAlloc::DetachGuard::~DetachGuard() noexcept
  617. {
  618. if (m_alloc)
  619. m_alloc->detach();
  620. }
  621. inline SlabAlloc* SlabAlloc::DetachGuard::release() noexcept
  622. {
  623. SlabAlloc* alloc = m_alloc;
  624. m_alloc = nullptr;
  625. return alloc;
  626. }
  627. inline bool SlabAlloc::ref_less_than_slab_ref_end(ref_type ref, const Slab& slab) noexcept
  628. {
  629. return ref < slab.ref_end;
  630. }
  631. inline size_t SlabAlloc::get_upper_section_boundary(size_t start_pos) const noexcept
  632. {
  633. return get_section_base(1 + get_section_index(start_pos));
  634. }
  635. inline size_t SlabAlloc::align_size_to_section_boundary(size_t size) const noexcept
  636. {
  637. if (matches_section_boundary(size))
  638. return size;
  639. else
  640. return get_upper_section_boundary(size);
  641. }
  642. inline size_t SlabAlloc::get_lower_section_boundary(size_t start_pos) const noexcept
  643. {
  644. return get_section_base(get_section_index(start_pos));
  645. }
  646. inline bool SlabAlloc::matches_section_boundary(size_t pos) const noexcept
  647. {
  648. auto boundary = get_lower_section_boundary(pos);
  649. return pos == boundary;
  650. }
  651. template <typename Func>
  652. void SlabAlloc::for_all_free_entries(Func f) const
  653. {
  654. ref_type ref = align_size_to_section_boundary(m_baseline.load(std::memory_order_relaxed));
  655. for (const auto& e : m_slabs) {
  656. BetweenBlocks* bb = reinterpret_cast<BetweenBlocks*>(e.addr);
  657. REALM_ASSERT(bb->block_before_size == 0);
  658. while (1) {
  659. int size = bb->block_after_size;
  660. f(ref, sizeof(BetweenBlocks));
  661. ref += sizeof(BetweenBlocks);
  662. if (size == 0) {
  663. break;
  664. }
  665. if (size > 0) { // freeblock.
  666. f(ref, size);
  667. bb = reinterpret_cast<BetweenBlocks*>(reinterpret_cast<char*>(bb) + sizeof(BetweenBlocks) + size);
  668. ref += size;
  669. }
  670. else {
  671. bb = reinterpret_cast<BetweenBlocks*>(reinterpret_cast<char*>(bb) + sizeof(BetweenBlocks) - size);
  672. ref -= size;
  673. }
  674. }
  675. // any gaps in ref-space is reported as a free block to the validator:
  676. auto next_ref = align_size_to_section_boundary(ref);
  677. if (next_ref > ref) {
  678. f(ref, next_ref - ref);
  679. ref = next_ref;
  680. }
  681. }
  682. }
  683. } // namespace realm
  684. #endif // REALM_ALLOC_SLAB_HPP