|
@@ -1,38 +1,37 @@
|
|
|
+use smallvec::SmallVec;
|
|
|
+
|
|
|
use super::*;
|
|
|
-use std::marker::PhantomData;
|
|
|
-use std::ops::{Deref, DerefMut};
|
|
|
+use core::marker::PhantomData;
|
|
|
+use core::ops::{Deref, DerefMut};
|
|
|
|
|
|
/// CursorState represents the current state of the cursor. Currently, there are two possible states:
|
|
|
/// 1. inactive: the state where the cursor is not positioned on any node.
|
|
|
/// 2. positioned on a node: this state includes information about the node the cursor is on,
|
|
|
/// as well as the offset of the entry that needs to be operated on within the slots of the current node.
|
|
|
-enum CursorState<'a, I, Operation>
|
|
|
+enum CursorState<'a, I, L, Operation>
|
|
|
where
|
|
|
I: ItemEntry,
|
|
|
+ L: XLock,
|
|
|
{
|
|
|
Inactive,
|
|
|
AtNode {
|
|
|
- node: &'a XNode<I, Operation>,
|
|
|
+ node: &'a XNode<I, L, Operation>,
|
|
|
operation_offset: u8,
|
|
|
},
|
|
|
}
|
|
|
|
|
|
-impl<'a, I: ItemEntry, Operation> CursorState<'a, I, Operation> {
|
|
|
+impl<'a, I: ItemEntry, L: XLock, Operation> CursorState<'a, I, L, Operation> {
|
|
|
fn default() -> Self {
|
|
|
Self::Inactive
|
|
|
}
|
|
|
|
|
|
- fn arrive_node(&mut self, node: &'a XNode<I, Operation>, operation_offset: u8) {
|
|
|
+ fn arrive_node(&mut self, node: &'a XNode<I, L, Operation>, operation_offset: u8) {
|
|
|
*self = Self::AtNode {
|
|
|
node,
|
|
|
operation_offset,
|
|
|
};
|
|
|
}
|
|
|
|
|
|
- fn is_inactive(&self) -> bool {
|
|
|
- matches!(self, Self::Inactive)
|
|
|
- }
|
|
|
-
|
|
|
fn is_at_node(&self) -> bool {
|
|
|
matches!(
|
|
|
self,
|
|
@@ -43,7 +42,7 @@ impl<'a, I: ItemEntry, Operation> CursorState<'a, I, Operation> {
|
|
|
)
|
|
|
}
|
|
|
|
|
|
- fn node_info(&self) -> Option<(&'a XNode<I, Operation>, u8)> {
|
|
|
+ fn node_info(&self) -> Option<(&'a XNode<I, L, Operation>, u8)> {
|
|
|
if let Self::AtNode {
|
|
|
node,
|
|
|
operation_offset,
|
|
@@ -66,39 +65,40 @@ impl<'a, I: ItemEntry, Operation> CursorState<'a, I, Operation> {
|
|
|
/// XEntry at the next index. If the Cursor perform reset or next and then have a target index that is not able to touch,
|
|
|
/// the Cursor's state will also set to `Inactive`.
|
|
|
///
|
|
|
-/// Hence, at any given moment a cursor will be positioned on the XNode and be ready to operate its target XEntry.
|
|
|
+/// Hence, at any given moment when no operation is being performed, a cursor will be positioned on the XNode and be ready to operate its target XEntry.
|
|
|
/// If not, it means that the cursor is not able to touch the target `XEntry`.
|
|
|
///
|
|
|
/// The cursor will also record all nodes passed from the head node to the target position in `passed_node`,
|
|
|
/// thereby assisting it in performing some operations that involve searching upwards.
|
|
|
///
|
|
|
/// Multiple Cursors are allowed to operate on a single XArray at the same time.
|
|
|
-pub struct Cursor<'a, I, M>
|
|
|
+pub struct Cursor<'a, I, L, M>
|
|
|
where
|
|
|
I: ItemEntry,
|
|
|
+ L: XLock,
|
|
|
M: ValidMark,
|
|
|
{
|
|
|
/// The `XArray` the cursor located in.
|
|
|
- xa: &'a XArray<I, M>,
|
|
|
+ xa: &'a XArray<I, L, M>,
|
|
|
/// The target index of the cursor in the belonged `XArray`.
|
|
|
index: u64,
|
|
|
/// Represents the current state of the cursor.
|
|
|
- state: CursorState<'a, I, ReadOnly>,
|
|
|
+ state: CursorState<'a, I, L, ReadOnly>,
|
|
|
/// Record add nodes passed from the head node to the target position.
|
|
|
- /// The index is the layer of the recorded node.
|
|
|
- passed_node: [Option<&'a XNode<I, ReadOnly>>; MAX_LAYER],
|
|
|
+ /// The index is the height of the recorded node.
|
|
|
+ ancestors: SmallVec<[&'a XNode<I, L, ReadOnly>; MAX_HEIGHT]>,
|
|
|
|
|
|
_marker: PhantomData<I>,
|
|
|
}
|
|
|
|
|
|
-impl<'a, I: ItemEntry, M: ValidMark> Cursor<'a, I, M> {
|
|
|
+impl<'a, I: ItemEntry, L: XLock, M: ValidMark> Cursor<'a, I, L, M> {
|
|
|
/// Create an `Cursor` to perform read related operations on the `XArray`.
|
|
|
- pub(crate) fn new(xa: &'a XArray<I, M>, index: u64) -> Self {
|
|
|
+ pub(super) fn new(xa: &'a XArray<I, L, M>, index: u64) -> Self {
|
|
|
let mut cursor = Self {
|
|
|
xa,
|
|
|
index,
|
|
|
state: CursorState::default(),
|
|
|
- passed_node: [None; MAX_LAYER],
|
|
|
+ ancestors: SmallVec::new(),
|
|
|
_marker: PhantomData,
|
|
|
};
|
|
|
|
|
@@ -108,7 +108,7 @@ impl<'a, I: ItemEntry, M: ValidMark> Cursor<'a, I, M> {
|
|
|
|
|
|
/// Move the `Cursor` to the `XNode`, and update the cursor's state based on its target index.
|
|
|
/// Return a reference to the `XEntry` within the slots of the current XNode that needs to be operated on.
|
|
|
- fn move_to(&mut self, node: &'a XNode<I, ReadOnly>) -> &'a XEntry<I> {
|
|
|
+ fn move_to(&mut self, node: &'a XNode<I, L, ReadOnly>) -> &'a XEntry<I, L> {
|
|
|
let (current_entry, offset) = {
|
|
|
let offset = node.entry_offset(self.index);
|
|
|
let current_entry = node.ref_node_entry(offset);
|
|
@@ -144,20 +144,25 @@ impl<'a, I: ItemEntry, M: ValidMark> Cursor<'a, I, M> {
|
|
|
operation_offset += 1;
|
|
|
while operation_offset == SLOT_SIZE as u8 {
|
|
|
operation_offset = current_node.offset_in_parent() + 1;
|
|
|
- let parent_layer = (*current_node.layer() + 1) as usize;
|
|
|
- self.passed_node[parent_layer - 1] = None;
|
|
|
- current_node = self.passed_node[parent_layer].unwrap();
|
|
|
+ if let Some(node) = self.ancestors.pop() {
|
|
|
+ current_node = node;
|
|
|
+ continue;
|
|
|
+ }
|
|
|
+
|
|
|
+ operation_offset = 0;
|
|
|
+ break;
|
|
|
}
|
|
|
self.state.arrive_node(current_node, operation_offset);
|
|
|
|
|
|
- while current_node.layer() != 0 {
|
|
|
+ while !current_node.is_leaf() {
|
|
|
let next_entry = current_node.ref_node_entry(operation_offset);
|
|
|
if !next_entry.is_node() {
|
|
|
self.init();
|
|
|
return;
|
|
|
}
|
|
|
+
|
|
|
let next_node = next_entry.as_node().unwrap();
|
|
|
- self.passed_node[*next_node.layer() as usize] = Some(next_node);
|
|
|
+ self.ancestors.push(current_node);
|
|
|
self.move_to(next_node);
|
|
|
(current_node, operation_offset) = self.state.node_info().unwrap();
|
|
|
}
|
|
@@ -181,7 +186,7 @@ impl<'a, I: ItemEntry, M: ValidMark> Cursor<'a, I, M> {
|
|
|
if let Some((current_node, operation_offset)) = self.state.node_info() {
|
|
|
let entry = current_node.ref_node_entry(operation_offset);
|
|
|
if entry.is_item() {
|
|
|
- return Some(unsafe { &*(entry as *const XEntry<I> as *const I) });
|
|
|
+ return Some(unsafe { &*(entry as *const XEntry<I, L> as *const I) });
|
|
|
}
|
|
|
}
|
|
|
None
|
|
@@ -191,7 +196,7 @@ impl<'a, I: ItemEntry, M: ValidMark> Cursor<'a, I, M> {
|
|
|
/// It then returns the reference to the `XEntry` stored in the slot corresponding to the target index.
|
|
|
/// A target operated XEntry must be an item entry.
|
|
|
/// If can not touch the target entry, the function will return `None`.
|
|
|
- fn traverse_to_target(&mut self) -> Option<&'a XEntry<I>> {
|
|
|
+ fn traverse_to_target(&mut self) -> Option<&'a XEntry<I, L>> {
|
|
|
if self.is_arrived() {
|
|
|
let (current_node, operation_offset) = self.state.node_info().unwrap();
|
|
|
return Some(current_node.ref_node_entry(operation_offset));
|
|
@@ -204,18 +209,15 @@ impl<'a, I: ItemEntry, M: ValidMark> Cursor<'a, I, M> {
|
|
|
self.move_to(self.xa.head().as_node().unwrap());
|
|
|
|
|
|
let (mut current_node, operation_offset) = self.state.node_info().unwrap();
|
|
|
- let mut current_layer = current_node.layer();
|
|
|
let mut operated_entry = current_node.ref_node_entry(operation_offset);
|
|
|
- while current_layer > 0 {
|
|
|
+ while !current_node.is_leaf() {
|
|
|
if !operated_entry.is_node() {
|
|
|
self.init();
|
|
|
return None;
|
|
|
}
|
|
|
-
|
|
|
- self.passed_node[*current_layer as usize] = Some(current_node);
|
|
|
+ self.ancestors.push(current_node);
|
|
|
|
|
|
current_node = operated_entry.as_node().unwrap();
|
|
|
- *current_layer -= 1;
|
|
|
operated_entry = self.move_to(current_node);
|
|
|
}
|
|
|
Some(operated_entry)
|
|
@@ -224,7 +226,7 @@ impl<'a, I: ItemEntry, M: ValidMark> Cursor<'a, I, M> {
|
|
|
/// Initialize the Cursor to its initial state.
|
|
|
pub fn init(&mut self) {
|
|
|
self.state = CursorState::default();
|
|
|
- self.passed_node = [None; MAX_LAYER];
|
|
|
+ self.ancestors = SmallVec::new();
|
|
|
}
|
|
|
|
|
|
/// Return the target index of the cursor.
|
|
@@ -250,44 +252,49 @@ impl<'a, I: ItemEntry, M: ValidMark> Cursor<'a, I, M> {
|
|
|
/// XEntry at the next index. If the `CursorMut` perform reset or next and then have a target index that is not able to touch,
|
|
|
/// the `CursorMut`'s state will also set to `Inactive`.
|
|
|
///
|
|
|
-/// Hence, at any given moment a `CursorMut` will be positioned on the XNode and be ready to operate its target XEntry.
|
|
|
+/// Hence, at any given moment when no operation is being performed, a `CursorMut` will be positioned on the XNode and be ready to operate its target XEntry.
|
|
|
/// If not, it means that the `CursorMut` is not able to touch the target `XEntry`. For this situation, the `CursorMut`
|
|
|
/// can invoke `store` method which will expand the XArray to guarantee to reach the target position.
|
|
|
///
|
|
|
/// The `CursorMut` will also record all nodes passed from the head node to the target position in passed_node,
|
|
|
/// thereby assisting it in performing some operations that involve searching upwards.
|
|
|
///
|
|
|
-/// **Features for COW (Copy-On-Write).** The CursorMut guarantees that all nodes it traverses during the process are exclusively owned by the current XArray.
|
|
|
+/// **Features for COW (Copy-On-Write).** The CursorMut guarantees that if it is exclusive, all nodes it traverses during the process are exclusively owned by the current XArray.
|
|
|
/// If it finds that the node it is about to access is shared with another XArray due to a COW clone, it will trigger a COW to copy and create an exclusive node for access.
|
|
|
/// Additionally, since it holds a mutable reference to the current XArray, it will not conflict with any other cursors on the XArray.
|
|
|
+/// CursorMut is set to exclusive when a modification is about to be performed
|
|
|
///
|
|
|
-/// When a CursorMut doing operation on XArray, it should not be affected by other CursorMuts or affect other Cursors.
|
|
|
-pub struct CursorMut<'a, I, M>
|
|
|
+/// When a CursorMut doing write operation on XArray, it should not be affected by other CursorMuts or affect other Cursors.
|
|
|
+pub struct CursorMut<'a, I, L, M>
|
|
|
where
|
|
|
I: ItemEntry,
|
|
|
+ L: XLock,
|
|
|
M: ValidMark,
|
|
|
{
|
|
|
/// The `XArray` the cursor located in.
|
|
|
- xa: &'a mut XArray<I, M>,
|
|
|
+ xa: &'a mut XArray<I, L, M>,
|
|
|
/// The target index of the cursor in the belonged `XArray`.
|
|
|
index: u64,
|
|
|
/// Represents the current state of the cursor.
|
|
|
- state: CursorState<'a, I, ReadWrite>,
|
|
|
+ state: CursorState<'a, I, L, ReadWrite>,
|
|
|
/// Record add nodes passed from the head node to the target position.
|
|
|
- /// The index is the layer of the recorded node.
|
|
|
- passed_node: [Option<&'a XNode<I, ReadWrite>>; MAX_LAYER],
|
|
|
+ /// The index is the height of the recorded node.
|
|
|
+ ancestors: SmallVec<[&'a XNode<I, L, ReadWrite>; MAX_HEIGHT]>,
|
|
|
+
|
|
|
+ is_exclusive: bool,
|
|
|
|
|
|
_marker: PhantomData<I>,
|
|
|
}
|
|
|
|
|
|
-impl<'a, I: ItemEntry, M: ValidMark> CursorMut<'a, I, M> {
|
|
|
+impl<'a, I: ItemEntry, L: XLock, M: ValidMark> CursorMut<'a, I, L, M> {
|
|
|
/// Create an `CursorMut` to perform read and write operations on the `XArray`.
|
|
|
- pub(crate) fn new(xa: &'a mut XArray<I, M>, index: u64) -> Self {
|
|
|
+ pub(super) fn new(xa: &'a mut XArray<I, L, M>, index: u64) -> Self {
|
|
|
let mut cursor = Self {
|
|
|
xa,
|
|
|
index,
|
|
|
state: CursorState::default(),
|
|
|
- passed_node: [None; MAX_LAYER],
|
|
|
+ ancestors: SmallVec::new(),
|
|
|
+ is_exclusive: false,
|
|
|
_marker: PhantomData,
|
|
|
};
|
|
|
|
|
@@ -297,10 +304,10 @@ impl<'a, I: ItemEntry, M: ValidMark> CursorMut<'a, I, M> {
|
|
|
|
|
|
/// Move the `CursorMut` to the `XNode`, and update the cursor's state based on its target index.
|
|
|
/// Return a reference to the `XEntry` within the slots of the current XNode that needs to be operated on next.
|
|
|
- fn move_to(&mut self, node: &'a XNode<I, ReadWrite>) -> &'a XEntry<I> {
|
|
|
+ fn move_to(&mut self, node: &'a XNode<I, L, ReadWrite>) -> &'a XEntry<I, L> {
|
|
|
let (current_entry, offset) = {
|
|
|
let offset = node.entry_offset(self.index);
|
|
|
- let current_entry = node.ref_node_entry(offset);
|
|
|
+ let current_entry = node.ref_node_entry(self.is_exclusive, offset);
|
|
|
(current_entry, offset)
|
|
|
};
|
|
|
self.state.arrive_node(node, offset);
|
|
@@ -311,6 +318,7 @@ impl<'a, I: ItemEntry, M: ValidMark> CursorMut<'a, I, M> {
|
|
|
pub fn reset_to(&mut self, index: u64) {
|
|
|
self.init();
|
|
|
self.index = index;
|
|
|
+ self.is_exclusive = false;
|
|
|
self.traverse_to_target();
|
|
|
}
|
|
|
|
|
@@ -320,6 +328,7 @@ impl<'a, I: ItemEntry, M: ValidMark> CursorMut<'a, I, M> {
|
|
|
/// the method returns the provided entry without making changes.
|
|
|
/// Otherwise, it replaces the current entry with the provided one and returns the old entry.
|
|
|
pub fn store(&mut self, item: I) -> Option<I> {
|
|
|
+ self.ensure_exclusive_before_modify();
|
|
|
let stored_entry = XEntry::from_item(item);
|
|
|
let target_entry = self.expand_and_traverse_to_target();
|
|
|
if stored_entry.raw() == target_entry.raw() {
|
|
@@ -349,20 +358,25 @@ impl<'a, I: ItemEntry, M: ValidMark> CursorMut<'a, I, M> {
|
|
|
operation_offset += 1;
|
|
|
while operation_offset == SLOT_SIZE as u8 {
|
|
|
operation_offset = current_node.offset_in_parent() + 1;
|
|
|
- let parent_layer = (*current_node.layer() + 1) as usize;
|
|
|
- self.passed_node[parent_layer - 1] = None;
|
|
|
- current_node = self.passed_node[parent_layer].unwrap();
|
|
|
+ if let Some(node) = self.ancestors.pop() {
|
|
|
+ current_node = node;
|
|
|
+ continue;
|
|
|
+ }
|
|
|
+
|
|
|
+ operation_offset = 0;
|
|
|
+ break;
|
|
|
}
|
|
|
self.state.arrive_node(current_node, operation_offset);
|
|
|
|
|
|
- while current_node.layer() != 0 {
|
|
|
- let next_entry = current_node.ref_node_entry(operation_offset);
|
|
|
+ while !current_node.is_leaf() {
|
|
|
+ let next_entry = current_node.ref_node_entry(self.is_exclusive, operation_offset);
|
|
|
if !next_entry.is_node() {
|
|
|
self.init();
|
|
|
return;
|
|
|
}
|
|
|
+
|
|
|
let next_node = next_entry.as_node_mut().unwrap();
|
|
|
- self.passed_node[*next_node.layer() as usize] = Some(next_node);
|
|
|
+ self.ancestors.push(current_node);
|
|
|
self.move_to(next_node);
|
|
|
(current_node, operation_offset) = self.state.node_info().unwrap();
|
|
|
}
|
|
@@ -374,25 +388,22 @@ impl<'a, I: ItemEntry, M: ValidMark> CursorMut<'a, I, M> {
|
|
|
/// This operation will also mark all nodes along the path from the head node to the target node with the input `mark`,
|
|
|
/// because a marked intermediate node should be equivalent to having a child node that is marked.
|
|
|
pub fn set_mark(&mut self, mark: M) -> Result<(), ()> {
|
|
|
+ self.ensure_exclusive_before_modify();
|
|
|
if let Some((current_node, operation_offset)) = self.state.node_info() {
|
|
|
- let item_entry = current_node.ref_node_entry(operation_offset);
|
|
|
+ let item_entry = current_node.ref_node_entry(self.is_exclusive, operation_offset);
|
|
|
if item_entry.is_null() {
|
|
|
return Err(());
|
|
|
}
|
|
|
|
|
|
current_node.set_mark(operation_offset, mark.index());
|
|
|
|
|
|
- let head_layer = *(self.xa.head().as_node().unwrap().layer()) as usize;
|
|
|
let mut offset_in_parent = current_node.offset_in_parent();
|
|
|
- let mut parent_layer = (*current_node.layer() + 1) as usize;
|
|
|
- while parent_layer <= head_layer {
|
|
|
- let parent_node = self.passed_node[parent_layer].unwrap();
|
|
|
- if parent_node.is_marked(offset_in_parent, mark.index()) {
|
|
|
+ for ancestor in self.ancestors.iter().rev() {
|
|
|
+ if ancestor.is_marked(offset_in_parent, mark.index()) {
|
|
|
break;
|
|
|
}
|
|
|
- parent_node.set_mark(offset_in_parent, mark.index());
|
|
|
- offset_in_parent = parent_node.offset_in_parent();
|
|
|
- parent_layer += 1;
|
|
|
+ ancestor.set_mark(offset_in_parent, mark.index());
|
|
|
+ offset_in_parent = ancestor.offset_in_parent();
|
|
|
}
|
|
|
Ok(())
|
|
|
} else {
|
|
@@ -406,25 +417,24 @@ impl<'a, I: ItemEntry, M: ValidMark> CursorMut<'a, I, M> {
|
|
|
/// This operation will also unset the input `mark` for all nodes along the path from the head node to the target node
|
|
|
/// if the input `mark` have not marked any of their children.
|
|
|
pub fn unset_mark(&mut self, mark: M) -> Result<(), ()> {
|
|
|
- if let Some((mut current_node, operation_offset)) = self.state.node_info() {
|
|
|
- let item_entry = current_node.ref_node_entry(operation_offset);
|
|
|
+ self.ensure_exclusive_before_modify();
|
|
|
+ if let Some((current_node, operation_offset)) = self.state.node_info() {
|
|
|
+ let item_entry = current_node.ref_node_entry(self.is_exclusive, operation_offset);
|
|
|
if item_entry.is_null() {
|
|
|
return Err(());
|
|
|
}
|
|
|
|
|
|
current_node.unset_mark(operation_offset, mark.index());
|
|
|
|
|
|
- let head_layer = *(self.xa.head().as_node().unwrap().layer()) as usize;
|
|
|
- let mut parent_layer = (*current_node.layer() + 1) as usize;
|
|
|
- while current_node.is_mark_clear(mark.index()) {
|
|
|
- let offset_in_parent = current_node.offset_in_parent();
|
|
|
- let parent_node = self.passed_node[parent_layer].unwrap();
|
|
|
- parent_node.unset_mark(offset_in_parent, mark.index());
|
|
|
+ if current_node.is_mark_clear(mark.index()) {
|
|
|
+ let mut offset_in_parent = current_node.offset_in_parent();
|
|
|
+ for ancestor in self.ancestors.iter().rev() {
|
|
|
+ ancestor.unset_mark(offset_in_parent, mark.index());
|
|
|
+ if !ancestor.is_mark_clear(mark.index()) {
|
|
|
+ break;
|
|
|
+ }
|
|
|
|
|
|
- current_node = parent_node;
|
|
|
- parent_layer += 1;
|
|
|
- if parent_layer > head_layer {
|
|
|
- break;
|
|
|
+ offset_in_parent = ancestor.offset_in_parent();
|
|
|
}
|
|
|
}
|
|
|
Ok(())
|
|
@@ -438,6 +448,7 @@ impl<'a, I: ItemEntry, M: ValidMark> CursorMut<'a, I, M> {
|
|
|
/// This is achieved by storing an empty `XEntry` at the target index using the `store` method.
|
|
|
/// The method returns the replaced `XEntry` that was previously stored at the target index.
|
|
|
pub fn remove(&mut self) -> Option<I> {
|
|
|
+ self.ensure_exclusive_before_modify();
|
|
|
if let Some((current_node, operation_offset)) = self.state.node_info() {
|
|
|
let old_entry = current_node.set_entry(operation_offset, XEntry::EMPTY);
|
|
|
return XEntry::into_item(old_entry);
|
|
@@ -449,32 +460,30 @@ impl<'a, I: ItemEntry, M: ValidMark> CursorMut<'a, I, M> {
|
|
|
/// It then returns the reference to the `XEntry` stored in the slot corresponding to the target index.
|
|
|
/// A target operated XEntry must be an item entry.
|
|
|
/// If can not touch the target entry, the function will return `None`.
|
|
|
- fn traverse_to_target(&mut self) -> Option<&'a XEntry<I>> {
|
|
|
+ fn traverse_to_target(&mut self) -> Option<&'a XEntry<I, L>> {
|
|
|
if self.is_arrived() {
|
|
|
let (current_node, operation_offset) = self.state.node_info().unwrap();
|
|
|
- return Some(current_node.ref_node_entry(operation_offset));
|
|
|
+ return Some(current_node.ref_node_entry(self.is_exclusive, operation_offset));
|
|
|
}
|
|
|
|
|
|
let max_index = self.xa.max_index();
|
|
|
if max_index < self.index || max_index == 0 {
|
|
|
return None;
|
|
|
}
|
|
|
- let head = self.xa.head_mut().as_node_mut().unwrap();
|
|
|
+ let head = self.xa.head_mut(self.is_exclusive).as_node_mut().unwrap();
|
|
|
self.move_to(head);
|
|
|
|
|
|
let (mut current_node, operation_offset) = self.state.node_info().unwrap();
|
|
|
- let mut current_layer = current_node.layer();
|
|
|
- let mut operated_entry = current_node.ref_node_entry(operation_offset);
|
|
|
- while current_layer > 0 {
|
|
|
+ let mut operated_entry = current_node.ref_node_entry(self.is_exclusive, operation_offset);
|
|
|
+ while !current_node.is_leaf() {
|
|
|
if !operated_entry.is_node() {
|
|
|
self.init();
|
|
|
return None;
|
|
|
}
|
|
|
|
|
|
- self.passed_node[*current_layer as usize] = Some(current_node);
|
|
|
+ self.ancestors.push(current_node);
|
|
|
|
|
|
current_node = operated_entry.as_node_mut().unwrap();
|
|
|
- *current_layer -= 1;
|
|
|
operated_entry = self.move_to(current_node);
|
|
|
}
|
|
|
Some(operated_entry)
|
|
@@ -483,74 +492,74 @@ impl<'a, I: ItemEntry, M: ValidMark> CursorMut<'a, I, M> {
|
|
|
/// Traverse the XArray and move to the node that can operate the target entry.
|
|
|
/// During the traverse, the cursor may modify the XArray to let itself be able to reach the target node.
|
|
|
///
|
|
|
- /// Before traverse, the cursor will first expand the layer of `XArray` to make sure it have enough capacity.
|
|
|
+ /// Before traverse, the cursor will first expand the height of `XArray` to make sure it have enough capacity.
|
|
|
/// During the traverse, the cursor will allocate new `XNode` and put it in the appropriate slot if needed.
|
|
|
///
|
|
|
/// It then returns the reference to the `XEntry` stored in the slot corresponding to the target index.
|
|
|
/// A target operated XEntry must be an item entry.
|
|
|
- fn expand_and_traverse_to_target(&mut self) -> &'a XEntry<I> {
|
|
|
+ fn expand_and_traverse_to_target(&mut self) -> &'a XEntry<I, L> {
|
|
|
if self.is_arrived() {
|
|
|
let (current_node, operation_offset) = self.state.node_info().unwrap();
|
|
|
- return current_node.ref_node_entry(operation_offset);
|
|
|
+ return current_node.ref_node_entry(self.is_exclusive, operation_offset);
|
|
|
}
|
|
|
|
|
|
- self.expand_layer();
|
|
|
- let head_ref = self.xa.head_mut().as_node_mut().unwrap();
|
|
|
+ self.expand_height();
|
|
|
+ let head_ref = self.xa.head_mut(self.is_exclusive).as_node_mut().unwrap();
|
|
|
self.move_to(head_ref);
|
|
|
|
|
|
let (mut current_node, operation_offset) = self.state.node_info().unwrap();
|
|
|
- let mut current_layer = current_node.layer();
|
|
|
- let mut operated_entry = current_node.ref_node_entry(operation_offset);
|
|
|
- while current_layer > 0 {
|
|
|
+ let mut operated_entry = current_node.ref_node_entry(self.is_exclusive, operation_offset);
|
|
|
+ while !current_node.is_leaf() {
|
|
|
+ let current_height = current_node.height();
|
|
|
+
|
|
|
if !operated_entry.is_node() {
|
|
|
let new_entry = {
|
|
|
let (current_node, operation_offset) = self.state.node_info().unwrap();
|
|
|
let new_owned_entry =
|
|
|
- self.alloc_node(Layer::new(*current_layer - 1), operation_offset);
|
|
|
+ self.alloc_node(Height::new(*current_height - 1), operation_offset);
|
|
|
let _ = current_node.set_entry(operation_offset, new_owned_entry);
|
|
|
- current_node.ref_node_entry(operation_offset)
|
|
|
+ current_node.ref_node_entry(self.is_exclusive, operation_offset)
|
|
|
};
|
|
|
operated_entry = new_entry;
|
|
|
}
|
|
|
|
|
|
- self.passed_node[*current_layer as usize] = Some(current_node);
|
|
|
+ self.ancestors.push(current_node);
|
|
|
|
|
|
current_node = operated_entry.as_node_mut().unwrap();
|
|
|
- *current_layer -= 1;
|
|
|
operated_entry = self.move_to(current_node);
|
|
|
}
|
|
|
operated_entry
|
|
|
}
|
|
|
|
|
|
- /// Increase the number of layers for XArray to expand its capacity, allowing it to accommodate the target index,
|
|
|
- /// and returns the layer of the final head node.
|
|
|
+ /// Increase the number of heights for XArray to expand its capacity, allowing it to accommodate the target index,
|
|
|
+ /// and returns the height of the final head node.
|
|
|
///
|
|
|
- /// If the head node of the XArray does not exist, allocate a new head node of appropriate layer directly.
|
|
|
+ /// If the head node of the XArray does not exist, allocate a new head node of appropriate height directly.
|
|
|
/// Otherwise, if needed, repeatedly insert new nodes on top of the current head node to serve as the new head.
|
|
|
- fn expand_layer(&mut self) -> Layer {
|
|
|
+ fn expand_height(&mut self) -> Height {
|
|
|
if self.xa.head().is_null() {
|
|
|
- let mut head_layer = Layer::new(0);
|
|
|
- while self.index > head_layer.max_index() {
|
|
|
- *head_layer += 1;
|
|
|
+ let mut head_height = Height::new(1);
|
|
|
+ while self.index > head_height.max_index() {
|
|
|
+ *head_height += 1;
|
|
|
}
|
|
|
- let head = self.alloc_node(head_layer, 0);
|
|
|
+ let head = self.alloc_node(head_height, 0);
|
|
|
self.xa.set_head(head);
|
|
|
- return head_layer;
|
|
|
+ return head_height;
|
|
|
} else {
|
|
|
loop {
|
|
|
- let head_layer = {
|
|
|
- let head = self.xa.head().as_node().unwrap();
|
|
|
- head.layer()
|
|
|
+ let head_height = {
|
|
|
+ let head = self.xa.head_mut(self.is_exclusive).as_node().unwrap();
|
|
|
+ head.height()
|
|
|
};
|
|
|
|
|
|
- if head_layer.max_index() > self.index {
|
|
|
- return head_layer;
|
|
|
+ if head_height.max_index() > self.index {
|
|
|
+ return head_height;
|
|
|
}
|
|
|
|
|
|
- let new_node_entry = self.alloc_node(Layer::new(*head_layer + 1), 0);
|
|
|
+ let new_node_entry = self.alloc_node(Height::new(*head_height + 1), 0);
|
|
|
let old_head_entry = self.xa.set_head(new_node_entry);
|
|
|
let old_head = old_head_entry.as_node_mut().unwrap();
|
|
|
- let new_head = self.xa.head_mut().as_node_mut().unwrap();
|
|
|
+ let new_head = self.xa.head_mut(self.is_exclusive).as_node_mut().unwrap();
|
|
|
for i in 0..3 {
|
|
|
if !old_head.mark(i).is_clear() {
|
|
|
new_head.set_mark(0, i);
|
|
@@ -561,23 +570,33 @@ impl<'a, I: ItemEntry, M: ValidMark> CursorMut<'a, I, M> {
|
|
|
}
|
|
|
}
|
|
|
|
|
|
- /// Allocate a new XNode with the specified layer and offset,
|
|
|
+ /// Allocate a new XNode with the specified height and offset,
|
|
|
/// then generate a node entry from it and return it to the caller.
|
|
|
- fn alloc_node(&mut self, layer: Layer, offset: u8) -> XEntry<I> {
|
|
|
- XEntry::from_node(XNode::<I, ReadWrite>::new(layer, offset))
|
|
|
+ fn alloc_node(&mut self, height: Height, offset: u8) -> XEntry<I, L> {
|
|
|
+ XEntry::from_node(XNode::<I, L, ReadWrite>::new(height, offset))
|
|
|
+ }
|
|
|
+
|
|
|
+ fn ensure_exclusive_before_modify(&mut self) {
|
|
|
+ if !self.is_exclusive {
|
|
|
+ self.is_exclusive = true;
|
|
|
+ if self.is_arrived() {
|
|
|
+ self.init();
|
|
|
+ self.traverse_to_target();
|
|
|
+ }
|
|
|
+ }
|
|
|
}
|
|
|
}
|
|
|
|
|
|
-impl<'a, I: ItemEntry, M: ValidMark> Deref for CursorMut<'a, I, M> {
|
|
|
- type Target = Cursor<'a, I, M>;
|
|
|
+impl<'a, I: ItemEntry, L: XLock, M: ValidMark> Deref for CursorMut<'a, I, L, M> {
|
|
|
+ type Target = Cursor<'a, I, L, M>;
|
|
|
|
|
|
fn deref(&self) -> &Self::Target {
|
|
|
- unsafe { &*(self as *const CursorMut<'a, I, M> as *const Cursor<'a, I, M>) }
|
|
|
+ unsafe { &*(self as *const CursorMut<'a, I, L, M> as *const Cursor<'a, I, L, M>) }
|
|
|
}
|
|
|
}
|
|
|
|
|
|
-impl<'a, I: ItemEntry, M: ValidMark> DerefMut for CursorMut<'a, I, M> {
|
|
|
+impl<'a, I: ItemEntry, L: XLock, M: ValidMark> DerefMut for CursorMut<'a, I, L, M> {
|
|
|
fn deref_mut(&mut self) -> &mut Self::Target {
|
|
|
- unsafe { &mut *(self as *const CursorMut<'a, I, M> as *mut Cursor<'a, I, M>) }
|
|
|
+ unsafe { &mut *(self as *const CursorMut<'a, I, L, M> as *mut Cursor<'a, I, L, M>) }
|
|
|
}
|
|
|
}
|