- Create a
ChronicleMap
Instance - In-memory Chronicle Map
- Persisted Chronicle Map
- ChronicleMap instance vs Chronicle Map data store
- Configure entries
- Single
ChronicleMap
instance per JVM - Recovery
- Key and Value Types
- Custom serializers
ChronicleMap
usage patterns- Close
ChronicleMap
- Behaviour Customization
- Entry checksums
This document describes the Chronicle Map tutorial supplied with the project.
Creating an instance of ChronicleMap
is a little more complex than just calling a constructor.
To create an instance you have to use the ChronicleMapBuilder
.
The following code snippet creates an in-memory Chronicle Map store, to store about 50,000 city name → postal code mappings.
It is accessible within a single JVM process; the process in which it was created. The data is accessible while the process is alive. When the process is terminated, the data is cleared.
import net.openhft.chronicle.map.*
.....
interface PostalCodeRange {
int minCode();
void minCode(int minCode);
int maxCode();
void maxCode(int maxCode);
}
ChronicleMapBuilder<CharSequence, PostalCodeRange> cityPostalCodesMapBuilder =
ChronicleMapBuilder.of(CharSequence.class, PostalCodeRange.class)
.name("city-postal-codes-map")
.averageKey("Amsterdam")
.entries(50_000);
ChronicleMap<CharSequence, PostalCodeRange> cityPostalCodes =
cityPostalCodesMapBuilder.create();
// Or shorter form, without builder variable extraction:
ChronicleMap<CharSequence, PostalCodeRange> cityPostalCodes = ChronicleMapBuilder
.of(CharSequence.class, PostalCodeRange.class)
.name("city-postal-codes-map")
.averageKey("Amsterdam")
.entries(50_000)
.create();
You can amend this code to create a persisted Chronicle Map by replacing .create()
calls with .createPersistedTo(cityPostalCodesFile)
.
Use a persisted Chronicle Map if you want it to, either:
-
outlive the process it was created within; for example, to support hot application redeployment.
-
be accessible from multiple concurrent processes on the same server.
-
persist the data to disk.
The cityPostalCodesFile
has to represent the same location on your server for all the Java processes that wish to access this Chronicle Map instance.
For example, OS.getTarget() + "/cityPostalCodes.dat"
.
The name and location of the file is entirely up to you.
Note
|
When you create a ChronicleMap instance with .createPersistedTo(file) , and the specified file already exists in the system, you open a view to the existing Chronicle Map data store from this JVM process, rather than creating a new Chronicle Map data store.
That means that the data store may already contain some entries.
No special action with the data is performed during such an operation.
If you want to clean up corrupted entries, and ensure that the data store is in correct state, see the Recovery section.
|
In this tutorial, the term ChronicleMap
instance (or simply ChronicleMap
) is used to refer to an on-heap object providing access to a Chronicle Map data store (or Chronicle Map key-value store, or Chronicle Map store, or simply Chronicle Map, with space between two words in contrast to ChronicleMap
), which could be purely in-memory, or persisted to disk.
Currently the Java implementation doesn’t allow creation of multiple accessor ChronicleMap
objects for a single in-memory Chronicle Map store; there is always a one-to-one relationship.
A persisted Chronicle Map store, however, does allow creation of multiple accessor ChronicleMap
instances; either within a single JVM process (although it is not recommended), or from concurrent JVM processes.
When no processes access the file, it could be moved to another location in the system, or even to another server. It could even run on different operating systems. When opened from another location you will observe the same data.
If you don’t need the Chronicle Map instance to survive a server restart (that is, you don’t need
persistence to disk; only multi-process access), mount the file on tmpfs. For example, on Linux
it is as easy as placing your file in /dev/shm
directory.
You must configure .entries(entries) — to support the maximum number of entries in the Chronicle Map. Try to configure the entries so that the created Chronicle Map is going to serve about 99% of the requests.
You should not put additional margin over the actual target number of entries. This bad practice was popularized by new HashMap(capacity)
and new HashSet(capacity)
constructors, which accept capacity, that should be multiplied by load factor to obtain the actual maximum expected number of entries in the container. ChronicleMap
and ChronicleSet
do not have a notion of a load factor.
See ChronicleMapBuilder#entries()
in Javadocs for more information.
Once a ChronicleMap
instance is created, its configurations are sealed and cannot be changed
using the ChronicleMapBuilder
instance.
If you want to access a Chronicle Map data store
concurrently within a Java process, you should not create a separate ChronicleMap
instance for each thread. Within the JVM environment, a ChronicleMap
instance is a ConcurrentMap
, and could be accessed concurrently the same way as, for example, a ConcurrentHashMap
.
If a process, accessing a persisted Chronicle Map, terminated abnormally, for example:
-
crashed
-
SIGKILL
ed -
terminated because the host operating system crashed
-
terminated because the host machine lost power
then the Chronicle Map may remain in an inaccessible and/or corrupted state.
The next time a process opens the Chronicle Map, it should be done using .recoverPersistedTo()
method in ChronicleMapBuilder
.
Unlike createPersistedTo()
, this method scans all the memory of the Chronicle Map store for
inconsistencies and, if any are found, it cleans them up.
The method .recoverPersistedTo()
needs to access the Chronicle Map exclusively. If a concurrent process is
accessing the Chronicle Map while another process is attempting to perform recovery, the results of
operations on the accessing process side, and results of recovery are unspecified; the data could be corrupted further. You must ensure that no other process is accessing the Chronicle Map store when
calling .recoverPersistedTo()
.
Example:
ChronicleMap<CharSequence, PostalCodeRange> cityPostalCodes = ChronicleMap
.of(CharSequence.class, PostalCodeRange.class)
.name("city-postal-codes-map")
.averageKey("Amsterdam")
.entries(50_000)
.recoverPersistedTo(cityPostalCodesFile, false);
The second parameter in recoverPersistedTo()
method is called
sameBuilderConfigAndLibraryVersion
. It has two possible values:
-
true
- ifChronicleMapBuilder
is configured in exactly the same way, as when the Chronicle Map (persisted to the given file) was created, and using the same version of the Chronicle Map library -
false
- if the initial configuration is not known, or the current version of Chronicle Map library could be different from the version originally used to create this Chronicle Map.
If sameBuilderConfigAndLibraryVersion
is true
, recoverPersistedTo()
"knows" all the right
configurations, and what should be written to the header. It checks if the recovered Chronicle Map’s
header memory (containing serialized configurations) is corrupted or not. If the header is
corrupted, it is overridden, and the recovery process continues.
If sameBuilderConfigAndLibraryVersion
is false
, recoverPersistedTo()
relies on the
configurations written to the Chronicle Map’s header, assuming it is not corrupted. If it is
corrupted, ChronicleHashRecoveryFailedException
is thrown.
Note
|
The subject header memory is never updated on ordinary operations with Chronicle Map, so it couldn’t be corrupted if an accessing process crashed, or the operating system crashed, or even the machine lost power. Only hardware, memory, or disk corruption, or a bug in the file system, could lead to Chronicle Map header memory corruption. |
.recoverPersistedTo()
is harmless if the previous process accessing the Chronicle Map terminated
normally; however this is a computationally expensive procedure that should generally be avoided.
Only call this method if the Map is actually corrupted, and not as part of normal usage. If it is called
concurrently it can silently corrupt the Map.
If the Chronicle Map is configured to store entry checksums along with entries, then the recovery procedure checks that the checksum is correct for each entry.
Otherwise, it assumes the entry is corrupted and deletes it from the Chronicle Map. If checksums are to be stored, the recovery procedure cannot guarantee correctness of the entry data. See [Entry checksums](#entry-checksums) section for more information.
The key, or value type, of ChronicleMap<K, V>
could be:
-
Types with best possible out-of-the-box support:
-
Any value interface
-
Any class implementing
Byteable
interface from Chronicle Bytes. -
Any class implementing
BytesMarshallable
. interface from Chronicle Bytes. The implementation class should have a public no-arg constructor. -
byte[]
andByteBuffer
-
CharSequence
,String
andStringBuilder
. Note that these char sequence types are serialized using UTF-8 encoding by default. If you need a different encoding, refer to the example in customCharSequence
encoding. -
Integer
,Long
andDouble
-
-
Types supported out-of-the-box, but that are not particularly efficiently. You may want to implement more efficient custom serializers for them:
-
Any class implementing
java.io.Externalizable
. The implementation class should have a publicno-arg
constructor. -
Any type implementing
java.io.Serializable
, including boxed primitive types (except those listed above) and array types. -
Types implementing
net.openhft.chronicle.wire.Marshallable
can be used as key or value types withMarshallableReaderWriter
. It’s much less efficient thanBytesMarshallable
, but it may bring some benefits due to its self-describing format. Handling schema changes (added or removed fields) comes out-of-the-box.NoteChronicle Map does not consider use of BytesMarshallable#usesSelfDescribingMessage
method.
-
-
Any other type, if custom serializers are provided.
Value interfaces are preferred as they do not generate garbage, and have close to zero serialization/deserialization costs. They are preferable even to boxed primitives. For example, try to use net.openhft.chronicle.core.values.IntValue
instead of Integer
.
Generally, you must provide hints for the ChronicleMapBuilder
with the average sizes of the keys and the values, which are going to be inserted into the ChronicleMap
. This is required in order to allocate the proper amount of shared memory. Do this using averageKey()
(preferred) or averageKeySize()
, and
averageValue()
or averageValueSize()
respectively.
In the example above, averageKey("Amsterdam")
is called, because it is assumed that "Amsterdam" (9 bytes in UTF-8 encoding) is the average length for city names. Some names are shorter (Tokyo, 5 bytes), some names are longer (San Francisco, 13 bytes).
Another example could be if values in your ChronicleMap
are adjacency lists of some social graph, where nodes are represented as long
identifiers, and adjacency lists are long[]
arrays. For example, if the average number of friends is 150, you could configure the ChronicleMap
as follows:
Map<Long, long[]> socialGraph = ChronicleMap
.of(Long.class, long[].class)
.name("social-graph-map")
.entries(1_000_000_000L)
.averageValue(new long[150])
.create();
You could omit specifying key, or value, average sizes, if their types are boxed Java primitives or value interfaces. They are constantly-sized and Chronicle Map knows about that.
If the key or value type is constantly sized, or keys or values only of a certain size appear in your Chronicle Map domain, then preferably you should configure constantKeySizeBySample()
or
constantValueSizeBySample()
, instead of averageKey()
or averageValue()
. For example:
ChronicleSet<UUID> uuids =
ChronicleSet.of(UUID.class)
.name("uuids")
// All UUIDs take 16 bytes.
.constantKeySizeBySample(UUID.randomUUID())
.entries(1_000_000)
.create();
Chronicle Map allows you to configure custom marshallers for key or value types which are not supported out-of-the-box.You can also serialize supported types like String
in some custom way (encoded other than UTF-8), or serialize supported types more efficiently than by default.
There are three pairs of serialization interfaces. Only one of them should be chosen in a single implementation, and supplied to the ChronicleMapBuilder
for the key or value type. These are:
-
Choose the most suitable pair of serialization interfaces; BytesWriter and BytesReader, SizedWriter and SizedReader, or DataAccess and SizedReader. Recommendations on which pair to choose are given in the linked sections, describing each pair.
-
If implementation of the writer or reader part is configuration-less, give it a
private
constructor, and define a singleINSTANCE
constant. A sole instance of this marshaller class in the JVM. ImplementReadResolvable
and returnINSTANCE
from thereadResolve()
method. Do not make the implementation a Javaenum
. -
If both the writer and reader are configuration-less, merge them into a single
-Marshaller
implementation class. -
Make best efforts to reuse
using
objects on the reader side (BytesReader
orSizedReader
); including nesting objects. -
Make best efforts to cache intermediate serialization results on the writer side while working with some object. For example, try not to make expensive computations in both
size()
andwrite()
methods of theSizedWriter
implementation. Rather, compute them and cache in an serializer instance field. -
Make best efforts to reuse intermediate objects that are used for reading or writing. Store them in instance fields of the serializer implementation.
-
If a serializer implementation is stateful, or has cache fields, implement
StatefulCopyable
.
See UnderstandingStatefulCopyable
for more information. -
Implement
writeMarshallable()
andreadMarshallable()
by writing and reading configuration fields (but not the state or cache fields) of the serializer instance one-by-one. Use the givenWireOut
/WireIn
object.
See [CustomCharSequence
encoding](#custom-charsequence-encoding) section for some non-trivial example of implementing these methods. See also Wire tutorial. -
Don’t forget to initialize transient/cache/state fields of the instance in the end of
readMarshallable()
implementation. This is needed, because before callingreadMarshallable()
, Wire framework creates a serializer instance by means ofUnsafe.allocateInstance()
rather than calling any constructor. -
If implementing
DataAccess
, consider implementation to beData
also, and returnthis
fromgetData()
method. -
Don’t forget to implement
equals()
,hashCode()
andtoString()
inData
implementation, returned fromDataAccess.getData()
method, regardless if this is actually the sameDataAccess
object, or a separate object. -
Except
DataAccess
which is also aData
, serializers shouldn’t override Object’sequals()
,hashCode()
andtoString()
(these methods are never called on serializers inside Chronicle Map library); they shouldn’t implementSerializable
orExternalizable
(but have to implementnet.openhft.chronicle.wire.Marshallable
); shouldn’t implementCloneable
(but have to implementStatefulCopyable
, if they are stateful or have cache fields). -
After implementing custom serializers, don’t forget to actually apply them to
ChronicleMapBuilder
bykeyMarshallers()
,keyReaderAndDataAccess()
,valueMarshallers()
orvalueReaderAndDataAccess()
methods.
ChronicleMap
supports all operations from:
-
Map
interfaces;get()
,put()
, etc, including methods added in Java 8, likecompute()
andmerge()
, and -
ConcurrentMap
interfaces;putIfAbsent()
,replace()
.
All operations, including those which include "two-steps", for example, compute()
, are correctly synchronized in terms of the ConcurrentMap
interface. This means that you could use a ChronicleMap
instance just like a HashMap
or ConcurrentHashMap
.
PostalCodeRange amsterdamCodes = Values.newHeapInstance(PostalCodeRange.class);
amsterdamCodes.minCode(1011);
amsterdamCodes.maxCode(1183);
cityPostalCodes.put("Amsterdam", amsterdamCodes);
...
PostalCodeRange amsterdamCodes = cityPostalCodes.get("Amsterdam");
However, this approach often generates garbage, because the values should be deserialized from off-heap memory to on-heap memory when the new value objects are allocated. There are several possibilities to reuse objects efficiently:
If you want to create a ChronicleMap
where keys are long
ids, use LongValue
instead of Long
key:
ChronicleMap<LongValue, Order> orders = ChronicleMap
.of(LongValue.class, Order.class)
.name("orders-map")
.entries(1_000_000)
.create();
LongValue key = Values.newHeapInstance(LongValue.class);
key.setValue(id);
orders.put(key, order);
...
long[] orderIds = ...
// Allocate a single heap instance for inserting all keys from the array.
// This could be a cached or ThreadLocal value as well, eliminating
// allocations altogether.
LongValue key = Values.newHeapInstance(LongValue.class);
for (long id : orderIds) {
// Reuse the heap instance for each key
key.setValue(id);
Order order = orders.get(key);
// process the order...
}
Use ChronicleMap#getUsing(K key, V using)
to reuse the value object. It works if the value type is CharSequence
. Pass StringBuilder
as the using
argument. For example:
ChronicleMap<LongValue, CharSequence> names = ...
StringBuilder name = new StringBuilder();
for (long id : ids) {
key.setValue(id);
names.getUsing(key, name);
// process the name...
}
In this case, calling names.getUsing(key, name)
is equivalent to:
name.setLength(0);
name.append(names.get(key));
The difference is that it doesn’t generate garbage. The value type is the value interface. Pass the heap instance to read the data into it without new object allocation:
ThreadLocal<PostalCodeRange> cachedPostalCodeRange =
ThreadLocal.withInitial(() -> Values.newHeapInstance(PostalCodeRange.class));
...
PostalCodeRange range = cachedPostalCodeRange.get();
cityPostalCodes.getUsing(city, range);
// process the range...
-
If the value type implements
BytesMarshallable
, orExternalizable
, thenChronicleMap
attempts to reuse the givenusing
object by deserializing the value into the given object. -
If custom marshaller is configured in the
ChronicleMapBuilder
via.valueMarshaller()
, thenChronicleMap
attempts to reuse the given object by calling thereadUsing()
method from the marshaller interface.
If ChronicleMap
fails to reuse the object in getUsing()
, it does no harm. It falls back to
object creation, as in the get()
method. In particular, even null
is allowed to be passed as
using
object. It allows a "lazy" using object initialization pattern:
// a field
PostalCodeRange cachedRange = null;
...
// in a method
cachedRange = cityPostalCodes.getUsing(city, cachedRange);
// process the range...
In this example, cachedRange
is null
initially. On the first getUsing()
call, the heap value
is allocated and saved in a cachedRange
field for later reuse.
Note
|
If the value type is a value interface, do not use flyweight implementation as the getUsing() argument. This is dangerous, because on reusing flyweight points to the ChronicleMap memory
directly, but the access is not synchronized. At best you could read inconsistent value state;
at worst you could corrupt the ChronicleMap memory.
|
For accessing the ChronicleMap
value memory directly use the following techniques.
try (ExternalMapQueryContext<CharSequence, PostalCodeRange, ?> c =
cityPostalCodes.queryContext("Amsterdam")) {
MapEntry<CharSequence, PostalCodeRange> entry = c.entry();
if (entry != null) {
PostalCodeRange range = entry.value().get();
// Access the off-heap memory directly, by calling range
// object getters.
// This is very rewarding, when the value has a lot of fields
// and expensive to copy to heap all of them, when you need to access
// just a few fields.
} else {
// city not found..
}
}
In this example, consistent graph edge addition and removals are implemented using multi-key queries:
public static boolean addEdge(
ChronicleMap<Integer, Set<Integer>> graph, int source, int target) {
if (source == target)
throw new IllegalArgumentException("loops are forbidden");
ExternalMapQueryContext<Integer, Set<Integer>, ?> sourceC = graph.queryContext(source);
ExternalMapQueryContext<Integer, Set<Integer>, ?> targetC = graph.queryContext(target);
// order for consistent lock acquisition => avoid dead lock
if (sourceC.segmentIndex() <= targetC.segmentIndex()) {
return innerAddEdge(source, sourceC, target, targetC);
} else {
return innerAddEdge(target, targetC, source, sourceC);
}
}
private static boolean innerAddEdge(
int source, ExternalMapQueryContext<Integer, Set<Integer>, ?> sourceContext,
int target, ExternalMapQueryContext<Integer, Set<Integer>, ?> targetContext) {
try (ExternalMapQueryContext<Integer, Set<Integer>, ?> sc = sourceContext) {
try (ExternalMapQueryContext<Integer, Set<Integer>, ?> tc = targetContext) {
sc.updateLock().lock();
tc.updateLock().lock();
MapEntry<Integer, Set<Integer>> sEntry = sc.entry();
if (sEntry != null) {
MapEntry<Integer, Set<Integer>> tEntry = tc.entry();
if (tEntry != null) {
return addEdgeBothPresent(sc, sEntry, source, tc, tEntry, target);
} else {
addEdgePresentAbsent(sc, sEntry, source, tc, target);
return true;
}
} else {
MapEntry<Integer, Set<Integer>> tEntry = tc.entry();
if (tEntry != null) {
addEdgePresentAbsent(tc, tEntry, target, sc, source);
} else {
addEdgeBothAbsent(sc, source, tc, target);
}
return true;
}
}
}
}
private static boolean addEdgeBothPresent(
MapQueryContext<Integer, Set<Integer>, ?> sc,
@NotNull MapEntry<Integer, Set<Integer>> sEntry, int source,
MapQueryContext<Integer, Set<Integer>, ?> tc,
@NotNull MapEntry<Integer, Set<Integer>> tEntry, int target) {
Set<Integer> sNeighbours = sEntry.value().get();
if (sNeighbours.add(target)) {
Set<Integer> tNeighbours = tEntry.value().get();
boolean added = tNeighbours.add(source);
assert added;
sEntry.doReplaceValue(sc.wrapValueAsData(sNeighbours));
tEntry.doReplaceValue(tc.wrapValueAsData(tNeighbours));
return true;
} else {
return false;
}
}
private static void addEdgePresentAbsent(
MapQueryContext<Integer, Set<Integer>, ?> sc,
@NotNull MapEntry<Integer, Set<Integer>> sEntry, int source,
MapQueryContext<Integer, Set<Integer>, ?> tc, int target) {
Set<Integer> sNeighbours = sEntry.value().get();
boolean added = sNeighbours.add(target);
assert added;
sEntry.doReplaceValue(sc.wrapValueAsData(sNeighbours));
addEdgeOneSide(tc, source);
}
private static void addEdgeBothAbsent(MapQueryContext<Integer, Set<Integer>, ?> sc, int source,
MapQueryContext<Integer, Set<Integer>, ?> tc, int target) {
addEdgeOneSide(sc, target);
addEdgeOneSide(tc, source);
}
private static void addEdgeOneSide(MapQueryContext<Integer, Set<Integer>, ?> tc, int source) {
Set<Integer> tNeighbours = new HashSet<>();
tNeighbours.add(source);
MapAbsentEntry<Integer, Set<Integer>> tAbsentEntry = tc.absentEntry();
assert tAbsentEntry != null;
tAbsentEntry.doInsert(tc.wrapValueAsData(tNeighbours));
}
public static boolean removeEdge(
ChronicleMap<Integer, Set<Integer>> graph, int source, int target) {
ExternalMapQueryContext<Integer, Set<Integer>, ?> sourceC = graph.queryContext(source);
ExternalMapQueryContext<Integer, Set<Integer>, ?> targetC = graph.queryContext(target);
// order for consistent lock acquisition => avoid dead lock
if (sourceC.segmentIndex() <= targetC.segmentIndex()) {
return innerRemoveEdge(source, sourceC, target, targetC);
} else {
return innerRemoveEdge(target, targetC, source, sourceC);
}
}
private static boolean innerRemoveEdge(
int source, ExternalMapQueryContext<Integer, Set<Integer>, ?> sourceContext,
int target, ExternalMapQueryContext<Integer, Set<Integer>, ?> targetContext) {
try (ExternalMapQueryContext<Integer, Set<Integer>, ?> sc = sourceContext) {
try (ExternalMapQueryContext<Integer, Set<Integer>, ?> tc = targetContext) {
sc.updateLock().lock();
MapEntry<Integer, Set<Integer>> sEntry = sc.entry();
if (sEntry == null)
return false;
Set<Integer> sNeighbours = sEntry.value().get();
if (!sNeighbours.remove(target))
return false;
tc.updateLock().lock();
MapEntry<Integer, Set<Integer>> tEntry = tc.entry();
if (tEntry == null)
throw new IllegalStateException("target node should be present in the graph");
Set<Integer> tNeighbours = tEntry.value().get();
if (!tNeighbours.remove(source))
throw new IllegalStateException("the target node have an edge to the source");
sEntry.doReplaceValue(sc.wrapValueAsData(sNeighbours));
tEntry.doReplaceValue(tc.wrapValueAsData(tNeighbours));
return true;
}
}
}
Usage:
HashSet<Integer> averageValue = new HashSet<>();
for (int i = 0; i < AVERAGE_CONNECTIVITY; i++) {
averageValue.add(i);
}
ChronicleMap<Integer, Set<Integer>> graph = ChronicleMapBuilder
.of(Integer.class, (Class<Set<Integer>>) (Class) Set.class)
.name("graph")
.entries(100)
.averageValue(averageValue)
.create();
addEdge(graph, 1, 2);
removeEdge(graph, 1, 2);
Unlike ConcurrentHashMap
, ChronicleMap
stores its data off-heap; often in a memory mapped file.
It is recommended that you call close()
when you have finished working with a ChronicleMap
.
map.close()
This is especially important when working with Chronicle Map replication, as failure to call close may prevent you from restarting a replicated map on the same port.
In the event that your application crashes, it may not
be possible to call close()
. Your operating system will usually close dangling ports automatically. So, although it is recommended that you close()
when you have finished with the map,
it is not something that you must do; it’s just something that we recommend you should do.
Warning
|
If you call close() too early before you have finished working with the map, this can cause
your JVM to crash. Close MUST be the last thing that you do with the map.
|
You can customize the behaviour of Chronicle Map.
See CM_Tutorial_Behaviour for more details.
Chronicle Map is able to store entry checksums along with entries. With entry checksums it is possible to identify partially written entries (in the case of operating system, or power failure), and corrupted entries (in the case of hardware, memory, or disk corruption) and clean them up during the recovery procedure.
Entry checksums are 32-bit
numbers, computed by a hash function with good avalanche effect.
Theoretically, there is still about a one-in-a-billion chance that after entry corruption, it passes the sum check.
By default, entry checksums are:
-
ON
if the Chronicle Map is persisted to disk (i. e. created viacreatePersistedTo()
method) -
OFF
if the Chronicle Map is purely in-memory.
Storing checksums for a purely in-memory Chronicle Map hardly makes any practical sense, but you might want to disable storing checksums for a persisted Chronicle Map by calling .checksumEntries(false)
on the ChronicleMapBuilder
used to create a map. It makes sense if you don’t need extra safety that checksums provide.
Entry checksums are computed automatically when an entry is inserted into a Chronicle Map, and
re-computed automatically on operations which update the whole value. For example, map.put()
,
map.replace()
, map.compute()
, mapEntry.doReplaceValue()
. See the MapEntry
interface in
Javadocs. If you update values directly, bypassing Chronicle Map logic, keeping the entry checksum up-to-date is also your responsibility.
It is strongly recommended to update off-heap memory of values directly only within a context, and update or write lock held. Within a context, you are provided with an entry object of MapEntry
type. To re-compute entry checksum manually, cast that object to ChecksumEntry
type and call the .updateChecksum()
method:
try (ChronicleMap<Integer, LongValue> map = ChronicleMap
.of(Integer.class, LongValue.class)
.entries(1)
// Entry checksums make sense only for persisted Chronicle Maps, and are ON by
// default for such maps
.createPersistedTo(file)) {
LongValue value = Values.newHeapInstance(LongValue.class);
value.setValue(42);
map.put(1, value);
try (ExternalMapQueryContext<Integer, LongValue, ?> c = map.queryContext(1)) {
// Update lock required for calling ChecksumEntry.checkSum()
c.updateLock().lock();
MapEntry<Integer, LongValue> entry = c.entry();
Assert.assertNotNull(entry);
ChecksumEntry checksumEntry = (ChecksumEntry) entry;
Assert.assertTrue(checksumEntry.checkSum());
// to access off-heap bytes, should call value().getUsing() with Native value
// provided. Simple get() return Heap value by default
LongValue nativeValue =
entry.value().getUsing(Values.newNativeReference(LongValue.class));
// This value bytes update bypass Chronicle Map internals, so checksum is not
// updated automatically
nativeValue.setValue(43);
Assert.assertFalse(checksumEntry.checkSum());
// Restore correct checksum
checksumEntry.updateChecksum();
Assert.assertTrue(checksumEntry.checkSum());
}
}