- May 25, 2018
-
-
Vladimir Davydov authored
This is required to rework garbage collection in vinyl.
-
Vladimir Davydov authored
We pass lsn of index alter/create records, let's pass lsn of drop record for consistency. This is also needed by vinyl to store it in vylog (see the next patch).
-
Vladimir Davydov authored
If an index was dropped and then recreated, then while replaying vylog we will reuse vy_lsm_recovery_info object corresponding to it. There's no reason why we do that instead of simply allocating a new object - amount of memory saved is negligible, but the code looks more complex. Let's simplify the code - whenever we see VY_LOG_CREATE_LSM, create a new vy_lsm_recovery_info object and replace the old incarnation if any in the hash map.
-
Vladimir Davydov authored
Do not use errinj as it is unreliable. Check that: - No memory is freed by immediately after space drop (WAL is off). - All memory is freed asynchronously after yield.
-
Vladislav Shpilevoy authored
Closes #3425
-
- May 24, 2018
-
-
Vladimir Davydov authored
When a memtx space is dropped or truncated, we delegate freeing tuples stored in it to a background fiber so as not to block the caller (and tx thread) for too long. Turns out it doesn't work out well for ephemeral spaces, which share the destruction code with normal spaces: the problem is the user might issue a lot of complex SQL SELECT statements that create a lot of ephemeral spaces and do not yield and hence don't give the garbage collection fiber a chance to clean up. There's a test that emulates this, 2.0:test/sql-tap/gh-3083-ephemeral-unref-tuples.test.lua. For this test to pass, let's run garbage collection procedure on demand, i.e. when any of memtx allocation functions fails to allocate memory. Follow-up #3408
-
Vladimir Davydov authored
Currently, the engine has not control over yields issued during asynchronous index destruction. As a result, it can't force gc when there's not enough memory. To fix that, let's make gc callback stateful: now it's supposed to free some objects and return true if there's still more objects to free or false otherwise. Yields are now done by the memtx engine itself after each gc callback invocation.
-
- May 22, 2018
-
-
Vladimir Davydov authored
No point in this level of indirection. We embed bps tree implementation into memtx_tree_index, why don't we do the same in case of hash index. A good side effect is that we can now define iterators in headers for both memtx_tree_index and memtx_hash_index, which is required to improve memtx garbage collection mechanism.
-
Vladimir Davydov authored
Since it is created when the memtx engine is initialized, we should destroy it on engine shutdown.
-
Vladimir Davydov authored
All functions that need them are now explicitly passed engine so we can consolidate all variables related to memtx engine state in one place.
-
Vladimir Davydov authored
We need this so that we can force garbage collection when we are short on memory. There are two such functions: one is used for allocating index extents, another for allocating tuples. Index allocating function has an opaque context so we simply reuse it for passing memtx engine to it. To pass memtx engine to tuple allocating function, we add an opaque engine specific pointer to tuple_format and set it to memtx_engine for memtx spaces.
-
Vladimir Davydov authored
The two files are too closely related: memtx_arena is defined and used in memtx_engine.c, but initialized in memtx_tuple.cc. Since memtx_tuple.cc is small, let's fold it into memtx_engine.c.
-
Vladimir Davydov authored
Postponing it until a memtx index is created for the first time saves us no memory or cpu, it only makes the code more difficult to follow.
-
- May 21, 2018
-
-
Vladislav Shpilevoy authored
-
Vladimir Davydov authored
When a memtx space is dropped or truncated, we have to unreference all tuples stored in it. Currently, we do it synchronously, thus blocking the tx thread. If a space is big, tx thread may remain blocked for several seconds, which is unacceptable. This patch makes drop/truncate hand actual work to a background fiber. Before this patch, drop of a space with 10M 64-byte records took more than 0.5 seconds. After this patch, it takes less than 1 millisecond. Closes #3408
-
Vladimir Davydov authored
Force major compaction of all ranges when index.compact() is called. Note, the function only triggers compaction, it doesn't wait until compaction is complete. Closes #3139
-
Vladimir Davydov authored
This patch adds index.compact() Lua method. The new method is backed by index_vtab::compact. Currently, it's a no-op for all kinds of indexes. It will be used by Vinyl engine in order to trigger major compaction. Part of #3139
-
- May 19, 2018
-
-
Konstantin Belyavskiy authored
This test falls from time to time, because .xlog may have a different number in a name (and using box.info.lsn is not an option here). Since it's setup of two masters, it could be one or two xlogs in a folder, so first get a list of all matching files and then delete the last one.
-
- May 18, 2018
-
-
Vladislav Shpilevoy authored
Simplify collation code.
-
- May 17, 2018
-
-
Vladislav Shpilevoy authored
utf8 is a module partially compatible with Lua 5.3 utf8 and lua-utf8 third party module. Partially means, that not all functions are implemented. The patch introduces these ones: upper, lower, len, char, sub, next. Len and char works exactly like in Lua 5.3. Other functions work like in lua-utf8, because they are not presented in Lua 5.3. Tarantool utf8 has extensions: * isupper/lower/alpha/digit, that check some property by a symbol or by its code; * cmp/casecmp, that compare two UTF8 strings. Closes #3290 Closes #3385 Closes #3081
-
Vladislav Shpilevoy authored
Collation fingerprint is a formatted string unique for a set of collation properties. Equal collations with different names have the same fingerprint. This new property is used to build collation fingerprint cache to use in Tarantool internals, where collation name does not matter. Fingerprint cache can never conflict or replace on insertion into it. It means, that, for example, utf8 module being created in this patchset, can fill collation cache with its own collations and it will affect neither users or other modules.
-
Vladislav Shpilevoy authored
In the issue #3290 the important problem appeared - Tarantool can not create completely internal collations with no ID, name, owner. Just for internal usage. Original struct coll can not be used for this since * it has fields that are not needed in internals; * collation name is public thing, and the collation cache uses it, so it would be necessary to forbid to a user usage of some system names; * when multiple collations has the same comparator and only their names/owners/IDs are different, the separate UCollator objects are created, but it would be good to be able to reference a single one. This patch renames coll to coll_id, coll_def to call_id_def and introduces coll - pure collation object with no any user defined things. Needed for #3290.
-
Vladimir Davydov authored
If a compacted run was created after the last checkpoint, it is not needed to recover from any checkpoint and hence can be deleted right away to save disk space. Closes #3407
-
- May 16, 2018
-
-
Konstantin Osipov authored
A follow up for the patch for gh-3051
-
Kirill Shcherbatov authored
The format was lost when performing update operations as new tuple was referenced to default runtime format. Fixes #3051
-
Kirill Shcherbatov authored
Tuple has been checked instead of new_tuple returned as box_tuple_update result.
-
Konstantin Osipov authored
No other changes.
-
Konstantin Osipov authored
No semantical changees.
-
Konstantin Osipov authored
No semantical changes.
-
Konstantin Belyavskiy authored
Small refactoring: remove 'enum replica_state' since reuse a subset from applier state machine 'enum replica_state' to check if we have achieved replication quorum and hence can leave read-only mode.
-
Vladislav Shpilevoy authored
Now netbox decodes iproto packet body twice: first, MessagePack -> Lua, then Lua -> box.tuple. Let's convert MessagePack to box.tuple objects without a round trip to Lua. This should reduce the amount of garbage produced by net.box. Part of #3333
-
- May 15, 2018
-
-
Konstantin Osipov authored
-
Vladimir Davydov authored
Improve the test by decreasing range_size so that it creates a lot of ranges for test indexes, not just one. This helped find bugs causing the crash described in #3393. Follow-up #3393
-
Vladimir Davydov authored
Although the bug in vy_task_dump_complete() due to which a tuple could be lost during dump was fixed, there still may be affected deployments as the bug was persisted on disk. To avoid occasional crashes on such deployments, let's make vinyl_iterator_secondary_next() skip tuples that are present in a secondary index but missing in the primary. Closes #3393
-
Vladimir Davydov authored
vy_task_dump_complete() creates a slice per each range overlapping with the newly written run. It uses vy_range_tree_psearch(min_key) to find the first overlapping range and nsearch(max_key) to find the range immediately following the last overlapping range. This is incorrect as nsearch rb tree method returns the element matching the search key if it is present in the tree. That is, if the max key written to a run turns out to be equal the beginning of a range, the slice won't be created for it and it will be silently and persistently lost. The issue manifests itself as crash in vinyl_iterator_secondary_next(), when we fail to find the tuple in the primary index corresponding to a statement found in a secondary index. Part of #3393
-
Vladimir Davydov authored
vy_run_iterator_seek() is supposed to check that the resulting statement matches the search key in case of ITER_EQ, but if the search key lies at the beginning of the slice, it doesn't. As a result, vy_point_lookup() may fail to find an existing tuple as demonstrated below. Suppose we are looking for key {10} in the primary index which consists of an empty mem and two runs: run 1: DELETE{15} run 2: INSERT{10} vy_run_iterator_next() returns DELETE{15} for run 1 because of the missing EQ check and vy_point_lookup() stops at run 1 (since the terminal statement is found) and mistakenly returns NULL. The issue manifests itself as crash in vinyl_iterator_secondary_next(), when we fail to find the tuple in the primary index corresponding to a statement found in a secondary index. Part of #3393
-
Kirill Yukhin authored
-
Alexander Turenko authored
Follows up #3396.
-
- May 14, 2018
-
-
Vladislav Shpilevoy authored
When a connection is closed, some of long-poll requests still may by in TX thread with non-discarded input. If a connection is closed, and then an input is discarded, then connection must not try to read new data. The bug was introduced here: f4d66dae by me. Closes #3400
-