Simplified distributed block storage with strong consistency, like in Ceph
Go to file
Vitaliy Filippov c777a0041a
Test / buildenv (push) Successful in 12s Details
Test / build (push) Successful in 2m47s Details
Test / test_cas (push) Successful in 10s Details
Test / make_test (push) Successful in 36s Details
Test / test_change_pg_count (push) Successful in 39s Details
Test / test_change_pg_size (push) Successful in 10s Details
Test / test_change_pg_count_ec (push) Successful in 34s Details
Test / test_create_nomaxid (push) Successful in 10s Details
Test / test_etcd_fail (push) Successful in 1m2s Details
Test / test_interrupted_rebalance_imm (push) Successful in 1m54s Details
Test / test_add_osd (push) Successful in 2m48s Details
Test / test_interrupted_rebalance (push) Successful in 2m1s Details
Test / test_failure_domain (push) Successful in 41s Details
Test / test_interrupted_rebalance_ec (push) Successful in 2m3s Details
Test / test_snapshot_ec (push) Successful in 25s Details
Test / test_snapshot (push) Successful in 1m16s Details
Test / test_minsize_1 (push) Successful in 18s Details
Test / test_interrupted_rebalance_ec_imm (push) Successful in 1m23s Details
Test / test_move_reappear (push) Successful in 21s Details
Test / test_rm (push) Successful in 16s Details
Test / test_snapshot_down (push) Successful in 30s Details
Test / test_snapshot_down_ec (push) Successful in 30s Details
Test / test_splitbrain (push) Successful in 25s Details
Test / test_snapshot_chain (push) Successful in 2m18s Details
Test / test_snapshot_chain_ec (push) Successful in 3m13s Details
Test / test_rebalance_verify_imm (push) Successful in 3m8s Details
Test / test_rebalance_verify (push) Successful in 3m41s Details
Test / test_switch_primary (push) Successful in 36s Details
Test / test_write (push) Successful in 40s Details
Test / test_write_no_same (push) Successful in 16s Details
Test / test_write_xor (push) Successful in 39s Details
Test / test_rebalance_verify_ec (push) Successful in 4m56s Details
Test / test_rebalance_verify_ec_imm (push) Successful in 4m21s Details
Test / test_heal_pg_size_2 (push) Successful in 4m15s Details
Test / test_heal_ec (push) Successful in 5m1s Details
Test / test_heal_csum_32k_dj (push) Successful in 5m32s Details
Test / test_heal_csum_32k (push) Successful in 5m38s Details
Test / test_heal_csum_4k_dmj (push) Successful in 5m43s Details
Test / test_scrub (push) Successful in 1m31s Details
Test / test_scrub_zero_osd_2 (push) Successful in 1m17s Details
Test / test_heal_csum_4k_dj (push) Successful in 5m57s Details
Test / test_scrub_xor (push) Successful in 30s Details
Test / test_scrub_pg_size_3 (push) Successful in 1m7s Details
Test / test_scrub_pg_size_6_pg_minsize_4_osd_count_6_ec (push) Successful in 41s Details
Test / test_scrub_ec (push) Successful in 24s Details
Test / test_heal_csum_32k_dmj (push) Successful in 3m56s Details
Test / test_heal_csum_4k (push) Successful in 3m16s Details
Release 1.4.4
A couple of fixes for EC pools

- Fix a segfault possible on partial EC overwrite in 1234 -> 5030 rebalance scenario
- Fix two problems leading to EC pools stalling on rebalance & parallel sudden stops
  of OSDs, for example during a sudden poweroff of a host:
  - Recovery auto-tuning (1.4.0 feature) could apply too large delays and stall
    the EC journal - fixed by limiting delays with a new recovery_tune_sleep_cutoff_us
    parameter (10 seconds by default) and applying recovery pauses before write
    operations, not after them, to not occupy space in the journal for long time
  - Dynamic journal space reservation (1.3.0 feature) wasn't accounting new writes
    when checking the limit so OSDs could still fill the journal fully and stall -
    fixed by including new writes into the limit
- Print etcd dbSize instead of dbSizeInUse in status
2024-02-11 16:23:08 +03:00
.gitea/workflows Raise test_snapshot_chain_ec timeout to 6 minutes 2024-02-11 16:13:52 +03:00
cpp-btree@8de8b467ac Fix "partly outside array bounds" warnings for GCC 12 in cpp-btree 2024-01-15 03:04:33 +03:00
csi Release 1.4.4 2024-02-11 16:23:08 +03:00
debian Release 1.4.4 2024-02-11 16:23:08 +03:00
docker Add tests to CI 2023-05-14 00:06:09 +03:00
docs Add cutoff threshold for recovery auto-tuning 2024-02-11 16:13:52 +03:00
json11@fd37016cf8 Return the new accidentally rolled back json11 commit ("allow trailing comma") 2023-11-07 15:49:23 +03:00
mon Release 1.4.4 2024-02-11 16:23:08 +03:00
patches Release 1.4.4 2024-02-11 16:23:08 +03:00
rpm Release 1.4.4 2024-02-11 16:23:08 +03:00
src Release 1.4.4 2024-02-11 16:23:08 +03:00
tests Fix flapping "scrub" test 2024-01-28 14:59:33 +03:00
.dockerignore Add the remaining build script for Debian 2021-02-26 01:59:18 +03:00
.gitignore Add .gitignore 2021-03-08 17:04:10 +03:00
.gitmodules Add cpp-btree and json11 submodules 2020-09-17 23:07:06 +03:00 Add CLA PR form 2024-01-14 16:48:24 +03:00 Add Contributor License Aggrement in Russian and English 2023-12-31 01:23:52 +03:00
CMakeLists.txt Release 1.4.4 2024-02-11 16:23:08 +03:00
GPL-2.0.txt Add license texts 2020-09-17 23:07:06 +03:00
LICENSE Add LICENSE 2021-04-10 17:44:12 +03:00 Forgot to add new parameter page to README 2023-10-28 13:39:53 +03:00 Forgot to add new parameter page to README 2023-10-28 13:39:53 +03:00
VNPL-1.1-RU.txt Fix typo 2023-01-27 01:52:02 +03:00
VNPL-1.1.txt Fix typo in VNPL-1.1 2022-06-24 01:34:25 +03:00 Use packages/ subdir instead of build/ for Docker package builds 2021-02-25 23:59:04 +03:00 Use packages/ subdir instead of build/ for Docker package builds 2021-02-25 23:59:04 +03:00
pull_request_template.yml Add CLA PR form 2024-01-14 16:48:24 +03:00


Читать на русском

The Idea

Make Clustered Block Storage Fast Again.

Vitastor is a distributed block SDS, direct replacement of Ceph RBD and internal SDS's of public clouds. However, in contrast to them, Vitastor is fast and simple at the same time. The only thing is it's slightly young :-).

Vitastor is architecturally similar to Ceph which means strong consistency, primary-replication, symmetric clustering and automatic data distribution over any number of drives of any size with configurable redundancy (replication or erasure codes/XOR).

Vitastor targets primarily SSD and SSD+HDD clusters with at least 10 Gbit/s network, supports TCP and RDMA and may achieve 4 KB read and write latency as low as ~0.1 ms with proper hardware which is ~10 times faster than other popular SDS's like Ceph or internal systems of public clouds.

Vitastor supports QEMU, NBD, NFS protocols, OpenStack, Proxmox, Kubernetes drivers. More drivers may be created easily.

Read more details below in the documentation.

Talks and presentations


Author and License

Copyright (c) Vitaliy Filippov (vitalif [at], 2019+

Join Vitastor Telegram Chat:

All server-side code (OSD, Monitor and so on) is licensed under the terms of Vitastor Network Public License 1.1 (VNPL 1.1), a copyleft license based on GNU GPLv3.0 with the additional "Network Interaction" clause which requires opensourcing all programs directly or indirectly interacting with Vitastor through a computer network and expressly designed to be used in conjunction with it ("Proxy Programs"). Proxy Programs may be made public not only under the terms of the same license, but also under the terms of any GPL-Compatible Free Software License, as listed by the Free Software Foundation. This is a stricter copyleft license than the Affero GPL.

Please note that VNPL doesn't require you to open the code of proprietary software running inside a VM if it's not specially designed to be used with Vitastor.

Basically, you can't use the software in a proprietary environment to provide its functionality to users without opensourcing all intermediary components standing between the user and Vitastor or purchasing a commercial license from the author 😀.

Client libraries (cluster_client and so on) are dual-licensed under the same VNPL 1.1 and also GNU GPL 2.0 or later to allow for compatibility with GPLed software like QEMU and fio.

You can find the full text of VNPL-1.1 in the file VNPL-1.1.txt. GPL 2.0 is also included in this repository as GPL-2.0.txt.