Simplified distributed block and file storage with strong consistency, like in Ceph
 
 
 
 
 
 
Go to file
Vitaliy Filippov f72f14e6a7
Test / buildenv (push) Successful in 9s Details
Test / build (push) Successful in 2m53s Details
Test / make_test (push) Successful in 32s Details
Test / test_add_osd (push) Successful in 1m27s Details
Test / test_cas (push) Successful in 8s Details
Test / test_change_pg_count (push) Successful in 40s Details
Test / test_change_pg_size (push) Successful in 8s Details
Test / test_create_nomaxid (push) Successful in 8s Details
Test / test_change_pg_count_ec (push) Successful in 39s Details
Test / test_etcd_fail (push) Successful in 53s Details
Test / test_interrupted_rebalance (push) Successful in 1m50s Details
Test / test_failure_domain (push) Successful in 43s Details
Test / test_interrupted_rebalance_ec (push) Successful in 2m3s Details
Test / test_interrupted_rebalance_imm (push) Successful in 2m9s Details
Test / test_snapshot (push) Successful in 24s Details
Test / test_minsize_1 (push) Successful in 14s Details
Test / test_snapshot_ec (push) Successful in 30s Details
Test / test_interrupted_rebalance_ec_imm (push) Successful in 1m24s Details
Test / test_rm (push) Successful in 16s Details
Test / test_snapshot_down (push) Successful in 23s Details
Test / test_snapshot_down_ec (push) Successful in 25s Details
Test / test_splitbrain (push) Successful in 21s Details
Test / test_snapshot_chain (push) Successful in 2m24s Details
Test / test_snapshot_chain_ec (push) Successful in 3m5s Details
Test / test_rebalance_verify_imm (push) Successful in 3m21s Details
Test / test_write (push) Successful in 36s Details
Test / test_rebalance_verify (push) Successful in 4m12s Details
Test / test_write_no_same (push) Successful in 15s Details
Test / test_write_xor (push) Successful in 52s Details
Test / test_rebalance_verify_ec_imm (push) Successful in 4m29s Details
Test / test_rebalance_verify_ec (push) Successful in 5m25s Details
Test / test_heal_pg_size_2 (push) Successful in 4m10s Details
Test / test_heal_ec (push) Successful in 4m46s Details
Test / test_heal_csum_32k_dmj (push) Successful in 5m31s Details
Test / test_heal_csum_32k_dj (push) Successful in 5m41s Details
Test / test_heal_csum_32k (push) Successful in 6m41s Details
Test / test_scrub (push) Successful in 1m13s Details
Test / test_heal_csum_4k_dmj (push) Successful in 6m53s Details
Test / test_scrub_xor (push) Successful in 54s Details
Test / test_scrub_zero_osd_2 (push) Successful in 58s Details
Test / test_heal_csum_4k_dj (push) Successful in 6m27s Details
Test / test_scrub_pg_size_6_pg_minsize_4_osd_count_6_ec (push) Successful in 1m15s Details
Test / test_scrub_pg_size_3 (push) Successful in 1m27s Details
Test / test_heal_csum_4k (push) Successful in 6m20s Details
Test / test_scrub_ec (push) Successful in 29s Details
Test / test_move_reappear (push) Successful in 17s Details
Clear old PG states, history, and OSD states on etcd state reload
Also add protection from etcd watcher messages being split into multiple websocket
messages - I'm not sure if etcd actually does that, but it's better to have extra
protection anyway.

Also check that all etcd watchers are started in the keepalive routine, otherwise
it sometimes tries to revive etcd watchers starting with revision=1 which obviously
always fails because this revision is nearly always compacted.

All these changes should fix an old rarely reproduced bug where SOMETIMES OSDs
didn't react to PG config changes which was leading to offline pools on node reboot.
It happened on the full reload of state from etcd.
2023-12-24 02:02:13 +03:00
.gitea/workflows Verify checksums in test_heal in different combinations 2023-07-29 12:17:18 +03:00
cpp-btree@45e6d1f131 Fix build under GCC 8 2022-05-10 12:26:47 +03:00
csi Fix incorrect error in CSI when searching for the device in /sys 2023-12-14 01:00:32 +03:00
debian Support building qemu 8.1 from bookworm-backports 2023-12-10 00:34:13 +03:00
docker Add tests to CI 2023-05-14 00:06:09 +03:00
docs Add vitastor-disk update-sb command 2023-12-14 01:11:42 +03:00
json11@fd37016cf8 Return the new accidentally rolled back json11 commit ("allow trailing comma") 2023-11-07 15:49:23 +03:00
mon Use the same etcd_ws_keepalive_interval in OSD and mon 2023-12-23 20:07:29 +03:00
patches Remove pve-storage-portal-dns-list format for vitastor_etcd_address 2023-12-20 02:22:06 +03:00
rpm Release 1.3.1 2023-12-04 18:35:09 +03:00
src Clear old PG states, history, and OSD states on etcd state reload 2023-12-24 02:02:13 +03:00
tests Support listening on non-127.0.0.1 in tests 2023-11-01 12:45:27 +03:00
.dockerignore Add the remaining build script for Debian 2021-02-26 01:59:18 +03:00
.gitignore Add .gitignore 2021-03-08 17:04:10 +03:00
.gitmodules Add cpp-btree and json11 submodules 2020-09-17 23:07:06 +03:00
CMakeLists.txt Release 1.3.1 2023-12-04 18:35:09 +03:00
GPL-2.0.txt Add license texts 2020-09-17 23:07:06 +03:00
LICENSE Add LICENSE 2021-04-10 17:44:12 +03:00
README-ru.md Forgot to add new parameter page to README 2023-10-28 13:39:53 +03:00
README.md Forgot to add new parameter page to README 2023-10-28 13:39:53 +03:00
VNPL-1.1-RU.txt Fix typo 2023-01-27 01:52:02 +03:00
VNPL-1.1.txt Fix typo in VNPL-1.1 2022-06-24 01:34:25 +03:00
copy-fio-includes.sh Use packages/ subdir instead of build/ for Docker package builds 2021-02-25 23:59:04 +03:00
copy-qemu-includes.sh Use packages/ subdir instead of build/ for Docker package builds 2021-02-25 23:59:04 +03:00

README.md

Vitastor

Читать на русском

The Idea

Make Clustered Block Storage Fast Again.

Vitastor is a distributed block SDS, direct replacement of Ceph RBD and internal SDS's of public clouds. However, in contrast to them, Vitastor is fast and simple at the same time. The only thing is it's slightly young :-).

Vitastor is architecturally similar to Ceph which means strong consistency, primary-replication, symmetric clustering and automatic data distribution over any number of drives of any size with configurable redundancy (replication or erasure codes/XOR).

Vitastor targets primarily SSD and SSD+HDD clusters with at least 10 Gbit/s network, supports TCP and RDMA and may achieve 4 KB read and write latency as low as ~0.1 ms with proper hardware which is ~10 times faster than other popular SDS's like Ceph or internal systems of public clouds.

Vitastor supports QEMU, NBD, NFS protocols, OpenStack, Proxmox, Kubernetes drivers. More drivers may be created easily.

Read more details below in the documentation.

Talks and presentations

Documentation

Author and License

Copyright (c) Vitaliy Filippov (vitalif [at] yourcmc.ru), 2019+

Join Vitastor Telegram Chat: https://t.me/vitastor

All server-side code (OSD, Monitor and so on) is licensed under the terms of Vitastor Network Public License 1.1 (VNPL 1.1), a copyleft license based on GNU GPLv3.0 with the additional "Network Interaction" clause which requires opensourcing all programs directly or indirectly interacting with Vitastor through a computer network and expressly designed to be used in conjunction with it ("Proxy Programs"). Proxy Programs may be made public not only under the terms of the same license, but also under the terms of any GPL-Compatible Free Software License, as listed by the Free Software Foundation. This is a stricter copyleft license than the Affero GPL.

Please note that VNPL doesn't require you to open the code of proprietary software running inside a VM if it's not specially designed to be used with Vitastor.

Basically, you can't use the software in a proprietary environment to provide its functionality to users without opensourcing all intermediary components standing between the user and Vitastor or purchasing a commercial license from the author 😀.

Client libraries (cluster_client and so on) are dual-licensed under the same VNPL 1.1 and also GNU GPL 2.0 or later to allow for compatibility with GPLed software like QEMU and fio.

You can find the full text of VNPL-1.1 in the file VNPL-1.1.txt. GPL 2.0 is also included in this repository as GPL-2.0.txt.