mirror of
https://github.com/ipfs/kubo.git
synced 2026-02-26 04:47:45 +08:00
docs: update badger section in config.md (#8662)
This should fix the issue of users thinking badger is "no-brainer faster choice" and then running into problems. Co-authored-by: Johnny <9611008+johnnymatthews@users.noreply.github.com>
This commit is contained in:
parent
bc7ddef8dc
commit
4ad316803e
@ -188,37 +188,36 @@ documented in `ipfs config profile --help`.
|
||||
|
||||
- `flatfs`
|
||||
|
||||
Configures the node to use the flatfs datastore.
|
||||
Configures the node to use the flatfs datastore. Flatfs is the default datastore.
|
||||
|
||||
This is the most battle-tested and reliable datastore, but it's significantly
|
||||
slower than the badger datastore. You should use this datastore if:
|
||||
This is the most battle-tested and reliable datastore.
|
||||
You should use this datastore if:
|
||||
|
||||
- You need a very simple and very reliable datastore and you trust your
|
||||
- You need a very simple and very reliable datastore, and you trust your
|
||||
filesystem. This datastore stores each block as a separate file in the
|
||||
underlying filesystem so it's unlikely to lose data unless there's an issue
|
||||
with the underlying file system.
|
||||
- You need to run garbage collection on a small (<= 10GiB) datastore. The
|
||||
default datastore, badger, can leave several gigabytes of data behind when
|
||||
garbage collecting.
|
||||
- You're concerned about memory usage. In its default configuration, badger can
|
||||
use up to several gigabytes of memory.
|
||||
- You need to run garbage collection in a way that reclaims free space as soon as possible.
|
||||
- You want to minimize memory usage.
|
||||
- You are ok with the default speed of data import, or prefer to use `--nocopy`.
|
||||
|
||||
This profile may only be applied when first initializing the node.
|
||||
|
||||
|
||||
- `badgerds`
|
||||
|
||||
Configures the node to use the badger datastore.
|
||||
|
||||
This is the fastest datastore. Use this datastore if performance, especially
|
||||
when adding many gigabytes of files, is critical. However:
|
||||
Configures the node to use the experimental badger datastore. Keep in mind that this **uses an outdated badger 1.x**.
|
||||
|
||||
Use this datastore if some aspects of performance,
|
||||
especially the speed of adding many gigabytes of files, are critical. However, be aware that:
|
||||
|
||||
- This datastore will not properly reclaim space when your datastore is
|
||||
smaller than several gigabytes. If you run IPFS with '--enable-gc' (you have
|
||||
enabled block-level garbage collection), you plan on storing very little data in
|
||||
smaller than several gigabytes. If you run IPFS with `--enable-gc`, you plan on storing very little data in
|
||||
your IPFS node, and disk usage is more critical than performance, consider using
|
||||
flatfs.
|
||||
- This datastore uses up to several gigabytes of memory.
|
||||
`flatfs`.
|
||||
- This datastore uses up to several gigabytes of memory.
|
||||
- Good for medium-size datastores, but may run into performance issues if your dataset is bigger than a terabyte.
|
||||
- The current implementation is based on old badger 1.x which is no longer supported by the upstream team.
|
||||
|
||||
This profile may only be applied when first initializing the node.
|
||||
|
||||
|
||||
Loading…
Reference in New Issue
Block a user