* v2.1.0.2 * restore tweaks to simlibp2p * fix: nil ref on size calc * fix: panic should induce shutdown from event_distributor * fix: friendlier initialization that requires less manual kickstarting for test/devnets * fix: fewer available shards than provers should choose shard length * fix: update stored worker registry, improve logging for debug mode * fix: shut the fuck up, peer log * qol: log value should be snake cased * fix:non-archive snap sync issues * fix: separate X448/Decaf448 signed keys, add onion key to registry * fix: overflow arithmetic on frame number comparison * fix: worker registration should be idempotent if inputs are same, otherwise permit updated records * fix: remove global prover state from size calculation * fix: divide by zero case * fix: eager prover * fix: broadcast listener default * qol: diagnostic data for peer authenticator * fix: master/worker connectivity issue in sparse networks tight coupling of peer and workers can sometimes interfere if mesh is sparse, so give workers a pseudoidentity but publish messages with the proper peer key * fix: reorder steps of join creation * fix: join verify frame source + ensure domain is properly padded (unnecessary but good for consistency) * fix: add delegate to protobuf <-> reified join conversion * fix: preempt prover from planning with no workers * fix: use the unallocated workers to generate a proof * qol: underflow causes join fail in first ten frames on test/devnets * qol: small logging tweaks for easier log correlation in debug mode * qol: use fisher-yates shuffle to ensure prover allocations are evenly distributed when scores are equal * qol: separate decisional logic on post-enrollment confirmation into consensus engine, proposer, and worker manager where relevant, refactor out scoring * reuse shard descriptors for both join planning and confirm/reject decisions * fix: add missing interface method and amend test blossomsub to use new peer id basis * fix: only check allocations if they exist * fix: pomw mint proof data needs to be hierarchically under global intrinsic domain * staging temporary state under diagnostics * fix: first phase of distributed lock refactoring * fix: compute intrinsic locking * fix: hypergraph intrinsic locking * fix: token intrinsic locking * fix: update execution engines to support new locking model * fix: adjust tests with new execution shape * fix: weave in lock/unlock semantics to liveness provider * fix lock fallthrough, add missing allocation update * qol: additional logging for diagnostics, also testnet/devnet handling for confirmations * fix: establish grace period on halt scenario to permit recovery * fix: support test/devnet defaults for coverage scenarios * fix: nil ref on consensus halts for non-archive nodes * fix: remove unnecessary prefix from prover ref * add test coverage for fork choice behaviors and replay – once passing, blocker (2) is resolved * fix: no fork replay on repeat for non-archive nodes, snap now behaves correctly * rollup of pre-liveness check lock interactions * ahead of tests, get the protobuf/metrics-related changes out so teams can prepare * add test coverage for distributed lock behaviors – once passing, blocker (3) is resolved * fix: blocker (3) * Dev docs improvements (#445) * Make install deps script more robust * Improve testing instructions * Worker node should stop upon OS SIGINT/SIGTERM signal (#447) * move pebble close to Stop() * move deferred Stop() to Start() * add core id to worker stop log message * create done os signal channel and stop worker upon message to it --------- Co-authored-by: Cassandra Heart <7929478+CassOnMars@users.noreply.github.com> --------- Co-authored-by: Daz <daz_the_corgi@proton.me> Co-authored-by: Black Swan <3999712+blacks1ne@users.noreply.github.com> |
||
|---|---|---|
| alias | ||
| bedlam | ||
| bls48581 | ||
| bulletproofs | ||
| channel | ||
| client | ||
| config | ||
| conntest | ||
| consensus | ||
| crates | ||
| dashboards/grafana | ||
| docker | ||
| emp-ot | ||
| emp-tool | ||
| ferret | ||
| go-libp2p | ||
| go-libp2p-blossomsub | ||
| go-libp2p-kad-dht | ||
| go-multiaddr | ||
| go-multiaddr-dns | ||
| hypergraph | ||
| nekryptology | ||
| node | ||
| pebble | ||
| protobufs | ||
| rpm | ||
| scripts | ||
| signers | ||
| types | ||
| utils | ||
| vdf | ||
| verenc | ||
| .dockerignore | ||
| .env.example | ||
| .gitattributes | ||
| .gitignore | ||
| ARCHITECTURE.md | ||
| Cargo.lock | ||
| Cargo.toml | ||
| CONTRIBUTING.md | ||
| DOCKER-README.md | ||
| Dockerfile.conntest.source | ||
| Dockerfile.release | ||
| Dockerfile.source | ||
| Dockerfile.source.dockerignore | ||
| Dockerfile.sourceavx512 | ||
| Dockerfile.vdf.source | ||
| Dockerfile.vdf.sourceavx512 | ||
| Dockerfile.vdf.sourcezen3 | ||
| Dockerfile.vdf.sourcezen4 | ||
| emp-install.py | ||
| go.mod | ||
| go.sum | ||
| install-qclient.sh | ||
| LICENSE | ||
| README.md | ||
| Taskfile.yaml | ||
| TESTING.md | ||
Quilibrium - 2.1 - Bloom
Quick Start
Running production nodes from source is no longer recommended given build complexity. Please refer to our release information to obtain the latest version.
Running From Source
Ensure you have all required dependencies.
Ubuntu Linux
For Ubuntu Linux, you can install these by running the following from the project root:
./scripts/install-deps-ubuntu.sh
macOS
Because Mac varies in terms of dependency management, we recommend installing Xcode for build toolchain, then use homebrew to install openssl. Other dependencies via homebrew are the dynamically linked version of the libraries, so we recommend manually fetching the required packages:
curl https://gmplib.org/download/gmp/gmp-6.3.0.tar.xz > /tmp/gmp.tar.xz
pushd /tmp/
tar xvf gmp.tar.xz
pushd gmp-6.3.0
./configure
make
make check
sudo make install
popd
git clone https://github.com/flintlib/flint.git
pushd flint
git checkout flint-3.0
./bootstrap.sh
./configure \
--prefix=/usr/local \
--with-gmp=/usr/local \
--with-mpfr=/usr/local \
--enable-static \
--disable-shared \
CFLAGS="-O3"
make
sudo make install
popd
popd
From there, you can trigger generation of all dependencies to build the node with:
task build_node_arm64_macos
gRPC/REST Support
If you want to enable gRPC/REST, add the following entries to your config.yml:
listenGrpcMultiaddr: <multiaddr>
listenRESTMultiaddr: <multiaddr>
Please note: this interface, while read-only, is unauthenticated and not rate- limited. It is recommended that you only enable if you are properly controlling access via firewall or only query via localhost.
Prometheus Metrics
Quilibrium nodes expose comprehensive Prometheus metrics for monitoring and observability. The metrics are organized across several subsystems:
Disk Monitoring (disk_monitor namespace)
Tracks disk usage and space metrics for the node's data directory.
disk_monitor_usage_percentage- Current disk usage percentagedisk_monitor_total_bytes- Total disk space in bytesdisk_monitor_used_bytes- Used disk space in bytesdisk_monitor_free_bytes- Free disk space in bytes
P2P Networking (blossomsub namespace)
Monitors the BlossomSub peer-to-peer protocol performance.
blossomsub_*_total- Various operation counters (add_peer, remove_peer, join, leave, graft, prune, etc.)blossomsub_*_messages- Message count histograms for IHave, IWant, IDontWant messages
Consensus Time Reel (quilibrium.time_reel subsystem)
Tracks consensus timing, fork choice, and blockchain tree operations.
frames_processed_total- Total frames processed (by type and status)equivocations_detected_total- Equivocation detection counterhead_changes_total- Blockchain head changes (advances vs reorganizations)reorganization_depth- Depth histogram of blockchain reorganizationstree_depth/tree_node_count- Current tree structure metricsfork_choice_evaluations_total- Fork choice algorithm executions
Dynamic Fees (quilibrium.dynamic_fees subsystem)
Monitors fee voting and calculation based on sliding window averages.
fee_votes_added_total/fee_votes_dropped_total- Fee vote trackingcurrent_fee_multiplier- Current calculated fee multipliersliding_window_size- Current number of votes in windowfee_vote_distribution- Distribution histogram of fee votes
Event Distribution (quilibrium.event_distributor subsystem)
Tracks internal event processing and distribution.
events_processed_total- Events processed by typesubscribers_count- Current active subscribersbroadcasts_total- Event broadcast counteruptime_seconds- Distributor uptime
Hypergraph State (quilibrium.hypergraph subsystem)
The most comprehensive metrics tracking the CRDT hypergraph operations.
Core Operations
add_vertex_total/remove_vertex_total- Vertex operationsadd_hyperedge_total/remove_hyperedge_total- Hyperedge operations*_duration_seconds- Operation timing histograms
Lookups and Queries
lookup_vertex_total/lookup_hyperedge_total- Lookup countersget_vertex_total/get_hyperedge_total- Get operation counters
Transactions
transaction_total- Transaction counters by statuscommit_total/commit_duration_seconds- Commit metrics
Proofs
traversal_proof_create_total/traversal_proof_verify_total- Proof operationstraversal_proof_duration_seconds- Proof timing
Execution Intrinsics (quilibrium.intrinsics subsystem)
Monitors the execution engine's intrinsic operations.
materialize_total/materialize_duration_seconds- State materializationinvoke_step_total/invoke_step_errors_total- Step executioncommit_total/commit_errors_total- State commitsstate_size_bytes- Current state size by intrinsic type
gRPC Metrics
Standard gRPC server and client metrics are automatically registered, including request duration, message sizes, and in-flight requests.
App Consensus Engine (quilibrium.app_consensus subsystem)
Monitors shard-specific consensus operations for application shards.
frames_processed_total- Total frames processed (by app_address and status)frame_processing_duration_seconds- Frame processing timeframe_validation_total- Frame validation resultsframe_proving_total/frame_proving_duration_seconds- Frame proving metricsframe_publishing_total/frame_publishing_duration_seconds- Frame publishing metricstransactions_collected_total- Transactions collected for framespending_messages_count- Current pending message countexecutors_registered- Current number of registered executorsengine_state- Current engine state (0=stopped through 7=stopping)current_difficulty- Current mining difficultycurrent_frame_number- Current frame number being processedtime_since_last_proven_frame_seconds- Time elapsed since last proven frame
Global Consensus Engine (quilibrium.global_consensus subsystem)
Monitors global consensus operations across all shards.
frames_processed_total- Total global frames processed (by status)frame_processing_duration_seconds- Global frame processing timeframe_validation_total- Global frame validation resultsframe_proving_total/frame_proving_duration_seconds- Global frame proving metricsframe_publishing_total/frame_publishing_duration_seconds- Global frame publishing metricsshard_commitments_collected- Number of shard commitments collectedshard_commitment_collection_duration_seconds- Time to collect shard commitmentsexecutors_registered- Current number of registered shard executorsengine_state- Current engine state (0=stopped through 7=stopping)current_difficulty- Current global consensus difficultycurrent_frame_number- Current global frame numbertime_since_last_proven_frame_seconds- Time elapsed since last proven global frameglobal_coordination_total/global_coordination_duration_seconds- Global coordination metricsstate_summaries_aggregated- Number of shard state summaries aggregated
Development
Please see the CONTRIBUTING.md file for more information on how to contribute to this repository.
License + Interpretation
Significant portions of Quilibrium's codebase depends on GPL-licensed code, mandating a minimum license of GPL, however Quilibrium is licensed as AGPL to accomodate the scenario in which a cloud provider may wish to coopt the network software. The AGPL allows such providers to do so, provided they are willing to contribute back the management code that interacts with the protocol and node software. To provide clarity, our interpretation is with respect to node provisioning and management tooling for deploying alternative networks, and not applications which are deployed to the network, mainnet status monitors, or container deployments of mainnet nodes from the public codebase.