Fork of Polkastats-backend-v3 to work with Social Network Energy. This repo is used to index data into postgres for providing a higher performance API to front end applications.
git clone https://github.com/Colm3na/polkastats-backend-v3.git
cd polkastats-backend-v3
npm install
To launch all docker containers at once:
npm run docker
To run them separately:
npm run docker:<container-name>
- substrate-node
- postgres
- graphql-engine
- crawler
- phragmen (temporarily disabled)
git pull
npm run docker:clean
npm run docker:build
npm run docker
To deploy the crawler and social network node to kubernetes see the /charts folder.
This crawler container listens to new blocks and fills the database. There are a number of processes executed within this container. Some of this processes are triggered based on time configuration that can be seen in this file: backend.config.js The crawler is able to detect and fill the gaps in postgres database by harvesting all the missing data, so it's safe and resilience against node outages or restarts.
This container includes an offline-phragmen binary. It is a forked modification of Kianenigma repository.
The crawler needs to wait for your substrate-node container to get synced before starting to collect data. You can use an already synced external RPC for instant testing by changing the environment variable WS_PROVIDER_URL in docker-compose.yml
file:
crawler:
image: polkastats-backend:latest
build:
context: ../../
dockerfile: ./docker/polkastats-backend/backend/Dockerfile
depends_on:
- "postgres"
- "substrate-node"
restart: on-failure
environment:
- NODE_ENV=production
- WS_PROVIDER_URL=wss://kusama-rpc.polkadot.io # Change this line
Just uncomment out the first one and comment the second and rebuild the dockers.
npm run docker:clean
npm run docker
Then browse to http://localhost:8082
Click on "Data" at the top menu
Then add all tables to the tracking process
From now on, hasura will be collecting and tracking all the changes in the data base.
In order to check it and see its power you could start a new subscription or just perform an example query such us this one:
- Block query example:
query {
block {
block_hash
block_author
block_number
block_author_name
current_era
current_index
new_accounts
session_length
session_per_era
session_progress
}
}
- Rewards query example:
query {
rewards {
era_index
era_rewards
stash_id
timestamp
}
}
- Validator by number of nominators example:
query {
validator_num_nominators {
block_number
nominators
timestamp
}
}
- Account query example:
query {
account {
account_id
balances
identity
}
}
- Block subscription example:
subscription {
block {
block_number
block_hash
current_era
current_index
}
}
- Validator active subscription example:
subscription MySubscription {
validator_active {
account_id
active
block_number
session_index
timestamp
}
}
- Account subscription example:
subscription MySubscription {
account {
account_id
balances
}
}
You can customize your configuration through the following environment variables:
Env name | Description | Default Value |
---|---|---|
WS_PROVIDER_URL | Substrate node to use | ws://substrate-node:9944 |
Env name | Description | Default Value |
---|---|---|
POSTGRES_USER | PostgreSQL username | polkastats |
POSTGRES_PASSWORD | PostgreSQL user password | polkastats |
POSTGRES_HOST | PostgreSQL host | postgres |
POSTGRES_PORT | PostgreSQL port | 5432 |
POSTGRES_DATABASE | PostgreSQL database name | polkastats |
Env name | Description | Default Value |
---|---|---|
CRAWLER_BLOCK_LISTENER_POLLING_TIME_MS | Polling time for block listener (ms) | 60000 |
CRAWLER_ACTIVE_ACCOUNTS_POLLING_TIME_MS | Polling time for accounts crawler (ms) | 600000 |
CRAWLER_PHRAGMEN_POLLING_TIME_MS | Polling time for phragmen executions (ms) | 300000 |
CRAWLER_PHRAGMEN_OUTPUT_DIR | Directory to store the phgramen JSON result | /tmp/phragmen |
CRAWLER_PHRAGMEN_BINARY_PATH | Path to the phragmen executable | /usr/app/polkastats-backend-v3/offline-phragme |
You can also disable specific crawlers with the following environment variables:
CRAWLER_SYSTEM_DISABLE
CRAWLER_BLOCK_LISTENER_DISABLE
CRAWLER_BLOCK_HARVESTER_DISABLE
CRAWLER_STAKING_DISABLE
CRAWLER_ACTIVE_ACCOUNTS_DISABLE
CRAWLER_CHAIN_DISABLE
CRAWLER_ERA_LISTENER_DISABLE
CRAWLER_PHRAGMEN_DISABLE
For instance, if you want to disable phgramen crawler just set CRAWLER_PHRAGMEN_DISABLE=true
.