This is the main repo for the AI bot we call Garbo. Garbo is a Discord bot that is powered by LLM:s to effectively fetch and extract GHG self-reported data from companies. It automates the process of data extraction, evaluation, and formatting, providing a streamlined workflow for handling environmental data.
Garbo is invoked through a set of commands in Discord and has a pipeline of tasks that will be started in order for her to both extract, evaluate and format the data autonomously.
We utilise an open source queue manager called BullMQ which relies on Redis. The data is then stored into DB and Wikidata.
Test the app in Discord channel #rapporter-att-granska by using the command /pdf and Garbo will be answering with a parsed JSON
Some of the following steps will be performed in parallel and most will be asynchronous. If a process is failed it's important to be able to restart it after a new code release so we can iterate on the prompts etc without having to restart the whole process again.
flowchart TB
PDF[PDF]
Cache{Is in cache?}
NLM[Parse PDF]
Tables[Extract Tables]
Emissions[Extract Emissions]
Industry[Industry]
Goals[Climate Goals]
Review[Discord Review]
Precheck --> GuessWikidata --> Emissions
Precheck --> FiscalYear --> Emissions
PDF --> Cache --(no)--> NLM --> Tables --> Precheck
Cache --(yes)--> Precheck
CheckDB{Exists in API?}
Emissions --(followUp)--> Industry --> CheckDB --(yes)--> Review --> API.Industry
CheckDB --(no)--> API.Industry
Emissions --(followUp)--> Scope1+2 --> CheckDB --(yes)--> Review --> API.Emissions
CheckDB --(no)--> API.Emissions
Emissions --(followUp)--> Scope3 --> CheckDB --(yes)--> Review --> API.Emissions
CheckDB --(no)--> API.Emissions
Emissions --(followUp)--> Biogenic --> CheckDB --(yes)--> Review --> API.Emissions
CheckDB --(no)--> API.Emissions
Emissions --(followUp)--> Goals --> CheckDB --(yes)--> Review --> API.Goals
CheckDB --(no)--> API.Goals
Emissions --(followUp)--> Initiatives --> CheckDB --(yes)--> Review --> API.Initiatives
CheckDB --(no)--> API.Initiatives
Emissions --(followUp)--> Turnover --> CheckDB --(yes)--> Review --> API.Economy
CheckDB --(no)--> API.Initiatives
Emissions --(followUp)--> Employees --> CheckDB --(yes)--> Review --> API.Economy
CheckDB --(no)--> API.Economy
Ensure you have Node.js version 22.0.0 or higher installed. You will also need Docker (or Podman) to run containers.
Make a copy of the file .env.example
and name it .env
. Fill it in using the instructions in the file.
npm i
Note
If you use a Linux-based operating system, you might need to install additional dependencies for the third-party package canvas
. Follow the instructions.
This project expects some containers running in the background to work properly. We use Postgres as our primary database, Redis for managing the queue system, ChromaDB for embeddings and the NLM ingestor for parsing PDF:s.
The simplest way to start the containers the first time is to run the following docker commands.
docker run -d -p 5432:5432 --name garbo_postgres -e POSTGRES_PASSWORD=mysecretpassword postgres
# These are only necessary to develop the AI pipeline. Feel free to skip them if you only plan to develop the frontend and/or the API.
docker run -d -p 6379:6379 --name garbo_redis redis
docker run -d -p 8000:8000 --name garbo_chroma chromadb/chroma
docker run -d -p 5001:5001 --name garbo_ingestor ghcr.io/nlmatics/nlm-ingestor
Next time, you can start the containers back up using
docker start garbo_postgres garbo_redis garbo_chroma garbo_ingestor
Or if you only plan to develop the frontend and/or the API, this is enough:
docker start garbo_postgres
You may want a graphical user interface to make it easier to manage your local containers. Podman desktop and Rancher desktop are both good alternatives
This applies migrations and seeding data needed for development.
npm run prisma migrate dev
Note
This step is very helpful to get a good starting point for developing and testing the frontend and/or the API. However, you may also skip it if you want to start with a clean database.
First, ask one of the Klimatkollen team members and they will send you a database backup.
Not required the first time: Delete the database to make sure it doesn't exist:
docker exec -i garbo_postgres dropdb -f -U postgres --if-exists garbo
Then, replace ~/Downloads/backup_garbo_XYZ.dump
with the path to your DB backup file and restore the database backup with the following command:
docker exec -i garbo_postgres pg_restore -C -v -d postgres -U postgres < ~/Downloads/backup_garbo_XYZ.dump
The code can be started in three main ways, depending on what you plan to develop/test/run locally.
Note
If you plan to develop the frontend and/or the API, this is the best way to get started:
npm run dev-api
If you plan to develop the AI pipeline, this is the recommended way to start the code.
First, run the following command to start the API and the queue system, including an admin dashboard to view progress, logs and more.
npm run dev-board
Now you can go to http://localhost:3000 and see the dashboard.
Then, open another terminal and start the AI pipeline and its workers, which are responsible for processing each report. These can be scaled horizontally.
npm run dev-workers
Get everything up and running with one command (with all output in one terminal).
npm run dev
Well done! You've now set up the garbo
backend and are ready to start development :)
These steps can be useful to test DB migrations with data similar to the production environment.
- Recommended: Create a local test DB. This allows you to keep your regular development DB intact.
docker run -d -p 5432:5432 --name garbo_test_postgres -e POSTGRES_PASSWORD=mysecretpassword postgres
Alternatively, make sure your local postgres container is running.
-
Ask one of the Klimatkollen team members and they will send you a database backup.
-
Delete the database if it exists:
docker exec -i garbo_test_postgres dropdb -f -U postgres --if-exists garbo
- Restore the backup. This will initially connect to the default
postgres
database without making any modifications and then create any databases if they do not exist
docker exec -i garbo_test_postgres pg_restore -C -v -d postgres -U postgres < ~/Downloads/backup_garbo_XYZ.dump
-
Test the DB migrations with
npm run prisma migrate dev
. -
Restart the Garbo API and workers and verify the migration was successful.
To run the tests, use the following command:
npm test
To run the application
docker run -d -p 3000:3000 ghcr.io/klimatbyran/garbo npm start
# start how many workers you want:
docker run -d ghcr.io/klimatbyran/garbo npm run workers
docker run -d ghcr.io/klimatbyran/garbo npm run workers
docker run -d ghcr.io/klimatbyran/garbo npm run workers
# first time you need to initialize the postgres database:
npm run prisma db push # create tables
npm run prisma db seed # seed the data with initial content
This application is deployed in production with Kubernetes and uses FluxCD as CD pipeline. The yaml files in the k8s directory are automatically synced to the cluster. If you want to run a fork of the application yourself - just add these helm charts as dependencies:
postgresql (bitnami)
redis (bitnami)
chromadb
metabase
To create secret in the k8s cluster - use this command to transfer your .env file as secret to the cluster:
kubectl create secret generic env --from-env-file=.env
We welcome contributions! Please see our CONTRIBUTING.md for guidelines on how to contribute to this project.
For any questions or issues, please contact the maintainers at [email protected] and you will get an invite to our Discord.
This project is licensed under the terms of the MIT License © Klimatbyrån Ideell Förening.