This repository contains the source code for insights.logicstar.ai.
We track all GitHub pull requests created since 2025-05-15 and analyze them to identify autonomous code agents.
Pull requests are classified using the following rules:
- Human: All PRs that do not match any known agent pattern.
- OpenAI Codex: PRs where the head branch starts with
codex/. - Google Jules: PRs where the first commit is authored by
google-labs-jules[bot]. - GitHub Copilot: PRs where the head branch starts with
copilot/. - Devin: PRs authored by
devin-ai-integration[bot]. - Cursor Agent: PRs where the head branch starts with
cursor/. - Claude Code: PRs where the first commit is authored by
claudeor the head branch starts withclaude/. - OpenHands: PRs where the first commit is authored by
openhands. - Codegen: PRs authored by
codegen-sh. - Tembo: PRs authored by
tembo-io. - Cosine: PRs where the head branch starts with
cosine/
The system consists of:
- A Next.js frontend for querying the database and visualizing insights.
- A Python backend for scraping GitHub pull requests and generating insights.
While we host a live version at insights.logicstar.ai, you can also deploy the system locally.
Ensure you have access to a PostgreSQL database.
Make sure you have npm and node installed. Then run:
npm install
npm run buildConfigure the environment via frontend/.env.local:
DATABASE_URL=postgres://<user>:<password>@localhost:5433/<db>
To start the frontend:
npm run startInstall the backend package in development mode:
pip3 install -e backendYou should now have access to the aitw CLI tool.
Set the following environment variables:
POSTGRES_CONNECT="dbname=<dbname> user=<user> password=<passwd> host=<host> port=<port>"
GOOGLE_APPLICATION_CREDENTIALS="<path_to_gcp_app_creds>"We use GCP for logging scraping worker activity. You can omit the credentials if you're not using GCP logging.
To start a scraping worker, ensure the environment is configured and run:
aitw scrape worker --token <GH_TOKEN>To backfill all PRs since 2025-05-15:
aitw scrape manager backfillTo derive insights from the current database:
aitw insights hourly
aitw insights daily To update with new PRs created or closed since the last update:
aitw scrape manager updateWe recommend running both the update and insights commands regularly (e.g., hourly).
2025-07-10: Added “Cite Us” section, released dataset, and added OpenHands, Cosine, Claude Code, and Codegen agents. 2025-07-08: Initial release of the dashboard.
