Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

In crawler cronjob, take declarations directly from postgres #1183

Closed
Tracked by #1167
michielbdejong opened this issue Sep 26, 2024 · 3 comments
Closed
Tracked by #1167

In crawler cronjob, take declarations directly from postgres #1183

michielbdejong opened this issue Sep 26, 2024 · 3 comments

Comments

@michielbdejong
Copy link
Member

No description provided.

@michielbdejong
Copy link
Member Author

This is an adaptation of milestone 1.2 which reads:

Adapt the ToS;DR webapp to obtain the list of services and documents to annotate from the API instead of the ToS;DR server database

@michielbdejong
Copy link
Member Author

Given the trouble we ran into with the use of very large git repositories, basically trying to use GitHub as a database server, we decided it's better if the crawl engine writes its metadata directly to the postgres database from which ToS;DR is already loading its data.

@michielbdejong
Copy link
Member Author

michielbdejong commented Sep 26, 2024

Working on this in the ota-to-postgres branch of this repo , and the postgres-backend branch of our fork of the OTA engine.

Got a test script working, reading the declarations from pg. Next step: use this as the declarations source for npx ota track

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant