Currently yaml files are manually added/updated via PR request. This requires substantial time to ensure the files are formatted correctly and the links are correct.
Crawlers such as https://crawlee.dev allow scripts to read an API or webpage, and extract data needed.
Running a crawler could be done in a pipeline, or as a developer step manually. However Crawlers can check many sites/pages and automate the work required to create data files.
Example workflow:

This would allow Crawlers to be used to generate yaml files, which could then be submitted via PRs and go through the validation/virus scanning checks.
Currently yaml files are manually added/updated via PR request. This requires substantial time to ensure the files are formatted correctly and the links are correct.
Crawlers such as https://crawlee.dev allow scripts to read an API or webpage, and extract data needed.
Running a crawler could be done in a pipeline, or as a developer step manually. However Crawlers can check many sites/pages and automate the work required to create data files.
Example workflow:

This would allow Crawlers to be used to generate yaml files, which could then be submitted via PRs and go through the validation/virus scanning checks.