-
Notifications
You must be signed in to change notification settings - Fork 3
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Figure out a full-featured scraping framework #11
Comments
The current definitions look like this: Line 17 in b1f03fb
|
IMHO it's a good idea to scrape the listings from manuals whenever possible. Manuals usually describe the most "official" features of the implementation - they tend to put emphasis on polished features, not experimental ones. So if a SRFI is mentioned in the manual, support is probably reasonably complete. |
Some of the manuals are generated programatically (STklos is one example). Maybe (I don't know) this could be used, somehow. |
Yes - in that case it doesn't matter whether we scrape the source or result. Whichever is easier to do. For STklos we currently scrape |
I added GitLab support to the scraper generator to integrate Loko and Kawa. I also added support for Snow Fort and For Racket, we can scrape |
Fantastic. Thanks you very much for continuing to work on this! |
No problem! (I had to neglect it for a while though. ^^;) I added Racket to the scraper generator (although it's not yet complete, see racket/srfi#10). Now only CHICKEN, Guile and MIT/GNU Scheme remain. |
CHICKEN down with d25c7b7! Thanks to the work of @diamond-lizard. |
Excellent work! |
Currently we use a hand-written scraper for these implementations:
Would be nice to extend the scraper generator in
listings.scm
so it can handle all of the above.The scraper generator currently writes Unix shell scripts that use curl, tar and grep, but that's just an implementation detail. If Scheme had a more fully developed archive framework, we could just as well skip the shell scripts and run the scrapers directly in Scheme.
The main thing is to discover a specification language for the scrapers.
The text was updated successfully, but these errors were encountered: