-
Notifications
You must be signed in to change notification settings - Fork 3
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Better URL analysis during -bake #26
Comments
I am very much confused on the issue of uploading a baked version and the difference between a baked version (which find the files with simple server) and the uploaded version, which does not find the files). I tried with my host and github.io - with the same result the baking process takes a a static file surprisingly, the simpleserver contracts the relative url in the with simpleserver I see
(witout the intervening folder) and with uploaded on github:
which means the relative url is relative to the file and then relative to the folder and the file not found. If i remove the constructed folder and rename the index file to I have constructed a small example as a model for academic homepage and have it running when served. you can find it at The source is It is very confusing and I cannot understand why sprinkles does not just copy the satic files when baking. Otherwise - I like it a lot and it is very reactive to changes (when one understands when one has to restart).. Next I will try your suggestions for the classification of blogs |
Further investigation shows that the addition of the folder is not the cause. Sprinkles only misses the links to the next step the chain down and does not insert the related folder with linked pages. This is easily corrected by copying the full static folder. |
The -bake command needs to analyze link hrefs and such to figure out what else to scrape. But the algorithm it uses is a bit too crude - anything not starting with
http:
orhttps:
is considered a local link, but this is false; things can use other protocols, such asmailto:
,javascript:
,file:
,data:
etc., which the scraper shouldn't touch.See #23 for example.
The text was updated successfully, but these errors were encountered: