angelo70 minutes to crawl 586 websites, crash-free. i'm going to put it on a schedule. capjamesg how often are you reindexing homepages? (i know you're indexing way more than that)
[tantek]I'm still hesitant to implement "read-of" because I don't want to separately implement "watch-of" and "listen-of". and whatever is the verb for "looking" at a photo or other static image.
vikanezrimayatried to implement asynchronous streaming templates in Rust. Started from a stream which throws byte chunks at its consumer. The code is several times larger than its output, which is concerning.
[snarfed]capjamesg it's not really a lot of manual work per domain, the crawler runs itself. every now and then a site is unusual and needs debugging, but most of the work will be before and after the crawl
angeloi seeded indieweb.rocks with 870 sites i got from omz13 in the form of a list of, correct me if i'm wrong, all domains with h-cards from the indiemap. after my own representative h-card parsing i found ~600 domains with representative h-cards. does that sound about right snarfed?
omz13angelo my list was initially seeded with domains listed in indie-map, chat-names, indieweb-ring; it was then reduced to sites with an h-card on their root page
omz13at some stage I will run iwstats again but enable representative h-card retrieval which is a bit more network heavy; I was waiting until my fetch library had some more battle-hardening