nertzy, petermolnar, alex11, jonnybarnes, lagash, geoffo, [Murray], tetov-irc, [KevinMarks] and AramZS joined the channel; petermolnar left the channel
#capjamesg_I want to run the IndieWeb Search crawler across multiple computers. How can I do this?
#capjamesg_I know very little about deployment but think k8s / Docker might be in the realm of that for which I am looking.
#capjamesg_Basically, I want an easy way to control computers running the crawler so I can suspend / rebuild / start crawlers easily.
#[manton]Docker and Kubernetes may be overkill to start with. Is it a script you run? I would start by manually running the processes you need on multiple servers and then automate the management of them when you need to. (This is Ruby-specific, but I use Sidekiq across a few servers to manage background tasks. Probably something similar for Python.)
#aaronpkAgreed, it's more a matter of how you organize tasks than an ops or language question
#[manton]Yep. I think we’d need to know more about your current setup… Also, if you’re not already running multiple crawlers on a single server, that would be another place to start.
#capjamesg_[manton] The crawler is multi-threaded so I'm limited in how many instances I can run by resources.
#capjamesg_I have a messaging queue set up for relaying domains to crawl and some other things.
#capjamesg_I can, after adding the right config, set up a crawler and have it send data to the main system from any machine. But I wanted to figure out how to automate this.
#[manton][jacky] I'm glad it's working for someone! No need to send it, but thanks. I have an old one from a few years ago… Maybe something is busted with my account now.