Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Given that the author was quite vague about the nature of this “pipeline” and that their product is an “AI-powered Startup Selection engine”, I have a suspicion they were probably crawling and scraping a whole bunch of pages for new startups. It’s possible that this was totally legit and it just looked like a ddos attack, or that it was something else entirely, but everyone here seems to have taken him at his word that what they were doing was actually above board.


I agree that we're not getting the full story.

On the other hand, completely shutting down all their services without a quick conversation...


Having been on the receiving end of terribly broken "pipelines" at startups wanting to hammer away at my resources, the right response is terminate first discuss later.

I know of a company that explicitly had a "call us to discuss first" clause in their contract with a smaller cloud provider. Everyone was on holiday and not answering the phones while their hacked account was being used to spin up dozens of boxes launching a DoS attack against a crypto scam site. Guess who had to eat the bill on that one?


Surely there was a support contract option where there were on-call people?


In case something looking like a DDoS, they tend to disable network access to droplets until you manually 'resolve' it. Instead of destroying the whole account.


My guess is that the whole story is trash considering I found out about it via promoted tweet. Someone paid to put that on my feed. Noise.


Even if there is legitimate cause to disable the account, I would prefer to see DO being more transparent with their process here.


What's wrong with scraping a bunch of pages. As long as they are following robots.txt, it's no big deal.


Or even better, have contracts with the companies. Maybe unlikely for them, but I think “scraping” is too often assumed to be “bad” in some way. The company I work for does a lot of web scraping, but we have contracts with our partners to scrape their websites. They may still have robots.txt that ask users not to scrape some areas, but we are allowed to bypass those.


HN always had a boner for web-crawler hate. How dare you automate downloading of public data others post online!




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: