Sukrit Kumar Vemula

SCRAPR - The data layer for the agentic web

SCRAPR is a new approach to web data extraction. Instead of relying on fragile DOM selectors or heavy browser automation, SCRAPR looks at how modern websites actually load their data and extracts structured responses directly from those sources. The goal is to make web data pipelines faster, more reliable, and easier to maintain. Right now SCRAPR is in early MVP and we’re looking for developers, data teams, and AI builders who need clean structured data from websites.

Add a comment

Replies

Best
Julian Collins

The interception approach is clever, way faster than spinning up a headless browser for every request. Have you thought about a batch endpoint where you can throw a list of URLs at it in one call? Anytime I've built a scraping pipeline for a project, the single-URL-at-a-time loop is where things get slow and annoying to manage.

Sukrit Kumar Vemula
@juelz Thanks Julian, really appreciate that! And yes — that’s a great point. Running things one URL at a time can definitely become slow when you’re building pipelines. There’s already support for batch-style requests where you can pass multiple URLs in one call, and I’m planning to expand that further so it works better for larger data pipelines.
Emad Ibrahim

Smart approach intercepting the underlying API calls instead of fighting the DOM. I've built data pipelines that relied on traditional scraping and the maintenance burden of broken selectors is brutal. Curious -- do you have plans for a schema definition layer where users can map the intercepted responses to a consistent output format? That would make it really useful for feeding structured data into AI workflows.

Sukrit Kumar Vemula
@emad_ibrahim Thanks Emad, really appreciate that. And yeah, the maintenance from broken selectors is exactly one of the main problems I wanted to avoid. A schema / mapping layer is definitely something I’ve been thinking about. Right now the focus is on getting clean structured responses out reliably, but adding a way for users to map or normalize outputs for pipelines and AI workflows would make a lot of sense.
Handuo

Really smart approach to web scraping. Focusing on where data actually comes from rather than relying on DOM selectors is a much more resilient strategy. Most scraping tools break the moment a site updates its frontend, so anchoring to underlying API calls makes a lot of sense.

Curious about how you handle rate limiting and sites that aggressively block automated access. Either way, congrats on the launch!

Sukrit Kumar Vemula
@handuo Thanks, really appreciate that! For things like rate limiting or stricter access controls, it really depends on how the specific site handles requests. SCRAPR focuses on keeping requests lightweight and behaving like a normal client rather than relying on heavy browser automation.
Joel Farthing

Great implementation! Is the live demo on the website operable? I can't seem to enter text into the fields. Early access requested!

Sukrit Kumar Vemula

@joel_farthing Thanks, really appreciate that!

The demo on the site is more of a preview right now, so the input fields aren’t fully interactive yet. I’m working on making a proper live demo soon.

Glad you requested early access — I’ll make sure you get access as we roll out the next version!

David Parrelli

Intercepting network calls instead of rendering pages is a smart approach. Way less fragile than the usual scraping setups. What kinds of sites have been trickiest to support so far?

Sukrit Kumar Vemula

@dparrelli Thanks, appreciate that!

Some of the trickier ones tend to be sites that generate requests dynamically or rely heavily on session-based flows, since those can behave differently depending on how the page loads.

But overall most modern sites still rely on some form of underlying data requests.

Arjun Chintapalli

Wait also @gabe how is this even allowed as per Product Hunt launch rules, this is just a vercel app website with a waitlist?

I thought the product hunt rules were that no waitlists.

Sukrit Kumar Vemula
@arjun_chintapalli Hey! The API itself is actually built and mostly ready. I'm a student, so for now I'm planning to host the first version on Vercel using their free serverless resources. The waitlist is mainly to manage early access while I finalize deployment and scaling. Appreciate the feedback!
suifeng

This is such a smart pivot from the usual DOM-parsing headaches! As a dev who's spent way too many hours fixing scrapers because of a tiny CSS change, focusing on the data responses directly sounds like a lifesaver. How do you handle sites with heavy anti-bot protections or obfuscated API endpoints?

Avinash S

The "data layer for the agentic web" framing is interesting - curious how you're handling anti-bot countermeasures that vary by target site. Are you routing through rotating proxies or using something more sophisticated on the infrastructure side? Asking because this seems like it gets complicated fast at scale.