A large portion of valuable web data sits behind login walls — dashboards, internal tools, admin panels, analytics views, private documentation, customer portals, and CRM systems. Until now, Spidra was limited to publicly available pages.
Today, we're introducing authenticated scraping and crawling, which allows Spidra to operate inside a logged-in session, just like a real user's browser.
Spidra uses cookie-based authentication, the same mechanism browsers use to maintain login sessions. Simply log in to the target website using your browser, copy the relevant cookies from DevTools, and pass them to Spidra. From that point on, all requests behave as if they're coming from your authenticated session.
This works in both the Playground and the API. Here's how it looks via the API:
curl --request POST \
--url https://api.spidra.io/api/scrape \
--header 'Content-Type: application/json' \
--header 'x-api-key: YOUR_API_KEY' \
--data '{
"urls": [
{ "url": "https://app.example.com/dashboard" }
],
"cookies": "session_id=abc123; auth_token=xyz789"
}'
Authenticated crawling works the same way — cookies are applied at the start of the crawl and the session is preserved across all discovered pages, so Spidra can navigate links, pagination, and internal sections without re-authentication issues.
The API supports both standard cookie strings and raw DevTools pastes. Spidra automatically detects the format and parses it accordingly. Authentication cookies are never stored — they are used transiently for the duration of a job and discarded immediately afterward.
Read the full blog here.
