Issues using Dynamic Page Scraper on modern websites (Investing, Yahoo, Redfin)

1. The problem

Hi SCM team,

I’m having consistent issues with the Dynamic Page Scraper. It works fine for light or simple HTML pages, but fails on many modern, heavily protected websites.

Here are some concrete examples:

  1. Investing.com
    https://www.investing.com/equities/nvidia-corp
  • Page loads visually
  • When trying to add selectors, it throws errors
  • Selectors never capture the actual text (empty or incorrect values)
  1. Redfin
    https://www.redfin.com/CA/Burbank/2031-N-Pass-Ave-91505/home/5284386
  • Returns a 403 error
  • Likely blocking bots / automation tools
  1. Yahoo Finance
    https://finance.yahoo.com/quote/NVDA/
  • Page loads, but element picker cannot capture any real content
  • Any selected element results in empty selectors / no actual text

There are more websites I’d like to scrape, but many of them fail in similar ways.

I understand this is challenging because these sites use strong protections (bot detection, automation blocking, JS rendering, etc.). However, it would be very helpful if SCM could support:

  • Custom User-Agent configuration
  • Additional browser fingerprint options
  • Better handling of JS-rendered content
  • Anti-bot mitigation options

Some scrapers are experimenting with open-source AI browser agents, which may provide ideas or inspiration:

I’m not suggesting copying these tools directly, but they may offer useful concepts for improving dynamic scraping reliability.
It would be good to know how to make the Dynamic Page Scraper work on the websites mentioned above.

Thank you

2. Screenshot or task log of the problem




There is 2 limitations.

1- right now the selector tools does fail on some complex sites. Ideally we would find a way to use chromes selection tool. FYI the one we use right now is a couple of years old.

2- as you pointed out some sites have anti scraping code that stops browsers from opening it.

I’m looking at puppeteer for 2 and will need to look for solution’s to 1.

Puppeteer might have some anti scraping mitigation code to reduce fingerprints.

I already have puppeteer working in scm it just requires some user testing.

2 Likes