Check `robots.txt` before starting. - Add `time.sleep(1)` or more between page requests. - Use a descriptive `User-Agent` header. - Limit the total number of requests per session. - Consider if the site has an API instead.