Respect robots.txt and rate limits - Add delays between requests (3+ seconds recommended) - Cache responses to avoid repeated requests - Include proper User-Agent header - Consider using their data export feature for large datasets