diff --git a/crawler/1_dump_listings.py b/crawler/1_dump_listings.py index 8379940..63f0667 100644 --- a/crawler/1_dump_listings.py +++ b/crawler/1_dump_listings.py @@ -100,7 +100,8 @@ async def _fetch_listings_with_semaphore( ) -> list[dict[str, Any]]: result = [] # we don't know how many pages we have but we stop as soon as there's no more - for page_id in range(999): + for page_id in range(1, 3): + # seems like all searches stop at 1500 entries (page_id * page_size) async with semaphore: try: listing_query_result = await listing_query(