I wrote an algorithm that scheduled crawling jobs efficiently across 24 hour window with different crawling frequency based on the posting frequency of each account. This lead to eliminating rate limit errors, efficient data crawling without any duplicates and the most important thing, no missing data.