Use robots.txt to block crawlers from "action URLs." This prevents wasted server resources from useless crawler hits. It's an age-old best practice that remains relevant today. Google's Gary Illyes ...
Results that may be inaccessible to you are currently showing.
Hide inaccessible results