You can prevent robots from crawling pagination pages on your website by using the "robots.txt" file. The "robots.txt" file is a simple text file that is placed in the root directory of your website, and it provides instructions to web robots (also known as "bots" or "crawlers") about which pages or sections of your website should not be crawled.
Here's an example of how you can use "robots.txt" to stop robots from crawling pagination pages:
User-agent: * Disallow: /page/
The "User-agent: *" line specifies that these instructions apply to all robots. The "Disallow: /page/" line tells robots not to crawl any pages that contain the "/page/" directory.
Note that the "robots.txt" file is just a request and robots are not required to follow the instructions in it. However, most well-behaved robots will respect the rules specified in the "robots.txt" file.