To ban the 360spider crawler using robots.txt, you can add the following lines to your robots.txt file:
User-Agent: 360Spider Disallow: /
This will instruct the 360spider crawler to not crawl any content on your website.
If you want to use .htaccess to block the 360spider crawler, you can add the following lines to your .htaccess file:
SetEnvIfNoCase User-Agent "^360Spider" bad_bot Deny from env=bad_bot
This will deny access to any user agent containing "360Spider" in the User-Agent header.
Remember to always backup your files before making any changes and test to ensure the blocking is working as expected.