Basic Steps To Avoid Red Flag On Your Scampage

Anoxymen

Member
Registered User
Joined
Jun 18, 2018
Messages
31
Likes
0
Points
6
Age
37
Stop Google from indexing your scampage
Step One Using robots.txt
What is a robots.txt file?
A robots.txt file tells search engine crawlers which pages or files the crawler can or can't request from your site. This is used mainly to avoid overloading your site with requests; it is not a mechanism for keeping a web page out of Google. To keep a web page out of Google, you should use noindex directives (Step two)
Note: If you’re using WordPress, you might see a robots.txt file when you go to yoursite.com/robots.txt, but you won’t be able to find it in your files.

This is because WordPress creates a virtual robots.txt file if there’s no robots.txt in the root directory.

If this happens to you, you’ll need to create a new robots.txt file.


Creating a robots.txt file

You can create a new robots.txt
Add the below code in the robot.txt file

User-agent: *
Disallow: /

this will block all search bots from indexing.
STEP TWO: noindex directives
Using the robots meta tag. Using the code below

<meta name="robots" content="noindex" />

Place the robots meta tag in the <head>section of a given page
Find the source code of the page you want to change, and make sure you’re in between the <head> tags.
image17-7.png

Then paste this line:

<meta name=”robots” content=”nofollow”>

So it should look like this:
image19-8.png

Make sure you’re not putting this line between any other tags––just the <head> tags.


for more info or help use the comments box or Inbox
 

FoxGroup

New Member
Registered User
Joined
Sep 20, 2019
Messages
1
Likes
0
Points
1
Man, I love U!

This is awesome. Never knew that something like this is possible
 
Top Bottom