Revision history for Mod033bRobotIndexing
Revision [23266]
Last edited on 2016-05-20 07:38:47 by JavaWoman [Replaces old-style internal links with new pipe-split links.]Additions:
**[[http://www.wakkawiki.com/CharlesNepote | CharlesNepote]]** @ [[http://www.wikini.net/ | WikiNi]]
**[[http://www.wakkawiki.com/MatthiasAppel | MatthiasAppel]]**
Hm, does anybody know how to make search robots not to index the ../edit ../revisions and ../referrers links on every page? You can search for WakkaWiki on [[http://www.google.de/search?q=+site:www.wakkawiki.com+wakkawiki&hl=de&lr=&ie=UTF-8&oe=UTF-8&start=50&sa=N | Google]] and see the result.
--[[http://www.wakkawiki.com/MatthiasAppel | MatthiasAppel]]
-- [[http://www.wakkawiki.com/CharlesNepote | CharlesNepote]]
--[[http://www.wakkawiki.com/MatthiasAppel | MatthiasAppel]]
''Did it work with robots.txt? And what does $ mean?? What is with referrers_sites?'' --[[http://www.wakkawiki.com/MoE | MoE]]
~& Yahoo! recently [[http://www.ysearchblog.com/archives/000372.html | announced wildcard support]], so the above mentioned solution works for Yahoo! Slurp as well. Up to now I haven't found a solution for msn. :( --DaC
On [[http://www.wikini.net/ | WikiNi]] we code that in [[http://cvsweb.tuxfamily.org/cvs/wikini/actions/header.php?rev=1.12&content-type=text/x-cvsweb-markup&cvsroot=wikini | header.php]] :
-- [[http://www.wakkawiki.com/CharlesNepote | CharlesNepote]]
**[[http://www.wakkawiki.com/MatthiasAppel | MatthiasAppel]]**
Hm, does anybody know how to make search robots not to index the ../edit ../revisions and ../referrers links on every page? You can search for WakkaWiki on [[http://www.google.de/search?q=+site:www.wakkawiki.com+wakkawiki&hl=de&lr=&ie=UTF-8&oe=UTF-8&start=50&sa=N | Google]] and see the result.
--[[http://www.wakkawiki.com/MatthiasAppel | MatthiasAppel]]
-- [[http://www.wakkawiki.com/CharlesNepote | CharlesNepote]]
--[[http://www.wakkawiki.com/MatthiasAppel | MatthiasAppel]]
''Did it work with robots.txt? And what does $ mean?? What is with referrers_sites?'' --[[http://www.wakkawiki.com/MoE | MoE]]
~& Yahoo! recently [[http://www.ysearchblog.com/archives/000372.html | announced wildcard support]], so the above mentioned solution works for Yahoo! Slurp as well. Up to now I haven't found a solution for msn. :( --DaC
On [[http://www.wikini.net/ | WikiNi]] we code that in [[http://cvsweb.tuxfamily.org/cvs/wikini/actions/header.php?rev=1.12&content-type=text/x-cvsweb-markup&cvsroot=wikini | header.php]] :
-- [[http://www.wakkawiki.com/CharlesNepote | CharlesNepote]]
Deletions:
**[[http://www.wakkawiki.com/MatthiasAppel MatthiasAppel]]**
Hm, does anybody know how to make search robots not to index the ../edit ../revisions and ../referrers links on every page? You can search for WakkaWiki on [[http://www.google.de/search?q=+site:www.wakkawiki.com+wakkawiki&hl=de&lr=&ie=UTF-8&oe=UTF-8&start=50&sa=N Google]] and see the result.
--[[http://www.wakkawiki.com/MatthiasAppel MatthiasAppel]]
-- [[http://www.wakkawiki.com/CharlesNepote CharlesNepote]]
--[[http://www.wakkawiki.com/MatthiasAppel MatthiasAppel]]
''Did it work with robots.txt? And what does $ mean?? What is with referrers_sites?'' --[[http://www.wakkawiki.com/MoE MoE]]
~& Yahoo! recently [[http://www.ysearchblog.com/archives/000372.html announced wildcard support]], so the above mentioned solution works for Yahoo! Slurp as well. Up to now I haven't found a solution for msn. :( --DaC
On [[http://www.wikini.net/ WikiNi]] we code that in [[http://cvsweb.tuxfamily.org/cvs/wikini/actions/header.php?rev=1.12&content-type=text/x-cvsweb-markup&cvsroot=wikini header.php]] :
-- [[http://www.wakkawiki.com/CharlesNepote CharlesNepote]]
Revision [19287]
Edited on 2008-01-28 00:14:45 by JavaWoman [Modified links pointing to docs server]No Differences
Additions:
**[[http://www.wakkawiki.com/CharlesNepote CharlesNepote]]** @ [[http://www.wikini.net/ WikiNi]]
**[[http://www.wakkawiki.com/MatthiasAppel MatthiasAppel]]**
**[[http://www.wakkawiki.com/MatthiasAppel MatthiasAppel]]**
Deletions:
[[http://www.wakkawiki.com/MatthiasAppel MatthiasAppel]]
Additions:
[[http://www.wikini.net/ WikiNi]] ([[http://www.wakkawiki.com/CharlesNepote CharlesNepote]])
--[[http://www.wakkawiki.com/MatthiasAppel MatthiasAppel]]
-- [[http://www.wakkawiki.com/CharlesNepote CharlesNepote]]
--[[http://www.wakkawiki.com/MatthiasAppel MatthiasAppel]]
''Did it work with robots.txt? And what does $ mean?? What is with referrers_sites?'' --[[http://www.wakkawiki.com/MoE MoE]]
-- [[http://www.wakkawiki.com/CharlesNepote CharlesNepote]]
--[[http://www.wakkawiki.com/MatthiasAppel MatthiasAppel]]
-- [[http://www.wakkawiki.com/CharlesNepote CharlesNepote]]
--[[http://www.wakkawiki.com/MatthiasAppel MatthiasAppel]]
''Did it work with robots.txt? And what does $ mean?? What is with referrers_sites?'' --[[http://www.wakkawiki.com/MoE MoE]]
-- [[http://www.wakkawiki.com/CharlesNepote CharlesNepote]]
Deletions:
--MatthiasAppel
-- CharlesNepote
--MatthiasAppel
''Did it work with robots.txt? And what does $ mean?? What is with referrers_sites?'' --MoE
-- CharlesNepote
Additions:
WikiNi (CharlesNepote)
[[http://www.wakkawiki.com/MatthiasAppel MatthiasAppel]]
[[http://www.wakkawiki.com/MatthiasAppel MatthiasAppel]]
Deletions:
Additions:
~& There is a fix for this problem at RobotsDotTxt. - BaxilDragon
Additions:
~& Yahoo! recently [[http://www.ysearchblog.com/archives/000372.html announced wildcard support]], so the above mentioned solution works for Yahoo! Slurp as well. Up to now I haven't found a solution for msn. :( --DaC
Deletions:
Additions:
==== Wikka Mod 033 ====
Type: Bug Fix / Feature
----
===Credit:===
WikiNi (CharlesNepote) -- http://www.wakkawiki.com/MatthiasAppel
----
This is the method used in Wikka:
In header.php :
%%(php)
<?php if ($this->GetMethod() != 'show') echo "<meta name=\"robots\" content=\"noindex, nofollow\" />\n";?>
%%
----
History from WakkaWiki........
Hm, does anybody know how to make search robots not to index the ../edit ../revisions and ../referrers links on every page? You can search for WakkaWiki on [[http://www.google.de/search?q=+site:www.wakkawiki.com+wakkawiki&hl=de&lr=&ie=UTF-8&oe=UTF-8&start=50&sa=N Google]] and see the result.
--MatthiasAppel
Yes. See : [[http://www.robotstxt.org/wc/exclusion.html]].
-- CharlesNepote
Does that mean, that I have to add a robots.txt with this content:
%%
User-agent: Googlebot
Disallow: /edit$
Disallow: /revisions$
Disallow: /referrers$
User-agent: *
Disallow: /edit
Disallow: /revisions
Disallow: /referrers
%%
--MatthiasAppel
''Did it work with robots.txt? And what does $ mean?? What is with referrers_sites?'' --MoE
Dont really know, but I think $ doesnt belong there. But you could use something like this:
%%
User-agent: googlebot
Disallow: /*/edit
Disallow: /*/revisions
Disallow: /*/referrers
Disallow: /*/referrers_sites
%% Googlebot understands the wildcard *. Other SE Bots won't.
But also like Charles' method noted below. --JanPiotrowski
~& Yahoo! recently [[http://www.ysearchblog.com/archives/000372.html announced wildcard support]] as well so the above mentioned solution works for Yahoo! Slurp as well. Up to now I haven't found a solution for msn. :( --DaC
On [[http://www.wikini.net/ WikiNi]] we code that in [[http://cvsweb.tuxfamily.org/cvs/wikini/actions/header.php?rev=1.12&content-type=text/x-cvsweb-markup&cvsroot=wikini header.php]] :
%%(php)
<?php if ($this->GetMethod() != 'show') echo "<meta name=\"robots\" content=\"noindex, nofollow\" />\n";?>
%%
It seem to work quite well.
-- CharlesNepote
You can't get corrects contents of robots.txt if you install wiki at the root of your web, ie, if urls to HomePage is like my_domain.com/HomePage, because /robots.txt will be redirected to /wakka.php?wakka=robots.txt
Type: Bug Fix / Feature
----
===Credit:===
WikiNi (CharlesNepote) -- http://www.wakkawiki.com/MatthiasAppel
----
This is the method used in Wikka:
In header.php :
%%(php)
<?php if ($this->GetMethod() != 'show') echo "<meta name=\"robots\" content=\"noindex, nofollow\" />\n";?>
%%
----
History from WakkaWiki........
Hm, does anybody know how to make search robots not to index the ../edit ../revisions and ../referrers links on every page? You can search for WakkaWiki on [[http://www.google.de/search?q=+site:www.wakkawiki.com+wakkawiki&hl=de&lr=&ie=UTF-8&oe=UTF-8&start=50&sa=N Google]] and see the result.
--MatthiasAppel
Yes. See : [[http://www.robotstxt.org/wc/exclusion.html]].
-- CharlesNepote
Does that mean, that I have to add a robots.txt with this content:
%%
User-agent: Googlebot
Disallow: /edit$
Disallow: /revisions$
Disallow: /referrers$
User-agent: *
Disallow: /edit
Disallow: /revisions
Disallow: /referrers
%%
--MatthiasAppel
''Did it work with robots.txt? And what does $ mean?? What is with referrers_sites?'' --MoE
Dont really know, but I think $ doesnt belong there. But you could use something like this:
%%
User-agent: googlebot
Disallow: /*/edit
Disallow: /*/revisions
Disallow: /*/referrers
Disallow: /*/referrers_sites
%% Googlebot understands the wildcard *. Other SE Bots won't.
But also like Charles' method noted below. --JanPiotrowski
~& Yahoo! recently [[http://www.ysearchblog.com/archives/000372.html announced wildcard support]] as well so the above mentioned solution works for Yahoo! Slurp as well. Up to now I haven't found a solution for msn. :( --DaC
On [[http://www.wikini.net/ WikiNi]] we code that in [[http://cvsweb.tuxfamily.org/cvs/wikini/actions/header.php?rev=1.12&content-type=text/x-cvsweb-markup&cvsroot=wikini header.php]] :
%%(php)
<?php if ($this->GetMethod() != 'show') echo "<meta name=\"robots\" content=\"noindex, nofollow\" />\n";?>
%%
It seem to work quite well.
-- CharlesNepote
You can't get corrects contents of robots.txt if you install wiki at the root of your web, ie, if urls to HomePage is like my_domain.com/HomePage, because /robots.txt will be redirected to /wakka.php?wakka=robots.txt
Deletions:
Type: Bug Fix / Feature
----
===Credit:===
WikiNi (CharlesNepote) -- http://www.wakkawiki.com/MatthiasAppel
----
This is the method used in Wikka:
In header.php :
%%(php)
<?php if ($this->GetMethod() != 'show') echo "<meta name=\"robots\" content=\"noindex, nofollow\" />\n";?>
%%
----
History from WakkaWiki........
Hm, does anybody know how to make search robots not to index the ../edit ../revisions and ../referrers links on every page? You can search for WakkaWiki on [[http://www.google.de/search?q=+site:www.wakkawiki.com+wakkawiki&hl=de&lr=&ie=UTF-8&oe=UTF-8&start=50&sa=N Google]] and see the result.
--MatthiasAppel
Yes. See : [[http://www.robotstxt.org/wc/exclusion.html]].
-- CharlesNepote
Does that mean, that I have to add a robots.txt with this content:
%%
User-agent: Googlebot
Disallow: /edit$
Disallow: /revisions$
Disallow: /referrers$
User-agent: *
Disallow: /edit
Disallow: /revisions
Disallow: /referrers
%%
--MatthiasAppel
''Did it work with robots.txt? And what does $ mean?? What is with referrers_sites?'' --MoE
Dont really know, but I think $ doesnt belong there. But you could use something like this:
%%
User-agent: googlebot
Disallow: /*/edit
Disallow: /*/revisions
Disallow: /*/referrers
Disallow: /*/referrers_sites
%% Googlebot understands the wildcard *. Other SE Bots won't.
But also like Charles' method noted below. --JanPiotrowski
On [[http://www.wikini.net/ WikiNi]] we code that in [[http://cvsweb.tuxfamily.org/cvs/wikini/actions/header.php?rev=1.12&content-type=text/x-cvsweb-markup&cvsroot=wikini header.php]] :
%%(php)
<?php if ($this->GetMethod() != 'show') echo "<meta name=\"robots\" content=\"noindex, nofollow\" />\n";?>
%%
It seem to work quite well.
-- CharlesNepote
You can't get corrects contents of robots.txt if you install wiki at the root of your web, ie, if urls to HomePage is like my_domain.com/HomePage, because /robots.txt will be redirected to /wakka.php?wakka=robots.txt
Additions:
%%(php)
%%(php)
%%(php)
Additions:
==== Wikka Mod 033 ====
Type: Bug Fix / Feature
----
===Credit:===
WikiNi (CharlesNepote) -- http://www.wakkawiki.com/MatthiasAppel
----
This is the method used in Wikka:
In header.php :
<?php if ($this->GetMethod() != 'show') echo "<meta name=\"robots\" content=\"noindex, nofollow\" />\n";?>
----
History from WakkaWiki........
Hm, does anybody know how to make search robots not to index the ../edit ../revisions and ../referrers links on every page? You can search for WakkaWiki on [[http://www.google.de/search?q=+site:www.wakkawiki.com+wakkawiki&hl=de&lr=&ie=UTF-8&oe=UTF-8&start=50&sa=N Google]] and see the result.
--MatthiasAppel
Yes. See : [[http://www.robotstxt.org/wc/exclusion.html]].
-- CharlesNepote
Does that mean, that I have to add a robots.txt with this content:
%%
User-agent: Googlebot
Disallow: /edit$
Disallow: /revisions$
Disallow: /referrers$
User-agent: *
Disallow: /edit
Disallow: /revisions
Disallow: /referrers
%%
--MatthiasAppel
''Did it work with robots.txt? And what does $ mean?? What is with referrers_sites?'' --MoE
Dont really know, but I think $ doesnt belong there. But you could use something like this:
%%
User-agent: googlebot
Disallow: /*/edit
Disallow: /*/revisions
Disallow: /*/referrers
Disallow: /*/referrers_sites
%% Googlebot understands the wildcard *. Other SE Bots won't.
But also like Charles' method noted below. --JanPiotrowski
On [[http://www.wikini.net/ WikiNi]] we code that in [[http://cvsweb.tuxfamily.org/cvs/wikini/actions/header.php?rev=1.12&content-type=text/x-cvsweb-markup&cvsroot=wikini header.php]] :
<?php if ($this->GetMethod() != 'show') echo "<meta name=\"robots\" content=\"noindex, nofollow\" />\n";?>
It seem to work quite well.
-- CharlesNepote
You can't get corrects contents of robots.txt if you install wiki at the root of your web, ie, if urls to HomePage is like my_domain.com/HomePage, because /robots.txt will be redirected to /wakka.php?wakka=robots.txt
Type: Bug Fix / Feature
----
===Credit:===
WikiNi (CharlesNepote) -- http://www.wakkawiki.com/MatthiasAppel
----
This is the method used in Wikka:
In header.php :
<?php if ($this->GetMethod() != 'show') echo "<meta name=\"robots\" content=\"noindex, nofollow\" />\n";?>
----
History from WakkaWiki........
Hm, does anybody know how to make search robots not to index the ../edit ../revisions and ../referrers links on every page? You can search for WakkaWiki on [[http://www.google.de/search?q=+site:www.wakkawiki.com+wakkawiki&hl=de&lr=&ie=UTF-8&oe=UTF-8&start=50&sa=N Google]] and see the result.
--MatthiasAppel
Yes. See : [[http://www.robotstxt.org/wc/exclusion.html]].
-- CharlesNepote
Does that mean, that I have to add a robots.txt with this content:
%%
User-agent: Googlebot
Disallow: /edit$
Disallow: /revisions$
Disallow: /referrers$
User-agent: *
Disallow: /edit
Disallow: /revisions
Disallow: /referrers
%%
--MatthiasAppel
''Did it work with robots.txt? And what does $ mean?? What is with referrers_sites?'' --MoE
Dont really know, but I think $ doesnt belong there. But you could use something like this:
%%
User-agent: googlebot
Disallow: /*/edit
Disallow: /*/revisions
Disallow: /*/referrers
Disallow: /*/referrers_sites
%% Googlebot understands the wildcard *. Other SE Bots won't.
But also like Charles' method noted below. --JanPiotrowski
On [[http://www.wikini.net/ WikiNi]] we code that in [[http://cvsweb.tuxfamily.org/cvs/wikini/actions/header.php?rev=1.12&content-type=text/x-cvsweb-markup&cvsroot=wikini header.php]] :
<?php if ($this->GetMethod() != 'show') echo "<meta name=\"robots\" content=\"noindex, nofollow\" />\n";?>
It seem to work quite well.
-- CharlesNepote
You can't get corrects contents of robots.txt if you install wiki at the root of your web, ie, if urls to HomePage is like my_domain.com/HomePage, because /robots.txt will be redirected to /wakka.php?wakka=robots.txt
Deletions:
Type: Bug Fix / Feature
----
===Credit:===
""CharlesNepote"" -- http://www.wakkawiki.com/MatthiasAppel
----
This is the method used in Wikka:
In header.php :
<?php if ($this->GetMethod() != 'show') echo "<meta name=\"robots\" content=\"noindex, nofollow\" />\n";?>
----
History from WakkaWiki........
Hm, does anybody know how to make search robots not to index the ../edit ../revisions and ../referrers links on every page? You can search for WakkaWiki on [[http://www.google.de/search?q=+site:www.wakkawiki.com+wakkawiki&hl=de&lr=&ie=UTF-8&oe=UTF-8&start=50&sa=N Google]] and see the result.
--MatthiasAppel
Yes. See : [[http://www.robotstxt.org/wc/exclusion.html]].
-- CharlesNepote
Does that mean, that I have to add a robots.txt with this content:
%%
User-agent: Googlebot
Disallow: /edit$
Disallow: /revisions$
Disallow: /referrers$
User-agent: *
Disallow: /edit
Disallow: /revisions
Disallow: /referrers
%%
--MatthiasAppel
''Did it work with robots.txt? And what does $ mean?? What is with referrers_sites?'' --MoE
Dont really know, but I think $ doesnt belong there. But you could use something like this:
%%
User-agent: googlebot
Disallow: /*/edit
Disallow: /*/revisions
Disallow: /*/referrers
Disallow: /*/referrers_sites
%% Googlebot understands the wildcard *. Other SE Bots won't.
But also like Charles' method noted below. --JanPiotrowski
On [[http://www.wikini.net/ WikiNi]] we code that in [[http://cvsweb.tuxfamily.org/cvs/wikini/actions/header.php?rev=1.12&content-type=text/x-cvsweb-markup&cvsroot=wikini header.php]] :
<?php if ($this->GetMethod() != 'show') echo "<meta name=\"robots\" content=\"noindex, nofollow\" />\n";?>
It seem to work quite well.
-- CharlesNepote
You can't get corrects contents of robots.txt if you install wiki at the root of your web, ie, if urls to HomePage is like my_domain.com/HomePage, because /robots.txt will be redirected to /wakka.php?wakka=robots.txt
Additions:
You can't get corrects contents of robots.txt if you install wiki at the root of your web, ie, if urls to HomePage is like my_domain.com/HomePage, because /robots.txt will be redirected to /wakka.php?wakka=robots.txt
See RobotsDotTxt and FaviconDotIco. --DotMG
See RobotsDotTxt and FaviconDotIco. --DotMG
Additions:
This is the method used in Wikka:
In header.php :
History from WakkaWiki........
In header.php :
History from WakkaWiki........
Additions:
Dont really know, but I think $ doesnt belong there. But you could use something like this:
%% Googlebot understands the wildcard *. Other SE Bots won't.
But also like Charles' method noted below. --JanPiotrowski
%% Googlebot understands the wildcard *. Other SE Bots won't.
But also like Charles' method noted below. --JanPiotrowski
Deletions:
%% Google understand the wildcard "*". --JanPiotrowski
Additions:
Dont know, but $ doesnt belong there. But you could use something like this:
User-agent: googlebot
Disallow: /*/edit
Disallow: /*/revisions
Disallow: /*/referrers
Disallow: /*/referrers_sites
%% Google understand the wildcard "*". --JanPiotrowski