Wiki source for Mod033bRobotIndexing


Show raw source

==== Wikka Mod 033 ====
Type: Bug Fix / Feature
----
===Credit:===
**[[http://web.archive.org/web/20040811012539/http://www.wakkawiki.com/CharlesNepote | CharlesNepote]]** @ [[http://www.wikini.net/ | WikiNi]]
**[[http://web.archive.org/web/20040810215014/http://www.wakkawiki.com/MatthiasAppel | MatthiasAppel]]**
----

This is the method used in Wikka:

In header.php :
%%(php)
<?php if ($this->GetMethod() != 'show') echo "<meta name=\"robots\" content=\"noindex, nofollow\" />\n";?>
%%

----

History from WakkaWiki........


Hm, does anybody know how to make search robots not to index the ../edit ../revisions and ../referrers links on every page? You can search for WakkaWiki on [[http://www.google.de/search?q=+site:www.wakkawiki.com+wakkawiki&hl=de&lr=&ie=UTF-8&oe=UTF-8&start=50&sa=N | Google]] and see the result.
--[[http://web.archive.org/web/20040810215014/http://www.wakkawiki.com/MatthiasAppel | MatthiasAppel]]

Yes. See : [[http://www.robotstxt.org/wc/exclusion.html]].
-- [[http://web.archive.org/web/20040811012539/http://www.wakkawiki.com/CharlesNepote | CharlesNepote]]

Does that mean, that I have to add a robots.txt with this content:
%%
User-agent: Googlebot
Disallow: /edit$
Disallow: /revisions$
Disallow: /referrers$

User-agent: *
Disallow: /edit
Disallow: /revisions
Disallow: /referrers
%%
--[[http://web.archive.org/web/20040810215014/http://www.wakkawiki.com/MatthiasAppel | MatthiasAppel]]
''Did it work with robots.txt? And what does $ mean?? What is with referrers_sites?'' --[[http://web.archive.org/web/20040619185958/http://www.wakkawiki.com/MoE | MoE]]

Dont really know, but I think $ doesnt belong there. But you could use something like this:
%%
User-agent: googlebot
Disallow: /*/edit
Disallow: /*/revisions
Disallow: /*/referrers
Disallow: /*/referrers_sites
%% Googlebot understands the wildcard *. Other SE Bots won't.
But also like Charles' method noted below. --JanPiotrowski

~& Yahoo! recently [[http://www.ysearchblog.com/archives/000372.html | announced wildcard support]], so the above mentioned solution works for Yahoo! Slurp as well. Up to now I haven't found a solution for msn. :( --DaC

On [[http://www.wikini.net/ | WikiNi]] we code that in [[http://cvsweb.tuxfamily.org/cvs/wikini/actions/header.php?rev=1.12&content-type=text/x-cvsweb-markup&cvsroot=wikini | header.php]] :
%%(php)
<?php if ($this->GetMethod() != 'show') echo "<meta name=\"robots\" content=\"noindex, nofollow\" />\n";?>
%%
It seem to work quite well.
-- [[http://web.archive.org/web/20040811012539/http://www.wakkawiki.com/CharlesNepote | CharlesNepote]]

You can't get corrects contents of robots.txt if you install wiki at the root of your web, ie, if urls to HomePage is like my_domain.com/HomePage, because /robots.txt will be redirected to /wakka.php?wakka=robots.txt
~& There is a fix for this problem at RobotsDotTxt. - BaxilDragon
See RobotsDotTxt and FaviconDotIco. --DotMG
Valid XHTML :: Valid CSS: :: Powered by WikkaWiki