Greenguy's Board

Greenguy's Board (http://www.greenguysboard.com/board/index.php)
-   Possible Cheaters (http://www.greenguysboard.com/board/forumdisplay.php?f=19)
-   -   robots txt to prevent crawling of freesites (http://www.greenguysboard.com/board/showthread.php?t=34504)

Jel 2006-09-16 02:28 AM

robots txt to prevent crawling of freesites
 
OK,

I have a guy submitting that on his root robots.txt has the following:

User-agent: *
Disallow: /gall/
Disallow: /gall1/
Disallow: /gall2/
Disallow: /gall3/
Disallow: /gall4/
Disallow: /gall5/
Disallow: /gall6/
Disallow: /gall7/
Disallow: /gall8/
Disallow: /gall9/
Disallow: /gall10/
Disallow: /gall11/
Disallow: /gall12/
Disallow: /gall13/
Disallow: /cgi-bin/
Disallow: /img/

domain: soccerwank.com (also sexcarrot.com with different directory names his freesites are in)

On the freesites themselves, in the head, is the meta:
meta name="robots" content="index, follow"

I was running a link checker and was getting flags with this message:
"The link was not checked due to robots exclusion rules. Check the link manually."
Hence me looking at the root robots file.

Seems very fishy to me, and this is titled 'possible cheaters' but I can't fathom whether this is an honest mistake, as obviously his freesites aren't going to get pickjed up by the SES, or just a way to glean traffic from LLs.

Carrie 2006-09-16 02:32 AM

Also looks like a way to turn recip links (A->B->A) into one-way links from the LLs to his domains. (One-way links being more valuable.)

ponygirl 2006-09-16 03:02 AM

Quote:

Originally Posted by Jel (Post 300037)
On the freesites themselves, in the head, is the meta:
meta name="robots" content="index, follow"

that is the thing that bugs me...why have that in the meta unless he forgot to take it out from a cut & paste off something else. Kinda tricky imo.

he's a member of the board, maybe we'll hear something.

Preacher 2006-09-16 12:29 PM

This isn't the only forum member doing this. Like Jel I'm not sure it's exactly cheating so I would like to hear more opinions on this.

ponygirl 2006-09-16 12:39 PM

Quote:

Originally Posted by Preacher (Post 300104)
This isn't the only forum member doing this. Like Jel I'm not sure it's exactly cheating so I would like to hear more opinions on this.

yes me too. I wouldn't consider it cheating, really - it's not breaking any rules, but it's misleading.

virgohippy 2006-09-18 03:18 AM

Quote:

Originally Posted by ponygirl (Post 300106)
yes me too. I wouldn't consider it cheating, really - it's not breaking any rules, but it's misleading.

If someone's going to take the extra time and energy to stuff their robots.txt file in an attempt to get more back then they're giving I would consider that following the rules, but still acting with malicious intent.

Malicious intent falls under the unspoken rule of, "I don't like your business practices, therefore, I don't want to do business with you."

In this situation, I probably wouldn't send a rejection email, or even ask what's up. I'd just silently make their sites dissappear with a quick click of the delete button.

Jel, thanks for bringing this issue up. As if I didn't already have enough to check for... |loony|

Carrie 2006-09-20 06:34 AM

Methinks folks don't look in this section often enough.
I think I'll send TT a note and tell him to look here; I like him and it surprises me that this kind of thing would be done on purpose. Maybe he's got a good explanation for it.
Either way, it certainly isn't in anyone's LL rules that it can't be done, so...? Weird situation.

picXX 2006-09-20 07:28 AM

Hi.

I am the owner of both soccerwank.com , sexcarrot.com and pornogata.com wich have the same robots.txt files and all the domains.

When i first started out building galleries and freesites, i was told to create a robots.txt file like that to prevent google from crawling thousands of duplicate galleryfiles and hundred of duplicated freesitefiles.

I am linking to the freesites on my mainsite, but offcourse that would not benefit all the LL i am submitting to.

I willl change the robots.txt files on all my domain asap, and prevent from crawling only the galleryfolders instead of the freesite directory folders.

I am sorry i have braught up the issue, cuz i was really not aware of it, i just followed some friends good advice.

picXX 2006-09-20 07:31 AM

And please dont see my as a possible cheater.. i have never even once tried to cheat fellow webmasters with intention.

picXX 2006-09-20 07:36 AM

I have now removed the disallow to all directories at these 3 domains.

However, i do know most galleribuilders that have a HUB site are doing the same, but that is not me, so just wanted to tell that my directories are open for spiders now.
:)

ponygirl 2006-09-20 07:40 AM

Quote:

Originally Posted by picXX (Post 300858)
Hi.

I am the owner of both soccerwank.com , sexcarrot.com and pornogata.com wich have the same robots.txt files and all the domains.

When i first started out building galleries and freesites, i was told to create a robots.txt file like that to prevent google from crawling thousands of duplicate galleryfiles and hundred of duplicated freesitefiles.

I am linking to the freesites on my mainsite, but offcourse that would not benefit all the LL i am submitting to.

I willl change the robots.txt files on all my domain asap, and prevent from crawling only the galleryfolders instead of the freesite directory folders.

I am sorry i have braught up the issue, cuz i was really not aware of it, i just followed some friends good advice.

I rather thought it was something like that.

Quote:

Originally Posted by picXX (Post 300862)
And please dont see my as a possible cheater.. i have never even once tried to cheat fellow webmasters with intention.

That's why I wanted to see what this was all about...I've never seen anything in your sites suggesting you cheat in any way, and I don't think you are a cheater either picXX. Thanks for coming by and clearing it up |thumb

picXX 2006-09-20 07:46 AM

Quote:

I don't think you are a cheater either picXX
Thank you... it gets my neckhair to raise and my body to freeze when i come over threads like this, even if this is the first thread i am accused for cheating, it is not wery comfortable to read that someone are considering you a cheat when all you do is to try to follow all the rules/tips/advice from experienced webmasters.

Thanks to Carrie allso, that braught this thread to my attention.

Virgohippy:
Quote:

In this situation, I probably wouldn't send a rejection email, or even ask what's up. I'd just silently make their sites dissappear with a quick click of the delete button.
In you're case, i would never knew why i never got a listing, and therefore never could have fixed it... well, that's not wery Kosher either?

Mr. Blue 2006-09-20 08:00 AM

Unfortunately picxx you were a victim of bad advice. I'd suggest you no longer listen to the person that gave you this advice in the first place :)

There's a lot of bad information tossed around on boards, etc, and you really have to be careful on who you listen to. This board is a good place to post questions about LL as you won't get steered wrong as you'll be getting the info straight from the horses (owners) mouth.

virgohippy 2006-09-20 03:27 PM

Quote:

Originally Posted by picXX (Post 300878)
In you're case, i would never knew why i never got a listing, and therefore never could have fixed it... well, that's not wery Kosher either?

Yeah... you're probably right.

With my next reincarnation, I'll make it a point to include a "if you don't see your site listed within a couple weeks, and you don't recieve a declined reason, contact me here..." :)

But I still refuse to get my hands dirty! |loony|

Jel 2006-09-21 03:08 AM

Quote:

Originally Posted by Mr. Blue (Post 300887)
Unfortunately picxx you were a victim of bad advice. I'd suggest you no longer listen to the person that gave you this advice in the first place :)

There's a lot of bad information tossed around on boards, etc, and you really have to be careful on who you listen to. This board is a good place to post questions about LL as you won't get steered wrong as you'll be getting the info straight from the horses (owners) mouth.

Spot on Mr. Blue. picXX thanks for clearing that up, and removing that robots file :)

Halfdeck 2006-09-22 11:37 AM

I know LL owners like to get all the backlinks they can get their hands on (so would I), but you're forgetting one thing:

Duplicate content.

Getting linkbacks from supplemental pages is not going to do anyone any good.

EDIT:

Not to mention low quality backlinks from free sites aren't going to make or break your ranking on Google (though MSN probably eats them up). In a few years, who knows, Google may ignore them altogether.

One way to look at recips is advertising your LL via increasing brand awareness. Approach them as means of inflating your SE position -- and you're in violation of Google guidelines.

Preventing duplicate content is a legitimate reason for disallowing mirrors, however. A large percentage of supps under a domain may negatively impact the entire domain.

ponygirl 2006-09-22 12:11 PM

Quote:

Originally Posted by Halfdeck (Post 301696)
I know LL owners like to get all the backlinks they can get their hands on (so would I), but you're forgetting one thing:

Duplicate content.

Getting linkbacks from supplemental pages is not going to do anyone any good.

EDIT:

Not to mention low quality backlinks from free sites aren't going to make or break your ranking on Google (though MSN probably eats them up). In a few years, who knows, Google may ignore them altogether.

One way to look at recips is advertising your LL via increasing brand awareness. Approach them as means of inflating your SE position -- and you're in violation of Google guidelines.

Preventing duplicate content is a legitimate reason for disallowing mirrors, however. A large percentage of supps under a domain may negatively impact the entire domain.

I admit to knowing next to nothing about SEO, but I have heard lots about dup content...so I'll just say that I don't care if there is a nofollow for that reason. I'm not going to decline submits for that if someone really thinks it helps them with SEs - maybe it does what do I know :D

BUT make sure your meta tags don't say anything different, that's all.

now I know this was an honest mistake but it's a good opportunity for everyone to make sure they check the little details :)

Mr. Blue 2006-09-22 04:02 PM

Quote:

Originally Posted by Halfdeck (Post 301696)
I know LL owners like to get all the backlinks they can get their hands on (so would I), but you're forgetting one thing:

Duplicate content.

Getting linkbacks from supplemental pages is not going to do anyone any good.

EDIT:

Not to mention low quality backlinks from free sites aren't going to make or break your ranking on Google (though MSN probably eats them up). In a few years, who knows, Google may ignore them altogether.

One way to look at recips is advertising your LL via increasing brand awareness. Approach them as means of inflating your SE position -- and you're in violation of Google guidelines.

Preventing duplicate content is a legitimate reason for disallowing mirrors, however. A large percentage of supps under a domain may negatively impact the entire domain.

This is one of the reasons when I was submitting freesites regularly I'd only submit to around 12 so there wouldn't be duplicate page penalties. If however I wanted to submit to more than 12, I'd just change the page content enough to avoid the duplicate page penalty...you can keep the template the same, change the sales text, alt tags, title tags, etc, etc, etc. enough where you wouldn't get penalized for a duplicate page penalty. A little extra work, but feasibly 1 freesite can be optimized for a whole bunch of keywords depending on how many LL you're submitting to and how much you're changing the page.

Preacher 2006-09-22 05:56 PM

Quote:

Originally Posted by Jel (Post 300037)
...I was running a link checker and was getting flags with this message:
"The link was not checked due to robots exclusion rules. Check the link manually."
Hence me looking at the root robots file.

This is actually more of the point that I was interested in. There is a submitter out there who is disallowing any bot that hits his free-sites, I believe through htaccess.

Therefore whenever my bot ran across his domains all of his free-sites would get flagged and pulled as unavailable and I'd have to re-add them manually. This was incredibly annoying.

I actually did this for a while but finally got tired of it and just left his sites in an error/delisted status. He no longer submits to me, but I still see his name pop-up from time to time.

virgohippy 2006-09-22 08:17 PM

Quote:

Originally Posted by Halfdeck (Post 301696)
Preventing duplicate content is a legitimate reason for disallowing mirrors, however. A large percentage of supps under a domain may negatively impact the entire domain.

From an SEO POV for a LL, which would be worse:

Linking to a number of freesites on a domain with tons of dissallowed pages?

or

Linking to a number of freesites which may or may not be flagged for spam?

Seems to me most submitters aren't able to produce and submit more than a small handful of mirrors anyway. |huh

Quote:

Originally Posted by Mr. Blue (Post 301749)
A little extra work, but feasibly 1 freesite can be optimized for a whole bunch of keywords depending on how many LL you're submitting to and how much you're changing the page.

As a submitter, I second your advice Mr. Blue. |thumb

Quote:

Originally Posted by Preacher (Post 301766)
I actually did this for a while but finally got tired of it and just left his sites in an error/delisted status...

Not only that, but the occasional bot check of existing listings must surely make regular check ups yet even more fun. |sad|

Halfdeck 2006-09-22 08:28 PM

Quote:

This is actually more of the point that I was interested in.
I hear you Preacher. However, META NOINDEX should solve that problem. It's also a more reliable alternative to robots.txt for keeping URLs out of Google's index.

Quote:

Originally Posted by Mr. Blue (Post 301749)
A little extra work, but feasibly 1 freesite can be optimized for a whole bunch of keywords depending on how many LL you're submitting to and how much you're changing the page.

Mr. Blue, that's as realistic a statement as saying it only takes a little extra work for LL owners to send hand-written rejection emails.

Quote:

Unfortunately picxx you were a victim of bad advice. I'd suggest you no longer listen to the person that gave you this advice in the first place
http://www.greenguysboard.com/board/...ghlight=robots

In that thread, I wrote:

Quote:

1000 incoming links are worthless if the links point to a page Google considers spam. If Google chooses an index.html with small LL linking to it over an index.html with links from penisbot, LOR, etc, then you're shit out of luck...I wouldn't leave it up to Google to decide that for me.

To customize your free sites to make them appear unique to SEs seems to be a waste of time to me.

One suggestion I saw to use robots.txt to prevent Google from indexing the duplicate content...so you'd get nice SE traffic off your primary index.html and then pull nice LL traffic off the duplicate doorways.

Personally, I'd build 1 site and submit to 20. The 20 LL will give you say 30,000 uniques over a few months and then 5,000 or whatever from SE a month.

If you spent time submitting to the other 80, lets be optimistic and say you'd get 5,000 uniques from them a month. Still, if you had to choose between 5k from google or 5k from LLs...which would you choose?
Linkster wrote:

Quote:

First - Im glad to see people thinking along these lines - unfortunately you are all giving Google a heck of alot more credit that its due.
If the pages are all on one domain - yeah Google will pick one and call the rest dupes - if you dont change the pages - if you change them AND - change the inside pages for each copy of the site - then you might have a chance of getting more than one copy in Google.

as far as the robots text thingie - you are submitting free sites to a LL - if the LL bans you because of this then they are not a LL - they are trying to build a SE hub - granted we would like to get some SE benefit but it sure isnt the main reason to run a LL - or make decisions on listing someone - there are definitely much better ways for a LL to get the phrases they want in a SE - which is why after testing niche recips we went back to the single recip.

I will repeat what I have said in other threads about this - the sites you submit to LLs should NOT be the copies you are trying to get in the SEs - a site that specifically is built for the SEs will do much better and can have a few more aggressive ways of getting surfers to buy - of course thats just my way of doing things but it seems that it has worked well for a bunch of us over the years

virgohippy 2006-09-22 08:48 PM

Quote:

Originally Posted by Halfdeck (Post 301782)

Interesting thread Halfdeck... now my head is spinning. |loony|

Halfdeck 2006-09-22 08:59 PM

Quote:

Seems to me most submitters aren't able to produce and submit more than a small handful of mirrors anyway.
It only takes two to tango.

Mirror sites and robots.txt disallow both lead to your LL likely getting no link juice from recips whatsoever. If you want decent backlinks, you might think about accepting only unique free sites. Even then, if you're linking to each other, chances are the link is completely ignored by Google.

virgohippy 2006-09-22 10:04 PM

Quote:

Originally Posted by Halfdeck (Post 301786)
Mirror sites and robots.txt disallow both lead to your LL likely getting no link juice from recips whatsoever. If you want decent backlinks, you might think about accepting only unique free sites. Even then, if you're linking to each other, chances are the link is completely ignored by Google.

Seems to me there's a bit of give somewhere in that statement. If that were true my LL should be worthless to google - right now it's only a notch or two above worthless. |loony|

But I see your point. In my own experiments I've noticed that backlinks from unique pages with backlinks from other unique pages gives a much higher return than backlinks from non-unique pages... well, from google at least. ;)

Mr. Blue 2006-09-23 02:56 AM

Quote:

Originally Posted by Halfdeck (Post 301782)
Mr. Blue, that's as realistic a statement as saying it only takes a little extra work for LL owners to send hand-written rejection emails.

That's not entirely true as if you're working on a template system it takes only a little bit more time to edit the pages for multiple LLs...hell, it doesn't really take much time to change the meta tags, sales text, and alt tags. If you run between 12 recips per freesite...if you're submitting to roughly 36 LLs...that's only 3 freesites you have to edit and optimize. It's not like submitted to tgps where you may have 2000+ pages to edit.

Now, I personally always submitted to a small LL grouping to remove the duplicate page penalty, but if there were more LLs that interested me, I would now do as I stated above.

As for whether it's good advice or not to use the robot.txt...if this is setting off people's scripts, if it's too much hassle for them to review your freesites, reviewers will do exactly do as preacher and jel did...they'll most likely not even bother reviewing or listing your sites.

So, there's a certain futility in following the robot.txt advice if it prevents you from easily getting listed at the LLs you're submitting to. As a submitter you have the options of either submitting to very few LL's (that was my choice when I was regularly submitting), Change the pages enough to avoid the duplicate page penalty (not hard to do when you're working with a template system), or use the robot.txt and not get listed on a number of sites you're submitting to.

Out of those options the most sane and easy one is to submit to a very small high quality group of linklists that you know will list you regularly.

You posted something Linkster said and maybe he can stop in and help me out on this point as I will gladly defer to his expertise in this area because I know he knows wayyyy more about this topic than I ever will. In that post you quoted, Linkster says that linksforsex went to a single recip, but a few months ago he went back to the category specific recips and that's still the case today. Now, why would a Link List switch back to category recips? As far as I can tell category recips have mainly been put in place for SEO? I'm not sure what benefit category specific recips would have other then SEO.

If LLs are using category specific recips, it seems they're doing so for SEO, if that's the case then it would behoove all involved to make google as happy as they can and take effort to remove the duplicate page penalty, but at the same time not negate the category recips that so many LL's use by using a robot.txt to block the engine from searching those pages. Essentially you are breaking a Link Lists rules because you're completely negating any benefit the Link List owner was trying to get by having category specific recips.


All times are GMT -4. The time now is 08:49 PM.

Powered by vBulletin® Version 3.8.1
Copyright ©2000 - 2025, Jelsoft Enterprises Ltd.
© Greenguy Marketing Inc