I have 3 domains
these 2
http://www.janhendrikx.be
http://www.standenbouw-jan.be
redirect to
http://www.ontwerpbureaujan.be/
In google webmasters I added ontwerpbureaujan
Do I have to add the others too? Or do I get duplicate content
Do I have to use canonical URLs? How?
'Standenbouw' is my main seo keyword, maybe I should add http://www.standenbouw-jan.be to webmasters, not ontwerpbureaujan ? ...
I'm sure janhendrikx.be & standenbouw-jan.be have 301 redirects to ontwerpbureaujan.be
If not, please get it done by your Web host else Google will actually consider them as three different websites and will crawl them all.
If they have 301 redirects, then you could add the others too and Google would know that they are the same site and hence wouldn't crawl all of them.
But you could still add all of them and it may increase your pagerank for the keyword and in that case you could mention http://www.ontwerpbureaujan.be/ as your preferred domain on the site settings area.
Here are a few recommendations by Google itself:
https://support.google.com/webmasters/answer/139066?hl=en#4
Good luck!
Related
I am just a learn in the field of SEO and i have a main domain and an addon domains. Both have separate websites. Consider main.com is my main domain and addon.com is my addon domain name which is pointed to a sub directory called "addon".
I can access addon.com by using the following 3 ways.
addon.com
main.com/addon
addon.main.com
Are these urls are indexed separately by search engines? If so how can i prevent this?
Does Search engine think main.com/addon as a page in the main.com?
I am not sure i need to worry about all these things or just leave it as it is. I searched to google but couldn't find a right answer.
It may be too late to answer. However, it may benefit others.
Primarydomain and subdomain or addon-domain will not be linked by the search engines automatically, unless you link them purposefully or inadvertently. Except all conditions are true:
Your web root normally public_html has no index page
Directory indexing of your web root is opened, eventually
exposing/linking your sub-folder -which is attached to your
addon-domain- to google and entire world.
In that scenario robots.txt solution is not recommended, because search engines may ignore robot.txt rules.
Reference
Google will only index pages if they are linked to or listed in the sitemap. You can stop the addon.main.com or main.com/addon being indexed by using noindex tags:
<META NAME="ROBOTS" CONTENT="NOINDEX, NOFOLLOW">
or disallowing it in the robots.txt
The search engine will consider main.com/addon as a page of main.com - if sites are completely separate i'd recommend using a separate domain (preferably a keyword rich domain) but it's up to you really
We have three domain names with the same content. For the three domains, it will return a 200 OK HTTP code. It will look like duplicates of the same content. If there is a canonical tag on every page it will be better.
The best would be to create a redirection on the subdomain panel in cpanel so that at least addon.main.com would redirect to addon.com
Then, you can add a robots.txt to the root path of the primary domain and add
user-agent:*
disallow:/
so that no robot will visit main.com/addon
Google gives less weight to subdomain hosted site of another domain.
Superbad for SEO
If you are hosting for SEO and love the convenience of cPanel, then forget hosting domains as addon domains.
#Vasanthan R.P.
Its an excellent question, often overlooked by SEO professionals. +1 for you
When I searching our web site on Google I found three sites with the same content show up. I always thought we were using only one site www.foo.com, but it turn out we have www.foo.net and www.foo.info with the same content as www.foo.com.
I know it is extremely bad to have the same content under different URL. And it seems we have being using three domains for years and I have not seen punitive blunt so far. What is going on? Is Google using new policy like this blog advocate?http://www.seodenver.com/duplicate-content-over-multiple-domains-seo-issues/ Or is it OK using DNS redirect? What should I do? Thanks
If you are managing the websites via Google Webmaster Tools, it is possible to specify the "primary domain".
However, the world of search engines doesn't stop with Google, so your best bet is to send a 301 redirect to your primary domain. For example.
www.foo.net should 301 redirect to www.foo.com
www.foo.net/bar should 301 redirect to www.foo.com/bar
and so on.
This will ensure that www.foo.com gets the entire score, rather than (potentially) a third of the score that you might get for link-backs (internal and external).
Look into canonical links, as documented by Google.
If your site has identical or vastly
similar content that's accessible
through multiple URLs, this format
provides you with more control over
the URL returned in search results. It
also helps to make sure that
properties such as link popularity are
consolidated to your preferred
version.
They explicitly state it will work cross-domain.
We have #1 and #2 spots we would like to keep, but because of the way things were jumbled we have to migrate to a new domain.
We do not want the new domain to be penalized for duplicate content, we want it to naturally take the spot on Google.
How do we tell Google our page has moved?
301 redirects are the recommended way to do this according to Google themselves. I tend to perform 301's using a .htaccess file (a few different methods here) but it can also be done using PHP like this:
header('Location: '.$newlocation, true, 301)
What seengee user has answered is perfect.
Have a look to Google Webmaster Tools too; there's a specific option for Address change.
Change of address
If you're planning to move your site to a new domain, use the Change of Address tool to tell Google about your new URL. This will help us update our index faster and smooth the transition for your users.
For best results, follow these steps:
Set up the new site
Review our guidelines for moving your site to a new domain. Set up your content on your new domain, then make sure all internal links point to the new domain.
Redirect all traffic from the old site
Use a 301 redirect to permanently redirect the pages on your old site to your new site. This tells users and search engines that your site has permanently moved. Ask webmasters to update their links to point to your new domain and make sure incoming links to your old site are redirected correctly using the 301 redirects.
Add your new site to Webmaster Tools
Make sure you have added and verified your new domain.
Tell us the URL of your new domain
Try this: https://www.google.com/webmasters/tools add both domains there and then, go Change of address.
Http 301 Request is used for Permanent Redirect the old website's content,URL and Links to New Website URL. Google Never Penalize the New Website for Duplicate content.
what about "Keyword" you use in meta for performing in SERP result. After redirection how you manage them.
position before Redirection , after redirection have same OR not.
AS SEO purpose don't make Redirection withour proper plan. First make proper optimization of "B" page before "A" redirect ( 301, 302 ) "B" page.
Dont make 302 Redirect , its spammy after 3 months, why within 20 Days SERP catche show "B" page instead of "A" page.
in the end : Before Redirection - Optimize the Destination URl - for Benefit SEOand business
Have a Nice Day ... Thanks
Google itself detect your web page if you implemented the 301, 302 or 404 redirection. But if you want to tell google that you have moved your page then:
Go to google webmaster tool
In Crawl section there is an option of Fetch as Google.
Just enter your url and check the fetch status, you will find that google will index your new url.
in google webmaster tools click on change of site address.
firstly you need to put content on new website where you want to redirect the website.
redirect usingold website by using HTTP 301 code and after verify in google webmaster tools.
Do 301 redirects. This will tell the search engines that the pages have moved and where they are now. This also associates the old URL with the new URL for Google which means all of your old incoming links will now be redirects to your new pages. Also use Google Webmaster Tools to submit XML sitemap.
In both Google and Bing Webmaster tools, you can notify a change of address. I will warn you, however, that when changing domains, you're not going to keep your current rankings.
Submit a change of address notice in Webmaster tools
Do 1:1 page redirecting (redirect all other pages that can't be mapped to a similar page to your homepage)
Submit an up-to-date sitemap
This should help:
https://support.google.com/webmasters/answer/83106?hl=en
http://moz.com/blog/achieving-an-seo-friendly-domain-migration-the-infographic
A friend of mine had a report done on an old site I built. One of the issues mentioned in the report was the following:
Your website needs a www resolve.
Currently you can go to
http://urmarialarts.com or
http://www.urmartialarts.com which
means in Google's eyes you have two
websites with the same content. You
should set the hosting up so you are
redirected
Is this the case? And if so do I need to setup a 301 redirect on one of the domains to point to the other?
Yes and yes, for further information: http://www.google.com/support/webmasters/bin/answer.py?hl=en&answer=44231
You should entirely avoid this problem by using a canonical link tag. http://googlewebmastercentral.blogspot.com/2009/02/specify-your-canonical.html
A few days ago we replaced our web site with an updated version. The original site's content was migrated to http://backup.example.com. Search engines do not know about the old site, and I do not want them to know.
While we were in the process of updating our site, Google crawled the old version.
Now when using Google to search for our web site, we get results for both the new and old sites (e.g., http://www.example.com and http://backup.example.com).
Here are my questions:
Can I update the backup site content with the new content? Then we can get rid all of old content. My concern is that Google will lower our page ranking due to duplicate content.
If I prevent the old site from being accessed, how long will it take for the information to clear out of Google's search results?
Can I use google disallow to block Google from the old web site.
You should probably put a robots.txt file in your backup site and tell robots not to crawl it at all. Google will obey the restrictions though not all crawlers will. You might want to check out the options available to you at Google's WebMaster Central. Ask Google and see if they will remove the errant links for you from their data.
you can always use robot.txt on backup.* site to disallow google to index it.
More info here: link text
Are the URL formats consistent enough between the backup and current site that you could redirect a given page on the backup site to its equivalent on the current one? If so you could do so, having the backup site send 301 Permanent Redirects to each of the equivalent pages on the site you actually want indexed. The redirecting pages should drop out of the index (after how much time, I do not know).
If not, definitely look into robots.txt as Zepplock mentioned. After setting the robots.txt you can expedite removal from Google's index with their Webmaster Tools
Also you can make a rule in your scripts to redirect with header 301 each page to new one
Robots.txt is a good suggestion but...Google doesn't always listen. Yea, that's right, they don't always listen.
So, disallow all spiders but....also put this in your header
<meta name="robots" content="noindex, nofollow, noarchive" />
It's better to be safe than sorry. Meta commands are like yelling at Google "I DONT WANT YOU TO DO THIS TO THIS PAGE". :)
Do both, save yourself some pain. :)
I suggest you to either add no index meta tag in all old page or just disallow by robots.txt. Best way to just blocked the by robots.txt. One thing more add the sitemap in new site and submit it in webmaster that improve your new website indexing.
Password protect your webpages or directories that you don't want web spiders to crawl/index by putting password protecting code in the .htaccess file (if present in your website's root directory on the server or create a new one and upload it).
The web spiders will never know that password and hence won't be able to index the protected directories or web pages.
you can block any particular urls in webmasters check once...even you can block using robots.txt....remove sitemap for your old backup site and put noindex no follow tag for all of your old backup pages...i too handled this situation for one of my client............