I have moved all images from a website to a content delivery network on another domain. And, as a result, lost all Google image search traffic. Is damage permanent or will traffic return, will images on another domain still allow my site to be in image search results? Maybe I should have moved images gradually? Any advise?
It's gone because Google can now no longer find the images it had previously spidered, google will of course find the new locations however there is no guarantee it will rank your images the same as before.
The best way to recover is to implement 301 redirects using your .htaccess file. Depending on how you've moved the images and if they are in the same or a different folder structure it may require a bit of work to fix.
Related
I'm parsing data from other website, and question wether it's better to download images and show them on my own or just links to website images I parsed . Is the link to image by default slower then image from own source?
Couldn't find answer to the simple question. If question is discussable and doesn't belong here, someone comment down please in order to delete it.
Some rules of thumb:
Don't display content on your page which you 'source' from another site without the other sites permission. ('Share this' links provided by youtube are okay, directly linking to the .flv file of someones video from another site to display on yours is not).
Don't copy content from other domains onto your domain without their permission first (doing so would be a copyright violation).
So to answer your question: You should copy the content onto your domain/host, but only if they have given permission to allow this kind of use.
Edit: I am interpreting your question as "I am taking content from another website [and putting it on my own] and I am wondering if I should link directly to their content ( tags pointing to the other domain) or if I should download/copy the content to my website and have my server handle everything?"
The "technical" answer is "it depends on how good your host is compared to the other host when serving content to the average visitor". Compare a page run by Google vs. the same thing run on a home server behind a 56k modem. It matters if you have broadband, but if you're on a 33.3k modem it doesn't.
My opencart website is not displaying images on way back machine of Alexa.com (How did
www.aaa------aa.com
look in the past?), I have checked multiple dates but images does not display anywhere.
My opencart images are also not getting crawled by google merchant center, it shows that it is because of the robots.txt but I have removed images from robots.txt file & still it shows same error.
My website is working fine otherwise and I am also getting orders but want to know if the above two issues are interrelated and what can be the best solution to this problem.
Thanks,
Gaurav
from what you are saying i think the problem with wayback machine is that you cache of images got cleared so ofcource most urls of images are dead, since opencart get 98% of images from cache after they get resized. So no problem there, i think this happens with every oencart setup that cached images are cleared from time to time, also in the same fashion please check images folder and sub-directories if you have any .htaccess limiting everything but local requests, also one last notice google usually takes from one day and up to update data regarding robots.txt and other data.
And if nothing works out try to specifically allow access to one directory with robots.txt
Hope i could help you.
I'm depending on my users uploading image links, not image files. So my site actually requests this links from these external sites by including them on an element and getting the src from the database.
Could sites detect they are receiving lots of "requests" (no idea if it's the right term) and block your url from doing more requests?
If they can, how many times would go unnoticed? Also are there policies like having to add a link to the source or something?
Sites can detect and block your IP address, especially if you are sending a lot of requests that could slow down their site and cause performance issues on their end. Having users post links shouldn't be any issue though.
Short answer: yes, hotlinking to files such as images can be detected and blocked. How many times can you do it? That depends on the limitations set by the website hosting the files you are hotlinking. There isn't really a definitive answer for every server about the limitations.
We host our Magento store internally with a not so amazing Internet connection. It's good enough to process the shop, but images take a while to upload to the visitor.
So what we have done is tweak the image.php file in app/code/core/Mage/Catalog/Model to point the images to a public space on RackSpace.
It's not an ideal method though with the admin console complaining that images don't exist and other random mess ups.
So I was wondering if anyone has implemented such a solution which I can use for our shop?
I imagine it will need to ignore the cache check as it cant upload to rackspace.
You can use the option of the backend of Magento
And redirect the trafic of images to another server(CDN).
There is one more solution, for example the module of one pica CDN with the option to select many CDN( amazon, custom cdn, etc)
http://www.magentocommerce.com/magento-connect/onepica-imagecdn.html
I have a client who wants to build a huge Image gallery website, and I am confused about how to structure the website for future Storage expansion.
Let me explain more...
Let us say that each user will upload his images to
website.com/Uploads/User/Images
Now creating the upload logic and displaying the images is not my issue here, my real problem is that say I have 200 GB hard Disk and if i have 20000 Customer where each client uploads 10 MB max, now as you see I will run out of space.
So how do I handle expansion in future without changing structure of web site, meaning that users will always upload to the same Path I have mentioned above, so obviously my front-end views will fetch images from same location too.
It may be stupid but I am lost on this. I mean, how guys like Facebook or other big sites do that ?
You can try using a cloud cdn(content delivery network), which will be dynamically expandable. amazon/rackspace, they are well known for this kind of service.
Ok after Tedious Searching, i have found the answer, basically it boils down to two Methods,
One Called Push, where you have to store the Files on the CDN Server by FTP, Api, etc...
The other called Pull Origin, where you dont have to change anything, you just Configure Your CDN to Fetch the Resources from your Servers, of course you have to Store files on Original Server First.
there is a lot more to it, but if anyone had my same wondering , just make Search about CDN Push or Pull in Google