In a News sitemap, we need to include URLs for all our articles, but only from the last three days. we shouldn't include older articles. Why we shouldn't old articles and if we keep the old articles what would happen?
I assume your talking about the XML sitemap for the crawling engine like google?
Yes, you can keep the old articles, even more, this is preferable.
Related
Good day to all.
I connected the tag manager to the site, but I want to be sure that it is on all pages.
This is an online home appliance store.
I want to understand the Laravel principle, how can I find all the pages?
Help me figure out the code
Thanks in advance.
For a quick check, I normally use http://www.gachecker.com/ Very useful, it not only indicates absent libraries, it also indicates double libraries. Pretty useful.
You can implement basic pageview tracking through GTM and in a few days see if there are missing pages in Analytics report.
You could use your sitemap or access log (grouped by page path) to check it manually or with a site crawling tool.
Finally, sure, you can go through Laravel's page templates and make sure GTM is referenced in each. I'm not an expert in PHP frameworks though.
If you want to see Laravel-only solution, you should ask your question with no reference to GTM.
I've created a multilanguage website and I tried to generate a sitemap with a common Sitemap Online Generator.
Unfortunally it crawled links like "en/en/..." or "it/de/en/..." (that not exist and are not correct of course). I'm afraid that Google could do the same.
I read all about tag (maybe the problem is there) and did lots of trying, but the resutl is always the same: lots of redundant links (en/apartments/en/apartments/torre)
Any suggestions?
Solved!
I wrote my problem to the sitemap generator webmaster. It was an error about its tool (it doesn't consider singular apices around base-tag's attribute).
I sent the sitemap to google and everything was ok.
I have my google news sitemap, which according to google's spec, should contain articles posted in last 2 days. Now, suppose there are no articles pushed in last 2 days for my site, my news sitemap would be empty. IS this the desired behaviour. Do I need to show something else, if no articles are posted in last 2 days. Will the webmaster tools invalidate the sitemap?
I don't see what other choice there is for a sitemapper if you have not created any news stories within last two days. You could perhaps choose not to update your news sitemap? :)
Will the webmaster tools invalidate the sitemap?
Yes, it will... But then, GWT judges it as any regular xml sitemap and does not account for specific Google News sitemap rules.
So the more important question is:
IS this the desired behaviour. Do I need to show something else, if no
articles are posted in last 2 days.
... to which I cannot find any clear answer anywhere :(
I've played around with Google Sitemaps on a couple sites. The lastmod, changefreq, and priority parameters are pretty cool in theory. But in practice I haven't seen these parameters affect much.
And most of my sites don't have a Google Sitemap and that has worked out fine. Google still crawls the site and finds all of my pages. The old meta robot and robots.txt mechanisms still work when you don't want a page (or directory) to be indexed. And I just leave every other page alone and as long as there's a link to it Google will find it.
So what reasons have you found to write a Google Sitemap? Is it worth it?
From the FAQ:
Sitemaps are particularly helpful if:
Your site has dynamic content.
Your site has pages that aren't easily
discovered by Googlebot during the
crawl process—for example, pages
featuring rich AJAX or images.
Your site is new and has few links to it.
(Googlebot crawls the web by
following links from one page to
another, so if your site isn't well
linked, it may be hard for us to
discover it.)
Your site has a large
archive of content pages that are not
well linked to each other, or are not
linked at all.
It also allows you to provide more granular information to Google about the relative importance of pages in your site and how often the spider should come back. And, as mentioned above, if Google deems your site important enough to show sublinks under in the search results, you can control what appears via sitemap.
I believe the "special links" in search results are generated from the google sitemap.
What do I mean by "special link"? Search for "apache", below the first result (Apache software foundation) there are two columns of links ("Apache Server", "Tomcat", "FAQ").
I guess it helps Google to prioritize their crawl? But in practice I was involved in a project where we used the gzip-ed large version of it where it helped massively. And AFAIK there is a nice integration with webmaster tools as well.
I am also curious about the topic, but does it cost anything to generate a sitemap?
In theory, anything that costs nothing and may have a potential gain, even if very small or very remote, can be defined as "worth it".
In addition, Google says: "Tell us about your pages with Sitemaps: which ones are the most important to you and how often they change. You can also let us know how you would like the URLs we index to appear." (Webmaster Tools)
I don't think that the bold statement above is possible with the traditional mechanisms that search engines use to discover URLs.
I am building a forum site where the post is retrieved on the same page as the listing via AJAX. When a new post is shown, the URI fragment is changed (ex: .php#1_This-is-the-first-post). Also the title and meta tags are changed.
My question is this. I have read that search engines aren't able to use #these-words. So therefore, my entire site won't be able to be indexed (as it will look like one page).
What can i do to get around this, or at least make my sub-pages be able to get indexed?
NOTE: I have built almost all of the site, so radically changes would be hard. SEO is my weakest geek-skill.
Add non-AJAX versions of every page, and link to them from your popups as "permalinks" (or whatever you want to call them). Not only aren't your pages available to search engines, they can't be bookmarked or emailed to friends. I recently worked with some designers on a site and talked them out of using an AJAX-only design. They ended up putting article "teasers" in popups and making users go to a page with a bookmarkable URL to read the complete texts.
As difficult as it may be, the "best" answer may be to re-architect your site to use the hash tag URL scheme more sparingly
Short of that, I'd suggest the following:
Create an alternative, non-hash based URL scheme. This is a must.
Create a site-map that allows search engines to find your existing pages through the new URL scheme.
Slowly port your site over. You might consider adding these deeper links on the page, or encourage users to share those links instead of the hash-based ones, etc.
Hope this helps!