Spam Bugs in analytics - debugging

I am getting 100 sessions on /error page on my all of three website. As it is coming from "not set" so unable to check the path of that traffic.
Is there any spam or how can I add any kind of filter when I don't have idea where the sessions are coming.
Please help.
error in analytics

Do you have the bot filtering enabled in your view? We find that it is not perfect but it's a good start.
Perhaps you can look at these request and see there is a set pattern; we have seen spam traffic coming from OS = Linux when typically our site visitors are mostly windows and mac. We know those are spams for sure; these requests are from weird, probably automatically generated screen size. Check city / province / OS etc of these sessions to see if there is a set pattern.
Do you have a view specific for reporting? If not, it will be a good practice to create a filtered view so that you can apply these filtering logics as needed. The idea is to keep an unfiltered view for diagnosis, and a filtered view for analysis and reporting. In the new view, you can create exclude filters to block the patterns. Make sure your filter is specific enough so that it won't block more than it should.
Check if your site has any soft or hard redirect for those pages; it could be why the error page does not capture any additional information. (It happens a lot for 404 redirects)

Related

Unknown Url Impression

I have a dev site running on Heroku. I put in google analytics. It was working fine. I decided to create a new tracking code so I could wipe out any previous data and start fresh.
I put in the new tracking code in my code, then pushed to Heroku. I noticed I was getting hits. But it was some weird url, looks like spam but what is this /www1.free-share-buttons.top ? why in the world is that getting logged?
This is most likely to be the result of so-called referral spam. You are getting hits from bots, which are targeting random UA identifiers, and not from actual visits and Analytics code on your website. To double-check this, you can view your Top pages reports in Analytics, by adding Hostname as secondary dimensions. You'll see (not set) next to these unknown pages, which means, they were pushed into Analytics from devices other than your site. There are several methods to prevent this happening, e.g. by adding an Analytics View filter, that prevents other Hostnames to appear in your reports.

Scraping pages that do not seem to have URLs

I'm trying to scrape these listings and provide more exposure for these job listings on a site that belongs to a client of mine. The issue is that I need to be able to link to the specific job listing in order for the job seeker to apply. This is the page I'm trying to save listing links from.
It would be ideal if I could save an address for the job seeker to click on to see the original listing and then apply.
What is this website doing to not feature a URL for these pages
Is it possible to provide a listing specific address
If that's possible how could I generate that address?
If I can't get a specific address I think I could get it so that the user clicks a link that triggers an internal script on my client's site which takes the listing ID and searches the site I found that listing on, and then redirects the user to that specific listing.
The downside to this is that the user will have to wait a little while depending on how far back the listing is on a directory. I could put some kind of progress bar with a pleasant "Searching for your listing! Thanks for being patient" message.
If I can avoid having to do this, though, that'd be great!
I'm using Nokogiri and Mechanize.
The page you refer to appears to be generated by an Oracle product, so one would think they'd be willing to construct a web form properly (and with reference to accessibility concerns). They haven't, so it occurs to me that either their engineer was having a bad day, or they are deliberately making it (slightly) harder to scrape.
The reason your browser shows no href when you hover over those links is that there isn't one. What the page does instead is to use JavaScript to capture the click event, populate a POST form with some hidden values, and call the submit method programmatically. This can cause problems with screen-readers and other accessibility devices, as well as causing problems with the way in which back buttons have to re-submit the page.
The good news is that constructions of this kind can usually be scraped by creating a form yourself, either using a real one on a third party page, or via a crawler library. If you post the right values to the target URI, reverse-engineered from examining the page's script, the resulting document should be the "linked" page you expect.

Retrieving .swf from website

I was wondering if you guys can help me out. Im making a website for an authorized Dish dealer. Ive been trying to retrieve the flash animation on dish.com. I was able to get it through firefox with 'Page Info' but all I get is a black rectangle without no animation. Its a .swf, any help is appreciated.
There are a couple of possible issues there.
There's a chance that the movie is loading something from that domain, which will not get loaded when the container movie is loaded in another domain due to cross-domain policy issues (and security sandbox restrictions).
Also, in its code the developer might have made it to check the URL and restricted some actions accordingly.
Another possibility is, it might be using some data that it gathers from the page source, or a server-side script which you might not be providing it.
These are the most likely possibilities I can think of, but if I get a chance to think more I'm sure I can come up with many others.

How to know quantity of users with turned off images in browser?

I'm working on the quite popular website, which looks good if user has turned on "Load images" option in his browser's settings.
When you try to open the website with "turned off images" option, it becomes not usable, many components won't work, because user won't see "important" buttons(we don't use standard OS buttons).
So, we can't understand and measure negative business impact of this mistake(absent alt/title attributes).
We can't set priority for this task - because we don't know how much such users comes to our website.
Please give me some advice how this problem can be solved?
Look in the logs for how many hits you get on a page without the subsequent requests from the browser for the other images.
Of course the browser might have images cached, so look for the first time you get a hit.
You can even use IP address for this, since it's OK if you throw out good data (that is, hits that are new that you disregard). The question is just: Of the hits you know are first-time, how many don't get images?
If this is a public page (i.e. not a web application that you've logged in to), also disregard search engine bots to the greatest extent possible; they usually won't retrieve images.

How do I Extend Blogengine.Net to collect statistics of visitors?

I love BlogEngine. But from what I can se it does not collect the standard information about the visitors I would like to see (referrer, browser-type and so on).
When I log in as Admin I have a menu item named "Referrer". I can choose a weekday and then I'll be presented with 1 or 2 rows with
"google.com 4 hits, "itmaskinen.se 6 hits" and so on, But that's not what I want to se, I want to se where my visitors come from, country, IP if possible, how many visitors and so on.
If someone of you are familiar with Blogengine.Net and can point me in the right direction to where I would put my own log-code or if you know any visitor-statistic-extension that can do it for me, I would be really happy to know. I prefer an extension, because if I make changes myself to BlogEngine it may break later updates I install.
Blogengine.Net is a blog software made in .Net found here: http://www.dotnetblogengine.net/
And yes, I prefer to take this question here rather then in the Blogengine.Net forum, you know why. ;)
(Anyone, feel free to edit my (bad) english in this post and after that delete this sentence)
This isn't an extension, but it's what I use to collect all my blogengine.net data and it should be upgrade safe.
When you log into the Blogengine.NET admin screens you can go to "Settings> Custome Code > Tracking Script", here you can put your http://www.google.com/analytics/ logging script. Google Analytics provides all the referrer, browser type, etc stuff you were wanting. And what's nice is you can then create additional accounts for other sites if you choose.
I use both Google Analytics and StatCounter to track visitor stats. I find that each one provides useful information that the other doesn't. And they're both free to a certain extent.
I place their javascript code int the site.master file of my custom BE.Net skin.
For Google Analytics I go a step further and pass the username of authenticated users as a custom variable. That way I can match users names up with the stats. To do this you can use the _setVar javascript method on the GA pageTracker like so:
<script type="text/javascript">
var pageTracker = _gat._getTracker("UA-129049-25");
var userDefinedValue = '<%= System.Web.Security.Membership.GetUser() != null ? System.Web.Security.Membership.GetUser().UserName : "" %>';
pageTracker._setVar(userDefinedValue);
pageTracker._trackPageview();
</script>
Anyone noticed that we miss all the hits coming from RSS readers? Syndication.axd does not run the analytics javascripts. So we miss the vast majority of viewers from the statistics. And we happily analyze that is just not impotant - ad-hoc visitors.
For the vast majority of cases, Google Analytics does just fine. It all depends on how much data you want. For example, if you want to keep note of IP addresses and resolve them to get domain names, and also highlight all visits to your blog from, say, your coworkers at the company where you work, you'd have to write some custom code yourself. However, it's all fairly primitive - these sorts of things are easily achievable using ASP.NET.
I set up gathering statistics on IIS web site of my BlogEngine instance and then analyze the logs using WebLog Expert - http://www.weblogexpert.com.
It is more reliable than google analytics, since I see really ALL requests that are coming to my IIS, no matter if this is a request to axd or to some static content. And, once I've found out that google was fooling me in the number of visits. After that I trust my IIS statistics much more than google.
There is a Widget which can be use to display Visits and Online Users Statistics.
You can find it from following links:
http://www.nuget.org/packages/Statistics/
http://www.itnerd.ir/post/2013/07/25/Visits-and-Online-Users-Statistics-widget-for-BlogEngine-2
but to see the instructions go to the second link.

Resources