Can I create a report that will list all previously run reports? - dynamics-crm

I'm using Dynamics CRM 2015 and want to create a report that will show all reports run in the last 12 months.
I've been using the Report Wizard and can't seem to find the entity that is created when a report is run. I can find when a report was created but not every time it has been run.
Example of expected results:
Report X
4/3/2019 Admin 1
4/2/2019 Admin 3
Report Y
4/3/2019 Admin 2
4/2/2019 Admin 1
I'm not worried about the formats, I will most likely tinker with it after. I simply want to find a way to display every instance any report has been run.

Since you're on CRM 2015, it would follow that your system is on-prem.
That means you're not able to use the relatively new Activity Logging a.k.a. Read Auditing that's available in D365 Online, which seems to have what you're looking for.
The out-of-box audit in CRM 2015 would give you some kind of "user access" auditing (i.e. when people login), but not show you specific report runs. It's really designed to capture changes to the data for audited entities.
As far as I know there is no entity record created when a user runs a report. Provided you were willing to hook into and/or replace all the report triggers throughout the system (i.e. in all ribbons), you could hypothetically build something to track report runs. But it seems like that would be cost prohibitive.
According to this article you should be able to pull this info out of the ReportServer DB. I'd quote the relevant parts here but it seems very involved - creating temp tables, etc.

Related

BulkDeleteFailureBase - lots of "Not Enough Privilege" records

I've taken over support of a CRM 2016 On-Premise system. I don't know the history of the particular instance, but I suspect it's been copied and/or imported many times.
The BulkDeleteFailureBase tables has just short of 2 million rows, almost all of which contain an error description like:
Not enough privilege to access the Microsoft Dynamics CRM object or
perform the requested operation. The current Organizationid '<GUID1>'
does not match with userOrTeam's organization id '<GUID2>'.
OrganisationBase has only one record with <GUID2> in it.
Has this happened because the instance has been copied/moved around incorrectly? If so, is this likely an indication more problems are heading my way in the future?
How can I recover from this?
BulkDeleteFailureBase is one of the system async jobs logging table where platform captures the run/success/failure logs.
Probably someone might have tried to clean the data like Plugin Trace log which were copied over from different DB backup/restore or CRM Org restoration. They used Bulk delete & all that fails, ended up here.
MS Support recommendation gives the script to clean those tables safely. Leaving it only gives you performance head-ache.

Alfresco - Download statistics and user permissions report

Hey to every alfresco pro out there!
Is there any way to create a report (graphical or textually, i don't care) to see the following information:
download count per file
how many times did user X download a specific file
which permissions do the users have
Are my goals easy to realize? Is there any plugin out there that i can use for this? (Already searched for some but couldn't find one) Hope that you can help me :)
mtzE
There is nothing out-of-the-box that is counting downloads. Maybe the audit service can be used to count reads, but you'll have to turn it on and configure it. Once turned on, the audit service writes records to a set of audit tables in your Alfresco database. You can then use any reporting tool to query those tables.
If you want to check the permissions a user has you can use something like OpenCMIS to connect to the repository, traverse a folder path, and then, for each object, you can inspect the ACL of that object to use as data in your report.
As Lista said, one way to create such reports is to use AAAR, but that is not required.

send email automatically in Microsoft dynamics CRM

I have field in entity Quote called new_date_expiration (type = Date).
I want send email if new_date_expiration equals date system in Microsoft dynamics CRM automatically.
any ideas?
thanks in advance.
You can do this using the CRM Workflow Automation Tool combined with an on-demand workflow. The setup process is quite simple:
Create a new workflow in CRM (make sure the "run on demand" option is turned on) which generates and sends your email.
Create an advanced find query on your entity which defines the criteria of when you want the email to be sent. In your case it sounds like the condition would be something like Expiration Date [equals] Today.
Once you are satisfied with the advanced find query, copy of the source FetchXML for it.
Configure the CRM Workflow Automation Tool (instructions on how to do this are included in the download package) to run the workflow created in the first step using the FetchXML you just copied.
Set your server to run the tool run daily at, for example, 8 AM using Windows Task Scheduler.
Once you have this configured, everyday at the specified time the results of the FetchXML query will be pulled and the provided workflow will be run against it. This effectively accomplishes exactly what you want without a bunch of timeouts.

Expired job executions in Spring Batch

Problem
Users can submit data to generate a report, which triggers a spring-batch job. If the same data is submitted (by the same user or another user) the same report should be generated such that Spring Batch doesn't start a new job, under the premise that the report has already been generated.
To make matters a little more complicated, generated reports expire after 90 days. The idea behind this is that the data gleaned from various web services used to build the report is likely out of date. Therefore, after 90 days the report should be re-generated using new data from those web services.
Questions
When a job has already run, how can I discover the job execution id for that job? This id is used in the URL to uniquely identify a report. JobExplorer is severely limited in querying Spring Batch data.
How can I trigger another instance of the job only after 90 days? The issue is that given duplicate job parameters, a JobInstanceAlreadyCompleteException will be thrown. Must I encode the 90 days has an extra identifying parameter, or is there an easier way?
Clean up old jobs must be done using business methods as well as for expired reports.
After this premise you can try a different path to solve your problem:
Every user launch a different job, with same report properties but
an extra job-parameters to make every jobs different
First step is to check - using business method - if there is a running job for that report ; in this case notify user he have to wait or retry later (use a decider)
Second step is to check if there is a completed - not expired - report using a business method;if so retrieve it and show to user (use a decider as for step 2)
Generate report (delete old, if necessary)
Show report to user.
Of course, generated report metadata tables are different from SB tables and should be accessed using DAO related to domain context (the report in your case).
Can this a valid alternative?

Storing ssrs reports in a file that can be called immediately

Hi Fellow SSRS Developers,
I have a scenario that I'm trying to tend to but need to know if what I want to do is even possible.
I have 4 reports that I would like to have run and then store the actual report in a file on a server. The reason for this need is because the response time on the reports is a bit long and I've done everything in SQL to speed it up.
What I want to have happen, is when a user clicks on the report name, instead of rendering the report on their screen I simply want to call the report that is already in a file so that it will load in lightning quick time.
Has anyone ever done this with SSRS and is it even possible?
Thanks,
Other than running reports on demand, there are two specific options: Running from a Cached report and running from a Snapshot.
You can see details on all of this in Setting Report Processing Properties.
Caching
From Books Online:
To enhance performance, you can specify a report (and data) to be
cached temporarily when a user runs the report. The cached copy is
subsequently available to other users who access the same report. With
this approach, if ten users open the report, only the first request
results in report processing. The report is subsequently cached, and
the remaining nine users view the cached report.
So here you can see that it is a specific user action that causes a stored report to be created.
See Report Caching in Reporting Services.
Snapshots
From Books Online:
A report snapshot is a report that contains layout information and
data that is retrieved at a specific point in time. You can run a
report as a report snapshot to prevent the report from being run at
arbitrary times (for example, during a scheduled backup). A report
snapshot is usually created and subsequently refreshed on a schedule,
allowing you to time exactly when report and data processing will
occur. If a report is based on queries that take a long time to run,
or on queries that use data from a data source that you prefer no one
access during certain hours, you should run the report as a snapshot.
Here you can see that these are these are generally set up on a regular schedule, i.e. independent of user activity.
See Creating, Modifying, and Deleting Snapshots in Report History.
In this case it seems like Snapshots might be your best option so you have more control of when the stored report is created. The main issue with Snapshots is that they need either stored credentials or an unattended execution account so might not be possible in all cases.

Resources