Dynamics CRM 2013: Audit logs not showing details - dynamics-crm

I have come across an odd error I have never encountered before in the audit logs:
As you can see there are no details displayed but rather just an image
When I click on a specific line I only get this:
I haven't the faintest idea of where to start looking
I am a system administrator in the system so I don't think it's a permissions issue.
I can also confirm that there are real values in the system where the audit logs display that image instead of a value.
It seems to be a widespread issue across almost every entity in the system too.
EDIT:
I have looked in the audit management logs and found this:
I guess this indicates that no logs have been deleted?

This occurs when auditing was switched off and on again. From that moment the system cannot guarantee that the audit trail shown is complete and therefore it displays a torn page symbol.

Related

Trace failed login attempts Windows Server

We have noticed ~15k failed login attempts a day on one of our admin-accounts in the domain.
The source server is found and the event type is "Network", the source is a DC that has not been touched (except WinUpd) for years so a virus seems unlikely but of course possible.
Is there a way to trace exactly what the failed attempts point at? We have recently changed FSMO roles between two other DCs in the domain, maybe that has something to do with it?
You can check the login failed attemps based in audit logon events local computer policy.
use the keyboard shortcut Windows Key + R and type:gpedit.msc in the Run line and hit Enter.
In Group Policy Editor, navigate to Windows Settings >> Security Settings >> Local Policy >> Audit Policy.
Then double click on Audit Logon Events.
From there, check the boxes to audit failed audit attempts and click OK.
There you go! Now you’ll be able to see the complete logon activities (failed l) for your Windows computer.
Please refer this one as well based on event id you can know exactly what the failed attempts point at. : https://social.technet.microsoft.com/Forums/en-US/f49cd4d6-a7d5-4213-8482-72d1d5306dab/windows-server-2012-r2-help-finding-failed-logon-attempts-source?forum=winserversecurity
Reference: https://www.groovypost.com/howto/pin-windows-8-start-screen-programs-desktop/

Oracle APEX 5.1.2: Feedback Region

I have a client who use the built in feedback system to get his notes, but suddenly when I tries to go to the feedback I got the error as described in the image below.
I tried to delete cookies in the system and nothing happens, is there any solutions.
Regards

BulkDeleteFailureBase - lots of "Not Enough Privilege" records

I've taken over support of a CRM 2016 On-Premise system. I don't know the history of the particular instance, but I suspect it's been copied and/or imported many times.
The BulkDeleteFailureBase tables has just short of 2 million rows, almost all of which contain an error description like:
Not enough privilege to access the Microsoft Dynamics CRM object or
perform the requested operation. The current Organizationid '<GUID1>'
does not match with userOrTeam's organization id '<GUID2>'.
OrganisationBase has only one record with <GUID2> in it.
Has this happened because the instance has been copied/moved around incorrectly? If so, is this likely an indication more problems are heading my way in the future?
How can I recover from this?
BulkDeleteFailureBase is one of the system async jobs logging table where platform captures the run/success/failure logs.
Probably someone might have tried to clean the data like Plugin Trace log which were copied over from different DB backup/restore or CRM Org restoration. They used Bulk delete & all that fails, ended up here.
MS Support recommendation gives the script to clean those tables safely. Leaving it only gives you performance head-ache.

Parse migration tool stuck after canceling

So I have started the migration process from parse db to a mongolab db. The migration needed to be canceled at the verify stage. After pressing cancel, the migration tool is now stuck on the verify phase with no way to restart.
Has anyone else ran into this issue or have any ideas on how to restart the migration tool? Thanks ahead of time.
What I have attempted:
Removing and creating a new MongoLab DB
Multiple browsers
Logout/login
Parse Migration Tool Image
This issue seems to have been fixed on the parse side. But if it still happens, you have contact parse directly using the job id.
https://github.com/ParsePlatform/parse-server/issues/42
We discussed this here and basically in order to deal with migration
issues, we need you to individually report a bug through
parse.com/help. This allows each of you to get individual support,
from a team who can escalate issues, and for us to be able to ask up
front for specific information which will help us handle your issue.
We're exposing the 'job id' for the migration in the UI, and that is
something you should provide if you're having a problem.
Going to close this issue. For problems involving migration, or any
issues not specific to parse-server, please visit
https://parse.com/help and find the bug icon / "Something wrong?"
section to start the bug reporting flow.

Read application log written on Windows Azure

I have 10 applications they have same logic to write the log on a text file located on the application root folder.
I have an application which reads the log files of all the applicaiton and shows details in a web page.
Can the same be achieved on Windows Azure? I don't want to use the 'DiagnosticMonitor' API's. As I cannot change logging logic of application.
Thanks,
Aman
Even if technically this is possible, this is not advisable as the Fabric Controller can re-create any role at a whim (well - with good reasons, but unpredictable none-the-less) and so whenever this happens you will lose any files stored locally on a role.
So - primarily you should be looking for a different place to store those logs, and there are many options, but all require that you change the logging logic of the application.
You could do this, but aside from the issue Yossi pointed out (the log would be ephemeral; it could get deleted at any time), you'd have a different log file on each role instance (VM). That means when you hit your web page to view the log, you'd see whatever happened to be on the log on that particular VM, instead of what you presumably want (a roll-up of the log files across all VMs).
Windows Azure Diagnostics could help, since you can configure it to copy log files off to blob storage (so no need to change the logging). But honestly I find Diagnostics a bit cumbersome for this. It will end up creating a lot of different blobs, and you'll have to change the log viewer to read all those blobs and combine them.
I personally would suggest writing a separate piece of code that monitors the log file and, for each new line, stores the line as an entity (row) in table storage. This bit of code could be launched as a startup task and just run continuously as a separate process (leaving everything else unchanged). Then modify the log viewer to read the last n entities from table storage and display them.
(I'm assuming you can modify the log viewer even if you can't modify the apps that log to the file.)
What about writing logs to something like azure storage table? Just need to define unique ParitionKey/RowKey, then you can easily retrieve the log for the web page.

Resources