Users not able to view resources in resource plan and also not able to add new resources to the resource plan in project server 2013 - ms-project-server-2013

I recently came across a situation where two of our users were not able to see the added resources in resource plan. Moreover they can add new resources but cant save them. All they see is an alert saying 'Project is no longer checked out to you'.It happens only for specific projects.
As the alert says it was checked out to someone else, I tried to force checkin the project/resource plan but it doesn't appear in the 'Force Checkin Enterprise Objects' itself.
I thought whether those users doesn't have enough privilege to do the operation, but they are part of Project Managers/Portfolio Mangers where they have full access to do this task.
Can someone provide some insights on what caused this issue to the users.

I see that the Resource plan was corrupted, and thats why users were not able to add/save it. I created a utility using PSI which deletes the Resource Plan for those projects.
I validated Project Schedule,Resource Utilization, capacity and none of them lost their value after deleting the resource plan. So the utility resolved the issue.
I am posting answer for my own question as it helps for someone else..

Related

How to automatically delete old Azure Blob Storage containers

Microsoft offers the Lifecycle Management service, which allows me to setup a rule for an action.
There, I can delete old blobs by setting up an expiration date. However, after deleting all the blobs the container remains there, forever empty.
Is there any configuration that also deletes de container whenever it is x days old and/or empty?
We don’t have container delete as part of life-cycle management now.We are in planning to add in the future.
You can share your feedback or suggestion here. All the feedback you share in these forums will be monitored and reviewed by the Microsoft engineering teams responsible for building Azure.

Cannot Access Business Rules in my Dynamics CRM Dev Environment

I'm having difficulties in creating / updating business rules. Every time I'm trying to create new Business Rules or Edit the Existing business rules. I'm getting an error without a error code. basically it's not letting me to look or create the business rules and sometimes even the processes(workflows) also.
I'm working on CRM 2016 on-premise (version-8.0.0.1088),I'm having a system Administrator's role. It's really tough to troubleshoot and find a solution for this issue has been struggling for a while with the issue.
I've imported an unmanaged solution recently. It's very annoying, spent a lot of time in researching but still no use, any help much appreciated!.
Error Log Details: (from Trace)
TargetInvocationException: Exception has been thrown by the target of an >invocation. ---> System.IO.FileNotFoundException: Could not load file or >assembly 'log4net, Version=1.2.15.0, Culture=neutral, >PublicKeyToken=669e0ddf0bb1aa2a' or one of its dependencies. The system cannot >find the file specified.
thanks to Federico Jousset who have promptly responded to my question. He has suggested me to enable the Trace log feature in CRM (on premise), which I wasn't aware of, and once done, it gave me reason why it was failing.
It looks like your custom dll deployed to dynamics crm is dependent on log4net, which is not found in server GAC/merged dll.

Active Directory Domain Services Auditing

I'll try to explain my goal as good as I can;
I want to trigger a script whenever there is a new computer added to a Organizational Unit.
To do this i need to activate the logging of this event under the local security policy/audit policy. I guess my question is, do I need to do this on all the domain controllers, or is it enough to do it one just one?
Also, is it possible to see the event from a member server with the Management Tools pack installed? As I don't want to put too much work on the Domain Controllers.
Here is the Microsoft article that gives 4 ways of tracking changes in Microsoft Active-Directory. You will find everything you need from configuring the eventlog to receiving notifications by way of different kind of polling.

How to copy the entire DB in CRM Dynamics?

It seems to me as a wise idea to test run my workflow on a local server before deploying in at the customer's. To be entirely sure, I'd like to copy all the data from their DB to my test organization (I have full access rights). The problem is that I can't see any straightforward way to export the whole shabang to a XML Spreadsheet.
What's the best way to export/import everything from/to a DB? The source and the target servers are not the same.
Of course I've got the option of backing up the clients DB and restore it, would the brown stuff hit the fan, but it'll far more professional if I won't have to.
The client's DB is in the cloud, which makes me suspect that perhaps I won't be able to access it at all and as far I can see, there's no way to back-up the data there. Am I missing it or is it that bad?
I fully agree that would be sensible. Usually we have a number development and test servers for all our work, generally we do not exactly mirror the data in the client database however.
We create a representative sample of data in our dev servers and then just move across the Crm solution for deployment.
As far as I know there is not straight forward way to get all the data, if you really want to do this I would suggest taking a back up of their database and importing to yours.
(As a side note, not all clients are happy for copies of their database - especially if its a live system - to be taken off site. Personally if it is a live database I wouldn’t put that risk on yourself, if the data gets lost or leaked you might suffer the consequences).
James raises good points about the business aspects of your request, however to get hold of the record-level data there are few options. The easiest by far is a wholesale export and import of the underlying SQL database. (For the record, the alternative is to do a data migration from live into a different db but this is no small task so I won't even entertain that any further here).
You mention that the client is using CRM Online ("...client's DB is in the cloud..."). You can raise a (free) support request with CRM Online Support who will provide you with a copy of the YourOrg_MSCRM database which can then be reimported into an on-premise deployment.
If you wish to simply have a test instance that has a copy of the Microsoft CRM Online organization, Microsoft does provide a means to do that. Depending on how many professional user licenses that the customer has, this may be free, but could be an extra cost and both instances would count against the storage limit for Microsoft CRM Online. You can see full details here - https://community.dynamics.com/crm/b/crmteamblog/archive/2014/03/20/introducing-sandbox-instances-in-crm-online.aspx . You can see steps on how to setup a sandbox instance here - https://technet.microsoft.com/en-us/library/dn467371.aspx "Add an instance to your subscription". This is something that I have used with one of our Microsoft CRM customers as it was a very good way to help validate the Scribe Online migration and customization changes we were making before moving those into production. The nice thing about doing it this way is that everything is still contained in the same Office 365 tenant and you can limit which users have access to the Sandbox organization, which is important for customers in knowing that their data is safe and not on some unknown server or machine.

How to do Continuous Integration with a live website without affecting users?

I have implemented Continuous Integration using TFS Version Control and TFS Build 2010. The compiled website project gets dropped in a shared folder with a version number.
Now I have a very basic question and may be a stupid question. When we normally deploy a website project from VS 2010 to a webserver it uploads App_Offline.htm file to the website folder so no requests are served to the user. After publish is completed that App_Offline.htm file is removed. During that period of time users see outage.
If we use CI on a live website then how can we eliminate that outage which appears to a user. I believe the whole point of CI is that users get to see newer features and the site is never down.
How is this accomplished? If we deploy website project to root folder then existing users will be affected and that is certainly no advisable.
I wanted to know what is the recommended practice with VS2010, TFS2010 Build & Version Control.
There's no real foolproof method for this, service up-time is never 100%, that's why people usually define it in 'nines'
But, if you had multiple web servers (Backup, fail-over, mirror etc.), you could roll out the update across them, so that as you update some servers, others will still be online (albeit with the old version) to serve users.
In general, only some of the largest websites have to worry so meticulously about being down for a few short minutes, so make sure you're focusing your energy in the right place ; )
Regarding taking down the site for the shortest time possible, the only way I've seen this done successfully is using multiple sites - either load balancing, or 2 sites on the same machine + swapping host headers after the release/warm up. But in most cases it's not worth the effort, releases shouldn't take down the site for more than a few seconds in which time there should be relatively few requests. You're better off trying a few things you can do to help your users live through a site release.
Move session out of proc.
If the users session lives in the app pool it will be lost when a new version is released, change the config to move it into a session server or the database.
Specify a machine key for the website
Viewstate (and cookies?) are encrypted using a key that is generated when a site starts, if a site restarts due to a release any users filling out a form will receive a invalid viewstate exception on postback. (Note: this may have other security implications)

Resources