Helllo,
I have a working and up to date environment for Dynamics v9 on-prem.
I want to create another environment which will be identical to the above mentioned one.
Let's say that I want to clone PROD env and create a new Test environment with the same data and customizations that I have in prod.
What are the proper steps ? There is a lack of resources about this topic so I might have posted this in a wrong place. In this case, sorry and please refer me to the correct forum if you know such.
My idea is that, yes I would back up the sql server and then restore it but what happens on the Dynamics side ? Do I need to import any kind of configuration from the Dynamics Deployment Configuration ?
Or can I create the new environment and then restore the prod database in this env ?
Thanks
Steps:
Backup the database of the existing D365 organization
Create a new D365 organization
Restore the new organization using the backup of step 1
The wizard will guide you through the process.
Related
I have a prod environment where the version of dynamics in 9.0.16.7 (prod, was not installed by me).
I have installed installed the new organization which is version 9.0.2.3034. (I have not applied any kind of updates, it was this version from the beginning).
Now I want to take all that's in prod to my new organization.
Steps I have performed:
Got the backup of the prod environment sql database.
Restored it in the sql of the dev server.
Opened Deployment manager -> Import organization -> DB was automatically detected -> did the mapping -> waited for the result.
The result I got looks like this:
Microsoft.Crm.CrmException: Database having version 9.0.16.7 is not supported for upgraded.
My question(s):
How can I restore the prod environment in my new organization ?
If I need to change the version of the new organization what are the steps that I need to take in order to achieve the result ?
Any kind of response will be awesome because I struggle finding any kind of resources about this topic.
Thanks
I found an update which upgraded the organization.
I imported organization and it succeeded.
I would like to know how to auto manage portal's custom code just like in TFS/VSTS?
At present ,I am using XRMToolbox to manage ,push or pull portal's code into CRM Instance but disadvantage is code checkin and checkout.
Can anyone help me in this to manage a code with auto pull and push option into CRM instance with checkin ,checkout options?
Thanks in Advance!
I'm afraid the XRMToolbox plugin doesn't support it yet.
Ref: https://github.com/MscrmTools/MscrmTools.PortalCodeEditor/issues/13
But there is no stopping you from creating your own pipeline - at the end of the day portal code is just bunch of Crm entities. Part of Crm SDK is configuration migration tool - last version is here:
https://www.nuget.org/packages/Microsoft.CrmSdk.XrmTooling.ConfigurationMigration.Wpf
So the idea is:
1) Get this tool
2) Define entities you want to backup & create schema xml file for them. I think you'd want adx_webpage, adx_webfile, adx_pagetemplate (and all attributes from them)
3) Export data using this schema - this exports them to .zip package that contains simple structure (schema file and data file); so you can unzip it and store in your git branch (pull)
4) For push zip this file and again use configuration migration tool to import the data
This gives you also an opportunity to have separate dev version of portal code and production version of portal code (which is always a good thing).
Portals code is made up of configuration changes to a solution (which can be extracted as xml) and data (records such as web pages, web roles etc.)
There are several tools available to help you source control both.
xrm-ci-framework provides automation tools to extract your CRM solution as xml, and then source control it. You can do this locally or in the cloud with Azure DevOps or other.
msbuild-xrm-sourcecontrol is similar. It integrates into Visual Studio to help you extract CRM customisations locally. It also has a partner project xrm-datamigration which helps you extract data from CRM, version control it and deploy it to other environments in your release pipeline. Both have documentation on the GitHub pages I've linked; this blog post is informative too.
We use CRM 2016 SP1. We deployed our latest changes from Dev to QA, staging and production by means of importing a managed solution file. As part of the solution file there is a custom Action workflow (Category = Action) with the publisher prefix set to our company name.
The workflow has been activated and working well in QA and staging but when trying to activate it on production CRM comes up with the error "Unexpected Error". Downloaded error details show the same message.
Upon investigating the workflow in production we realised that its publisher prefix had been changed to "new_". To be more specific, the Process Name property starts with the correct prefix name but the Unique Name property starts with "new_".
We had not made any changes to the workflow in Dev and it was working fine in production prior to the deployment.
So far my research on the Internet on how/why this change has come about and how to fix it has been in vain. So any help is greatly appreciated.
Regards
This is identified as an issue in 2016 SP1.
Check Link Here
The solution is to Recreate the Process with correct Prefix and delete the one which start with new_ that to before final deployment/packaging/release.
It is observed that even both processes has different prefix they are identified as same process.
Please see this link for the solution
https://community.dynamics.com/crm/f/117/p/243572/680715#680715
Does anyone know about this problem: Any new fields I add work fine in the local back office, but when I use Webmatrix to publish to the server (discountASP.net) fields don't show up. I did a view source in the browser and they're just not there!
For example, #Umbraco.Field("comments")
Thanks!
Daniel
If you add new field, they are only added in the database. That means you would need to update the database on the production website. Webmatrix doesn't do this for you (by default).
There are a few ways to handle this scenario:
copy your database to the production server (i would advice against this, because you might overwrite content and media changes on the production server)
create the fields manually on the production server (easy solution)
use a commercial package like courrier (personally i believe it's a good solution, only if you have a content staging workflow)
use a free package like usync (http://our.umbraco.org/projects/developer-tools/usync)
We've been experimenting with Octopus Deploy on a development PC and now want to transfer the environment we've created onto our main Octopus Deploy server (which is used by other teams and already has a few environment set up on it).
So we would like to backup/restore this one environment. However, it looks like Octopus only allows you to backup/restore the entire database.
Is it possible to move a single environment from one Octopus server to another using backup/restore or another means?
What worked for me was simply doing the following in order:
Shutting down Octopus service so that no transaction going through.
Copy the raven database (usually stored in Program Files\Data) to your new server.
Install the new Octopus server and during the setup, in the Storage Tab, specify the location of your data location copied in the second step above.
The Octopus developer, Paul, mentions the great thing about RavenDB is the installation. It requires no services running like SQL. It's just a copy paste of the data itself and great for installation and portability.
There's currently no way to backup/restore just part of the database - you'd need to restore a full backup, and then delete the information you don't need.
Octopus 2.0 (which is now a public beta) has a comprehensive REST API so it would be possible to use that API to fetch a subset of information and import it to your new Octopus server.