SQL Server 2016 PolyBase install - installation

I'm trying to install and Run PolyBase in a brand new SQL Server environment - the installation runs fine, but I can't get the PolyBase services to start.
In SQL Server Configuration manager I see the state as 'Change Pending...', and task manager shows the services as 'starting'
If I try to configure PolyBase in SSMS it sits there for a while then bombs out with a null reference exception.
I can kill the processes in task manager but still no dice when I try to restart them - sits there for ages before finally returning 'the request failed or the service did not respond in timely fashion'
This is on a newly provisioned Azure VM, using the Developer version of 2016, so as far as I can tell the environment is completely clean - worth noting that I've tried this on two separate environments with the same result.
I've got plenty of SQL Server experience, but I've never used PolyBase before, so I accept that I might be missing something, but I would expect a fresh out-of-the-box install of SQL server to at least let me run the services...
Hopefully I'm doing something stupid and one of you can steer me in the right direction. Thanks in advance for any help.
CMR

Try enabling TCP/IP for SQL Server Network (by default only Shared Memory is enabled)

PolyBase head nodes are only supported in SQL Server 2016 Enterprise Edition.
https://msdn.microsoft.com/en-us/library/cc645993.aspx

It might be worth installing Service Pack 1 and trying again, because Polybase is now a Feature in Standard Edition. https://blogs.msdn.microsoft.com/sqlreleaseservices/sql-server-2016-service-pack-1-sp1-released/

Related

Self-Hosted Integration Runtime times out after ADF pipeline loads a few tables

I have recently installed Integration Runtime for a local server which hosts an Access DB. The idea is to pull data from it and store in in Azure SQL DB. I have done the following:
Integration Runtime Services Installed on local machine hosting the MS Access DB and connected to it using ODBC
Created linked services in ADF to connect to the DB
Created Datasets for source and destination DB for each table required. One for the source msaccess and one for the target in azure sql db
Created a pipeline to copy the data from the source and sink into the asql db mentioned in step #3
Basically, all the connections work however when I trigger my pipeline to load around 10 of these tables, it runs and loads the first two and then fails afterwards by timing out. I must restart the Integration Runtime everytime to get it back up and running otherwise I can no longer query the tables.
To mitigate this I figured there was too much traffic and the server needed to rest between calls so I added wait timers in between each step of the pipelines but not much success. It did help a bit but that might be coincidence.
The error log in the monitor spits out at the failed step is :
Error: 2200
ErrorCode=UserErrorFailedToConnectOdbcSource,'Type=Microsoft.DataTransfer.Common.Shared.HybridDeliveryException,Message=,Source=Microsoft.DataTransfer.Runtime.GenericOdbcConnectors,''Type=System.Data.Odbc.OdbcException,Message=,Source=,'
btw the integration runtime install service running is 5.12.7984.1 and the version of access installed is through office 365 x64. The exact MS access driver is 16.14430.20006. The OS is Windows Server 2019
I am getting the exact same error. To start, I did check the event viewer logs and saw some errors to do with access. So I gave the user running the IR more access to the registry keys/general log on as a service rights. This helped a little, but I am still stuck with the same problem.
When copying from an Access DB located on the SHIR itself to flat files in lake storage, I encountered the same error.
Removing Office 365 from the machine, then re-installing the Access runtime solved the problem.
This answer is from #ezaidi comments above.

Solaris 11 development environment on Cloud or hosting

I got in charge to create a website based on Java. Production environment specs include Solaris 11.1, OHS 12.1.3, WebLogic 12.1.3, Java 1.7.0_51 and Oracle Database 11.2.
I want to create a server on some cloud or hosting service as Development environment with the same specs to avoid migration problems to Production. I also think this approach helps to give my team a single server where they can work and have some testers/client to visit the site.
Normally I would use a local Development environment but a lot of people is involved and differences with Production can become a problem at migration.
I checked http://www.polarhome.com/ but I don't know if it will fit all specs needed. I looked at Windows Azure and Google Cloud with no success. AWS maybe? I also checked https://cloud.oracle.com but I don't understand if they already offer what I need.
Do you know any providers to create my Development environment or another approach/suggestion to develop this project??
Thanks!
EDIT.
To clarify, the client's Production environment already exists and is running somewhere. My project will be installed on that environment when development is finished. I personally think that developing on any VM with WebLogic 12.1.X and Oracle Database 11.X should be enough, but I've never done it so I wanted to follow client's advise on having a Development environment similar to Production.
Do you think I can just create a VM on any OS and just install WebLogic 12 with Oracle Database 11?? Any suggestions to avoid migration issues if I take that route?
I think that develop a new website from scratch thinking to use the architecture proposed by you is a nonsense. I think that if you will use cloud services like PaaS you will do something better.
In any case, you can find solaris VMs on Cloudsigma , Entic and Oracle Public Cloud

Exception while using Windows Azure Caching : No such host is known

I am trying to get started with Azure and am trying to use the Caching feature. I created a cloud service project and added a Cache worker role and a web role. I installed "Windows Azure Caching" nuget into projects for both the roles and added the name of cache worker role as identifier in DataCacheClients element in web.config of the web role.
I added the following code into the web role:
DataCacheFactory cf = new DataCacheFactory();
DataCache c = cf.GetDefaultCache();
When I try to run this locally on the emulator, I get the following exception:
ErrorCode<ERRCA0017>:SubStatus<ES0006>:There is a temporary failure.
Please retry later. (One or more specified cache servers are unavailable,
which could be caused by busy network or servers. For on-premises cache clusters,
also verify the following conditions. Ensure that security permission has been granted
for this client account, and check that the AppFabric Caching Service is allowed through
the firewall on all cache hosts. Also the MaxBufferSize on the server must be greater
than or equal to the serialized object size sent from the client.).
Additional Information : The client was trying to communicate with the server: net.tcp://MvcWebRole1:24233.
Inner Exception : No such host is known
Can you please tell me what I am missing here?
Azure SDK used : v2.0
Timing of your question couldn't be better. We also faced exactly the same issue and were scrathing our head as to what the problem could be. We had one project where everything worked perfectly fine and in one we were getting the same error. Based on our research, we have identified the problem with the Nuget package for caching. It seems a new version (2.1.0.0) was released yesterday and we found that if we install that package, we get this error. Can you check the package version in your case? The documentation states that this new version can only be used with the latest SDK (2.1) released today.
One solution would be to uninstall version 2.1.0.0 and install version 2.0.0.0. To install version 2.0.0.0, open Package Manager Console (View --> Other Windows --> Package Manager Console) and type following command there:
Install-Package Microsoft.WindowsAzure.Caching -Version 2.0.0.0
This fixed our problem. Hopefully it should fix yours too.
Here is a link to the Windows Azure Cloud Integration Engineering blog on how to deal with this same issue. They recommend upgrading to Azure SDK v 2.1 or rolling back as the accepted answer states.
http://blogs.msdn.com/b/cie/archive/2013/08/08/windows-azure-caching-2-1-0-0-no-such-host-is-known.aspx
This exception could also occur under compute emulator if for the role in the client library configuration there is no cache configured. In my case this happened on purpose, since the cache emulator has some problems that can slow down the test and debugging on the service.
In previous version of Windows Azure Caching, in this scenario the construction of DataCacheFactory would fail with an exception (handled by my code); with Windows Azure Caching 2.1 (and Azure SDK 2.1) in this same scenario the code would consider the role name as a server address and thus on DataCache construction would try to communicate with the non existing cache -- this leads to a 3 minute wait and the exception reported in the question.
I have changed my code to detect the new situation for this scenario -- you can find more detail in this SO question.

When installing Tableau Server getting error

When I am installing Tableau server I am seeing this server. I am the administrator of the system but still I was seeing the message.Please help me in this regard
It's often safer to use the default NTAuthority\NetworkService account if you're installing Tableau Server for the first time, since this is (almost always) guaranteed to work and can always be changed later.
If you do want to proceed with using SriHarsha-PC\SriHarsha as the Run As account, then take a look at the following link from the Tableau Software Knowledge Base which lists all of the permissions that your chosen account will need in order to run Tableau Server correctly.
Tableau Server Run As account permissions
If that does not provide sufficient information, then create a support request and Tableau Technical Support will try and help resolve the issue.

How to actually configure debugging in CFBuilder

I have ColdFusion Builder 2.0.0 installed and I am trying to look at the much vaunted step debugging. However, I cannot seem to get it to work as I don't have my site / JRun install setup in the naive way the examples show.
I am using version 9,0,1,274733 of ColdFusion and my configuration is as follows:-
Installed as multi-server version with Jrun here:- c:\Apps\JRun4
application files are here:- d:\websites\my.website.com
web root is here d:\websites\my.website.com\www
core library of CFCs is here d:\websites\frameworks\core which is mapped in CF as core
I have read this watched this http://help.adobe.com/en_US/ColdFusionBuilder/Using/WS0ef8c004658c1089-31c11ef1121cdfd6aa0-7fff.html and this http://forta.com/blog/index.cfm/2007/5/30/CF8-Debugger-Getting-Started and watched this https://experts.adobeconnect.com/_a204547676/p33029638/?launcher=false&fcsContent=true&pbMode=normal but I get stuck at the point after you have configured RDS and you are setting up the server for your project.
Now I am pretty sure the above is correct, when I move to the next page in the wizard I get the following:-
Now I as I understand it my Server Home should be c:\Apps\JRun4 and my Document root should be d:\websites\my.website.com
This all looks like it is going to be fine until you actually try and debug when I get
followed by
I can confirm that the server is running and RDS is enabled as in the RDS Dataview I can see all my databases.
Any help would be gratefully received as this is very frustrating and the documentation is very lacking.
There is a video tutorial as well that you may want to check and see if that helps. http://blogs.adobe.com/anand/2011/01/learn-how-to-debug-coldfusion-applications-using-coldfusion-builder-2.html
You need to specify the RDS username/password and the "application server name". If you are using the base instance that was installed when you setup the multiserver install of CF that is "cfusion", otherwise its the name of the instance you are using.
The RDS username is most likely "admin" unless you setup custom users for RDS. The password is the RDS password you specified when you installed CF.

Resources