Mathematica: setting "SetDirectory" to a remote server? - wolfram-mathematica

How do I set "SetDirectory" to a remote server with Mathematica?
I have tried SetDirectory["//server/folder/"] but without success. I also tried putting "smb:" before the // but it doesn't work. I get:
SetDirectory::cdir: Cannot set current directory
Looked at the documentation, but could not find help. Any suggestion? The server is password protected. I was expecting a username/password pop-up window, and I wonder if it's possible to set that up.
Thanks

I found a better way: I leave the Mathematica notebook file within the remote directory instead. That way I only have to go through password protection once.

Related

MATLAB - Using urlwrite with https Site Not Working in OS X [duplicate]

Anybody know if it's possible?
I'm trying to get the data by using the following code
url = 'https://cgwb.nci.nih.gov/cgi-bin/hgTracks';
params = {'org','Human','db','hg18','position','EGFR'};
urltxt = urlread(url,'get',params);
but get the error
??? Error using ==> urlread at 111
Error downloading URL. Your network connection may be down or your proxy settings improperly configured.
If I substitute https to http, it works, but I get "301 Moved Permanently" page with the above https-link.
The link in browser works properly in both cases (redirecting http request). The site does not require any authentication.
May be there are other ways than urlread?
Sorry, I found the answer on SO:
Handling an invalid security certificate using MATLAB's urlread command
Will test and remove if needed.
UPDATE:
It really works. DO you think I should delete the question?
An alternative solution that worked for me:
(ps.: I'm using Fedora Linux, Matlab 2017a. Not sure if it will work in a simply way for windows users).
The following command line in Matlab that I used to get the data as:
AllDataURL=urlread('https://bittrex.com/api/v1.1/public/getmarketsummaries');
was successfully replaced by the following command line:
[status,AllDataURL]=dos('curl https://bittrex.com/api/v1.1/public/getmarketsummaries');
Although the result value for the variable 'status' is zero, the data into variable AllDataURL is exactly the same as the previous one when using urlread.
Hope it helps.

Unable to get the configuration file WcmApiConfig.properties

I have an issue with FileNet-p8:
All simply, I do not have access to a WorkFlow already designed. I have always a popup asking me to authenticate , I use the same Login/psw to access to the Workplace. But always I get the same message :
Unable to get the configuration file WcmApiConfig.properties
I'm workin with the IBM JVM 1.6 and the Firefox browser.
Thanks.
In our environment, you must use IE in order to avoid that error message. We have not been able to get FireFox to work with WorkplaceXT or PCC.
Even when launching PCC from ACCPE we need to use IE.
If you see the login screen, don't even bother with your credentials. It simply will not work.

Why do I get a blank page instead of Joomla?

I am just beginning with Joomla on a linux fedora server (running apache,mysql,php), running in local. I have also installed phpmyadmin.
I have dowloaded/unizip joomla on var/www/html/joomla1 and created a database for joomla1.
Then restarted the server. Selinux is set to permissive.
However when browsing at http://localhost/joomla1/ i simply get a blank page
Any idea what's wrong ? How to fix this ?
Check your database permissions (does the user have permissions to access the data), check your error logs on the server.
Maybe this URL is right?
http://localhost/html/joomla1/
Check /var/log/httpd/error_log
Look for anything saying "Fatal:"
This will give you an indication of whats causing the problem

MOSS search crawl fails with "Access is denied ..."

Recently the search crawler stopped working on my MOSS installation. The message in the crawl log is
Access is denied. Check that the Default Content Access Account has access to this content, or add a crawl rule to crawl this content. (The item was deleted because it was either not found or the crawler was denied access to it.)
The default content account is an admin on the site collection that I am trying to crawl.
Almost every result for this error on Google tells me to add the DisableLoobackCheck registry key with a value of 1. I have done this and rebooted and the error continues.
The "Do not allow Basic Authentication" checkbox in my crawl rule screen is unchecked.
Is there anything else that could be causing this error? Something with file system or database permissions maybe?
Edit: All signs seem to indicate that the "DisableLoopbackCheck" should fix this, but it doesn't seem to work. Could I be doing something wrong when I enable this?
I'm doing it in My Computer\HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\Control\Lsa, where I create a new DWORD key called DisableLoopbackCheck and give it the hex value 1.
It turned out not to be related to DisableLoopbackCheck. The problem was that the search was accessing the site through its external URL. You are supposedly not supposed to be able to access a site from within a server using the same URL that you use to reach it from the outside, at least in pre-SP1 MOSS. But I was doing this for about two years somehow. MS Support tells me they don't quite understand how it was ever working. So it looks like I ran into an issue that should have been manifesting all along. I'm not sure what caused it to appear suddenly, maybe some routine patching of the server. The solution was to extend the web application so it was accessible internally through the machine name, then point the crawler at that.

Upload images problem: IO error. (Error #2038)

I'm using script which is uploading files to server via flash component. Sometimes, very rarely, when trying to upload images via Firefox I get following error: IO error #2038. Searching on the net I could find reason why is it really happening to me. But I found solution for my case:
I open IE6, do the same thing there (photos are always uploaded without problem) and the when I try again in Firefox problem disappears. If someone had similar problems maybe this could help or maybe this hint could help to someone discovering cause of the problem :)
I just had this error also using ASP.Net MVC and I discovered I just had my handler namespace wrong/missspellled in the web config section where the handlers go. HTH...
After far more time than I'd like to admit, here is the solution. Change the following variables in your php.ini to be large enough to handle your largest file:
; Maximum size of POST data that PHP will accept.
post_max_size = 32M
; Maximum allowed size for uploaded files.
upload_max_filesize = 32M
; Whether to allow HTTP file uploads. (this is on by default)
file_uploads = On
Since files are sent as POST data, both the first variables need to be set. Otherwise, there will be an error, but you won't see what the error is, except IO error #2038.
If you didn't add quote around the option names, try doing that. Sometimes in IE, quoting the option names fixes the issue. It would help if you posted some example code though.
Also make sure that the path to the server being used by the upload component is correct. If the path is a relative path, try to fully qualify it. IE can do some "magic" with relative paths that the other browsers cannot do, and sometimes trying the code from one browser to the other can tweak things.
Also, could be permissions...and having IE go there first could manipulate the permissions in a way that the other browser(s) then work "correctly".
I solved the problem.
In my office I use a proxy and I must change the settings to No Proxy.
Or if you do not use proxy, try to change in php.ini:
Hostname:Port localhost:0
to:
Hostname:Port localhost:80

Resources