Remote project with Zend Studio editor - zend-studio

I have just start using Zend Studio - 7.2.0 and i have some confusion.
I have attached one image where you can see working with remove server need a connection to that server, i have done it successfully but now when i download the remote project, it show two sub level 1. My Home and 2. Root ,
Can any one tell me what is the difference between theses two and what is there use.
Thanks

That simply means that the home directory is the same directory as the lowest directory that the SFTP connection can go. If your home directory was /home/user, My Home would go there and Root would go to /

Related

How to run projects when xampp and project folders are in two different computers?

I want to place and run my project from the second computer when xampp is installed on first computer, is it possible to map between xampp and my project which are present in two different pc's. If it is possible how to achieve that?
Yes, We can do that. After a lot of research I got it. If we want to access the project from any other external path(shared network) rather than our PC general location where xampp is located, do the following.
1) Connect to a shared network by mapping the path in your PC connecting to a shared folder
If you are using windows7 go to file explorer and check "Map Network Drive" on the top.
2) After that go to xampp control panel-> apache-> config-> search for "c:/xampp/htdocs"(where you are accessing your previous project) -> now change it to the new path where your project currently hosted/placed(shared network/any other folder).
3) stop the sevices and restart the xampp that's it you can now access your project folder which is located in different folder/server.

PhpStorm Project Settings per User?

In my company 3 developers use PhpStorm for developing WebApps.
Everyone needs to open (not simultaneously) a project an other user created from time to time.
Let's say user A uses a deployment server named server A and user B uses server b.
User A created the project in the deployment.xml server A is saved.
Now user B needs to open the project and creates server B. Now server B is saved in deployment.xml
If user A want to reopen the project again, he has to set the server back to server A.
Same goes for the workspace.xml.
My question is if it possible to use different project settings for user A and user B?
If it is possible what do I need to change in my settings?
If not how can 3 users work together (no VCS)?
Thanks in advance
In the case where no settings are to be shared between developers, only the project files, PhpStorm accepts reading project files from a directory outside of the classical 'project directory'. Here is one approach to that; each developer is to:
Create and open a new empty project from PhpStorm, in a folder on the local hard drive. Only project settings will be stored here.
Open the Settings dialog, choose 'Directories' from the left pane, then click 'Add Content Root' (at the top of the rightmost pane).
Choose the folder on the network share that contains the project files.
Remove the original 'content root' corresponding to the project settings folder on the local hard drive.
Each developer can now open their locally created directory in PhpStorm, and any files opened/created will exist on the shared folder.
A related question: PhpStorm - How can I avoid creating the .idea folder?
Please note that working with files that are not on a local filesystem, especially when they are editable by multiple developers, is not recommended. Even for small projects, storing a Git repository on the network share, with each developer git pulling when beginning work and git pushing when done, is a small investment that will pay large dividends. Here is a good introduction: http://www.gitguys.com/topics/creating-a-shared-repository-users-sharing-the-repository/, but the key command is git init --bare reponame in the setup phase.

Magento upgrade process - archiving extensions

I'm attempting to upgrade a Magento 1.7.0.2 website to 1.8.1.0 following the process given by Magento on their website: http://www.magentocommerce.com/knowledge-base/entry/ce18-upgrade-roadmap?icid=BLOG_magentoce1810release_upgradeinstructions
I follow this up to step 2b on the above link where it says:
Archive the file system.
This includes the media directory and subdirectories; all extensions and customizations; and all custom themes.
I understand how to make a zip of the media directory and theme directory files. However, I don't understand how to archive/move the extensions? We have about 10 extensions installed on the website we're upgrading; some were installed by copying zip files on to the webspace, whereas others were installed via Magento Connect.
The files for the extensions are all over the webspace (in line with Magento's huge directory structure) so I don't see how it's possible to identify all of these files and move them across.
Does anyone know what the best way of doing this is?
The steps on the Magento blog are a little odd. Does anybody really do it that way?
Surely the way to do it is:
0. clone your entire production system onto a new server - call it the back up
1. clone your entire production system onto another new server - call it the dev server
3. Via Magento Connect, upgrade the dev server to 1.8
4. Download all the Magento 1.8 files to your PC and then upload them all to the dev server just to be sure all the files are in place
5, Run lots of tests and iron out any problems
6. Either switch your store URL to point to the dev server or put your live store in maintenance mode while you repeat steps 1-6 on the live server; if it screws up see step 0.
Your point is correct : it is very hard to extract the modules which is why I say always build up from your current state rather than trying to build up form a fresh Magento 1.8.
The way I did it eventually was following steps similar to these:
http://www.magentocommerce.com/boards/viewthread/283184/
The point I was missing was the directories that needed manually moving to the "upgrade" version of the website (the "dev server" Malachy's post).
Steps were:
Install a clean copy of 1.8 on a dev server.
Make a copy of the live server (1.7) database on to the dev server.
Move these folders from the live server:
app/design/frontend/mytheme
app/etc/local.xml
app/etc/modules
app/code/community
app/code/local
js
skin/frontend/mytheme
media
Change app/etc/local.xml to reference the 1.7 database, but on the dev server.
Change the base_url (nonsecure & secure) settings in the config table of the DB.
Go to the URL of the dev website, wait for the upgrade scripts to run.
Test / fix - repeat as much as required.
Put the dev copy of the site live.
Caveat with this is that if your live site is getting orders frequently, you'll not get a copy of that as your database will become out-of-sync as soon as you copy it during step 2.

Path too long error when building a windows azure service

I have been trying to publish my service to windows azure. The service consists of a single webRole, however I have added remote login functionality published it and built it a few times, and now all the sudden it will not build. The reason it gives is that
Details below:
"Error 56 The specified path, file name, or both are too long. The fully qualified file name must be less than 260 characters, and the directory name must be less than 248 characters. C:\Program Files (x86)\MSBuild\Microsoft\Cloud Service\1.0\Visual Studio 10.0\Microsoft.CloudService.targets 202 5 FileSystemCreator"
I have gone on all the forums, I have used CSPack command line for packaging the service which is fine but I'm having a really hard time configuring the certificate for remote desktop connect and I would like to take advantage of this feature as I am creating some websites in the onStart event and I would like to peek into IIS. Some microsoft employees do agree that this is a bug and the have promised a fix this issue, refer to post . I am using VS2010 and I do not know how to fix this bug.
Can anyone please help, or point me to a place where I can get any help.
I ran into the same problem with a new solution.
Note that, unlike Eugenio Pace's response suggests, the error occurs only when deploying to Azure (and not when running the project in the Azure Compute Emulator).
Try adding the following line to the first property group of your Windows Azure Visual Studio Project file (*.ccproj):
<ServiceOutputDirectory>C:\Azure\</ServiceOutputDirectory>
The trailing slash (for whatever path you select) appears to be required. This folder will be deleted each time you create a package if it exists.
This setting seems to redirect the working folder for the package to a shorter base path, preventing the path too long error.
Credit goes to: http://govada.blogspot.com/2011/12/windows-azure-package-build-error.html
Perhaps the local folder used to store temporary development fabric is too long. See Windows Azure - Resolving "The Path is too long after being fully qualified" Error Message.
I was having this problem as well when deploying a Node.js project to Azure.
To fix it, I had to change my "TEMP" and "TMP" user environment variables to something shorter than their default values.
In my case, they were pointing by default to %USERPROFILE%\AppData\Local\Temp, changing them to C:\Temp solved it.
Make sure you restart Windows after.
The better solution may be to create a symbolic link to your project folder. This doesn't require moving files or changing system variables. Open up the command prompt as an administrator and run this:
mklink /D C:\Dev C:\Users\danzo\Source\Workspaces
Obviously you can change "C:\Dev" to whatever you want it to be and you'll need to change the longer path above to the root directory of your soltions/projects folder.
Same problem happened to me when I try Packaging an Umbraco project for Azure (https://github.com/WindowsAzure-Accelerators/wa-accelerator-umbraco/wiki/Deployment), I found the solution is to: Copy and rename the long-name path and folder to "C:\someshortname".
(solution was suggested by this: link)
I tried all the above 2 approaches:
-change TEMP and TMP enviromental variables
-<ServiceOutputDirectory> path
and didn't work.
In my case, I had to move the whole project to a shorter path C:\ and worked.
I'm using W7 and VS12.
When you run a cloud service on the development fabric, the development fabric uses a temporary folder to store a number of files including local storage locations, cached binaries, configuration, diagnostics information and cached compiled web site content.
By default this location is: C:\Users\\AppData\Local\dftmp
Credit goes to Jim Nakashima of Microsoft :
https://blogs.msdn.microsoft.com/jnak/2010/01/14/windows-azure-resolving-the-path-is-too-long-after-being-fully-qualified-error-message/
In order to change the temporary folder, a user environment variable has to be created :
It is named _CSRUN_STATE_DIRECTORY
Give it a value of short named directory like :
c:\AzureTemp
Don't forget to restart Visual Studio in order to have the environmennt variables to be read again
It fixed many compilations problem !

Notepad++ with Local and FTP synchronizer. [?]

Does the Notepad++ have Local and FTP synchronizer by any plugin?? because I develop websites using PHP and notepad++ has all the features I like and its really lightweight but I had to switch to Netbeans because I use a web hosting but I always like to save the code in my computer too. and netbeans can do that, even anything you insert locally in the folder it automatically adds the folder and the files in the FTP server which is great. but if the notepad++ has the feature to at least update the files that we are saving in notepad++ in both local and ftp server I would be so glad, I search that for a long time, but I can't use netbeans anymore I lose way too much time, netbeans is really heavy!
Thanks!
NppFTP: a plugin that allows FTP,
FTPS, FTPES and SFTP communications.
Very useful for web development.
Author: harrybharry
Homepage: http://sourceforge.net/projects/nppftp/
Install it from Plugin Manager
These are NppFTP plugin panel and toolbar button
Open profile settings dialog
Then configure profiles
Just wanted to post this here for anybody looking for the same solution I was looking for... (and I think helps answer this question more thoroughly).
I keep an exact replica of my public_html directory on my local machine. I wanted to be able to double click a file on the remote server and live edit so that I had a mirrored copy on my local machine. Note: if you are looking for functionality similar to Dreamweaver's site manager... there is a feature request for that. This solution only allows your local files to get updated when you edit a remote file.
So here goes the basic connection settings (pretty standard):
h: some.ftphost.com
u: some_ftp_user
p: a_very_secure_password
d: /public_html
Then, here is where the magic comes in. Under the "cache" tab for the ftp profile, add the following:
Local path: E:\Path\to\your\local\server\public_html
External path: /public_html
The external path should be the same as the "initial directory" in your connection settings. Hope this makes sense. Please ask questions if you have any.
Then what you need is rather FTP_synchronize
double-clicking file will open it for
editing and saving file (in usual way)
will update it on server .
I actually prefer the way that Notepad++ works with FTP compared to Netbeans. Notepad++ always treats the remote file as the master copy. So when you open it, it first downloads it and stores it in the local cache. Netbeans however always opens the local copy first - you have to explicitly synchronise with the external server to pull down the files from the server. If you're working with other developers - its much better to use the server copy so that you pull down any changes by other developers.
If you want the synchronisation try these steps:
In Notepad++ | NppFTP | Global Settings | Set the Global cache to be C:\inetpub\wwwroot\%USERNAME%#%HOSTNAME% which works for IIS or change the directory to your webserver root directory
You will have to make sure Notepad++ has permissions to create directories in your server root
Download one file from the server using NppFTP so that you can see what the directory struction looks like you can probably put just %HOSTNAME% e.g. C:\inetpub\wwwroot\domain.com
Then use Filezilla to download all the files into that directory - you can also use Filezilla to check for synchronisation changes.
Then use NppFTP which will download the files into that structure.
You should then be able to access the files through localhost/domain.com
If you're the only one working on the project that should then be enough, but if there are others, or if you make changes elsewhere you can use Filezilla to check the file timestamps to synchronise.
The default install for Notepad++ has a Plugin called NppFTP on the Plugins menu. I don't know how feature-full it is, however.
Get NppFTP
Connect to ftp.xxxx.com
Double click file to open
Edit changes
Save with automatic upload

Resources