Closed. This question needs details or clarity. It is not currently accepting answers.
Want to improve this question? Add details and clarify the problem by editing this post.
Closed 7 years ago.
Improve this question
I'd like to setup and use Laravel as a PHP framework. I faced with Homestead.
I don't know what is this.
is this essential for Laravel and the projects that i make with it ?
Sniping the answer here, just in case others have the same question:
Homestead is nothing more than a Vagrant Environment
What Vagrant does is spin up virtual servers on your local machine that you may then interact with as if they were remote servers. The idea behind this is two-fold:
Vagrant Environments can be tailored specifically to mirror - or come close to matching - the production environment on which your web application will be hosted.
Vagrant files can be version controlled so that team members working on the same project are guaranteed to be working in the same environment.
With all that said, all Homestead is is a Vagrant box that has a basic LEMP stack and a bunch of other server components configured for you out of the box that will enable you to quickly provision new sites and databases through a simple YAML configuration.
While it would certainly be easier to spin up a Laravel/Lumen installation for local development within a Homestead environment, you could really do it within any kind of virtual machine or locally installed HTTP server.
If you really wanted to, you could also used PHP's built-in server to quickly spin up a Laravel application as well by running php artisan serve in a Laravel 4 application, or php -S localhost:8000 -t public in Laravel 5.
But ultimately, no, Homestead is not a necessity. Only a nicety.
Related
Closed. This question needs details or clarity. It is not currently accepting answers.
Want to improve this question? Add details and clarify the problem by editing this post.
Closed 5 days ago.
Improve this question
The official documentation for deploying TYPO3 states "The composer install command should not be run on the live environment."
Why is this? As far as I know deployer - a well received deployment tool for TYPO3 - does exactly this.
TYPO3 Recommends Composer as the Main Installation method.
but its good practice to run "composer install" on your local machine (or a CI Pipline like Github Actions) then move the downloaded code to the final destination server. and i think deployer is dooing exaclty that.
this has multiple advantages:
less dependancies on the production system. (one less software there to be exploited)
possiblity to run unit tests before deployment
add additional build steps for your deployment (like compilin sass
or javascripts)
ther is no need for credentials of private repositories on the production server
Closed. This question needs details or clarity. It is not currently accepting answers.
Want to improve this question? Add details and clarify the problem by editing this post.
Closed 2 years ago.
Improve this question
I have a localhost Laravel project I want to deploy with Deployer. Currently I do it manually using an FTP tool and only sync the app, resources folders and seems to work just fine.
I want to use Deployer or some other tool I can run from terminal to sync or upload new files to the server.
Can someone help with a recipe or advice?
Do I need rsync setup using deployer or is there a way to do it without recipe/rsync.php?
Here are the steps I want configured(for now):
connect to the server, I have ssh access and I can probably configure a key
setup the 2 3 folders I want to sync, as well as files that need to be ignored.
These seem like simple tasks but for some reason I have a hard time setting them up.
Thank you
I don't know if this questions is still pending for answer, but one alternative is using some versioning tool like git, you only watch some folders and ignore the remaining. and with the basic recipe you can deploy a github/gilab/bitbucket project.
A more in dept explanation on this topic can be found [enter link description here]here1.
Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
This question does not appear to be about a specific programming problem, a software algorithm, or software tools primarily used by programmers. If you believe the question would be on-topic on another Stack Exchange site, you can leave a comment to explain where the question may be able to be answered.
Closed 7 years ago.
Improve this question
I have a Azure VM running Win server 2012 r2 datacenter. I have installed Active Directory Domain Services Role onto that VM. Now i want to Clone it with the same ADDS settings to create new VMs.
I tried Sysprep, but later i found out that sysprep doesnt support ADDS.
Is there any possible way ?
Thank You.
You can't. You have to install and promote each domain controller in your forest.
However, using powershell + Azure Custom Script VM Extension you can automate the process.
Custom Script extension for Windows allows you to run PowerShell
scripts on a remote VM, without logging into it. The scripts can be
run after provisioning the VM or any time during the lifecycle of the
VM without requiring to open any additional ports on the VM. The most
common use case for Custom Script Extension include running,
installing, and configuring additional software on the VM after it's
provisioned.
More info:
Custom Script extension for Windows virtual machines
Step-by-Step: Auto-join a VM to ADDS in the Azure Cloud
Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 8 years ago.
Improve this question
I have to install php application in a one computer and access these application from others computer in a LAN. is it possible?
Yes. It is possible.
Install PHP and configure the PHP application on one computer on the LAN. Let's call this machine myserver.local. Then test out the application on the same workstation to make sure it's working fine. You'd probably be using a URL like this:
http://localhost/appname/index.php
Once you're sure that it works fine, then you can use the following to access the application from the other computers in the LAN:
http://myserver.local/appname/index.php
Hope this helps.
Yes you can.
If the computer you are hosting the php application is a part of the intranet then all you have to do is send the application link or the ip to others and they can access the application.
Yes you can,
You can use the server IP Address(IP of machine where server software's are installed) to access the content on remote machine.
Closed. This question is off-topic. It is not currently accepting answers.
Want to improve this question? Update the question so it's on-topic for Stack Overflow.
Closed 10 years ago.
Improve this question
I am trying to install project management application called Redmine on Windows Azure Virtual Machine and access it from internet. Redmine is based on Ruby and is using a database so I also had to install them too. Actually I managed to install everything and when I connect to virtual machine by Remote Desktop I can access Redmine there by typing localhost/redmine in browser. The question is what should I do now to access it from outside? (for example home pc or anybody in internet). My virtual machine has got an address like myapp.cloudapp.com but there is no way I can access redmine by typing myapp.cloudapp.com/redmine or something like this.
Well, I can't exactly speak for Azure as I haven't used it, but there should be a way to forward port 80 through to your VM. You should also make sure IIS is setup to listen on 0.0.0.0:80 and not 127.0.0.1:80
Edit: Usually port forwarding is setup in the control panel for a cloud provider, so I would start by checking whatever interface Azure gives you to configure your VM.