Devilbox Auto DNS on Windows - windows

I started using devilbox (https://github.com/cytopia/devilbox) which gives me a great and really easy to setup LAMP server.
Following the docs for Windows Auto DNS, you should make 127.0.0.1 your default DNS of your "active network adapter". However:
Setting DNS in my Wireless Adapter works but leads to losing Internet Access when docker/devilbox isn't running (as the DNS Server isn't running at this point)
Setting it on vEthernet Adapter (Hyber-v Virtual Ethernet adapter, probably created by Docker) works too but resets itself upon restarting
Setting it within Docker > Dashbord > Settings has no effect at all (probably because it is only for within docker and not on the
host)
I can eighter set the alternate DNS in Solution 1 to e.g. 8.8.8.8 OR I have to edit DNS the vEthternet everytime I start devilbox. I therefore highly suspect that I'm missing something, and as the official docs are thin on this topic:
How can I setup Devilboxes DNS for Windows permanently?

Related

Cannot access other machines on remote network with OpenVPN

I am trying to set up OpenVPN so that I can access machines inside an Azure subnet from my pc which is outside Azure.
I have successfully installed OpenVPN on both server (Windows Server 2019) and pc (Windows 10) using the instructions here: https://community.openvpn.net/openvpn/wiki/Easy_Windows_Guide?__cf_chl_jschl_tk__=pmd_889e3e419b8b865ffd4da6e493bef6df0782273e-1629275604-0-gqNtZGzNAfijcnBszQgi, and I can successfully connect from client to server, however, I cannot connect to any other machine on the Azure subnet upon which the server is sitting.
The server and the other machines I want to connect to are on a 10.0.0.0 subnet, and the VPN is coming up on the 10.8.0.0 network as I would expect from the examples.
I have enabled IP routing on the server as recommended in the OpenVPN FAQ but this has not fixed the issue.
I have also added a 'push "route 10.0.0.0 255.255.255.0"' line to the server config, and I can see from the client log (and the client routing table) that this has been executed, but I am still unable to connect to other machines in the subnet.
I was looking into using Tap instead of Tun, but when I dug into at what was actually being used, it looks as if as if both ends are using the Tap adaptor anyway, even though I have specified 'dev tun' in both the client and the server configs.
I have had bit of a trawl but can't find anything about the Tap adaptor when the Tun adaptor has been configured, so that is a bit of a mystery.
The only other thing that I have read is that it might be necessary to set up a route back to the OpenVPN subnet on the gateway server for 10.0.0.0, but that's not a server I control as it's part of the Azure infrastructure.
What do I have to do to get access to other machines on the 10.0.0.0 subnet? And why is the Tap adaptor being selected despite the config specifying the Tun adaptor ?
I made a number of other changes before I finally got it sorted out - I do not know if they were all necessary but in addition to the above:
I changed 'dev tun' to 'dev tap' in the server and client configs.
I followed the instructions here NAT-hack to add NAT to the server.
And finally, I added 'route 10.0.0.0 255.255.255.0 10.8.0.1' to the
server config file.

Hyper-V - No Guest Internet

I'm using a Windows Server 2019 server from Microsoft Azure. I have installed Hyper-V with the management tools and am now trying to setup a Ubuntu 14.04LTS VM inside of Hyper-V. I create a new External Switch however when setting up the Ubuntu instance there it tells me that there is an issue with the DHCP server. When I check my network connections for the newly created switch there is IPV4 connectivity, with packets being sent and received.
Things that I have tried:
-Ensured that inside of the Virtual Switch Manager that in my newly created virtual switch, the external network is selected as my main internet adapter. And that the "Allow management os to share this network adapter".
-Tried selecting "Internal Network"
-Inside of the ubuntu server, tried dchlient -r and dhclient eth0 to try to retrieve a new IP (For internal network this has worked, but without internet access, for external ip it hangs on the command dhclient eth0)
-In Network Connections I have tried bridging the two connections (out of desperation, "bricks" the vps causing me to not be able to rdp, must create a new azure vm)
-I have tried right clicking the main "ethernet" inside the Network Connections and allowing sharing options to other users.
Please help as I'm quite lost as to why the Ubuntu Guest is not connecting to the internet with the new network switch
The virtual switch connection types are very confusingly named, and sharing is somewhat flaky. I've had best results with the Internal Network. Most of the time it just works but there are certain situations when the NAT service breaks and you will have to restart some combination of the host and guest machines, possibly both -- I've not found a way to just restart the virtual switch service without restarting the host OS.
If you need to use the External Network type, be aware that your actual physical router will be in the mix, so you must make sure that it is properly configured, especially if you're doing MAC filtering on your router and the guest is not using the hardware MAC. This usually happens because of the virtualisation process itself, even if it's not something you've configured in the guest.

vmWare Workstation External Accessibility Issue

I'm running Windows Server 2012 w/ vmWare Workstation. I've built a GitLab VM on Centos 7 that's totally setup and accessible on my local network. It's configured using Bridged Mode so it has it's own IP from the DHCP Server.
I use No-IP to connect to my Network externally which has been working great for several years now. I have port-forwarding setup within my router to forward traffic for the GitLab webUI to the GitLab VM, but it's not accessible externally. I even tried setting up the port forwarding to direct the traffic to the Windows Server and then setup internal port forwarding w/ netsh on the Windows Server to forward the traffic to the GitLab VM, making sure I opened the port on the Windows Firewall (even tried disabling it), but I still can't get to the GitLab VM externally. AFAIK running a VM w/ a Bridged adapter should essentially be like it is just another physical machine on the network.
Now, I am running IIS on the Windows Server, but when I specify a specific port using my public No-IP Domain, the router should detect the traffic on that port and forward it according to the rules that I have setup, correct? IIS shouldn't be interfering with any traffic on other ports with the external Domain.
I'm totally stumped on this on and searching around the web really hasn't helped much.
So it turns out that I did everything 100% correctly with setting up port forwarding right to the IP of the VM, but my workplace blocks just about every port except for 80 and 443. Tested connectivity from an AWS box and everything is accessible exactly as designed.
Now I just feel like an idiot, but hey, I figured it out.

vmware: unable to ping using NAT IP address, but can ping physical IP

I've been searching around and found no similar issues have been asked.
My desktop (windows 7) is in LAN, has IP (192.168.3.121).
I installed vmware (windows 7), using NAT connection (172.168.174.128). The guest is able to access internet without problem.
In guest, if I "ping 192.168.3.121", this will be ok.
If I "ping 192.168.3.xxx", will also be ok.
If I "ping 192.168.174.2" (DNS), this will be ok.
If I "ping 192.168.174.1", can't ping.
Note my host Vmnet8's ip is indeed 192.168.174.1. Ping from host also doesn't work.
From host:
"ping 192.168.174.128" (guest IP), does not work
"ping 192.168.174.2", does not work
Both host/guest windows are installed without any other "security/defender/firewall" related softwares.
Anywhere can go wrong?
Thanks.
If you configured your vmware VM to run in a virtualized NAT network, then you will not be able to access/ping your VM from the Host, or anywhere else for that matter, without configuring port forwarding for that virtual NAT network.
If you would like to be able to access your VM from your host you can either:
A) Change the mode of the network adapter for the VM to a bridged adapter. This will make the VM act as if it is just another computer on the same network your Desktop is and will be accessible at an IP such as 192.168.3.122
or
B) Add a Host-Only network adapter to the VM. This second NIC will be connected to a network that has no internet access, but is connected to the host and any other VMs on the same host-only network
Also, check the firewall settings to allow inbound ICMP inside the VM.
You mean that you cannot ping to the VMnet8 interface of your physical PC.
Maybe it is not activated.
It should be activated first by issuing the command at the cmd prompt with the admin's privilege.
C:\Windows\system32>netsh interface set interface name="VMware Network Adapter V
Mnet8" admin=ENABLED
I have seen this issue with two different windows 10 machine & two different version of vmware workstation ( 15 & 16).
One way it works is I start the wireshark & under capture options I select on VMWARE8(default for workstation/need to adjust according to your NAT Interface) & than start ping from My Local Machine to NAT IP of the VM .
It takes time but it works. I do not what triggers this .
My initial thought was it's one of the Windows 10 upgrade but with two different version of windows 10 & this old issue resurfaced.
Navigate to "Control Panel\Network and Internet\Network Connections".
Disable and Enable the VMware Network Adapter VMnet8 and try again.
I was having same problem after the latest release of VirtualBox-6.1.6-137129-Win update. "NAT Network" on any of the guest machine was not working. So i downgraded my virtual box version to 6 and works fine with all my VMs but problem remains in Kali Linux.
NAT Network was successful in giving IP to kali machine but Internet was not working. Problem i found was somehow gateway of kali wasn't set. Then i configured both Gateway and DNS manually and it worked for me.
Make sure that you Uncheck the option "Use this Connection only for resources on this network."
Kali manual configuration for IP, DNS and Gateway

Google Compute Engine IIS Webfarm

I'm trying to setup a Win2008R2 IIS webfarm on Google Compute Engine.
I've got the machine setup, however when I try to add it to a network load balancer pool, the balancer consistently reports the machine as unhealthy - even if i disable healt checks. I have a single forward rule setup for port 80.
I've tried different size instances in different regions/zones to no avail. Traffic into the load balancer never makes it to my instance, and the instance is always report as unhealthy.
For the firewall I went ahead and added a blanket rule so 0.0.0.0/0 can access all local net services (ICMP;TCP:1-65535; UDP:1-65535) and I've disabled windows firewall.
Anyone have any experience getting this working?
Spoke with google support. "Known issue with windows instances - check back in 6 months." In the mean time, use linux or setup your own NLB within your project.
Strange that it is not working for you. I replicated your situation and I am getting to the machine with no issues. The load balancer is forwarding traffic as expected and it reaches the system who is marked as healthy in the Lb pool.
You may want to add the following rule to the windows firewall with advanced security(make sure you use the "advanced security" one and not the default):
Inbound rule > New port > port 80
Once this is done, from your machine you can curl or telnet to the address while running a netstat on the Windows system and you should see the LB forwarding rule IP making requests :
$ curl IP (locally)
$ netstat (on the windows machine)
Hope this helps !

Resources