My working env. is: Windows 10 Pro Up-to-Date (13/01/2020) and Docker 19.03.5
Trying to map a volume to a folder location from dockers to windows FS is not working as read on this discussion:
https://github.com/docker-library/mariadb/issues/152
So as suggested, I use the docker managed volumes to persists my data. It actually works but what I would like to know if there is a way to explore these volumes that are handles directly by docker.
By now I think that the file where everythins is stored is:
C:\ProgramData\DockerDesktop\vm-data\DockerDesktop.vhdx
Which I tried to open with wome vhdx reader without success. I think it may be possible to mount it but apparently I have not rights to do so. Maybe I'm even wrong on the file name.
Any help is appreciated. Thanks on advance!!
Related
I am designing a python script which stores some configurations in a file when it is first run. I want to make it so that the procedure will almost be same on Ubuntu and windows alike. I noticed Ubuntu has something like ./config file where I can store configurations however I believe there is nothing like that on windows. Whats the best place to store such information such that the path will almost be the same in both os?
Thanks in advance.
I need some straight answers about this as the current docker info and general Web info mixes hyperv and vmware info up.
I have installed docker on my windows 10 pro machine. I do not have vmware/virtual box installed I don't need it I have hyperv? I can use docker on a Linux Ubuntu box fairly well and I (think!) I understand volumes. Sorry for the background...
I am developing a node app and I simply want to have a volume within my Linux container mapped to a local directory on my windows machine, this should be simple but everytime I run my (lets say alpine linux) container '-v /c/Users:/somedata' the directory within the Linux /somedata directory is empty?
I just don't get this functionality on Windows? If you have a decent link I would be very grateful as I have been going over the docker info for two days and I feel I am nowhere!
If volumes are not supported between Windows and Linux because of the OS differences would the answer be to use copy within a Docker file? And simply copy my dev files into the container being created?
MANY MANY THANKS IN ADVANCE!
I have been able to get a link to take place, but I don't really know what the rules are, yet.
(I am using Docker for windows 1.12.0-beta21 (build: 5971) )
You have to share the drive(s) in your docker settings (this may require logging in)
The one or two times I've gotten it to work, I used
-v //d/vms/mysql:/var/lib/mysql
(where I have a folder D:\vms\mysql)
(note the "//d" to indicate the drive letter)
I am trying to reproduce this with a different setup, though, and I am not having any luck. Hopefully the next release will make this even easier for us!
I am newbie in Big Data space. I downloaded cloudera-quickstart-vm-5.7.0-0-virtualbox from Cloudera and got Oracle virtual box where i have to setup remote space. But i am not able to run .vmdk file using VM. I am able to get it using .ovf file. I am not sure what is the issue here. Do i need to change some configuration for running this on machine (network or something else).
Adding error print in my post.
enter image description here
Other users have reported similar problems using the .vmdk file. The workaround, which you have already done, is to import the .ovf file instead. The user experience should be the same, either way.
I was messing with systemd in my Vagrant box and I think I did something wrong since the box won't boot anymore. What I need to do is to remove certain files under /etc/systemd in my Vagrant box. However since I can't boot it, I have to remove those files from the host, but I don't know how to do it.
The VM provider is virtualbox. Both the host and the guest are ubuntu 15.04.
Thanks in advance!
option 1: you can try to add vagrant ssh -c 'cmd to run' in your vagrant file, but as you said it really does not boot, I give it really low chance to success.
option 2: create a new VM and add the hard drive from the old box to the new box (add new hard drive from existing disk), boot it and you might need to mount the new hard drive then you should be able to access your data so make the change in your /etc/systemd or whatever files you need and save. once you have made all your fix, you should be able to reboot from the old vm.
I ended up just creating a new VM but I'm fairly certain that this would work: you just mount the virtual disk file (.vmdk in my case) as a hard drive and make the fix. For different formats of virtual disks you need different tools to mount them, which I'm not going to cover here.
I'm an EC2 beginner. I was able to setup a working EC2 instance for my site. The problem is that I want to use a different AMI (a CentOS one).
I'm wondering what's the exact way to transfer files from one EBS to another EBS?
To be clear, I've researched online and I see that the best way to do this is to mount the EBS to the new instance, and copy the files then. My problem is that I haven't seen any clear, step-by-step instructions on how to do this.
I'm hoping you guys can give me this, as I don't want to mess up my working EC2 instance by using rsync to sync the files in between it and the new instance. (Unless this is an acceptable way to do it, then by all means please let me know)
Thanks!
For transferring files/folders to Ec2 instance, here are the suggested steps
Windows to linux
You can download and install winscp http://download.cnet.com/WinSCP/3000-2160_4-10400769.html, using this application you can transfer the files/folders
Linux to Windows
you can execute the following command,
scp -i key_pairfile_location zip_file_name.zip root#machine_name:target_folder
Hope this helps !