Forgot Amazon EC2 instance [windows 2008 server] password. [closed] - amazon-ec2

Closed. This question is off-topic. It is not currently accepting answers.
Want to improve this question? Update the question so it's on-topic for Stack Overflow.
Closed 9 years ago.
Improve this question
I cannot retrieve the password for my EC2 instance running Windows 2008 server. I want this particular instance and my get password does not work.
Any work around how I can retrieve the password.

Your server is launched from an AMI most probably that is the reason you cannot retrieve your password. It has happened to me in past. In short what you have to do is create an image of the existing instance and then launch a new instance using that image.
create a snapshot with ec2-create-snapshot
attach the snapshot to a new instance.

Related

Bash scripting to copy [closed]

Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
We don’t allow questions seeking recommendations for books, tools, software libraries, and more. You can edit the question so it can be answered with facts and citations.
Closed last year.
The community reviewed whether to reopen this question last year and left it closed:
Original close reason(s) were not resolved
Improve this question
Please share any bash script to run s3 copy commands. I have already tried separate Script to call the rclone command for each agency to backup EC2 instance windows server data to S3.
The command below syncs the current directory to an S3 bucket using a named profile.
aws s3 sync . $S3_BUCKET_URL --profile $YOUR_PROFILE_BRO
The next command syncs the S3 bucket to the current directory using a named profile.
aws s3 sync $S3_BUCKET_URL . --profile $YOUR_PROFILE_BRO
The next command copies a file (file.txt) from your machine to S3 using the default profile of your machine.
aws s3 cp file.txt s3://my-bucket/
Of course, you need an aws-cli and AWS credential pair (secret key & secret key id) to make this work from your machine or on-premise networks. If you want to copy from EC2 to S3, you can assign IAM roles that posses permissions to write/read objects to that EC2 and you should be good to go.

Install Ansible on a provisioned machine using Terraform [closed]

Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 1 year ago.
Improve this question
I just provisioned an EC2 instance using terraform.
Now I'd like to install ansible on said machine in an automated way using terraform.
Is there a way to do so?
Just want to clarify:
Usually there is no need to install ansible on a machine after provisioning it since there isn't a daemon or agent required to be able to execute ansible tasks on a remote machine.
The usual workflow would be to provision machines with terraform and then configure them via ansible from a remote machine that can reach the newly created ones. So it only ever needs one machine that has ansible installed (which could as well be your local machine in case you can directly connect to the provisioned instances).
However, to answer your question on how you can install ansible (or anything else) on a machine that you just provisioned with terraform:
There is a provisioner called remote-exec: https://www.terraform.io/docs/language/resources/provisioners/remote-exec.html
Just use it within your instance declaration.
It comes with an argument called inline which you can use to write down your installation commands. There are many different ways of installing ansible, depending on your setup some of those may not even work, therefore I cannot tell you which one is the best for you.
However, to give you an example of how to use the provisioner:
resource "aws_instance" "my-instance" {
instance_type = "t3.micro"
# ...
provisioner "remote-exec" {
inline = [
"apt update",
"apt install <package>",
]
}
}

Can I look at files my code makes on heroku? [closed]

Closed. This question needs details or clarity. It is not currently accepting answers.
Want to improve this question? Add details and clarify the problem by editing this post.
Closed 2 years ago.
Improve this question
I have a discord bot that saves JSON files on the dir he is in so it could work on more than one server without colliding.
I finished my code and I uploaded it to heroku for hosting. The thing is , when I ran the code locally I could see the files that were being created for each server for testing but now I don't know how to reach them.
Is there a way to check all the files I have in heroku without downloading everything down?
You can install Heroku CLI and then access your files using:
heroku login
heroku run bash -a APPNAME
But be aware that Heroku uses an ephemeral filesystem. This means that your local filesystem is only accessible to a single dyno, and once the dyno is stopped, restarted, or moved, all files on the local filesystem are destroyed.
You could use a service like Amazon S3 to store your files in a more permanent way.

How to fix this Error "The following arguments are required: command" [closed]

Closed. This question needs details or clarity. It is not currently accepting answers.
Want to improve this question? Add details and clarify the problem by editing this post.
Closed 2 years ago.
Improve this question
I am beginner on AWS. I am working on Windows environment. Anybody knows "How to fix the error below"?. I tried to use "aws s3 la" and "aws ec2 describe-intances", but this doesn't fix the error.
"aws s3 la" and "aws ec2 describe-intances"
The commands don't work because they have spelling mistakes. They should be:
aws s3 ls
and
aws ec2 describe-instances

Is it possible to access S3 objects from on-prem servers with automatically updating AWS credentials? [closed]

Closed. This question is opinion-based. It is not currently accepting answers.
Want to improve this question? Update the question so it can be answered with facts and citations by editing this post.
Closed 2 years ago.
Improve this question
I have a list of zip files stored in an S3 bucket that I would like to access from my web app which runs on an IIS server, outside AWS (on-prem). What would be the best way to handle the AWS credentials (I have a total of 4 environments - 4 IIS servers I need to maintain).
I know that I can manually set up an AWS credential profile (using an IAM user) using the AWS CLI on each environment but I was wondering if there would be a better way to handle the AWS credentials that are stored on the servers. For example, our organization has an IAM user policy that expires the credentials every 90 days. That means that there would be a bit of maintenance overhead requiring me to update the IAM user credentials all the machines every 90 days.
Is there a way to automate the above process so that the IAM user credentials are updated automatically?
You would need to create such a process yourself. There is nothing within AWS that can update credentials stored on your own computers.
IAM Users can have two active credentials, so the process would be:
Generate second set of credentials on the IAM User
Update credentials stored in your applications
Disable/delete the first set of credentials from the IAM User
This gives some time for applications to update to the new credentials while the old credentials are still valid. These steps can be automated via API calls to AWS, but your own application would need to initiate the steps and update the credentials on your servers.

Resources