Connection refused to quickstart.cloudera:8020 [closed] - hadoop

Closed. This question needs details or clarity. It is not currently accepting answers.
Want to improve this question? Add details and clarify the problem by editing this post.
Closed 2 years ago.
Improve this question
I'm using Cloudera-quickstart 5.5.0 virtualbox
Trying to run this on terminal. As you can below, there is an exception. I've searched for solution to solve this on internet and found something.
1-) configuring core-site.xml file. https://datashine.wordpress.com/2014/09/06/java-net-connectexception-connection-refused-for-more-details-see-httpwiki-apache-orghadoopconnectionrefused/
But I can only open this file readable and haven't been able to change it. It seems I need to be root or hdfs user (su hdfs -) but it asks me for a password which I don't know.

A network configuration is not your problem. You don't need to touch any configurations in the VM, you need to start the services. In this image, for example. The HDFS service on the left is disabled, and I get the same error on that last command.
You have to start Cloudera Manager and start ZooKeeper, YARN, and HDFS (in that order).
To open Cloudera Manager, go to http://quickstart.cloudera:7180 in Firefox on the VM.
Then start the mentioned services.
After you start the services, you can use HDFS commands.

Related

How do I set up Deployer to sync only specified folders from localhost to production? [closed]

Closed. This question needs details or clarity. It is not currently accepting answers.
Want to improve this question? Add details and clarify the problem by editing this post.
Closed 2 years ago.
Improve this question
I have a localhost Laravel project I want to deploy with Deployer. Currently I do it manually using an FTP tool and only sync the app, resources folders and seems to work just fine.
I want to use Deployer or some other tool I can run from terminal to sync or upload new files to the server.
Can someone help with a recipe or advice?
Do I need rsync setup using deployer or is there a way to do it without recipe/rsync.php?
Here are the steps I want configured(for now):
connect to the server, I have ssh access and I can probably configure a key
setup the 2 3 folders I want to sync, as well as files that need to be ignored.
These seem like simple tasks but for some reason I have a hard time setting them up.
Thank you
I don't know if this questions is still pending for answer, but one alternative is using some versioning tool like git, you only watch some folders and ignore the remaining. and with the basic recipe you can deploy a github/gilab/bitbucket project.
A more in dept explanation on this topic can be found [enter link description here]here1.

Files transfer to HDFS [closed]

Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 3 years ago.
Improve this question
I need to bring the files (zip, csv, xml etc) from windows share location to HDFS. Which is the best approach ? I have kafka - flume - hdfs in mind. Please suggest the efficient way.
I tried getting the files to Kafka consumer.
producer.send(
new ProducerRecord(topicName,key,value),
Expect an efficient approach
Kafka is not designed to send files, only individual messages of up to 1MB, by default.
You can install NFS Gateway in Hadoop, then you should be able to copy directly from the windows share to HDFS without any streaming technology, only a scheduled script on the windows machine, or externally ran
Or you can mount the windows share on some Hadoop node, and schedule a Cron job if you need continuous file delivery - https://superuser.com/a/1439984/475508
Other solutions I've seen use tools like Nifi / Streamsets which can be used to read/move files
https://community.hortonworks.com/articles/26089/windows-share-nifi-hdfs-a-practical-guide.html

How to use processor 'GetFile' in Apache NIFI to get the file of network-attached disks [closed]

Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 6 years ago.
Improve this question
I just start using Apache NIFI,what i want to do is using the processor 'GetFile' to get some file from a remoting network-attached machine to my local disks,but i don't know how to configure it in the settings of the processor,and i can not find any documents about this question, Any help is appreciated, thanks.
There is extensive documentation on the Apache NiFi website and within your running instance of NiFi, you can right-click on any processor and select "Usage" to see this documentation inline.
To configure any processor, right-click and select "Configuration", then switch to the "Properties" tab. In GetFile, you need to provide the path to the directory you want to monitor as the Input Directory property, and the file name or pattern you want to retrieve as the File Filter. If this is a specific file known a priori, you can provide a literal name. If it is a pattern (i.e. all CSV files), you can use a pattern like [^\.].*\.csv. You should use the same input path as you would use to browse to the files on the host operating system.

How can you change the file being redirected to while script is still running? [closed]

Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 6 years ago.
Improve this question
Assume I have this script running continually:
node myprogram.js > logfile.log
If I want to make the output dump to a new log file every day without stopping or restarting "node myprogram.js", what should I do?
For example, every day, I want to see the logs saved as 2015-12-01.log, 2015-12-02.log, 2015-12-03.log, etc, and not have the logs be dumped into a single file.
I would use logrotate its the pre-installed utility most linux OS's use for what you are talking about plus more, typical default settings would involve automatically compressing log files of a certain age and then eventually deleting the oldest log files.
The utility runs automatically once a day and performs log rotations as per a configuration you define.
I would prefer this question in server fault sister site. Nonetheless, there are many tools to use. Check out logrotate / rotatelogs.

tomcat 7 instance not starting on ubuntu [closed]

Closed. This question is off-topic. It is not currently accepting answers.
Want to improve this question? Update the question so it's on-topic for Stack Overflow.
Closed 10 years ago.
Improve this question
I'm new to Tomcat, Apache and Ubuntu and i have no idea where to look for logs.
I've followed the article letter by letter of how to create multiple instances of tomcat but ive git a glitch and im not sure how to go about debugging it. Three quarters down the page im supposed to be kick starting the instance but all i get back is [fail], no error message. Im not sure where to look or what i can do to deug this.
https://github.com/getrailo/railo/wiki/Ubuntu-12.04-Railo-4-Beta-with-Tomcat7-multi-instance
Any help appreciated. thank you.
I've finally managed to get where i want and thats to have multiple instances of tomcat all running at the same time. I wouldn't have been able to do this without the article mentioned in these posts, so big thanks to all their hard work.
The reason i was getting [fail] message when trying to start up tomcat, was because i followed the java options steps based on my hardware, that cost me a day!! When reset to its default i was able to start/stop my instances with no errors. I would advise you to use the default (JAVA_OPTS="-Djava.awt.headless=true -Xmx128m -XX:+UseConcMarkSweepGC") then tweak when you have everything running.
If you get to the point where you are getting permission or folder not found error when trying to browse your instance, it is because tomcat7 doesn't have ownership to your working folder (follow this to know how to create your working folder with railo deployment).
Hope this saves someone hours/days.

Resources