Print the characters from string - bash
I have a file which contains words-
abfiuf.com abdbhj.co.in abcahjkl.org.in.2 abciuf zasdg cbhjk asjk
including other contents. My Requirement is -The word which starts with abfiuf,
abdbhj, abcahjkl, abciuf ,.... cut the two character from middle like below.
abfiuf - fi
abdbhj - db
abjcahjkl - ca
abciuf - ci
I have tried below-
First to get the matching word-
cat /etc/xyz.txt|grep -Eo \<(abfiuf|abdbhj|abjcahjkl|abciuf)\S*'|cut -f1 -d"."
But unable to cut before and after matching "fi" , "db" , "ca" , "ci" from words .
Tried with sed command - sed 's/^.*fi/fi/' -> working only for one word removing before. But How to cut multiple char before & after from words ?
EDIT2: Since OP told he/she only want to print matching strings value only if that is the case then one should try following.
awk 'match($0,/fi|db|ca|ci/){print substr($0,RSTART,RLENGTH)}' Input_file
OR in case you want to print a message with line number that DO NOT have any match found then try following.
awk 'match($0,/fi|db|ca|ci/){print substr($0,RSTART,RLENGTH);next} {print "Line number " FNR " is NOT having any matching value in it."}' Input_file
Assuming that you need to print only 3rd and 4th character if that is the case then try following.
awk '{print substr($0,3,2)}' Input_file
EDIT: Now I am assuming that you DO NOT want to hard code the position to print from lines if that is the case then try following, which will first calculate the length of line and will print 2 characters starting from its middle letter.
awk '{len=length($0)/2;print substr($0,len,2)}' Input_file
Related
How do i get the value present in first double quotes?
I'm currently writing a bash script to get the first value among the many comma separated strings. I have a file that looks like this - name things: "water bottle","40","new phone cover",10 place I just need to return the value in first double quotes. water bottle The value in first double quotes can be one word/two words. That is, water bottle can be sometimes replaced with pen. I tried - awk '/:/ {print $2}' But this just gives water I wanted to comma separate it, but there's colon(:) after things. So, I'm not sure how to separate it. How do i get the value present in first double quotes? EDIT: SOLUTION: I used the below code since I particularly wanted to use awk - awk '/:/' test.txt | cut -d\" -f2
A solution using the cut utility could be cut -d\" -f2 infile > outfile
Using gnu awk you could make use of a capture group, and use a negated character class to not cross the , as that is the field delimiter. awk 'match($0, /^[^",:]*:[^",]*"([^"]*)"/, a) {print a[1]}' file Output water bottle The pattern matches ^ Start of string [^",:]*:Optionally match any value except " and , and :, then match : [^",]* Optionally match any value except " and , "([^"]*)" Capture in group 1 the value between double quotes If the value is always between double quotes, a short option to get the desired result could be setting the field separator to " and check if group 1 contains a colon, although technically you can also get water bottle if there is only a leading double quote and not closing one. awk -F'"' '$1 ~ /:/ {print $2}' file
With your shown samples, please try following awk code. awk '/^things:/ && match($0,/"[^"]*/){print substr($0,RSTART+1,RLENGTH-1)}' Input_file Explanation: In awk program checking if line starts with things: AND using match function to match everything between 1st and 2nd " and printing them accordingly.
Solution 1: awk You can use a single awk command: awk -F\" 'index($1, ":"){print $2}' test.txt > outfile See the online demo. The -F\" sets the field separator to a " char, index($1, ":") condition makes sure Field 1 contains a : char (no regex needed) and then {print $2} prints the second field value. Solution 2: awk + cut You can use awk + cut: awk '/:/' test.txt | cut -d\" -f2 > outfile With awk '/:/' test.txt, you will extract line(s) containing : char, and then the piped cut -d\" -f2 command will split the string with " as a separator and return the second item. See the online demo. Solution 3: sed Alternatively, you can use sed: sed -n 's/^[^"]*"\([^"]*\)".*/\1/p' file > outfile See the online demo: #!/bin/bash s='name things: "water bottle","40","new phone cover",10 place' sed -n 's/^[^"]*"\([^"]*\)".*/\1/p' <<< "$s" # => water bottle The command means -n - the option suppresses the default line output ^[^"]*"\([^"]*\)".* - a POSIX BRE regex pattern that matches ^ - start of string [^"]* - zero or more chars other than " " - a " char \([^"]*\) - Group 1 (\1 refers to this value): any zero or more chars other than " ".* - a " char and the rest of the string. \1 replaces the match with Group 1 value p - only prints the result of a successful substitution.
Searching for a string between two characters
I need to find two numbers from lines which look like this >Chr14:453901-458800 I have a large quantity of those lines mixed with lines that doesn't contain ":" so we can search for colon to find the line with numbers. Every line have different numbers. I need to find both numbers after ":" which are separated by "-" then substract the first number from the second one and print result on the screen for each line I'd like this to be done using awk I managed to do something like this: awk -e '$1 ~ /\:/ {print $0}' file.txt but it's nowhere near the end result For this example i showed above my result would be: 4899 Because it is the result of 458800 - 453901 = 4899 I can't figure it out on my own and would appreciate some help
With GNU awk. Separate the row into multiple columns using the : and - separators. In each row containing :, subtract the contents of column 2 from the contents of column 3 and print result. awk -F '[:-]' '/:/{print $3-$2}' file Output: 4899
Using awk $ awk -F: '/:/ {split($2,a,"-"); print a[2] - a[1]}' input_file 4899
Extract the last three columns from a text file with awk
I have a .txt file like this: ENST00000000442 64073050 64074640 64073208 64074651 ESRRA ENST00000000233 127228399 127228552 ARF5 ENST00000003100 91763679 91763844 CYP51A1 I want to get only the last 3 columns of each line. as you see some times there are some empty lines between 2 lines which must be ignored. here is the output that I want to make: 64073208 64074651 ESRRA 127228399 127228552 ARF5 91763679 91763844 CYP51A1 awk '/a/ {print $1- "\t" $-2 "\t" $-3}' file.txt. it does not return what I want. do you know how to correct the command?
Following awk may help you in same. awk 'NF{print $(NF-2),$(NF-1),$NF}' OFS="\t" Input_file Output will be as follows. 64073208 64074651 ESRRA 127228399 127228552 ARF5 91763679 91763844 CYP51A1 EDIT: Adding explanation of command too now.(NOTE this following command is for only explanation purposes one should run above command only to get the results) awk 'NF ###Checking here condition NF(where NF is a out of the box variable for awk which tells number of fields in a line of a Input_file which is being read). ###So checking here if a line is NOT NULL or having number of fields value, if yes then do following. { print $(NF-2),$(NF-1),$NF###Printing values of $(NF-2) which means 3rd last field from current line then $(NF-1) 2nd last field from line and $NF means last field of current line. } ' OFS="\t" Input_file ###Setting OFS(output field separator) as TAB here and mentioning the Input_file here.
You can use sed too sed -E '/^$/d;s/.*\t(([^\t]*[\t|$]){2})/\1/' infile
With some piping: $ cat file | tr -s '\n' | rev | cut -f 1-3 | rev 64073208 64074651 ESRRA 127228399 127228552 ARF5 91763679 91763844 CYP51A1 First, cat the file to tr to squeeze out repeted \ns to get rid of empty lines. Then reverse the lines, cut the first three fields and reverse again. You could replace the useless cat with the first rev.
print 1st string of a line if last 5 strings match input
I have a requirement to print the first string of a line if last 5 strings match specific input. Example: Specified input is 2 India;1;2;3;4;5;6 Japan;1;2;2;2;2;2 China;2;2;2;2 England;2;2;2;2;2 Expected Output: Japan England As you can see, China is excluded as it doesn't meet the requirement (last 5 digits have to be matched with the input).
grep ';2;2;2;2;2$' file | cut -d';' -f1 $ in a regex stands for "end of line", so grep will print all the lines that end in the given string -d';' tells cut to delimit columns by semicolons -f1 outputs the first column
You could use awk: awk -F';' -v v="2" -v count=5 ' { c=0; for(i=2;i<=NF;i++){ if($i == v) c++ if(c>=count){print $1;next} } }' file where v is the value to match count is the maximum number of value to print the wanted string the for loop is parsing all fields delimited with a ; in order to find a match This script doesn't need the 5 values 2 to be consecutive.
With sed: sed -n 's/^\([^;]*\).*;2;2;2;2;2$/\1/p' file It captures and output non ; first characters in lines ending with ;2;2;2;2;2 It can be shortened with GNU sed to: sed -nE 's/^([^;]*).*(;2){5}$/\1/p' file
awk -F\; '/;2;2;2;2;2$/{print $1}' file Japan England
Append and replace using awk/sed
I have this file: 2016,05,P,0002 ,CJGLOPSD8 00,BBF,BBDFTP999,051000100,GBP, , -2705248.00 00,BBF,BBDFTP999,059999998,GBP, , -3479679.38 00,BBF,BBDFTP999,061505141,GBP, , -0.40 00,BBF,BBDFTP999,061505142,GBP, , 6207621.00 00,BBF,BBDFTP999,061505405,GBP, , -0.16 00,BBF,BBDFTP999,061552000,GBP, , -0.24 00,BBF,BBDFTP999,061559010,GBP, , -0.44 00,BBF,BBDFTP999,062108021,GBP, , -0.34 00,BBF,BBDFTP999,063502007,GBP, , -0.28 I want to programmatically (in unix, or informatica if possible) grab the first two fields in the top row, concatenate them, append them to the end of each line and remove that first row. Like so: 00,BBF,BBDFTP999,051000100,GBP,,-2705248.00,201605 00,BBF,BBDFTP999,059999998,GBP,,-3479679.38,201605 00,BBF,BBDFTP999,061505141,GBP,,-0.40,201605 00,BBF,BBDFTP999,061505142,GBP,,6207621.00,201605 00,BBF,BBDFTP999,061505405,GBP,,-0.16,201605 00,BBF,BBDFTP999,061552000,GBP,,-0.24,201605 00,BBF,BBDFTP999,061559010,GBP,,-0.44,201605 00,BBF,BBDFTP999,062108021,GBP,,-0.34,201605 00,BBF,BBDFTP999,063502007,GBP,,-0.28,201605 This is my current attempt: awk -vvar1=`cat OF\ OPSDOWN8.CSV | head -1 | cut -d',' -f1` -vvar2=`cat OF\ OPSDOWN8.CSV | head -1 | cut -d',' -f2` 'BEGIN {FS=OFS=","} {print $0, var 1var2}' OF\ OPSDOWN8.CSV> OF_OPSDOWN8.csv Any pointers? I've tried looking around the forum but can only find answers to part of my question. Thanks for your help.
Use this awk: awk 'BEGIN{FS=OFS=","} NR==1{val=$1$2;next} {gsub(/ */,"");print $0,val}' file Explanation: BEGIN{FS=OFS=","} - This block will set FS (Field Separator) and OFS (Output Field Separator) as ,. NR==1 - Working with line number 1. Here, $1 and $2 denotes field number. print $0,val - Printing $0 (whole line) and stored value from val.
I would use the following awk command: awk 'NR==1{d=$1$2;next}{$(NF+1)=d;gsub(/[[:space:]]/,"")}1' FS=, OFS=, file Explanation: NR==1{d=$1$2;next} applies on line 1 and set's a variable d(ate) to the value of the first and the second field. The variable is being used when processing the remaining lines. next tells awk to go ahead with the next line right away without processing further instructions on this line. {$(NF+1)=d;gsub(/[[:space:]]/,"")}1 appends a new field to the line (NF is the number of fields, assigning d to $(NF+1) effectively adds a field. gsub() is used to removing spaces. 1 at the end always evaluates to true and makes awk print the modified line. FS=, is a command line argument. It set's the input field delimiter to ,. OFS=, is a command line argument. It set's the output field delimiter to ,. Output: 00,BBF,BBDFTP999,051000100,GBP,,-2705248.00,201605 00,BBF,BBDFTP999,059999998,GBP,,-3479679.38,201605 00,BBF,BBDFTP999,061505141,GBP,,-0.40,201605 00,BBF,BBDFTP999,061505142,GBP,,6207621.00,201605 00,BBF,BBDFTP999,061505405,GBP,,-0.16,201605 00,BBF,BBDFTP999,061552000,GBP,,-0.24,201605 00,BBF,BBDFTP999,061559010,GBP,,-0.44,201605 00,BBF,BBDFTP999,062108021,GBP,,-0.34,201605 00,BBF,BBDFTP999,063502007,GBP,,-0.28,201605
With sed : sed '1{s/\([^,]*\),\([^,]*\),.*/\1\2/;h;d};/.*/G;s/\n/,/;s/ //g' file in ERE mode : sed -r '1{s/([^,]*),([^,]*),.*/\1\2/;h;d};/.*/G;s/\n/,/;s/ //g' file Output : 00,BBF,BBDFTP999,051000100,GBP,,-2705248.00,201605 00,BBF,BBDFTP999,059999998,GBP,,-3479679.38,201605 00,BBF,BBDFTP999,061505141,GBP,,-0.40,201605 00,BBF,BBDFTP999,061505142,GBP,,6207621.00,201605 00,BBF,BBDFTP999,061505405,GBP,,-0.16,201605 00,BBF,BBDFTP999,061552000,GBP,,-0.24,201605 00,BBF,BBDFTP999,061559010,GBP,,-0.44,201605 00,BBF,BBDFTP999,062108021,GBP,,-0.34,201605 00,BBF,BBDFTP999,063502007,GBP,,-0.28,201605
This might work for you (GNU sed): sed '1s/,//;1s/,.*//;1h;1d;s/ //g;G;s/\n/,/' file For the first line only: remove the first comma, remove from the next comma to the end of the line, store the amended line in the hold space (HS) and then delete the current line (the d abruptly ends processing). For subsequent lines: remove all spaces, append the HS and replace the newline (from the G command) with a comma. Or if you prefer: sed '1{s/,//;s/,.*//;h;d};s/ //g;G;s/\n/,/' file
If you want to use Informatica for this, use two Source Qualifiers. Read the file twice - just one line in one SQ (filter out the rest) and in the second SQ read the whole file except the first line (skip header). Join the two on dummy port and you're done.