Split file with 800,000 columns - bash
I want to split a file of genomic data with 800,000 columns and 40,000 rows into a series of files with 100 columns each, total size 118GB.
I am currently running the following bash script, multithread 15 times:
infile="$1"
start=$2
end=$3
step=$(($4-1))
for((curr=$start, start=$start, end=$end; curr+step <= end; curr+=step+1)); do
cut -f$curr-$((curr+step)) "$infile" > "${infile}.$curr" -d' '
done
However, judging by current progress of the script, it will take 300 days to complete the split?!
Is there a more efficient way to column wise split a space-delimited file into smaller chunks?
Try this awk script:
awk -v cols=100 '{
f = 1
for (i = 1; i <= NF; i++) {
printf "%s%s", $i, (i % cols && i < NF ? OFS : ORS) > (FILENAME "." f)
f=int(i/cols)+1
}
}' largefile
I expect it to be faster than the shell script in the question.
Related
Loop to create a a DF from values in bash
Im creating various text files from a file like this: Chrom_x,Pos,Ref,Alt,RawScore,PHRED,ID,Chrom_y 10,113934,A,C,0.18943,5.682,rs10904494,10 10,126070,C,T,0.030435000000000007,3.102,rs11591988,10 10,135656,T,G,0.128584,4.732,rs10904561,10 10,135853,A,G,0.264891,6.755,rs7906287,10 10,148325,A,G,0.175257,5.4670000000000005,rs9419557,10 10,151997,T,C,-0.21169,0.664,rs9286070,10 10,158202,C,T,-0.30357,0.35700000000000004,rs9419478,10 10,158946,C,T,2.03221,19.99,rs11253562,10 10,159076,G,A,1.403107,15.73,rs4881551,10 What I am trying to do is extract, in bash, all values beetwen two values: gawk '$6>=0 && $NF<=5 {print $0}' file.csv > 0_5.txt And create files from 6 to 10, from 11 to 15... from 95 to 100. I was thinking in creating a loop for this with something like #!/usr/bin/env bash n=( 0,5,6,10...) if i in n: gawk '$6>=n && $NF<=n+1 {print $0}' file.csv > n_n+1.txt and so on. How can i convert this as a loop and create files with this specific values.
While you could use a shell loop to provide inputs to an awk script, you could also just use awk to natively split the values into buckets and write the lines to those "bucket" files itself: awk -F, ' NR > 1 { i=int((($6 - 1) / 5)) fname=(i*5) "_" (i+1)*5 ".txt" print $0 > fname }' < input The code skips the header line (NR > 1) and then computes a "bucket index" by dividing the value in column six by five. The filename is then constructed by multiplying that index (and its increment) by five. The whole line is then printed to that filename. To use a shell loop (and call awk 20 times on the input), you could use something like this: for((i=0; i <= 19; i++)) do floor=$((i * 5)) ceiling=$(( (i+1) * 5)) awk -F, -v floor="$floor" -v ceiling="$ceiling" \ 'NR > 1 && $6 >= floor && $6 < ceiling { print }' < input \ > "${floor}_${ceiling}.txt" done The basic idea is the same; here, we're creating the bucket index with the outer loop and then passing the range into awk as the floor and ceiling variables. We're only asking awk to print the matching lines; the output from awk is captured by the shell as a redirection into the appropriate file.
Using one awk output into another awk command
I have one file (excel file) which has some columns (not fixed, changes dynamically) and I need to get values for couple of particular columns. I'm able to get the columns using one awk command and then printing rows using these columns numbers into another awk command. Is there any way I can combine into one? awk -F',' ' {for(i=1;i < 9;i++) {if($i ~ /CLIENT_ID/) {print i}}} {for(s=1;s < 2;s++) {if($s ~ /SEC_DESC/) {print s}}} ' <file.csv> | awk -F "," '!($5~/...[0-9]L/ && $21~/FUT /) {print $0}' <file.csv> Gives me output as 5 and 9 for columns (client_idandsec_desc`), which is their column number (this changes with different files). Now using this column number, I get the desired output as follows: awk -F "," '!($5~/...[0-9]L/ && $21~/FUT /) {print $0}' <file.csv> How can I combine these into one command? Pass a variable from the first to the second? Input (csv file having various dynamic columns, interested in following two columns) CLIENT_ID SEC_DESC USZ256 FUT DEC 16 U.S. USZ256L FUT DEC 16 U.S. BONDS WNZ256 FUT DEC 16 CBX WNZ256L FUT DEC 16 CBX BONDS Output give me rows- 2 and 4 that matched my regex pattern in second awk command (having column numbers as 5 & 21). These column numbers changes as per file so first have to get the column number using first awl and then giving it as input to second awk.
I think I got it. awk -F',' ' NR == 1 { for (i=1; i<=NF; ++i) { if ($i == "CLIENT_ID") cl_col = i if ($i == "SEC_DESC") sec_col = i } } NR > 1 && !($cl_col ~ /...[0-9]L/ && $sec_col ~ /FUT /) {print $0} ' RED_FUT_TST.csv
To solve your problem you can test when you're processing the first row, and put the logic to discover the column numbers there. Then when you are processing the data rows, use the column numbers from the first step. (NR is an awk built-in variable containing the record number being processed. NF is the number of columns.) E.g.: $ cat red.awk NR == 1 { for (i=1; i<=NF; ++i) { if ($i == "CLIENT_ID") cl_col = i; if ($i == "SEC_DESC") sec_col = i; } } NR > 1 && $cl_col ~ /...[0-9]L/ && $sec_col ~ /FUT / $ awk -F'\t' -f red.awk RED_FUT_TST.csv USZ256L FUT DEC 16 U.S. BONDS WNZ256L FUT DEC 16 CBX BONDS
Awk - find the minimum in a given row
I have a file organized in rows and columns. I want to find the minimum in a given row, for example row number 4, and then transfer the corresponding column number in a bash variable (lev). However the small code I wrote is not working lev=`echo - |awk '{ m=100; l=1; {If (NR==4) for(i=2;i<=NF;i++) { if( $i <m) m=$i; l=i } } print l }' file.txt`
There are multiple things wrong with your script. Perhaps you can figure out using this sample. $ lev=$(awk 'NR==4{min=99999; for(i=1;i<=NF;i++) if($i < min) {min=$i; ix=i} print ix}' file.txt)
How to efficiently sum two columns in a file with 270,000+ rows in bash
I have two columns in a file, and I want to automate summing both values per row for example read write 5 6 read write 10 2 read write 23 44 I want to then sum the "read" and "write" of each row. Eventually after summing, I'm finding the max sum and putting that max value in a file. I feel like I have to use grep -v to rid of the column headers per row, which like stated in the answers, makes the code inefficient since I'm grepping the entire file just to read a line. I currently have this in a bash script (within a for loop where $x is the file name) to sum the columns line by line lines=`grep -v READ $x|wc -l | awk '{print $1}'` line_num=1 arr_num=0 while [ $line_num -le $lines ] do arr[$arr_num]=`grep -v READ $x | sed $line_num'q;d' | awk '{print $2 + $3}'` echo $line_num line_num=$[$line_num+1] arr_num=$[$arr_num+1] done However, the file to be summed has 270,000+ rows. The script has been running for a few hours now, and it is nowhere near finished. Is there a more efficient way to write this so that it does not take so long?
Use awk instead and take advantage of modulus function: awk '!(NR%2){print $1+$2}' infile
awk is probably faster, but the idiomatic bash way to do this is something like: while read -a line; do # read each line one-by-one, into an array # use arithmetic expansion to add col 1 and 2 echo "$(( ${line[0]} + ${line[1]} ))" done < <(grep -v READ input.txt) Note the file input file is only read once (by grep) and the number of externally forked programs is kept to a minimum (just grep, called only once for the whole input file). The rest of the commands are bash builtins. Using the <( ) process substition, in case variables set in the while loop are required out of scope of the while loop. Otherwise a | pipe could be used.
Your question is pretty verbose, yet your goal is not clear. The way I read it, your numbers are on every second line, and you want only to find the maximum sum. Given that: awk ' NR%2 == 1 {next} NR == 2 {max = $1+$2; next} $1+$2 > max {max = $1+$2} END {print max} ' filename
You could also use a pipeline with tools that implicitly loop over the input like so: grep -v read INFILE | tr -s ' ' + | bc | sort -rn | head -1 > OUTFILE This assumes there are spaces between your read and write data values.
Why not run: awk 'NR==1 { print "sum"; next } { print $1 + $2 }' You can afford to run it on the file while the other script it still running. It'll be complete in a few seconds at most (prediction). When you're confident it's right, you can kill the other process. You can use Perl or Python instead of awk if you prefer. Your code is running grep, sed and awk on each line of the input file; that's damnably expensive. And it isn't even writing the data to a file; it is creating an array in Bash's memory that'll need to be printed to the output file later.
Assuming that it's always one 'header' row followed by one 'data' row: awk ' BEGIN{ max = 0 } { if( NR%2 == 0 ){ sum = $1 + $2; if( sum > max ) { max = sum } } } END{ print max }' input.txt Or simply trim out all lines that do not conform to what you want: grep '^[0-9]\+\s\+[0-9]\+$' input.txt | awk ' BEGIN{ max = 0 } { sum = $1 + $2; if( sum > max ) { max = sum } } END{ print max }' input.txt
Comparison shell script for large text/csv files - improvement needed
My task is the following - I have two CSV files: File 1 (~9.000.000 records): type(text),number,status(text),serial(number),data1,data2,data3 File 2 (~6000 records): serial_range_start(number),serial_range_end(number),info1(text),info2(text) The goal is to add to each entry in File 1 the corresponding info1 and info2 from File 2: type(text),number,status(text),serial(number),data1,data2,data3,info1(text),info2(text) I use the following script: #!/bin/bash USER="file1.csv" RANGE="file2.csv" for SN in `cat $USER | awk -F , '{print $4}'` do #echo \n "$SN" for LINE in `cat $RANGE` do i=`grep $LINE $RANGE| awk -F, '{print $1}'` #echo \n "i= " "$i" j=`grep $LINE $RANGE| awk -F, '{print $2}'` #echo \n "j= " "$j" k=`echo $SN` #echo \n "k= " "$k" if [ $k -ge $i -a $k -le $j ] then echo `grep $SN $USER`,`grep $i $RANGE| cut -d',' -f3-4` >> result.csv break #else #echo `grep $SN $USER`,`echo 'N/A','N/A'` >> result.csv fi done done The script works rather good on small files but I'm sure there is a way to optimize it because I am running it on an i5 laptop with 4GB of RAM. I am a newbie in shell scripting and I came up with this script after hours and hours of research, trial and error but now I am out of ideas. note: not all the info in file 1 can be found in file. Thank you! Adrian. FILE EXAMPLES and additional info: File 1 example: prep,28620026059,Active,123452010988759,No,No,No post,28619823474,Active,123453458466109,Yes,No,No post,28619823474,Inactive,123453395270941,Yes,Yes,Yes File 2 example: 123452010988750,123452010988759,promo32,1.11 123453458466100,123453458466199,promo64,2.22 123450000000000,123450000000010,standard128,3.333 Result example (currently): prep,28620026059,Active,123452010988759,No,No,No,promo32,1.11 post,28619823474,Active,123453458466109,Yes,No,No,promo64,2.22 Result example (nice to have): prep,28620026059,Active,123452010988759,No,No,No,promo32,1.11 post,28619823474,Active,123453458466109,Yes,No,No,promo64,2.22 post,28619823474,Inactive,123453395270941,Yes,Yes,Yes,NA,NA File 1 is sorted after the 4th column File 2 is sorted after the first column. File 2 does not have ranges that overlap Not all the info in file 1 can be found in a range in file 2 Thanks again! LE: The script provided by Jonathan seems to have an issue on some records, as follows: file 2: 123456780737000,123456780737012,ONE 32,1.11 123456780016000,123456780025999,ONE 64,2.22 file 1: Postpaid,24987326427,Active,123456780737009,Yes,Yes,Yes Postpaid,54234564719,Active,123456780017674,Yes,Yes,Yes The output is the following: Postpaid,24987326427,Active,123456780737009,Yes,Yes,Yes,ONE 32,1.11 Postpaid,54234564719,Active,123456780017674,Yes,Yes,Yes,ONE 32,1.11 and it should be: Postpaid,24987326427,Active,123456780737009,Yes,Yes,Yes,ONE 32,1.11 Postpaid,54234564719,Active,123456780017674,Yes,Yes,Yes,ONE 64,2.22 It seems that it returns 0 and writes the info on first record from file2...
I think this will work reasonably well: awk -F, 'BEGIN { n = 0; OFS = ","; } NR==FNR { lo[n] = $1; hi[n] = $2; i1[n] = $3; i2[n] = $4; n++ } NR!=FNR { for (i = 0; i < n; i++) { if ($4 >= lo[i] && $4 <= hi[i]) { print $1, $2, $3, $4, $5, $6, $7, i1[i], i2[i]; break; } } }' file2 file1 Given file2 containing: 1,10,xyz,pqr 11,20,abc,def 21,30,ambidextrous,warthog and file1 containing: A,123,X2,1,data01_1,data01_2,data01_3 A,123,X2,2,data02_1,data02_2,data02_3 A,123,X2,3,data03_1,data03_2,data03_3 A,123,X2,4,data04_1,data04_2,data04_3 A,123,X2,5,data05_1,data05_2,data05_3 A,123,X2,6,data06_1,data06_2,data06_3 A,123,X2,7,data07_1,data07_2,data07_3 A,123,X2,8,data08_1,data08_2,data08_3 A,123,X2,9,data09_1,data09_2,data09_3 A,123,X2,10,data10_1,data10_2,data10_3 A,123,X2,11,data11_1,data11_2,data11_3 A,123,X2,12,data12_1,data12_2,data12_3 A,123,X2,13,data13_1,data13_2,data13_3 A,123,X2,14,data14_1,data14_2,data14_3 A,123,X2,15,data15_1,data15_2,data15_3 A,123,X2,16,data16_1,data16_2,data16_3 A,123,X2,17,data17_1,data17_2,data17_3 A,123,X2,18,data18_1,data18_2,data18_3 A,123,X2,19,data19_1,data19_2,data19_3 A,223,X2,20,data20_1,data20_2,data20_3 A,223,X2,21,data21_1,data21_2,data21_3 A,223,X2,22,data22_1,data22_2,data22_3 A,223,X2,23,data23_1,data23_2,data23_3 A,223,X2,24,data24_1,data24_2,data24_3 A,223,X2,25,data25_1,data25_2,data25_3 A,223,X2,26,data26_1,data26_2,data26_3 A,223,X2,27,data27_1,data27_2,data27_3 A,223,X2,28,data28_1,data28_2,data28_3 A,223,X2,29,data29_1,data29_2,data29_3 the output of the command is: A,123,X2,1,data01_1,data01_2,data01_3,xyz,pqr A,123,X2,2,data02_1,data02_2,data02_3,xyz,pqr A,123,X2,3,data03_1,data03_2,data03_3,xyz,pqr A,123,X2,4,data04_1,data04_2,data04_3,xyz,pqr A,123,X2,5,data05_1,data05_2,data05_3,xyz,pqr A,123,X2,6,data06_1,data06_2,data06_3,xyz,pqr A,123,X2,7,data07_1,data07_2,data07_3,xyz,pqr A,123,X2,8,data08_1,data08_2,data08_3,xyz,pqr A,123,X2,9,data09_1,data09_2,data09_3,xyz,pqr A,123,X2,10,data10_1,data10_2,data10_3,xyz,pqr A,123,X2,11,data11_1,data11_2,data11_3,abc,def A,123,X2,12,data12_1,data12_2,data12_3,abc,def A,123,X2,13,data13_1,data13_2,data13_3,abc,def A,123,X2,14,data14_1,data14_2,data14_3,abc,def A,123,X2,15,data15_1,data15_2,data15_3,abc,def A,123,X2,16,data16_1,data16_2,data16_3,abc,def A,123,X2,17,data17_1,data17_2,data17_3,abc,def A,123,X2,18,data18_1,data18_2,data18_3,abc,def A,123,X2,19,data19_1,data19_2,data19_3,abc,def A,223,X2,20,data20_1,data20_2,data20_3,abc,def A,223,X2,21,data21_1,data21_2,data21_3,ambidextrous,warthog A,223,X2,22,data22_1,data22_2,data22_3,ambidextrous,warthog A,223,X2,23,data23_1,data23_2,data23_3,ambidextrous,warthog A,223,X2,24,data24_1,data24_2,data24_3,ambidextrous,warthog A,223,X2,25,data25_1,data25_2,data25_3,ambidextrous,warthog A,223,X2,26,data26_1,data26_2,data26_3,ambidextrous,warthog A,223,X2,27,data27_1,data27_2,data27_3,ambidextrous,warthog A,223,X2,28,data28_1,data28_2,data28_3,ambidextrous,warthog A,223,X2,29,data29_1,data29_2,data29_3,ambidextrous,warthog This uses a linear search on the list of ranges; you can write functions in awk and a binary search looking for the correct range would perform better on 6,000 entries. That part, though, is an optimization — exercise for the reader. Remember that the first rule of optimization is: don't. The second rule of optimization (for experts only) is: don't do it yet. Demonstrate that it is a problem. This code shouldn't take all that much longer than the time it takes to copy the 9,000,000 record file (somewhat longer, but not disastrously so). Note, though, that if the file1 data is sorted, the tail of the processing will take longer than the start because of the linear search. If the serial numbers are in a random order, then it will all take about the same time on average. If your CSV data has commas embedded in the text fields, then awk is no longer suitable; you need a tool with explicit support for CSV format — Perl and Python both have suitable modules. Answer to Exercise for the Reader awk -F, 'BEGIN { n = 0; OFS = ","; } NR==FNR { lo[n] = $1; hi[n] = $2; i1[n] = $3; i2[n] = $4; n++ } NR!=FNR { i = search($4) print $1, $2, $3, $4, $5, $6, $7, i1[i], i2[i]; } function search(i, l, h, m) { l = 0; h = n - 1; while (l <= h) { m = int((l + h)/2); if (i >= lo[m] && i <= hi[m]) return m; else if (i < lo[m]) h = m - 1; else l = m + 1; } return 0; # Should not get here }' file2 file1 Not all that hard to write the binary search. This gives the same result as the original script on the sample data. It has not been exhaustively tested, but appears to work. Note that the code does not really handle missing ranges in file2; it assumes that the ranges are contiguous but non-overlapping and in sorted order and cover all the values that can appear in the serial column of file1. If those assumptions are not valid, you get erratic behaviour until you fix either the code or the data.
In Unix you may use the join command (type 'man join' for more information) which can be configured to work similarly to join operation in databases. That may help you to add the information from File2 to File1.