Process Substitution Without Shell? |
November 28th, 2023 |
shell, tech |
Say I have a program that reads two files together in a single pass [1] and writes something out. The inputs you have are compressed, so you'll need to decompress them, and the output needs to be compressed before you write it out to storage. You could do:
# download the files aws s3 cp "$path1" . aws s3 cp "$path2" . # decompress the files gunzip "$file1" gunzip "$file2" # run the command cmd -1 "$file1" -2 "$file2" > "$fileOut" # compress the output gzip "$fileOut" # upload the output aws s3 cp "$fileOut.gz" "$pathOut"
This works, but for large files it's slow and needs too much space. We're waiting for each step to finish before starting the next, and we're storing some very large intermediate files on the local machine.
Instead, we'd like to stream the inputs down, decompress them as we go, compress the output as it comes out, and stream the output back up. In bash this is reasonably straightforward to write:
cmd -1 <(aws s3 cp "$path1" - | gunzip) \ -2 <(aws s3 cp "$path2" - | gunzip) \ | gzip | aws s3 cp - "$pathOut"
This uses almost no disk space and it parallelizes the decompression, command, and recompression. But it's also shell...
I tend to use python for this kind of thing, where I'm gluing things together and want it to be clear what I'm doing. It seems like it should be possible to do this sort of thing with the subprocess module, but while I've played with it a few times I haven't figured it out. I'd like an API like:
pipeline = subprocess.Pipeline() dl1 = pipeline.process( ["aws", "s3", "cp", path1, "-"]) gunzip1 = pipeline.process( ["gunzip"], stdin=dl1.stdout) dl2 = pipeline.process( ["aws", "s3", "cp", path2, "-"]) gunzip2 = pipeline.process( ["gunzip"], stdin=dl2.stdout) cmd = pipeline.process( ["cmd", "-1", dl1.stdout, "-2", dl2.stdout]) gzip = pipeline.process( ["gzip"], stdin=cmd.stdout) pipeline.process( ["aws", "s3", "cp", "-", pathOut], stdin=gzip.stdout) pipeline.check_call()
Or:
from subprocess import check_call, PIPE, InputFile check_call([ "cmd", "-1", InputFile([ "aws", "s3", "cp", path1, "-", PIPE, "gunzip"]), "-2", InputFile([ "aws", "s3", "cp", path2, "-", PIPE, "gunzip"]), PIPE, "gzip", PIPE, "aws", "s3", "cp", "-", pathOut])
These are 5x and 3x the length of the bash version, but I'd be willing to put up with that for having something that's more robust. The difference would also be smaller in practice as the commands would typically have a lot of arguments.
I see these stack overflow answers suggesting named pipes, but they seem awkward, hard to read, and easy to get wrong. Is there a better way? Should I just stick with bash when doing something bash is this good a fit for, especially now that people can paste my code into an LLM and get an explanation of what it's doing?
[1] Interleaved fastq
files, where the Nth record in file 1
corresponds to the Nth record in file 2.
Comment via: facebook, lesswrong, mastodon