Op 10/03/2022 om 13:16 schreef Loris Bennett:
Hi,
I have a command which produces output like the
following:
Job ID: 9431211
Cluster: curta
User/Group: build/staff
State: COMPLETED (exit code 0)
Nodes: 1
Cores per node: 8
CPU Utilized: 01:30:53
CPU Efficiency: 83.63% of 01:48:40 core-walltime
Job Wall-clock time: 00:13:35
Memory Utilized: 6.45 GB
Memory Efficiency: 80.68% of 8.00 GB
I want to parse this and am using subprocess.Popen and accessing the
contents via Popen.stdout. However, for testing purposes I want to save
various possible outputs of the command as text files and use those as
inputs.
What format should I use to pass data to the actual parsing function?
Is this a command you run, produces that output, and then stops (as
opposed to a long-running program that from time to time generates a
bunch of output)?
Because in that case I would use subprocess.run() with
capture_output=True instead of subprocess.Popen(). subprocess.run()
returns a CompletedProcess instance wich has stdout and stderr members
that contain the captured output as byte sequences or strings, depending
on the parameters you passed.
So in your case I would simply read the content of each text file as a
whole into a string, and use subprocess.run() to get the command's
output also as a string. Then you can have a parse function that accepts
such strings, and works exactly the same for the text files as for the
command output. Your parse function can then use splitlines() to access
the lines individually. The data size is very small so it's not a
problem to have it all in memory at the same time (i.e. no need to worry
about trying to stream it instead).
--
"Don't Panic."
-- Douglas Adams, The Hitchhiker's Guide to the Galaxy
--
https://mail.python.org/mailman/listinfo/python-list