Added a pandas script for cleaning up csv.

This commit is contained in:
Tristan Smith 2024-05-26 23:51:32 -04:00
parent 1e9929f753
commit 2e6f06eb1d

20
organize_csv.py Normal file
View file

@ -0,0 +1,20 @@
# Not sure how this will act on a number of machines.
import pandas as pd
df = pd.read_csv('results.csv')
df = df.drop_duplicates()
df = df.sort_values(by=['Worker ID', 'IP Address'])
# Group by Worker ID, IP Address, Log File, and Error Type
df['Error Count'] = 1
df_grouped = df.groupby(['Worker ID', 'IP Address', 'Log File', 'Error Type'], as_index=False).agg({
'Error Message': 'first',
'Error Count': 'sum'
})
df_grouped.to_csv('cleaned_results_grouped.csv', index=False)
print(df_grouped)