-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy pathmain_EventAnalysis_AllFrames.py
171 lines (133 loc) · 7.32 KB
/
main_EventAnalysis_AllFrames.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
import os
import pandas as pd
import numpy as np
from tqdm import tqdm
# Parameters
PAT_NOW = "S23_199"
FACEDX_CSV_DIRECTORY = os.path.abspath(f'/home/jgopal/NAS/Analysis/outputs_Combined/{PAT_NOW}/')
OPENFACE_CSV_DIRECTORY = os.path.abspath(f'/home/jgopal/NAS/Analysis/outputs_OpenFace/{PAT_NOW}/')
OUTPUT_DIRECTORY = os.path.abspath(f'/home/jgopal/NAS/Analysis/outputs_EventAnalysis/')
OUTPUT_CSV = os.path.join(OUTPUT_DIRECTORY, f'combined_events_{PAT_NOW}.csv')
META_DATA_CSV_PATH = os.path.join(os.path.abspath(f'/home/jgopal/NAS/Analysis/outputs_EventAnalysis/'), f'chosen_thresholds_{PAT_NOW}.csv')
# Create the output directory if it doesn't exist
os.makedirs(OUTPUT_DIRECTORY, exist_ok=True)
# Load the thresholds from the meta data CSV
thresholds_df = pd.read_csv(META_DATA_CSV_PATH)
EVENT_THRESHOLDS = dict(zip(thresholds_df['Emotion'], thresholds_df['Threshold']))
MIN_EVENT_LENGTH = 2 # Minimum length of each event in frames
MERGE_TIME = 3 # Maximum frames apart to consider merging events
FACEDX_FPS = 5 # FPS after down sampling
VIDEO_FPS = 30 # FPS of original video (for time stamps!)
# Function to detect and merge events
def detect_events(emotion_df, au_df, openface_df):
events = []
for emotion, threshold in EVENT_THRESHOLDS.items():
emotion_values = emotion_df[emotion].values
frames = emotion_df['frame'].values
# Identify start and end frames of events
above_threshold = emotion_values >= threshold
start_indices = np.where((above_threshold[:-1] == False) & (above_threshold[1:] == True))[0] + 1
end_indices = np.where((above_threshold[:-1] == True) & (above_threshold[1:] == False))[0]
if len(start_indices) == 0 or len(end_indices) == 0:
continue
if start_indices[0] > end_indices[0]:
end_indices = end_indices[1:]
if start_indices[-1] > end_indices[-1]:
start_indices = start_indices[:-1]
for start_frame, end_frame in zip(frames[start_indices], frames[end_indices]):
start_frame = int(start_frame) # Ensure it's an integer
end_frame = int(end_frame) # Ensure it's an integer
# Calculate event length based on the available rows in hse or ogau CSV
event_length = len(au_df[(au_df['frame'] >= start_frame) & (au_df['frame'] <= end_frame)])
if event_length < MIN_EVENT_LENGTH:
continue
# Merge close by events
if events and start_frame - events[-1]['End Frame'] <= MERGE_TIME:
# Ensure 'Duration in Seconds' exists before updating it
if 'Duration in Seconds' not in events[-1]:
events[-1]['Duration in Seconds'] = 0
events[-1]['End Frame'] = end_frame
events[-1]['Duration in Seconds'] = round(events[-1]['Duration in Seconds'] + event_length / FACEDX_FPS, 1)
continue
minutes = int((start_frame // VIDEO_FPS) // 60)
seconds = int((start_frame // VIDEO_FPS) - (60 * minutes))
if minutes >= 60:
hours = int(minutes // 60)
minutes = minutes % 60
start_time = f"{hours:02d}:{minutes:02d}:{seconds:02d}"
else:
start_time = f"{minutes:02d}:{seconds:02d}"
duration = round(event_length / FACEDX_FPS, 1)
event_data = {
'Filename': video_file,
'Start Time': start_time,
'Duration in Seconds': duration, # Ensure duration is set initially
'Event Type': emotion,
'End Frame': end_frame
}
# Add the full range of frames (at 5 FPS) for this event
for frame in range(start_frame, end_frame + 1):
# Convert frame columns to integers for consistent comparison
au_df['frame'] = au_df['frame'].astype(int)
emotion_df['frame'] = emotion_df['frame'].astype(int)
openface_df['frame'] = openface_df['frame'].astype(int)
# Use integer comparison for frames
frame_au = au_df[au_df['frame'] == int(frame)].drop(['frame', 'timestamp', 'success'], axis=1)
frame_emotion = emotion_df[emotion_df['frame'] == int(frame)].drop(['frame', 'timestamp', 'success'], axis=1)
# Check if the frame exists in the data, otherwise skip this frame
if frame_au.empty or frame_emotion.empty:
continue
# Handle potential variations in column names (with or without leading space)
au45_r_col = ' AU45_r' if ' AU45_r' in openface_df.columns else 'AU45_r'
au45_c_col = ' AU45_c' if ' AU45_c' in openface_df.columns else 'AU45_c'
frame_openface = openface_df[openface_df['frame'] == frame][[au45_r_col, au45_c_col]]
# If no OpenFace data is found, set the OpenFace columns to NaN
if frame_openface.empty:
frame_openface = pd.DataFrame({'OpenFace_AU45_r': [np.nan], 'OpenFace_AU45_c': [np.nan]})
else:
frame_openface = frame_openface.rename(columns={au45_r_col: 'OpenFace_AU45_r', au45_c_col: 'OpenFace_AU45_c'})
frame_data = event_data.copy()
frame_data['Frame'] = frame
frame_data.update(frame_au.to_dict(orient='records')[0] if not frame_au.empty else {})
frame_data.update(frame_emotion.to_dict(orient='records')[0] if not frame_emotion.empty else {})
frame_data.update(frame_openface.to_dict(orient='records')[0] if not frame_openface.empty else {})
events.append(frame_data)
return events
# Process each video file
all_events = []
# Loop through the subfolders in the given CSV directory
for subfolder in tqdm(os.listdir(FACEDX_CSV_DIRECTORY)):
video_file = subfolder
# Load emotion and AU CSVs
emotion_csv_path = os.path.join(FACEDX_CSV_DIRECTORY, subfolder, 'outputs_hse.csv')
au_csv_path = os.path.join(FACEDX_CSV_DIRECTORY, subfolder, 'outputs_ogau.csv')
openface_csv_path = os.path.join(OPENFACE_CSV_DIRECTORY, f'{subfolder[:-4]}.csv')
if not os.path.exists(emotion_csv_path) or not os.path.exists(au_csv_path):
print(f"Skipping {video_file}: missing CSV files.")
continue
if os.path.getsize(emotion_csv_path) == 0 or os.path.getsize(au_csv_path) == 0:
print(f"Skipping {video_file}: empty CSV files.")
continue
if os.path.getsize(openface_csv_path) == 0:
print(f"Skipping {video_file}: empty OpenFace CSV file.")
continue
try:
emotion_df = pd.read_csv(emotion_csv_path)
au_df = pd.read_csv(au_csv_path)
openface_df = pd.read_csv(openface_csv_path)
except pd.errors.EmptyDataError:
print(f"Skipping {video_file}: empty CSV files.")
continue
except OSError as e:
print(f"Skipping {video_file}: OSError - {e}")
continue
# Detect events in the video
video_events = detect_events(emotion_df, au_df, openface_df)
all_events.extend(video_events)
# Remove 'End Frame' before saving
for event in all_events:
event.pop('End Frame', None)
# Save all events to a single CSV file
events_df = pd.DataFrame(all_events)
events_df.to_csv(OUTPUT_CSV, index=False)
print(f"Events saved to {OUTPUT_CSV}")