Skip to content

Commit f928de5

Browse files
committed
Updates from testing: make fewer assumptions about output file locations, minor cleanup of results dict.
Signed-off-by: rlratzel <[email protected]>
1 parent c688b05 commit f928de5

File tree

4 files changed

+13
-16
lines changed

4 files changed

+13
-16
lines changed

benchmarking/run.py

Lines changed: 3 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -109,7 +109,7 @@ def run_entry( # noqa: PLR0915
109109
(session_entry_path / d).absolute() for d in ["scratch", "ray_cluster", "logs", "benchmark_results"]
110110
]
111111

112-
cmd = entry.get_command_to_run(session_entry_path, dataset_resolver)
112+
cmd = entry.get_command_to_run(session_entry_path, benchmark_results_path, dataset_resolver)
113113
run_id = result.get("run_id", f"{entry.name}-{int(time.time())}")
114114

115115
try:
@@ -141,7 +141,7 @@ def run_entry( # noqa: PLR0915
141141
logger.warning(f"\t\t⏰ Timed out after {entry.timeout_s}s")
142142
logger.info(f"\t\tLogs found in {logs_path}")
143143

144-
run_data = {
144+
result.update({
145145
"cmd": cmd,
146146
"started_at": started_at,
147147
"ended_at": time.time(),
@@ -151,11 +151,10 @@ def run_entry( # noqa: PLR0915
151151
"timed_out": completed["timed_out"],
152152
"logs_dir": logs_path,
153153
"success": success,
154-
}
154+
})
155155
ray_data = {}
156156
script_persisted_data = get_entry_script_persisted_data(benchmark_results_path)
157157
result.update({
158-
"run_data": run_data,
159158
"ray_data": ray_data,
160159
"script_persisted_data": script_persisted_data,
161160
})

benchmarking/runner/matrix.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -21,11 +21,11 @@ class MatrixEntry:
2121
# If set, overrides the session-level delete_scratch setting for this entry
2222
delete_scratch: bool | None = None
2323

24-
def get_command_to_run(self, session_entry_path: Path, resolver: DatasetResolver) -> str:
24+
def get_command_to_run(self, session_entry_path: Path, benchmark_results_path: Path, resolver: DatasetResolver) -> str:
2525
if self.script:
2626
script_path = self.script_base_dir / self.script
2727
# FIXME: should --benchmark-results-path always be passed?
28-
cmd = f"python {script_path} {self.args or ''} --benchmark-results-path" + " {session_entry_dir}/benchmark_results"
28+
cmd = f"python {script_path} {self.args or ''} --benchmark-results-path={benchmark_results_path}"
2929

3030
cmd = self.substitute_datasets_in_cmd(cmd, resolver)
3131
cmd = self.substitute_template_placeholders(cmd, session_entry_path)

benchmarking/runner/sinks/slack_sink.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -150,7 +150,7 @@ def _post_style1(self) -> None:
150150
data = [
151151
("name", result["name"]),
152152
("success", result["success"]),
153-
("runtime", f"{result['run_data']['exec_time_s']:.2f} s"),
153+
("runtime", f"{result['exec_time_s']:.2f} s"),
154154
]
155155
left, right = zip(*data)
156156
right = [str(val) for val in right]
@@ -207,7 +207,7 @@ def _post_style2(self) -> None:
207207
data = [
208208
("name", result["name"]),
209209
("success", result["success"]),
210-
("runtime", f"{result['run_data']['exec_time_s']:.2f} s"),
210+
("runtime", f"{result['exec_time_s']:.2f} s"),
211211
]
212212
for (var, val) in data:
213213
row = [

benchmarking/scripts/test_benchmark.py

Lines changed: 6 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -33,10 +33,8 @@ def run_demo_benchmark( # noqa: PLR0913
3333
from nemo_curator.backends.experimental.ray_data import RayDataExecutor
3434

3535
executor = RayDataExecutor()
36-
if use_ray_data_settings:
37-
from ray.data import DataContext
38-
39-
DataContext.get_current().target_max_block_size = 1
36+
#from ray.data import DataContext
37+
#DataContext.get_current().target_max_block_size = 1
4038

4139
elif executor_name == "xenna":
4240
from nemo_curator.backends.xenna import XennaExecutor
@@ -55,15 +53,15 @@ def run_demo_benchmark( # noqa: PLR0913
5553
try:
5654
time.sleep(10)
5755
output_tasks = []
56+
run_time_taken = time.perf_counter() - run_start_time
57+
num_removed = 0
5858
logger.success(f"Benchmark completed in {run_time_taken:.2f}s")
5959
success = True
6060
except Exception as e: # noqa: BLE001
6161
logger.error(f"Benchmark failed: {e}")
6262
output_tasks = []
6363
success = False
64-
finally:
65-
run_time_taken = time.perf_counter() - run_start_time
66-
num_removed = 0
64+
6765
return {
6866
"params": {
6967
"executor": executor_name,
@@ -125,7 +123,7 @@ def main() -> int:
125123
"tasks": [],
126124
}
127125
finally:
128-
write_results(results, args.output_path)
126+
write_results(results, args.benchmark_results_path)
129127

130128
# Return proper exit code based on success
131129
return 0 if results["metrics"]["is_success"] else 1

0 commit comments

Comments
 (0)