Skip to content

Commit

Permalink
fix
Browse files Browse the repository at this point in the history
  • Loading branch information
IlyasMoutawwakil committed Feb 19, 2024
1 parent 2fe3d85 commit 3211d16
Show file tree
Hide file tree
Showing 4 changed files with 11 additions and 15 deletions.
6 changes: 1 addition & 5 deletions Makefile
Original file line number Diff line number Diff line change
@@ -1,9 +1,5 @@
# List of targets that are not associated with files
.PHONY: quality style install \
build_docker_cpu, build_docker_cuda, build_docker_rocm, \
test_cli_cpu_pytorch, test_cli_rocm_pytorch, \
test_cli_cpu_neural_compressor, test_cli_cpu_onnxruntime, test_cli_cpu_openvino, \
test_api_cpu, test_api_cuda, test_api_rocm, test_api_misc
.PHONY: quality style install build_docker_cpu build_docker_cuda build_docker_rocm test_cli_cpu_neural_compressor test_cli_cpu_onnxruntime test_cli_cpu_openvino test_cli_cpu_pytorch test_cli_rocm_pytorch test_cli_cuda_pytorch test_api_cpu test_api_cuda test_api_rocm test_api_misc

quality:
ruff check .
Expand Down
16 changes: 8 additions & 8 deletions optimum_benchmark/benchmarks/inference/benchmark.py
Original file line number Diff line number Diff line change
Expand Up @@ -79,25 +79,25 @@ def run(self, backend: Backend[BackendConfigT]) -> None:
)

if backend.config.task in TEXT_GENERATION_TASKS:
LOGGER.info("\t+ Generating and preparing Text Generation input")
LOGGER.info("\t+ Generating and preparing Text Generation inputs")
self.text_generation_inputs = self.input_generator()
self.text_generation_inputs = backend.prepare_inputs(self.text_generation_inputs)
self.text_generation_inputs = {"input_ids": self.text_generation_inputs["input_ids"]}
LOGGER.info("\t+ Updating Text Generation kwargs with default values")
self.config.generate_kwargs = {**TEXT_GENERATION_KWARGS, **self.config.generate_kwargs}
LOGGER.info("\t+ Initializing Text Generation report")
self.report = TextGenerationReport(prefill=BenchmarkMeasurements(), decode=BenchmarkMeasurements())

elif backend.config.task in IMAGE_DIFFUSION_TASKS:
LOGGER.info("\t+ Generating and preparing Image Diffusion input")
LOGGER.info("\t+ Generating Image Diffusion inputs")
self.image_diffusion_inputs = self.input_generator()
self.image_diffusion_inputs = backend.prepare_inputs(self.image_diffusion_inputs)
LOGGER.info("\t+ Updating Image Diffusion kwargs with default values")
self.config.call_kwargs = {**IMAGE_DIFFUSION_KWARGS, **self.config.call_kwargs}
LOGGER.info("\t+ Initializing Image Diffusion report")
self.report = ImageDiffusionReport(call=BenchmarkMeasurements())

else:
LOGGER.info("\t+ Generating and preparing Inference input")
LOGGER.info("\t+ Generating and preparing Inference inputs")
self.inference_inputs = self.input_generator()
self.inference_inputs = backend.prepare_inputs(self.inference_inputs)
LOGGER.info("\t+ Initializing Inference report")
Expand Down Expand Up @@ -180,7 +180,7 @@ def run_image_diffusion_memory_tracking(self, backend: Backend):
LOGGER.info("\t+ Running memory tracking")
self.memory_tracker.reset()
with self.memory_tracker.track():
_ = backend.call(self.image_diffusion_inputs, self.config.forward_kwargs)
_ = backend.call(self.image_diffusion_inputs, self.config.call_kwargs)

self.report.call.memory = self.memory_tracker.get_max_memory()

Expand Down Expand Up @@ -224,7 +224,7 @@ def run_image_diffusion_latency_tracking(self, backend: Backend):
self.latency_tracker.reset()
while self.latency_tracker.get_elapsed_time() < self.config.duration:
with self.latency_tracker.track():
_ = backend.call(self.image_diffusion_inputs, self.config.forward_kwargs)
_ = backend.call(self.image_diffusion_inputs, self.config.call_kwargs)

self.report.call.latency = self.latency_tracker.get_latency()
self.report.call.throughput = Throughput.from_latency(
Expand Down Expand Up @@ -268,7 +268,7 @@ def run_image_diffusion_energy_tracking(self, backend: Backend):
LOGGER.info("\t+ Running energy tracking")
self.energy_tracker.reset()
with self.energy_tracker.track():
_ = backend.call(self.image_diffusion_inputs, self.config.forward_kwargs)
_ = backend.call(self.image_diffusion_inputs, self.config.call_kwargs)

self.report.call.energy = self.energy_tracker.get_energy()
self.report.call.efficiency = Efficiency.from_energy(
Expand Down Expand Up @@ -296,7 +296,7 @@ def prefill_volume(self) -> int: # in tokens

@property
def call_volume(self) -> int: # in images
return self.config.input_shapes["batch_size"] * self.config.forward_kwargs["num_images_per_prompt"]
return self.config.input_shapes["batch_size"] * self.config.call_kwargs["num_images_per_prompt"]

@property
def decode_volume(self) -> int: # in tokens
Expand Down
2 changes: 1 addition & 1 deletion tests/configs/_bert_sweep_.yaml
Original file line number Diff line number Diff line change
@@ -1,5 +1,5 @@
hydra:
sweeper:
params:
backend.model: hf-internal-testing/tiny-random-bert
backend.task: fill-mask,text-classification,token-classification,question-answering
backend.model: hf-internal-testing/tiny-random-bert,hf-internal-testing/tiny-random-roberta
2 changes: 1 addition & 1 deletion tests/configs/_diffusers_.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -4,5 +4,5 @@ backend:
model: hf-internal-testing/tiny-stable-diffusion-torch

benchmark:
forward_kwargs:
call_kwargs:
num_inference_steps: 2

0 comments on commit 3211d16

Please sign in to comment.