Profile Job Results

Jobs
jp01x7e2g
Results Ready
Name
whisper_medium_decoder
Target Device
  • QCS8550 (Proxy)
  • Android 12
  • Qualcomm® QCS8550
Creator
ai-hub-support@qti.qualcomm.com
Target Model
Input Specs
input_ids: int32[1, 1]
attention_mask: float16[1, 1, 1, 200]
k_cache_self_0_in: float16[16, 1, 64, 199]
v_cache_self_0_in: float16[16, 1, 199, 64]
k_cache_self_1_in: float16[16, 1, 64, 199]
v_cache_self_1_in: float16[16, 1, 199, 64]
k_cache_self_2_in: float16[16, 1, 64, 199]
v_cache_self_2_in: float16[16, 1, 199, 64]
k_cache_self_3_in: float16[16, 1, 64, 199]
v_cache_self_3_in: float16[16, 1, 199, 64]
k_cache_self_4_in: float16[16, 1, 64, 199]
v_cache_self_4_in: float16[16, 1, 199, 64]
k_cache_self_5_in: float16[16, 1, 64, 199]
v_cache_self_5_in: float16[16, 1, 199, 64]
k_cache_self_6_in: float16[16, 1, 64, 199]
v_cache_self_6_in: float16[16, 1, 199, 64]
k_cache_self_7_in: float16[16, 1, 64, 199]
v_cache_self_7_in: float16[16, 1, 199, 64]
k_cache_self_8_in: float16[16, 1, 64, 199]
v_cache_self_8_in: float16[16, 1, 199, 64]
k_cache_self_9_in: float16[16, 1, 64, 199]
v_cache_self_9_in: float16[16, 1, 199, 64]
k_cache_self_10_in: float16[16, 1, 64, 199]
v_cache_self_10_in: float16[16, 1, 199, 64]
k_cache_self_11_in: float16[16, 1, 64, 199]
v_cache_self_11_in: float16[16, 1, 199, 64]
k_cache_self_12_in: float16[16, 1, 64, 199]
v_cache_self_12_in: float16[16, 1, 199, 64]
k_cache_self_13_in: float16[16, 1, 64, 199]
v_cache_self_13_in: float16[16, 1, 199, 64]
k_cache_self_14_in: float16[16, 1, 64, 199]
v_cache_self_14_in: float16[16, 1, 199, 64]
k_cache_self_15_in: float16[16, 1, 64, 199]
v_cache_self_15_in: float16[16, 1, 199, 64]
k_cache_self_16_in: float16[16, 1, 64, 199]
v_cache_self_16_in: float16[16, 1, 199, 64]
k_cache_self_17_in: float16[16, 1, 64, 199]
v_cache_self_17_in: float16[16, 1, 199, 64]
k_cache_self_18_in: float16[16, 1, 64, 199]
v_cache_self_18_in: float16[16, 1, 199, 64]
k_cache_self_19_in: float16[16, 1, 64, 199]
v_cache_self_19_in: float16[16, 1, 199, 64]
k_cache_self_20_in: float16[16, 1, 64, 199]
v_cache_self_20_in: float16[16, 1, 199, 64]
k_cache_self_21_in: float16[16, 1, 64, 199]
v_cache_self_21_in: float16[16, 1, 199, 64]
k_cache_self_22_in: float16[16, 1, 64, 199]
v_cache_self_22_in: float16[16, 1, 199, 64]
k_cache_self_23_in: float16[16, 1, 64, 199]
v_cache_self_23_in: float16[16, 1, 199, 64]
k_cache_cross_0: float16[16, 1, 64, 1500]
v_cache_cross_0: float16[16, 1, 1500, 64]
k_cache_cross_1: float16[16, 1, 64, 1500]
v_cache_cross_1: float16[16, 1, 1500, 64]
k_cache_cross_2: float16[16, 1, 64, 1500]
v_cache_cross_2: float16[16, 1, 1500, 64]
k_cache_cross_3: float16[16, 1, 64, 1500]
v_cache_cross_3: float16[16, 1, 1500, 64]
k_cache_cross_4: float16[16, 1, 64, 1500]
v_cache_cross_4: float16[16, 1, 1500, 64]
k_cache_cross_5: float16[16, 1, 64, 1500]
v_cache_cross_5: float16[16, 1, 1500, 64]
k_cache_cross_6: float16[16, 1, 64, 1500]
v_cache_cross_6: float16[16, 1, 1500, 64]
k_cache_cross_7: float16[16, 1, 64, 1500]
v_cache_cross_7: float16[16, 1, 1500, 64]
k_cache_cross_8: float16[16, 1, 64, 1500]
v_cache_cross_8: float16[16, 1, 1500, 64]
k_cache_cross_9: float16[16, 1, 64, 1500]
v_cache_cross_9: float16[16, 1, 1500, 64]
k_cache_cross_10: float16[16, 1, 64, 1500]
v_cache_cross_10: float16[16, 1, 1500, 64]
k_cache_cross_11: float16[16, 1, 64, 1500]
v_cache_cross_11: float16[16, 1, 1500, 64]
k_cache_cross_12: float16[16, 1, 64, 1500]
v_cache_cross_12: float16[16, 1, 1500, 64]
k_cache_cross_13: float16[16, 1, 64, 1500]
v_cache_cross_13: float16[16, 1, 1500, 64]
k_cache_cross_14: float16[16, 1, 64, 1500]
v_cache_cross_14: float16[16, 1, 1500, 64]
k_cache_cross_15: float16[16, 1, 64, 1500]
v_cache_cross_15: float16[16, 1, 1500, 64]
k_cache_cross_16: float16[16, 1, 64, 1500]
v_cache_cross_16: float16[16, 1, 1500, 64]
k_cache_cross_17: float16[16, 1, 64, 1500]
v_cache_cross_17: float16[16, 1, 1500, 64]
k_cache_cross_18: float16[16, 1, 64, 1500]
v_cache_cross_18: float16[16, 1, 1500, 64]
k_cache_cross_19: float16[16, 1, 64, 1500]
v_cache_cross_19: float16[16, 1, 1500, 64]
k_cache_cross_20: float16[16, 1, 64, 1500]
v_cache_cross_20: float16[16, 1, 1500, 64]
k_cache_cross_21: float16[16, 1, 64, 1500]
v_cache_cross_21: float16[16, 1, 1500, 64]
k_cache_cross_22: float16[16, 1, 64, 1500]
v_cache_cross_22: float16[16, 1, 1500, 64]
k_cache_cross_23: float16[16, 1, 64, 1500]
v_cache_cross_23: float16[16, 1, 1500, 64]
position_ids: int32[1]
Completion Time
4/18/2026, 9:07:02 AM
Options
--qairt_version 2.42
Versions
  • ONNX Runtime: 1.24.3
  • QAIRT: v2.42.0.251225135753_193295
  • Android: 13 (TP1A.220624.014)
  • AI Hub: aihub-2026.04.13.0
Estimated Inference Time
33.9 ms
Estimated Peak Memory Usage
160 ‑ 162 MB
Compute Units
NPU
5938
StageTimeMemory
First App Load
2.80 s872‑873 MB
Subsequent App Load
1.19 s882‑885 MB
Inference
33.9 ms160‑162 MB
ONNX RuntimeValue
execution_modeSEQUENTIAL
intra_op_num_threads0
inter_op_num_threads0
enable_memory_patternfalse
enable_cpu_memory_arenafalse
graph_optimization_levelENABLE_ALL
QNN Execution ProviderValue
htp_performance_mode"burst"
htp_graph_finalization_optimization_mode"3"
enable_htp_fp16_precision"1"
capture_network_visualizationsfalse
context_priority"normal"
offload_graph_io_quantization"1"

Sign up to run this model on a hosted Qualcomm® device!

Run on device