Aria Ghora Prabono

Technical Profile
ML/AI Research Computer Vision • NLP • Time-Series • Transfer Learning PyTorch, Transformers, CLIP, Diffusion Models
Performance CUDA • ONNX • Compute Shaders • Quantization Kernel fusion, memory bandwidth optimization
Languages Python • C • Rust • Odin ML model development lifecycle, FFI bindings, SIMD-optimized computation, bare metal when needed
Infrastructure Docker • AWS • MLflow • DuckDB Cloud deployment, reproducible environments, experiment tracking, ETL
Trajectory
2025–Present AI Principal Engineer EAGLYS, Tokyo, Japan
2025–2025 Project AI Lead
2022–2025 AI Research Engineer
2021–2022 ML Engineer Nomura Research Institute, Jakarta, Indonesia
2021–Present Independent Consultant Self
2016–2021 Integrated MS-PhD, Industrial Information Systems Engineering HUFS, Yongin, South Korea
Technical Philosophy
Core Principle Complexity is debt. Every abstraction must earn its place. Every dependency is liability.
Methodology Measure first, optimize on evidence. Start from fundamentals, not frameworks.
Performance Most slowness is broken defaults. Fix those first, then go deep.
Avoid Premature abstraction • Optimization without profiling • Accepting "that's just how it is"
GitHub LinkedIn Email