π PhD in Informatics (Shizuoka University, 2024)
π¬ Research Focus: NLP β’ Multimodal ML β’ Vision-Language Models β’ Explainable AI
π Currently: Product Development Engineer at Pi Photonics, Japan
-
Cross-Lingual Vision-Language Models: Evaluating CLIP's performance on Japanese vs English memes (repo)
-
Dual-Rationale Attention Mechanisms: Novel architecture for sarcasm-aware hate speech detection
-
Industrial AI: OAK-D camera-based safety monitoring systems
Currently developing AI systems for industrial applications:
- Edge AI for Safety Monitoring: Camera-based crane detection/ wave detection/ human detection on embedded platforms
- Real-time Video Analysis: Feature extraction and object detection on STM32 microcontrollers
- Firmware Development: Implementing ML inference pipelines on resource-constrained devices
Combining research expertise in deep learning with hands-on embedded systems deployment.
-
Mamun, M.B., Tsunakawa, T., Nishida, M., Nishimura, M. (2024). "Hate Speech Detection by Using Rationales for Judging Sarcasm." Applied Sciences, 14(11), 4898.
-
Mamun, M.B. (2025). "Sarcasm-Aware Hate Speech Detection Using Rationales." Journal of the Japanese Society for Artificial Intelligence, 40(1), 1058-1059.
Deep Learning & NLP: PyTorch β’ Transformers (BERT, RoBERTa, DeBERTa, CLIP) β’ Attention Mechanisms β’ Multi-task Learning β’ Vision-Language Pre-training
Computer Vision & Edge AI: OpenCV β’ OAK-D Depth Cameras β’ Real-time Object Detection β’ Video Processing β’ Embedded Inference
Embedded Systems: STM32 β’ Raspberry Pi β’ Firmware Development β’ Real-time Feature Extraction β’ Industrial Safety Applications
Tools & Frameworks: Hugging Face Transformers β’ scikit-learn β’ TensorFlow β’ C++ β’ Python
π‘ Note: My current industry work at Pi Photonics is maintained on the company's private GitLab server. This GitHub showcases my independent research in multimodal ML and NLP.
