A multidisciplinary research group advancing the frontiers of artificial intelligence, brain-computer interfaces, and neuro-inspired systems — to amplify human potential and drive positive change.
AImplifier R&D is a multidisciplinary team committed to pushing the frontiers of artificial intelligence to drive positive change across humanity, education, and science. We strive to understand and amplify intelligence — both human and machine — and harness its power to shape a better, more equitable world.
Bridging AI and brain-computer interfaces to deepen our understanding of intelligence, cognition, and neuro-diseases like epilepsy.
Reading Groups, AI talks, and competitions that nurture the next generation of tech talent across AI, data science, and biomedical innovation.
Building efficient models for solar energy prediction, seizure detection, sleep disorder diagnostics, and climate policy optimization.
From EEG diffusion models to reinforcement learning environments and training-free model merging, our work spans neuroscience, machine learning, and agentic AI.
A self-supervised diffusion model for learning rich EEG signal representations, featuring a specialized State-Space Model architecture and latent fusion transformer — applied to downstream tasks such as seizure detection.
View on GitHubAn open-source platform democratising EEG and brain source imaging for epilepsy diagnosis and neurological research, integrating AI and cloud computing to make advanced neuroimaging accessible in clinical practice.
A physics-based reinforcement learning environment simulating competitive pickleball gameplay, where AI agents learn to play and compete autonomously — serving as the challenge domain for a national AI, HCI & BCI competition.
A training-free framework for merging multiple fine-tuned models across different tasks, mitigating performance degradation through neuronal subspace decomposition to enable efficient multi-task model composition.
View on GitHubLightweight adaptation strategies enabling agentic AI systems to generalise rapidly across diverse tasks and environments.
Research into alignment, robustness, and adversarial defence for autonomous AI agents operating in high-stakes settings.
Deploying multi-agent workflows for real-world enterprise automation, decision support, and intelligent process orchestration.
Our work appears in top-tier venues including ACL and EMNLP, covering model merging, EEG representation learning, and sleep stage classification.
Puah, J.H., Goh, S.K., Zhang, Z., Ye, Z., Chan, C.K., Lim, K.S., Fong, S.L. and Woon, K.S., 2025. EEGDM: EEG Representation Learning via Generative Diffusion Model. arXiv:2508.14086.
Chin, B.W.H., Yew, Y.T., Wu, H., Liang, L., Chan, C.K., Zain, N.M., Samdin, S.B. and Goh, S.K., 2025. SleepDIFFormer: Sleep Stage Classification via Multivariate Differential Transformer. arXiv:2508.15215.
Fang, Z., Du, G., Yu, S., Guo, Y., Zhang, Y., Cao, Y., Li, J., Tang, H.K. and Goh, S.K., 2025. To See a World in a Spark of Neuron: Disentangling Multi-task Interference for Training-free Model Merging. arXiv:2503.05320.
Du, G., Fang, Z., Li, J., Li, J., Jiang, R., Yu, S., Guo, Y., Chen, Y., Goh, S.K., Tang, H.K. and He, D., 2025. Neural Parameter Search for Slimmer Fine-Tuned Models and Better Transfer. arXiv:2505.18713.
Li, J., Du, G., Li, J., Goh, S.K., Wang, W., Wang, Y., Liu, F., Tang, H.K., Alharbi, S., He, D. and Zhang, M., 2025. Multi-modality Expansion and Retention for LLMs through Parameter Merging and Decoupling. arXiv:2505.17110.
A diverse group of researchers, advisors, and students working at the intersection of AI, neuroscience, and systems engineering.
We partner with leading universities and medical centres worldwide to advance our research and expand its impact.