AI Algorithms

AI and Machine Learning are rapidly evolving, driven by advances in data, computing, and algorithms. Research for 2023-2024 focuses on enhancing large language and foundation models, with efforts to improve their efficiency, automation, and integration into real-world tasks. Key areas include fine-tuning models for enterprise applications, advancing AI safety and trust, and exploring deep learning theories like self-supervised learning, reinforcement learning, and generative AI. These innovations are shaping the future of AI across industries, from natural language processing to ethical considerations.

Projects

2025

Project RPI Principal Investigators IBM Principal Investigators
Theoretical and Algorithmic Foundations of In-Context Learning and Chain of Thought Using Properly Trained Transformer Models Meng Wang Songtao Lu, Pin-Yu Chen, Xiaodong Cui
Key-Value Cache Compression for Memory-Efficient Large Language Model Inference Mohammad Mohammadi Amiri Pin-Yu Chen, Tejaswini Pedapati, Subhajit Chaudhury
Associative Energy-based Diffusion Models for Generative AI Mohammed J. Zaki Dmitry Krotov, Rogerio S. Feris
Fast Inference and Alignment for Large Multi-modal Models Koushik Kar, Tianyi Chen Parikshit Ram, Nathalie Baracaldo, Yi Zhou, Horst Samulowitz, Ken Wong
Meta-Transfer-Learning for Tabular Data Distillation, Generation, and Predictive Modeling Oshani Seneviratne Horst Samulowitz, Yi Zhou, Parikshit Ram
Interpretable Foundation Models for General-Purpose Time Series Analysis Agung Julius Lam Nguyen
Strategic AI: Enhancing Large Language Model Agents with Decision-Making Algorithms for Advanced Reasoning and Planning Santiago Paternain Dharmashankar Subramanian
Large Language Models as Planning Domain Model Generators Selmer Bringsjord Kavitha Srinivas, Harsha Kokel, Michael Katz, Shirin Sohrabi

2024

Project RPI Principal Investigators IBM Principal Investigators
Resource-Effective Fine-Tuning of Large Language Models Mohammad Mohammadi Amiri Pin-Yu Chen, Tejaswini Pedapati, Subhajit Chaudhury
Testing LLM Safety via Causal Reasoning Ali Tajer Prasanna Sattigeri, Dennis Wei, Dmitrity Katz-Rogozhnikov
Theoretical and Algorithmic Foundations of In-Context Learning Using Properly Trained Transformer Models Meng Wang Songtao Lu, Pin-Yu Chen
Unlearning: Dynamics of Membership Privacy and Inference Attacks Against Large Language Models Lei Yu Magdon Ismail, Nathalie Baracaldo, Ling Cai
Control-Based Reinforcement Learning Santiago Paternain Mark Squillante; Chai Wah Wu
Correctors and Selectors: Building An Ecosystem for LLM Alignment Alex Gittens Mikhail Yurochkin, Mayank Agarwal
Data Distillation in Tabular Date: A Foundation Model Approach Oshani Seneviratne, Inwon Kang Horst Samulowitz, Parikshit Ram, Yi Zhou
Energy Transformer for Foundational Models Mohammed Zaki Dmitry Krotov, Benjamin Hoover, Hendrik Strobelt
FIT: Fast Inference using Transformer models Koushik Kar, Tianyi Chen Parikshit Ram, Nathalie Baracaldo, Yi Zhou, Soham Dan, Horst Samulowitz
Foundational Models for Understanding Tabular Data Through AI Automation Jianxi Gao Kavitha Srinivas, Tejsawini Pedapati, Horst Samulowitz, Pin-Yu Chen
Multi-Objective Training of Foundation Acoustic Models for Automatic Speech Recognition Tianyi Chen, Mei Si Xiaodong Cui, Brian Kingsbury, Songtao Lu
Back to top