Welcome to the Q4C (QForce) Series

Queens For Computing
Queens College CUNY Computer Science Colloquium


This colloquium is intended to bring together Computer Science and Data Science researchers in the tri-state area (especially in NYC) and to foster collaboration. We welcome talks on any topic of interest to the CS community, including theory, algorithms, machine learning, and data science. If you are interested in attending in-person or online, or would like to give a talk, please contact the organizers.


  1. Monday, 09/15/2025, 12:15PM - 1:30PM
    Room: Science Building, C205
    Speaker: Zining Zhu, Stevens Institute of Technology

    Title: Improving LLM reasoning with mechanistic interpretability insights

    Abstract: LLMs have shown great success in tasks requiring reasoning skills, and LLM-based agentic reasoning systems have been used in scenarios including explanation, coding, problem-solving, among many others. Meanwhile, many improvement areas have also been identified, and we look forward to improving the reasoning capabilities. In recent years, mechanistic interpretability offers useful tools and intuitions. In this talk, I introduce some of our recent attempts to apply the tools and intuitions to improving LLM reasoning, and share some exciting preliminary findings along this avenue.

    Zining is an assistant professor at Stevens Institute of Technology department of computer science. He directs the Explainable and Controllable AI lab. The lab’s research involve understanding the mechanisms and abilities of AIs, and incorporating the findings into controlling the AIs. Zining looks forward to building safe, trustworthy agentic AIs that can assist humans to discover knowledge and better perform high-stake tasks. Zining has received paper award at NAACL.

  2. Monday, 09/29/2025, 12:15PM - 1:30PM
    Room: Science Building, C205
    Speaker: Hanfei Yu, Stevens Institute of Technology

    Title: Serverless Computing for AI Systems

    Abstract: Serverless computing is emerging as the next-generation computing paradigm, bridging HPC and cloud cyberinfrastructures with its ease of deployment, instant scalability, and pay-as-you-go pricing. As AI reshapes industries and academia, serverless computing is increasingly explored for large-scale AI training and inference. However, traditional serverless architectures are not optimized for AI workloads, introducing critical performance bottlenecks that hinder their direct applicability. Rethinking serverless computing from an algorithm-system co-design perspective is essential to unlocking its full potential for AI systems.

    In this talk, I will present two case studies—Stellaris and FineMoE—that demonstrate the necessity of algorithm-system co-design for serverless AI workloads. Stellaris introduces a generic asynchronous learning paradigm for distributed deep reinforcement learning (DRL) training, leveraging serverless computing to achieve higher efficiency and lower costs. FineMoE optimizes Mixture-of-Experts (MoE) serving through fine-grained expert offloading, significantly improving memory efficiency while maintaining low inference latency. I will also discuss key challenges and future directions in serverless computing for AI, highlighting opportunities for optimizing AI workloads across cloud and HPC environments.

    Bio: Hanfei Yu is a fifth-year Ph.D. student in the Department of Electrical and Computer Engineering at Stevens Institute of Technology, advised by Prof. Hao Wang. He received his M.S. in Computer Science and Systems from the University of Washington Tacoma and his B.S. in Electronic Engineering from Shanghai Jiao Tong University. Hanfei's research focuses on serverless computing, reinforcement learning systems, LLM serving, and large-scale AI/ML systems. His work aims to develop efficient serverless AI ecosystems that integrate cloud and HPC resources to optimize AI workloads. His research has explored AI/ML-driven techniques to enhance serverless computing efficiency and the design of optimized serverless infrastructures for AI training and inference. He was a Research Intern at Microsoft Azure Research Systems and the Microsoft M365 Systems Innovation Group. Hanfei's contributions have been recognized with the ACM SoCC'24 Best Paper Award and as an ACM/IEEE SC'24 Best Student Paper Finalist. He was also selected as one of the 2025 MLCommons ML and Systems Rising Stars.

  3. Monday, 10/06/2025, 12:15PM - 1:30PM Room: Science Building, C205
    Speaker: Jonathan Gryak, Queens College CUNY

    Title:

    Abstract:

  4. Monday, 10/20/2025, 12:15PM - 1:30PM Room: Science Building, C205
    Speaker: Peter Heller, Queens College CUNY

    Title:

    Abstract:

  5. Monday, 11/03/2025, 12:15PM - 1:30PM Room: Science Building, C205
    Speaker: Yifan Hu, Rutgers University

    Title: Contextual Stochastic Bilevel Optimization and Reinforcement Learning

    Abstract: We introduce contextual stochastic bilevel optimization (CSBO) - a stochastic bilevel optimization framework with the lower-level problem minimizing an expectation conditioned on contextual information and on the upper-level decision variable. We also assume that there may be multiple (or even infinitely many) followers at the lower level. CSBO encapsulates important applications such as meta-learning, personalized federated learning, end-to-end learning, and Wasserstein distributionally robust optimization with side information as special cases. Due to the contextual information at the lower level, existing single-loop methods for classical stochastic bilevel optimization are not applicable. We thus propose an efficient double-loop gradient method based on the Multilevel Monte-Carlo (MLMC) technique. When specialized to stochastic nonconvex optimization, the sample complexity of our method matches existing lower bounds. We further discuss contextual bilevel reinforcement learning which finds various applications from model design, mechanism design, and preference optimization.

    Bio: Yifan Hu is an assistant professor in the Department of Statistics, Rutgers University (New Brunswick, NJ). He was as a postdoc researcher at EPFL and ETH Zurich in Switzerland from 2022 to 2025. Prior to that, he obtained PhD in Operations Research from University of Illinois at Urbana-Champaign in 2022. His research interests lie in optimal decision-making under uncertainty, with an intersection of optimization, statistics, and operations research. In particular, he studies problems arising from reinforcement learning, preference optimization, causal inference and discovery, and operations management, aiming to build new models and develop simple-to-implement methods with provable guarantees.

  6. Monday, 12/01/2025, 12:15PM - 1:30PM Room: Science Building, C205
    Speaker: Jia Xu, Stevens Institute of Technology

    Title:

    Abstract:




The seminar is organized by Jun Li
Email Contact: jun.li@qc.cuny.edu