Chen Feng
-
Institute Associate Professor
Chen Feng is an Associate Professor at NYU. He is interested in active and collaborative robot perception and learning through multidisciplinary use-inspired research that originates from construction, manufacturing, and transportation domains. Before NYU, Chen was a research scientist in the computer vision group at Mitsubishi Electric Research Labs (MERL) in Cambridge, MA, focusing on localization, mapping, and deep learning for self-driving cars and robotics, and invented several patented algorithms. Chen holds a bachelor’s degree in geospatial engineering from Wuhan University in China, a master’s degree in electrical engineering, and a Ph.D. in civil engineering, both from the University of Michigan at Ann Arbor. Chen publishes in and reviews for prestigious AI and Robotics venues like CVPR/ICCV/ICRA/IROS, and he also serves as an associate editor for IEEE Robotics and Automation Letters (RA-L). Chen is a recipient of the NSF CAREER Award in 2023. More information on his research can be found at https://ai4ce.github.io/.
Chen leads the AI4CE lab (pronounced as "A-I-force") with students from several departments (CSE, MAE, CUE, ECE) and schools (Tandon and Courant) at NYU. The AI4CE lab conducts multidisciplinary use-inspired research to develop novel algorithms and systems for intelligent agents to accurately understand and efficiently interact with materials and humans in dynamic and unstructured environments. We aim to fundamentally advance Robotics and AI in areas such as localization, mapping, navigation, mobile manipulation, and scene understanding, to address infrastructure challenges on Earth and beyond, including construction robotics, manufacturing automation, and autonomous vehicles.
Research News
New AI Language-Vision Models Transform Traffic Video Analysis to Improve Road Safety
New York City's thousands of traffic cameras capture endless hours of footage each day, but analyzing that video to identify safety problems and implement improvements typically requires resources that most transportation agencies don't have.
Now, researchers at NYU Tandon School of Engineering have developed an artificial intelligence system that can automatically identify collisions and near-misses in existing traffic video by combining language reasoning and visual intelligence, potentially transforming how cities improve road safety without major new investments.
Published in the journal Accident Analysis and Prevention, the research won New York City's Vision Zero Research Award, an annual recognition of work that aligns with the City's road safety priorities and offers actionable insights. Professor Kaan Ozbay, the paper's senior author, presented the study at the eighth annual Research on the Road symposium on November 19.
The work exemplifies cross-disciplinary collaboration between computer vision experts from NYU's new Center for Robotics and Embodied Intelligence and transportation safety researchers at NYU Tandon's C2SMART center, where Ozbay serves as Director.
By automatically identifying where and when collisions and near-misses occur, the team’s system — called SeeUnsafe — can help transportation agencies pinpoint dangerous intersections and road conditions that need intervention before more serious accidents happen. It leverages pre-trained AI models that can understand both images and text, representing one of the first applications of multimodal large language models to analyze long-form traffic videos.
"You have a thousand cameras running 24/7 in New York City. Having people examine and analyze all that footage manually is untenable," Ozbay said. "SeeUnsafe gives city officials a highly effective way to take full advantage of that existing investment."
"Agencies don't need to be computer vision experts. They can use this technology without the need to collect and label their own data to train an AI-based video analysis model," added NYU Tandon Associate Professor Chen Feng, a co-founding director of the Center for Robotics and Embodied Intelligence, and paper co-author.
Tested on the Toyota Woven Traffic Safety dataset, SeeUnsafe outperformed other models, correctly classifying videos as collisions, near-misses, or normal traffic 76.71% of the time. The system can also identify which specific road users were involved in critical events, with success rates reaching up to 87.5%.
Traditionally, traffic safety interventions are implemented only after accidents occur. By analyzing patterns of near-misses — such as vehicles passing too close to pedestrians or performing risky maneuvers at intersections — agencies can proactively identify danger zones. This approach enables the implementation of preventive measures like improved signage, optimized signal timing, and redesigned road layouts before serious accidents take place.
The system generates “road safety reports” — natural language explanations for its decisions, describing factors like weather conditions, traffic volume, and the specific movements that led to near-misses or collisions.
While the system has limitations, including sensitivity to object tracking accuracy and challenges with low-light conditions, it establishes a foundation for using AI to “understand” road safety context from vast amounts of traffic footage. The researchers suggest the approach could extend to in-vehicle dash cameras, potentially enabling real-time risk assessment from a driver's perspective.
The research adds to a growing body of work from C2SMART that can improve New York City's transportation systems. Recent projects include studying how heavy electric trucks could strain the city's roads and bridges, analyzing how speed cameras change driver behavior across different neighborhoods, developing a “digital twin” that can find smarter routing to reduce FDNY response times, and a multi-year collaboration with the City to monitor the Brooklyn-Queens Expressway for damage-causing overweight vehicles.
In addition to Ozbay and Feng, the paper's authors are lead author Ruixuan Zhang, a Ph.D. student in transportation engineering at NYU Tandon; Beichen Wang and Juexiao Zhang, both graduate students from NYU's Courant Institute of Mathematical Sciences; and Zilin Bian, a recent NYU Tandon Ph.D. graduate now an assistant professor at Rochester Institute of Technology.
Funding for the research came from the National Science Foundation and the U.S. Department of Transportation's University Transportation Centers Program.
Ruixuan Zhang, Beichen Wang, Juexiao Zhang, Zilin Bian, Chen Feng, Kaan Ozbay,
When language and vision meet road safety: Leveraging multimodal large language models for video-based traffic accident analysis, Accident Analysis & Prevention, Volume 219,2025, 108077,ISSN 0001-4575,https://doi.org/10.1016/j.aap.2025.108077.
AI-powered and Robot-assisted Manufacturing for Modular Construction
Semiha Ergan, assistant professor in the Departments of Civil and Urban Engineering, and Computer Science and Engineering and Chen Feng, assistant professor in the departments of Civil and Urban and Mechanical and Aerospace Engineering will lead this project.
Modular construction, with an established record of accelerating projects and reducing costs, is a revolutionary way to transform the construction industry. However, new construction capabilities are needed to perform modular construction at scale, where, as is the case in factories, the industry suffers from the dependency on skilled labor. Among the challenges this project aims to address:
- Every project is unique and requires efficiency and accuracy in recognition and handling workpieces
- Design and production-line changes are common, and require design standardization and optimization of modules, and
- Production lines are complex in space and time, and necessitate the guidance of workers while processing design and installation information accurately
To focus on these challenges while exploring modular construction within the context of Future Manufacturing (FM), this project exploits opportunities at the intersection of AI/robotics/building information modeling and manufacturing, with the potential to increase the scalability of modular construction.
The research will pioneer initial formulations to enable (a) high throughput in manufacturing through the definition and evaluation of processes that embrace real-time workpiece semantic grounding and in-situ AR-robotic assistance, (b) feasibility studies of optimizing and standardizing the design of modules, and utilization of a cyber-infrastructure for their standardization, (c) prototyping cyber-infrastructures as both novel ways of forming academia and industry partnerships, and data infrastructures to accelerate data-driven adaption in FM for modular construction, and (d) synergistic activities with a two-year institution to train and educate FM workforce for the potential of FM and technologies evaluated.
The team argues that, while the evaluations of technologies will focus on the modular construction, the proposed technologies could make manufacturing industries more competitive, particularly heavy manufacturing industries that share similar challenges such as agricultural, mining, and ship building. The project will therefore enhance U.S. competitiveness in production, bolster economic growth, educate students, and influence workforce behavior towards efficiency and accuracy with the skills required for leadership in FM.