Extended Reality (XR) offers immersive experiences and has broad applications across education, training, manufacturing, transportation, and healthcare. In this talk, we first present several of our developed XR prototypes and highlight the growing need for natural language interaction within XR environments -- such as generating and animating 3D scenes directly from spoken input. Achieving this vision requires seamless integration of Large Language Models (LLMs), such as GPT-4, with XR systems. However, the complexity of XR environments makes it difficult to accurately extract relevant contextual data and scene/object parameters from an overwhelming volume of XR artifacts. This not only increases costs under pay-per-use LLM models, but also elevates the risk of generation errors and privacy concerns. To address these challenges, we introduce a novel framework for creating interactive XR worlds using structured JSON data generated by LLMs. Unlike prior work that focuses on generating executable code, LLMER translates natural language input into lightweight and interpretable JSON representations, significantly reducing the risk of application crashes and minimizing processing latency. It adopts a multi-stage pipeline that delivers only essential, context-aware information tailored to the user’s request, and incorporates multiple modules optimized for various XR tasks.
Bin Li is an associate professor at the School of Electrical Engineering and Computer Science at Penn State. He received Ph.D. degree in Electrical and Computer Engineering from The Ohio State University (OSU). He was a Postdoctoral Researcher in the Coordinated Science Lab at the University of Illinois at Urbana-Champaign (UIUC). His research focuses on immersive networking and computing, digital twins, mobile edge computing, stochastic network optimization, and machine learning. He received the NSF Career Award and Google Faculty Research Award. He has extensive experience in the development and optimization of extended reality (XR) systems and won Best Poster/Demo Awards at IEEE INFOCOM, ACM MobiHoc, and IEEE VR. His research has been sponsored by NSF, ARO, DoE, and DARPA.

