Towards Ubiquitous Mobile Connectivity in Physical Internet
ABSTRACT:Physical Internet envisions renovating digital transportation network in a way to make physical goods delivered with efficiency,
resilience and sustainability in modern logistics. One key factor towards Physical Internet is traceability – with accurate, timely,
and reliable acquisition of goods status in full supply chain and its operational cycle – essentially empowering ubiquitous mobile
connectivity in: 1) high density package location tracking in warehouse inventory, 2) high available V2X cooperation in (last-time)
autonomous urban delivery, and 3) high mobility freight status streaming in networked transportation. This talk will introduce the design,
implementation and deployment experience of our mobile RFID, VILD and multipath networking system for improving scalability, availability
and reliability in logistics, vehicular and high-speed railway networks, towards the vision of (ubiquitous mobile connectivity in) Physical Internet.
Bin:Prof. Chenren Xu is a Boya Young Fellow Associate Professor (with early tenure) and Deputy Director of Institute of Networking and
Energy-efficient Computing in the School of Computer Science at Peking University (PKU) where he directs Wireless AI for Science (WAIS) Lab. His
research interests span wireless, networking and system, with a current focus on backscatter communication for low power IoT connectivity, future
mobile Internet for high mobility data networking, and collaborative edge intelligence system for mobile and IoT computing. He earned his Ph.D. from
WINLAB, Rutgers University, and worked as postdoctoral fellow in Carnegie Mellon University and visiting scholars in AT&T Shannon Labs and
Microsoft Research. He has been active as TPC and OC members in top networking research venues including SIGCOMM, NSDI, MobiCom, MobiSys,
SenSys and INFOCOM. He is an Editor of ACM IMWUT and on the Steering Committee of UbiComp and HotMobile. He is a recipient of NSFC Excellent
Young Scientists Fund (2020), ACM SIGCOMM China Rising Star (2020), Alibaba DAMO Academy Young Fellow (2018) and CCF-Intel Young Faculty (2017)
awards. His work has been featured in MIT Technology Review.
Efficient Multi-modal LLM
ABSTRACT:This talk presents efficient multi-modal LLM innovations across the full stack.
I’ll first present VILA, a visual language model pre-training recipe beyond visual instruction tuning, enabling multi-image reasoning and in-context learning capabilities. Followed by SmoothQuant and AWQ for LLM quantization, and the TinyChat inference library. AWQ and TinyChat enable VILA 3B deployable on Jetson Orin Nano, bringing new capabilities for mobile vision applications. Second, I’ll present efficient representation learning, including EfficientViT for high-resolution vision, accelerating SAM by 48x without performance loss; and condition-aware neural networks for adding control to diffusion models. Third, I’ll present StreamingLLM, a KV cache optimization technique for long conversation and LongLoRA, using sparse, shifted attention for long-context LLM.
Bin:Song Han is an associate professor at MIT EECS. He received his PhD degree from Stanford University. He proposed the “Deep Compression” technique including pruning and quantization that is widely used for efficient AI computing, and “Efficient Inference Engine” that first brought weight sparsity to modern AI chips, which is a top-5 cited paper in 50 years of ISCA. He pioneered the TinyML research that brings deep learning to IoT devices. His team’s recent work on large language model quantization and acceleration (SmoothQuant, AWQ, StreamingLLM) improved the efficiency of LLM inference, adopted by NVIDIA TensorRT-LLM. Song received best paper awards at ICLR and FPGA, faculty awards from Amazon, Facebook, NVIDIA, Samsung and SONY. Song was named “35 Innovators Under 35” by MIT Technology Review, NSF CAREER Award, and Sloan Research Fellowship. Song was the cofounder of DeePhi (now part of AMD), and cofounder of OmniML (now part of NVIDIA). Song developed the EfficientML.ai course to disseminate efficient ML research.