Transformers, Diffusion, and Quantum Computing
Latest research summaries in ML, Robotics, CV, NLP and AI
Welcome to today's edition of State of AI 👋 And a warm welcome to our 21 new subscribers since last edition!
Let's get into it 👇
Contents
FlowSpec: Continuous Pipelined Speculative Decoding for Efficient Distributed LLM Inference
MAPS: Advancing Multi-Modal Reasoning in Expert-Level Physical Science
Less is Enough: Training-Free Video Diffusion Acceleration via Runtime-Adaptive Caching
Bootstrapping Grounded Chain-of-Thought in Multimodal LLMs for Data-Efficient Model Adaptation
StructTransform: A Scalable Attack Surface for Safety-Aligned Large Language Models
Gradient-Based Model Fingerprinting for LLM Similarity Detection and Family Classification
Batch-Max: Higher LLM Throughput using Larger Batch Sizes and KV Cache Compression
GPAS: Accelerating Convergence of LLM Pretraining via Gradient-Preserving Activation Scaling
MultiGen: Using Multimodal Generation in Simulation to Learn Multimodal Policies in Real
Large Language Model-Driven Closed-Loop UAV Operation with Semantic Observations
Enabling Population-Level Parallelism in Tree-Based Genetic Programming for Comprehensive GPU Acceleration
Authors: Zhihong Wu, Lishuang Wang, Kebin Sun, Zhuozhao Li, Ran Cheng
Source and references: https://arxiv.org/abs/2501.17168v4
Introduction
This paper presents EvoGP, a high-performance framework that enables comprehensive GPU acceleration of Tree-based Genetic Programming (TGP) through population-level parallel execution.
Keep reading with a 7-day free trial
Subscribe to State of AI to keep reading this post and get 7 days of free access to the full post archives.