+

Research Posts

Agent S: An Open Agentic Framework that Uses Computers Like a Human

Papers with Code
Reporter Naomi Wilson

By Naomi Wilson

Posted on: October 14, 2024

Agent S: An Open Agentic Framework that Uses Computers Like a Human

We present Agent S, an open agentic framework that enables autonomous interaction with computers through a Graphical User Interface (GUI), aimed at transforming human-computer interaction by automating complex, multi-step tasks. Agent S aims to address three key challenges in automating computer tas...

Read More

SageAttention: Accurate 8-Bit Attention for Plug-and-play Inference Acceleration

Papers with Code
Reporter Javier Vásquez

By Javier Vásquez

Posted on: October 07, 2024

SageAttention: Accurate 8-Bit Attention for Plug-and-play Inference Acceleration

The transformer architecture predominates across various models. As the heart of the transformer, attention has a computational complexity of O(N^2), compared to O(N) for linear transformations. When handling large sequence lengths, attention becomes the primary time-consuming component. Although qu...

Read More

On Uncertainty In Natural Language Processing

Papers with Code
Reporter Naomi Wilson

By Naomi Wilson

Posted on: October 07, 2024

On Uncertainty In Natural Language Processing

The last decade in deep learning has brought on increasingly capable systems that are deployed on a wide variety of applications. In natural language processing, the field has been transformed by a number of breakthroughs including large language models, which are used in increasingly many user-faci...

Read More

Lightning UQ Box: A Comprehensive Framework for Uncertainty Quantification in Deep Learning

Papers with Code
Reporter Kate Martin

By Kate Martin

Posted on: October 07, 2024

Lightning UQ Box: A Comprehensive Framework for Uncertainty Quantification in Deep Learning

Uncertainty quantification (UQ) is an essential tool for applying deep neural networks (DNNs) to real world tasks, as it attaches a degree of confidence to DNN outputs. However, despite its benefits, UQ is often left out of the standard DNN workflow due to the additional technical knowledge required...

Read More

Exploring the Benefit of Activation Sparsity in Pre-training

Papers with Code
Reporter Kate Martin

By Kate Martin

Posted on: October 07, 2024

Exploring the Benefit of Activation Sparsity in Pre-training

Pre-trained Transformers inherently possess the characteristic of sparse activation, where only a small fraction of the neurons are activated for each token. While sparse activation has been explored through post-training methods, its potential in pre-training remains untapped. In this work, we firs...

Read More

Choices are More Important than Efforts: LLM Enables Efficient Multi-Agent Exploration

Papers with Code
Reporter Naomi Wilson

By Naomi Wilson

Posted on: October 04, 2024

Choices are More Important than Efforts: LLM Enables Efficient Multi-Agent Exploration

With expansive state-action spaces, efficient multi-agent exploration remains a longstanding challenge in reinforcement learning. Although pursuing novelty, diversity, or uncertainty attracts increasing attention, redundant efforts brought by exploration without proper guidance choices poses a pract...

Read More

CoT-ST: Enhancing LLM-based Speech Translation with Multimodal Chain-of-Thought

Papers with Code
Reporter Javier Vásquez

By Javier Vásquez

Posted on: October 02, 2024

CoT-ST: Enhancing LLM-based Speech Translation with Multimodal Chain-of-Thought

Speech Language Models (SLMs) have demonstrated impressive performance on speech translation tasks. However, existing research primarily focuses on direct instruction fine-tuning and often overlooks the inherent reasoning capabilities of SLMs. In this paper, we introduce a three-stage training frame...

Read More

Simple and Fast Distillation of Diffusion Models

Papers with Code
Reporter Naomi Wilson

By Naomi Wilson

Posted on: October 02, 2024

Simple and Fast Distillation of Diffusion Models

Diffusion-based generative models have demonstrated their powerful performance across various tasks, but this comes at a cost of the slow sampling speed. To achieve both efficient and high-quality synthesis, various distillation-based accelerated sampling methods have been developed recently. Howeve...

Read More