Our research on Large Language Models focuses on architectural improvements, efficient training methods, and practical applications. We investigate how LLMs can be integrated with embodied systems to enable more natural and efficient interactions between agents and their environments.
This research introduces a novel approach for accelerating inference in large language models through the use of residual decoding heads. By optimizing the decoding process, we achieve significant performance improvements without sacrificing model quality, enabling more efficient deployment of LLMs in resource-constrained environments.
We explore methods for fine-tuning large language models while preserving data privacy. Our approach uses synthetic data generation techniques to create training examples that maintain utility while protecting sensitive information, allowing for specialized model adaptation without compromising user privacy.

← Back to Research