Events

Sparsity In Deep Neural Nets

In this talk, we will introduce a novel sparsity in deep neural nets. Lorem ipsum dolor sit amet, consectetur adipiscing elit. Nulla quis commodo augue. Nam et lacinia sapien. Maecenas pellentesque erat quis ante congue, at accumsan ipsum ullamcorper. Etiam iaculis euismod ex, vitae interdum ipsum imperdiet non. Vestibulum bibendum egestas rhoncus. Suspendisse ut lacinia ligula, vitae elementum mi. Nullam accumsan nunc sapien. Aliquam turpis diam, suscipit in faucibus ac, consequat mollis orci. Donec sit amet auctor risus.

Apr 1, 2024

Sparsity In Deep Neural Nets

Large Language Models (LLMs) have captured the attention of the tech world with their remarkable common-sense reasoning and generalizability. However, their large size and server transfer requirements can make them resource-intensive and slow, which is problematic for use in mobile or wearable devices like smart glasses and smart watches. Moreover, on-device computing could offer a solution to privacy concerns by keeping sensitive data, such as text messages or photos, on the device itself. To tackle these challenges, we’ve developed a more compact language model, ranging from 0.5B to 1.4B parameters. This model is designed to run on-device, providing a competitive performance for conversational grounded tasks, while also managing latency and memory usage effectively.

Apr 1, 2024

LLM-Based Assistive Recommendation

Large Language Models (LLMs) have demonstrated impressive capabilities for natural language understanding and generation. What if we power recommender systems with LLMs that can reason through user activities on the platform, and can help plan recommendations accordingly? Equipped with UXR insights that people use nuanced language to describe their long-term interest journeys, and they want more user agency in their recommender platforms, we pinpoint a combination of LLMs and personalization as a key ingredient for the next generation of assistive recommendation.

Apr 1, 2024