<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Nelson-Bighetti | 2025 Conference on Predictive Inference in Sports</title><link>https://predinfsports.netlify.app/author/nelson-bighetti/</link><atom:link href="https://predinfsports.netlify.app/author/nelson-bighetti/index.xml" rel="self" type="application/rss+xml"/><description>Nelson-Bighetti</description><generator>Hugo Blox Builder (https://hugoblox.com)</generator><language>en-us</language><lastBuildDate>Mon, 01 Apr 2024 00:00:00 +0000</lastBuildDate><image><url>https://predinfsports.netlify.app/media/logo_hu18d0b29f94f4a496c38e2169fb1d271f_138914_300x300_fit_lanczos_3.png</url><title>Nelson-Bighetti</title><link>https://predinfsports.netlify.app/author/nelson-bighetti/</link></image><item><title>Sparsity In Deep Neural Nets</title><link>https://predinfsports.netlify.app/event/on-device-llm/</link><pubDate>Mon, 01 Apr 2024 00:00:00 +0000</pubDate><guid>https://predinfsports.netlify.app/event/on-device-llm/</guid><description>&lt;p>Large Language Models (LLMs) have captured the attention of the tech world with their remarkable common-sense reasoning and generalizability. However, their large size and server transfer requirements can make them resource-intensive and slow, which is problematic for use in mobile or wearable devices like smart glasses and smart watches. Moreover, on-device computing could offer a solution to privacy concerns by keeping sensitive data, such as text messages or photos, on the device itself. To tackle these challenges, we’ve developed a more compact language model, ranging from 0.5B to 1.4B parameters. This model is designed to run on-device, providing a competitive performance for conversational grounded tasks, while also managing latency and memory usage effectively.&lt;/p></description></item></channel></rss>