Projects per year
Abstract
This paper investigates whether Large Language Models (LLMs), fine-tuned on synthetic but domain-representative data, can perform the twofold task of (i) slot and intent detection and (ii) natural language response generation for a smart home assistant, while running solely on resource-limited, CPU-only edge hardware. We fine-tune LLMs to produce both JSON action calls and text responses. Our experiments show that 16-bit and 8-bit quantized variants preserve high accuracy on slot and intent detection and maintain strong semantic coherence in generated text, while the 4-bit model, while retaining generative fluency, suffers a noticeable drop in device-service classification accuracy. Further evaluations on noisy human (non-synthetic) prompts and out-of-domain intents confirm the models' generalization ability, obtaining around 80--86\% accuracy. While the average inference time is 5--6 seconds per query -- acceptable for one-shot commands but suboptimal for multi-turn dialogue -- our results affirm that an on-device LLM can effectively unify command interpretation and flexible response generation for home automation without relying on specialized hardware.
Original language | English |
---|---|
Title of host publication | Proceedings of the Tenth Workshop on Noisy and User-generated Text (W-NUT 2025) |
Publisher | Association for Computational Linguistics |
Publication date | 29 Apr 2025 |
DOIs | |
Publication status | Published - 29 Apr 2025 |
Keywords
- LLM
- Home Assistant
Fingerprint
Dive into the research topics of 'On-Device LLMs for Home Assistant: Dual Role in Intent Detection and Response Generation'. Together they form a unique fingerprint.Projects
- 1 Active
-
Digital Twins for Abundant Feedback: Novel Feedback Paradigms via Explainable Multilingual Natural Language Processing
Bjerva, J. (PI), Lindsay, E. (PI) & Zhang, M. (Project Participant)
01/01/2024 → 31/12/2025
Project: Research