2024
DOI: 10.3390/s24020347
|View full text |Cite
|
Sign up to set email alerts
|

Using Large Language Models to Enhance the Reusability of Sensor Data

Alberto Berenguer,
Adriana Morejón,
David Tomás
et al.

Abstract: The Internet of Things generates vast data volumes via diverse sensors, yet its potential remains unexploited for innovative data-driven products and services. Limitations arise from sensor-dependent data handling by manufacturers and user companies, hindering third-party access and comprehension. Initiatives like the European Data Act aim to enable high-quality access to sensor-generated data by regulating accuracy, completeness, and relevance while respecting intellectual property rights. Despite data availa… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
3

Relationship

1
2

Authors

Journals

citations
Cited by 3 publications
(3 citation statements)
references
References 24 publications
0
3
0
Order By: Relevance
“…As described in Section 3, the system comprises two primary components: data gathering and data retrieval. Concerning the data gathering component, prior work by the authors [49] evaluated various LLMs, including ChatGPT and Llama 2, to assess their performance in transforming raw sensor data into a structured format. The highest precision and recall achieved were 93.51% and 85.33%, respectively, demonstrating their suitability for the task.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…As described in Section 3, the system comprises two primary components: data gathering and data retrieval. Concerning the data gathering component, prior work by the authors [49] evaluated various LLMs, including ChatGPT and Llama 2, to assess their performance in transforming raw sensor data into a structured format. The highest precision and recall achieved were 93.51% and 85.33%, respectively, demonstrating their suitability for the task.…”
Section: Discussionmentioning
confidence: 99%
“…Subsequently, an LLM (GPT-3.5 Turbo from OpenAI) was employed to transform these datasets into CSV format, as described in Ref. [49]. The average number of tokens in each input file was 98,885.87, so most of the files to be transformed by the LLM had to be split into several files that met the maximum input tokens of the model (16,385 tokens in the case of GPT-3.5 Turbo).…”
Section: Datasetmentioning
confidence: 99%
“…Still, it is too early to make assumptions about LLM performance in the healthcare sphere. The well-known problem of LLM hallucination [39] limits not only the reliability and trustworthiness of the generated results but also creates a relatively new problem of responsibility for LLM-generated mistakes, wrong advice and externally coherent but medically meaningless solutions. GPT-4 demonstrated high interpretative abilities but showed reasoning inconsistencies, while Falcon and LLaMA 2 reached significant accuracy yet with insufficient explanatory reasoning [40].…”
Section: Limitationsmentioning
confidence: 99%