Proceedings of the SC '23 Workshops of the International Conference on High Performance Computing, Network, Storage, and Analys 2023
DOI: 10.1145/3624062.3624172
|View full text |Cite
|
Sign up to set email alerts
|

HPC-GPT: Integrating Large Language Model for High-Performance Computing

Xianzhong Ding,
Le Chen,
Murali Emani
et al.

Abstract: Large Language Models (LLMs), including the LLaMA model, have exhibited their efficacy across various general-domain natural language processing (NLP) tasks. However, their performance in highperformance computing (HPC) domain tasks has been less than optimal due to the specialized expertise required to interpret the model's responses. In response to this challenge, we propose HPC-GPT, a novel LLaMA-based model that has been supervised finetuning using generated QA (Question-Answer) instances for the HPC domai… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
2
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 10 publications
(3 citation statements)
references
References 30 publications
0
3
0
Order By: Relevance
“…The implementation of hardware accelerators was shown to drastically reduce training times and improve throughput, enabling more rapid iteration and development cycles [56,57]. Custom accelerators, designed specifically for machine learning workloads, were evaluated for their ability to handle the demands of large language models [11,58,59]. The advantages of leveraging hardware acceleration include enhancing real-time inference capabilities, making large models more practical for deployment in latency-sensitive applications [60,61].…”
Section: Hardware Accelerationmentioning
confidence: 99%
“…The implementation of hardware accelerators was shown to drastically reduce training times and improve throughput, enabling more rapid iteration and development cycles [56,57]. Custom accelerators, designed specifically for machine learning workloads, were evaluated for their ability to handle the demands of large language models [11,58,59]. The advantages of leveraging hardware acceleration include enhancing real-time inference capabilities, making large models more practical for deployment in latency-sensitive applications [60,61].…”
Section: Hardware Accelerationmentioning
confidence: 99%
“…Research has also looked into the integration of multimodal data processing capabilities, allowing LLMs to handle a variety of input types beyond text [45,46]. Lastly, there is a growing interest in developing lightweight models that maintain high performance while being less resource-intensive [47,48].…”
Section: Advancements In Llm Architectures For Information Retrievalmentioning
confidence: 99%
“…Both approaches have been instrumental in advancing the efficiency of LLM inference [18,19]. Further research has explored the use of mixed precision training and inference, where different parts of the model operate at varying levels of precision [20][21][22]. This method has been found to balance the trade-offs between speed and accuracy, offering a practical solution for accelerating LLM inference [23].…”
Section: Introductionmentioning
confidence: 99%