Proceedings of the 16th ACM International Conference on Computing Frontiers 2019
DOI: 10.1145/3310273.3323047
|View full text |Cite
|
Sign up to set email alerts
|

nGraph-HE

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
8
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 111 publications
(16 citation statements)
references
References 15 publications
0
8
0
Order By: Relevance
“…It is limited to the interaction between the processor and memory leaving aside accelerator devices such as the GPU. In order to solve or significantly reduce this problem, methods have emerged such as the optimization of the directed asynchronous graph created at the time of the execution of the model, as proposed by Le [15] and Boemer [3] or, methods of working with sparses matrices [20] to reduce memory consumption during training processes. They have also considered changing the way, memory levels are used during training as mentioned in Rhu [21] and Lim [17] works but they also present bottlenecks due to the high level of communication that occurs through the PCIe bus.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…It is limited to the interaction between the processor and memory leaving aside accelerator devices such as the GPU. In order to solve or significantly reduce this problem, methods have emerged such as the optimization of the directed asynchronous graph created at the time of the execution of the model, as proposed by Le [15] and Boemer [3] or, methods of working with sparses matrices [20] to reduce memory consumption during training processes. They have also considered changing the way, memory levels are used during training as mentioned in Rhu [21] and Lim [17] works but they also present bottlenecks due to the high level of communication that occurs through the PCIe bus.…”
Section: Discussionmentioning
confidence: 99%
“…All of the above have forced accelerator designers for deep neural networks to use highcost memory solutions such as HBM (High Bandwidth Memory) used in Google TPUs [11]. Other solutions have been proposed to overcome these limitations such as the development of new techniques to improve training by working directly on the neural network graph [3,15] or working with sparse matrices [20]. Designing specialized dense nodes for the effective use of accelerators has also been done.…”
Section: Related Workmentioning
confidence: 99%
“…The nGraph-HE framework, proposed by Boemer et al [75] in 2019, is based on Intel's nGraph ML compiler [76] and translates standard TensorFlow computations into arithmetic circuits in BFV or CKKS using the SEAL library. It enables inference on pre-trained models over encrypted inputs, applying FHE-specific optimizations (e.g., constant folding, SIMD-packing, and graph-level optimizations such as lazy rescaling and depth-aware encoding), and run-time optimizations (e.g., bypassing special plaintext values).…”
Section: G Ngraph-hementioning
confidence: 99%
“…This technique can directly adapt computational operations to encrypted data. A large number of methods have been proposed that automatically classify data encrypted with secure computation [ 4 , 5 , 6 ]. These methods can protect test data; however, the encrypted data can hardly be compressed.…”
Section: Introductionmentioning
confidence: 99%