2011 International Conference on High Performance Computing &Amp; Simulation 2011
DOI: 10.1109/hpcsim.2011.5999896
|View full text |Cite
|
Sign up to set email alerts
|

Optimisation and parallelisation of the partitioning around medoids function in R

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2013
2013
2018
2018

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(4 citation statements)
references
References 11 publications
0
4
0
Order By: Relevance
“…We clustered this distance matrix using the Partition Around Medoids (PAM) clustering algorithm (Park and Jun, 2009). We implemented clustering in R using the ConsensusClusterPlus package (Wilkerson and Hayes, 2010) from Bioconductor with the ppam function from Sprint package to perform parallel PAM (Piotrowski et al, 2011). We set the number of clusters to match the individual ADAGE model (e.g.…”
Section: Star Methodsmentioning
confidence: 99%
“…We clustered this distance matrix using the Partition Around Medoids (PAM) clustering algorithm (Park and Jun, 2009). We implemented clustering in R using the ConsensusClusterPlus package (Wilkerson and Hayes, 2010) from Bioconductor with the ppam function from Sprint package to perform parallel PAM (Piotrowski et al, 2011). We set the number of clusters to match the individual ADAGE model (e.g.…”
Section: Star Methodsmentioning
confidence: 99%
“…Clearly Figure 3 shows that the solution adopted with the relatively fast EBS exposed to all nodes by the NFS on the master node is probably not efficient enough and is the main performance bottleneck. This can be seen when comparing these results with previous benchmarks conducted on a local cluster that achieved much better scaling [11]. These were performed on a shared memory cluster with 8 dual-core processors, where each process had simultaneous access to data stored on a disk.…”
Section: Resultsmentioning
confidence: 67%
“…Here all the functions show relatively good speedup as the number of the processes increases. Petrou [35] and Piotrowski et al [11] have previously shown that, with a suitably sized dataset, pcor and ppam exhibit near linear scaling on up to 32 processes on a Cray XT supercomputer consisting of 1416 compute blades each with four quad core processor sockets where the CPUs were AMD 2.3 GHz Opteron chips with 8 GB of memory. These blades were connected via the proprietary CRAY SEASTAR2 interconnect with access to high performance parallel I/O disk storage as opposed to network attached storage.…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation