?? jgfreducebench.java
字號:
/*************************************************************************** ** Java Grande Forum Benchmark Suite - MPJ Version 1.0 ** ** produced by ** ** Java Grande Benchmarking Project ** ** at ** ** Edinburgh Parallel Computing Centre ** * * email: epcc-javagrande@epcc.ed.ac.uk ** ** ** This version copyright (c) The University of Edinburgh, 2001. ** All rights reserved. ** ***************************************************************************/package jgf_mpj_benchmarks.section1;import java.io.*;//import jgfutil.*; import jgf_mpj_benchmarks.jgfutil.*; import mpi.*;public class JGFReduceBench implements JGFSection1{ public static int nprocess; public static int rank; private static final int INITSIZE = 1; private static final int MAXSIZE = 1000000; private static final double TARGETTIME = 10.0; private static final int MLOOPSIZE = 2; private static final int SMAX = 5000000; private static final int SMIN = 4; public JGFReduceBench(int rank, int nprocess) { this.rank = rank; this.nprocess=nprocess; } public void JGFrun() throws MPIException { int size,i,l,m_size; double logsize; double b_time; b_time = 0.0; double [] time = new double[1]; m_size = 0; logsize = Math.log((double) SMAX) - Math.log((double) SMIN);/* Reduce an array of doubles *//* Create the timers */ if(rank==0){ JGFInstrumentor.addTimer("Section1:Reduce:Double", "bytes"); JGFInstrumentor.addTimer("Section1:Reduce:Barrier", "barriers"); }/* loop over no of different message sizes */ for(l=0;l<MLOOPSIZE;l++){/* Initialize the sending data */ m_size = (int)(Math.exp(Math.log((double)SMIN)+(double) ((double) l/(double) MLOOPSIZE*logsize))); double [] send_arr = new double[m_size]; double [] recv_arr = new double[m_size]; time[0] = 0.0; size=INITSIZE; MPI.COMM_WORLD.Barrier();/* Start the timer */ while (time[0] < TARGETTIME && size < MAXSIZE){ if(rank==0){ JGFInstrumentor.resetTimer("Section1:Reduce:Double"); JGFInstrumentor.startTimer("Section1:Reduce:Double"); }/* Carryout the broadcast operation */ for (int k=0; k<size; k++){ MPI.COMM_WORLD.Reduce(send_arr,0,recv_arr,0,send_arr.length,MPI.DOUBLE,MPI.SUM,0); MPI.COMM_WORLD.Barrier(); }/* Stop the timer */ if(rank==0){ JGFInstrumentor.stopTimer("Section1:Reduce:Double"); time[0] = JGFInstrumentor.readTimer("Section1:Reduce:Double"); JGFInstrumentor.addOpsToTimer("Section1:Reduce:Double",(double) size*send_arr.length*8); }/* Broadcast time to the other processes */ MPI.COMM_WORLD.Barrier(); MPI.COMM_WORLD.Bcast(time,0,1,MPI.DOUBLE,0); size *=2; } size /=2;/* determine the cost of the Barrier, subtract the cost and write out the performance time */ MPI.COMM_WORLD.Barrier(); if(rank==0) { JGFInstrumentor.resetTimer("Section1:Reduce:Barrier"); JGFInstrumentor.startTimer("Section1:Reduce:Barrier"); } for (int k=0; k<size; k++){ MPI.COMM_WORLD.Barrier(); } if(rank==0) { JGFInstrumentor.stopTimer("Section1:Reduce:Barrier"); b_time = JGFInstrumentor.readTimer("Section1:Reduce:Barrier"); JGFInstrumentor.addTimeToTimer("Section1:Reduce:Double", -b_time); JGFInstrumentor.printperfTimer("Section1:Reduce:Double",send_arr.length); } } } public static void main(String[] argv) throws MPIException{ } public JGFReduceBench(String[] argv) throws Exception {/* Initialise MPI */ MPI.Init(argv); rank = MPI.COMM_WORLD.Rank(); nprocess = MPI.COMM_WORLD.Size(); if(rank==0){ JGFInstrumentor.printHeader(1,0,nprocess); } JGFReduceBench rd = new JGFReduceBench(rank, nprocess); rd.JGFrun();/* Finalise MPI */ MPI.Finalize(); }}
?? 快捷鍵說明
復制代碼
Ctrl + C
搜索代碼
Ctrl + F
全屏模式
F11
切換主題
Ctrl + Shift + D
顯示快捷鍵
?
增大字號
Ctrl + =
減小字號
Ctrl + -