Mercurial > cgi-bin > hgwebdir.cgi > VMS > VMS_Implementations > SSR_impls > SSR__MC_shared_impl
changeset 67:b5b5323b4177 Holistic_Model
added debugging stuff
| author | Some Random Person <seanhalle@yahoo.com> |
|---|---|
| date | Fri, 16 Mar 2012 23:41:31 -0700 |
| parents | 51b8f15fea29 |
| children | 81a0f076b12e |
| files | SSR.c SSR.h SSR_Counter_Recording.c SSR_Counter_Recording.h SSR_Measurement.h SSR_PluginFns.c SSR_Request_Handlers.c |
| diffstat | 7 files changed, 75 insertions(+), 73 deletions(-) [+] |
line diff
1.1 --- a/SSR.c Wed Mar 14 23:22:00 2012 -0700 1.2 +++ b/SSR.c Fri Mar 16 23:41:31 2012 -0700 1.3 @@ -29,8 +29,8 @@ 1.4 * A: thinking pin the coreCtlrs for all of BLIS -- let Master arbitrate 1.5 * among library, DKU, WT, FoR -- all the patterns in terms of virtual 1.6 * processors (or equivalently work-units), so Master picks which virt procr 1.7 - * from which portions of app (DKU, WT, FoR) onto which sched slots 1.8 - *Might even do hierarchy of masters -- group of sched slots for each core 1.9 + * from which portions of app (DKU, WT, FoR) onto which anim slots 1.10 + *Might even do hierarchy of masters -- group of anim slots for each core 1.11 * has its own master, that keeps generated work local 1.12 * single-reader-single-writer sync everywhere -- no atomic primitives 1.13 * Might have the different assigners talk to each other, to negotiate 1.14 @@ -186,7 +186,7 @@ 1.15 1.16 //Hook up the semantic layer's plug-ins to the Master virt procr 1.17 _VMSMasterEnv->requestHandler = &SSR__Request_Handler; 1.18 - _VMSMasterEnv->slaveAssigner = &SSR__assign_slaveVP; 1.19 + _VMSMasterEnv->slaveAssigner = &SSR__assign_slaveVP_to_slot; 1.20 #ifdef HOLISTIC__TURN_ON_PERF_COUNTERS 1.21 _VMSMasterEnv->counterHandler = &SSR__counter_handler; 1.22 #endif 1.23 @@ -562,7 +562,7 @@ 1.24 1.25 void * 1.26 SSR__receive_type_to( const int type, SlaveVP *receivePr ) 1.27 - { 1.28 + { DEBUG__printf1(dbgRqstHdlr,"WL: receive type to: %d", receivePr->slaveID); 1.29 SSRSemReq reqData; 1.30 1.31 reqData.receivePr = receivePr; 1.32 @@ -585,7 +585,8 @@ 1.33 */ 1.34 void * 1.35 SSR__receive_from_to( SlaveVP *sendPr, SlaveVP *receivePr ) 1.36 - { SSRSemReq reqData; 1.37 + { DEBUG__printf2(dbgRqstHdlr,"WL: receive from %d to: %d", sendPr->slaveID, receivePr->slaveID); 1.38 + SSRSemReq reqData; 1.39 1.40 //hash on the receiver, 'cause always know it, but sometimes want to 1.41 // receive from anonymous sender
2.1 --- a/SSR.h Wed Mar 14 23:22:00 2012 -0700 2.2 +++ b/SSR.h Fri Mar 16 23:41:31 2012 -0700 2.3 @@ -33,7 +33,7 @@ 2.4 //=========================================================================== 2.5 2.6 /*Semantic layer-specific data sent inside a request from lib called in app 2.7 - * to request handler called in MasterLoop 2.8 + * to request handler called in AnimationMaster 2.9 */ 2.10 2.11 typedef struct 2.12 @@ -253,7 +253,7 @@ 2.13 SSR__Request_Handler( SlaveVP *requestingPr, void *_semEnv ); 2.14 2.15 SlaveVP * 2.16 -SSR__assign_slaveVP( void *_semEnv, int coreNum, SchedSlot *slot ); 2.17 +SSR__assign_slaveVP_to_slot( void *_semEnv, AnimSlot *slot ); 2.18 2.19 SlaveVP* 2.20 SSR__create_procr_helper( TopLevelFnPtr fnPtr, void *initData,
3.1 --- a/SSR_Counter_Recording.c Wed Mar 14 23:22:00 2012 -0700 3.2 +++ b/SSR_Counter_Recording.c Fri Mar 16 23:41:31 2012 -0700 3.3 @@ -31,9 +31,10 @@ 3.4 void SSR__counter_handler(int evt_type, int vpid, int task, SlaveVP* pr, uint64 cycles, uint64 instrs) 3.5 { 3.6 3.7 - if (pr->isMasterVP || pr->isShutdownVP) { //Save only values for actual work 3.8 + if (pr->typeOfVP == Master || pr->typeOfVP == Shutdown) 3.9 + { //Only save values for application work, done in a SlaveVP 3.10 return; 3.11 - } 3.12 + } 3.13 3.14 SSRSemEnv *semanticEnv = _VMSMasterEnv->semanticEnv; 3.15 3.16 @@ -47,7 +48,7 @@ 3.17 3.18 if(pr){ 3.19 e.coreID = pr->coreAnimatedBy; 3.20 - e.slot = pr->schedSlot; 3.21 + e.slot = pr->animSlotAssignedTo; 3.22 } else { 3.23 e.coreID = -1; 3.24 e.slot = NULL;
4.1 --- a/SSR_Counter_Recording.h Wed Mar 14 23:22:00 2012 -0700 4.2 +++ b/SSR_Counter_Recording.h Fri Mar 16 23:41:31 2012 -0700 4.3 @@ -13,7 +13,7 @@ 4.4 typedef struct { 4.5 int event_type; 4.6 int coreID; 4.7 - SchedSlot* slot; 4.8 + AnimSlot* slot; 4.9 int vp; 4.10 int task; 4.11 uint64 cycles;
5.1 --- a/SSR_Measurement.h Wed Mar 14 23:22:00 2012 -0700 5.2 +++ b/SSR_Measurement.h Fri Mar 16 23:41:31 2012 -0700 5.3 @@ -71,7 +71,7 @@ 5.4 5.5 #else //===================== turned off ========================== 5.6 5.7 - #define MEAS__Make_Meas_Hists_for_Language 5.8 + #define MEAS__Make_Meas_Hists_for_Language 5.9 #define Meas_startSendFromTo 5.10 #define Meas_endSendFromTo 5.11 #define Meas_startSendOfType
6.1 --- a/SSR_PluginFns.c Wed Mar 14 23:22:00 2012 -0700 6.2 +++ b/SSR_PluginFns.c Fri Mar 16 23:41:31 2012 -0700 6.3 @@ -28,61 +28,62 @@ 6.4 //============================== Assigner ================================== 6.5 // 6.6 /*For SSR, assigning a slave simply takes the next work-unit off the 6.7 - * ready-to-go work-unit queue and assigns it to the slaveToSched. 6.8 + * ready-to-go work-unit queue and assigns it to the offered slot. 6.9 *If the ready-to-go work-unit queue is empty, then nothing to assign 6.10 - * to the slave -- return FALSE to let Master loop know scheduling that 6.11 - * slave failed. 6.12 + * to the animation slot -- return FALSE to let Master loop know assigning 6.13 + * that slot failed. 6.14 */ 6.15 SlaveVP * 6.16 -SSR__assign_slaveVP( void *_semEnv, int coreNum, SchedSlot *slot ) 6.17 - { SlaveVP *schedPr; 6.18 +SSR__assign_slaveVP_to_slot( void *_semEnv, AnimSlot *slot ) 6.19 + { SlaveVP *assignPr; 6.20 SSRSemEnv *semEnv; 6.21 - int32 slotNum; 6.22 + int32 coreNum, slotNum; 6.23 6.24 + coreNum = slot->coreOfSlot; 6.25 slotNum = slot->slotIdx; 6.26 6.27 semEnv = (SSRSemEnv *)_semEnv; 6.28 6.29 - schedPr = readPrivQ( semEnv->readyVPQs[coreNum] ); 6.30 + assignPr = readPrivQ( semEnv->readyVPQs[coreNum] ); 6.31 //Note, using a non-blocking queue -- it returns NULL if queue empty 6.32 - if(!schedPr){ 6.33 - schedPr = semEnv->idlePr[coreNum][slotNum]; 6.34 + if(!assignPr){ 6.35 + assignPr = semEnv->idlePr[coreNum][slotNum]; 6.36 //things that would normally happen in resume(), but these VPs never go there 6.37 #ifdef HOLISTIC__TURN_ON_OBSERVE_UCC 6.38 - schedPr->numTimesAssigned++; //Somewhere here! 6.39 + assignPr->assignCount++; //Somewhere here! 6.40 Unit newu; 6.41 - newu.vp = schedPr->slaveID; 6.42 - newu.task = schedPr->numTimesAssigned; 6.43 + newu.vp = assignPr->slaveID; 6.44 + newu.task = assignPr->assignCount; 6.45 addToListOfArrays(Unit,newu,semEnv->unitList); 6.46 6.47 - if (schedPr->numTimesAssigned > 1){ 6.48 + if (assignPr->assignCount > 1){ 6.49 Dependency newd; 6.50 - newd.from_vp = schedPr->slaveID; 6.51 - newd.from_task = schedPr->numTimesAssigned - 1; 6.52 - newd.to_vp = schedPr->slaveID; 6.53 - newd.to_task = schedPr->numTimesAssigned; 6.54 + newd.from_vp = assignPr->slaveID; 6.55 + newd.from_task = assignPr->assignCount - 1; 6.56 + newd.to_vp = assignPr->slaveID; 6.57 + newd.to_task = assignPr->assignCount; 6.58 addToListOfArrays(Dependency, newd ,semEnv->ctlDependenciesList); 6.59 } 6.60 #endif 6.61 } 6.62 #ifdef HOLISTIC__TURN_ON_OBSERVE_UCC 6.63 - if (schedPr) { 6.64 - //schedPr->numTimesAssigned++; 6.65 + if (assignPr) { 6.66 + //assignPr->numTimesAssigned++; 6.67 Unit prev_in_slot = semEnv->last_in_slot[coreNum * NUM_SCHED_SLOTS + slotNum]; 6.68 if(prev_in_slot.vp != 0){ 6.69 Dependency newd; 6.70 newd.from_vp = prev_in_slot.vp; 6.71 newd.from_task = prev_in_slot.task; 6.72 - newd.to_vp = schedPr->slaveID; 6.73 - newd.to_task = schedPr->numTimesAssigned; 6.74 + newd.to_vp = assignPr->slaveID; 6.75 + newd.to_task = assignPr->assignCount; 6.76 addToListOfArrays(Dependency,newd,semEnv->hwArcs); 6.77 } 6.78 - prev_in_slot.vp = schedPr->slaveID; 6.79 - prev_in_slot.task = schedPr->numTimesAssigned; 6.80 + prev_in_slot.vp = assignPr->slaveID; 6.81 + prev_in_slot.task = assignPr->assignCount; 6.82 semEnv->last_in_slot[coreNum * NUM_SCHED_SLOTS + slotNum] = prev_in_slot; 6.83 } 6.84 #endif 6.85 - return( schedPr ); 6.86 + return( assignPr ); 6.87 } 6.88 6.89 6.90 @@ -177,7 +178,7 @@ 6.91 void 6.92 handleDissipate( SlaveVP *requestingPr, SSRSemEnv *semEnv ) 6.93 { 6.94 - DEBUG_Print1(dbgRqstHdlr,"Dissipate request from processor %d\n",requestingPr->slaveID) 6.95 + DEBUG__printf1(dbgRqstHdlr,"Dissipate request from processor %d",requestingPr->slaveID) 6.96 //free any semantic data allocated to the virt procr 6.97 VMS_PI__free( requestingPr->semanticData ); 6.98 6.99 @@ -239,19 +240,18 @@ 6.100 { SSRSemReq *semReq; 6.101 SlaveVP *newPr; 6.102 6.103 - DEBUG_Print1(dbgRqstHdlr,"Create request from processor %d ",requestingPr->slaveID) 6.104 6.105 semReq = VMS_PI__take_sem_reqst_from( req ); 6.106 6.107 newPr = SSR__create_procr_helper( semReq->fnPtr, semReq->initData, semEnv, 6.108 semReq->coreToAssignOnto ); 6.109 6.110 - DEBUG_Print1(dbgRqstHdlr,"(new VP: %d)\n",newPr->slaveID) 6.111 + DEBUG__printf2(dbgRqstHdlr,"Create from: %d, new VP: %d", requestingPr->slaveID, newPr->slaveID) 6.112 6.113 #ifdef HOLISTIC__TURN_ON_OBSERVE_UCC 6.114 Dependency newd; 6.115 newd.from_vp = requestingPr->slaveID; 6.116 - newd.from_task = requestingPr->numTimesAssigned; 6.117 + newd.from_task = requestingPr->assignCount; 6.118 newd.to_vp = newPr->slaveID; 6.119 newd.to_task = 1; 6.120 //addToListOfArraysDependency(newd,semEnv->commDependenciesList); 6.121 @@ -278,18 +278,18 @@ 6.122 */ 6.123 #endif 6.124 #ifdef HOLISTIC__TURN_ON_OBSERVE_UCC 6.125 - procr->numTimesAssigned++; //Somewhere here! 6.126 + procr->assignCount++; //Somewhere here! 6.127 Unit newu; 6.128 newu.vp = procr->slaveID; 6.129 - newu.task = procr->numTimesAssigned; 6.130 + newu.task = procr->assignCount; 6.131 addToListOfArrays(Unit,newu,semEnv->unitList); 6.132 6.133 - if (procr->numTimesAssigned > 1){ 6.134 + if (procr->assignCount > 1){ 6.135 Dependency newd; 6.136 newd.from_vp = procr->slaveID; 6.137 - newd.from_task = procr->numTimesAssigned - 1; 6.138 + newd.from_task = procr->assignCount - 1; 6.139 newd.to_vp = procr->slaveID; 6.140 - newd.to_task = procr->numTimesAssigned; 6.141 + newd.to_task = procr->assignCount; 6.142 addToListOfArrays(Dependency, newd ,semEnv->ctlDependenciesList); 6.143 } 6.144 #endif
7.1 --- a/SSR_Request_Handlers.c Wed Mar 14 23:22:00 2012 -0700 7.2 +++ b/SSR_Request_Handlers.c Fri Mar 16 23:41:31 2012 -0700 7.3 @@ -89,7 +89,7 @@ 7.4 HashEntry *entry; 7.5 HashTable *commHashTbl = semEnv->commHashTbl; 7.6 7.7 - DEBUG_Print1(dbgRqstHdlr,"SendType request from processor %d\n",semReq->sendPr->slaveID) 7.8 + DEBUG__printf1(dbgRqstHdlr,"SendType request from processor %d",semReq->sendPr->slaveID) 7.9 7.10 receivePr = semReq->receivePr; //For "send", know both send & recv procrs 7.11 sendPr = semReq->sendPr; 7.12 @@ -120,7 +120,7 @@ 7.13 SSRSemReq *clonedReq = cloneReq( semReq ); 7.14 clonedReq-> nextReqInHashEntry = waitingReq->nextReqInHashEntry; 7.15 waitingReq->nextReqInHashEntry = clonedReq; 7.16 - DEBUG_Print2( dbgRqstHdlr, "linked requests: %p, %p | ", clonedReq,\ 7.17 + DEBUG__printf2( dbgRqstHdlr, "linked requests: %p, %p ", clonedReq,\ 7.18 waitingReq ) 7.19 return; 7.20 } 7.21 @@ -129,9 +129,9 @@ 7.22 #ifdef HOLISTIC__TURN_ON_OBSERVE_UCC 7.23 Dependency newd; 7.24 newd.from_vp = sendPr->slaveID; 7.25 - newd.from_task = sendPr->numTimesAssigned; 7.26 + newd.from_task = sendPr->assignCount; 7.27 newd.to_vp = receivePr->slaveID; 7.28 - newd.to_task = receivePr->numTimesAssigned +1; 7.29 + newd.to_task = receivePr->assignCount +1; 7.30 //(newd,semEnv->commDependenciesList); 7.31 addToListOfArrays(Dependency,newd,semEnv->dynDependenciesList); 7.32 int32 groupId = semReq->msgType; 7.33 @@ -143,10 +143,10 @@ 7.34 } 7.35 Unit u; 7.36 u.vp = sendPr->slaveID; 7.37 - u.task = sendPr->numTimesAssigned; 7.38 + u.task = sendPr->assignCount; 7.39 addToListOfArrays(Unit,u,semEnv->ntonGroups[groupId]->senders); 7.40 u.vp = receivePr->slaveID; 7.41 - u.task = receivePr->numTimesAssigned +1; 7.42 + u.task = receivePr->assignCount +1; 7.43 addToListOfArrays(Unit,u,semEnv->ntonGroups[groupId]->receivers); 7.44 #endif 7.45 7.46 @@ -185,7 +185,7 @@ 7.47 HashEntry *entry; 7.48 HashTable *commHashTbl = semEnv->commHashTbl; 7.49 7.50 - DEBUG_Print2(dbgRqstHdlr,"SendFromTo request from processor %d to %d\n",semReq->sendPr->slaveID,semReq->receivePr->slaveID) 7.51 + DEBUG__printf2(dbgRqstHdlr,"SendFromTo request from processor %d to %d",semReq->sendPr->slaveID,semReq->receivePr->slaveID) 7.52 7.53 receivePr = semReq->receivePr; //For "send", know both send & recv procrs 7.54 sendPr = semReq->sendPr; 7.55 @@ -209,9 +209,9 @@ 7.56 #ifdef HOLISTIC__TURN_ON_OBSERVE_UCC 7.57 Dependency newd; 7.58 newd.from_vp = sendPr->slaveID; 7.59 - newd.from_task = sendPr->numTimesAssigned; 7.60 + newd.from_task = sendPr->assignCount; 7.61 newd.to_vp = receivePr->slaveID; 7.62 - newd.to_task = receivePr->numTimesAssigned +1; 7.63 + newd.to_task = receivePr->assignCount +1; 7.64 //addToListOfArraysDependency(newd,semEnv->commDependenciesList); 7.65 addToListOfArrays(Dependency,newd,semEnv->commDependenciesList); 7.66 #endif 7.67 @@ -285,7 +285,7 @@ 7.68 7.69 receivePr = semReq->receivePr; 7.70 7.71 - DEBUG_Print1(dbgRqstHdlr,"ReceiveType request from processor %d\n",receivePr->slaveID) 7.72 + DEBUG__printf1(dbgRqstHdlr,"ReceiveType request from processor %d",receivePr->slaveID) 7.73 7.74 key[0] = (int)receivePr->slaveID; 7.75 key[1] = (int)(semReq->msgType); 7.76 @@ -318,9 +318,9 @@ 7.77 #ifdef HOLISTIC__TURN_ON_OBSERVE_UCC 7.78 Dependency newd; 7.79 newd.from_vp = sendPr->slaveID; 7.80 - newd.from_task = sendPr->numTimesAssigned; 7.81 + newd.from_task = sendPr->assignCount; 7.82 newd.to_vp = receivePr->slaveID; 7.83 - newd.to_task = receivePr->numTimesAssigned +1; 7.84 + newd.to_task = receivePr->assignCount +1; 7.85 //addToListOfArraysDependency(newd,semEnv->commDependenciesList); 7.86 addToListOfArrays(Dependency,newd,semEnv->dynDependenciesList); 7.87 int32 groupId = semReq->msgType; 7.88 @@ -332,10 +332,10 @@ 7.89 } 7.90 Unit u; 7.91 u.vp = sendPr->slaveID; 7.92 - u.task = sendPr->numTimesAssigned; 7.93 + u.task = sendPr->assignCount; 7.94 addToListOfArrays(Unit,u,semEnv->ntonGroups[groupId]->senders); 7.95 u.vp = receivePr->slaveID; 7.96 - u.task = receivePr->numTimesAssigned +1; 7.97 + u.task = receivePr->assignCount +1; 7.98 addToListOfArrays(Unit,u,semEnv->ntonGroups[groupId]->receivers); 7.99 #endif 7.100 7.101 @@ -352,13 +352,13 @@ 7.102 */ 7.103 void 7.104 handleReceiveFromTo( SSRSemReq *semReq, SSRSemEnv *semEnv) 7.105 - { SlaveVP *sendPr, *receivePr; 7.106 + { SlaveVP *sendPr, *receivePr; 7.107 int key[] = {0,0,0}; 7.108 - SSRSemReq *waitingReq; 7.109 + SSRSemReq *waitingReq; 7.110 HashEntry *entry; 7.111 HashTable *commHashTbl = semEnv->commHashTbl; 7.112 7.113 - DEBUG_Print2(dbgRqstHdlr,"ReceiveFromTo request from processor %d to %d\n",semReq->sendPr->slaveID,semReq->receivePr->slaveID) 7.114 + DEBUG__printf2(dbgRqstHdlr,"ReceiveFromTo %d : %d",semReq->sendPr->slaveID,semReq->receivePr->slaveID) 7.115 7.116 receivePr = semReq->receivePr; 7.117 sendPr = semReq->sendPr; //for receive from-to, know send procr 7.118 @@ -378,9 +378,9 @@ 7.119 #ifdef HOLISTIC__TURN_ON_OBSERVE_UCC 7.120 Dependency newd; 7.121 newd.from_vp = sendPr->slaveID; 7.122 - newd.from_task = sendPr->numTimesAssigned; 7.123 + newd.from_task = sendPr->assignCount; 7.124 newd.to_vp = receivePr->slaveID; 7.125 - newd.to_task = receivePr->numTimesAssigned +1; 7.126 + newd.to_task = receivePr->assignCount +1; 7.127 //addToListOfArraysDependency(newd,semEnv->commDependenciesList); 7.128 addToListOfArrays(Dependency,newd,semEnv->commDependenciesList); 7.129 #endif 7.130 @@ -427,7 +427,7 @@ 7.131 handleMalloc( SSRSemReq *semReq, SlaveVP *requestingPr, SSRSemEnv *semEnv ) 7.132 { void *ptr; 7.133 7.134 - DEBUG_Print1(dbgRqstHdlr,"Malloc request from processor %d\n",requestingPr->slaveID) 7.135 + DEBUG__printf1(dbgRqstHdlr,"Malloc request from processor %d",requestingPr->slaveID) 7.136 7.137 ptr = VMS_PI__malloc( semReq->sizeToMalloc ); 7.138 requestingPr->dataRetFromReq = ptr; 7.139 @@ -439,7 +439,7 @@ 7.140 void 7.141 handleFree( SSRSemReq *semReq, SlaveVP *requestingPr, SSRSemEnv *semEnv ) 7.142 { 7.143 - DEBUG_Print1(dbgRqstHdlr,"Free request from processor %d\n",requestingPr->slaveID) 7.144 + DEBUG__printf1(dbgRqstHdlr,"Free request from processor %d",requestingPr->slaveID) 7.145 VMS_PI__free( semReq->ptrToFree ); 7.146 resume_slaveVP( requestingPr, semEnv ); 7.147 } 7.148 @@ -477,7 +477,7 @@ 7.149 handleStartFnSingleton( SSRSemReq *semReq, SlaveVP *requestingPr, 7.150 SSRSemEnv *semEnv ) 7.151 { SSRSingleton *singleton; 7.152 - DEBUG_Print1(dbgRqstHdlr,"StartFnSingleton request from processor %d\n",requestingPr->slaveID) 7.153 + DEBUG__printf1(dbgRqstHdlr,"StartFnSingleton request from processor %d",requestingPr->slaveID) 7.154 7.155 singleton = &(semEnv->fnSingletons[ semReq->singletonID ]); 7.156 handleStartSingleton_helper( singleton, requestingPr, semEnv ); 7.157 @@ -487,7 +487,7 @@ 7.158 SSRSemEnv *semEnv ) 7.159 { SSRSingleton *singleton; 7.160 7.161 - DEBUG_Print1(dbgRqstHdlr,"StartDataSingleton request from processor %d\n",requestingPr->slaveID) 7.162 + DEBUG__printf1(dbgRqstHdlr,"StartDataSingleton request from processor %d",requestingPr->slaveID) 7.163 if( *(semReq->singletonPtrAddr) == NULL ) 7.164 { singleton = VMS_PI__malloc( sizeof(SSRSingleton) ); 7.165 singleton->waitQ = makeVMSQ(); 7.166 @@ -534,7 +534,7 @@ 7.167 { 7.168 SSRSingleton *singleton; 7.169 7.170 - DEBUG_Print1(dbgRqstHdlr,"EndFnSingleton request from processor %d\n",requestingPr->slaveID) 7.171 + DEBUG__printf1(dbgRqstHdlr,"EndFnSingleton request from processor %d",requestingPr->slaveID) 7.172 7.173 singleton = &(semEnv->fnSingletons[ semReq->singletonID ]); 7.174 handleEndSingleton_helper( singleton, requestingPr, semEnv ); 7.175 @@ -545,7 +545,7 @@ 7.176 { 7.177 SSRSingleton *singleton; 7.178 7.179 - DEBUG_Print1(dbgRqstHdlr,"EndDataSingleton request from processor %d\n",requestingPr->slaveID) 7.180 + DEBUG__printf1(dbgRqstHdlr,"EndDataSingleton request from processor %d",requestingPr->slaveID) 7.181 7.182 singleton = *(semReq->singletonPtrAddr); 7.183 handleEndSingleton_helper( singleton, requestingPr, semEnv ); 7.184 @@ -558,7 +558,7 @@ 7.185 void 7.186 handleAtomic( SSRSemReq *semReq, SlaveVP *requestingPr, SSRSemEnv *semEnv ) 7.187 { 7.188 - DEBUG_Print1(dbgRqstHdlr,"Atomic request from processor %d\n",requestingPr->slaveID) 7.189 + DEBUG__printf1(dbgRqstHdlr,"Atomic request from processor %d",requestingPr->slaveID) 7.190 semReq->fnToExecInMaster( semReq->dataForFn ); 7.191 resume_slaveVP( requestingPr, semEnv ); 7.192 } 7.193 @@ -583,7 +583,7 @@ 7.194 { SSRSemData *semData; 7.195 TransListElem *nextTransElem; 7.196 7.197 - DEBUG_Print1(dbgRqstHdlr,"TransStart request from processor %d\n",requestingPr->slaveID) 7.198 + DEBUG__printf1(dbgRqstHdlr,"TransStart request from processor %d",requestingPr->slaveID) 7.199 7.200 //check ordering of entering transactions is correct 7.201 semData = requestingPr->semanticData; 7.202 @@ -637,7 +637,7 @@ 7.203 SSRTrans *transStruc; 7.204 TransListElem *lastTrans; 7.205 7.206 - DEBUG_Print1(dbgRqstHdlr,"TransEnd request from processor %d\n",requestingPr->slaveID) 7.207 + DEBUG__printf1(dbgRqstHdlr,"TransEnd request from processor %d",requestingPr->slaveID) 7.208 7.209 transStruc = &(semEnv->transactionStrucs[ semReq->transID ]); 7.210
