forked from KxSystems/nano
-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy pathmthread.sh
executable file
·328 lines (302 loc) · 8.24 KB
/
mthread.sh
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
#!/bin/bash
VERS=v1.18
# usage mthread.sh numthreads keep|delete [compress]
HERE=`pwd`
echo please set PATH and QHOME at top of this script, to include q dirs, then comment out these two lines and re-run mthread.sh
exit
export QHOME=$HOME/q
export PATH=$PATH:$QHOME/l64
export QBIN="$QHOME/l64/q"
#
DATE=`date +%m%d:%H%M`
HOST=`uname -n`
PARLIST=${HERE}/partitions
MYID=`id -u`
declare -a array
if [ $# -lt 2 ]
then
echo "Usage: mthread #numberthreads keep|delete [compress]"
exit
fi
if [ $1 -le 0 ]
then
echo "Usage: mthread #numberthreads keep|delete [compress]"
exit
fi
if [ $1 -gt 128 ]
then
echo "This test does not qualify results above 64 processes"
exit
fi
if [ -f ./flush.sh ]
then
:
else
echo "flush.sh is missing"
echo "please copy and edit one of the supplied flush-* prototype scripts into the file flush.sh, edit it for your configuration, and try again"
exit
fi
echo "flushing buffer cache before file creations"
if [ $MYID -eq 0 ]
then
echo umount and flush.sh
./flush.sh
else
echo umount and flush via sudo ./flush.sh
sudo ./flush.sh
fi
touch ${HERE}/sync-$HOST
NUMSEGS=`wc -l $PARLIST | awk '{print $1}'`
array=(`cat $PARLIST`)
TARGETROOT=`dirname ${array[0]}`
NUMTHREADS=$1
j=0
CW=$(expr 800 / $NUMTHREADS)
mkdir -p ${HERE}/${DATE}
# WRITE TEST .......................
cd $HERE
echo Starting write...
echo "version $VERS" >> ${HERE}/${DATE}/aggregates-${HOST}
echo "numprocs $NUMTHREADS" >> ${HERE}/${DATE}/aggregates-${HOST}
echo "disk partitions $NUMSEGS" >> ${HERE}/${DATE}/aggregates-${HOST}
if [ "$3" = "compress" ]
then
echo "Compressed data" >> ${HERE}/${DATE}/aggregates-${HOST}
fi
START=$(date +%s%3N)
# important that this it outside this loop with "q prepare", as first time after a mount as the
# fs may take a long time to start (S3 sync) and we want the wrtte processes to run in parallel
for i in `seq $NUMTHREADS`
do
mkdir -p ${array[$j]}/${HOST}.${i}/${DATE}
j=`expr $j + 1`
if [ $j -ge $NUMSEGS ]
then
j=0
fi
done
j=0
for i in `seq $NUMTHREADS`
do
cd ${array[$j]}/${HOST}.${i}/${DATE}
if [ "$3" = "compress" ]
then
${QBIN} ${HERE}/io.q -prepare -compress -threads $NUMTHREADS | tee ${HERE}/${DATE}/RES-${HOST}-${NUMTHREADS}t-${i} &
else
${QBIN} ${HERE}/io.q -prepare -threads $NUMTHREADS | tee ${HERE}/${DATE}/RES-${HOST}-${NUMTHREADS}t-${i} &
fi
cd -
j=`expr $j + 1`
if [ $j -ge $NUMSEGS ]
then
j=0
fi
done
wait
echo "Files created, flushing buffer cache....."
if [ $MYID -eq 0 ]
then
./flush.sh
else
sudo ./flush.sh
fi
sleep 1
rm ${HERE}/sync-$HOST
#
# sync up across multiple host testing...
#
while [ `ls -l ${HERE}/sync-* 2> /dev/null | wc -l` -ne 0 ]
do
sleep 0.5
done
# air gap for any remote stats collection....
cd ${HERE}
THRU=$(grep 'sync write' ${HERE}/${DATE}/RES-${HOST}-${NUMTHREADS}t-* | awk '{print $5}' | awk '{printf "%.0f\n",$1}' | sort -n | head -1)
THRU=$(echo $THRU | awk '{printf "%.0f",$1}')
THRU=$(expr $THRU \* $NUMTHREADS)
echo "Total Write Rate(sync): " $THRU " MiB/sec" 2>&1 | tee -a ${HERE}/${DATE}/aggregates-${HOST}
THRU=$(grep 'create list' ${HERE}/${DATE}/RES-${HOST}-${NUMTHREADS}t-* | awk '{print $4}' | awk '{printf "%.0f\n",$1}' | sort -n | head -1)
THRU=$(echo $THRU | awk '{printf "%.0f",$1}')
THRU=$(expr $THRU \* $NUMTHREADS)
echo "Total create list rate: " $THRU " MiB/sec" 2>&1 | tee -a ${HERE}/${DATE}/aggregates-${HOST}
sleep 5
#
# READ test................................................
#
echo Starting read tests...
#
# simple semaphore for completion checking for all hosts ...
#
touch ${HERE}/sync2-$HOST
j=0
cd $HERE
for i in `seq $NUMTHREADS`
do
cd ${array[$j]}/${HOST}.${i}/${DATE}
${QBIN} ${HERE}/io.q -read -threads $NUMTHREADS >> ${HERE}/${DATE}/RES-${HOST}-${NUMTHREADS}t-${i} 2>&1 &
j=`expr $j + 1`
if [ $j -ge $NUMSEGS ]
then
j=0
fi
done
wait
#
j=0
ELAPSED=$(grep 'End thread -23! mapped read' ${HERE}/${DATE}/RES-${HOST}-${NUMTHREADS}t-* | awk '{print $6}'| awk '{printf "%.3f\n",$1/1000}' | sort -nr | head -1)
WALKIES=$(grep 'End thread walklist' ${HERE}/${DATE}/RES-${HOST}-${NUMTHREADS}t-* | awk '{print $4}'| awk '{printf "%.3f\n",$1/1000}' | sort -nr | head -1)
# Use filesize which is direct from real filesize from q, e.g compressed data
SIZE=$(grep '^filesize' ${HERE}/${DATE}/RES-${HOST}-${NUMTHREADS}t-1 | awk '{print $2}')
SIZE=$(echo $SIZE | awk '{printf "%.0f",$1}')
# this is where we catch the process aggregation...
SIZE=$(expr $SIZE \* $NUMTHREADS )
echo $SIZE " " $ELAPSED " " $WALKIES | tee -a ${HERE}/${DATE}/aggregates-${HOST}
echo $SIZE $ELAPSED | awk '{$1=sprintf("%5.2f",$1/$2);print "Streaming Read(mapped) Rate: ", $1," MiB/sec"}' | tee -a ${HERE}/${DATE}/aggregates-${HOST}
echo $SIZE $WALKIES | awk '{$1=sprintf("%5.2f",$1/$2);print "Walking List Rate: ", $1," MiB/sec"}' | tee -a ${HERE}/${DATE}/aggregates-${HOST}
rm -f ${HERE}/sync-$HOST
#
# sync up across multiple host testing...
#
while [ `ls -l ${HERE}/sync-* 2> /dev/null | wc -l` -ne 0 ]
do
sleep 0.5
done
# air gap for any remote stats collection....
sleep 5
#
# REREAD test for fresh kdb+ session, without flush, cached in kernel buffer, re-mapped...
#
echo "Starting Re-Read (Cache) tests..."
#
# simple semaphore for completion checking for all hosts ...
#
touch ${HERE}/sync2-$HOST
j=0
cd $HERE
for i in `seq $NUMTHREADS`
do
cd ${array[$j]}/${HOST}.${i}/${DATE}
${QBIN} ${HERE}/io.q -reread -threads $NUMTHREADS >> ${HERE}/${DATE}/RES-${HOST}-${NUMTHREADS}t-${i} 2>&1 &
j=`expr $j + 1`
if [ $j -ge $NUMSEGS ]
then
j=0
fi
done
wait
#
j=0
ELAPSED=$(grep 'End thread -23! mapped reread' ${HERE}/${DATE}/RES-${HOST}-${NUMTHREADS}t-* | awk '{print $6}'| awk '{printf "%.3f\n",$1/1000}' | sort -nr | head -1)
# Use filesize which is direct from real filesize from q, e.g compressed data
SIZE=$(grep '^filesize' ${HERE}/${DATE}/RES-${HOST}-${NUMTHREADS}t-1 | awk '{print $2}')
SIZE=$(echo $SIZE | awk '{printf "%.0f",$1}')
# this is where we catch the process aggregation...
SIZE=$(expr $SIZE \* $NUMTHREADS )
echo $SIZE $ELAPSED
echo $SIZE $ELAPSED | awk '{$1=sprintf("%5.2f",$1/$2);print "Streaming ReRead(mapped) Rate: ", $1," MiB/sec"}' | tee -a ${HERE}/${DATE}/aggregates-${HOST}
#
# sync up across multiple host testing...
#
while [ `ls -l ${HERE}/sync-* 2> /dev/null | wc -l` -ne 0 ]
do
sleep 0.5
done
# air gap for any remote stats collection....
sleep 5
#
# META DATA tests.......................................................
#
echo Starting metadata tests...
#
# simple semaphore for completion checking for all hosts ...
#
touch ${HERE}/sync2-$HOST
j=0
cd $HERE
for i in `seq $NUMTHREADS`
do
cd ${array[$j]}/${HOST}.${i}/${DATE}
${QBIN} ${HERE}/io.q -meta -threads $NUMTHREADS >> ${HERE}/${DATE}/RES-${HOST}-${NUMTHREADS}t-${i} 2>&1 &
j=`expr $j + 1`
if [ $j -ge $NUMSEGS ]
then
j=0
fi
done
wait
#
rm -f ${HERE}/sync2-$HOST
while [ `ls -l ${HERE}/sync2-* 2> /dev/null | wc -l` -ne 0 ]
do
sleep 0.5
done
echo "flushing buffer cache....."
cd $HERE
if [ $MYID -eq 0 ]
then
./flush.sh
else
sudo ./flush.sh
fi
for FUNC in random1m random64k random1mu random64ku
do
#
#
touch ${HERE}/sync2-$HOST
j=0
cd $HERE
echo -n "$FUNC ..." | tee -a ${HERE}/${DATE}/aggregates-${HOST}
sleep 5
for i in `seq $NUMTHREADS`
do
cd ${array[$j]}/${HOST}.${i}/${DATE}
${QBIN} ${HERE}/io.q -${FUNC} -threads $NUMTHREADS >> ${HERE}/${DATE}/RES-${HOST}-${NUMTHREADS}t-${i} 2>&1 &
j=`expr $j + 1`
if [ $j -ge $NUMSEGS ]
then
j=0
fi
done
START=$(date +%s%3N)
wait
# 100 M longs x 8 bytes...
FINISH=$(date +%s%3N)
ELAPSED=$(expr $FINISH - $START)
SIZE=$(( 800 * $NUMTHREADS ))
echo $SIZE $ELAPSED | awk '{$1=sprintf("%5.2f",$1/($2/1000));print ": ", $1," MiB/sec"}' | tee -a ${HERE}/${DATE}/aggregates-${HOST}
rm ${HERE}/sync2-$HOST
while [ `ls -l ${HERE}/sync2-* 2> /dev/null | wc -l` -ne 0 ]
do
sleep 0.5
done
#
#
sleep 3
echo "flushing buffer cache after running test ${FUNC} ....."
cd $HERE
if [ $MYID -eq 0 ]
then
./flush.sh
else
sudo ./flush.sh
fi
done
#
# an air gap for any storage stats gathering before unlinks go out ...
#
sleep 5
if [ "$2" = "delete" ]
then
echo "cleaning up DB..."
j=0
for i in `seq $NUMTHREADS`
do
rm -rf ${array[$j]}/${HOST}.${i}/${DATE}
j=`expr $j + 1`
if [ $j -ge $NUMSEGS ]
then
j=0
fi
done
fi
rm -rf ${HERE}/sync-*