Merge pull request #1 from vitalif/master

16 objects per thread + generate random offsets for every op
master
rumanzo 2019-02-27 20:44:51 +03:00 committed by GitHub
commit 4cac3d310f
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
1 changed files with 66 additions and 61 deletions

View File

@ -13,25 +13,7 @@ import (
"time" "time"
) )
//future feature func bench(cephconn *Cephconnection, osddevice Device, buffs *[][]byte, bs int64, objsize int64, params *Params,
func makeoffsets(threads int64, bs int64, objsize int64) [][]int64 {
var offsets [][]int64
for i := int64(0); i < threads; i++ {
s1 := rand.NewSource(i)
r1 := rand.New(s1)
localoffsets := make([]int64, 0, objsize-bs)
for i := int64(0); i < objsize-bs; i += bs {
localoffsets = append(localoffsets, i)
}
r1.Shuffle(len(localoffsets), func(i, j int) {
localoffsets[i], localoffsets[j] = localoffsets[j], localoffsets[i]
})
offsets = append(offsets, localoffsets)
}
return offsets
}
func bench(cephconn *Cephconnection, osddevice Device, buffs *[][]byte, offset [][]int64, params *Params,
wg *sync.WaitGroup, result chan string) { wg *sync.WaitGroup, result chan string) {
defer wg.Done() defer wg.Done()
threadresult := make(chan []time.Duration, params.threadsCount) threadresult := make(chan []time.Duration, params.threadsCount)
@ -43,14 +25,14 @@ func bench(cephconn *Cephconnection, osddevice Device, buffs *[][]byte, offset [
} }
}() }()
// calculate object for each thread // calculate object for each thread
for suffix := 0; len(objectnames) < int(params.threadsCount); suffix++ { for suffix := 0; len(objectnames) < int(params.threadsCount)*16; suffix++ {
name := "bench_" + strconv.Itoa(suffix) name := "bench_" + strconv.Itoa(suffix)
if osddevice.ID == GetObjActingPrimary(cephconn, *params, name) { if osddevice.ID == GetObjActingPrimary(cephconn, *params, name) {
objectnames = append(objectnames, name) objectnames = append(objectnames, name)
} }
} }
for i, j := 0, 0; i < int(params.threadsCount); i, j = i+1, j+2 { for i := 0; i < int(params.threadsCount); i++ {
go bench_thread(cephconn, osddevice, (*buffs)[j:j+2], offset[i], params, threadresult, objectnames[i]) go bench_thread(cephconn, osddevice, (*buffs)[i*2:i*2+2], bs, objsize, params, threadresult, objectnames[i*16:i*16+16])
} }
for i := int64(0); i < params.threadsCount; i++ { for i := int64(0); i < params.threadsCount; i++ {
for _, lat := range <-threadresult { for _, lat := range <-threadresult {
@ -58,16 +40,25 @@ func bench(cephconn *Cephconnection, osddevice Device, buffs *[][]byte, offset [
} }
} }
close(threadresult) close(threadresult)
latencygrade := map[int]int{} latencygrade := map[int64]int{}
latencytotal := int64(0)
for _, lat := range osdlatencies { for _, lat := range osdlatencies {
micro := lat.Nanoseconds()/1000
rounded := micro
switch { switch {
case lat < time.Millisecond*10: case micro < 1000: // 0-1ms round to 0.1ms
latencygrade[int(lat.Round(time.Millisecond).Nanoseconds()/1000000)]++ rounded = (micro/100)*100
case lat < time.Millisecond*100: case micro < 10000: // 2-10ms round to 1ms
latencygrade[int(lat.Round(time.Millisecond*10)/1000000)]++ rounded = (micro/1000)*1000
default: case micro < 100000: // 10-100ms round to 10ms
latencygrade[int(lat.Round(time.Millisecond*100)/1000000)]++ rounded = (micro/10000)*10000
case micro < 1000000: // 100-1000ms round to 100ms
rounded = (micro/100000)*100000
default: // 1000+ms round to 1s
rounded = (micro/1000000)*1000000
} }
latencytotal += micro
latencygrade[rounded]++
} }
var buffer bytes.Buffer var buffer bytes.Buffer
@ -100,62 +91,77 @@ func bench(cephconn *Cephconnection, osddevice Device, buffs *[][]byte, offset [
} }
} }
latencytotal = latencytotal/int64(len(osdlatencies))
// iops = 1s / latency
iops := 1000000 / latencytotal
// avg speed = iops * block size / 1 MB
avgspeed := (1000000 / float64(latencytotal) * float64(params.blocksize) / 1024 / 1024)
avgline := fmt.Sprintf("Avg iops: %-5v Avg speed: %.3f MB/s\n\n", iops, avgspeed)
switch {
case iops < 80:
buffer.WriteString(darkred(avgline))
case iops < 200:
buffer.WriteString(red(avgline))
case iops < 500:
buffer.WriteString(yellow(avgline))
default:
buffer.WriteString(green(avgline))
}
//sort latencies //sort latencies
var keys []int var keys []int64
for k := range latencygrade { for k := range latencygrade {
keys = append(keys, k) keys = append(keys, k)
} }
sort.Ints(keys) sort.Slice(keys, func(i, j int) bool { return keys[i] < keys[j] })
for _, k := range keys { for _, k := range keys {
var blocks bytes.Buffer var blocks bytes.Buffer
var mseconds string var mseconds string
switch { switch {
case 10 <= k && k < 20: case k < 1000:
mseconds = green(fmt.Sprintf("[%v-%v]", k, k+9)) mseconds = green(fmt.Sprintf("[%.1f-%.1f)", float64(k)/1000, 0.1+float64(k)/1000))
case 20 <= k && k < 100: case k < 2000:
mseconds = red(fmt.Sprintf("[%v-%v]", k, k+9)) mseconds = yellow(fmt.Sprintf("[%.1f-%.1f)", float64(k)/1000, 0.1+float64(k)/1000))
case k >= 100: case k < 10000:
mseconds = darkred(fmt.Sprintf("[%v-%v]", k, k+99)) mseconds = yellow(fmt.Sprintf("[%3v-%3v)", k/1000, 1+k/1000))
case k < 100000:
mseconds = red(fmt.Sprintf("[%3v-%3v)", k/1000, 10+k/1000))
case k < 1000000:
mseconds = darkred(fmt.Sprintf("[%3v-%3v]", k/1000, 99+k/1000))
default: default:
mseconds = green(k) mseconds = darkred(fmt.Sprintf("[%2vs-%2vs]", k/1000000, 1+k/1000000))
} }
for i := 0; i < 50*(latencygrade[k]*100/len(osdlatencies))/100; i++ { for i := 0; i < 50*(latencygrade[k]*100/len(osdlatencies))/100; i++ {
blocks.WriteString("#") blocks.WriteString("#")
} }
iops := latencygrade[k] / int(params.duration.Seconds())
avgspeed := (float64(latencygrade[k]) * float64(params.blocksize) / float64(params.duration.Seconds())) / 1024 / 1024 //mb/sec
megabyteswritten := (float64(latencygrade[k]) * float64(params.blocksize)) / 1024 / 1024 megabyteswritten := (float64(latencygrade[k]) * float64(params.blocksize)) / 1024 / 1024
buffer.WriteString(fmt.Sprintf("%+9v ms: [%-50v] Count: %-5v IOPS: %-5v Avg speed: %-6.3f Mb/Sec Summary written: %6.3f Mb\n", buffer.WriteString(fmt.Sprintf("%+9v ms: [%-50v] Count: %-5v Total written: %6.3f MB\n",
mseconds, blocks.String(), latencygrade[k], iops, avgspeed, megabyteswritten)) mseconds, blocks.String(), latencygrade[k], megabyteswritten))
} }
result <- buffer.String() result <- buffer.String()
} }
func bench_thread(cephconn *Cephconnection, osddevice Device, buffs [][]byte, offsets []int64, params *Params, func bench_thread(cephconn *Cephconnection, osddevice Device, buffs [][]byte, bs int64, objsize int64, params *Params,
result chan []time.Duration, objname string) { result chan []time.Duration, objnames []string) {
starttime := time.Now() starttime := time.Now()
var latencies []time.Duration var latencies []time.Duration
endtime := starttime.Add(params.duration) endtime := starttime.Add(params.duration)
n := 0 n := 0
for { for {
if time.Now().After(endtime) { offset := rand.Int63n(objsize/bs) * bs
objname := objnames[rand.Int31n(int32(len(objnames)))]
startwritetime := time.Now()
if startwritetime.After(endtime) {
break break
} }
for _, offset := range offsets { err := cephconn.ioctx.Write(objname, buffs[n], uint64(offset))
if time.Now().Before(endtime) { endwritetime := time.Now()
startwritetime := time.Now() if err != nil {
err := cephconn.ioctx.Write(objname, buffs[n], uint64(offset)) log.Printf("Can't write obj: %v, osd: %v", objname, osddevice.Name)
endwritetime := time.Now() continue
if err != nil {
log.Printf("Can't write obj: %v, osd: %v", objname, osddevice.Name)
continue
}
latencies = append(latencies, endwritetime.Sub(startwritetime))
} else {
break
}
} }
latencies = append(latencies, endwritetime.Sub(startwritetime))
if n == 0 { if n == 0 {
n++ n++
} else { } else {
@ -184,16 +190,15 @@ func main() {
} }
} }
osddevices := GetOsds(cephconn, params) osddevices := GetOsds(cephconn, params)
offsets := makeoffsets(params.threadsCount, params.blocksize, params.objectsize)
var wg sync.WaitGroup var wg sync.WaitGroup
results := make(chan string, len(osddevices)*int(params.threadsCount)) results := make(chan string, len(osddevices)*int(params.threadsCount))
for _, osd := range osddevices { for _, osd := range osddevices {
wg.Add(1) wg.Add(1)
if params.parallel == true { if params.parallel == true {
go bench(cephconn, osd, &buffs, offsets, &params, &wg, results) go bench(cephconn, osd, &buffs, params.blocksize, params.objectsize, &params, &wg, results)
} else { } else {
bench(cephconn, osd, &buffs, offsets, &params, &wg, results) bench(cephconn, osd, &buffs, params.blocksize, params.objectsize, &params, &wg, results)
log.Println(<-results) log.Println(<-results)
} }