Merge pull request #1 from vitalif/master
16 objects per thread + generate random offsets for every opmaster
commit
4cac3d310f
127
ceph-gobench.go
127
ceph-gobench.go
|
@ -13,25 +13,7 @@ import (
|
||||||
"time"
|
"time"
|
||||||
)
|
)
|
||||||
|
|
||||||
//future feature
|
func bench(cephconn *Cephconnection, osddevice Device, buffs *[][]byte, bs int64, objsize int64, params *Params,
|
||||||
func makeoffsets(threads int64, bs int64, objsize int64) [][]int64 {
|
|
||||||
var offsets [][]int64
|
|
||||||
for i := int64(0); i < threads; i++ {
|
|
||||||
s1 := rand.NewSource(i)
|
|
||||||
r1 := rand.New(s1)
|
|
||||||
localoffsets := make([]int64, 0, objsize-bs)
|
|
||||||
for i := int64(0); i < objsize-bs; i += bs {
|
|
||||||
localoffsets = append(localoffsets, i)
|
|
||||||
}
|
|
||||||
r1.Shuffle(len(localoffsets), func(i, j int) {
|
|
||||||
localoffsets[i], localoffsets[j] = localoffsets[j], localoffsets[i]
|
|
||||||
})
|
|
||||||
offsets = append(offsets, localoffsets)
|
|
||||||
}
|
|
||||||
return offsets
|
|
||||||
}
|
|
||||||
|
|
||||||
func bench(cephconn *Cephconnection, osddevice Device, buffs *[][]byte, offset [][]int64, params *Params,
|
|
||||||
wg *sync.WaitGroup, result chan string) {
|
wg *sync.WaitGroup, result chan string) {
|
||||||
defer wg.Done()
|
defer wg.Done()
|
||||||
threadresult := make(chan []time.Duration, params.threadsCount)
|
threadresult := make(chan []time.Duration, params.threadsCount)
|
||||||
|
@ -43,14 +25,14 @@ func bench(cephconn *Cephconnection, osddevice Device, buffs *[][]byte, offset [
|
||||||
}
|
}
|
||||||
}()
|
}()
|
||||||
// calculate object for each thread
|
// calculate object for each thread
|
||||||
for suffix := 0; len(objectnames) < int(params.threadsCount); suffix++ {
|
for suffix := 0; len(objectnames) < int(params.threadsCount)*16; suffix++ {
|
||||||
name := "bench_" + strconv.Itoa(suffix)
|
name := "bench_" + strconv.Itoa(suffix)
|
||||||
if osddevice.ID == GetObjActingPrimary(cephconn, *params, name) {
|
if osddevice.ID == GetObjActingPrimary(cephconn, *params, name) {
|
||||||
objectnames = append(objectnames, name)
|
objectnames = append(objectnames, name)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
for i, j := 0, 0; i < int(params.threadsCount); i, j = i+1, j+2 {
|
for i := 0; i < int(params.threadsCount); i++ {
|
||||||
go bench_thread(cephconn, osddevice, (*buffs)[j:j+2], offset[i], params, threadresult, objectnames[i])
|
go bench_thread(cephconn, osddevice, (*buffs)[i*2:i*2+2], bs, objsize, params, threadresult, objectnames[i*16:i*16+16])
|
||||||
}
|
}
|
||||||
for i := int64(0); i < params.threadsCount; i++ {
|
for i := int64(0); i < params.threadsCount; i++ {
|
||||||
for _, lat := range <-threadresult {
|
for _, lat := range <-threadresult {
|
||||||
|
@ -58,16 +40,25 @@ func bench(cephconn *Cephconnection, osddevice Device, buffs *[][]byte, offset [
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
close(threadresult)
|
close(threadresult)
|
||||||
latencygrade := map[int]int{}
|
latencygrade := map[int64]int{}
|
||||||
|
latencytotal := int64(0)
|
||||||
for _, lat := range osdlatencies {
|
for _, lat := range osdlatencies {
|
||||||
|
micro := lat.Nanoseconds()/1000
|
||||||
|
rounded := micro
|
||||||
switch {
|
switch {
|
||||||
case lat < time.Millisecond*10:
|
case micro < 1000: // 0-1ms round to 0.1ms
|
||||||
latencygrade[int(lat.Round(time.Millisecond).Nanoseconds()/1000000)]++
|
rounded = (micro/100)*100
|
||||||
case lat < time.Millisecond*100:
|
case micro < 10000: // 2-10ms round to 1ms
|
||||||
latencygrade[int(lat.Round(time.Millisecond*10)/1000000)]++
|
rounded = (micro/1000)*1000
|
||||||
default:
|
case micro < 100000: // 10-100ms round to 10ms
|
||||||
latencygrade[int(lat.Round(time.Millisecond*100)/1000000)]++
|
rounded = (micro/10000)*10000
|
||||||
|
case micro < 1000000: // 100-1000ms round to 100ms
|
||||||
|
rounded = (micro/100000)*100000
|
||||||
|
default: // 1000+ms round to 1s
|
||||||
|
rounded = (micro/1000000)*1000000
|
||||||
}
|
}
|
||||||
|
latencytotal += micro
|
||||||
|
latencygrade[rounded]++
|
||||||
}
|
}
|
||||||
|
|
||||||
var buffer bytes.Buffer
|
var buffer bytes.Buffer
|
||||||
|
@ -100,62 +91,77 @@ func bench(cephconn *Cephconnection, osddevice Device, buffs *[][]byte, offset [
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
latencytotal = latencytotal/int64(len(osdlatencies))
|
||||||
|
// iops = 1s / latency
|
||||||
|
iops := 1000000 / latencytotal
|
||||||
|
// avg speed = iops * block size / 1 MB
|
||||||
|
avgspeed := (1000000 / float64(latencytotal) * float64(params.blocksize) / 1024 / 1024)
|
||||||
|
avgline := fmt.Sprintf("Avg iops: %-5v Avg speed: %.3f MB/s\n\n", iops, avgspeed)
|
||||||
|
switch {
|
||||||
|
case iops < 80:
|
||||||
|
buffer.WriteString(darkred(avgline))
|
||||||
|
case iops < 200:
|
||||||
|
buffer.WriteString(red(avgline))
|
||||||
|
case iops < 500:
|
||||||
|
buffer.WriteString(yellow(avgline))
|
||||||
|
default:
|
||||||
|
buffer.WriteString(green(avgline))
|
||||||
|
}
|
||||||
|
|
||||||
//sort latencies
|
//sort latencies
|
||||||
var keys []int
|
var keys []int64
|
||||||
for k := range latencygrade {
|
for k := range latencygrade {
|
||||||
keys = append(keys, k)
|
keys = append(keys, k)
|
||||||
}
|
}
|
||||||
sort.Ints(keys)
|
sort.Slice(keys, func(i, j int) bool { return keys[i] < keys[j] })
|
||||||
for _, k := range keys {
|
for _, k := range keys {
|
||||||
var blocks bytes.Buffer
|
var blocks bytes.Buffer
|
||||||
var mseconds string
|
var mseconds string
|
||||||
switch {
|
switch {
|
||||||
case 10 <= k && k < 20:
|
case k < 1000:
|
||||||
mseconds = green(fmt.Sprintf("[%v-%v]", k, k+9))
|
mseconds = green(fmt.Sprintf("[%.1f-%.1f)", float64(k)/1000, 0.1+float64(k)/1000))
|
||||||
case 20 <= k && k < 100:
|
case k < 2000:
|
||||||
mseconds = red(fmt.Sprintf("[%v-%v]", k, k+9))
|
mseconds = yellow(fmt.Sprintf("[%.1f-%.1f)", float64(k)/1000, 0.1+float64(k)/1000))
|
||||||
case k >= 100:
|
case k < 10000:
|
||||||
mseconds = darkred(fmt.Sprintf("[%v-%v]", k, k+99))
|
mseconds = yellow(fmt.Sprintf("[%3v-%3v)", k/1000, 1+k/1000))
|
||||||
|
case k < 100000:
|
||||||
|
mseconds = red(fmt.Sprintf("[%3v-%3v)", k/1000, 10+k/1000))
|
||||||
|
case k < 1000000:
|
||||||
|
mseconds = darkred(fmt.Sprintf("[%3v-%3v]", k/1000, 99+k/1000))
|
||||||
default:
|
default:
|
||||||
mseconds = green(k)
|
mseconds = darkred(fmt.Sprintf("[%2vs-%2vs]", k/1000000, 1+k/1000000))
|
||||||
}
|
}
|
||||||
for i := 0; i < 50*(latencygrade[k]*100/len(osdlatencies))/100; i++ {
|
for i := 0; i < 50*(latencygrade[k]*100/len(osdlatencies))/100; i++ {
|
||||||
blocks.WriteString("#")
|
blocks.WriteString("#")
|
||||||
}
|
}
|
||||||
iops := latencygrade[k] / int(params.duration.Seconds())
|
|
||||||
avgspeed := (float64(latencygrade[k]) * float64(params.blocksize) / float64(params.duration.Seconds())) / 1024 / 1024 //mb/sec
|
|
||||||
megabyteswritten := (float64(latencygrade[k]) * float64(params.blocksize)) / 1024 / 1024
|
megabyteswritten := (float64(latencygrade[k]) * float64(params.blocksize)) / 1024 / 1024
|
||||||
buffer.WriteString(fmt.Sprintf("%+9v ms: [%-50v] Count: %-5v IOPS: %-5v Avg speed: %-6.3f Mb/Sec Summary written: %6.3f Mb\n",
|
buffer.WriteString(fmt.Sprintf("%+9v ms: [%-50v] Count: %-5v Total written: %6.3f MB\n",
|
||||||
mseconds, blocks.String(), latencygrade[k], iops, avgspeed, megabyteswritten))
|
mseconds, blocks.String(), latencygrade[k], megabyteswritten))
|
||||||
}
|
}
|
||||||
result <- buffer.String()
|
result <- buffer.String()
|
||||||
}
|
}
|
||||||
|
|
||||||
func bench_thread(cephconn *Cephconnection, osddevice Device, buffs [][]byte, offsets []int64, params *Params,
|
func bench_thread(cephconn *Cephconnection, osddevice Device, buffs [][]byte, bs int64, objsize int64, params *Params,
|
||||||
result chan []time.Duration, objname string) {
|
result chan []time.Duration, objnames []string) {
|
||||||
|
|
||||||
starttime := time.Now()
|
starttime := time.Now()
|
||||||
var latencies []time.Duration
|
var latencies []time.Duration
|
||||||
endtime := starttime.Add(params.duration)
|
endtime := starttime.Add(params.duration)
|
||||||
n := 0
|
n := 0
|
||||||
for {
|
for {
|
||||||
if time.Now().After(endtime) {
|
offset := rand.Int63n(objsize/bs) * bs
|
||||||
|
objname := objnames[rand.Int31n(int32(len(objnames)))]
|
||||||
|
startwritetime := time.Now()
|
||||||
|
if startwritetime.After(endtime) {
|
||||||
break
|
break
|
||||||
}
|
}
|
||||||
for _, offset := range offsets {
|
err := cephconn.ioctx.Write(objname, buffs[n], uint64(offset))
|
||||||
if time.Now().Before(endtime) {
|
endwritetime := time.Now()
|
||||||
startwritetime := time.Now()
|
if err != nil {
|
||||||
err := cephconn.ioctx.Write(objname, buffs[n], uint64(offset))
|
log.Printf("Can't write obj: %v, osd: %v", objname, osddevice.Name)
|
||||||
endwritetime := time.Now()
|
continue
|
||||||
if err != nil {
|
|
||||||
log.Printf("Can't write obj: %v, osd: %v", objname, osddevice.Name)
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
latencies = append(latencies, endwritetime.Sub(startwritetime))
|
|
||||||
} else {
|
|
||||||
break
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
latencies = append(latencies, endwritetime.Sub(startwritetime))
|
||||||
if n == 0 {
|
if n == 0 {
|
||||||
n++
|
n++
|
||||||
} else {
|
} else {
|
||||||
|
@ -184,16 +190,15 @@ func main() {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
osddevices := GetOsds(cephconn, params)
|
osddevices := GetOsds(cephconn, params)
|
||||||
offsets := makeoffsets(params.threadsCount, params.blocksize, params.objectsize)
|
|
||||||
|
|
||||||
var wg sync.WaitGroup
|
var wg sync.WaitGroup
|
||||||
results := make(chan string, len(osddevices)*int(params.threadsCount))
|
results := make(chan string, len(osddevices)*int(params.threadsCount))
|
||||||
for _, osd := range osddevices {
|
for _, osd := range osddevices {
|
||||||
wg.Add(1)
|
wg.Add(1)
|
||||||
if params.parallel == true {
|
if params.parallel == true {
|
||||||
go bench(cephconn, osd, &buffs, offsets, ¶ms, &wg, results)
|
go bench(cephconn, osd, &buffs, params.blocksize, params.objectsize, ¶ms, &wg, results)
|
||||||
} else {
|
} else {
|
||||||
bench(cephconn, osd, &buffs, offsets, ¶ms, &wg, results)
|
bench(cephconn, osd, &buffs, params.blocksize, params.objectsize, ¶ms, &wg, results)
|
||||||
log.Println(<-results)
|
log.Println(<-results)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue