2019-02-14 15:01:33 +03:00
|
|
|
package main
|
|
|
|
|
|
|
|
import (
|
2019-02-27 13:46:44 +03:00
|
|
|
"bytes"
|
2019-02-19 15:46:00 +03:00
|
|
|
"fmt"
|
2019-02-27 15:04:48 +03:00
|
|
|
"github.com/fatih/color"
|
2019-02-14 15:01:33 +03:00
|
|
|
"log"
|
2019-02-14 17:35:18 +03:00
|
|
|
"math/rand"
|
2019-02-27 13:46:44 +03:00
|
|
|
"sort"
|
2019-02-26 18:09:59 +03:00
|
|
|
"strconv"
|
2019-02-27 15:04:48 +03:00
|
|
|
"strings"
|
2019-02-22 17:58:31 +03:00
|
|
|
"sync"
|
2019-02-14 19:18:43 +03:00
|
|
|
"time"
|
2019-02-14 15:01:33 +03:00
|
|
|
)
|
|
|
|
|
2019-02-27 21:12:08 +03:00
|
|
|
func bench(cephconn *Cephconnection, osddevice Device, buffs *[][]byte, startbuf *[]byte, params *Params,
|
2019-02-27 13:46:44 +03:00
|
|
|
wg *sync.WaitGroup, result chan string) {
|
2019-02-22 17:58:31 +03:00
|
|
|
defer wg.Done()
|
2019-02-27 13:46:44 +03:00
|
|
|
threadresult := make(chan []time.Duration, params.threadsCount)
|
|
|
|
var objectnames []string
|
|
|
|
var osdlatencies []time.Duration
|
2019-02-27 11:26:40 +03:00
|
|
|
defer func() {
|
|
|
|
for _, object := range objectnames {
|
|
|
|
cephconn.ioctx.Delete(object)
|
|
|
|
}
|
|
|
|
}()
|
2019-02-26 18:09:59 +03:00
|
|
|
// calculate object for each thread
|
2019-02-27 17:53:07 +03:00
|
|
|
for suffix := 0; len(objectnames) < int(params.threadsCount)*16; suffix++ {
|
2019-02-26 18:09:59 +03:00
|
|
|
name := "bench_" + strconv.Itoa(suffix)
|
|
|
|
if osddevice.ID == GetObjActingPrimary(cephconn, *params, name) {
|
|
|
|
objectnames = append(objectnames, name)
|
2019-02-27 21:12:08 +03:00
|
|
|
if err := cephconn.ioctx.WriteFull(name, *startbuf); err != nil {
|
|
|
|
log.Printf("Can't write object: %v, osd: %v", name, osddevice.Name)
|
|
|
|
}
|
2019-02-26 18:09:59 +03:00
|
|
|
}
|
|
|
|
}
|
2019-02-27 17:53:07 +03:00
|
|
|
for i := 0; i < int(params.threadsCount); i++ {
|
2019-02-27 21:12:08 +03:00
|
|
|
go BenchThread(cephconn, osddevice, (*buffs)[i*2:i*2+2], params, threadresult, objectnames[i*16:i*16+16])
|
2019-02-26 13:20:51 +03:00
|
|
|
}
|
|
|
|
for i := int64(0); i < params.threadsCount; i++ {
|
2019-02-27 13:46:44 +03:00
|
|
|
for _, lat := range <-threadresult {
|
|
|
|
osdlatencies = append(osdlatencies, lat)
|
|
|
|
}
|
2019-02-22 17:58:31 +03:00
|
|
|
}
|
2019-02-26 13:20:51 +03:00
|
|
|
close(threadresult)
|
2019-02-27 18:39:13 +03:00
|
|
|
latencygrade := map[int64]int{}
|
|
|
|
latencytotal := int64(0)
|
2019-02-27 13:46:44 +03:00
|
|
|
for _, lat := range osdlatencies {
|
2019-02-27 21:12:08 +03:00
|
|
|
micro := lat.Nanoseconds() / 1000
|
2019-02-27 18:39:13 +03:00
|
|
|
rounded := micro
|
2019-02-27 13:46:44 +03:00
|
|
|
switch {
|
2019-02-27 18:39:13 +03:00
|
|
|
case micro < 1000: // 0-1ms round to 0.1ms
|
2019-02-27 21:12:08 +03:00
|
|
|
rounded = (micro / 100) * 100
|
2019-02-27 18:39:13 +03:00
|
|
|
case micro < 10000: // 2-10ms round to 1ms
|
2019-02-27 21:12:08 +03:00
|
|
|
rounded = (micro / 1000) * 1000
|
2019-02-27 18:39:13 +03:00
|
|
|
case micro < 100000: // 10-100ms round to 10ms
|
2019-02-27 21:12:08 +03:00
|
|
|
rounded = (micro / 10000) * 10000
|
2019-02-27 18:39:13 +03:00
|
|
|
case micro < 1000000: // 100-1000ms round to 100ms
|
2019-02-27 21:12:08 +03:00
|
|
|
rounded = (micro / 100000) * 100000
|
2019-02-27 18:39:13 +03:00
|
|
|
default: // 1000+ms round to 1s
|
2019-02-27 21:12:08 +03:00
|
|
|
rounded = (micro / 1000000) * 1000000
|
2019-02-27 13:46:44 +03:00
|
|
|
}
|
2019-02-27 18:39:13 +03:00
|
|
|
latencytotal += micro
|
|
|
|
latencygrade[rounded]++
|
2019-02-27 13:46:44 +03:00
|
|
|
}
|
2019-02-27 15:49:36 +03:00
|
|
|
|
2019-02-27 13:46:44 +03:00
|
|
|
var buffer bytes.Buffer
|
2019-02-27 15:49:36 +03:00
|
|
|
|
|
|
|
//color info
|
2019-02-27 15:04:48 +03:00
|
|
|
yellow := color.New(color.FgHiYellow).SprintFunc()
|
2019-02-27 15:49:36 +03:00
|
|
|
red := color.New(color.FgHiRed).SprintFunc()
|
|
|
|
darkred := color.New(color.FgRed).SprintFunc()
|
2019-02-27 15:04:48 +03:00
|
|
|
green := color.New(color.FgHiGreen).SprintFunc()
|
2019-02-27 15:49:36 +03:00
|
|
|
darkgreen := color.New(color.FgGreen).SprintFunc()
|
2019-02-27 15:04:48 +03:00
|
|
|
buffer.WriteString(fmt.Sprintf("Bench result for %v\n", osddevice.Name))
|
|
|
|
infos := map[string]string{"front_addr": strings.Split(osddevice.Info.FrontAddr, "/")[0],
|
|
|
|
"ceph_release/version": osddevice.Info.CephRelease + "/" + osddevice.Info.CephVersionShort, "cpu": osddevice.Info.CPU,
|
|
|
|
"hostname": osddevice.Info.Hostname, "default_device_class": osddevice.Info.DefaultDeviceClass, "devices": osddevice.Info.Devices,
|
|
|
|
"distro_description": osddevice.Info.DistroDescription, "journal_rotational": osddevice.Info.JournalRotational,
|
|
|
|
"rotational": osddevice.Info.Rotational, "kernel_version": osddevice.Info.KernelVersion, "mem_swap_kb": osddevice.Info.MemSwapKb,
|
|
|
|
"mem_total_kb": osddevice.Info.MemTotalKb, "osd_data": osddevice.Info.OsdData, "osd_objectstore": osddevice.Info.OsdObjectstore}
|
|
|
|
infonum := 1
|
|
|
|
var infokeys []string
|
|
|
|
for k := range infos {
|
|
|
|
infokeys = append(infokeys, k)
|
|
|
|
}
|
|
|
|
sort.Strings(infokeys)
|
2019-02-27 15:49:36 +03:00
|
|
|
buffer.WriteString(fmt.Sprintf("%-30v %-45v", darkgreen("osdname"), red(osddevice.Name)))
|
2019-02-27 15:04:48 +03:00
|
|
|
for _, key := range infokeys {
|
|
|
|
infonum++
|
2019-02-27 15:49:36 +03:00
|
|
|
buffer.WriteString(fmt.Sprintf("%-30v %-45v", darkgreen(key), yellow(infos[key])))
|
2019-02-27 15:04:48 +03:00
|
|
|
if (infonum % 3) == 0 {
|
|
|
|
buffer.WriteString("\n")
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2019-02-27 21:12:08 +03:00
|
|
|
latencytotal = latencytotal / int64(len(osdlatencies))
|
2019-02-27 18:54:24 +03:00
|
|
|
// iops = 1s / latency
|
|
|
|
iops := 1000000 / latencytotal
|
2019-02-27 18:43:28 +03:00
|
|
|
// avg speed = iops * block size / 1 MB
|
|
|
|
avgspeed := (1000000 / float64(latencytotal) * float64(params.blocksize) / 1024 / 1024)
|
2019-02-27 18:54:24 +03:00
|
|
|
avgline := fmt.Sprintf("Avg iops: %-5v Avg speed: %.3f MB/s\n\n", iops, avgspeed)
|
|
|
|
switch {
|
|
|
|
case iops < 80:
|
|
|
|
buffer.WriteString(darkred(avgline))
|
|
|
|
case iops < 200:
|
|
|
|
buffer.WriteString(red(avgline))
|
|
|
|
case iops < 500:
|
|
|
|
buffer.WriteString(yellow(avgline))
|
|
|
|
default:
|
|
|
|
buffer.WriteString(green(avgline))
|
|
|
|
}
|
2019-02-27 18:43:28 +03:00
|
|
|
|
2019-02-27 15:49:36 +03:00
|
|
|
//sort latencies
|
2019-02-27 18:39:13 +03:00
|
|
|
var keys []int64
|
2019-02-27 15:49:36 +03:00
|
|
|
for k := range latencygrade {
|
|
|
|
keys = append(keys, k)
|
|
|
|
}
|
2019-02-27 18:39:13 +03:00
|
|
|
sort.Slice(keys, func(i, j int) bool { return keys[i] < keys[j] })
|
2019-02-27 13:46:44 +03:00
|
|
|
for _, k := range keys {
|
|
|
|
var blocks bytes.Buffer
|
2019-02-27 15:49:36 +03:00
|
|
|
var mseconds string
|
|
|
|
switch {
|
2019-02-27 18:39:13 +03:00
|
|
|
case k < 1000:
|
2019-02-27 22:09:53 +03:00
|
|
|
mseconds = green(fmt.Sprintf("[%.1f-%.1f)", float64(k)/1000, 0.1+float64(k)/1000))
|
2019-02-27 18:39:13 +03:00
|
|
|
case k < 2000:
|
2019-02-27 22:09:53 +03:00
|
|
|
mseconds = yellow(fmt.Sprintf("[%.1f-%.1f)", float64(k)/1000, 0.1+float64(k)/1000))
|
2019-02-27 21:57:38 +03:00
|
|
|
case k < 9000:
|
2019-02-27 22:09:53 +03:00
|
|
|
mseconds = yellow(fmt.Sprintf("[%.1f-%.1f)", float64(k/1000), float64(1+k/1000)))
|
2019-02-27 18:39:13 +03:00
|
|
|
case k < 10000:
|
2019-02-27 22:57:47 +03:00
|
|
|
mseconds = color.YellowString(fmt.Sprintf("[%.1f-%v)", float64(k/1000), 1+k/1000))
|
2019-02-27 18:39:13 +03:00
|
|
|
case k < 100000:
|
2019-02-27 22:09:53 +03:00
|
|
|
mseconds = red(fmt.Sprintf("[%3v-%3v)", k/1000, 10+k/1000))
|
2019-02-27 18:39:13 +03:00
|
|
|
case k < 1000000:
|
2019-02-27 22:09:53 +03:00
|
|
|
mseconds = darkred(fmt.Sprintf("[%3v-%3v]", k/1000, 99+k/1000))
|
2019-02-27 15:49:36 +03:00
|
|
|
default:
|
2019-02-27 22:09:53 +03:00
|
|
|
mseconds = darkred(fmt.Sprintf("[%2vs-%2vs]", k/1000000, 1+k/1000000))
|
2019-02-27 15:49:36 +03:00
|
|
|
}
|
2019-02-27 13:46:44 +03:00
|
|
|
for i := 0; i < 50*(latencygrade[k]*100/len(osdlatencies))/100; i++ {
|
|
|
|
blocks.WriteString("#")
|
|
|
|
}
|
|
|
|
megabyteswritten := (float64(latencygrade[k]) * float64(params.blocksize)) / 1024 / 1024
|
2019-02-27 22:57:47 +03:00
|
|
|
buffer.WriteString(fmt.Sprintf("%-20v ms: [%-50v] Count: %-5v Total written: %6.3f MB\n",
|
2019-02-27 18:43:28 +03:00
|
|
|
mseconds, blocks.String(), latencygrade[k], megabyteswritten))
|
2019-02-27 13:46:44 +03:00
|
|
|
}
|
|
|
|
result <- buffer.String()
|
2019-02-22 17:58:31 +03:00
|
|
|
}
|
|
|
|
|
2019-02-27 21:12:08 +03:00
|
|
|
func BenchThread(cephconn *Cephconnection, osddevice Device, buffs [][]byte, params *Params,
|
2019-02-27 17:53:07 +03:00
|
|
|
result chan []time.Duration, objnames []string) {
|
2019-02-27 11:26:40 +03:00
|
|
|
|
2019-02-26 18:09:59 +03:00
|
|
|
starttime := time.Now()
|
|
|
|
var latencies []time.Duration
|
|
|
|
endtime := starttime.Add(params.duration)
|
|
|
|
n := 0
|
|
|
|
for {
|
2019-02-27 21:12:08 +03:00
|
|
|
offset := rand.Int63n(params.objectsize/params.blocksize) * params.blocksize
|
2019-02-27 17:53:07 +03:00
|
|
|
objname := objnames[rand.Int31n(int32(len(objnames)))]
|
|
|
|
startwritetime := time.Now()
|
|
|
|
if startwritetime.After(endtime) {
|
2019-02-26 18:09:59 +03:00
|
|
|
break
|
|
|
|
}
|
2019-02-27 17:53:07 +03:00
|
|
|
err := cephconn.ioctx.Write(objname, buffs[n], uint64(offset))
|
|
|
|
endwritetime := time.Now()
|
|
|
|
if err != nil {
|
2019-02-27 21:12:08 +03:00
|
|
|
log.Printf("Can't write object: %v, osd: %v", objname, osddevice.Name)
|
2019-02-27 17:53:07 +03:00
|
|
|
continue
|
2019-02-26 18:09:59 +03:00
|
|
|
}
|
2019-02-27 17:53:07 +03:00
|
|
|
latencies = append(latencies, endwritetime.Sub(startwritetime))
|
2019-02-26 18:09:59 +03:00
|
|
|
if n == 0 {
|
|
|
|
n++
|
|
|
|
} else {
|
|
|
|
n = 0
|
|
|
|
}
|
|
|
|
}
|
2019-02-27 13:46:44 +03:00
|
|
|
result <- latencies
|
2019-02-19 15:46:00 +03:00
|
|
|
}
|
|
|
|
|
2019-02-14 15:01:33 +03:00
|
|
|
func main() {
|
2019-02-14 17:35:18 +03:00
|
|
|
params := Route()
|
2019-02-15 20:04:32 +03:00
|
|
|
cephconn := connectioninit(params)
|
2019-02-14 17:35:18 +03:00
|
|
|
defer cephconn.conn.Shutdown()
|
2019-02-14 17:59:40 +03:00
|
|
|
|
2019-02-14 19:18:43 +03:00
|
|
|
// https://tracker.ceph.com/issues/24114
|
|
|
|
time.Sleep(time.Millisecond * 100)
|
|
|
|
|
2019-02-14 17:35:18 +03:00
|
|
|
var buffs [][]byte
|
2019-02-22 17:58:31 +03:00
|
|
|
for i := int64(0); i < 2*params.threadsCount; i++ {
|
2019-02-14 17:59:40 +03:00
|
|
|
buffs = append(buffs, make([]byte, params.blocksize))
|
2019-02-14 17:35:18 +03:00
|
|
|
}
|
2019-02-27 21:12:08 +03:00
|
|
|
startbuff := make([]byte, params.objectsize)
|
2019-02-14 17:35:18 +03:00
|
|
|
for num := range buffs {
|
|
|
|
_, err := rand.Read(buffs[num])
|
|
|
|
if err != nil {
|
|
|
|
log.Fatalln(err)
|
|
|
|
}
|
|
|
|
}
|
2019-02-22 17:58:31 +03:00
|
|
|
osddevices := GetOsds(cephconn, params)
|
|
|
|
|
|
|
|
var wg sync.WaitGroup
|
2019-02-27 13:46:44 +03:00
|
|
|
results := make(chan string, len(osddevices)*int(params.threadsCount))
|
2019-02-26 13:20:51 +03:00
|
|
|
for _, osd := range osddevices {
|
|
|
|
wg.Add(1)
|
|
|
|
if params.parallel == true {
|
2019-02-27 21:12:08 +03:00
|
|
|
go bench(cephconn, osd, &buffs, &startbuff, ¶ms, &wg, results)
|
2019-02-26 13:20:51 +03:00
|
|
|
} else {
|
2019-02-27 21:12:08 +03:00
|
|
|
bench(cephconn, osd, &buffs, &startbuff, ¶ms, &wg, results)
|
2019-02-27 13:46:44 +03:00
|
|
|
log.Println(<-results)
|
2019-02-26 13:20:51 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
if params.parallel == true {
|
|
|
|
go func() {
|
|
|
|
wg.Wait()
|
|
|
|
close(results)
|
|
|
|
}()
|
|
|
|
|
|
|
|
for message := range results {
|
2019-02-27 15:49:36 +03:00
|
|
|
log.Println(message)
|
2019-02-27 13:46:44 +03:00
|
|
|
|
2019-02-22 17:58:31 +03:00
|
|
|
}
|
|
|
|
}
|
2019-02-15 20:04:32 +03:00
|
|
|
|
2019-02-14 15:01:33 +03:00
|
|
|
}
|