Clean up. Pre write objects. Go fmt
parent
4cac3d310f
commit
8a10c8a255
|
@ -13,7 +13,7 @@ import (
|
||||||
"time"
|
"time"
|
||||||
)
|
)
|
||||||
|
|
||||||
func bench(cephconn *Cephconnection, osddevice Device, buffs *[][]byte, bs int64, objsize int64, params *Params,
|
func bench(cephconn *Cephconnection, osddevice Device, buffs *[][]byte, startbuf *[]byte, params *Params,
|
||||||
wg *sync.WaitGroup, result chan string) {
|
wg *sync.WaitGroup, result chan string) {
|
||||||
defer wg.Done()
|
defer wg.Done()
|
||||||
threadresult := make(chan []time.Duration, params.threadsCount)
|
threadresult := make(chan []time.Duration, params.threadsCount)
|
||||||
|
@ -29,10 +29,13 @@ func bench(cephconn *Cephconnection, osddevice Device, buffs *[][]byte, bs int64
|
||||||
name := "bench_" + strconv.Itoa(suffix)
|
name := "bench_" + strconv.Itoa(suffix)
|
||||||
if osddevice.ID == GetObjActingPrimary(cephconn, *params, name) {
|
if osddevice.ID == GetObjActingPrimary(cephconn, *params, name) {
|
||||||
objectnames = append(objectnames, name)
|
objectnames = append(objectnames, name)
|
||||||
|
if err := cephconn.ioctx.WriteFull(name, *startbuf); err != nil {
|
||||||
|
log.Printf("Can't write object: %v, osd: %v", name, osddevice.Name)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
for i := 0; i < int(params.threadsCount); i++ {
|
for i := 0; i < int(params.threadsCount); i++ {
|
||||||
go bench_thread(cephconn, osddevice, (*buffs)[i*2:i*2+2], bs, objsize, params, threadresult, objectnames[i*16:i*16+16])
|
go BenchThread(cephconn, osddevice, (*buffs)[i*2:i*2+2], params, threadresult, objectnames[i*16:i*16+16])
|
||||||
}
|
}
|
||||||
for i := int64(0); i < params.threadsCount; i++ {
|
for i := int64(0); i < params.threadsCount; i++ {
|
||||||
for _, lat := range <-threadresult {
|
for _, lat := range <-threadresult {
|
||||||
|
@ -43,19 +46,19 @@ func bench(cephconn *Cephconnection, osddevice Device, buffs *[][]byte, bs int64
|
||||||
latencygrade := map[int64]int{}
|
latencygrade := map[int64]int{}
|
||||||
latencytotal := int64(0)
|
latencytotal := int64(0)
|
||||||
for _, lat := range osdlatencies {
|
for _, lat := range osdlatencies {
|
||||||
micro := lat.Nanoseconds()/1000
|
micro := lat.Nanoseconds() / 1000
|
||||||
rounded := micro
|
rounded := micro
|
||||||
switch {
|
switch {
|
||||||
case micro < 1000: // 0-1ms round to 0.1ms
|
case micro < 1000: // 0-1ms round to 0.1ms
|
||||||
rounded = (micro/100)*100
|
rounded = (micro / 100) * 100
|
||||||
case micro < 10000: // 2-10ms round to 1ms
|
case micro < 10000: // 2-10ms round to 1ms
|
||||||
rounded = (micro/1000)*1000
|
rounded = (micro / 1000) * 1000
|
||||||
case micro < 100000: // 10-100ms round to 10ms
|
case micro < 100000: // 10-100ms round to 10ms
|
||||||
rounded = (micro/10000)*10000
|
rounded = (micro / 10000) * 10000
|
||||||
case micro < 1000000: // 100-1000ms round to 100ms
|
case micro < 1000000: // 100-1000ms round to 100ms
|
||||||
rounded = (micro/100000)*100000
|
rounded = (micro / 100000) * 100000
|
||||||
default: // 1000+ms round to 1s
|
default: // 1000+ms round to 1s
|
||||||
rounded = (micro/1000000)*1000000
|
rounded = (micro / 1000000) * 1000000
|
||||||
}
|
}
|
||||||
latencytotal += micro
|
latencytotal += micro
|
||||||
latencygrade[rounded]++
|
latencygrade[rounded]++
|
||||||
|
@ -91,7 +94,7 @@ func bench(cephconn *Cephconnection, osddevice Device, buffs *[][]byte, bs int64
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
latencytotal = latencytotal/int64(len(osdlatencies))
|
latencytotal = latencytotal / int64(len(osdlatencies))
|
||||||
// iops = 1s / latency
|
// iops = 1s / latency
|
||||||
iops := 1000000 / latencytotal
|
iops := 1000000 / latencytotal
|
||||||
// avg speed = iops * block size / 1 MB
|
// avg speed = iops * block size / 1 MB
|
||||||
|
@ -123,7 +126,7 @@ func bench(cephconn *Cephconnection, osddevice Device, buffs *[][]byte, bs int64
|
||||||
case k < 2000:
|
case k < 2000:
|
||||||
mseconds = yellow(fmt.Sprintf("[%.1f-%.1f)", float64(k)/1000, 0.1+float64(k)/1000))
|
mseconds = yellow(fmt.Sprintf("[%.1f-%.1f)", float64(k)/1000, 0.1+float64(k)/1000))
|
||||||
case k < 10000:
|
case k < 10000:
|
||||||
mseconds = yellow(fmt.Sprintf("[%3v-%3v)", k/1000, 1+k/1000))
|
mseconds = yellow(fmt.Sprintf("[%.1f-%.1f)", float64(k/1000), float64(1+k/1000)))
|
||||||
case k < 100000:
|
case k < 100000:
|
||||||
mseconds = red(fmt.Sprintf("[%3v-%3v)", k/1000, 10+k/1000))
|
mseconds = red(fmt.Sprintf("[%3v-%3v)", k/1000, 10+k/1000))
|
||||||
case k < 1000000:
|
case k < 1000000:
|
||||||
|
@ -141,7 +144,7 @@ func bench(cephconn *Cephconnection, osddevice Device, buffs *[][]byte, bs int64
|
||||||
result <- buffer.String()
|
result <- buffer.String()
|
||||||
}
|
}
|
||||||
|
|
||||||
func bench_thread(cephconn *Cephconnection, osddevice Device, buffs [][]byte, bs int64, objsize int64, params *Params,
|
func BenchThread(cephconn *Cephconnection, osddevice Device, buffs [][]byte, params *Params,
|
||||||
result chan []time.Duration, objnames []string) {
|
result chan []time.Duration, objnames []string) {
|
||||||
|
|
||||||
starttime := time.Now()
|
starttime := time.Now()
|
||||||
|
@ -149,7 +152,7 @@ func bench_thread(cephconn *Cephconnection, osddevice Device, buffs [][]byte, bs
|
||||||
endtime := starttime.Add(params.duration)
|
endtime := starttime.Add(params.duration)
|
||||||
n := 0
|
n := 0
|
||||||
for {
|
for {
|
||||||
offset := rand.Int63n(objsize/bs) * bs
|
offset := rand.Int63n(params.objectsize/params.blocksize) * params.blocksize
|
||||||
objname := objnames[rand.Int31n(int32(len(objnames)))]
|
objname := objnames[rand.Int31n(int32(len(objnames)))]
|
||||||
startwritetime := time.Now()
|
startwritetime := time.Now()
|
||||||
if startwritetime.After(endtime) {
|
if startwritetime.After(endtime) {
|
||||||
|
@ -158,7 +161,7 @@ func bench_thread(cephconn *Cephconnection, osddevice Device, buffs [][]byte, bs
|
||||||
err := cephconn.ioctx.Write(objname, buffs[n], uint64(offset))
|
err := cephconn.ioctx.Write(objname, buffs[n], uint64(offset))
|
||||||
endwritetime := time.Now()
|
endwritetime := time.Now()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.Printf("Can't write obj: %v, osd: %v", objname, osddevice.Name)
|
log.Printf("Can't write object: %v, osd: %v", objname, osddevice.Name)
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
latencies = append(latencies, endwritetime.Sub(startwritetime))
|
latencies = append(latencies, endwritetime.Sub(startwritetime))
|
||||||
|
@ -183,6 +186,8 @@ func main() {
|
||||||
for i := int64(0); i < 2*params.threadsCount; i++ {
|
for i := int64(0); i < 2*params.threadsCount; i++ {
|
||||||
buffs = append(buffs, make([]byte, params.blocksize))
|
buffs = append(buffs, make([]byte, params.blocksize))
|
||||||
}
|
}
|
||||||
|
startbuff := make([]byte, params.objectsize)
|
||||||
|
rand.Read(startbuff)
|
||||||
for num := range buffs {
|
for num := range buffs {
|
||||||
_, err := rand.Read(buffs[num])
|
_, err := rand.Read(buffs[num])
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
@ -196,9 +201,9 @@ func main() {
|
||||||
for _, osd := range osddevices {
|
for _, osd := range osddevices {
|
||||||
wg.Add(1)
|
wg.Add(1)
|
||||||
if params.parallel == true {
|
if params.parallel == true {
|
||||||
go bench(cephconn, osd, &buffs, params.blocksize, params.objectsize, ¶ms, &wg, results)
|
go bench(cephconn, osd, &buffs, &startbuff, ¶ms, &wg, results)
|
||||||
} else {
|
} else {
|
||||||
bench(cephconn, osd, &buffs, params.blocksize, params.objectsize, ¶ms, &wg, results)
|
bench(cephconn, osd, &buffs, &startbuff, ¶ms, &wg, results)
|
||||||
log.Println(<-results)
|
log.Println(<-results)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue