Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add results of fio benchmark #9

Merged
merged 1 commit into from
Oct 30, 2024
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
61 changes: 61 additions & 0 deletions results/ocp_emptydir_fio_3G-1800_202410xx/fs-performance.yml
Original file line number Diff line number Diff line change
@@ -0,0 +1,61 @@
---
# This is a job file for kubernetes to run the fs-performance tests
#
# Usage:
# - Modify the PVC to obtain the type of storage you want to test.
# This includes adjusting the accessModes and storageClass.
# - Apply this file: kubectl apply -f fs-performance.yml
# - When the job is done, read the pod log for the results.
# - Clean up: kubctl delete -f fs-performance.yml

# kind: PersistentVolumeClaim
# apiVersion: v1
# metadata:
# name: fs-perf-target
# spec:
# # To test a particular type of storage, set the name of the StorageClass here.
# # storageClassName: gp2
# accessModes: ["ReadWriteOnce"]
# resources:
# requests:
# storage: 400Gi

---

apiVersion: batch/v1
kind: Job
metadata:
name: fs-performance
spec:
template:
metadata:
name: fs-performance
spec:
containers:
- name: fs-performance
image: quay.io/johnstrunk/fs-performance:latest
env:
# TARGET_PATH must match the path for the volumeMount, below.
- name: BENCHMARKS
value: fio
- name: FIO_CAPACITY_MB
value: "3000"
- name: FIO_RUNTIME
value: "1800"
- name: ITERATIONS
value: "1"
- name: TARGET_PATH
value: "/local"
volumeMounts:
# - name: target
# mountPath: /target
- name: local
mountPath: /local
restartPolicy: Never
volumes:
- name: local
emptyDir:
sizeLimit: 4Gi
# - name: target
# persistentVolumeClaim:
# claimName: fs-perf-target
17 changes: 17 additions & 0 deletions results/ocp_emptydir_fio_3G-1800_202410xx/logs_3G_1800.txt
Original file line number Diff line number Diff line change
@@ -0,0 +1,17 @@
Configuration:
List of benchmarks to run: fio (BENCHMARKS)
Target path for tests: /local (TARGET_PATH)
Number of test iterations to run: 1 (ITERATIONS)
Random startup delay (s): 0 (STARTUP_DELAY)
Random delay between iterations (s): 0 (RAND_THINK)
Delete contents of target dir on startup: 0 (DELETE_FIRST)
File size for fio benchmark: 3000 (FIO_CAPACITY_MB)
Runtime for individual fio tests (s): 1800 (FIO_RUNTIME)
Git repo to use for clone test: https://github.com/eclipse/che (CLONE_REPO)
Benchmark: fio
Max write bandwidth: 282 MiB/s
Max read bandwidth: 530 MiB/s
Write I/O latency: 0.034 ms (50%=0.031, 90%=0.034, 95%=0.038, 99%=0.045)
Read I/O latency: 0.111 ms (50%=0.113, 90%=0.129, 95%=0.134, 99%=0.151)
Max write throughput: 62305 IOPS
Max read throughput: 93520 IOPS
56 changes: 56 additions & 0 deletions results/ocp_nese_fio_100M-1800-x10_20241018/fs-performance.yml
Original file line number Diff line number Diff line change
@@ -0,0 +1,56 @@
---
# This is a job file for kubernetes to run the fs-performance tests
#
# Usage:
# - Modify the PVC to obtain the type of storage you want to test.
# This includes adjusting the accessModes and storageClass.
# - Apply this file: kubectl apply -f fs-performance.yml
# - When the job is done, read the pod log for the results.
# - Clean up: kubctl delete -f fs-performance.yml

kind: PersistentVolumeClaim
apiVersion: v1
metadata:
name: fs-perf-target
spec:
# To test a particular type of storage, set the name of the StorageClass here.
# storageClassName: gp2
accessModes: ["ReadWriteOnce"]
resources:
requests:
storage: 900Gi

---

apiVersion: batch/v1
kind: Job
metadata:
name: fs-performance
spec:
template:
metadata:
name: fs-performance
spec:
containers:
- name: fs-performance
image: quay.io/johnstrunk/fs-performance:latest
env:
# TARGET_PATH must match the path for the volumeMount, below.
- name: BENCHMARKS
value: fio
- name: FIO_CAPACITY_MB
value: "100"
- name: FIO_RUNTIME
value: "1800"
- name: ITERATIONS
value: "10"
- name: TARGET_PATH
value: "/target"
volumeMounts:
- name: target
mountPath: /target
restartPolicy: Never
volumes:
- name: target
persistentVolumeClaim:
claimName: fs-perf-target
80 changes: 80 additions & 0 deletions results/ocp_nese_fio_100M-1800-x10_20241018/logs_100M_1800_x10.txt
Original file line number Diff line number Diff line change
@@ -0,0 +1,80 @@
Configuration:
List of benchmarks to run: fio (BENCHMARKS)
Target path for tests: /target (TARGET_PATH)
Number of test iterations to run: 10 (ITERATIONS)
Random startup delay (s): 0 (STARTUP_DELAY)
Random delay between iterations (s): 0 (RAND_THINK)
Delete contents of target dir on startup: 0 (DELETE_FIRST)
File size for fio benchmark: 100 (FIO_CAPACITY_MB)
Runtime for individual fio tests (s): 1800 (FIO_RUNTIME)
Git repo to use for clone test: https://github.com/eclipse/che (CLONE_REPO)
Benchmark: fio
Max write bandwidth: 152 MiB/s
Max read bandwidth: 2133 MiB/s
Write I/O latency: 39.005 ms (50%=5.603, 90%=117.965, 95%=191.889, 99%=383.779)
Read I/O latency: 0.568 ms (50%=0.387, 90%=0.578, 95%=0.627, 99%=1.466)
Max write throughput: 541 IOPS
Max read throughput: 32186 IOPS
Benchmark: fio
Max write bandwidth: 260 MiB/s
Max read bandwidth: 2035 MiB/s
Write I/O latency: 24.175 ms (50%=3.129, 90%=77.07, 95%=137.363, 99%=283.116)
Read I/O latency: 0.597 ms (50%=0.375, 90%=0.561, 95%=0.618, 99%=5.538)
Max write throughput: 621 IOPS
Max read throughput: 36637 IOPS
Benchmark: fio
Max write bandwidth: 179 MiB/s
Max read bandwidth: 2143 MiB/s
Write I/O latency: 36.6 ms (50%=5.014, 90%=111.673, 95%=185.598, 99%=371.196)
Read I/O latency: 1.11 ms (50%=0.391, 90%=0.643, 95%=3.097, 99%=19.792)
Max write throughput: 519 IOPS
Max read throughput: 18951 IOPS
Benchmark: fio
Max write bandwidth: 156 MiB/s
Max read bandwidth: 2172 MiB/s
Write I/O latency: 42.281 ms (50%=6.849, 90%=104.333, 95%=179.306, 99%=513.802)
Read I/O latency: 0.585 ms (50%=0.399, 90%=0.578, 95%=0.635, 99%=3.097)
Max write throughput: 627 IOPS
Max read throughput: 69947 IOPS
Benchmark: fio
Max write bandwidth: 182 MiB/s
Max read bandwidth: 2010 MiB/s
Write I/O latency: 36.366 ms (50%=3.457, 90%=121.111, 95%=196.084, 99%=354.419)
Read I/O latency: 0.802 ms (50%=0.412, 90%=0.602, 95%=0.692, 99%=12.78)
Max write throughput: 435 IOPS
Max read throughput: 53536 IOPS
Benchmark: fio
Max write bandwidth: 240 MiB/s
Max read bandwidth: 1513 MiB/s
Write I/O latency: 46.898 ms (50%=4.293, 90%=158.335, 95%=238.027, 99%=425.722)
Read I/O latency: 1.003 ms (50%=0.42, 90%=0.627, 95%=1.057, 99%=17.433)
Max write throughput: 600 IOPS
Max read throughput: 29808 IOPS
Benchmark: fio
Max write bandwidth: 204 MiB/s
Max read bandwidth: 2115 MiB/s
Write I/O latency: 36.413 ms (50%=5.276, 90%=113.77, 95%=177.209, 99%=312.476)
Read I/O latency: 1.175 ms (50%=0.395, 90%=0.651, 95%=3.621, 99%=21.103)
Max write throughput: 515 IOPS
Max read throughput: 36933 IOPS
Benchmark: fio
Max write bandwidth: 170 MiB/s
Max read bandwidth: 1796 MiB/s
Write I/O latency: 36.145 ms (50%=7.569, 90%=106.43, 95%=166.724, 99%=325.059)
Read I/O latency: 0.579 ms (50%=0.362, 90%=0.586, 95%=0.643, 99%=4.145)
Max write throughput: 390 IOPS
Max read throughput: 45681 IOPS
Benchmark: fio
Max write bandwidth: 216 MiB/s
Max read bandwidth: 1894 MiB/s
Write I/O latency: 35.45 ms (50%=4.227, 90%=109.576, 95%=170.918, 99%=320.864)
Read I/O latency: 1.764 ms (50%=0.453, 90%=1.09, 95%=9.765, 99%=28.443)
Max write throughput: 555 IOPS
Max read throughput: 31513 IOPS
Benchmark: fio
Max write bandwidth: 190 MiB/s
Max read bandwidth: 1785 MiB/s
Write I/O latency: 40.727 ms (50%=5.276, 90%=131.596, 95%=202.375, 99%=367.002)
Read I/O latency: 1.408 ms (50%=0.428, 90%=0.733, 95%=6.652, 99%=22.938)
Max write throughput: 418 IOPS
Max read throughput: 29350 IOPS
56 changes: 56 additions & 0 deletions results/ocp_nese_fio_300G-1800_202410xx/fs-performance.yml
Original file line number Diff line number Diff line change
@@ -0,0 +1,56 @@
---
# This is a job file for kubernetes to run the fs-performance tests
#
# Usage:
# - Modify the PVC to obtain the type of storage you want to test.
# This includes adjusting the accessModes and storageClass.
# - Apply this file: kubectl apply -f fs-performance.yml
# - When the job is done, read the pod log for the results.
# - Clean up: kubctl delete -f fs-performance.yml

kind: PersistentVolumeClaim
apiVersion: v1
metadata:
name: fs-perf-target
spec:
# To test a particular type of storage, set the name of the StorageClass here.
# storageClassName: gp2
accessModes: ["ReadWriteOnce"]
resources:
requests:
storage: 400Gi

---

apiVersion: batch/v1
kind: Job
metadata:
name: fs-performance
spec:
template:
metadata:
name: fs-performance
spec:
containers:
- name: fs-performance
image: quay.io/johnstrunk/fs-performance:latest
env:
# TARGET_PATH must match the path for the volumeMount, below.
- name: BENCHMARKS
value: fio
- name: FIO_CAPACITY_MB
value: "300000"
- name: FIO_RUNTIME
value: "1800"
- name: ITERATIONS
value: "1"
- name: TARGET_PATH
value: "/target"
volumeMounts:
- name: target
mountPath: /target
restartPolicy: Never
volumes:
- name: target
persistentVolumeClaim:
claimName: fs-perf-target
17 changes: 17 additions & 0 deletions results/ocp_nese_fio_300G-1800_202410xx/logs_300G_1800.txt
Original file line number Diff line number Diff line change
@@ -0,0 +1,17 @@
Configuration:
List of benchmarks to run: fio (BENCHMARKS)
Target path for tests: /target (TARGET_PATH)
Number of test iterations to run: 1 (ITERATIONS)
Random startup delay (s): 0 (STARTUP_DELAY)
Random delay between iterations (s): 0 (RAND_THINK)
Delete contents of target dir on startup: 0 (DELETE_FIRST)
File size for fio benchmark: 300000 (FIO_CAPACITY_MB)
Runtime for individual fio tests (s): 1800 (FIO_RUNTIME)
Git repo to use for clone test: https://github.com/eclipse/che (CLONE_REPO)
Benchmark: fio
Max write bandwidth: 219 MiB/s
Max read bandwidth: 690 MiB/s
Write I/O latency: 38.702 ms (50%=5.21, 90%=121.111, 95%=183.501, 99%=337.641)
Read I/O latency: 17.62 ms (50%=13.435, 90%=27.656, 95%=43.778, 99%=109.576)
Max write throughput: 819 IOPS
Max read throughput: 1929 IOPS
56 changes: 56 additions & 0 deletions results/ocp_nese_fio_600G-1800-x10_20241017/fs-performance.yml
Original file line number Diff line number Diff line change
@@ -0,0 +1,56 @@
---
# This is a job file for kubernetes to run the fs-performance tests
#
# Usage:
# - Modify the PVC to obtain the type of storage you want to test.
# This includes adjusting the accessModes and storageClass.
# - Apply this file: kubectl apply -f fs-performance.yml
# - When the job is done, read the pod log for the results.
# - Clean up: kubctl delete -f fs-performance.yml

kind: PersistentVolumeClaim
apiVersion: v1
metadata:
name: fs-perf-target
spec:
# To test a particular type of storage, set the name of the StorageClass here.
# storageClassName: gp2
accessModes: ["ReadWriteOnce"]
resources:
requests:
storage: 900Gi

---

apiVersion: batch/v1
kind: Job
metadata:
name: fs-performance
spec:
template:
metadata:
name: fs-performance
spec:
containers:
- name: fs-performance
image: quay.io/johnstrunk/fs-performance:latest
env:
# TARGET_PATH must match the path for the volumeMount, below.
- name: BENCHMARKS
value: fio
- name: FIO_CAPACITY_MB
value: "600"
- name: FIO_RUNTIME
value: "1800"
- name: ITERATIONS
value: "10"
- name: TARGET_PATH
value: "/target"
volumeMounts:
- name: target
mountPath: /target
restartPolicy: Never
volumes:
- name: target
persistentVolumeClaim:
claimName: fs-perf-target
Loading