Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

...

TestConditionData InputComments
UnitTest

src/vos/vea/tests/vea_ut

src/eio/smd/tests/smd_ut

Look at the DAOS-1246 and include all unit tests for NVMe

daos_epoch_discard() will also create the more fragmentation. So add unit test specific to that API call.


Exiting Functional Test
Identify and run existing functional test cases with NVMeManual Test is in Progress, once it's done will update the test case information here.
I/OCreate the Pool with small NVMe size (0/1G/48G)

Write/Read data (<4K) 1B/1K/4K [Random/Sequential] and make sure it does not use NVMe

Write/Read data with non standard sizes 4025,259K,1.1M,30.22M [Random/Sequential] and make sure it does not use NVMe


For second case, data sizes can be generated random instead of predefined fixed size.

Write/Read data (>4K) 1M/16G [Random/Sequential] and make sure it use NVMe

Write/Read data with non standard sizes 4025,259K,1.1M,30.22M [Random/Sequential] and make sure it does not use NVMe

For second case, data sizes can be generated random instead of predefined fixed size.
Create the Pool with Large NVMe size (1TB/2TB)

Write/Read data (<4K) 1B/1K/4K [Random/Sequential] and make sure it does not use NVMe

Write/Read data with non standard sizes 4025,259K,1.1M,30.22M [Random/Sequential] and make sure it does not use NVMe

For second case, data sizes can be generated random instead of predefined fixed size.

Write/Read data (>4K) 1M/16G/1TB [Random/Sequential] and make sure it use NVMe

Write/Read data with non standard sizes 4025,259K,1.1M,30.22M [Random/Sequential] and make sure it does not use NVMe

For second case, data sizes can be generated random instead of predefined fixed size.
Unaligned IO

Try using the offset from API or use the core Python API to modify the existing Array and read through.

The test code daos_run_io_conf.c will be doing similar thing so worth to use same.

I/O with Server/System restartCreate the Pool with small NVMe size (0/1G/48G)Write/Read data (<4K) 1B/1K/4KWrite the data/ Stop Server/Start Server/Read data back and check data integrity.

Write/Read data (>4K) 1M/16G Write the data/ Reboot the node/ Start server/read data back and check data integrity.
Create the Pool with Large NVMe size (1TB/2TB)Write/Read data (<4K) 1B/1K/4KWrite the data/ Stop Server/Start Server/Read data back and check data integrity.

Write/Read data (>4K) 1M/16G Write the data/ Reboot the node/ Start server/read data back and check data integrity.
Create the Pool with Large NVMe size (1TB/2TB)Write single IOR data set, Read single IOR data setKill the server while IO is doing Write and start the server. IO should continue after server start? Do the same when read is in progress
Create the Pool with Large NVMe size (1TB/2TB)Write multiple IOR data-sets, Read multiple IOR data-sets, Read-Write TogetherKill the server while IO is doing multiple write and start the server. IO should continue after server start? Do the same when Multiple Read is in progress.

PMDK may have some API to get the allocated but it wont be easy to use as it is. There is Pool Query which will provide in future which can be track the size to verify memory leak.

We may need fault injection layer to control the PMDK transaction with server start/stop. 

Example:

1> Create 3G pool

2> Write 1G data and before it goes for transaction commit shutdown the server.

3> Next server start it should claim the full 3G space 


Re-written data fetch validation

Write the data on NVMe >4K

Re-write using the same array with small size ~1-2 bytes which will go through SCM.

Do this and change ~100 bytes with the different data

Do fetch which will combined the record and verify it.

Do the similar thing like writing small data set to SCM and overwrite large data to NVMe, and validate the content.

When overwriting the data will be kind of new epoch entry getting created. But it will use the old data set and update the new bytes value only.

During fetch Epoch will aggregated and provide the result with modified bytes.


Re-written data fetch validation

Write the data on SCM <4K

Extend the data set using the same array with larger size >8K which will go through NVMe

Do this and repeat for few times with the different data

Fetch the dataset which will combined the records, Verify all the old+new records.

Verify there is no leak after PMDK Transaction (This can not be tested until we get some hooks to control PMDK transaction)

PMDK may have some API to get the allocated but it wont be easy to use as it is. There is Pool Query which will provide in future which can be track the size to verify memory leak.

Example:

1> Create 3G pool

2> Write 1G data and before it goes for transaction commit shutdown the server.

3> Next when server start verify it should claim the full 3G space.

We need fault injection layer to control the PMDK transaction with server start/stop. 

Large number of Pool with Server/System restartCreate the large number of pools (10000) with different NVMe sizesWrite mixed data across all the pools (1K/4K/1M/1T)Write the data/ Stop Server/Start Server/Read data back and check data integrity. 
Write the data/ Reboot the node/ Start server/read data back and check data integrity.
Pool CapacityCreate the NVMe pool size 1GBWrite IO > 1GB which should failed with ENOM SPACE
Create pool same size as NVMe drive
Write IO till pool is getting filled up, once the Drive is full it should not allow to write more data with ENOM_SPACE
Create the pool with maximum of NVMe size and delete.
Run this in loop for example if NVMe is 2TB, create the pool size of 1TB, 500GB, 500GB, delete all the pools. Do this in a loop and make sure pool creation work and size can be reclaimed.
Verify the Fragmentation.

Create the pool with 12GB.

Add 8GB size of data on NVMe and 4G on SCM. IO has to be written with different container (~100) with different sizes.

use more smaller sizes for NVMe (4K-32K) and for SCM (8Bytes-4K) to have more fragmentation.



Fragmentation information will be needed from pool query in future to validate some test area. 

Pool ExtendExtend the single pool to multiple targetsCreate the few data set on single pool (1K/4K/1M/1T). Extend the pool to all target at once.Verify the data integrity after pool extension done

Create the few data set with single pools (1K/4K/1M/1T). Extend the pool target one by one, for example 6 server so pool created with 2 and start extending the pool to 4 servers one by oneVerify the data integrity after pool extension done
Extend the multiple pools to targetsCreate the few data set on different pools (1K/4K/1M/1T). Extend the pools to all target at once.Verify the data integrity after pool extension done

Create the few data set with single pools (1K/4K/1M/1T). Extend the pool target one by one, for example 6 server so pool created with 2 and start extending the pool to 4 servers one by oneVerify the data integrity after pool extension done
Pool ExcludeExclude the target from PoolCreate the few data set on different pools (1K/4K/1M/1T). Exclude the pools from all target one by one.Add target to pool and verify the data integrity after pool excluded
NVMe rebuildSingle drive RebuildUse 4 server minimum and load 50% of Drives. Shutdown the single server or Eject and make sure the data is getting rebuilt on another NVMe drive
ObjectCreate the large number of object update/fetch with different Object ID in single pool created on NVMe
Verify the objects are getting created and data are not corrupted
Create the large number of object in multiple pool created on NVMe (Pools size 1M/1G/1T)
Verify the objects are getting created and data are not corrupted
PerformanceCompare the DAOS performanceRun performance utility (TBD) without DAOS and IOR with DAOS.Performance

SPDK can be used to measure the raw performance of the drive.

daos_perf for measuring VOS performance and DAOS performance with below mode.

daos_perf VOS raw performance

daos_perf Echo mode

daos_perf DAOS mode.

Sizes To be covered for Performance: 4K,16K,1M,8M,32M.

Performance measurement for Read/Write and IOPS.

It's better to get all this different number in single graph for different sizes for better comparison.

daos_perf testRun daos_perf test with VOS and with DAOS
Verify performance Degradation over time.

Create pool, update to fill the pool with many container, Destroy the container. Different IOR runs with sizes can be run in parallel.

Run single IOR job to validate the performance number in-between.

This need to be done for hours/days (Fill the container delete the container).

Measure the performance throughout and at the end to validate it's not drop at any point.

We need to measure performance degradation over time. This will be a good exercise to verify whether our TRIM support is efficient (require 3DNAND)


With more Fragmentation

1> Create the small data set with Epoch1, 2, 3

2> Discard 3rd Epoch

3> Run single IOR to verify the performance.

Do the above steps for ~10000 times. Make sure at the end the discarded fragmentation is not creating the issue for new write and even performing up to the level for same IOR run

Use smaller sizes for NVMe (4K-32K) and for SCM (8Bytes-4K) to have more fragmentation.

MetadataCreate the pool of small size (NVMe size 1G)Run Mdtest to fill the Pool with metadataAfter fool it should not allow any data to be written on pool Even if NVMe has the space ? 
Control Plane/Management for NVMe
  • NVMe SSD discovery with "discover" bindings
  • NVMe SSD burn-in with "burnin" bindings
  • NVMe SSD configuration with  "configuration" bindings
TBDTBD
Control Plane/Management for NVMePro-active action based on telemetry data (rebalancing)
  • evicting SSD based on high temperature
  • wear-leveling data
TBDTBD
Control Plane/Management for NVMe
  • SSD firmware image update
TBDTBD