Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

fix(nvmf): fixing duplicate QID error #1616

Merged
merged 1 commit into from
Apr 3, 2024
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
207 changes: 207 additions & 0 deletions io-engine/tests/nexus_rebuild_parallel.rs
Original file line number Diff line number Diff line change
@@ -0,0 +1,207 @@
pub mod common;

use std::time::{Duration, Instant};

use common::{
compose::{
rpc::v1::{nexus::ChildState, GrpcConnect, Status},
Binary,
Builder,
},
nexus::NexusBuilder,
pool::PoolBuilder,
replica::ReplicaBuilder,
};
use tonic::Code;

struct Volume {
replicas: Vec<ReplicaBuilder>,
nex: NexusBuilder,
}

/// Runs multiple rebuild jobs parallel.
#[tokio::test]
async fn nexus_rebuild_parallel() {
common::composer_init();

const R: usize = 3; // Number of replicas / nodes.
const N: usize = 20; // Number f volumes
const REPL_SIZE: u64 = 50;
const NEXUS_SIZE: u64 = REPL_SIZE;
const POOL_SIZE: u64 = REPL_SIZE * N as u64 + REPL_SIZE;
const DISK_SIZE: u64 = POOL_SIZE * 1024 * 1024;
const DISK_NAME: &str = "disk";

// Create pool data file.
for r in 0 .. R {
let name = format!("/tmp/{DISK_NAME}_{r}");
common::delete_file(&[name.clone()]);
common::truncate_file_bytes(&name, DISK_SIZE);
}

let test = Builder::new()
.name("cargo-test")
.network("10.1.0.0/16")
.unwrap()
.add_container_bin(
"ms_0",
Binary::from_dbg("io-engine")
.with_args(vec!["-l", "1,2", "-Fcolor,compact,nodate,host"])
.with_bind("/tmp", "/host/tmp"),
)
.add_container_bin(
"ms_1",
Binary::from_dbg("io-engine")
.with_args(vec!["-l", "3,4", "-Fcolor,compact,nodate,host"])
.with_bind("/tmp", "/host/tmp"),
)
.add_container_bin(
"ms_2",
Binary::from_dbg("io-engine")
.with_args(vec!["-l", "5,6", "-Fcolor,compact,nodate,host"])
.with_bind("/tmp", "/host/tmp"),
)
.with_clean(true)
.build()
.await
.unwrap();

let conn = GrpcConnect::new(&test);

let ms_nex = conn.grpc_handle_shared("ms_2").await.unwrap();

let ms = [
conn.grpc_handle_shared("ms_0").await.unwrap(),
conn.grpc_handle_shared("ms_1").await.unwrap(),
conn.grpc_handle_shared("ms_2").await.unwrap(),
];

let mut pools = Vec::new();

// Create pools.
for (r, ms) in ms.iter().enumerate() {
let bdev = format!("aio:///host/tmp/{DISK_NAME}_{r}?blk_size=512");
let mut pool = PoolBuilder::new(ms.clone())
.with_name(&format!("p{r}"))
.with_new_uuid()
.with_bdev(&bdev);
pool.create().await.unwrap();
pools.push(pool);
}

let mut vols = Vec::new();

for i in 0 .. N {
// Create R replicas on the pools.
let mut replicas = Vec::new();
for r in 0 .. R {
let mut repl = ReplicaBuilder::new(ms[r].clone())
.with_pool(&pools[r])
.with_name(&format!("v{i}r{r}"))
.with_new_uuid()
.with_size_mb(REPL_SIZE)
.with_thin(false);
repl.create().await.unwrap();
repl.share().await.unwrap();
replicas.push(repl);
}

// Create the nexus with 2 replicas @ 1,2.
let mut nex = NexusBuilder::new(ms_nex.clone())
.with_name(&format!("v{i}"))
.with_new_uuid()
.with_size_mb(NEXUS_SIZE)
.with_replica(&replicas[1])
.with_replica(&replicas[2]);

nex.create().await.unwrap();

vols.push(Volume {
replicas,
nex,
});
}

// Adding replicas / starting rebuilds.
for vol in &vols {
vol.nex.add_replica(&vol.replicas[0], false).await.unwrap();
}

monitor_volumes(&vols, Duration::from_secs(30))
.await
.expect("All volumes must go online");

// Delete test files.
for r in 0 .. R {
let name = format!("/tmp/{DISK_NAME}_{r}");
common::delete_file(&[name.clone()]);
}
}

/// Monitors and prints volume states.
async fn monitor_volumes(
vols: &Vec<Volume>,
timeout: Duration,
) -> Result<(), Status> {
println!("\nMonitoring {n} volumes", n = vols.len());

let start = Instant::now();

loop {
let mut vols_degraded = false;
let mut vols_failed = false;

for vol in vols {
let mut s = String::new();

let n = vol.nex.get_nexus().await.unwrap();
for c in n.children.iter() {
match c.state() {
ChildState::Faulted => {
s = format!("{s} FAILED | ");
vols_failed = true;
}
ChildState::Online => {
s = format!("{s} ONLINE | ");
}
ChildState::Unknown => {
s = format!("{s} UNKNOWN | ");
vols_degraded = true;
}
ChildState::Degraded => {
s = format!(
"{s} REBUILD {p:02} | ",
p = c.rebuild_progress
);
vols_degraded = true;
}
}
}

println!(" {n}: {s}", n = n.name)
}
println!("-");

if vols_failed {
println!("One or more volumes failed");
return Err(Status::new(
Code::Internal,
"One or more volumes failed".to_string(),
));
}

if !vols_degraded {
println!("All volumes are online");
return Ok(());
}

if start.elapsed() > timeout {
return Err(Status::new(
Code::Cancelled,
"Waiting for volumes to go Online state timed out".to_string(),
));
}

tokio::time::sleep(std::time::Duration::from_millis(1000)).await;
}
}
6 changes: 3 additions & 3 deletions nix/pkgs/libspdk/default.nix
Original file line number Diff line number Diff line change
Expand Up @@ -56,13 +56,13 @@ let
# 7. Copy SHA256 from 'got' of the error message to 'sha256' field.
# 8. 'nix-shell' build must now succeed.
drvAttrs = rec {
version = "23.05-c210d86";
version = "23.05-e051030";

src = fetchFromGitHub {
owner = "openebs";
repo = "spdk";
rev = "c210d8650a2fcb44d3a83bb4df2b9bebbb20f833";
sha256 = "sha256-6LJquwT6WOPv/GK9IFkCjOdx17488vTK0Fy1f+rSOJU=";
rev = "e0510300f2c5b005a9ab32f0ab42740ac65b23ca";
sha256 = "sha256-pGeIiCjndKf82BVHIYhn6lXL2fNmKhjX11W7K9GqFvs=";
fetchSubmodules = true;
};

Expand Down
Loading