Skip to content

Commit

Permalink
chore(build): bump toolchain to 2022-06-09 (#3100)
Browse files Browse the repository at this point in the history
Signed-off-by: Alex Chi <iskyzh@gmail.com>
  • Loading branch information
skyzh authored Jun 10, 2022
1 parent 86c7df1 commit 9a1638e
Show file tree
Hide file tree
Showing 23 changed files with 43 additions and 61 deletions.
18 changes: 9 additions & 9 deletions ci/Dockerfile
Original file line number Diff line number Diff line change
Expand Up @@ -6,7 +6,8 @@ ARG RUST_TOOLCHAIN

RUN apt-get update -yy && \
DEBIAN_FRONTEND=noninteractive apt-get -y install make build-essential cmake protobuf-compiler curl \
openssl libssl-dev libcurl4-openssl-dev pkg-config bash openjdk-11-jdk wget unzip git tmux lld -yy
openssl libssl-dev libcurl4-openssl-dev pkg-config bash openjdk-11-jdk wget unzip git tmux lld -yy \
&& rm -rf /var/lib/{apt,dpkg,cache,log}/

SHELL ["/bin/bash", "-c"]

Expand All @@ -21,14 +22,13 @@ WORKDIR /risingwave

ENV PATH /root/.cargo/bin/:$PATH

# install build tools
RUN cargo install cargo-llvm-cov cargo-nextest cargo-udeps cargo-hakari cargo-sort cargo-make

# pre-cache required crates
RUN git clone https://github.com/singularity-data/risingwave && cd risingwave && cargo fetch && cd .. && rm -rf risingwave

# add required rustup components
RUN rustup component add rustfmt llvm-tools-preview clippy

# install sqllogictest
RUN cargo install --git https://github.com/risinglightdb/sqllogictest-rs --features bin
# install build tools
RUN cargo install cargo-llvm-cov cargo-nextest cargo-udeps cargo-hakari cargo-sort cargo-make cargo-cache \
&& cargo install --git https://github.com/risinglightdb/sqllogictest-rs --features bin \
&& cargo cache -a

# pre-cache required crates
RUN git clone https://github.com/singularity-data/risingwave && cd risingwave && cargo fetch && cd .. && rm -rf risingwave && cargo cache -g
3 changes: 2 additions & 1 deletion ci/build-ci-image.sh
Original file line number Diff line number Diff line change
Expand Up @@ -9,8 +9,9 @@ cd "$DIR"
cat ../rust-toolchain
# shellcheck disable=SC2155
export RUST_TOOLCHAIN=$(cat ../rust-toolchain)
export BUILD_ENV_VERSION=v20220609
export BUILD_ENV_VERSION=v20220610
export BUILD_TAG="public.ecr.aws/x5u3w5h6/rw-build-env:${BUILD_ENV_VERSION}"

docker build -t ${BUILD_TAG} --build-arg "RUST_TOOLCHAIN=${RUST_TOOLCHAIN}" .
aws ecr-public get-login-password --region us-east-1 | docker login --username AWS --password-stdin public.ecr.aws/x5u3w5h6
docker push ${BUILD_TAG}
2 changes: 1 addition & 1 deletion ci/docker-compose.yml
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
version: "3.9"
services:
rw-build-env:
image: public.ecr.aws/x5u3w5h6/rw-build-env:v202206091256
image: public.ecr.aws/x5u3w5h6/rw-build-env:v20220610
# build:
# context: ..
# dockerfile: ./ci/Dockerfile
Expand Down
2 changes: 1 addition & 1 deletion ci/scripts/unit-test.sh
Original file line number Diff line number Diff line change
Expand Up @@ -13,7 +13,7 @@ cargo doc --document-private-items --no-deps

echo "--- Run unit tests with coverage"
# use tee to disable progress bar
cargo llvm-cov nextest --lcov --output-path lcov.info --features failpoints -- --no-fail-fast | tee
cargo llvm-cov nextest --lcov --output-path lcov.info --features failpoints -- --no-fail-fast 2> >(tee)

echo "--- Run doctest"
cargo test --doc
Expand Down
4 changes: 2 additions & 2 deletions e2e_test/batch/functions/concat.slt
Original file line number Diff line number Diff line change
Expand Up @@ -9,12 +9,12 @@ ab
query T
select concat(NULL);
----

(empty)

query T
select concat(NULL, NULL);
----

(empty)

query T
select concat(1, 1.01);
Expand Down
2 changes: 1 addition & 1 deletion rust-toolchain
Original file line number Diff line number Diff line change
@@ -1 +1 @@
nightly-2022-05-24
nightly-2022-06-09
6 changes: 2 additions & 4 deletions src/connector/src/kinesis/source/reader.rs
Original file line number Diff line number Diff line change
Expand Up @@ -150,13 +150,11 @@ impl KinesisSplitReader {
async fn get_records(
&mut self,
) -> core::result::Result<GetRecordsOutput, SdkError<GetRecordsError>> {
let resp = self
.client
self.client
.get_records()
.set_shard_iterator(self.shard_iter.take())
.send()
.await;
resp
.await
}
}

Expand Down
2 changes: 1 addition & 1 deletion src/expr/src/vector_op/agg/general_sorted_grouper.rs
Original file line number Diff line number Diff line change
Expand Up @@ -111,7 +111,7 @@ impl EqGroups {
use std::collections::BinaryHeap;
let mut heap = BinaryHeap::new();
for (ci, column) in columns.iter().enumerate() {
if let Some(ri) = column.starting_indices().get(0) {
if let Some(ri) = column.starting_indices().first() {
heap.push(Reverse((ri, ci, 0)));
}
}
Expand Down
2 changes: 1 addition & 1 deletion src/expr/src/vector_op/ascii.rs
Original file line number Diff line number Diff line change
Expand Up @@ -16,7 +16,7 @@ use crate::Result;

#[inline(always)]
pub fn ascii(s: &str) -> Result<i32> {
Ok(s.as_bytes().get(0).map(|x| *x as i32).unwrap_or(0))
Ok(s.as_bytes().first().map(|x| *x as i32).unwrap_or(0))
}

#[cfg(test)]
Expand Down
2 changes: 1 addition & 1 deletion src/frontend/src/binder/expr/mod.rs
Original file line number Diff line number Diff line change
Expand Up @@ -204,7 +204,7 @@ impl Binder {
if return_type.is_numeric() {
return Ok(expr);
}
return Err(ErrorCode::InvalidInputSyntax(format!("+ {:?}", return_type)).into());
Err(ErrorCode::InvalidInputSyntax(format!("+ {:?}", return_type)).into())
}

pub(super) fn bind_trim(
Expand Down
2 changes: 1 addition & 1 deletion src/frontend/src/expr/function_call.rs
Original file line number Diff line number Diff line change
Expand Up @@ -57,7 +57,7 @@ impl std::fmt::Debug for FunctionCall {
ExprType::Cast => {
assert_eq!(self.inputs.len(), 1);
self.inputs[0].fmt(f)?;
return write!(f, "::{:?}", self.return_type);
write!(f, "::{:?}", self.return_type)
}
ExprType::Add => debug_binary_op(f, "+", &self.inputs),
ExprType::Subtract => debug_binary_op(f, "-", &self.inputs),
Expand Down
2 changes: 1 addition & 1 deletion src/frontend/src/scheduler/distributed/stage.rs
Original file line number Diff line number Diff line change
Expand Up @@ -393,7 +393,7 @@ impl StageRunner {
let children = execution_plan_node
.children
.iter()
.map(|e| self.convert_plan_node(&*e, task_id))
.map(|e| self.convert_plan_node(e, task_id))
.collect();

PlanNodeProst {
Expand Down
2 changes: 1 addition & 1 deletion src/frontend/src/scheduler/local.rs
Original file line number Diff line number Diff line change
Expand Up @@ -114,7 +114,7 @@ impl LocalQueryExecution {
let children = execution_plan_node
.children
.iter()
.map(|e| self.convert_plan_node(&*e))
.map(|e| self.convert_plan_node(e))
.collect::<Result<Vec<PlanNodeProst>>>()?;

Ok(PlanNodeProst {
Expand Down
4 changes: 2 additions & 2 deletions src/frontend/test_runner/src/lib.rs
Original file line number Diff line number Diff line change
Expand Up @@ -490,12 +490,12 @@ fn check_err(ctx: &str, expected_err: &Option<String>, actual_err: &Option<Strin
if expected_err == actual_err {
Ok(())
} else {
return Err(anyhow!(
Err(anyhow!(
"Expected {context} error: {}\n Actual {context} error: {}",
expected_err,
actual_err,
context = ctx
));
))
}
}
}
Expand Down
2 changes: 1 addition & 1 deletion src/meta/src/barrier/command.rs
Original file line number Diff line number Diff line change
Expand Up @@ -219,7 +219,7 @@ where
table_fragments,
dispatches,
table_sink_map,
source_state: _source_state,
source_state: _,
} => {
let mut dependent_table_actors = Vec::with_capacity(table_sink_map.len());
for (table_id, actors) in table_sink_map {
Expand Down
4 changes: 2 additions & 2 deletions src/meta/src/barrier/recovery.rs
Original file line number Diff line number Diff line change
Expand Up @@ -116,14 +116,14 @@ where
.expect("Retry until recovery success.");
debug!("recovery success");

return (
(
new_epoch,
self.fragment_manager.all_chain_actor_ids().await,
responses
.into_iter()
.flat_map(|r| r.create_mview_progress)
.collect(),
);
)
}

/// Sync all sources in compute nodes, the local source manager in compute nodes may be dirty
Expand Down
4 changes: 2 additions & 2 deletions src/prost/helpers/src/generate.rs
Original file line number Diff line number Diff line change
Expand Up @@ -124,10 +124,10 @@ pub fn implement(field: &Field) -> TokenStream2 {
}
}

return quote! {
quote! {
#[inline(always)]
pub fn #getter_fn_name(&self) -> &#ty {
&self.#field_name
}
};
}
}
1 change: 0 additions & 1 deletion src/prost/helpers/src/lib.rs
Original file line number Diff line number Diff line change
Expand Up @@ -25,7 +25,6 @@ use proc_macro2::TokenStream as TokenStream2;
use proc_macro_error::{proc_macro_error, ResultExt};
use syn::{DataStruct, DeriveInput};

#[cfg_attr(coverage, no_coverage)]
mod generate;

// This attribute will be placed before any pb types, including messages and enums.
Expand Down
2 changes: 1 addition & 1 deletion src/sqlparser/src/parser.rs
Original file line number Diff line number Diff line change
Expand Up @@ -1356,7 +1356,7 @@ impl Parser {
let all = self.parse_keyword(Keyword::ALL);
let distinct = self.parse_keyword(Keyword::DISTINCT);
if all && distinct {
return parser_err!("Cannot specify both ALL and DISTINCT".to_string());
parser_err!("Cannot specify both ALL and DISTINCT".to_string())
} else {
Ok(distinct)
}
Expand Down
6 changes: 3 additions & 3 deletions src/storage/src/hummock/compactor_tests.rs
Original file line number Diff line number Diff line change
Expand Up @@ -48,15 +48,15 @@ mod tests {
..Default::default()
});
let sstable_store = mock_sstable_store();
let storage = HummockStorage::with_default_stats(

HummockStorage::with_default_stats(
options.clone(),
sstable_store,
hummock_meta_client.clone(),
Arc::new(StateStoreMetrics::unused()),
)
.await
.unwrap();
storage
.unwrap()
}

#[tokio::test]
Expand Down
4 changes: 2 additions & 2 deletions src/storage/src/hummock/iterator/forward_user.rs
Original file line number Diff line number Diff line change
Expand Up @@ -249,8 +249,8 @@ impl UserIterator {
// Handle multi-version
self.last_key.clear();
// Handle range scan when key > end_key
let res = self.next().await;
res

self.next().await
}

/// Indicates whether the iterator can be used.
Expand Down
22 changes: 3 additions & 19 deletions src/storage/src/table/cell_based_table.rs
Original file line number Diff line number Diff line change
Expand Up @@ -64,6 +64,7 @@ pub struct CellBasedTable<S: StateStore> {
column_ids: Vec<ColumnId>,

/// Statistics.
#[allow(dead_code)]
stats: Arc<StateStoreMetrics>,

/// Indices of distribution keys in pk for computing value meta. None if value meta is not
Expand Down Expand Up @@ -337,13 +338,7 @@ impl<S: StateStore> CellBasedTable<S> {

// The returned iterator will iterate data from a snapshot corresponding to the given `epoch`
pub async fn iter(&self, epoch: u64) -> StorageResult<CellBasedTableRowIter<S>> {
CellBasedTableRowIter::new(
&self.keyspace,
self.column_descs.clone(),
epoch,
self.stats.clone(),
)
.await
CellBasedTableRowIter::new(&self.keyspace, self.column_descs.clone(), epoch).await
}

pub async fn iter_with_pk(
Expand All @@ -355,7 +350,6 @@ impl<S: StateStore> CellBasedTable<S> {
self.keyspace.clone(),
self.column_descs.clone(),
epoch,
self.stats.clone(),
pk_descs,
)
.await
Expand Down Expand Up @@ -432,7 +426,6 @@ impl<S: StateStore> CellBasedTable<S> {
self.column_descs.clone(),
(start_key, end_key),
epoch,
self.stats.clone(),
)
.await
}
Expand All @@ -458,7 +451,6 @@ impl<S: StateStore> CellBasedTable<S> {
self.column_descs.clone(),
(start_key, next_key),
epoch,
self.stats.clone(),
)
.await
}
Expand Down Expand Up @@ -488,16 +480,13 @@ pub struct CellBasedTableRowIter<S: StateStore> {
iter: StripPrefixIterator<S::Iter>,
/// Cell-based row deserializer
cell_based_row_deserializer: CellBasedRowDeserializer,
/// Statistics
_stats: Arc<StateStoreMetrics>,
}

impl<S: StateStore> CellBasedTableRowIter<S> {
pub(super) async fn new(
keyspace: &Keyspace<S>,
table_descs: Vec<ColumnDesc>,
epoch: u64,
_stats: Arc<StateStoreMetrics>,
) -> StorageResult<Self> {
keyspace.state_store().wait_epoch(epoch).await?;

Expand All @@ -508,7 +497,6 @@ impl<S: StateStore> CellBasedTableRowIter<S> {
let iter = Self {
iter,
cell_based_row_deserializer,
_stats,
};
Ok(iter)
}
Expand All @@ -518,7 +506,6 @@ impl<S: StateStore> CellBasedTableRowIter<S> {
table_descs: Vec<ColumnDesc>,
serialized_pk_bounds: R,
epoch: u64,
_stats: Arc<StateStoreMetrics>,
) -> StorageResult<Self>
where
R: RangeBounds<B> + Send,
Expand All @@ -534,7 +521,6 @@ impl<S: StateStore> CellBasedTableRowIter<S> {
let iter = Self {
iter,
cell_based_row_deserializer,
_stats,
};
Ok(iter)
}
Expand Down Expand Up @@ -599,11 +585,9 @@ impl<S: StateStore> DedupPkCellBasedTableRowIter<S> {
keyspace: Keyspace<S>,
table_descs: Vec<ColumnDesc>,
epoch: u64,
_stats: Arc<StateStoreMetrics>,
pk_descs: &[OrderedColumnDesc],
) -> StorageResult<Self> {
let inner =
CellBasedTableRowIter::new(&keyspace, table_descs.clone(), epoch, _stats).await?;
let inner = CellBasedTableRowIter::new(&keyspace, table_descs.clone(), epoch).await?;

let (data_types, order_types) = pk_descs
.iter()
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -291,7 +291,8 @@ mod tests {
.map(|(ord, idx)| OrderPair::new(idx, ord))
.collect::<Vec<_>>();
let sort_key_serializer = OrderedArraysSerializer::new(order_pairs);
let managed_state = ManagedStringAggState::new(

ManagedStringAggState::new(
keyspace.clone(),
row_count,
sort_key_indices,
Expand All @@ -300,8 +301,7 @@ mod tests {
sort_key_serializer,
)
.await
.unwrap();
managed_state
.unwrap()
}

#[tokio::test]
Expand Down

0 comments on commit 9a1638e

Please sign in to comment.