Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Avoid RowConverter for multi column grouping (10% faster clickbench queries) #12269

Merged
merged 31 commits into from
Sep 24, 2024
Merged
Show file tree
Hide file tree
Changes from 29 commits
Commits
Show all changes
31 commits
Select commit Hold shift + click to select a range
37d1382
row like group values to avoid rowconverter
jayzhan211 Aug 31, 2024
f40a164
comment out unused
jayzhan211 Aug 31, 2024
9c61a8b
implement to Arrow's builder
jayzhan211 Sep 14, 2024
f20780e
cleanup
jayzhan211 Sep 14, 2024
d3f54ed
switch back to vector
jayzhan211 Sep 15, 2024
6e0b179
clippy
jayzhan211 Sep 15, 2024
10d3d18
optimize for non-null
jayzhan211 Sep 16, 2024
7602a18
use truncate
jayzhan211 Sep 16, 2024
7a4dbd5
cleanup
jayzhan211 Sep 16, 2024
05fb466
cleanup
jayzhan211 Sep 16, 2024
8efcf07
fix first N bug
jayzhan211 Sep 16, 2024
bc16d55
fix null check
jayzhan211 Sep 17, 2024
fb94485
fast path null
jayzhan211 Sep 17, 2024
07ed966
fix bug
jayzhan211 Sep 17, 2024
925166d
fmt
jayzhan211 Sep 17, 2024
f59d11e
fix error
jayzhan211 Sep 17, 2024
e9f9abc
clippy
jayzhan211 Sep 17, 2024
720c343
adjust spill mode max mem
jayzhan211 Sep 18, 2024
b6cd012
Merge branch 'main' of https://github.com/apache/datafusion into row-…
jayzhan211 Sep 18, 2024
5b8aceb
revert test_create_external_table_with_terminator_with_newlines_in_va…
jayzhan211 Sep 18, 2024
ed9b78e
fix null handle bug
jayzhan211 Sep 18, 2024
fb1b745
cleanup
jayzhan211 Sep 18, 2024
2cab4c2
support binary
jayzhan211 Sep 18, 2024
0090008
add binary test
jayzhan211 Sep 18, 2024
e6d1890
use Vec<T> instead of Option<Vec<T>>
jayzhan211 Sep 21, 2024
77efb1a
add test and doc
jayzhan211 Sep 21, 2024
ef274a4
debug assert
jayzhan211 Sep 21, 2024
864bdfa
mv & rename
jayzhan211 Sep 21, 2024
103c45d
fix take_n logic
jayzhan211 Sep 22, 2024
0bd5251
address comment
jayzhan211 Sep 23, 2024
bf5dcc4
cleanup
jayzhan211 Sep 23, 2024
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion benchmarks/queries/clickbench/queries.sql
Original file line number Diff line number Diff line change
Expand Up @@ -40,4 +40,4 @@ SELECT "URL", COUNT(*) AS PageViews FROM hits WHERE "CounterID" = 62 AND "EventD
SELECT "TraficSourceID", "SearchEngineID", "AdvEngineID", CASE WHEN ("SearchEngineID" = 0 AND "AdvEngineID" = 0) THEN "Referer" ELSE '' END AS Src, "URL" AS Dst, COUNT(*) AS PageViews FROM hits WHERE "CounterID" = 62 AND "EventDate"::INT::DATE >= '2013-07-01' AND "EventDate"::INT::DATE <= '2013-07-31' AND "IsRefresh" = 0 GROUP BY "TraficSourceID", "SearchEngineID", "AdvEngineID", Src, Dst ORDER BY PageViews DESC LIMIT 10 OFFSET 1000;
SELECT "URLHash", "EventDate"::INT::DATE, COUNT(*) AS PageViews FROM hits WHERE "CounterID" = 62 AND "EventDate"::INT::DATE >= '2013-07-01' AND "EventDate"::INT::DATE <= '2013-07-31' AND "IsRefresh" = 0 AND "TraficSourceID" IN (-1, 6) AND "RefererHash" = 3594120000172545465 GROUP BY "URLHash", "EventDate"::INT::DATE ORDER BY PageViews DESC LIMIT 10 OFFSET 100;
SELECT "WindowClientWidth", "WindowClientHeight", COUNT(*) AS PageViews FROM hits WHERE "CounterID" = 62 AND "EventDate"::INT::DATE >= '2013-07-01' AND "EventDate"::INT::DATE <= '2013-07-31' AND "IsRefresh" = 0 AND "DontCountHits" = 0 AND "URLHash" = 2868770270353813622 GROUP BY "WindowClientWidth", "WindowClientHeight" ORDER BY PageViews DESC LIMIT 10 OFFSET 10000;
SELECT DATE_TRUNC('minute', to_timestamp_seconds("EventTime")) AS M, COUNT(*) AS PageViews FROM hits WHERE "CounterID" = 62 AND "EventDate"::INT::DATE >= '2013-07-14' AND "EventDate"::INT::DATE <= '2013-07-15' AND "IsRefresh" = 0 AND "DontCountHits" = 0 GROUP BY DATE_TRUNC('minute', to_timestamp_seconds("EventTime")) ORDER BY DATE_TRUNC('minute', M) LIMIT 10 OFFSET 1000;
SELECT DATE_TRUNC('minute', to_timestamp_seconds("EventTime")) AS M, COUNT(*) AS PageViews FROM hits WHERE "CounterID" = 62 AND "EventDate"::INT::DATE >= '2013-07-14' AND "EventDate"::INT::DATE <= '2013-07-15' AND "IsRefresh" = 0 AND "DontCountHits" = 0 GROUP BY DATE_TRUNC('minute', to_timestamp_seconds("EventTime")) ORDER BY DATE_TRUNC('minute', M) LIMIT 10 OFFSET 1000;
4 changes: 2 additions & 2 deletions datafusion/physical-expr-common/src/binary_map.rs
Original file line number Diff line number Diff line change
Expand Up @@ -235,9 +235,9 @@ where
}

/// The size, in number of entries, of the initial hash table
const INITIAL_MAP_CAPACITY: usize = 128;
pub(super) const INITIAL_MAP_CAPACITY: usize = 128;
/// The initial size, in bytes, of the string data
const INITIAL_BUFFER_CAPACITY: usize = 8 * 1024;
pub const INITIAL_BUFFER_CAPACITY: usize = 8 * 1024;
impl<O: OffsetSizeTrait, V> ArrowBytesMap<O, V>
where
V: Debug + PartialEq + Eq + Clone + Copy + Default,
Expand Down
315 changes: 315 additions & 0 deletions datafusion/physical-plan/src/aggregates/group_values/column_wise.rs
Original file line number Diff line number Diff line change
@@ -0,0 +1,315 @@
// Licensed to the Apache Software Foundation (ASF) under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
// regarding copyright ownership. The ASF licenses this file
// to you under the Apache License, Version 2.0 (the
// "License"); you may not use this file except in compliance
// with the License. You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing,
// software distributed under the License is distributed on an
// "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
// KIND, either express or implied. See the License for the
// specific language governing permissions and limitations
// under the License.

use crate::aggregates::group_values::group_value_row::{
ArrayRowEq, ByteGroupValueBuilder, PrimitiveGroupValueBuilder,
};
use crate::aggregates::group_values::GroupValues;
use ahash::RandomState;
use arrow::compute::cast;
use arrow::datatypes::{
Date32Type, Date64Type, Float32Type, Float64Type, Int16Type, Int32Type, Int64Type,
Int8Type, UInt16Type, UInt32Type, UInt64Type, UInt8Type,
};
use arrow::record_batch::RecordBatch;
use arrow_array::{Array, ArrayRef};
use arrow_schema::{DataType, SchemaRef};
use datafusion_common::hash_utils::create_hashes;
use datafusion_common::{DataFusionError, Result};
use datafusion_execution::memory_pool::proxy::{RawTableAllocExt, VecAllocExt};
use datafusion_expr::EmitTo;
use datafusion_physical_expr::binary_map::OutputType;

use hashbrown::raw::RawTable;

/// Compare GroupValue Rows column by column
pub struct GroupValuesColumn {
/// The output schema
schema: SchemaRef,

/// Logically maps group values to a group_index in
/// [`Self::group_values`] and in each accumulator
///
/// Uses the raw API of hashbrown to avoid actually storing the
/// keys (group values) in the table
///
/// keys: u64 hashes of the GroupValue
/// values: (hash, group_index)
map: RawTable<(u64, usize)>,

/// The size of `map` in bytes
map_size: usize,

/// The actual group by values, stored column-wise. Compare from
/// the left to right, each column is stored as `ArrayRowEq`.
/// This is shown faster than the row format
group_values: Vec<Box<dyn ArrayRowEq>>,

/// reused buffer to store hashes
hashes_buffer: Vec<u64>,

/// Random state for creating hashes
random_state: RandomState,
}

impl GroupValuesColumn {
pub fn try_new(schema: SchemaRef) -> Result<Self> {
let map = RawTable::with_capacity(0);
Ok(Self {
schema,
map,
map_size: 0,
group_values: vec![],
hashes_buffer: Default::default(),
random_state: Default::default(),
})
}
}

impl GroupValues for GroupValuesColumn {
fn intern(&mut self, cols: &[ArrayRef], groups: &mut Vec<usize>) -> Result<()> {
let n_rows = cols[0].len();

if self.group_values.is_empty() {
let len = cols.len();
let mut v = Vec::with_capacity(len);
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I think we should give len a better name like n_cols or just "inline" it like

Suggested change
let len = cols.len();
let mut v = Vec::with_capacity(len);
let mut v = Vec::with_capacity(cols.len());

because for me it currently just hurts readability.


for f in self.schema.fields().iter() {
let nullable = f.is_nullable();
match f.data_type() {
&DataType::Int8 => {
let b = PrimitiveGroupValueBuilder::<Int8Type>::new(nullable);
v.push(Box::new(b) as _)
}
&DataType::Int16 => {
let b = PrimitiveGroupValueBuilder::<Int16Type>::new(nullable);
v.push(Box::new(b) as _)
}
&DataType::Int32 => {
let b = PrimitiveGroupValueBuilder::<Int32Type>::new(nullable);
v.push(Box::new(b) as _)
}
&DataType::Int64 => {
let b = PrimitiveGroupValueBuilder::<Int64Type>::new(nullable);
v.push(Box::new(b) as _)
}
&DataType::UInt8 => {
let b = PrimitiveGroupValueBuilder::<UInt8Type>::new(nullable);
v.push(Box::new(b) as _)
}
&DataType::UInt16 => {
let b = PrimitiveGroupValueBuilder::<UInt16Type>::new(nullable);
v.push(Box::new(b) as _)
}
&DataType::UInt32 => {
let b = PrimitiveGroupValueBuilder::<UInt32Type>::new(nullable);
v.push(Box::new(b) as _)
}
&DataType::UInt64 => {
let b = PrimitiveGroupValueBuilder::<UInt64Type>::new(nullable);
v.push(Box::new(b) as _)
}
&DataType::Float32 => {
let b = PrimitiveGroupValueBuilder::<Float32Type>::new(nullable);
v.push(Box::new(b) as _)
}
&DataType::Float64 => {
let b = PrimitiveGroupValueBuilder::<Float64Type>::new(nullable);
v.push(Box::new(b) as _)
}
&DataType::Date32 => {
let b = PrimitiveGroupValueBuilder::<Date32Type>::new(nullable);
v.push(Box::new(b) as _)
}
&DataType::Date64 => {
let b = PrimitiveGroupValueBuilder::<Date64Type>::new(nullable);
v.push(Box::new(b) as _)
}
&DataType::Utf8 => {
let b = ByteGroupValueBuilder::<i32>::new(OutputType::Utf8);
v.push(Box::new(b) as _)
}
&DataType::LargeUtf8 => {
let b = ByteGroupValueBuilder::<i64>::new(OutputType::Utf8);
v.push(Box::new(b) as _)
}
&DataType::Binary => {
let b = ByteGroupValueBuilder::<i32>::new(OutputType::Binary);
v.push(Box::new(b) as _)
}
&DataType::LargeBinary => {
let b = ByteGroupValueBuilder::<i64>::new(OutputType::Binary);
v.push(Box::new(b) as _)
}
dt => todo!("{dt} not impl"),
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

It would be nice if this was an internal error rather than a panic

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

in #12620

}
}
self.group_values = v;
}

// tracks to which group each of the input rows belongs
groups.clear();

// 1.1 Calculate the group keys for the group values
let batch_hashes = &mut self.hashes_buffer;
batch_hashes.clear();
batch_hashes.resize(n_rows, 0);
create_hashes(cols, &self.random_state, batch_hashes)?;

for (row, &target_hash) in batch_hashes.iter().enumerate() {
let entry = self.map.get_mut(target_hash, |(exist_hash, group_idx)| {
// Somewhat surprisingly, this closure can be called even if the
// hash doesn't match, so check the hash first with an integer
// comparison first avoid the more expensive comparison with
// group value. https://github.com/apache/datafusion/pull/11718
if target_hash != *exist_hash {
return false;
}

fn check_row_equal(
array_row: &dyn ArrayRowEq,
lhs_row: usize,
array: &ArrayRef,
rhs_row: usize,
) -> bool {
array_row.equal_to(lhs_row, array, rhs_row)
}

for (i, group_val) in self.group_values.iter().enumerate() {
if !check_row_equal(group_val.as_ref(), *group_idx, &cols[i], row) {
return false;
}
}

true
});

let group_idx = match entry {
// Existing group_index for this group value
Some((_hash, group_idx)) => *group_idx,
// 1.2 Need to create new entry for the group
None => {
// Add new entry to aggr_state and save newly created index
// let group_idx = group_values.num_rows();
// group_values.push(group_rows.row(row));

let mut checklen = 0;
let group_idx = self.group_values[0].len();
for (i, group_value) in self.group_values.iter_mut().enumerate() {
group_value.append_val(&cols[i], row);
let len = group_value.len();
if i == 0 {
checklen = len;
} else {
debug_assert_eq!(checklen, len);
}
}

// for hasher function, use precomputed hash value
self.map.insert_accounted(
(target_hash, group_idx),
|(hash, _group_index)| *hash,
&mut self.map_size,
);
group_idx
}
};
groups.push(group_idx);
}

Ok(())
}

fn size(&self) -> usize {
let group_values_size = self.group_values.len();
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

My understanding is that size should return allocated size in bytes, so this does not look correct for self.group_values. I would expect us need to do something like

        let group_values_size = self.group_values.iter().map(|column| column.size()).sum()

and add a size method to ArrayRowEq so it can do some size calculation based on the field.

group_values_size + self.map_size + self.hashes_buffer.allocated_size()
}

fn is_empty(&self) -> bool {
self.len() == 0
}

fn len(&self) -> usize {
if self.group_values.is_empty() {
return 0;
}

self.group_values[0].len()
}

fn emit(&mut self, emit_to: EmitTo) -> Result<Vec<ArrayRef>> {
let mut output = match emit_to {
EmitTo::All => {
let group_values = std::mem::take(&mut self.group_values);
debug_assert!(self.group_values.is_empty());

group_values
.into_iter()
.map(|v| v.build())
.collect::<Vec<_>>()
}
EmitTo::First(n) => {
let output = self
.group_values
.iter_mut()
.map(|v| v.take_n(n))
.collect::<Vec<_>>();

// SAFETY: self.map outlives iterator and is not modified concurrently
unsafe {
for bucket in self.map.iter() {
// Decrement group index by n
match bucket.as_ref().1.checked_sub(n) {
// Group index was >= n, shift value down
Some(sub) => bucket.as_mut().1 = sub,
// Group index was < n, so remove from table
None => self.map.erase(bucket),
}
}
}

output
}
};

// TODO: Materialize dictionaries in group keys (#7647)
for (field, array) in self.schema.fields.iter().zip(&mut output) {
let expected = field.data_type();
if let DataType::Dictionary(_, v) = expected {
let actual = array.data_type();
if v.as_ref() != actual {
return Err(DataFusionError::Internal(format!(
"Converted group rows expected dictionary of {v} got {actual}"
)));
}
*array = cast(array.as_ref(), expected)?;
}
}

Ok(output)
}

fn clear_shrink(&mut self, batch: &RecordBatch) {
let count = batch.num_rows();
self.group_values.clear();
self.map.clear();
self.map.shrink_to(count, |_| 0); // hasher does not matter since the map is cleared
self.map_size = self.map.capacity() * std::mem::size_of::<(u64, usize)>();
self.hashes_buffer.clear();
self.hashes_buffer.shrink_to(count);
}
}
Loading