Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion be/src/exec/csv_scan_node.h
Original file line number Diff line number Diff line change
Expand Up @@ -70,7 +70,7 @@ class CsvScanNode : public ScanNode {
const SlotDescriptor* slot, RuntimeState* state,
std::stringstream* error_msg);

// split one line into fields, check every fields, fill every filed into tuple
// split one line into fields, check every fields, fill every field into tuple
bool split_check_fill(const std::string& line, RuntimeState* state);

void fill_fix_length_string(
Expand Down
2 changes: 1 addition & 1 deletion be/src/exec/parquet_reader.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -92,7 +92,7 @@ Status ParquetReaderWrap::init_parquet_reader(const std::vector<SlotDescriptor*>
for (int i = 0; i < _parquet_column_ids.size(); i++) {
std::shared_ptr<arrow::Field> field = field_schema->field(i);
if (!field) {
LOG(WARNING) << "Get filed schema failed. Column order:" << i;
LOG(WARNING) << "Get field schema failed. Column order:" << i;
return Status::InternalError(status.ToString());
}
_parquet_column_type.emplace_back(field->type()->id());
Expand Down
2 changes: 1 addition & 1 deletion be/src/olap/data_dir.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -125,7 +125,7 @@ Status DataDir::_init_cluster_id() {
if (fp == NULL) {
RETURN_NOT_OK_STATUS_WITH_WARN(
Status::IOError(Substitute("failed to open cluster id file $0", cluster_id_path)),
"open file filed");
"open file failed");
}

int lock_res = flock(fp->_fileno, LOCK_EX | LOCK_NB);
Expand Down
6 changes: 3 additions & 3 deletions be/src/olap/field.h
Original file line number Diff line number Diff line change
Expand Up @@ -78,7 +78,7 @@ class Field {
// todo(kks): Unify AggregateInfo::init method and Field::agg_init method

// This function will initialize destination with source.
// This functionn differs copy functionn in that if this filed
// This functionn differs copy functionn in that if this field
// contain aggregate information, this functionn will initialize
// destination in aggregate format, and update with srouce content.
virtual void agg_init(RowCursorCell* dst, const RowCursorCell& src, MemPool* mem_pool, ObjectPool* agg_pool) const {
Expand Down Expand Up @@ -189,12 +189,12 @@ class Field {
_type_info->deep_copy(dst->mutable_cell_ptr(), src.cell_ptr(), pool);
}

// deep copy filed content from `src` to `dst` without null-byte
// deep copy field content from `src` to `dst` without null-byte
inline void deep_copy_content(char* dst, const char* src, MemPool* mem_pool) const {
_type_info->deep_copy(dst, src, mem_pool);
}

// shallow copy filed content from `src` to `dst` without null-byte.
// shallow copy field content from `src` to `dst` without null-byte.
// for string like type, shallow copy only copies Slice, not the actual data pointed by slice.
inline void shallow_copy_content(char* dst, const char* src) const {
_type_info->shallow_copy(dst, src);
Expand Down
2 changes: 1 addition & 1 deletion be/src/olap/rowset/column_reader.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -678,7 +678,7 @@ ColumnReader* ColumnReader::create(uint32_t column_id,
case OLAP_FIELD_TYPE_LIST:
case OLAP_FIELD_TYPE_MAP:
default: {
LOG(WARNING) << "unspported filed type. field=" << column.name()
LOG(WARNING) << "unspported field type. field=" << column.name()
<< ", type=" << column.type();
break;
}
Expand Down
2 changes: 1 addition & 1 deletion be/src/olap/rowset/column_writer.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -122,7 +122,7 @@ ColumnWriter* ColumnWriter::create(uint32_t column_id,
case OLAP_FIELD_TYPE_LIST:
case OLAP_FIELD_TYPE_MAP:
default: {
LOG(WARNING) << "Unspported filed type. field=" << column.name()
LOG(WARNING) << "Unspported field type. field=" << column.name()
<< ", type=" << column.type();
break;
}
Expand Down