Skip to content

Commit

Permalink
[branch-2.1] Picks "[opt](autoinc) Remove some restrictions on schema…
Browse files Browse the repository at this point in the history
… change on table that has auto-increment column apache#40280" (apache#41096)

picks apache#40280
  • Loading branch information
bobhan1 authored Sep 23, 2024
1 parent 9dc55f9 commit 7d64c8c
Show file tree
Hide file tree
Showing 5 changed files with 140 additions and 12 deletions.
2 changes: 2 additions & 0 deletions be/src/olap/schema_change.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -1121,6 +1121,8 @@ Status SchemaChangeHandler::_convert_historical_rowsets(const SchemaChangeParams
// b. Generate historical data converter
auto sc_procedure = get_sc_procedure(changer, sc_sorting, sc_directly);

DBUG_EXECUTE_IF("SchemaChangeJob::_convert_historical_rowsets.block", DBUG_BLOCK);

// c.Convert historical data
bool have_failure_rowset = false;
for (const auto& rs_reader : sc_params.ref_rowset_readers) {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -718,12 +718,6 @@ private boolean processModifyColumn(ModifyColumnClause alterClause, OlapTable ol
} // end for handling other indices

if (typeChanged && !lightSchemaChange) {
Optional<Column> autoIncCol = olapTable.getBaseSchema(true).stream()
.filter(col -> col.isAutoInc()).findFirst();
if (autoIncCol.isPresent()) {
throw new DdlException("Can not modify column " + modColumn.getName() + " becasue table "
+ olapTable.getName() + " has auto-increment column " + autoIncCol.get().getName());
}
/*
* In new alter table process (AlterJobV2), any modified columns are treated as new columns.
* But the modified columns' name does not changed. So in order to distinguish this, we will add
Expand Down
Original file line number Diff line number Diff line change
@@ -0,0 +1,9 @@
-- This file is automatically generated. You should know what you did if you want to edit this
-- !sql --

-- !sql --

-- !sql --

-- !sql --

Original file line number Diff line number Diff line change
@@ -0,0 +1,129 @@
// Licensed to the Apache Software Foundation (ASF) under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
// regarding copyright ownership. The ASF licenses this file
// to you under the Apache License, Version 2.0 (the
// "License"); you may not use this file except in compliance
// with the License. You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing,
// software distributed under the License is distributed on an
// "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
// KIND, either express or implied. See the License for the
// specific language governing permissions and limitations
// under the License.

import org.junit.Assert
import java.util.concurrent.TimeUnit
import java.util.concurrent.atomic.AtomicBoolean

suite("test_auto_inc_schema_change_double_write", "nonConcurrent") {
def backends = sql_return_maparray('show backends')
def replicaNum = 0
for (def be : backends) {
def alive = be.Alive.toBoolean()
def decommissioned = be.SystemDecommissioned.toBoolean()
if (alive && !decommissioned) {
replicaNum++
}
}
assertTrue(replicaNum > 0)
if (isCloudMode()) {
replicaNum = 1
}

def block_convert_historical_rowsets = {
if (isCloudMode()) {
GetDebugPoint().enableDebugPointForAllBEs("CloudSchemaChangeJob::_convert_historical_rowsets.block")
} else {
GetDebugPoint().enableDebugPointForAllBEs("SchemaChangeJob::_convert_historical_rowsets.block")
}
}

def unblock = {
if (isCloudMode()) {
GetDebugPoint().disableDebugPointForAllBEs("CloudSchemaChangeJob::_convert_historical_rowsets.block")
} else {
GetDebugPoint().disableDebugPointForAllBEs("SchemaChangeJob::_convert_historical_rowsets.block")
}
}

for (def model : ["UNIQUE", "DUPLICATE"]) {
try {
GetDebugPoint().clearDebugPointsForAllFEs()
GetDebugPoint().clearDebugPointsForAllBEs()
def tableName = "test_auto_inc_schema_change_double_write"
def table1 = "${tableName}_${model}"
sql "DROP TABLE IF EXISTS ${table1} FORCE;"
sql """ CREATE TABLE IF NOT EXISTS ${table1} (
`k1` BIGINT NOT NULL AUTO_INCREMENT,
`c1` int,
`c2` int,
`c3` int,
`c4` int
)${model} KEY(k1)
DISTRIBUTED BY HASH(k1) BUCKETS 1
PROPERTIES (
"disable_auto_compaction" = "true",
"replication_num" = "${replicaNum}"); """

sql """insert into ${table1}(c1,c2,c3,c4) select number,number,number,number from numbers("number"="5000"); """
sql "sync;"
qt_sql "select count(*) from ${table1} group by k1 having count(*) > 1;"

block_convert_historical_rowsets()

AtomicBoolean stopped = new AtomicBoolean(false)

def iters = 3
def rows = 500
def thread_num = 4
def t1 = Thread.start {
def threads = []
(1..thread_num).each { id1 ->
threads.add(Thread.start {
while (!stopped.get()) {
(1..iters).each { id2 ->
sql """insert into ${table1}(c1,c2,c3,c4) select number,number,number,number from numbers("number"="${rows}");"""
}
Thread.sleep(200)
}
})
}

threads.each { thread -> thread.join() }
}

Thread.sleep(3000)

sql "alter table ${table1} modify column c3 varchar(100) null;"

Thread.sleep(3000);

unblock()

def t2 = Thread.start {
waitForSchemaChangeDone {
sql """SHOW ALTER TABLE COLUMN WHERE TableName='${table1}' ORDER BY createtime DESC LIMIT 1"""
time 20000
}
}

Thread.sleep(5000);
stopped.set(true)
t1.join()
t2.join()

qt_sql "select count(*) from ${table1} group by k1 having count(*) > 1;"

} catch(Exception e) {
logger.info(e.getMessage())
throw e
} finally {
GetDebugPoint().clearDebugPointsForAllFEs()
GetDebugPoint().clearDebugPointsForAllBEs()
}
}
}
Original file line number Diff line number Diff line change
Expand Up @@ -53,12 +53,6 @@ suite("test_schema_change_auto_inc") {
exception "Can't modify the column[id2]'s auto-increment attribute."
}

// schema change that invoke double write on a table which has auto-increment column is forbidden
test {
sql "alter table ${table1} modify column value VARCHAR(20) NOT NULL"
exception "Can not modify column value becasue table ${table1} has auto-increment column id"
}

sql """ insert into ${table1}(name, value, id2) values("A", 999, 1), ("B", 888, 2), ("C", 777, 3);"""
qt_sql "select count(distinct id) from ${table1};"

Expand Down

0 comments on commit 7d64c8c

Please sign in to comment.