@@ -79,12 +79,12 @@ func init() {
7979 dockerPool .Config .MaxTotal = runParallelInt
8080}
8181
82- const dbNameAtomic = "_test#$.ДБ_atomic_"
83- const dbNameOrdinary = "_test#$.ДБ_ordinary_"
82+ const dbNameAtomic = "_test#$.ДБ_atomic_/issue \\ _1091 "
83+ const dbNameOrdinary = "_test#$.ДБ_ordinary_/issue \\ _1091 "
8484const dbNameMySQL = "mysql_db"
8585const dbNamePostgreSQL = "pgsql_db"
86- const Issue331Atomic = "_issue331._atomic_"
87- const Issue331Ordinary = "_issue331.ordinary_"
86+ const Issue331Issue1091Atomic = "_issue331._atomic_/issue \\ _1091 "
87+ const Issue331Issue1091Ordinary = "_issue331.ordinary_/issue \\ _1091 "
8888
8989type TestDataStruct struct {
9090 Database string
@@ -152,28 +152,28 @@ var defaultTestData = []TestDataStruct{
152152 Fields : []string {"TimeStamp" , "Item" },
153153 OrderBy : "TimeStamp" ,
154154 }, {
155- Database : Issue331Atomic , DatabaseEngine : "Atomic" ,
156- Name : Issue331Atomic , // need cover fix https://github.com/Altinity/clickhouse-backup/issues/331
157- Schema : fmt .Sprintf ("(`%s` UInt64, Col1 String, Col2 String, Col3 String, Col4 String, Col5 String) ENGINE = MergeTree PARTITION BY `%s` ORDER BY (`%s`, Col1, Col2, Col3, Col4, Col5) SETTINGS index_granularity = 8192" , Issue331Atomic , Issue331Atomic , Issue331Atomic ),
155+ Database : Issue331Issue1091Atomic , DatabaseEngine : "Atomic" ,
156+ Name : Issue331Issue1091Atomic , // need cover fix https://github.com/Altinity/clickhouse-backup/issues/331
157+ Schema : fmt .Sprintf ("(`%s` UInt64, Col1 String, Col2 String, Col3 String, Col4 String, Col5 String) ENGINE = MergeTree PARTITION BY `%s` ORDER BY (`%s`, Col1, Col2, Col3, Col4, Col5) SETTINGS index_granularity = 8192" , Issue331Issue1091Atomic , Issue331Issue1091Atomic , Issue331Issue1091Atomic ),
158158 Rows : func () []map [string ]interface {} {
159159 var result []map [string ]interface {}
160160 for i := 0 ; i < 100 ; i ++ {
161- result = append (result , map [string ]interface {}{Issue331Atomic : uint64 (i ), "Col1" : "Text1" , "Col2" : "Text2" , "Col3" : "Text3" , "Col4" : "Text4" , "Col5" : "Text5" })
161+ result = append (result , map [string ]interface {}{Issue331Issue1091Atomic : uint64 (i ), "Col1" : "Text1" , "Col2" : "Text2" , "Col3" : "Text3" , "Col4" : "Text4" , "Col5" : "Text5" })
162162 }
163163 return result
164164 }(),
165- Fields : []string {Issue331Atomic , "Col1" , "Col2" , "Col3" , "Col4" , "Col5" },
166- OrderBy : Issue331Atomic + "_{test}" ,
165+ Fields : []string {Issue331Issue1091Atomic , "Col1" , "Col2" , "Col3" , "Col4" , "Col5" },
166+ OrderBy : Issue331Issue1091Atomic + "_{test}" ,
167167 }, {
168- Database : Issue331Ordinary , DatabaseEngine : "Ordinary" ,
169- Name : Issue331Ordinary , // need cover fix https://github.com/Altinity/clickhouse-backup/issues/331
170- Schema : fmt .Sprintf ("(`%s` String, order_time DateTime, amount Float64) ENGINE = MergeTree() PARTITION BY toYYYYMM(order_time) ORDER BY (order_time, `%s`)" , Issue331Ordinary , Issue331Ordinary ),
168+ Database : Issue331Issue1091Ordinary , DatabaseEngine : "Ordinary" ,
169+ Name : Issue331Issue1091Ordinary , // need cover fix https://github.com/Altinity/clickhouse-backup/issues/331
170+ Schema : fmt .Sprintf ("(`%s` String, order_time DateTime, amount Float64) ENGINE = MergeTree() PARTITION BY toYYYYMM(order_time) ORDER BY (order_time, `%s`)" , Issue331Issue1091Ordinary , Issue331Issue1091Ordinary ),
171171 Rows : []map [string ]interface {}{
172- {Issue331Ordinary : "1" , "order_time" : toTS ("2010-01-01 00:00:00" ), "amount" : 1.0 },
173- {Issue331Ordinary : "2" , "order_time" : toTS ("2010-02-01 00:00:00" ), "amount" : 2.0 },
172+ {Issue331Issue1091Ordinary : "1" , "order_time" : toTS ("2010-01-01 00:00:00" ), "amount" : 1.0 },
173+ {Issue331Issue1091Ordinary : "2" , "order_time" : toTS ("2010-02-01 00:00:00" ), "amount" : 2.0 },
174174 },
175- Fields : []string {Issue331Ordinary , "order_time" , "amount" },
176- OrderBy : Issue331Ordinary + "_{test}" ,
175+ Fields : []string {Issue331Issue1091Ordinary , "order_time" , "amount" },
176+ OrderBy : Issue331Issue1091Ordinary + "_{test}" ,
177177 }, {
178178 Database : dbNameOrdinary , DatabaseEngine : "Ordinary" ,
179179 Name : "yuzhichang_table3" ,
@@ -320,17 +320,17 @@ var defaultTestData = []TestDataStruct{
320320 " (`%s` UInt64, Col1 String, Col2 String, Col3 String, Col4 String, Col5 String) PRIMARY KEY `%s` " +
321321 " SOURCE(CLICKHOUSE(host 'localhost' port 9000 db '%s' table '%s' user 'default' password ''))" +
322322 " LAYOUT(HASHED()) LIFETIME(60)" ,
323- Issue331Atomic , Issue331Atomic , Issue331Atomic , Issue331Atomic ), // same table and name need cover fix https://github.com/Altinity/clickhouse-backup/issues/331
323+ Issue331Issue1091Atomic , Issue331Issue1091Atomic , Issue331Issue1091Atomic , Issue331Issue1091Atomic ), // same table and name need cover fix https://github.com/Altinity/clickhouse-backup/issues/331
324324 SkipInsert : true ,
325325 Rows : func () []map [string ]interface {} {
326326 var result []map [string ]interface {}
327327 for i := 0 ; i < 100 ; i ++ {
328- result = append (result , map [string ]interface {}{Issue331Atomic : uint64 (i ), "Col1" : "Text1" , "Col2" : "Text2" , "Col3" : "Text3" , "Col4" : "Text4" , "Col5" : "Text5" })
328+ result = append (result , map [string ]interface {}{Issue331Issue1091Atomic : uint64 (i ), "Col1" : "Text1" , "Col2" : "Text2" , "Col3" : "Text3" , "Col4" : "Text4" , "Col5" : "Text5" })
329329 }
330330 return result
331331 }(),
332332 Fields : []string {},
333- OrderBy : Issue331Atomic + "_{test}" ,
333+ OrderBy : Issue331Issue1091Atomic + "_{test}" ,
334334 },
335335 {
336336 Database : dbNameMySQL , DatabaseEngine : "MySQL('mysql:3306','mysql','root','root')" ,
@@ -389,28 +389,28 @@ var defaultIncrementData = []TestDataStruct{
389389 Fields : []string {"TimeStamp" , "Item" },
390390 OrderBy : "TimeStamp" ,
391391 }, {
392- Database : Issue331Atomic , DatabaseEngine : "Atomic" ,
393- Name : Issue331Atomic , // need cover fix https://github.com/Altinity/clickhouse-backup/issues/331
394- Schema : fmt .Sprintf ("(`%s` UInt64, Col1 String, Col2 String, Col3 String, Col4 String, Col5 String) ENGINE = MergeTree PARTITION BY `%s` ORDER BY (`%s`, Col1, Col2, Col3, Col4, Col5) SETTINGS index_granularity = 8192" , Issue331Atomic , Issue331Atomic , Issue331Atomic ),
392+ Database : Issue331Issue1091Atomic , DatabaseEngine : "Atomic" ,
393+ Name : Issue331Issue1091Atomic , // need cover fix https://github.com/Altinity/clickhouse-backup/issues/331
394+ Schema : fmt .Sprintf ("(`%s` UInt64, Col1 String, Col2 String, Col3 String, Col4 String, Col5 String) ENGINE = MergeTree PARTITION BY `%s` ORDER BY (`%s`, Col1, Col2, Col3, Col4, Col5) SETTINGS index_granularity = 8192" , Issue331Issue1091Atomic , Issue331Issue1091Atomic , Issue331Issue1091Atomic ),
395395 Rows : func () []map [string ]interface {} {
396396 var result []map [string ]interface {}
397397 for i := 200 ; i < 220 ; i ++ {
398- result = append (result , map [string ]interface {}{Issue331Atomic : uint64 (i ), "Col1" : "Text1" , "Col2" : "Text2" , "Col3" : "Text3" , "Col4" : "Text4" , "Col5" : "Text5" })
398+ result = append (result , map [string ]interface {}{Issue331Issue1091Atomic : uint64 (i ), "Col1" : "Text1" , "Col2" : "Text2" , "Col3" : "Text3" , "Col4" : "Text4" , "Col5" : "Text5" })
399399 }
400400 return result
401401 }(),
402- Fields : []string {Issue331Atomic , "Col1" , "Col2" , "Col3" , "Col4" , "Col5" },
403- OrderBy : Issue331Atomic + "_{test}" ,
402+ Fields : []string {Issue331Issue1091Atomic , "Col1" , "Col2" , "Col3" , "Col4" , "Col5" },
403+ OrderBy : Issue331Issue1091Atomic + "_{test}" ,
404404 }, {
405- Database : Issue331Ordinary , DatabaseEngine : "Ordinary" ,
406- Name : Issue331Ordinary , // need cover fix https://github.com/Altinity/clickhouse-backup/issues/331
407- Schema : fmt .Sprintf ("(`%s` String, order_time DateTime, amount Float64) ENGINE = MergeTree() PARTITION BY toYYYYMM(order_time) ORDER BY (order_time, `%s`)" , Issue331Ordinary , Issue331Ordinary ),
405+ Database : Issue331Issue1091Ordinary , DatabaseEngine : "Ordinary" ,
406+ Name : Issue331Issue1091Ordinary , // need cover fix https://github.com/Altinity/clickhouse-backup/issues/331
407+ Schema : fmt .Sprintf ("(`%s` String, order_time DateTime, amount Float64) ENGINE = MergeTree() PARTITION BY toYYYYMM(order_time) ORDER BY (order_time, `%s`)" , Issue331Issue1091Ordinary , Issue331Issue1091Ordinary ),
408408 Rows : []map [string ]interface {}{
409- {Issue331Ordinary : "3" , "order_time" : toTS ("2010-03-01 00:00:00" ), "amount" : 3.0 },
410- {Issue331Ordinary : "4" , "order_time" : toTS ("2010-04-01 00:00:00" ), "amount" : 4.0 },
409+ {Issue331Issue1091Ordinary : "3" , "order_time" : toTS ("2010-03-01 00:00:00" ), "amount" : 3.0 },
410+ {Issue331Issue1091Ordinary : "4" , "order_time" : toTS ("2010-04-01 00:00:00" ), "amount" : 4.0 },
411411 },
412- Fields : []string {Issue331Ordinary , "order_time" , "amount" },
413- OrderBy : Issue331Ordinary + "_{test}" ,
412+ Fields : []string {Issue331Issue1091Ordinary , "order_time" , "amount" },
413+ OrderBy : Issue331Issue1091Ordinary + "_{test}" ,
414414 }, {
415415 Database : dbNameOrdinary , DatabaseEngine : "Ordinary" ,
416416 Name : "yuzhichang_table3" ,
@@ -873,7 +873,7 @@ func TestS3NoDeletePermission(t *testing.T) {
873873 env .DockerExecNoError (r , "clickhouse-backup" , "clickhouse-backup" , "restore_remote" , "no_delete_backup" )
874874 env .DockerExecNoError (r , "clickhouse-backup" , "clickhouse-backup" , "delete" , "local" , "no_delete_backup" )
875875 r .Error (env .DockerExec ("clickhouse-backup" , "clickhouse-backup" , "delete" , "remote" , "no_delete_backup" ))
876- databaseList := []string {dbNameOrdinary , dbNameAtomic , dbNameMySQL , dbNamePostgreSQL , Issue331Atomic , Issue331Ordinary }
876+ databaseList := []string {dbNameOrdinary , dbNameAtomic , dbNameMySQL , dbNamePostgreSQL , Issue331Issue1091Atomic , Issue331Issue1091Ordinary }
877877 dropDatabasesFromTestDataDataSet (t , r , env , databaseList )
878878 r .NoError (env .DockerCP ("config-s3.yml" , "clickhouse-backup:/etc/clickhouse-backup/config.yml" ))
879879 env .DockerExecNoError (r , "clickhouse-backup" , "clickhouse-backup" , "delete" , "remote" , "no_delete_backup" )
@@ -2038,7 +2038,7 @@ func TestKeepBackupRemoteAndDiffFromRemote(t *testing.T) {
20382038 for i := 0 ; i < 5 ; i ++ {
20392039 backupNames [i ] = fmt .Sprintf ("keep_remote_backup_%d" , i )
20402040 }
2041- databaseList := []string {dbNameOrdinary , dbNameAtomic , dbNameMySQL , dbNamePostgreSQL , Issue331Atomic , Issue331Ordinary }
2041+ databaseList := []string {dbNameOrdinary , dbNameAtomic , dbNameMySQL , dbNamePostgreSQL , Issue331Issue1091Atomic , Issue331Issue1091Ordinary }
20422042 fullCleanup (t , r , env , backupNames , []string {"remote" , "local" }, databaseList , false , false , "config-s3.yml" )
20432043 incrementData := defaultIncrementData
20442044 generateTestData (t , r , env , "S3" , false , defaultTestData )
@@ -2073,7 +2073,7 @@ func TestKeepBackupRemoteAndDiffFromRemote(t *testing.T) {
20732073 }
20742074 env .DockerExecNoError (r , "clickhouse-backup" , "clickhouse-backup" , "-c" , "/etc/clickhouse-backup/config-s3.yml" , "restore" , "--rm" , latestIncrementBackup )
20752075 var res uint64
2076- r .NoError (env .ch .SelectSingleRowNoCtx (& res , fmt .Sprintf ("SELECT count() FROM `%s_%s`.`%s_%s`" , Issue331Atomic , t .Name (), Issue331Atomic , t .Name ())))
2076+ r .NoError (env .ch .SelectSingleRowNoCtx (& res , fmt .Sprintf ("SELECT count() FROM `%s_%s`.`%s_%s`" , Issue331Issue1091Atomic , t .Name (), Issue331Issue1091Atomic , t .Name ())))
20772077 r .Equal (uint64 (100 + 20 * 4 ), res )
20782078 fullCleanup (t , r , env , []string {latestIncrementBackup }, []string {"local" }, nil , true , true , "config-s3.yml" )
20792079 fullCleanup (t , r , env , backupNames , []string {"remote" }, databaseList , true , true , "config-s3.yml" )
@@ -2617,7 +2617,7 @@ func (env *TestEnvironment) runMainIntegrationScenario(t *testing.T, remoteStora
26172617 fullBackupName := fmt .Sprintf ("%s_full_%d" , t .Name (), rand .Int ())
26182618 incrementBackupName := fmt .Sprintf ("%s_increment_%d" , t .Name (), rand .Int ())
26192619 incrementBackupName2 := fmt .Sprintf ("%s_increment2_%d" , t .Name (), rand .Int ())
2620- databaseList := []string {dbNameOrdinary , dbNameAtomic , dbNameMySQL , dbNamePostgreSQL , Issue331Atomic , Issue331Ordinary }
2620+ databaseList := []string {dbNameOrdinary , dbNameAtomic , dbNameMySQL , dbNamePostgreSQL , Issue331Issue1091Atomic , Issue331Issue1091Ordinary }
26212621 tablesPattern := fmt .Sprintf ("*_%s.*" , t .Name ())
26222622 log .Debug ().Msg ("Clean before start" )
26232623 fullCleanup (t , r , env , []string {fullBackupName , incrementBackupName }, []string {"remote" , "local" }, databaseList , false , false , backupConfig )
0 commit comments