Merge branch 'master' into incorrect-datetime-1970
This commit is contained in:
commit
7d3514be08
2
.github/PULL_REQUEST_TEMPLATE.md
vendored
2
.github/PULL_REQUEST_TEMPLATE.md
vendored
@ -11,7 +11,7 @@ Related issue: https://github.com/github/gh-ost/issues/0123456789
|
|||||||
|
|
||||||
### Description
|
### Description
|
||||||
|
|
||||||
This PR [briefly explain what is does]
|
This PR [briefly explain what it does]
|
||||||
|
|
||||||
> In case this PR introduced Go code changes:
|
> In case this PR introduced Go code changes:
|
||||||
|
|
||||||
|
11
.travis.yml
11
.travis.yml
@ -8,13 +8,22 @@ os:
|
|||||||
|
|
||||||
env:
|
env:
|
||||||
- MYSQL_USER=root
|
- MYSQL_USER=root
|
||||||
|
- CURRENT_CI_ENV=travis
|
||||||
|
|
||||||
|
addons:
|
||||||
|
apt:
|
||||||
|
packages:
|
||||||
|
- git
|
||||||
|
- numactl
|
||||||
|
- libaio1
|
||||||
|
|
||||||
before_install:
|
before_install:
|
||||||
- mysql -e 'CREATE DATABASE IF NOT EXISTS test;'
|
- mysql -e 'CREATE DATABASE IF NOT EXISTS test;'
|
||||||
|
|
||||||
install: true
|
install: true
|
||||||
|
|
||||||
script: script/cibuild
|
script:
|
||||||
|
- script/cibuild
|
||||||
|
|
||||||
notifications:
|
notifications:
|
||||||
email: false
|
email: false
|
||||||
|
@ -94,7 +94,7 @@ Please see [Coding gh-ost](doc/coding-ghost.md) for a guide to getting started d
|
|||||||
|
|
||||||
[Download latest release here](https://github.com/github/gh-ost/releases/latest)
|
[Download latest release here](https://github.com/github/gh-ost/releases/latest)
|
||||||
|
|
||||||
`gh-ost` is a Go project; it is built with Go `1.8` (though `1.7` should work as well). To build on your own, use either:
|
`gh-ost` is a Go project; it is built with Go `1.9` and above. To build on your own, use either:
|
||||||
- [script/build](https://github.com/github/gh-ost/blob/master/script/build) - this is the same build script used by CI hence the authoritative; artifact is `./bin/gh-ost` binary.
|
- [script/build](https://github.com/github/gh-ost/blob/master/script/build) - this is the same build script used by CI hence the authoritative; artifact is `./bin/gh-ost` binary.
|
||||||
- [build.sh](https://github.com/github/gh-ost/blob/master/build.sh) for building `tar.gz` artifacts in `/tmp/gh-ost`
|
- [build.sh](https://github.com/github/gh-ost/blob/master/build.sh) for building `tar.gz` artifacts in `/tmp/gh-ost`
|
||||||
|
|
||||||
@ -107,3 +107,5 @@ Generally speaking, `master` branch is stable, but only [releases](https://githu
|
|||||||
- [@ggunson](https://github.com/ggunson)
|
- [@ggunson](https://github.com/ggunson)
|
||||||
- [@tomkrouper](https://github.com/tomkrouper)
|
- [@tomkrouper](https://github.com/tomkrouper)
|
||||||
- [@shlomi-noach](https://github.com/shlomi-noach)
|
- [@shlomi-noach](https://github.com/shlomi-noach)
|
||||||
|
- [@jessbreckenridge](https://github.com/jessbreckenridge)
|
||||||
|
- [@gtowey](https://github.com/gtowey)
|
||||||
|
@ -1 +1 @@
|
|||||||
1.0.42
|
1.0.45
|
||||||
|
2
build.sh
2
build.sh
@ -10,7 +10,7 @@ function build {
|
|||||||
GOOS=$3
|
GOOS=$3
|
||||||
GOARCH=$4
|
GOARCH=$4
|
||||||
|
|
||||||
if [[ $(go version | egrep "go1[.][012345678]") ]]; then
|
if ! go version | egrep -q 'go(1[.]9|1[.]1[0-9])' ; then
|
||||||
echo "go version is too low. Must use 1.9 or above"
|
echo "go version is too low. Must use 1.9 or above"
|
||||||
exit 1
|
exit 1
|
||||||
fi
|
fi
|
||||||
|
@ -1,4 +1,4 @@
|
|||||||
`gh-ost` has been updated to work with Amazon RDS however due to GitHub not relying using AWS for databases, this documentation is community driven so if you find a bug please [open an issue][new_issue]!
|
`gh-ost` has been updated to work with Amazon RDS however due to GitHub not using AWS for databases, this documentation is community driven so if you find a bug please [open an issue][new_issue]!
|
||||||
|
|
||||||
# Amazon RDS
|
# Amazon RDS
|
||||||
|
|
||||||
|
@ -52,3 +52,5 @@ The `SUPER` privilege is required for `STOP SLAVE`, `START SLAVE` operations. Th
|
|||||||
- If you have en `enum` field as part of your migration key (typically the `PRIMARY KEY`), migration performance will be degraded and potentially bad. [Read more](https://github.com/github/gh-ost/pull/277#issuecomment-254811520)
|
- If you have en `enum` field as part of your migration key (typically the `PRIMARY KEY`), migration performance will be degraded and potentially bad. [Read more](https://github.com/github/gh-ost/pull/277#issuecomment-254811520)
|
||||||
|
|
||||||
- Migrating a `FEDERATED` table is unsupported and is irrelevant to the problem `gh-ost` tackles.
|
- Migrating a `FEDERATED` table is unsupported and is irrelevant to the problem `gh-ost` tackles.
|
||||||
|
|
||||||
|
- `ALTER TABLE ... RENAME TO some_other_name` is not supported (and you shouldn't use `gh-ost` for such a trivial operation).
|
@ -28,23 +28,23 @@ type RowsEstimateMethod string
|
|||||||
|
|
||||||
const (
|
const (
|
||||||
TableStatusRowsEstimate RowsEstimateMethod = "TableStatusRowsEstimate"
|
TableStatusRowsEstimate RowsEstimateMethod = "TableStatusRowsEstimate"
|
||||||
ExplainRowsEstimate = "ExplainRowsEstimate"
|
ExplainRowsEstimate RowsEstimateMethod = "ExplainRowsEstimate"
|
||||||
CountRowsEstimate = "CountRowsEstimate"
|
CountRowsEstimate RowsEstimateMethod = "CountRowsEstimate"
|
||||||
)
|
)
|
||||||
|
|
||||||
type CutOver int
|
type CutOver int
|
||||||
|
|
||||||
const (
|
const (
|
||||||
CutOverAtomic CutOver = iota
|
CutOverAtomic CutOver = iota
|
||||||
CutOverTwoStep = iota
|
CutOverTwoStep
|
||||||
)
|
)
|
||||||
|
|
||||||
type ThrottleReasonHint string
|
type ThrottleReasonHint string
|
||||||
|
|
||||||
const (
|
const (
|
||||||
NoThrottleReasonHint ThrottleReasonHint = "NoThrottleReasonHint"
|
NoThrottleReasonHint ThrottleReasonHint = "NoThrottleReasonHint"
|
||||||
UserCommandThrottleReasonHint = "UserCommandThrottleReasonHint"
|
UserCommandThrottleReasonHint ThrottleReasonHint = "UserCommandThrottleReasonHint"
|
||||||
LeavingHibernationThrottleReasonHint = "LeavingHibernationThrottleReasonHint"
|
LeavingHibernationThrottleReasonHint ThrottleReasonHint = "LeavingHibernationThrottleReasonHint"
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
@ -91,6 +91,7 @@ type MigrationContext struct {
|
|||||||
SkipRenamedColumns bool
|
SkipRenamedColumns bool
|
||||||
IsTungsten bool
|
IsTungsten bool
|
||||||
DiscardForeignKeys bool
|
DiscardForeignKeys bool
|
||||||
|
AliyunRDS bool
|
||||||
|
|
||||||
config ContextConfig
|
config ContextConfig
|
||||||
configMutex *sync.Mutex
|
configMutex *sync.Mutex
|
||||||
@ -118,6 +119,8 @@ type MigrationContext struct {
|
|||||||
CriticalLoadHibernateSeconds int64
|
CriticalLoadHibernateSeconds int64
|
||||||
PostponeCutOverFlagFile string
|
PostponeCutOverFlagFile string
|
||||||
CutOverLockTimeoutSeconds int64
|
CutOverLockTimeoutSeconds int64
|
||||||
|
CutOverExponentialBackoff bool
|
||||||
|
ExponentialBackoffMaxInterval int64
|
||||||
ForceNamedCutOverCommand bool
|
ForceNamedCutOverCommand bool
|
||||||
PanicFlagFile string
|
PanicFlagFile string
|
||||||
HooksPath string
|
HooksPath string
|
||||||
@ -341,6 +344,14 @@ func (this *MigrationContext) SetCutOverLockTimeoutSeconds(timeoutSeconds int64)
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (this *MigrationContext) SetExponentialBackoffMaxInterval(intervalSeconds int64) error {
|
||||||
|
if intervalSeconds < 2 {
|
||||||
|
return fmt.Errorf("Minimal maximum interval is 2sec. Timeout remains at %d", this.ExponentialBackoffMaxInterval)
|
||||||
|
}
|
||||||
|
this.ExponentialBackoffMaxInterval = intervalSeconds
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
func (this *MigrationContext) SetDefaultNumRetries(retries int64) {
|
func (this *MigrationContext) SetDefaultNumRetries(retries int64) {
|
||||||
this.throttleMutex.Lock()
|
this.throttleMutex.Lock()
|
||||||
defer this.throttleMutex.Unlock()
|
defer this.throttleMutex.Unlock()
|
||||||
|
@ -64,17 +64,26 @@ func StringContainsAll(s string, substrings ...string) bool {
|
|||||||
return nonEmptyStringsFound
|
return nonEmptyStringsFound
|
||||||
}
|
}
|
||||||
|
|
||||||
func ValidateConnection(db *gosql.DB, connectionConfig *mysql.ConnectionConfig) (string, error) {
|
func ValidateConnection(db *gosql.DB, connectionConfig *mysql.ConnectionConfig, migrationContext *MigrationContext) (string, error) {
|
||||||
query := `select @@global.port, @@global.version`
|
versionQuery := `select @@global.version`
|
||||||
var port, extraPort int
|
var port, extraPort int
|
||||||
var version string
|
var version string
|
||||||
if err := db.QueryRow(query).Scan(&port, &version); err != nil {
|
if err := db.QueryRow(versionQuery).Scan(&version); err != nil {
|
||||||
return "", err
|
return "", err
|
||||||
}
|
}
|
||||||
extraPortQuery := `select @@global.extra_port`
|
extraPortQuery := `select @@global.extra_port`
|
||||||
if err := db.QueryRow(extraPortQuery).Scan(&extraPort); err != nil {
|
if err := db.QueryRow(extraPortQuery).Scan(&extraPort); err != nil {
|
||||||
// swallow this error. not all servers support extra_port
|
// swallow this error. not all servers support extra_port
|
||||||
}
|
}
|
||||||
|
// AliyunRDS set users port to "NULL", replace it by gh-ost param
|
||||||
|
if migrationContext.AliyunRDS {
|
||||||
|
port = connectionConfig.Key.Port
|
||||||
|
} else {
|
||||||
|
portQuery := `select @@global.port`
|
||||||
|
if err := db.QueryRow(portQuery).Scan(&port); err != nil {
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
if connectionConfig.Key.Port == port || (extraPort > 0 && connectionConfig.Key.Port == extraPort) {
|
if connectionConfig.Key.Port == port || (extraPort > 0 && connectionConfig.Key.Port == extraPort) {
|
||||||
log.Infof("connection validated on %+v", connectionConfig.Key)
|
log.Infof("connection validated on %+v", connectionConfig.Key)
|
||||||
|
@ -7,17 +7,18 @@ package binlog
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"fmt"
|
"fmt"
|
||||||
"github.com/github/gh-ost/go/sql"
|
|
||||||
"strings"
|
"strings"
|
||||||
|
|
||||||
|
"github.com/github/gh-ost/go/sql"
|
||||||
)
|
)
|
||||||
|
|
||||||
type EventDML string
|
type EventDML string
|
||||||
|
|
||||||
const (
|
const (
|
||||||
NotDML EventDML = "NoDML"
|
NotDML EventDML = "NoDML"
|
||||||
InsertDML = "Insert"
|
InsertDML EventDML = "Insert"
|
||||||
UpdateDML = "Update"
|
UpdateDML EventDML = "Update"
|
||||||
DeleteDML = "Delete"
|
DeleteDML EventDML = "Delete"
|
||||||
)
|
)
|
||||||
|
|
||||||
func ToEventDML(description string) EventDML {
|
func ToEventDML(description string) EventDML {
|
||||||
|
@ -44,7 +44,6 @@ func acceptSignals(migrationContext *base.MigrationContext) {
|
|||||||
// main is the application's entry point. It will either spawn a CLI or HTTP interfaces.
|
// main is the application's entry point. It will either spawn a CLI or HTTP interfaces.
|
||||||
func main() {
|
func main() {
|
||||||
migrationContext := base.NewMigrationContext()
|
migrationContext := base.NewMigrationContext()
|
||||||
|
|
||||||
flag.StringVar(&migrationContext.InspectorConnectionConfig.Key.Hostname, "host", "127.0.0.1", "MySQL hostname (preferably a replica, not the master)")
|
flag.StringVar(&migrationContext.InspectorConnectionConfig.Key.Hostname, "host", "127.0.0.1", "MySQL hostname (preferably a replica, not the master)")
|
||||||
flag.StringVar(&migrationContext.AssumeMasterHostname, "assume-master-host", "", "(optional) explicitly tell gh-ost the identity of the master. Format: some.host.com[:port] This is useful in master-master setups where you wish to pick an explicit master, or in a tungsten-replicator where gh-ost is unable to determine the master")
|
flag.StringVar(&migrationContext.AssumeMasterHostname, "assume-master-host", "", "(optional) explicitly tell gh-ost the identity of the master. Format: some.host.com[:port] This is useful in master-master setups where you wish to pick an explicit master, or in a tungsten-replicator where gh-ost is unable to determine the master")
|
||||||
flag.IntVar(&migrationContext.InspectorConnectionConfig.Key.Port, "port", 3306, "MySQL port (preferably a replica, not the master)")
|
flag.IntVar(&migrationContext.InspectorConnectionConfig.Key.Port, "port", 3306, "MySQL port (preferably a replica, not the master)")
|
||||||
@ -68,6 +67,7 @@ func main() {
|
|||||||
flag.BoolVar(&migrationContext.IsTungsten, "tungsten", false, "explicitly let gh-ost know that you are running on a tungsten-replication based topology (you are likely to also provide --assume-master-host)")
|
flag.BoolVar(&migrationContext.IsTungsten, "tungsten", false, "explicitly let gh-ost know that you are running on a tungsten-replication based topology (you are likely to also provide --assume-master-host)")
|
||||||
flag.BoolVar(&migrationContext.DiscardForeignKeys, "discard-foreign-keys", false, "DANGER! This flag will migrate a table that has foreign keys and will NOT create foreign keys on the ghost table, thus your altered table will have NO foreign keys. This is useful for intentional dropping of foreign keys")
|
flag.BoolVar(&migrationContext.DiscardForeignKeys, "discard-foreign-keys", false, "DANGER! This flag will migrate a table that has foreign keys and will NOT create foreign keys on the ghost table, thus your altered table will have NO foreign keys. This is useful for intentional dropping of foreign keys")
|
||||||
flag.BoolVar(&migrationContext.SkipForeignKeyChecks, "skip-foreign-key-checks", false, "set to 'true' when you know for certain there are no foreign keys on your table, and wish to skip the time it takes for gh-ost to verify that")
|
flag.BoolVar(&migrationContext.SkipForeignKeyChecks, "skip-foreign-key-checks", false, "set to 'true' when you know for certain there are no foreign keys on your table, and wish to skip the time it takes for gh-ost to verify that")
|
||||||
|
flag.BoolVar(&migrationContext.AliyunRDS, "aliyun-rds", false, "set to 'true' when you execute on Aliyun RDS.")
|
||||||
|
|
||||||
executeFlag := flag.Bool("execute", false, "actually execute the alter & migrate the table. Default is noop: do some tests and exit")
|
executeFlag := flag.Bool("execute", false, "actually execute the alter & migrate the table. Default is noop: do some tests and exit")
|
||||||
flag.BoolVar(&migrationContext.TestOnReplica, "test-on-replica", false, "Have the migration run on a replica, not on the master. At the end of migration replication is stopped, and tables are swapped and immediately swap-revert. Replication remains stopped and you can compare the two tables for building trust")
|
flag.BoolVar(&migrationContext.TestOnReplica, "test-on-replica", false, "Have the migration run on a replica, not on the master. At the end of migration replication is stopped, and tables are swapped and immediately swap-revert. Replication remains stopped and you can compare the two tables for building trust")
|
||||||
@ -83,6 +83,8 @@ func main() {
|
|||||||
|
|
||||||
flag.BoolVar(&migrationContext.SwitchToRowBinlogFormat, "switch-to-rbr", false, "let this tool automatically switch binary log format to 'ROW' on the replica, if needed. The format will NOT be switched back. I'm too scared to do that, and wish to protect you if you happen to execute another migration while this one is running")
|
flag.BoolVar(&migrationContext.SwitchToRowBinlogFormat, "switch-to-rbr", false, "let this tool automatically switch binary log format to 'ROW' on the replica, if needed. The format will NOT be switched back. I'm too scared to do that, and wish to protect you if you happen to execute another migration while this one is running")
|
||||||
flag.BoolVar(&migrationContext.AssumeRBR, "assume-rbr", false, "set to 'true' when you know for certain your server uses 'ROW' binlog_format. gh-ost is unable to tell, event after reading binlog_format, whether the replication process does indeed use 'ROW', and restarts replication to be certain RBR setting is applied. Such operation requires SUPER privileges which you might not have. Setting this flag avoids restarting replication and you can proceed to use gh-ost without SUPER privileges")
|
flag.BoolVar(&migrationContext.AssumeRBR, "assume-rbr", false, "set to 'true' when you know for certain your server uses 'ROW' binlog_format. gh-ost is unable to tell, event after reading binlog_format, whether the replication process does indeed use 'ROW', and restarts replication to be certain RBR setting is applied. Such operation requires SUPER privileges which you might not have. Setting this flag avoids restarting replication and you can proceed to use gh-ost without SUPER privileges")
|
||||||
|
flag.BoolVar(&migrationContext.CutOverExponentialBackoff, "cut-over-exponential-backoff", false, "Wait exponentially longer intervals between failed cut-over attempts. Wait intervals obey a maximum configurable with 'exponential-backoff-max-interval').")
|
||||||
|
exponentialBackoffMaxInterval := flag.Int64("exponential-backoff-max-interval", 64, "Maximum number of seconds to wait between attempts when performing various operations with exponential backoff.")
|
||||||
chunkSize := flag.Int64("chunk-size", 1000, "amount of rows to handle in each iteration (allowed range: 100-100,000)")
|
chunkSize := flag.Int64("chunk-size", 1000, "amount of rows to handle in each iteration (allowed range: 100-100,000)")
|
||||||
dmlBatchSize := flag.Int64("dml-batch-size", 10, "batch size for DML events to apply in a single transaction (range 1-100)")
|
dmlBatchSize := flag.Int64("dml-batch-size", 10, "batch size for DML events to apply in a single transaction (range 1-100)")
|
||||||
defaultRetries := flag.Int64("default-retries", 60, "Default number of retries for various operations before panicking")
|
defaultRetries := flag.Int64("default-retries", 60, "Default number of retries for various operations before panicking")
|
||||||
@ -238,6 +240,9 @@ func main() {
|
|||||||
if err := migrationContext.SetCutOverLockTimeoutSeconds(*cutOverLockTimeoutSeconds); err != nil {
|
if err := migrationContext.SetCutOverLockTimeoutSeconds(*cutOverLockTimeoutSeconds); err != nil {
|
||||||
log.Errore(err)
|
log.Errore(err)
|
||||||
}
|
}
|
||||||
|
if err := migrationContext.SetExponentialBackoffMaxInterval(*exponentialBackoffMaxInterval); err != nil {
|
||||||
|
log.Errore(err)
|
||||||
|
}
|
||||||
|
|
||||||
log.Infof("starting gh-ost %+v", AppVersion)
|
log.Infof("starting gh-ost %+v", AppVersion)
|
||||||
acceptSignals(migrationContext)
|
acceptSignals(migrationContext)
|
||||||
|
@ -78,21 +78,23 @@ func (this *Applier) InitDBConnections() (err error) {
|
|||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
this.singletonDB.SetMaxOpenConns(1)
|
this.singletonDB.SetMaxOpenConns(1)
|
||||||
version, err := base.ValidateConnection(this.db, this.connectionConfig)
|
version, err := base.ValidateConnection(this.db, this.connectionConfig, this.migrationContext)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
if _, err := base.ValidateConnection(this.singletonDB, this.connectionConfig); err != nil {
|
if _, err := base.ValidateConnection(this.singletonDB, this.connectionConfig, this.migrationContext); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
this.migrationContext.ApplierMySQLVersion = version
|
this.migrationContext.ApplierMySQLVersion = version
|
||||||
if err := this.validateAndReadTimeZone(); err != nil {
|
if err := this.validateAndReadTimeZone(); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
if impliedKey, err := mysql.GetInstanceKey(this.db); err != nil {
|
if !this.migrationContext.AliyunRDS {
|
||||||
return err
|
if impliedKey, err := mysql.GetInstanceKey(this.db); err != nil {
|
||||||
} else {
|
return err
|
||||||
this.connectionConfig.ImpliedKey = impliedKey
|
} else {
|
||||||
|
this.connectionConfig.ImpliedKey = impliedKey
|
||||||
|
}
|
||||||
}
|
}
|
||||||
if err := this.readTableColumns(); err != nil {
|
if err := this.readTableColumns(); err != nil {
|
||||||
return err
|
return err
|
||||||
|
@ -53,10 +53,12 @@ func (this *Inspector) InitDBConnections() (err error) {
|
|||||||
if err := this.validateConnection(); err != nil {
|
if err := this.validateConnection(); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
if impliedKey, err := mysql.GetInstanceKey(this.db); err != nil {
|
if !this.migrationContext.AliyunRDS {
|
||||||
return err
|
if impliedKey, err := mysql.GetInstanceKey(this.db); err != nil {
|
||||||
} else {
|
return err
|
||||||
this.connectionConfig.ImpliedKey = impliedKey
|
} else {
|
||||||
|
this.connectionConfig.ImpliedKey = impliedKey
|
||||||
|
}
|
||||||
}
|
}
|
||||||
if err := this.validateGrants(); err != nil {
|
if err := this.validateGrants(); err != nil {
|
||||||
return err
|
return err
|
||||||
@ -163,11 +165,6 @@ func (this *Inspector) inspectOriginalAndGhostTables() (err error) {
|
|||||||
return fmt.Errorf("Chosen key (%s) has nullable columns. Bailing out. To force this operation to continue, supply --allow-nullable-unique-key flag. Only do so if you are certain there are no actual NULL values in this key. As long as there aren't, migration should be fine. NULL values in columns of this key will corrupt migration's data", this.migrationContext.UniqueKey)
|
return fmt.Errorf("Chosen key (%s) has nullable columns. Bailing out. To force this operation to continue, supply --allow-nullable-unique-key flag. Only do so if you are certain there are no actual NULL values in this key. As long as there aren't, migration should be fine. NULL values in columns of this key will corrupt migration's data", this.migrationContext.UniqueKey)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if !this.migrationContext.UniqueKey.IsPrimary() {
|
|
||||||
if this.migrationContext.OriginalBinlogRowImage != "FULL" {
|
|
||||||
return fmt.Errorf("binlog_row_image is '%s' and chosen key is %s, which is not the primary key. This operation cannot proceed. You may `set global binlog_row_image='full'` and try again", this.migrationContext.OriginalBinlogRowImage, this.migrationContext.UniqueKey)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
this.migrationContext.SharedColumns, this.migrationContext.MappedSharedColumns = this.getSharedColumns(this.migrationContext.OriginalTableColumns, this.migrationContext.GhostTableColumns, this.migrationContext.ColumnRenameMap)
|
this.migrationContext.SharedColumns, this.migrationContext.MappedSharedColumns = this.getSharedColumns(this.migrationContext.OriginalTableColumns, this.migrationContext.GhostTableColumns, this.migrationContext.ColumnRenameMap)
|
||||||
log.Infof("Shared columns are %s", this.migrationContext.SharedColumns)
|
log.Infof("Shared columns are %s", this.migrationContext.SharedColumns)
|
||||||
@ -203,7 +200,7 @@ func (this *Inspector) validateConnection() error {
|
|||||||
return fmt.Errorf("MySQL replication length limited to 32 characters. See https://dev.mysql.com/doc/refman/5.7/en/assigning-passwords.html")
|
return fmt.Errorf("MySQL replication length limited to 32 characters. See https://dev.mysql.com/doc/refman/5.7/en/assigning-passwords.html")
|
||||||
}
|
}
|
||||||
|
|
||||||
version, err := base.ValidateConnection(this.db, this.connectionConfig)
|
version, err := base.ValidateConnection(this.db, this.connectionConfig, this.migrationContext)
|
||||||
this.migrationContext.InspectorMySQLVersion = version
|
this.migrationContext.InspectorMySQLVersion = version
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
@ -356,6 +353,9 @@ func (this *Inspector) validateBinlogs() error {
|
|||||||
this.migrationContext.OriginalBinlogRowImage = "FULL"
|
this.migrationContext.OriginalBinlogRowImage = "FULL"
|
||||||
}
|
}
|
||||||
this.migrationContext.OriginalBinlogRowImage = strings.ToUpper(this.migrationContext.OriginalBinlogRowImage)
|
this.migrationContext.OriginalBinlogRowImage = strings.ToUpper(this.migrationContext.OriginalBinlogRowImage)
|
||||||
|
if this.migrationContext.OriginalBinlogRowImage != "FULL" {
|
||||||
|
return fmt.Errorf("%s:%d has '%s' binlog_row_image, and only 'FULL' is supported. This operation cannot proceed. You may `set global binlog_row_image='full'` and try again", this.connectionConfig.Key.Hostname, this.connectionConfig.Key.Port, this.migrationContext.OriginalBinlogRowImage)
|
||||||
|
}
|
||||||
|
|
||||||
log.Infof("binary logs validated on %s:%d", this.connectionConfig.Key.Hostname, this.connectionConfig.Key.Port)
|
log.Infof("binary logs validated on %s:%d", this.connectionConfig.Key.Hostname, this.connectionConfig.Key.Port)
|
||||||
return nil
|
return nil
|
||||||
|
@ -149,6 +149,34 @@ func (this *Migrator) retryOperation(operation func() error, notFatalHint ...boo
|
|||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// `retryOperationWithExponentialBackoff` attempts running given function, waiting 2^(n-1)
|
||||||
|
// seconds between each attempt, where `n` is the running number of attempts. Exits
|
||||||
|
// as soon as the function returns with non-error, or as soon as `MaxRetries`
|
||||||
|
// attempts are reached. Wait intervals between attempts obey a maximum of
|
||||||
|
// `ExponentialBackoffMaxInterval`.
|
||||||
|
func (this *Migrator) retryOperationWithExponentialBackoff(operation func() error, notFatalHint ...bool) (err error) {
|
||||||
|
var interval int64
|
||||||
|
maxRetries := int(this.migrationContext.MaxRetries())
|
||||||
|
maxInterval := this.migrationContext.ExponentialBackoffMaxInterval
|
||||||
|
for i := 0; i < maxRetries; i++ {
|
||||||
|
newInterval := int64(math.Exp2(float64(i - 1)))
|
||||||
|
if newInterval <= maxInterval {
|
||||||
|
interval = newInterval
|
||||||
|
}
|
||||||
|
if i != 0 {
|
||||||
|
time.Sleep(time.Duration(interval) * time.Second)
|
||||||
|
}
|
||||||
|
err = operation()
|
||||||
|
if err == nil {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if len(notFatalHint) == 0 {
|
||||||
|
this.migrationContext.PanicAbort <- err
|
||||||
|
}
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
// executeAndThrottleOnError executes a given function. If it errors, it
|
// executeAndThrottleOnError executes a given function. If it errors, it
|
||||||
// throttles.
|
// throttles.
|
||||||
func (this *Migrator) executeAndThrottleOnError(operation func() error) (err error) {
|
func (this *Migrator) executeAndThrottleOnError(operation func() error) (err error) {
|
||||||
@ -227,7 +255,11 @@ func (this *Migrator) listenOnPanicAbort() {
|
|||||||
// validateStatement validates the `alter` statement meets criteria.
|
// validateStatement validates the `alter` statement meets criteria.
|
||||||
// At this time this means:
|
// At this time this means:
|
||||||
// - column renames are approved
|
// - column renames are approved
|
||||||
|
// - no table rename allowed
|
||||||
func (this *Migrator) validateStatement() (err error) {
|
func (this *Migrator) validateStatement() (err error) {
|
||||||
|
if this.parser.IsRenameTable() {
|
||||||
|
return fmt.Errorf("ALTER statement seems to RENAME the table. This is not supported, and you should run your RENAME outside gh-ost.")
|
||||||
|
}
|
||||||
if this.parser.HasNonTrivialRenames() && !this.migrationContext.SkipRenamedColumns {
|
if this.parser.HasNonTrivialRenames() && !this.migrationContext.SkipRenamedColumns {
|
||||||
this.migrationContext.ColumnRenameMap = this.parser.GetNonTrivialRenames()
|
this.migrationContext.ColumnRenameMap = this.parser.GetNonTrivialRenames()
|
||||||
if !this.migrationContext.ApproveRenamedColumns {
|
if !this.migrationContext.ApproveRenamedColumns {
|
||||||
@ -372,7 +404,13 @@ func (this *Migrator) Migrate() (err error) {
|
|||||||
if err := this.hooksExecutor.onBeforeCutOver(); err != nil {
|
if err := this.hooksExecutor.onBeforeCutOver(); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
if err := this.retryOperation(this.cutOver); err != nil {
|
var retrier func(func() error, ...bool) error
|
||||||
|
if this.migrationContext.CutOverExponentialBackoff {
|
||||||
|
retrier = this.retryOperationWithExponentialBackoff
|
||||||
|
} else {
|
||||||
|
retrier = this.retryOperation
|
||||||
|
}
|
||||||
|
if err := retrier(this.cutOver); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
atomic.StoreInt64(&this.migrationContext.CutOverCompleteFlag, 1)
|
atomic.StoreInt64(&this.migrationContext.CutOverCompleteFlag, 1)
|
||||||
|
@ -130,6 +130,9 @@ func (this *Server) applyServerCommand(command string, writer *bufio.Writer) (pr
|
|||||||
arg := ""
|
arg := ""
|
||||||
if len(tokens) > 1 {
|
if len(tokens) > 1 {
|
||||||
arg = strings.TrimSpace(tokens[1])
|
arg = strings.TrimSpace(tokens[1])
|
||||||
|
if unquoted, err := strconv.Unquote(arg); err == nil {
|
||||||
|
arg = unquoted
|
||||||
|
}
|
||||||
}
|
}
|
||||||
argIsQuestion := (arg == "?")
|
argIsQuestion := (arg == "?")
|
||||||
throttleHint := "# Note: you may only throttle for as long as your binary logs are not purged\n"
|
throttleHint := "# Note: you may only throttle for as long as your binary logs are not purged\n"
|
||||||
|
@ -107,7 +107,7 @@ func (this *EventsStreamer) InitDBConnections() (err error) {
|
|||||||
if this.db, _, err = mysql.GetDB(this.migrationContext.Uuid, EventsStreamerUri); err != nil {
|
if this.db, _, err = mysql.GetDB(this.migrationContext.Uuid, EventsStreamerUri); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
if _, err := base.ValidateConnection(this.db, this.connectionConfig); err != nil {
|
if _, err := base.ValidateConnection(this.db, this.connectionConfig, this.migrationContext); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
if err := this.readCurrentBinlogCoordinates(); err != nil {
|
if err := this.readCurrentBinlogCoordinates(); err != nil {
|
||||||
|
@ -56,5 +56,6 @@ func (this *ConnectionConfig) GetDBUri(databaseName string) string {
|
|||||||
// Wrap IPv6 literals in square brackets
|
// Wrap IPv6 literals in square brackets
|
||||||
hostname = fmt.Sprintf("[%s]", hostname)
|
hostname = fmt.Sprintf("[%s]", hostname)
|
||||||
}
|
}
|
||||||
return fmt.Sprintf("%s:%s@tcp(%s:%d)/%s?interpolateParams=true&autocommit=true&charset=utf8mb4,utf8,latin1", this.User, this.Password, hostname, this.Key.Port, databaseName)
|
interpolateParams := true
|
||||||
|
return fmt.Sprintf("%s:%s@tcp(%s:%d)/%s?interpolateParams=%t&autocommit=true&charset=utf8mb4,utf8,latin1", this.User, this.Password, hostname, this.Key.Port, databaseName, interpolateParams)
|
||||||
}
|
}
|
||||||
|
@ -55,3 +55,13 @@ func TestDuplicate(t *testing.T) {
|
|||||||
test.S(t).ExpectEquals(dup.User, "gromit")
|
test.S(t).ExpectEquals(dup.User, "gromit")
|
||||||
test.S(t).ExpectEquals(dup.Password, "penguin")
|
test.S(t).ExpectEquals(dup.Password, "penguin")
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func TestGetDBUri(t *testing.T) {
|
||||||
|
c := NewConnectionConfig()
|
||||||
|
c.Key = InstanceKey{Hostname: "myhost", Port: 3306}
|
||||||
|
c.User = "gromit"
|
||||||
|
c.Password = "penguin"
|
||||||
|
|
||||||
|
uri := c.GetDBUri("test")
|
||||||
|
test.S(t).ExpectEquals(uri, "gromit:penguin@tcp(myhost:3306)/test?interpolateParams=true&autocommit=true&charset=utf8mb4,utf8,latin1")
|
||||||
|
}
|
||||||
|
@ -15,11 +15,11 @@ type ValueComparisonSign string
|
|||||||
|
|
||||||
const (
|
const (
|
||||||
LessThanComparisonSign ValueComparisonSign = "<"
|
LessThanComparisonSign ValueComparisonSign = "<"
|
||||||
LessThanOrEqualsComparisonSign = "<="
|
LessThanOrEqualsComparisonSign ValueComparisonSign = "<="
|
||||||
EqualsComparisonSign = "="
|
EqualsComparisonSign ValueComparisonSign = "="
|
||||||
GreaterThanOrEqualsComparisonSign = ">="
|
GreaterThanOrEqualsComparisonSign ValueComparisonSign = ">="
|
||||||
GreaterThanComparisonSign = ">"
|
GreaterThanComparisonSign ValueComparisonSign = ">"
|
||||||
NotEqualsComparisonSign = "!="
|
NotEqualsComparisonSign ValueComparisonSign = "!="
|
||||||
)
|
)
|
||||||
|
|
||||||
// EscapeName will escape a db/table/column/... name by wrapping with backticks.
|
// EscapeName will escape a db/table/column/... name by wrapping with backticks.
|
||||||
|
@ -8,6 +8,7 @@ package sql
|
|||||||
import (
|
import (
|
||||||
"golang.org/x/text/encoding"
|
"golang.org/x/text/encoding"
|
||||||
"golang.org/x/text/encoding/charmap"
|
"golang.org/x/text/encoding/charmap"
|
||||||
|
"golang.org/x/text/encoding/simplifiedchinese"
|
||||||
)
|
)
|
||||||
|
|
||||||
type charsetEncoding map[string]encoding.Encoding
|
type charsetEncoding map[string]encoding.Encoding
|
||||||
@ -18,4 +19,5 @@ func init() {
|
|||||||
charsetEncodingMap = make(map[string]encoding.Encoding)
|
charsetEncodingMap = make(map[string]encoding.Encoding)
|
||||||
// Begin mappings
|
// Begin mappings
|
||||||
charsetEncodingMap["latin1"] = charmap.Windows1252
|
charsetEncodingMap["latin1"] = charmap.Windows1252
|
||||||
|
charsetEncodingMap["gbk"] = simplifiedchinese.GBK
|
||||||
}
|
}
|
||||||
|
@ -15,11 +15,13 @@ var (
|
|||||||
sanitizeQuotesRegexp = regexp.MustCompile("('[^']*')")
|
sanitizeQuotesRegexp = regexp.MustCompile("('[^']*')")
|
||||||
renameColumnRegexp = regexp.MustCompile(`(?i)\bchange\s+(column\s+|)([\S]+)\s+([\S]+)\s+`)
|
renameColumnRegexp = regexp.MustCompile(`(?i)\bchange\s+(column\s+|)([\S]+)\s+([\S]+)\s+`)
|
||||||
dropColumnRegexp = regexp.MustCompile(`(?i)\bdrop\s+(column\s+|)([\S]+)$`)
|
dropColumnRegexp = regexp.MustCompile(`(?i)\bdrop\s+(column\s+|)([\S]+)$`)
|
||||||
|
renameTableRegexp = regexp.MustCompile(`(?i)\brename\s+(to|as)\s+`)
|
||||||
)
|
)
|
||||||
|
|
||||||
type Parser struct {
|
type Parser struct {
|
||||||
columnRenameMap map[string]string
|
columnRenameMap map[string]string
|
||||||
droppedColumns map[string]bool
|
droppedColumns map[string]bool
|
||||||
|
isRenameTable bool
|
||||||
}
|
}
|
||||||
|
|
||||||
func NewParser() *Parser {
|
func NewParser() *Parser {
|
||||||
@ -86,6 +88,12 @@ func (this *Parser) parseAlterToken(alterToken string) (err error) {
|
|||||||
this.droppedColumns[submatch[2]] = true
|
this.droppedColumns[submatch[2]] = true
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
{
|
||||||
|
// rename table
|
||||||
|
if renameTableRegexp.MatchString(alterToken) {
|
||||||
|
this.isRenameTable = true
|
||||||
|
}
|
||||||
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -115,3 +123,7 @@ func (this *Parser) HasNonTrivialRenames() bool {
|
|||||||
func (this *Parser) DroppedColumnsMap() map[string]bool {
|
func (this *Parser) DroppedColumnsMap() map[string]bool {
|
||||||
return this.droppedColumns
|
return this.droppedColumns
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (this *Parser) IsRenameTable() bool {
|
||||||
|
return this.isRenameTable
|
||||||
|
}
|
||||||
|
@ -159,3 +159,42 @@ func TestParseAlterStatementDroppedColumns(t *testing.T) {
|
|||||||
test.S(t).ExpectTrue(parser.droppedColumns["b"])
|
test.S(t).ExpectTrue(parser.droppedColumns["b"])
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func TestParseAlterStatementRenameTable(t *testing.T) {
|
||||||
|
|
||||||
|
{
|
||||||
|
parser := NewParser()
|
||||||
|
statement := "drop column b"
|
||||||
|
err := parser.ParseAlterStatement(statement)
|
||||||
|
test.S(t).ExpectNil(err)
|
||||||
|
test.S(t).ExpectFalse(parser.isRenameTable)
|
||||||
|
}
|
||||||
|
{
|
||||||
|
parser := NewParser()
|
||||||
|
statement := "rename as something_else"
|
||||||
|
err := parser.ParseAlterStatement(statement)
|
||||||
|
test.S(t).ExpectNil(err)
|
||||||
|
test.S(t).ExpectTrue(parser.isRenameTable)
|
||||||
|
}
|
||||||
|
{
|
||||||
|
parser := NewParser()
|
||||||
|
statement := "drop column b, rename as something_else"
|
||||||
|
err := parser.ParseAlterStatement(statement)
|
||||||
|
test.S(t).ExpectNil(err)
|
||||||
|
test.S(t).ExpectTrue(parser.isRenameTable)
|
||||||
|
}
|
||||||
|
{
|
||||||
|
parser := NewParser()
|
||||||
|
statement := "engine=innodb rename as something_else"
|
||||||
|
err := parser.ParseAlterStatement(statement)
|
||||||
|
test.S(t).ExpectNil(err)
|
||||||
|
test.S(t).ExpectTrue(parser.isRenameTable)
|
||||||
|
}
|
||||||
|
{
|
||||||
|
parser := NewParser()
|
||||||
|
statement := "rename as something_else, engine=innodb"
|
||||||
|
err := parser.ParseAlterStatement(statement)
|
||||||
|
test.S(t).ExpectNil(err)
|
||||||
|
test.S(t).ExpectTrue(parser.isRenameTable)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
@ -15,13 +15,13 @@ import (
|
|||||||
type ColumnType int
|
type ColumnType int
|
||||||
|
|
||||||
const (
|
const (
|
||||||
UnknownColumnType ColumnType = iota
|
UnknownColumnType ColumnType = iota
|
||||||
TimestampColumnType = iota
|
TimestampColumnType
|
||||||
DateTimeColumnType = iota
|
DateTimeColumnType
|
||||||
EnumColumnType = iota
|
EnumColumnType
|
||||||
MediumIntColumnType = iota
|
MediumIntColumnType
|
||||||
JSONColumnType = iota
|
JSONColumnType
|
||||||
FloatColumnType = iota
|
FloatColumnType
|
||||||
)
|
)
|
||||||
|
|
||||||
const maxMediumintUnsigned int32 = 16777215
|
const maxMediumintUnsigned int32 = 16777215
|
||||||
|
@ -0,0 +1 @@
|
|||||||
|
(5.5)
|
1
localtests/datetime-submillis/ignore_versions
Normal file
1
localtests/datetime-submillis/ignore_versions
Normal file
@ -0,0 +1 @@
|
|||||||
|
(5.5)
|
1
localtests/datetime-to-timestamp-pk-fail/ignore_versions
Normal file
1
localtests/datetime-to-timestamp-pk-fail/ignore_versions
Normal file
@ -0,0 +1 @@
|
|||||||
|
(5.5)
|
1
localtests/datetime/ignore_versions
Normal file
1
localtests/datetime/ignore_versions
Normal file
@ -0,0 +1 @@
|
|||||||
|
(5.5)
|
22
localtests/fail-rename-table/create.sql
Normal file
22
localtests/fail-rename-table/create.sql
Normal file
@ -0,0 +1,22 @@
|
|||||||
|
drop table if exists gh_ost_test;
|
||||||
|
create table gh_ost_test (
|
||||||
|
id int auto_increment,
|
||||||
|
i int not null,
|
||||||
|
ts timestamp,
|
||||||
|
primary key(id)
|
||||||
|
) auto_increment=1;
|
||||||
|
|
||||||
|
drop event if exists gh_ost_test;
|
||||||
|
delimiter ;;
|
||||||
|
create event gh_ost_test
|
||||||
|
on schedule every 1 second
|
||||||
|
starts current_timestamp
|
||||||
|
ends current_timestamp + interval 60 second
|
||||||
|
on completion not preserve
|
||||||
|
enable
|
||||||
|
do
|
||||||
|
begin
|
||||||
|
insert into gh_ost_test values (null, 11, now());
|
||||||
|
insert into gh_ost_test values (null, 13, now());
|
||||||
|
insert into gh_ost_test values (null, 17, now());
|
||||||
|
end ;;
|
1
localtests/fail-rename-table/expect_failure
Normal file
1
localtests/fail-rename-table/expect_failure
Normal file
@ -0,0 +1 @@
|
|||||||
|
ALTER statement seems to RENAME the table
|
1
localtests/fail-rename-table/extra_args
Normal file
1
localtests/fail-rename-table/extra_args
Normal file
@ -0,0 +1 @@
|
|||||||
|
--alter="rename as something_else"
|
25
localtests/gbk-charset/create.sql
Normal file
25
localtests/gbk-charset/create.sql
Normal file
@ -0,0 +1,25 @@
|
|||||||
|
drop table if exists gh_ost_test;
|
||||||
|
create table gh_ost_test (
|
||||||
|
id int(11) NOT NULL AUTO_INCREMENT,
|
||||||
|
name varchar(512) DEFAULT NULL,
|
||||||
|
v varchar(255) DEFAULT NULL COMMENT '添加普通列测试',
|
||||||
|
PRIMARY KEY (id)
|
||||||
|
) ENGINE=InnoDB AUTO_INCREMENT=1 DEFAULT CHARSET=gbk;
|
||||||
|
|
||||||
|
insert into gh_ost_test values (null, 'gbk-test-initial', '添加普通列测试-添加普通列测试');
|
||||||
|
insert into gh_ost_test values (null, 'gbk-test-initial', '添加普通列测试-添加普通列测试');
|
||||||
|
|
||||||
|
drop event if exists gh_ost_test;
|
||||||
|
delimiter ;;
|
||||||
|
create event gh_ost_test
|
||||||
|
on schedule every 1 second
|
||||||
|
starts current_timestamp
|
||||||
|
ends current_timestamp + interval 60 second
|
||||||
|
on completion not preserve
|
||||||
|
enable
|
||||||
|
do
|
||||||
|
begin
|
||||||
|
insert into gh_ost_test (name) values ('gbk-test-default');
|
||||||
|
insert into gh_ost_test values (null, 'gbk-test', '添加普通列测试-添加普通列测试');
|
||||||
|
update gh_ost_test set v='添加普通列测试' where v='添加普通列测试-添加普通列测试' order by id desc limit 1;
|
||||||
|
end ;;
|
0
localtests/gbk-charset/extra_args
Normal file
0
localtests/gbk-charset/extra_args
Normal file
21
localtests/geometry57/create.sql
Normal file
21
localtests/geometry57/create.sql
Normal file
@ -0,0 +1,21 @@
|
|||||||
|
drop table if exists gh_ost_test;
|
||||||
|
create table gh_ost_test (
|
||||||
|
id int auto_increment,
|
||||||
|
g geometry,
|
||||||
|
primary key(id)
|
||||||
|
) auto_increment=1;
|
||||||
|
|
||||||
|
drop event if exists gh_ost_test;
|
||||||
|
delimiter ;;
|
||||||
|
create event gh_ost_test
|
||||||
|
on schedule every 1 second
|
||||||
|
starts current_timestamp
|
||||||
|
ends current_timestamp + interval 60 second
|
||||||
|
on completion not preserve
|
||||||
|
enable
|
||||||
|
do
|
||||||
|
begin
|
||||||
|
insert into gh_ost_test values (null, ST_GeomFromText('POINT(1 1)'));
|
||||||
|
insert into gh_ost_test values (null, ST_GeomFromText('POINT(2 2)'));
|
||||||
|
insert into gh_ost_test values (null, ST_GeomFromText('POINT(3 3)'));
|
||||||
|
end ;;
|
1
localtests/geometry57/ignore_versions
Normal file
1
localtests/geometry57/ignore_versions
Normal file
@ -0,0 +1 @@
|
|||||||
|
(5.5|5.6)
|
1
localtests/json57/ignore_versions
Normal file
1
localtests/json57/ignore_versions
Normal file
@ -0,0 +1 @@
|
|||||||
|
(5.5|5.6)
|
@ -19,7 +19,8 @@ create event gh_ost_test
|
|||||||
begin
|
begin
|
||||||
insert into gh_ost_test (id, i, j) values (null, 11, '"sometext"');
|
insert into gh_ost_test (id, i, j) values (null, 11, '"sometext"');
|
||||||
insert into gh_ost_test (id, i, j) values (null, 13, '{"key":"val"}');
|
insert into gh_ost_test (id, i, j) values (null, 13, '{"key":"val"}');
|
||||||
insert into gh_ost_test (id, i, j) values (null, 17, '{"is-it": true, "count": 3, "elements": []}');
|
insert into gh_ost_test (id, i, j) values (null, 17, '{"is-it": true, "count": 3, "elements": []}');
|
||||||
|
insert into gh_ost_test (id, i, j) values (null, 19, '{"text":"Lorem ipsum dolor sit amet, consectetuer adipiscing elit. Aenean commodo ligula eget dolor. Aenean massa. Cum sociis natoque penatibus et magnis dis parturient montes, nascetur ridiculus mus. Donec quam felis, ultricies nec, pellentesque eu, pretium quis, sem. Nulla consequat massa quis enim. Donec pede justo, fringilla vel, aliquet nec, vulputate eget, arcu. In enim justo, rhoncus ut, imperdiet a, venenatis vitae, justo. Nullam dictum felis eu pede mollis pretium. Integer tincidunt. Cras dapibus. Vivamus elementum semper nisi. Aenean vulputate eleifend tellus. Aenean leo ligula, porttitor eu, consequat vitae, eleifend ac, enim. Aliquam lorem ante, dapibus in, viverra quis, feugiat a, tellus. Phasellus viverra nulla ut metus varius laoreet. Quisque rutrum. Aenean imperdiet. Etiam ultricies nisi vel augue. Curabitur ullamcorper ultricies nisi. Nam eget dui. Etiam rhoncus. Maecenas tempus, tellus eget condimentum rhoncus, sem quam semper libero, sit amet adipiscing sem neque sed ipsum. Nam quam nunc, blandit vel, luctus pulvinar, hendrerit id, lorem. Maecenas nec odio et ante tincidunt tempus. Donec vitae sapien ut libero venenatis faucibus. Nullam quis ante. Etiam sit amet orci eget eros faucibus tincidunt. Duis leo. Sed fringilla mauris sit amet nibh. Donec sodales sagittis magna. Sed consequat, leo eget bibendum sodales, augue velit cursus nunc, quis gravida magna mi a libero. Fusce vulputate eleifend sapien. Vestibulum purus quam, scelerisque ut, mollis sed, nonummy id, metus. Nullam accumsan lorem in dui. Cras ultricies mi eu turpis hendrerit fringilla. Vestibulum ante ipsum primis in faucibus orci luctus et ultrices posuere cubilia Curae; In ac dui quis mi consectetuer lacinia. Nam pretium turpis et arcu. Duis arcu tortor, suscipit eget, imperdiet nec, imperdiet iaculis, ipsum. Sed aliquam ultrices mauris. Integer ante arcu, accumsan a, consectetuer eget, posuere ut, mauris. Praesent adipiscing. Phasellus ullamcorper ipsum rutrum nunc. Nunc nonummy metus. Vestibulum volutpat pretium libero. Cras id dui. Aenean ut eros et nisl sagittis vestibulum. Nullam nulla eros, ultricies sit amet, nonummy id, imperdiet feugiat, pede. Sed lectus. Donec mollis hendrerit risus. Phasellus nec sem in justo pellentesque facilisis. Etiam imperdiet imperdiet orci. Nunc nec neque. Phasellus leo dolor, tempus non, auctor et, hendrerit quis, nisi. Curabitur ligula sapien, tincidunt non, euismod vitae, posuere imperdiet, leo. Maecenas malesuada. Praesent congue erat at massa. Sed cursus turpis vitae tortor. Donec posuere vulputate arcu. Phasellus accumsan cursus velit. Vestibulum ante ipsum primis in faucibus orci luctus et ultrices posuere cubilia Curae; Sed aliquam, nisi quis porttitor congue, elit erat euismod orci, ac"}');
|
||||||
|
|
||||||
update gh_ost_test set j = '{"updated": 11}', updated = 1 where i = 11 and updated = 0;
|
update gh_ost_test set j = '{"updated": 11}', updated = 1 where i = 11 and updated = 0;
|
||||||
update gh_ost_test set j = json_set(j, '$.count', 13, '$.id', id), updated = 1 where i = 13 and updated = 0;
|
update gh_ost_test set j = json_set(j, '$.count', 13, '$.id', id), updated = 1 where i = 13 and updated = 0;
|
||||||
|
1
localtests/json57dml/ignore_versions
Normal file
1
localtests/json57dml/ignore_versions
Normal file
@ -0,0 +1 @@
|
|||||||
|
(5.5|5.6)
|
22
localtests/spatial57/create.sql
Normal file
22
localtests/spatial57/create.sql
Normal file
@ -0,0 +1,22 @@
|
|||||||
|
drop table if exists gh_ost_test;
|
||||||
|
create table gh_ost_test (
|
||||||
|
id int auto_increment,
|
||||||
|
g geometry,
|
||||||
|
pt point,
|
||||||
|
primary key(id)
|
||||||
|
) auto_increment=1;
|
||||||
|
|
||||||
|
drop event if exists gh_ost_test;
|
||||||
|
delimiter ;;
|
||||||
|
create event gh_ost_test
|
||||||
|
on schedule every 1 second
|
||||||
|
starts current_timestamp
|
||||||
|
ends current_timestamp + interval 60 second
|
||||||
|
on completion not preserve
|
||||||
|
enable
|
||||||
|
do
|
||||||
|
begin
|
||||||
|
insert into gh_ost_test values (null, ST_GeomFromText('POINT(1 1)'), POINT(10,10));
|
||||||
|
insert into gh_ost_test values (null, ST_GeomFromText('POINT(2 2)'), POINT(20,20));
|
||||||
|
insert into gh_ost_test values (null, ST_GeomFromText('POINT(3 3)'), POINT(30,30));
|
||||||
|
end ;;
|
1
localtests/spatial57/ignore_versions
Normal file
1
localtests/spatial57/ignore_versions
Normal file
@ -0,0 +1 @@
|
|||||||
|
(5.5|5.6)
|
1
localtests/swap-pk-uk/ignore_versions
Normal file
1
localtests/swap-pk-uk/ignore_versions
Normal file
@ -0,0 +1 @@
|
|||||||
|
(5.5)
|
1
localtests/swap-uk-uk/ignore_versions
Normal file
1
localtests/swap-uk-uk/ignore_versions
Normal file
@ -0,0 +1 @@
|
|||||||
|
(5.5)
|
@ -9,23 +9,43 @@
|
|||||||
|
|
||||||
tests_path=$(dirname $0)
|
tests_path=$(dirname $0)
|
||||||
test_logfile=/tmp/gh-ost-test.log
|
test_logfile=/tmp/gh-ost-test.log
|
||||||
ghost_binary=/tmp/gh-ost-test
|
default_ghost_binary=/tmp/gh-ost-test
|
||||||
|
ghost_binary=""
|
||||||
exec_command_file=/tmp/gh-ost-test.bash
|
exec_command_file=/tmp/gh-ost-test.bash
|
||||||
orig_content_output_file=/gh-ost-test.orig.content.csv
|
orig_content_output_file=/tmp/gh-ost-test.orig.content.csv
|
||||||
ghost_content_output_file=/gh-ost-test.ghost.content.csv
|
ghost_content_output_file=/tmp/gh-ost-test.ghost.content.csv
|
||||||
test_pattern="${1:-.}"
|
|
||||||
|
|
||||||
master_host=
|
master_host=
|
||||||
master_port=
|
master_port=
|
||||||
replica_host=
|
replica_host=
|
||||||
replica_port=
|
replica_port=
|
||||||
|
|
||||||
|
OPTIND=1
|
||||||
|
while getopts "b:" OPTION
|
||||||
|
do
|
||||||
|
case $OPTION in
|
||||||
|
b)
|
||||||
|
ghost_binary="$OPTARG"
|
||||||
|
;;
|
||||||
|
esac
|
||||||
|
done
|
||||||
|
shift $((OPTIND-1))
|
||||||
|
|
||||||
|
test_pattern="${1:-.}"
|
||||||
|
|
||||||
verify_master_and_replica() {
|
verify_master_and_replica() {
|
||||||
if [ "$(gh-ost-test-mysql-master -e "select 1" -ss)" != "1" ] ; then
|
if [ "$(gh-ost-test-mysql-master -e "select 1" -ss)" != "1" ] ; then
|
||||||
echo "Cannot verify gh-ost-test-mysql-master"
|
echo "Cannot verify gh-ost-test-mysql-master"
|
||||||
exit 1
|
exit 1
|
||||||
fi
|
fi
|
||||||
read master_host master_port <<< $(gh-ost-test-mysql-master -e "select @@hostname, @@port" -ss)
|
read master_host master_port <<< $(gh-ost-test-mysql-master -e "select @@hostname, @@port" -ss)
|
||||||
|
[ "$master_host" == "$(hostname)" ] && master_host="127.0.0.1"
|
||||||
|
echo "# master verified at $master_host:$master_port"
|
||||||
|
if ! gh-ost-test-mysql-master -e "set global event_scheduler := 1" ; then
|
||||||
|
echo "Cannot enable event_scheduler on master"
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
|
||||||
if [ "$(gh-ost-test-mysql-replica -e "select 1" -ss)" != "1" ] ; then
|
if [ "$(gh-ost-test-mysql-replica -e "select 1" -ss)" != "1" ] ; then
|
||||||
echo "Cannot verify gh-ost-test-mysql-replica"
|
echo "Cannot verify gh-ost-test-mysql-replica"
|
||||||
exit 1
|
exit 1
|
||||||
@ -35,6 +55,8 @@ verify_master_and_replica() {
|
|||||||
exit 1
|
exit 1
|
||||||
fi
|
fi
|
||||||
read replica_host replica_port <<< $(gh-ost-test-mysql-replica -e "select @@hostname, @@port" -ss)
|
read replica_host replica_port <<< $(gh-ost-test-mysql-replica -e "select @@hostname, @@port" -ss)
|
||||||
|
[ "$replica_host" == "$(hostname)" ] && replica_host="127.0.0.1"
|
||||||
|
echo "# replica verified at $replica_host:$replica_port"
|
||||||
}
|
}
|
||||||
|
|
||||||
exec_cmd() {
|
exec_cmd() {
|
||||||
@ -66,6 +88,15 @@ test_single() {
|
|||||||
local test_name
|
local test_name
|
||||||
test_name="$1"
|
test_name="$1"
|
||||||
|
|
||||||
|
if [ -f $tests_path/$test_name/ignore_versions ] ; then
|
||||||
|
ignore_versions=$(cat $tests_path/$test_name/ignore_versions)
|
||||||
|
mysql_version=$(gh-ost-test-mysql-master -s -s -e "select @@version")
|
||||||
|
if echo "$mysql_version" | egrep -q "^${ignore_versions}" ; then
|
||||||
|
echo -n "Skipping: $test_name"
|
||||||
|
return 0
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
|
||||||
echo -n "Testing: $test_name"
|
echo -n "Testing: $test_name"
|
||||||
|
|
||||||
echo_dot
|
echo_dot
|
||||||
@ -98,6 +129,7 @@ test_single() {
|
|||||||
--password=gh-ost \
|
--password=gh-ost \
|
||||||
--host=$replica_host \
|
--host=$replica_host \
|
||||||
--port=$replica_port \
|
--port=$replica_port \
|
||||||
|
--assume-master-host=${master_host}:${master_port}
|
||||||
--database=test \
|
--database=test \
|
||||||
--table=gh_ost_test \
|
--table=gh_ost_test \
|
||||||
--alter='engine=innodb' \
|
--alter='engine=innodb' \
|
||||||
@ -148,7 +180,8 @@ test_single() {
|
|||||||
|
|
||||||
if [ $execution_result -ne 0 ] ; then
|
if [ $execution_result -ne 0 ] ; then
|
||||||
echo
|
echo
|
||||||
echo "ERROR $test_name execution failure. cat $test_logfile"
|
echo "ERROR $test_name execution failure. cat $test_logfile:"
|
||||||
|
cat $test_logfile
|
||||||
return 1
|
return 1
|
||||||
fi
|
fi
|
||||||
|
|
||||||
@ -170,7 +203,12 @@ test_single() {
|
|||||||
|
|
||||||
build_binary() {
|
build_binary() {
|
||||||
echo "Building"
|
echo "Building"
|
||||||
rm -f $ghost_binary
|
rm -f $default_ghost_binary
|
||||||
|
[ "$ghost_binary" == "" ] && ghost_binary="$default_ghost_binary"
|
||||||
|
if [ -f "$ghost_binary" ] ; then
|
||||||
|
echo "Using binary: $ghost_binary"
|
||||||
|
return 0
|
||||||
|
fi
|
||||||
go build -o $ghost_binary go/cmd/gh-ost/main.go
|
go build -o $ghost_binary go/cmd/gh-ost/main.go
|
||||||
if [ $? -ne 0 ] ; then
|
if [ $? -ne 0 ] ; then
|
||||||
echo "Build failure"
|
echo "Build failure"
|
||||||
|
1
localtests/timestamp-to-datetime/ignore_versions
Normal file
1
localtests/timestamp-to-datetime/ignore_versions
Normal file
@ -0,0 +1 @@
|
|||||||
|
(5.5)
|
1
localtests/timestamp/ignore_versions
Normal file
1
localtests/timestamp/ignore_versions
Normal file
@ -0,0 +1 @@
|
|||||||
|
(5.5)
|
1
localtests/tz-datetime-ts/ignore_versions
Normal file
1
localtests/tz-datetime-ts/ignore_versions
Normal file
@ -0,0 +1 @@
|
|||||||
|
(5.5)
|
1
localtests/tz/ignore_versions
Normal file
1
localtests/tz/ignore_versions
Normal file
@ -0,0 +1 @@
|
|||||||
|
(5.5)
|
66
script/cibuild-gh-ost-replica-tests
Executable file
66
script/cibuild-gh-ost-replica-tests
Executable file
@ -0,0 +1,66 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
|
||||||
|
set -e
|
||||||
|
|
||||||
|
whoami
|
||||||
|
|
||||||
|
# Clone gh-ost-ci-env
|
||||||
|
# Only clone if not already running locally at latest commit
|
||||||
|
remote_commit=$(git ls-remote https://github.com/github/gh-ost-ci-env.git HEAD | cut -f1)
|
||||||
|
local_commit="unknown"
|
||||||
|
[ -d "gh-ost-ci-env" ] && local_commit=$(cd gh-ost-ci-env && git log --format="%H" -n 1)
|
||||||
|
|
||||||
|
echo "remote commit is: $remote_commit"
|
||||||
|
echo "local commit is: $local_commit"
|
||||||
|
|
||||||
|
if [ "$remote_commit" != "$local_commit" ] ; then
|
||||||
|
rm -rf ./gh-ost-ci-env
|
||||||
|
git clone https://github.com/github/gh-ost-ci-env.git
|
||||||
|
fi
|
||||||
|
|
||||||
|
test_mysql_version() {
|
||||||
|
local mysql_version
|
||||||
|
mysql_version="$1"
|
||||||
|
|
||||||
|
echo "##### Testing $mysql_version"
|
||||||
|
|
||||||
|
echo "### Setting up sandbox for $mysql_version"
|
||||||
|
|
||||||
|
find sandboxes -name "stop_all" | bash
|
||||||
|
|
||||||
|
mkdir -p sandbox/binary
|
||||||
|
rm -rf sandbox/binary/*
|
||||||
|
gh-ost-ci-env/bin/linux/dbdeployer unpack gh-ost-ci-env/mysql-tarballs/"$mysql_version".tar.gz --unpack-version="$mysql_version" --sandbox-binary ${PWD}/sandbox/binary
|
||||||
|
|
||||||
|
mkdir -p sandboxes
|
||||||
|
rm -rf sandboxes/*
|
||||||
|
|
||||||
|
if echo "$mysql_version" | egrep "5[.]5[.]" ; then
|
||||||
|
gtid=""
|
||||||
|
else
|
||||||
|
gtid="--gtid"
|
||||||
|
fi
|
||||||
|
gh-ost-ci-env/bin/linux/dbdeployer deploy replication "$mysql_version" --nodes 2 --sandbox-binary ${PWD}/sandbox/binary --sandbox-home ${PWD}/sandboxes ${gtid} --my-cnf-options log_slave_updates --my-cnf-options log_bin --my-cnf-options binlog_format=ROW --sandbox-directory rsandbox
|
||||||
|
|
||||||
|
sed '/sandboxes/d' -i gh-ost-ci-env/bin/gh-ost-test-mysql-master
|
||||||
|
echo 'sandboxes/rsandbox/m "$@"' >> gh-ost-ci-env/bin/gh-ost-test-mysql-master
|
||||||
|
|
||||||
|
sed '/sandboxes/d' -i gh-ost-ci-env/bin/gh-ost-test-mysql-replica
|
||||||
|
echo 'sandboxes/rsandbox/s1 "$@"' >> gh-ost-ci-env/bin/gh-ost-test-mysql-replica
|
||||||
|
|
||||||
|
export PATH="${PWD}/gh-ost-ci-env/bin/:${PATH}"
|
||||||
|
|
||||||
|
gh-ost-test-mysql-master -uroot -e "grant all on *.* to 'gh-ost'@'%' identified by 'gh-ost'"
|
||||||
|
|
||||||
|
echo "### Running gh-ost tests for $mysql_version"
|
||||||
|
./localtests/test.sh -b bin/gh-ost
|
||||||
|
|
||||||
|
find sandboxes -name "stop_all" | bash
|
||||||
|
}
|
||||||
|
|
||||||
|
echo "Building..."
|
||||||
|
. script/build
|
||||||
|
# Test all versions:
|
||||||
|
find gh-ost-ci-env/mysql-tarballs/ -name "*.tar.gz" | while read f ; do basename $f ".tar.gz" ; done | sort -r | while read mysql_version ; do
|
||||||
|
test_mysql_version "$mysql_version"
|
||||||
|
done
|
2
vendor/github.com/siddontang/go-mysql/cmd/go-mysqlbinlog/main.go
generated
vendored
2
vendor/github.com/siddontang/go-mysql/cmd/go-mysqlbinlog/main.go
generated
vendored
@ -41,7 +41,7 @@ func main() {
|
|||||||
Port: uint16(*port),
|
Port: uint16(*port),
|
||||||
User: *user,
|
User: *user,
|
||||||
Password: *password,
|
Password: *password,
|
||||||
RawModeEanbled: *rawMode,
|
RawModeEnabled: *rawMode,
|
||||||
SemiSyncEnabled: *semiSync,
|
SemiSyncEnabled: *semiSync,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
6
vendor/github.com/siddontang/go-mysql/replication/binlogsyncer.go
generated
vendored
6
vendor/github.com/siddontang/go-mysql/replication/binlogsyncer.go
generated
vendored
@ -43,8 +43,8 @@ type BinlogSyncerConfig struct {
|
|||||||
// SemiSyncEnabled enables semi-sync or not.
|
// SemiSyncEnabled enables semi-sync or not.
|
||||||
SemiSyncEnabled bool
|
SemiSyncEnabled bool
|
||||||
|
|
||||||
// RawModeEanbled is for not parsing binlog event.
|
// RawModeEnabled is for not parsing binlog event.
|
||||||
RawModeEanbled bool
|
RawModeEnabled bool
|
||||||
|
|
||||||
// If not nil, use the provided tls.Config to connect to the database using TLS/SSL.
|
// If not nil, use the provided tls.Config to connect to the database using TLS/SSL.
|
||||||
TLSConfig *tls.Config
|
TLSConfig *tls.Config
|
||||||
@ -83,7 +83,7 @@ func NewBinlogSyncer(cfg BinlogSyncerConfig) *BinlogSyncer {
|
|||||||
|
|
||||||
b.cfg = cfg
|
b.cfg = cfg
|
||||||
b.parser = NewBinlogParser()
|
b.parser = NewBinlogParser()
|
||||||
b.parser.SetRawMode(b.cfg.RawModeEanbled)
|
b.parser.SetRawMode(b.cfg.RawModeEnabled)
|
||||||
|
|
||||||
b.running = false
|
b.running = false
|
||||||
b.ctx, b.cancel = context.WithCancel(context.Background())
|
b.ctx, b.cancel = context.WithCancel(context.Background())
|
||||||
|
8
vendor/github.com/siddontang/go-mysql/replication/json_binary.go
generated
vendored
8
vendor/github.com/siddontang/go-mysql/replication/json_binary.go
generated
vendored
@ -338,7 +338,7 @@ func (d *jsonBinaryDecoder) decodeString(data []byte) string {
|
|||||||
|
|
||||||
l, n := d.decodeVariableLength(data)
|
l, n := d.decodeVariableLength(data)
|
||||||
|
|
||||||
if d.isDataShort(data, int(l)+n) {
|
if d.isDataShort(data, l+n) {
|
||||||
return ""
|
return ""
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -358,11 +358,11 @@ func (d *jsonBinaryDecoder) decodeOpaque(data []byte) interface{} {
|
|||||||
|
|
||||||
l, n := d.decodeVariableLength(data)
|
l, n := d.decodeVariableLength(data)
|
||||||
|
|
||||||
if d.isDataShort(data, int(l)+n) {
|
if d.isDataShort(data, l+n) {
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
data = data[n : int(l)+n]
|
data = data[n : l+n]
|
||||||
|
|
||||||
switch tp {
|
switch tp {
|
||||||
case MYSQL_TYPE_NEWDECIMAL:
|
case MYSQL_TYPE_NEWDECIMAL:
|
||||||
@ -459,7 +459,7 @@ func (d *jsonBinaryDecoder) decodeVariableLength(data []byte) (int, int) {
|
|||||||
length := uint64(0)
|
length := uint64(0)
|
||||||
for ; pos < maxCount; pos++ {
|
for ; pos < maxCount; pos++ {
|
||||||
v := data[pos]
|
v := data[pos]
|
||||||
length = (length << 7) + uint64(v&0x7F)
|
length |= uint64(v&0x7F) << uint(7*pos)
|
||||||
|
|
||||||
if v&0x80 == 0 {
|
if v&0x80 == 0 {
|
||||||
if length > math.MaxUint32 {
|
if length > math.MaxUint32 {
|
||||||
|
13
vendor/github.com/siddontang/go-mysql/replication/row_event_test.go
generated
vendored
13
vendor/github.com/siddontang/go-mysql/replication/row_event_test.go
generated
vendored
@ -403,6 +403,8 @@ func (_ *testDecodeSuite) TestParseJson(c *C) {
|
|||||||
|
|
||||||
// INSERT INTO `t10` (`c2`) VALUES (1);
|
// INSERT INTO `t10` (`c2`) VALUES (1);
|
||||||
// INSERT INTO `t10` (`c1`, `c2`) VALUES ('{"key1": "value1", "key2": "value2"}', 1);
|
// INSERT INTO `t10` (`c1`, `c2`) VALUES ('{"key1": "value1", "key2": "value2"}', 1);
|
||||||
|
// test json deserialization
|
||||||
|
// INSERT INTO `t10`(`c1`,`c2`) VALUES ('{"text":"Lorem ipsum dolor sit amet, consectetuer adipiscing elit. Aenean commodo ligula eget dolor. Aenean massa. Cum sociis natoque penatibus et magnis dis parturient montes, nascetur ridiculus mus. Donec quam felis, ultricies nec, pellentesque eu, pretium quis, sem. Nulla consequat massa quis enim. Donec pede justo, fringilla vel, aliquet nec, vulputate eget, arcu. In enim justo, rhoncus ut, imperdiet a, venenatis vitae, justo. Nullam dictum felis eu pede mollis pretium. Integer tincidunt. Cras dapibus. Vivamus elementum semper nisi. Aenean vulputate eleifend tellus. Aenean leo ligula, porttitor eu, consequat vitae, eleifend ac, enim. Aliquam lorem ante, dapibus in, viverra quis, feugiat a, tellus. Phasellus viverra nulla ut metus varius laoreet. Quisque rutrum. Aenean imperdiet. Etiam ultricies nisi vel augue. Curabitur ullamcorper ultricies nisi. Nam eget dui. Etiam rhoncus. Maecenas tempus, tellus eget condimentum rhoncus, sem quam semper libero, sit amet adipiscing sem neque sed ipsum. Nam quam nunc, blandit vel, luctus pulvinar, hendrerit id, lorem. Maecenas nec odio et ante tincidunt tempus. Donec vitae sapien ut libero venenatis faucibus. Nullam quis ante. Etiam sit amet orci eget eros faucibus tincidunt. Duis leo. Sed fringilla mauris sit amet nibh. Donec sodales sagittis magna. Sed consequat, leo eget bibendum sodales, augue velit cursus nunc, quis gravida magna mi a libero. Fusce vulputate eleifend sapien. Vestibulum purus quam, scelerisque ut, mollis sed, nonummy id, metus. Nullam accumsan lorem in dui. Cras ultricies mi eu turpis hendrerit fringilla. Vestibulum ante ipsum primis in faucibus orci luctus et ultrices posuere cubilia Curae; In ac dui quis mi consectetuer lacinia. Nam pretium turpis et arcu. Duis arcu tortor, suscipit eget, imperdiet nec, imperdiet iaculis, ipsum. Sed aliquam ultrices mauris. Integer ante arcu, accumsan a, consectetuer eget, posuere ut, mauris. Praesent adipiscing. Phasellus ullamcorper ipsum rutrum nunc. Nunc nonummy metus. Vestibulum volutpat pretium libero. Cras id dui. Aenean ut eros et nisl sagittis vestibulum. Nullam nulla eros, ultricies sit amet, nonummy id, imperdiet feugiat, pede. Sed lectus. Donec mollis hendrerit risus. Phasellus nec sem in justo pellentesque facilisis. Etiam imperdiet imperdiet orci. Nunc nec neque. Phasellus leo dolor, tempus non, auctor et, hendrerit quis, nisi. Curabitur ligula sapien, tincidunt non, euismod vitae, posuere imperdiet, leo. Maecenas malesuada. Praesent congue erat at massa. Sed cursus turpis vitae tortor. Donec posuere vulputate arcu. Phasellus accumsan cursus velit. Vestibulum ante ipsum primis in faucibus orci luctus et ultrices posuere cubilia Curae; Sed aliquam, nisi quis porttitor congue, elit erat euismod orci, ac"}',101);
|
||||||
|
|
||||||
tableMapEventData := []byte("m\x00\x00\x00\x00\x00\x01\x00\x04test\x00\x03t10\x00\x02\xf5\xf6\x03\x04\n\x00\x03")
|
tableMapEventData := []byte("m\x00\x00\x00\x00\x00\x01\x00\x04test\x00\x03t10\x00\x02\xf5\xf6\x03\x04\n\x00\x03")
|
||||||
|
|
||||||
@ -428,4 +430,15 @@ func (_ *testDecodeSuite) TestParseJson(c *C) {
|
|||||||
c.Assert(err, IsNil)
|
c.Assert(err, IsNil)
|
||||||
c.Assert(rows.Rows[0][1], Equals, float64(1))
|
c.Assert(rows.Rows[0][1], Equals, float64(1))
|
||||||
}
|
}
|
||||||
|
|
||||||
|
longTbls := [][]byte{
|
||||||
|
[]byte("m\x00\x00\x00\x00\x00\x01\x00\x02\x00\x02\xff\xfc\xd0\n\x00\x00\x00\x01\x00\xcf\n\v\x00\x04\x00\f\x0f\x00text\xbe\x15Lorem ipsum dolor sit amet, consectetuer adipiscing elit. Aenean commodo ligula eget dolor. Aenean massa. Cum sociis natoque penatibus et magnis dis parturient montes, nascetur ridiculus mus. Donec quam felis, ultricies nec, pellentesque eu, pretium quis, sem. Nulla consequat massa quis enim. Donec pede justo, fringilla vel, aliquet nec, vulputate eget, arcu. In enim justo, rhoncus ut, imperdiet a, venenatis vitae, justo. Nullam dictum felis eu pede mollis pretium. Integer tincidunt. Cras dapibus. Vivamus elementum semper nisi. Aenean vulputate eleifend tellus. Aenean leo ligula, porttitor eu, consequat vitae, eleifend ac, enim. Aliquam lorem ante, dapibus in, viverra quis, feugiat a, tellus. Phasellus viverra nulla ut metus varius laoreet. Quisque rutrum. Aenean imperdiet. Etiam ultricies nisi vel augue. Curabitur ullamcorper ultricies nisi. Nam eget dui. Etiam rhoncus. Maecenas tempus, tellus eget condimentum rhoncus, sem quam semper libero, sit amet adipiscing sem neque sed ipsum. Nam quam nunc, blandit vel, luctus pulvinar, hendrerit id, lorem. Maecenas nec odio et ante tincidunt tempus. Donec vitae sapien ut libero venenatis faucibus. Nullam quis ante. Etiam sit amet orci eget eros faucibus tincidunt. Duis leo. Sed fringilla mauris sit amet nibh. Donec sodales sagittis magna. Sed consequat, leo eget bibendum sodales, augue velit cursus nunc, quis gravida magna mi a libero. Fusce vulputate eleifend sapien. Vestibulum purus quam, scelerisque ut, mollis sed, nonummy id, metus. Nullam accumsan lorem in dui. Cras ultricies mi eu turpis hendrerit fringilla. Vestibulum ante ipsum primis in faucibus orci luctus et ultrices posuere cubilia Curae; In ac dui quis mi consectetuer lacinia. Nam pretium turpis et arcu. Duis arcu tortor, suscipit eget, imperdiet nec, imperdiet iaculis, ipsum. Sed aliquam ultrices mauris. Integer ante arcu, accumsan a, consectetuer eget, posuere ut, mauris. Praesent adipiscing. Phasellus ullamcorper ipsum rutrum nunc. Nunc nonummy metus. Vestibulum volutpat pretium libero. Cras id dui. Aenean ut eros et nisl sagittis vestibulum. Nullam nulla eros, ultricies sit amet, nonummy id, imperdiet feugiat, pede. Sed lectus. Donec mollis hendrerit risus. Phasellus nec sem in justo pellentesque facilisis. Etiam imperdiet imperdiet orci. Nunc nec neque. Phasellus leo dolor, tempus non, auctor et, hendrerit quis, nisi. Curabitur ligula sapien, tincidunt non, euismod vitae, posuere imperdiet, leo. Maecenas malesuada. Praesent congue erat at massa. Sed cursus turpis vitae tortor. Donec posuere vulputate arcu. Phasellus accumsan cursus velit. Vestibulum ante ipsum primis in faucibus orci luctus et ultrices posuere cubilia Curae; Sed aliquam, nisi quis porttitor congue, elit erat euismod orci, ac\x80\x00\x00\x00e"),
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, ltbl := range longTbls {
|
||||||
|
rows.Rows = nil
|
||||||
|
err = rows.Decode(ltbl)
|
||||||
|
c.Assert(err, IsNil)
|
||||||
|
c.Assert(rows.Rows[0][1], Equals, float64(101))
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
Loading…
Reference in New Issue
Block a user