[MXS-1148] CoreDump using maxadmin Created: 2017-02-27 Updated: 2017-03-02 Resolved: 2017-03-02 |
|
| Status: | Closed |
| Project: | MariaDB MaxScale |
| Component/s: | maxadmin, maxinfo |
| Affects Version/s: | 2.0.4 |
| Fix Version/s: | 2.1.0 |
| Type: | Bug | Priority: | Major |
| Reporter: | VAROQUI Stephane | Assignee: | Esa Korhonen |
| Resolution: | Cannot Reproduce | Votes: | 0 |
| Labels: | None | ||
| Sprint: | 2017-28, 2017-29 |
| Description |
|
What the replication manager do is in monitoring go routine pool the servers lists and the monitoring list During switchover we do
After a signifiant number or retried we are able to crash the maxscale. Nota that i do open and close the connection in failover and monitoring code 2017-02-27 10:34:05 notice : Server changed state: server1[192.168.0.41:5055]: server_down. [Running] -> [Down] |
| Comments |
| Comment by Johan Wikman [ 2017-02-28 ] |
|
Downgrading to major. Will be investigated of course, but the circumstances are somewhat unusual. |
| Comment by VAROQUI Stephane [ 2017-02-28 ] |
|
Ok i get more infos on this , this happens not with maxadmin. but with cobination of maxinfo as well , the first report description was misslinding due to a typo in the conf of replication-manager config |
| Comment by Esa Korhonen [ 2017-02-28 ] |
|
Hello, Stephane. I'm trying to reproduce your crash. |
| Comment by VAROQUI Stephane [ 2017-02-28 ] |
|
Hello Esa , yes just /servers and /monitors around 2s interval polling On the other end on a 2 nodes cluster I do in long timeout loop err = m.Command("clear server " + cluster.master.MxsServerName + " slave") if err != nil { cluster.LogPrint("ERROR: MaxScale client could not send command:%s", err) } if cluster.conf.MxsMonitor == false { err = m.Command("set server " + s.MxsServerName + " slave") } err = m.Command("set server " + oldmaster.MxsServerName + " slave") } } like inverting roles from master and slave |
| Comment by VAROQUI Stephane [ 2017-02-28 ] |
|
Note that i don't really noted the status of the monitor at that time it may have been stopped or may have been running |
| Comment by Esa Korhonen [ 2017-03-02 ] |
|
I didn't manage to replicate this (only got valgrind to detect one invalid read/write). Server state setting code has been modified for versions 2.1.X and the cause for this crash may well be fixed in the new versions. If this crash keeps occurring, you may wish to test with 2.1.0. |