Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

stability: panic on RangeLookup dispatched to correct range #6000

Closed
mberhault opened this issue Apr 12, 2016 · 4 comments
Closed

stability: panic on RangeLookup dispatched to correct range #6000

mberhault opened this issue Apr 12, 2016 · 4 comments
Labels
S-1-stability Severe stability issues that can be fixed by upgrading, but usually don’t resolve by restarting
Milestone

Comments

@mberhault
Copy link
Contributor

build sha: c382dd7
Beta cluster with photos + block_writer load. Restarted from scratch earlier today.

~1h after a restart, encountered the following:

panic: RangeLookup dispatched to correct range, but no matching RangeDescriptor was found: /Meta2/Table/55/1/8752590996753796817/"88e08dd7-d1b6-4aeb-ad46-1880b351e731"/6611/0

1: running [Created by grpc.(*Server).serveStreams.func1 @ server.go:324]
              panic.go:464            panic(#1, #18)
    storage   store.go:1557           (*Store).Send.func1(0xc833bd1258, 0xc833bd12c8, 0x14449b721ca69e59, 0, 0xc833bd12c0, #5)
              panic.go:426            panic(#1, #18)
    storage   replica_command.go:909  (*Replica).RangeLookup(#12, #19, #9, #22, 0, #4, 0x4, 0x1, 0, 0, ...)
    storage   replica_command.go:129  (*Replica).executeCmd(#12, #20, #17, 0, 0, 0, #19, #9, 0, #22, ...)
    storage   replica.go:1784         (*Replica).executeBatch(#12, #20, #17, 0, 0, #19, #9, 0, #22, 0, ...)
    storage   replica.go:1071         (*Replica).addReadOnlyCmd(#12, #20, #17, #22, 0, #4, 0x4, 0x1, 0, 0, ...)
    storage   replica.go:765          (*Replica).Send(#12, #20, #17, #22, 0, #4, 0x4, 0x1, 0, 0, ...)
    storage   store.go:1639           (*Store).Send(#5, #20, #17, #22, 0, #4, 0x4, 0x1, 0, 0, ...)
    storage   stores.go:177           (*Stores).Send(#6, #20, 0xc82bb06c30, 0, 0, #4, 0x4, 0x1, 0, 0, ...)
    server    node.go:729             (*Node).Batch.func2()
    stop      stopper.go:166          (*Stopper).RunTask(#8, 0xc833bd1870, 0)
    server    node.go:741             (*Node).Batch(#7, #20, #16, 0xc8271b2150, 0, 0, 0)
    roachpb   api.pb.go:1306          _Internal_Batch_Handler(#2, #7, #20, #16, 0xc833082a00, 0, 0, 0, 0)
    grpc      server.go:497           (*Server).processUnaryRPC(#11, #21, #14, #15, #10, #3, 0, 0, 0)
    grpc      server.go:646           (*Server).handleStream(#11, #21, #14, #15, 0)
    grpc      server.go:323           (*Server).serveStreams.func1.1(#13, #11, #21, #14, #15)

Full log (from ec2-52-91-234-29.compute-1.amazonaws.com):
node4.log.parse.txt

@tbg
Copy link
Member

tbg commented Apr 12, 2016

Could you save the data? Only that node is ok. Please don't restart (you
can after taking the dump, but it'll likely happen again).

On Tue, Apr 12, 2016 at 9:45 AM marc [email protected] wrote:

build sha: c382dd7
c382dd7
Beta cluster with photos + block_writer load. Restarted from scratch
earlier today.

~1h after a restart, encountered the following:

panic: RangeLookup dispatched to correct range, but no matching RangeDescriptor was found: /Meta2/Table/55/1/8752590996753796817/"88e08dd7-d1b6-4aeb-ad46-1880b351e731"/6611/0

1: running [Created by grpc.(_Server).serveStreams.func1 @ server.go:324]
panic.go:464 panic(#1, #18)
storage store.go:1557 (_Store).Send.func1(0xc833bd1258, 0xc833bd12c8, 0x14449b721ca69e59, 0, 0xc833bd12c0, #5)
panic.go:426 panic(#1, #18)
storage replica_command.go:909 (_Replica).RangeLookup(#12, #19, #9, #22, 0, #4, 0x4, 0x1, 0, 0, ...)
storage replica_command.go:129 (_Replica).executeCmd(#12, #20, #17, 0, 0, 0, #19, #9, 0, #22, ...)
storage replica.go:1784 (_Replica).executeBatch(#12, #20, #17, 0, 0, #19, #9, 0, #22, 0, ...)
storage replica.go:1071 (_Replica).addReadOnlyCmd(#12, #20, #17, #22, 0, #4, 0x4, 0x1, 0, 0, ...)
storage replica.go:765 (_Replica).Send(#12, #20, #17, #22, 0, #4, 0x4, 0x1, 0, 0, ...)
storage store.go:1639 (_Store).Send(#5, #20, #17, #22, 0, #4, 0x4, 0x1, 0, 0, ...)
storage stores.go:177 (_Stores).Send(#6, #20, 0xc82bb06c30, 0, 0, #4, 0x4, 0x1, 0, 0, ...)
server node.go:729 (_Node).Batch.func2()
stop stopper.go:166 (_Stopper).RunTask(#8, 0xc833bd1870, 0)
server node.go:741 (_Node).Batch(#7, #20, #16, 0xc8271b2150, 0, 0, 0)
roachpb api.pb.go:1306 _Internal_Batch_Handler(#2, #7, #20, #16, 0xc833082a00, 0, 0, 0, 0)
grpc server.go:497 (_Server).processUnaryRPC(#11, #21, #14, #15, #10, #3, 0, 0, 0)
grpc server.go:646 (_Server).handleStream(#11, #21, #14, #15, 0)
grpc server.go:323 (*Server).serveStreams.func1.1(#13, #11, #21, #14, #15)

Full log (from ec2-52-91-234-29.compute-1.amazonaws.com):
node4.log.parse.txt
https://github.com/cockroachdb/cockroach/files/215109/node4.log.parse.txt


You are receiving this because you are subscribed to this thread.
Reply to this email directly or view it on GitHub
#6000

-- Tobias

@mberhault
Copy link
Contributor Author

yeah, it's still down, but I'm backing up with suffix .6000 on ec2-52-91-234-29.compute-1.amazonaws.com

@mberhault mberhault modified the milestone: Q2 Apr 17, 2016
@tamird tamird added the S-1-stability Severe stability issues that can be fixed by upgrading, but usually don’t resolve by restarting label May 5, 2016
@tamird
Copy link
Contributor

tamird commented Jul 2, 2016

Is this data still available somewhere?

@petermattis petermattis modified the milestones: Q2, Q3 Jul 11, 2016
@tamird
Copy link
Contributor

tamird commented Oct 4, 2016

This looks like a meta2 gap? Closing as dupe of #9265.

@tamird tamird closed this as completed Oct 4, 2016
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
S-1-stability Severe stability issues that can be fixed by upgrading, but usually don’t resolve by restarting
Projects
None yet
Development

No branches or pull requests

4 participants