-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Bump lodash.merge from 4.6.1 to 4.6.2 in /devel/electron5/files #12
base: master
Are you sure you want to change the base?
Bump lodash.merge from 4.6.1 to 4.6.2 in /devel/electron5/files #12
Conversation
Bumps [lodash.merge](https://github.com/lodash/lodash) from 4.6.1 to 4.6.2. - [Release notes](https://github.com/lodash/lodash/releases) - [Commits](https://github.com/lodash/lodash/commits) Signed-off-by: dependabot[bot] <[email protected]>
No PORTEPOCH bump because this port wasn't stable to begin with. * thread #9, name = 'yuzu:CPUThread', stop reason = signal SIGABRT * frame #0: 0x0000000804146a8a libc.so.7`__sys_thr_kill at thr_kill.S:4 frame #1: 0x0000000804146424 libc.so.7`__raise(s=6) at raise.c:52:10 frame #2: 0x00000008040aef19 libc.so.7`abort at abort.c:67:8 frame #3: 0x00000008038f39b9 libcxxrt.so.1`report_failure(err=<unavailable>, thrown_exception=0x00000009d701aa88) at exception.cc:719:5 frame #4: 0x00000008038c34dc libc++.so.1`std::__1::__throw_system_error(ev=11, what_arg="mutex lock failed") at system_error.cpp:287:5 frame #5: 0x00000008038a834d libc++.so.1`std::__1::mutex::lock(this=<unavailable>) at mutex.cpp:35:9 frame #6: 0x0000000000dbb534 yuzu`std::__1::unique_lock<std::__1::mutex>::unique_lock(this=0x00000009c68f1d90, __m=0x0000000805984918) at __mutex_base:119:61 frame #7: 0x000000000136167d yuzu`Service::NVFlinger::NVFlinger::Lock(this=0x000000080c8c6958) at nvflinger.h:90:16 frame #8: 0x00000000014c5ab4 yuzu`Service::VI::IHOSBinderDriver::TransactParcel(this=0x00000009d536e6f8, thread=std::__1::shared_ptr<Kernel::Thread>::element_type @ 0x000000090faedc20 strong=9 weak=2, ctx=0x00000009d536e310, reason=Signal)::'lambda'(std::__1::shared_ptr<Kernel::Thread>, Kernel::HLERequestContext&, Kernel::ThreadWakeupReason)::operator()(std::__1::shared_ptr<Kernel::Thread>, Kernel::HLERequestContext&, Kernel::ThreadWakeupReason) const at vi.cpp:554:37 frame #9: 0x00000000014c59f5 yuzu`decltype(__f=0x00000009d536e6f8, __args=nullptr, __args=0x00000009d536e310, __args=0x00000009c68f2004)::'lambda'(std::__1::shared_ptr<Kernel::Thread>, Kernel::HLERequestContext&, Kernel::ThreadWakeupReason)&>(fp)(std::__1::forward<std::__1::shared_ptr<Kernel::Thread> >(fp0), std::__1::forward<Kernel::HLERequestContext&>(fp0), std::__1::forward<Kernel::ThreadWakeupReason>(fp0))) std::__1::__invoke<Service::VI::IHOSBinderDriver::TransactParcel(Kernel::HLERequestContext&)::'lambda'(std::__1::shared_ptr<Kernel::Thread>, Kernel::HLERequestContext&, Kernel::ThreadWakeupReason)&, std::__1::shared_ptr<Kernel::Thread>, Kernel::HLERequestContext&, Kernel::ThreadWakeupReason>(Service::VI::IHOSBinderDriver::TransactParcel(Kernel::HLERequestContext&)::'lambda'(std::__1::shared_ptr<Kernel::Thread>, Kernel::HLERequestContext&, Kernel::ThreadWakeupReason)&, std::__1::shared_ptr<Kernel::Thread>&&, Kernel::HLERequestContext&, Kernel::ThreadWakeupReason&&) at type_traits:3539:1 frame #10: 0x00000000014c594c yuzu`void std::__1::__invoke_void_return_wrapper<void>::__call<Service::VI::IHOSBinderDriver::TransactParcel(__args=0x00000009d536e6f8, __args=nullptr, __args=0x00000009d536e310, __args=0x00000009c68f2004)::'lambda'(std::__1::shared_ptr<Kernel::Thread>, Kernel::HLERequestContext&, Kernel::ThreadWakeupReason)&, std::__1::shared_ptr<Kernel::Thread>, Kernel::HLERequestContext&, Kernel::ThreadWakeupReason>(Service::VI::IHOSBinderDriver::TransactParcel(Kernel::HLERequestContext&)::'lambda'(std::__1::shared_ptr<Kernel::Thread>, Kernel::HLERequestContext&, Kernel::ThreadWakeupReason)&, std::__1::shared_ptr<Kernel::Thread>&&, Kernel::HLERequestContext&, Kernel::ThreadWakeupReason&&) at __functional_base:348:9 frame #11: 0x00000000014c58dc yuzu`std::__1::__function::__alloc_func<Service::VI::IHOSBinderDriver::TransactParcel(Kernel::HLERequestContext&)::'lambda'(std::__1::shared_ptr<Kernel::Thread>, Kernel::HLERequestContext&, Kernel::ThreadWakeupReason), std::__1::allocator<Service::VI::IHOSBinderDriver::TransactParcel(Kernel::HLERequestContext&)::'lambda'(std::__1::shared_ptr<Kernel::Thread>, Kernel::HLERequestContext&, Kernel::ThreadWakeupReason)>, void (std::__1::shared_ptr<Kernel::Thread>, Kernel::HLERequestContext&, Kernel::ThreadWakeupReason)>::operator(this=0x00000009d536e6f8, __arg=nullptr, __arg=0x00000009d536e310, __arg=0x00000009c68f2004)(std::__1::shared_ptr<Kernel::Thread>&&, Kernel::HLERequestContext&, Kernel::ThreadWakeupReason&&) at functional:1540:16 frame #12: 0x00000000014c480d yuzu`std::__1::__function::__func<Service::VI::IHOSBinderDriver::TransactParcel(Kernel::HLERequestContext&)::'lambda'(std::__1::shared_ptr<Kernel::Thread>, Kernel::HLERequestContext&, Kernel::ThreadWakeupReason), std::__1::allocator<Service::VI::IHOSBinderDriver::TransactParcel(Kernel::HLERequestContext&)::'lambda'(std::__1::shared_ptr<Kernel::Thread>, Kernel::HLERequestContext&, Kernel::ThreadWakeupReason)>, void (std::__1::shared_ptr<Kernel::Thread>, Kernel::HLERequestContext&, Kernel::ThreadWakeupReason)>::operator(this=0x00000009d536e6f0, __arg=nullptr, __arg=0x00000009d536e310, __arg=0x00000009c68f2004)(std::__1::shared_ptr<Kernel::Thread>&&, Kernel::HLERequestContext&, Kernel::ThreadWakeupReason&&) at functional:1714:12 frame #13: 0x0000000001116862 yuzu`std::__1::__function::__value_func<void (std::__1::shared_ptr<Kernel::Thread>, Kernel::HLERequestContext&, Kernel::ThreadWakeupReason)>::operator(this=0x00000009d536e6f0, __args=nullptr, __args=0x00000009d536e310, __args=0x00000009c68f2004)(std::__1::shared_ptr<Kernel::Thread>&&, Kernel::HLERequestContext&, Kernel::ThreadWakeupReason&&) const at functional:1867:16 frame #14: 0x00000000011167bc yuzu`std::__1::function<void (std::__1::shared_ptr<Kernel::Thread>, Kernel::HLERequestContext&, Kernel::ThreadWakeupReason)>::operator(this= Lambda in File vi.cpp at Line 552, __arg=<unavailable>, __arg=0x00000009d536e310, __arg=Signal)(std::__1::shared_ptr<Kernel::Thread>, Kernel::HLERequestContext&, Kernel::ThreadWakeupReason) const at functional:2473:12 frame #15: 0x000000000110a6a4 yuzu`Kernel::HLERequestContext::SleepClientThread(this=0x00000009d536e310, thread=std::__1::shared_ptr<Kernel::Thread>::element_type @ 0x000000090faedc20 strong=9 weak=2)>&&, std::__1::shared_ptr<Kernel::WritableEvent>)::$_0::operator()(std::__1::shared_ptr<Kernel::Thread>) at hle_ipc.cpp:67:17 frame #16: 0x000000000110a5b1 yuzu`decltype(__f=0x00000009d536e310, __args=nullptr)>&&, std::__1::shared_ptr<Kernel::WritableEvent>)::$_0&>(fp)(std::__1::forward<std::__1::shared_ptr<Kernel::Thread> >(fp0))) std::__1::__invoke<Kernel::HLERequestContext::SleepClientThread(std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, unsigned long, std::__1::function<void (std::__1::shared_ptr<Kernel::Thread>, Kernel::HLERequestContext&, Kernel::ThreadWakeupReason)>&&, std::__1::shared_ptr<Kernel::WritableEvent>)::$_0&, std::__1::shared_ptr<Kernel::Thread> >(Kernel::HLERequestContext::SleepClientThread(std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, unsigned long, std::__1::function<void (std::__1::shared_ptr<Kernel::Thread>, Kernel::HLERequestContext&, Kernel::ThreadWakeupReason)>&&, std::__1::shared_ptr<Kernel::WritableEvent>)::$_0&, std::__1::shared_ptr<Kernel::Thread>&&) at type_traits:3539:1 frame #17: 0x000000000110a532 yuzu`bool std::__1::__invoke_void_return_wrapper<bool>::__call<Kernel::HLERequestContext::SleepClientThread(__args=0x00000009d536e310, __args=nullptr)>&&, std::__1::shared_ptr<Kernel::WritableEvent>)::$_0&, std::__1::shared_ptr<Kernel::Thread> >(Kernel::HLERequestContext::SleepClientThread(std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, unsigned long, std::__1::function<void (std::__1::shared_ptr<Kernel::Thread>, Kernel::HLERequestContext&, Kernel::ThreadWakeupReason)>&&, std::__1::shared_ptr<Kernel::WritableEvent>)::$_0&, std::__1::shared_ptr<Kernel::Thread>&&) at __functional_base:317:16 frame #18: 0x000000000110a4f2 yuzu`std::__1::__function::__alloc_func<Kernel::HLERequestContext::SleepClientThread(std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, unsigned long, std::__1::function<void (std::__1::shared_ptr<Kernel::Thread>, Kernel::HLERequestContext&, Kernel::ThreadWakeupReason)>&&, std::__1::shared_ptr<Kernel::WritableEvent>)::$_0, std::__1::allocator<Kernel::HLERequestContext::SleepClientThread(std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, unsigned long, std::__1::function<void (std::__1::shared_ptr<Kernel::Thread>, Kernel::HLERequestContext&, Kernel::ThreadWakeupReason)>&&, std::__1::shared_ptr<Kernel::WritableEvent>)::$_0>, bool (std::__1::shared_ptr<Kernel::Thread>)>::operator(this=0x00000009d536e310, __arg=nullptr)(std::__1::shared_ptr<Kernel::Thread>&&) at functional:1540:16 frame #19: 0x00000000011094b3 yuzu`std::__1::__function::__func<Kernel::HLERequestContext::SleepClientThread(std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, unsigned long, std::__1::function<void (std::__1::shared_ptr<Kernel::Thread>, Kernel::HLERequestContext&, Kernel::ThreadWakeupReason)>&&, std::__1::shared_ptr<Kernel::WritableEvent>)::$_0, std::__1::allocator<Kernel::HLERequestContext::SleepClientThread(std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, unsigned long, std::__1::function<void (std::__1::shared_ptr<Kernel::Thread>, Kernel::HLERequestContext&, Kernel::ThreadWakeupReason)>&&, std::__1::shared_ptr<Kernel::WritableEvent>)::$_0>, bool (std::__1::shared_ptr<Kernel::Thread>)>::operator(this=0x00000009d536e300, __arg=nullptr)(std::__1::shared_ptr<Kernel::Thread>&&) at functional:1714:12 frame #20: 0x00000000011834ed yuzu`std::__1::__function::__value_func<bool (std::__1::shared_ptr<Kernel::Thread>)>::operator(this=0x000000090faee1c0, __args=nullptr)(std::__1::shared_ptr<Kernel::Thread>&&) const at functional:1867:16 frame #21: 0x0000000001180c18 yuzu`std::__1::function<bool (std::__1::shared_ptr<Kernel::Thread>)>::operator(this=0x000000090faee1c0, __arg=<unavailable>)(std::__1::shared_ptr<Kernel::Thread>) const at functional:2473:12 frame #22: 0x000000000117edb2 yuzu`Kernel::Thread::InvokeHLECallback(this=0x000000090faedc20, thread=nullptr) at thread.cpp:403:12 frame #23: 0x00000000011929d7 yuzu`Kernel::Svc::SendSyncRequest(system=0x000000000252f3d8, handle=622615) at svc.cpp:365:17 frame #24: 0x000000000118b3b5 yuzu`void Kernel::SvcWrap64<&(Kernel::Svc::SendSyncRequest(Core::System&, unsigned int))>(system=0x000000000252f3d8) at svc_wrap.h:50:24 frame #25: 0x000000000118a334 yuzu`Kernel::Svc::Call(system=0x000000000252f3d8, immediate=33) at svc.cpp:2649:13 frame #26: 0x00000000011a60e3 yuzu`Core::DynarmicCallbacks64::CallSVC(this=0x00000009c657df60, swi=33) at arm_dynarmic_64.cpp:123:9 frame #27: 0x00000000023f2c74 yuzu`Dynarmic::Backend::X64::impl::ThunkBuilder<void (Dynarmic::A64::UserCallbacks::*)(unsigned int), &(Dynarmic::A64::UserCallbacks::CallSVC(unsigned int))>::Thunk(this_=0x00000009c657df60, args=33) at devirtualize.h:28:16
While upstream issue is in userland driver FreeBSD kernel driver may fail to handle it gracefully e.g., $ sway $ mpv --hwdec=vaapi /path/to/file pid 12449 (sway), jid 0, uid 1234: exited on signal 11 WARNING !drm_modeset_is_locked(&crtc->mutex) failed at drm-kmod-drm_v5.4.62_3/drivers/gpu/drm/drm_atomic_helper.c:621 #0 0xffffffff81f3e833 at linux_dump_stack+0x23 #1 0xffffffff81ea43e3 at drm_atomic_helper_check_modeset+0xb3 #2 0xffffffff8259492c at intel_atomic_check+0x8c #3 0xffffffff81ea3379 at drm_atomic_check_only+0x3f9 #4 0xffffffff81ea37c3 at drm_atomic_commit+0x13 #5 0xffffffff81eb0c3f at drm_client_modeset_commit_atomic+0x1af #6 0xffffffff81eb093a at drm_client_modeset_commit_force+0x6a #7 0xffffffff81efa12a at drm_fb_helper_restore_fbdev_mode_unlocked+0x7a #8 0xffffffff81ef2386 at vt_kms_postswitch+0x166 #9 0xffffffff8076b66b at vt_window_switch+0x12b #10 0xffffffff807688ff at vtterm_cngrab+0x1f #11 0xffffffff80805a36 at cngrab+0x16 #12 0xffffffff8086295c at vpanic+0xec #13 0xffffffff80862863 at panic+0x43 #14 0xffffffff81f5d142 at dma_buf_poll+0x12 #15 0xffffffff808d2959 at kern_poll+0x329 #16 0xffffffff808d2620 at sys_poll+0x50 #17 0xffffffff80b3353b at amd64_syscall+0xfb
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
25 similar comments
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
29 similar comments
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Dependabot tried to update this pull request, but something went wrong. We're looking into it, but in the meantime you can retry the update by commenting |
Bumps lodash.merge from 4.6.1 to 4.6.2.
Commits
Dependabot will resolve any conflicts with this PR as long as you don't alter it yourself. You can also trigger a rebase manually by commenting
@dependabot rebase
.Dependabot commands and options
You can trigger Dependabot actions by commenting on this PR:
@dependabot rebase
will rebase this PR@dependabot recreate
will recreate this PR, overwriting any edits that have been made to it@dependabot merge
will merge this PR after your CI passes on it@dependabot squash and merge
will squash and merge this PR after your CI passes on it@dependabot cancel merge
will cancel a previously requested merge and block automerging@dependabot reopen
will reopen this PR if it is closed@dependabot close
will close this PR and stop Dependabot recreating it. You can achieve the same result by closing it manually@dependabot ignore this major version
will close this PR and stop Dependabot creating any more for this major version (unless you reopen the PR or upgrade to it yourself)@dependabot ignore this minor version
will close this PR and stop Dependabot creating any more for this minor version (unless you reopen the PR or upgrade to it yourself)@dependabot ignore this dependency
will close this PR and stop Dependabot creating any more for this dependency (unless you reopen the PR or upgrade to it yourself)@dependabot use these labels
will set the current labels as the default for future PRs for this repo and language@dependabot use these reviewers
will set the current reviewers as the default for future PRs for this repo and language@dependabot use these assignees
will set the current assignees as the default for future PRs for this repo and language@dependabot use this milestone
will set the current milestone as the default for future PRs for this repo and languageYou can disable automated security fix PRs for this repo from the Security Alerts page.