Created attachment 1133203 [details] pstack of evolution-source-registry Description of problem: I have GNOME set up to sync my calendar with Google. Today I noticed that the evolution-source-registry seems to be stuck in an endless loop. It's eating 100% CPU time on one core. Version-Release number of selected component (if applicable): evolution 3.18.5.1 How reproducible: I haven't killed it yet or rebooted, so I don't know how reproducible it is. Steps to Reproduce: 1. Set up Google Account to sync calender 2. never actually use evolution for email 3. reboot a few times Additional info: I have attached the output of `pstack <pid>` and also saved the gcore on my computer (900 MB though).
Thanks for a bug report. This sounds similar to an upstream bug [1]. The attached backtrace shows all threads idle, polling/waiting to do something. An application in this state might not cause any high CPU usage. You can catch a set of backtraces when using this command: for i in {1..10}; do \ gdb --batch --ex "t a a bt" -pid=`pidof evolution-source-registry` &>bt$i.txt; \ sleep 0.1; \ done Eventually play with the sleep argument to set the best interval between the captures. Please check the bt?.txt for any private information, like passwords, email address, server addresses,... I usually search for "pass" at least (quotes for clarity only). Also make sure you'll have installed a debuginfo package for the evolution-data-server (it's enough to have installed that one, without dependencies), thus the backtrace will contain line numbers from the evolution-data-server itself. [1] https://bugzilla.gnome.org/show_bug.cgi?id=761450
Thanks! Unfortunately I couldn't reproduce this yet. When it happens again, I will get the backtraces and reopen this bug :)
Created attachment 1133918 [details] Backtraces using gdb
Okay I've rebooted just now and it happened again. I've attached the txt files from your gdb loop. There's now also a seconds process called "evolution-calender-factory-subprocess". Should I attach backtraces of this one, too? Thanks in advance!
Thanks for the update. All the backtraces are the same, in a poll calls, thus "waiting for orders". I do not know which 'top' or 'system monitor' tells you that the soruce registry is using the CPU extensively, but I'd rather expect the calendar factory being the cause here, specifically the subprocess for your calendar(s). See the upstream bug I gave a link to in comment #1, there are some findings which I believe are related to you too. I tend to close this in a favour of this upstream bug report.
This message is a reminder that Fedora 23 is nearing its end of life. Approximately 4 (four) weeks from now Fedora will stop maintaining and issuing updates for Fedora 23. It is Fedora's policy to close all bug reports from releases that are no longer maintained. At that time this bug will be closed as EOL if it remains open with a Fedora 'version' of '23'. Package Maintainer: If you wish for this bug to remain open because you plan to fix it in a currently maintained version, simply change the 'version' to a later Fedora version. Thank you for reporting this issue and we are sorry that we were not able to fix it before Fedora 23 is end of life. If you would still like to see this bug fixed and are able to reproduce it against a later version of Fedora, you are encouraged change the 'version' to a later Fedora version prior this bug is closed as described in the policy above. Although we aim to fix as many bugs as possible during every release's lifetime, sometimes those efforts are overtaken by events. Often a more recent Fedora release includes newer upstream software that fixes bugs or makes them obsolete.
Fedora 23 changed to end-of-life (EOL) status on 2016-12-20. Fedora 23 is no longer maintained, which means that it will not receive any further security or bug fix updates. As a result we are closing this bug. If you can reproduce this bug against a currently maintained version of Fedora please feel free to reopen this bug against that version. If you are unable to reopen this bug, please file a new report against the current release. If you experience problems, please add a comment to this bug. Thank you for reporting this bug and we are sorry it could not be fixed.