[lldb-dev] How to debug LLDB server?
Greg Clayton via lldb-dev
lldb-dev at lists.llvm.org
Wed Oct 7 11:04:45 PDT 2015
Most calls for lldb-server should use an instance variable GDBRemoteCommunication::m_packet_timeout which you could then modify. But this timeout you are talking about is the time that the expression can take when running. I would just bump these up temporarily while you are debugging to avoid the timeouts. Just don't check it in.
So for GDB Remote packets, we already bump the timeout up in the GDBRemoteCommunication constructor:
#ifdef LLDB_CONFIGURATION_DEBUG
m_packet_timeout (1000),
#else
m_packet_timeout (1),
#endif
Anything else is probably expression timeouts and you will need to manually bump those up in order to debug, or you could do the same thing as the GDB Remote in InferiorCallPOSIX.cpp:
#ifdef LLDB_CONFIGURATION_DEBUG
options.SetTimeoutUsec(50000000);
#else
options.SetTimeoutUsec(500000);
#endif
> On Oct 7, 2015, at 10:33 AM, Eugene Birukov via lldb-dev <lldb-dev at lists.llvm.org> wrote:
>
> Hello,
>
> I am trying to see what is going inside LLDB server 3.7.0 but there are a lot of timeouts scattered everywhere. Say, InferiorCallPOSIX.cpp:74 sets hard-coded timeout to 500,000us, etc. These timeouts fire if I spend any time on breakpoint inside server and make debugging experience miserable. Is there any way to turn them all off?
>
> BTW, I am using LLDB as a C++ API, not as standalone program, but I have debugger attached to it and can alter its memory state.
>
> Thanks,
> Eugene
>
> _______________________________________________
> lldb-dev mailing list
> lldb-dev at lists.llvm.org
> http://lists.llvm.org/cgi-bin/mailman/listinfo/lldb-dev
More information about the lldb-dev
mailing list