%PDF- %PDF-
Direktori : /usr/share/gdb/python/gdb/dap/ |
Current File : //usr/share/gdb/python/gdb/dap/server.py |
# Copyright 2022-2024 Free Software Foundation, Inc. # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation; either version 3 of the License, or # (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. import functools import heapq import inspect import json import threading from contextlib import contextmanager import gdb from .io import read_json, start_json_writer from .startup import ( DAPException, DAPQueue, LogLevel, exec_and_log, in_dap_thread, in_gdb_thread, log, log_stack, start_thread, thread_log, ) from .typecheck import type_check # Map capability names to values. _capabilities = {} # Map command names to callables. _commands = {} # The global server. _server = None # A subclass of Exception that is used solely for reporting that a # request needs the inferior to be stopped, but it is not stopped. class NotStoppedException(Exception): pass # This is used to handle cancellation requests. It tracks all the # needed state, so that we can cancel both requests that are in flight # as well as queued requests. class CancellationHandler: def __init__(self): # Methods on this class acquire this lock before proceeding. self.lock = threading.Lock() # The request currently being handled, or None. self.in_flight_dap_thread = None self.in_flight_gdb_thread = None self.reqs = [] def starting(self, req): """Call at the start of the given request.""" with self.lock: self.in_flight_dap_thread = req def done(self, req): """Indicate that the request is done.""" with self.lock: self.in_flight_dap_thread = None def cancel(self, req): """Call to cancel a request. If the request has already finished, this is ignored. If the request is in flight, it is interrupted. If the request has not yet been seen, the cancellation is queued.""" with self.lock: if req == self.in_flight_gdb_thread: gdb.interrupt() else: # We don't actually ignore the request here, but in # the 'starting' method. This way we don't have to # track as much state. Also, this implementation has # the weird property that a request can be cancelled # before it is even sent. It didn't seem worthwhile # to try to check for this. heapq.heappush(self.reqs, req) @contextmanager def interruptable_region(self, req): """Return a new context manager that sets in_flight_gdb_thread to REQ.""" if req is None: # No request is handled in the region, just execute the region. yield return try: with self.lock: # If the request is cancelled, don't execute the region. while len(self.reqs) > 0 and self.reqs[0] <= req: if heapq.heappop(self.reqs) == req: raise KeyboardInterrupt() # Request is being handled by the gdb thread. self.in_flight_gdb_thread = req # Execute region. This may be interrupted by gdb.interrupt. yield finally: with self.lock: # Request is no longer handled by the gdb thread. self.in_flight_gdb_thread = None class Server: """The DAP server class.""" def __init__(self, in_stream, out_stream, child_stream): self.in_stream = in_stream self.out_stream = out_stream self.child_stream = child_stream self.delayed_events = [] # This queue accepts JSON objects that are then sent to the # DAP client. Writing is done in a separate thread to avoid # blocking the read loop. self.write_queue = DAPQueue() # Reading is also done in a separate thread, and a queue of # requests is kept. self.read_queue = DAPQueue() self.done = False self.canceller = CancellationHandler() global _server _server = self # Treat PARAMS as a JSON-RPC request and perform its action. # PARAMS is just a dictionary from the JSON. @in_dap_thread def _handle_command(self, params): req = params["seq"] result = { "request_seq": req, "type": "response", "command": params["command"], } try: self.canceller.starting(req) if "arguments" in params: args = params["arguments"] else: args = {} global _commands body = _commands[params["command"]](**args) if body is not None: result["body"] = body result["success"] = True except NotStoppedException: # This is an expected exception, and the result is clearly # visible in the log, so do not log it. result["success"] = False result["message"] = "notStopped" except KeyboardInterrupt: # This can only happen when a request has been canceled. result["success"] = False result["message"] = "cancelled" except DAPException as e: # Don't normally want to see this, as it interferes with # the test suite. log_stack(LogLevel.FULL) result["success"] = False result["message"] = str(e) except BaseException as e: log_stack() result["success"] = False result["message"] = str(e) self.canceller.done(req) return result # Read inferior output and sends OutputEvents to the client. It # is run in its own thread. def _read_inferior_output(self): while True: line = self.child_stream.readline() self.send_event( "output", { "category": "stdout", "output": line, }, ) # Send OBJ to the client, logging first if needed. def _send_json(self, obj): log("WROTE: <<<" + json.dumps(obj) + ">>>") self.write_queue.put(obj) # This is run in a separate thread and simply reads requests from # the client and puts them into a queue. A separate thread is # used so that 'cancel' requests can be handled -- the DAP thread # will normally block, waiting for each request to complete. def _reader_thread(self): while True: cmd = read_json(self.in_stream) if cmd is None: break log("READ: <<<" + json.dumps(cmd) + ">>>") # Be extra paranoid about the form here. If anything is # missing, it will be put in the queue and then an error # issued by ordinary request processing. if ( "command" in cmd and cmd["command"] == "cancel" and "arguments" in cmd # gdb does not implement progress, so there's no need # to check for progressId. and "requestId" in cmd["arguments"] ): self.canceller.cancel(cmd["arguments"]["requestId"]) self.read_queue.put(cmd) # When we hit EOF, signal it with None. self.read_queue.put(None) @in_dap_thread def main_loop(self): """The main loop of the DAP server.""" # Before looping, start the thread that writes JSON to the # client, and the thread that reads output from the inferior. start_thread("output reader", self._read_inferior_output) json_writer = start_json_writer(self.out_stream, self.write_queue) start_thread("JSON reader", self._reader_thread) while not self.done: cmd = self.read_queue.get() # A None value here means the reader hit EOF. if cmd is None: break result = self._handle_command(cmd) self._send_json(result) events = self.delayed_events self.delayed_events = [] for event, body in events: self.send_event(event, body) # Got the terminate request. This is handled by the # JSON-writing thread, so that we can ensure that all # responses are flushed to the client before exiting. self.write_queue.put(None) json_writer.join() send_gdb("quit") @in_dap_thread def send_event_later(self, event, body=None): """Send a DAP event back to the client, but only after the current request has completed.""" self.delayed_events.append((event, body)) # Note that this does not need to be run in any particular thread, # because it just creates an object and writes it to a thread-safe # queue. def send_event(self, event, body=None): """Send an event to the DAP client. EVENT is the name of the event, a string. BODY is the body of the event, an arbitrary object.""" obj = { "type": "event", "event": event, } if body is not None: obj["body"] = body self._send_json(obj) def shutdown(self): """Request that the server shut down.""" # Just set a flag. This operation is complicated because we # want to write the result of the request before exiting. See # main_loop. self.done = True def send_event(event, body=None): """Send an event to the DAP client. EVENT is the name of the event, a string. BODY is the body of the event, an arbitrary object.""" global _server _server.send_event(event, body) # A helper decorator that checks whether the inferior is running. def _check_not_running(func): @functools.wraps(func) def check(*args, **kwargs): # Import this as late as possible. This is done to avoid # circular imports. from .events import inferior_running if inferior_running: raise NotStoppedException() return func(*args, **kwargs) return check def request( name: str, *, response: bool = True, on_dap_thread: bool = False, expect_stopped: bool = True ): """A decorator for DAP requests. This registers the function as the implementation of the DAP request NAME. By default, the function is invoked in the gdb thread, and its result is returned as the 'body' of the DAP response. Some keyword arguments are provided as well: If RESPONSE is False, the result of the function will not be waited for and no 'body' will be in the response. If ON_DAP_THREAD is True, the function will be invoked in the DAP thread. When ON_DAP_THREAD is True, RESPONSE may not be False. If EXPECT_STOPPED is True (the default), then the request will fail with the 'notStopped' reason if it is processed while the inferior is running. When EXPECT_STOPPED is False, the request will proceed regardless of the inferior's state. """ # Validate the parameters. assert not on_dap_thread or response def wrap(func): code = func.__code__ # We don't permit requests to have positional arguments. try: assert code.co_posonlyargcount == 0 except AttributeError: # Attribute co_posonlyargcount is supported starting python 3.8. pass assert code.co_argcount == 0 # A request must have a **args parameter. assert code.co_flags & inspect.CO_VARKEYWORDS # Type-check the calls. func = type_check(func) # Verify that the function is run on the correct thread. if on_dap_thread: cmd = in_dap_thread(func) else: func = in_gdb_thread(func) if response: def sync_call(**args): return send_gdb_with_response(lambda: func(**args)) cmd = sync_call else: def non_sync_call(**args): return send_gdb(lambda: func(**args)) cmd = non_sync_call # If needed, check that the inferior is not running. This # wrapping is done last, so the check is done first, before # trying to dispatch the request to another thread. if expect_stopped: cmd = _check_not_running(cmd) global _commands assert name not in _commands _commands[name] = cmd return cmd return wrap def capability(name, value=True): """A decorator that indicates that the wrapper function implements the DAP capability NAME.""" def wrap(func): global _capabilities assert name not in _capabilities _capabilities[name] = value return func return wrap def client_bool_capability(name): """Return the value of a boolean client capability. If the capability was not specified, or did not have boolean type, False is returned.""" global _server if name in _server.config and isinstance(_server.config[name], bool): return _server.config[name] return False @request("initialize", on_dap_thread=True) def initialize(**args): global _server, _capabilities _server.config = args _server.send_event_later("initialized") return _capabilities.copy() @request("terminate", expect_stopped=False) @capability("supportsTerminateRequest") def terminate(**args): exec_and_log("kill") @request("disconnect", on_dap_thread=True, expect_stopped=False) @capability("supportTerminateDebuggee") def disconnect(*, terminateDebuggee: bool = False, **args): if terminateDebuggee: send_gdb_with_response("kill") _server.shutdown() @request("cancel", on_dap_thread=True, expect_stopped=False) @capability("supportsCancelRequest") def cancel(**args): # If a 'cancel' request can actually be satisfied, it will be # handled specially in the reader thread. However, in order to # construct a proper response, the request is also added to the # command queue and so ends up here. Additionally, the spec says: # The cancel request may return an error if it could not cancel # an operation but a client should refrain from presenting this # error to end users. # ... which gdb takes to mean that it is fine for all cancel # requests to report success. return None class Invoker(object): """A simple class that can invoke a gdb command.""" def __init__(self, cmd): self.cmd = cmd # This is invoked in the gdb thread to run the command. @in_gdb_thread def __call__(self): exec_and_log(self.cmd) class Cancellable(object): def __init__(self, fn, result_q=None): self.fn = fn self.result_q = result_q with _server.canceller.lock: self.req = _server.canceller.in_flight_dap_thread # This is invoked in the gdb thread to run self.fn. @in_gdb_thread def __call__(self): try: with _server.canceller.interruptable_region(self.req): val = self.fn() if self.result_q is not None: self.result_q.put(val) except (Exception, KeyboardInterrupt) as e: if self.result_q is not None: # Pass result or exception to caller. self.result_q.put(e) elif isinstance(e, KeyboardInterrupt): # Fn was cancelled. pass else: # Exception happened. Ignore and log it. err_string = "%s, %s" % (e, type(e)) thread_log("caught exception: " + err_string) log_stack() def send_gdb(cmd): """Send CMD to the gdb thread. CMD can be either a function or a string. If it is a string, it is passed to gdb.execute.""" if isinstance(cmd, str): cmd = Invoker(cmd) # Post the event and don't wait for the result. gdb.post_event(Cancellable(cmd)) def send_gdb_with_response(fn): """Send FN to the gdb thread and return its result. If FN is a string, it is passed to gdb.execute and None is returned as the result. If FN throws an exception, this function will throw the same exception in the calling thread. """ if isinstance(fn, str): fn = Invoker(fn) # Post the event and wait for the result in result_q. result_q = DAPQueue() gdb.post_event(Cancellable(fn, result_q)) val = result_q.get() if isinstance(val, (Exception, KeyboardInterrupt)): raise val return val