mirror of
https://github.com/PaddlePaddle/FastDeploy.git
synced 2026-04-23 00:17:25 +08:00
3651113ee5
* remove ENABLE_V1_DATA_PROCESSOR * fix unit test * fix unit test
586 lines
22 KiB
Python
586 lines
22 KiB
Python
"""
|
|
# Copyright (c) 2025 PaddlePaddle Authors. All Rights Reserved.
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License"
|
|
# you may not use this file except in compliance with the License.
|
|
# You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
"""
|
|
|
|
import os
|
|
import threading
|
|
import time
|
|
from abc import ABC, abstractmethod
|
|
from collections import defaultdict
|
|
from multiprocessing.reduction import ForkingPickler
|
|
|
|
import msgpack
|
|
import zmq
|
|
from zmq.utils import jsonapi
|
|
|
|
from fastdeploy import envs
|
|
from fastdeploy.metrics.metrics import main_process_metrics
|
|
from fastdeploy.metrics.stats import ZMQMetricsStats
|
|
from fastdeploy.utils import llm_logger
|
|
|
|
|
|
def _msgpack_default(obj):
|
|
"""Fallback serializer for types msgpack cannot handle natively."""
|
|
if hasattr(obj, "tolist"):
|
|
# paddle.Tensor, numpy.ndarray
|
|
return obj.tolist()
|
|
if hasattr(obj, "_asdict"):
|
|
# NamedTuple (LogprobsTensors, LogprobsLists, etc.)
|
|
return list(obj)
|
|
if hasattr(obj, "__dataclass_fields__"):
|
|
# dataclass (Logprob, SpeculateMetrics, etc.)
|
|
from dataclasses import asdict
|
|
|
|
return asdict(obj)
|
|
raise TypeError(f"Object of type {type(obj).__name__} is not msgpack serializable")
|
|
|
|
|
|
class ZmqServerBase(ABC):
|
|
"""
|
|
ZmqServerBase
|
|
"""
|
|
|
|
def __init__(self):
|
|
self.cached_results = defaultdict(list)
|
|
self.response_token_lock = threading.Lock()
|
|
self.address = None
|
|
self.response_handle_per_step = None
|
|
self.response_handle_name_per_step = None
|
|
self.batch_id_per_step = 0
|
|
|
|
@abstractmethod
|
|
def _create_socket(self):
|
|
"""Abstract method to create and return a ZeroMQ socket."""
|
|
pass
|
|
|
|
def _ensure_socket(self):
|
|
"""Ensure the socket is created before use."""
|
|
if self.socket is None:
|
|
self.socket: zmq.Socket = self._create_socket()
|
|
|
|
def send_json(self, data, flags: int = 0):
|
|
"""Send a Python object as a message using json to serialize.
|
|
|
|
Keyword arguments are passed on to json.dumps
|
|
|
|
Parameters
|
|
----------
|
|
obj : Python object
|
|
The Python object to send
|
|
flags : int
|
|
Any valid flags for :func:`Socket.send`
|
|
"""
|
|
_zmq_metrics_stats = ZMQMetricsStats()
|
|
try:
|
|
# package data with meta information
|
|
envelope = {"__meta": {"send_ts": time.perf_counter()}, "data": data}
|
|
msg = jsonapi.dumps(envelope)
|
|
|
|
# collect zmq send metrics
|
|
_zmq_metrics_stats.msg_bytes_send_total += len(msg)
|
|
|
|
return self.socket.send(msg, flags=flags)
|
|
except Exception as e:
|
|
_zmq_metrics_stats.msg_send_failed_total += 1
|
|
raise e
|
|
finally:
|
|
# collect zmq send metrics
|
|
_zmq_metrics_stats.msg_send_total += 1
|
|
main_process_metrics.record_zmq_stats(_zmq_metrics_stats, self.address)
|
|
|
|
def recv_json(self, flags: int = 0):
|
|
"""
|
|
Receive a JSON-serializable object from the socket.
|
|
"""
|
|
self._ensure_socket()
|
|
_zmq_metrics_stats = ZMQMetricsStats()
|
|
try:
|
|
# receive from socket
|
|
msg = self.socket.recv(flags=flags)
|
|
data_dict = self.socket._deserialize(msg, lambda buf: jsonapi.loads(buf))
|
|
|
|
# collect zmq recv metrics
|
|
_zmq_metrics_stats.msg_bytes_recv_total += len(msg)
|
|
_zmq_metrics_stats.msg_recv_total += 1
|
|
|
|
# first check if the received msg is a dict
|
|
if isinstance(data_dict, dict):
|
|
# then check if the dict has "__meta" key
|
|
if "__meta" in data_dict and "send_ts" in data_dict["__meta"]:
|
|
# if so, calculate the delay
|
|
_zmq_metrics_stats.zmq_latency = time.perf_counter() - data_dict["__meta"]["send_ts"]
|
|
return data_dict["data"]
|
|
return data_dict
|
|
finally:
|
|
main_process_metrics.record_zmq_stats(_zmq_metrics_stats, self.address)
|
|
|
|
def send_pyobj(self, data, flags: int = 0):
|
|
"""
|
|
Send a Pickle-serializable object over the socket.
|
|
"""
|
|
self._ensure_socket()
|
|
_zmq_metrics_stats = ZMQMetricsStats()
|
|
try:
|
|
envelope = {"__meta": {"send_ts": time.perf_counter()}, "data": data}
|
|
data_bytes = ForkingPickler.dumps(envelope)
|
|
_zmq_metrics_stats.msg_bytes_send_total += len(data_bytes)
|
|
self.socket.send(data_bytes, copy=False, flags=flags)
|
|
except Exception as e:
|
|
_zmq_metrics_stats.msg_send_failed_total += 1
|
|
raise e
|
|
finally:
|
|
_zmq_metrics_stats.msg_send_total += 1
|
|
main_process_metrics.record_zmq_stats(_zmq_metrics_stats, self.address)
|
|
|
|
def recv_pyobj(self, flags: int = 0):
|
|
"""
|
|
Receive a Pickle-serializable object from the socket.
|
|
"""
|
|
_zmq_metrics_stats = ZMQMetricsStats()
|
|
self._ensure_socket()
|
|
data_bytes = self.socket.recv(flags=flags)
|
|
envelope = ForkingPickler.loads(data_bytes)
|
|
if isinstance(envelope, dict):
|
|
if "__meta" in envelope and "send_ts" in envelope["__meta"]:
|
|
_zmq_metrics_stats.msg_recv_total += 1
|
|
_zmq_metrics_stats.msg_bytes_recv_total += len(data_bytes)
|
|
_zmq_metrics_stats.zmq_latency = time.perf_counter() - envelope["__meta"]["send_ts"]
|
|
main_process_metrics.record_zmq_stats(_zmq_metrics_stats, self.address)
|
|
return envelope["data"]
|
|
return envelope
|
|
|
|
def pack_aggregated_data(self, data):
|
|
"""
|
|
Aggregate multiple responses into one and send them to the client.
|
|
"""
|
|
result = data[0]
|
|
if len(data) > 1:
|
|
for response in data[1:]:
|
|
result.add(response)
|
|
result = ForkingPickler.dumps([result.to_dict()])
|
|
return result
|
|
|
|
def receive_json_once(self, block=False):
|
|
"""
|
|
Receive a single message from the socket.
|
|
"""
|
|
self._ensure_socket()
|
|
if self.socket is None or self.socket.closed:
|
|
return "zmp socket has closed", None
|
|
try:
|
|
flags = zmq.NOBLOCK if not block else 0
|
|
return None, self.recv_json(flags=flags)
|
|
except zmq.Again:
|
|
return None, None
|
|
except Exception as e:
|
|
self.close()
|
|
llm_logger.warning(f"{e}")
|
|
return str(e), None
|
|
|
|
def receive_pyobj_once(self, block=False):
|
|
"""
|
|
Receive a single message from the socket.
|
|
"""
|
|
self._ensure_socket()
|
|
if self.socket is None or self.socket.closed:
|
|
return "zmp socket has closed", None
|
|
try:
|
|
flags = zmq.NOBLOCK if not block else 0
|
|
return None, self.recv_pyobj(flags=flags)
|
|
except zmq.Again:
|
|
return None, None
|
|
except Exception as e:
|
|
self.close()
|
|
llm_logger.warning(f"{e}")
|
|
return str(e), None
|
|
|
|
def recv_result_handle(self):
|
|
while self.running:
|
|
try:
|
|
with self.response_token_lock:
|
|
client, _, request_id = self.socket.recv_multipart(flags=zmq.NOBLOCK)
|
|
req_id_str = request_id.decode("utf-8")
|
|
if envs.FD_ENABLE_INTERNAL_ADAPTER:
|
|
with self.mutex:
|
|
self.response_handle_per_step = client
|
|
else:
|
|
need_send_after_finished_inference = False
|
|
with self.mutex:
|
|
self.req_dict[req_id_str] = client
|
|
if req_id_str in self.cached_results:
|
|
if self.cached_results[req_id_str][-1][-1].finished:
|
|
need_send_after_finished_inference = True
|
|
if need_send_after_finished_inference:
|
|
self.send_response(req_id_str, [])
|
|
llm_logger.info(f"send_multipart finished, req_id: {req_id_str}")
|
|
self.req_dict.pop(req_id_str, None)
|
|
|
|
except zmq.Again:
|
|
time.sleep(0.001)
|
|
continue
|
|
except zmq.error.ZMQError as e:
|
|
llm_logger.error(f"recv_result_handle get zmq error: {e}")
|
|
break
|
|
except Exception as e:
|
|
llm_logger.error(f"recv_result_handle get unknown exception: {e}")
|
|
continue
|
|
|
|
def _send_response_per_step(self, batch_id, data):
|
|
"""
|
|
Send generated token result to client.
|
|
"""
|
|
self._ensure_socket()
|
|
if self.socket is None:
|
|
raise RuntimeError("Router socket not created. Call create_router() first.")
|
|
need_send_data = []
|
|
with self.mutex:
|
|
if self.response_handle_per_step is None:
|
|
self.cached_results["data"].extend(data)
|
|
else:
|
|
need_send_data = self.cached_results["data"]
|
|
self.cached_results["data"] = []
|
|
if self.response_handle_per_step is not None:
|
|
try:
|
|
if data:
|
|
need_send_data.extend(data)
|
|
start_send = time.time()
|
|
result = msgpack.packb(
|
|
[[response.to_dict() for response in send_data_list] for send_data_list in need_send_data]
|
|
)
|
|
with self.response_token_lock:
|
|
self.socket.send_multipart([self.response_handle_per_step, b"", result])
|
|
llm_logger.info(
|
|
f"send_multipart result: step {self.batch_id_per_step} lens {len(need_send_data)} elapse: {time.time()-start_send}"
|
|
)
|
|
self.batch_id_per_step += 1
|
|
|
|
except Exception as e:
|
|
llm_logger.error(f"Send result to zmq client failed: {e}")
|
|
|
|
def _send_response_per_query(self, req_id, data):
|
|
"""
|
|
Send generated token result to client.
|
|
"""
|
|
self._ensure_socket()
|
|
if self.socket is None:
|
|
raise RuntimeError("Router socket not created. Call create_router() first.")
|
|
new_data = []
|
|
has_result_handle = False
|
|
with self.mutex:
|
|
if req_id not in self.req_dict:
|
|
llm_logger.warning(
|
|
f"req_id '{req_id}' not in req_dict, caching response. Available req_ids: {list(self.req_dict.keys())}"
|
|
)
|
|
self.cached_results[req_id].append(data)
|
|
else:
|
|
has_result_handle = True
|
|
if req_id in self.cached_results:
|
|
for history_data in self.cached_results[req_id]:
|
|
new_data.extend(history_data)
|
|
llm_logger.info(
|
|
f"get request {req_id} result handle after cached result, total cached length {len(self.cached_results[req_id])}"
|
|
)
|
|
del self.cached_results[req_id]
|
|
if has_result_handle:
|
|
try:
|
|
new_data.extend(data)
|
|
start_send = time.time()
|
|
if self.aggregate_send:
|
|
result = self.pack_aggregated_data(new_data)
|
|
else:
|
|
result = ForkingPickler.dumps([response.to_dict() for response in new_data])
|
|
with self.response_token_lock:
|
|
|
|
_zmq_metrics_stats = ZMQMetricsStats()
|
|
try:
|
|
self.socket.send_multipart([self.req_dict[req_id], b"", result], copy=False)
|
|
_zmq_metrics_stats.msg_bytes_send_total += len(result)
|
|
except Exception as e:
|
|
_zmq_metrics_stats.msg_send_failed_total += 1
|
|
raise e
|
|
finally:
|
|
_zmq_metrics_stats.msg_send_total += 1
|
|
main_process_metrics.record_zmq_stats(_zmq_metrics_stats, self.address)
|
|
llm_logger.debug(
|
|
f"send_multipart result: {req_id} len {len(new_data)} elapse: {time.time()-start_send}"
|
|
)
|
|
|
|
except Exception as e:
|
|
llm_logger.error(f"Send result to zmq client failed: {e}")
|
|
|
|
if data and data[-1].finished:
|
|
with self.mutex:
|
|
if req_id in self.req_dict:
|
|
llm_logger.info(f"send_multipart finished, req_id: {req_id}")
|
|
self.req_dict.pop(req_id, None)
|
|
|
|
def _send_batch_response(self, batch_data, worker_pid=None):
|
|
"""
|
|
Batch send responses for multiple requests to a specific worker.
|
|
batch_data: List[List[output, ...], ...]
|
|
worker_pid: Target worker process ID for routing.
|
|
"""
|
|
if worker_pid is not None:
|
|
sock = self._get_worker_push_socket(worker_pid)
|
|
elif self.socket is not None:
|
|
sock = self.socket
|
|
else:
|
|
llm_logger.error("No PUSH socket available: worker_pid is required in batch mode")
|
|
return
|
|
|
|
metrics_address = self.address or self.worker_push_addresses.get(worker_pid, "unknown")
|
|
|
|
try:
|
|
result = msgpack.packb(
|
|
[[output.to_dict() for output in outputs] for outputs in batch_data],
|
|
default=_msgpack_default,
|
|
)
|
|
result_len = len(result)
|
|
|
|
# Only hold lock for the actual socket send
|
|
start_send = time.time()
|
|
_zmq_metrics_stats = ZMQMetricsStats()
|
|
with self.response_token_lock:
|
|
try:
|
|
sock.send(result, copy=False)
|
|
_zmq_metrics_stats.msg_bytes_send_total += result_len
|
|
except Exception as e:
|
|
_zmq_metrics_stats.msg_send_failed_total += 1
|
|
raise e
|
|
finally:
|
|
_zmq_metrics_stats.msg_send_total += 1
|
|
main_process_metrics.record_zmq_stats(_zmq_metrics_stats, metrics_address)
|
|
|
|
llm_logger.debug(
|
|
f"send_batch to worker {worker_pid}: {len(batch_data)} requests, elapse: {time.time() - start_send}"
|
|
)
|
|
|
|
except Exception as e:
|
|
llm_logger.error(f"Send batch response to worker {worker_pid} failed: {e}")
|
|
|
|
def send_response(self, req_id, data, worker_pid=None):
|
|
"""
|
|
Unified response sending interface.
|
|
|
|
Args:
|
|
req_id: Request ID (None for batch mode)
|
|
data: Response data
|
|
- Internal Adapter mode: List[List[output]] (batch per step)
|
|
- Batch mode (ZMQ_SEND_BATCH_DATA=1): List[List[output]], req_id is always None
|
|
- Per-query mode (ZMQ_SEND_BATCH_DATA=0): List[output], sent via ROUTER per request
|
|
worker_pid: Target worker process ID (batch mode only)
|
|
"""
|
|
if envs.FD_ENABLE_INTERNAL_ADAPTER:
|
|
self._send_response_per_step(req_id, data)
|
|
elif envs.ZMQ_SEND_BATCH_DATA:
|
|
self._send_batch_response(data, worker_pid=worker_pid)
|
|
else:
|
|
self._send_response_per_query(req_id, data)
|
|
|
|
@abstractmethod
|
|
def close(self):
|
|
pass
|
|
|
|
def __exit__(self, exc_type, exc_val, exc_tb):
|
|
self.close()
|
|
|
|
|
|
class ZmqIpcServer(ZmqServerBase):
|
|
"""
|
|
ZmqIpcServer, used when FD_ENABLE_INTERNAL_ADAPTER=0
|
|
"""
|
|
|
|
def __init__(self, name, mode):
|
|
super(ZmqIpcServer, self).__init__()
|
|
self.name = name
|
|
self.mode = mode
|
|
self.cached_results = defaultdict(list)
|
|
self.ZMQ_SNDHWM = int(envs.FD_ZMQ_SNDHWM)
|
|
self.aggregate_send = envs.FD_USE_AGGREGATE_SEND
|
|
self.mutex = threading.Lock()
|
|
self.response_token_lock = threading.Lock()
|
|
self.req_dict = dict()
|
|
self.running = True
|
|
self.context = zmq.Context()
|
|
|
|
# Per-worker PUSH sockets for batch mode (worker_pid -> zmq.Socket)
|
|
self.worker_push_sockets = {}
|
|
self.worker_push_addresses = {}
|
|
self.worker_push_lock = threading.Lock()
|
|
|
|
if mode == zmq.PUSH:
|
|
# Batch mode: don't bind a single socket here.
|
|
# Per-worker PUSH sockets are created on demand via _get_worker_push_socket().
|
|
self.push_name_prefix = name
|
|
self.socket = None
|
|
self.address = None
|
|
self.file_name = None
|
|
return
|
|
|
|
if mode == zmq.PULL:
|
|
self.file_name = f"/dev/shm/{name}.socket"
|
|
elif mode == zmq.ROUTER:
|
|
self.file_name = f"/dev/shm/router_{name}.ipc"
|
|
else:
|
|
raise ValueError(f"Unsupported ZMQ mode: {mode}")
|
|
self._create_socket()
|
|
|
|
def _create_socket(self):
|
|
"""create and return a ZeroMQ socket."""
|
|
self.socket = self.context.socket(self.mode)
|
|
self.socket.setsockopt(zmq.SNDHWM, self.ZMQ_SNDHWM)
|
|
self.socket.setsockopt(zmq.SNDTIMEO, -1)
|
|
self.address = f"ipc://{self.file_name}"
|
|
self.socket.bind(self.address)
|
|
return self.socket
|
|
|
|
def _get_worker_push_socket(self, worker_pid):
|
|
"""Get or create a PUSH socket for the given worker (batch mode only)."""
|
|
with self.worker_push_lock:
|
|
if worker_pid in self.worker_push_sockets:
|
|
return self.worker_push_sockets[worker_pid]
|
|
|
|
sock = self.context.socket(zmq.PUSH)
|
|
sock.setsockopt(zmq.SNDHWM, self.ZMQ_SNDHWM)
|
|
sock.setsockopt(zmq.SNDTIMEO, -1)
|
|
address = f"ipc:///dev/shm/response_{self.push_name_prefix}_w{worker_pid}.pull"
|
|
sock.connect(address)
|
|
self.worker_push_sockets[worker_pid] = sock
|
|
self.worker_push_addresses[worker_pid] = address
|
|
llm_logger.info(f"PUSH connected to worker {worker_pid}: {address}")
|
|
return sock
|
|
|
|
def _clear_ipc(self, name):
|
|
"""
|
|
Remove the IPC file with the given name.
|
|
"""
|
|
if os.path.exists(name):
|
|
try:
|
|
os.remove(name)
|
|
except OSError as e:
|
|
llm_logger.warning(f"Failed to remove IPC file {name} - {e}")
|
|
|
|
def close(self):
|
|
"""
|
|
Close the socket and context, and remove the IPC files.
|
|
"""
|
|
if not self.running:
|
|
return
|
|
|
|
self.running = False
|
|
llm_logger.info("ZMQ server is closing connection...")
|
|
try:
|
|
# Close per-worker PUSH sockets (batch mode)
|
|
with self.worker_push_lock:
|
|
for wpid, sock in self.worker_push_sockets.items():
|
|
try:
|
|
sock.close()
|
|
except Exception:
|
|
pass
|
|
self.worker_push_sockets.clear()
|
|
self.worker_push_addresses.clear()
|
|
|
|
if self.socket is not None and not self.socket.closed:
|
|
self.socket.close()
|
|
if not self.context.closed:
|
|
self.context.term()
|
|
if self.file_name:
|
|
self._clear_ipc(self.file_name)
|
|
except Exception as e:
|
|
llm_logger.warning(f"ZMQ server failed to close connection - {e}")
|
|
return
|
|
|
|
|
|
class ZmqTcpServer(ZmqServerBase):
|
|
"""
|
|
ZmqTcpServer, used when FD_ENABLE_INTERNAL_ADAPTER=1
|
|
"""
|
|
|
|
def __init__(self, port, mode):
|
|
super(ZmqTcpServer, self).__init__()
|
|
self.mode = mode
|
|
self.port = port
|
|
self.cached_results = defaultdict(list)
|
|
self.ZMQ_SNDHWM = int(envs.FD_ZMQ_SNDHWM)
|
|
self.aggregate_send = envs.FD_USE_AGGREGATE_SEND
|
|
|
|
self.mutex = threading.Lock()
|
|
self.req_dict = dict()
|
|
self.running = True
|
|
self.context = zmq.Context()
|
|
self._create_socket()
|
|
self.response_token_lock = threading.Lock()
|
|
|
|
def _create_socket(self):
|
|
"""create and return a ZeroMQ socket."""
|
|
self.socket = self.context.socket(self.mode)
|
|
self.socket.setsockopt(zmq.SNDHWM, self.ZMQ_SNDHWM)
|
|
self.socket.setsockopt(zmq.SNDTIMEO, -1)
|
|
self.address = f"tcp://*:{self.port}"
|
|
self.socket.bind(self.address)
|
|
return self.socket
|
|
|
|
def recv_control_cmd(self):
|
|
"""
|
|
Receive control command from client
|
|
"""
|
|
self._ensure_socket()
|
|
try:
|
|
client, _, task_data = self.socket.recv_multipart(flags=zmq.NOBLOCK)
|
|
task = msgpack.unpackb(task_data)
|
|
task_id_str = task["task_id"]
|
|
except zmq.Again:
|
|
return None
|
|
with self.mutex:
|
|
self.req_dict[task_id_str] = client
|
|
return task
|
|
|
|
def response_for_control_cmd(self, task_id, result):
|
|
"""
|
|
Send command result back to client.
|
|
"""
|
|
self._ensure_socket()
|
|
if self.socket is None:
|
|
raise RuntimeError("Router socket not created.")
|
|
try:
|
|
result = msgpack.packb(result)
|
|
self.socket.send_multipart([self.req_dict[task_id], b"", result])
|
|
|
|
except Exception as e:
|
|
llm_logger.error(f"Send result to zmq client failed: {e}")
|
|
|
|
with self.mutex:
|
|
self.req_dict.pop(task_id, None)
|
|
llm_logger.debug(f"response control cmd finished, task_id: {task_id}")
|
|
|
|
def close(self):
|
|
"""
|
|
Close the socket and context.
|
|
"""
|
|
if not self.running:
|
|
return
|
|
|
|
self.running = False
|
|
llm_logger.info("ZMQ server is closing connection...")
|
|
try:
|
|
if self.socket is not None and not self.socket.closed:
|
|
self.socket.close()
|
|
if not self.context.closed:
|
|
self.context.term()
|
|
|
|
except Exception as e:
|
|
llm_logger.warning(f"ZMQ server failed to close connection - {e}")
|
|
return
|