Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

feat: support server sent event #333

Merged
merged 13 commits into from
Jun 15, 2023
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
39 changes: 31 additions & 8 deletions Cargo.lock

Some generated files are not rendered by default. Learn more about how customized files appear on GitHub.

3 changes: 2 additions & 1 deletion Cargo.toml
Original file line number Diff line number Diff line change
Expand Up @@ -17,13 +17,14 @@ bytes = "1"
tracing = "0.1"
tracing-subscriber = { version = "0.3", features = ["local-time", "json"] }
tokio = { version = "1", features = ["rt", "rt-multi-thread", "time", "macros", "sync", "signal", "io-util"] }
derive_more = { version = "0.99", features = ["display", "error"] }
derive_more = { version = "0.99", features = ["display", "error", "from"] }
# MPMS that only one consumer sees each message & async
async-channel = { version = "1" }
once_cell = "1.18"
prometheus-client = "0.21.1"
argh = "0.1"
axum = "0.6.18"
async-stream = "0.3.5"
utoipa = "3.3.0"
serde_json = "1.0.96"
serde = "1.0.163"
31 changes: 31 additions & 0 deletions examples/server_side_event/client.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,31 @@
# Copyright 2023 MOSEC Authors
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.

import httpx
from httpx_sse import connect_sse

with httpx.Client() as client:
with connect_sse(
client, "POST", "http://127.0.0.1:8000/sse_inference", json={"text": "mosec"}
) as event_source:
for sse in event_source.iter_sse():
print(f"Event({sse.event}): {sse.data}")

# error handling
with httpx.Client() as client:
with connect_sse(
client, "POST", "http://127.0.0.1:8000/sse_inference", json={"error": "mosec"}
) as event_source:
for sse in event_source.iter_sse():
print(f"Event({sse.event}): {sse.data}")
45 changes: 45 additions & 0 deletions examples/server_side_event/server.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,45 @@
# Copyright 2023 MOSEC Authors
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.

from mosec import Server, ValidationError, Worker, get_logger

logger = get_logger()


class Preprocess(Worker):
def forward(self, data):
text = data.get("text")
if text is None:
raise ValidationError("text is required")
return text


class Inference(Worker):
def forward(self, data):
epoch = 5
for i in range(epoch):
for j in range(len(data)):
self.send_stream_event(
f"inference: ({i + 1}/{epoch}) {data[j]}", index=j
)

# this return value will be ignored
return data


if __name__ == "__main__":
server = Server()
server.append_worker(Preprocess)
server.append_worker(Inference, max_batch_size=2)
server.run()
27 changes: 27 additions & 0 deletions mosec/coordinator.py
Original file line number Diff line number Diff line change
Expand Up @@ -15,9 +15,11 @@
"""The Coordinator is used to control the data flow between `Worker` and `Server`."""

import os
import queue
import signal
import socket
import struct
import threading
import time
import traceback
from contextlib import contextmanager
Expand Down Expand Up @@ -106,8 +108,13 @@ def __init__(
self.worker.worker_id = worker_id
self.worker.max_batch_size = max_batch_size
self.worker.stage = stage
self.stream: queue.SimpleQueue = queue.SimpleQueue()
self.semaphore: threading.Semaphore = threading.Semaphore()
self.worker._stream_queue = self.stream
self.worker._stream_semaphore = self.semaphore
self.timeout = timeout
self.name = f"<{stage_id}|{worker.__name__}|{worker_id}>"
self.current_ids: Sequence[bytes] = []

self.protocol = Protocol(
name=self.name,
Expand All @@ -132,6 +139,7 @@ def __init__(

self.warmup()
self.init_protocol()
threading.Thread(target=self.streaming, daemon=True).start()
self.run()

def init_protocol(self):
Expand All @@ -157,6 +165,19 @@ def init_protocol(self):
self.protocol.addr,
)

def streaming(self):
"""Send stream data from the worker to the server through the socket."""
while not self.shutdown.is_set():
try:
text, index = self.stream.get(timeout=self.timeout)
# encode the text with UTF-8
payloads = (text.encode(),)
ids = (self.current_ids[index],)
self.protocol.send(HTTPStautsCode.STREAM_EVENT, ids, payloads)
self.semaphore.release()
except queue.Empty:
continue

def warmup(self):
"""Warmup to allocate resources (useful for GPU workload)[Optional]."""
need_warmup = self.worker.example or self.worker.multi_examples
Expand Down Expand Up @@ -263,6 +284,8 @@ def coordinate(self):
while not self.shutdown.is_set():
try:
_, ids, payloads = protocol_recv()
Copy link
Member

@lkevinzc lkevinzc Jun 14, 2023

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Since both normal server and sse server will fetch request from this protocol_recv, chances are that sse server will get data from /inference request when clients query the wrong endpoint. In this case, the server will process the data but the client will get nothing.

Similarly, if the client query a normal server with /sse_inference, weird thing will happen.

We'd better to handle this by sending error for now; for future we should not have both endpoints exposed by default, but use multi-route kind of feature to support the specified endpoint only when the server implements it.

Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

  1. impl sse but called /inference => cannot find the stream event channel (info log), but return nothing to the client

  2. impl inference but called /sse_inference => processed but return nothing

Shall we return an error if it cannot find the stream event channel? This might not be 100% accurate.

Not sure how to deal with the 2. under the current design.

Copy link
Member

@lkevinzc lkevinzc Jun 15, 2023

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Is it possible to do a check at server side, knowing whether it is a normal sever or sse server, then send this info to rust side. In doing so, the rust side will know whether it should reject the requests from non-implemented route. Another benefit is that these rejected requests will never go to python side for processing.

A simple possible implementation is to let users set a class attribute to the sse worker? Only when it is set the send_stream_event can be called normally.

Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This might need to implement an SSE worker class. But passing the type of worker might require a more complex config file than the current command line arguments.

# expose here to be used by stream event
self.current_ids = ids
except socket.timeout:
continue
except (struct.error, OSError) as err:
Expand Down Expand Up @@ -299,10 +322,14 @@ def coordinate(self):
payloads = ["inference internal error".encode()] * length

try:
# pylint: disable=consider-using-with
self.semaphore.acquire(timeout=self.timeout)
protocol_send(status, ids, payloads)
except OSError as err:
logger.error("%s failed to send to socket: %s", self.name, err)
break
finally:
self.semaphore.release()

self.protocol.close()
time.sleep(CONN_CHECK_INTERVAL)
4 changes: 2 additions & 2 deletions mosec/errors.py
Original file line number Diff line number Diff line change
Expand Up @@ -29,8 +29,8 @@
class MosecError(Exception):
"""Mosec basic exception."""

code = HTTPStautsCode.INTERNAL_ERROR
msg = "mosec error"
code: HTTPStautsCode = HTTPStautsCode.INTERNAL_ERROR
msg: str = "mosec error"


class ClientError(MosecError):
Expand Down
6 changes: 5 additions & 1 deletion mosec/protocol.py
Original file line number Diff line number Diff line change
Expand Up @@ -38,6 +38,9 @@ class HTTPStautsCode(IntFlag):
INTERNAL_ERROR = 8 # 500
TIMEOUT_ERROR = 16 # 408

# special one, indicate that it's a SSE
STREAM_EVENT = 32768


class Protocol:
"""IPC protocol.
Expand Down Expand Up @@ -130,9 +133,10 @@ def send(self, flag: int, ids: Sequence[bytes], payloads: Sequence[bytes]):
self.socket.sendall(data.getbuffer())
if logger.isEnabledFor(logging.DEBUG):
logger.debug(
"%s sent %d tasks with ids: %s",
"%s sent %d(%d) tasks with ids: %s",
self.name,
len(ids),
flag,
struct.unpack("!" + "I" * len(ids), b"".join(ids)),
)

Expand Down
26 changes: 24 additions & 2 deletions mosec/worker.py
Original file line number Diff line number Diff line change
Expand Up @@ -22,16 +22,22 @@
4. data processing
"""

from __future__ import annotations

import abc
import json
import pickle
from typing import Any, Dict, Sequence, Tuple
from typing import TYPE_CHECKING, Any, Dict, Optional, Sequence, Tuple

from mosec.errors import DecodingError, EncodingError
from mosec.utils import ParseTarget

MOSEC_REF_TEMPLATE = "#/components/schemas/{name}"

if TYPE_CHECKING:
from queue import SimpleQueue
from threading import Semaphore


class Worker(abc.ABC):
"""MOSEC worker interface.
Expand Down Expand Up @@ -62,6 +68,8 @@ class Worker(abc.ABC):
_worker_id: int = 0
_stage: str = ""
_max_batch_size: int = 1
_stream_queue: Optional[SimpleQueue] = None
_stream_semaphore: Optional[Semaphore] = None

def __init__(self):
"""Initialize the worker.
Expand All @@ -85,6 +93,20 @@ def deserialize_ipc(self, data: bytes) -> Any:
"""
return pickle.loads(data)

def send_stream_event(self, text: str, index: int = 0):
"""Send a stream event to the client.

Args:
text: the text to be sent, needs to be UTF-8 compatible
index: the index of the stream event. For single request, this will always
be 0. For dynamic batch request, this should be the index of the
request in the batch.
"""
if self._stream_queue is None or self._stream_semaphore is None:
raise RuntimeError("the worker stream or semaphore is not initialized")
self._stream_semaphore.acquire()
self._stream_queue.put((text, index))

@property
def stage(self) -> str:
"""Return the stage name."""
Expand Down Expand Up @@ -185,7 +207,7 @@ def forward(self, data: Any) -> Any:
- for a single-stage worker, data will go through
``<deserialize> -> <forward> -> <serialize>``

- for a multi-stage worker that is neithor `ingress` not `egress`, data
- for a multi-stage worker that is neither `ingress` not `egress`, data
will go through ``<deserialize_ipc> -> <forward> -> <serialize_ipc>``
"""
raise NotImplementedError
Expand Down
1 change: 1 addition & 0 deletions pyproject.toml
Original file line number Diff line number Diff line change
Expand Up @@ -78,6 +78,7 @@ output-format = "colorized"
reports = "no"
score = "yes"
max-args = 7
max-attributes = 10

[tool.pylint.messages_control]
disable = []
Expand Down
1 change: 1 addition & 0 deletions requirements/dev.txt
Original file line number Diff line number Diff line change
Expand Up @@ -12,3 +12,4 @@ pre-commit>=2.15.0
msgpack>=1.0.5
numpy>=1.24
httpx==0.24.1
httpx-sse==0.3.0
4 changes: 4 additions & 0 deletions rustfmt.toml
Original file line number Diff line number Diff line change
@@ -1,2 +1,6 @@
unstable_features = true
reorder_imports = true
format_strings = true
imports_granularity = "Module"
group_imports = "StdExternalCrate"
reorder_impl_items = true
Loading