Skip to content

Commit

Permalink
Add http server to JetStream (#115)
Browse files Browse the repository at this point in the history
* Add http server to JetStream

* Add generate api and cleanup

* Add unit tests

* format & deps

* type & lint

* Merge refactor

* fix refactor
  • Loading branch information
JoeZijunZhou authored Jul 23, 2024
1 parent bd6d013 commit 1830342
Show file tree
Hide file tree
Showing 12 changed files with 450 additions and 38 deletions.
92 changes: 57 additions & 35 deletions jetstream/core/server_lib.py
Original file line number Diff line number Diff line change
Expand Up @@ -93,40 +93,25 @@ def wait_for_termination(self) -> None:
self.stop()


def run(
port: int,
def create_driver(
config: Type[config_lib.ServerConfig],
devices: Any,
credentials: Any = grpc.insecure_server_credentials(),
threads: int | None = None,
jax_padding: bool = True,
metrics_server_config: config_lib.MetricsServerConfig | None = None,
enable_jax_profiler: bool = False,
jax_profiler_port: int = 9999,
metrics_collector: JetstreamMetricsCollector | None = None,
enable_model_warmup: bool = False,
) -> JetStreamServer:
"""Runs a server with a specified config.
):
"""Creates a driver with a specified config.
Args:
port: Port on which the server will be made available.
config: A ServerConfig to config engine, model, device slices, etc.
devices: Device objects, will be used to get engine with proper slicing.
credentials: Should use grpc credentials by default.
threads: Number of RPC handlers worker threads. This should be at least
equal to the decoding batch size to fully saturate the decoding queue.
jax_padding: The flag to enable JAX padding during tokenization.
metrics_server_config: The config to enable Promethus metric server.
enable_jax_profiler: The flag to enable JAX profiler server.
jax_profiler_port: The port JAX profiler server (default to 9999).
metrics_collector: The JetStream Promethus metric collector.
enable_model_warmup: The flag to enable model server warmup with AOT.
Returns:
JetStreamServer that wraps the grpc server and orchestrator driver.
An orchestrator driver.
"""

server_start_time = time.time()

logging.info("Kicking off gRPC server.")
engines = config_lib.get_engines(config, devices=devices)
prefill_params = [pe.load_params() for pe in engines.prefill_engines]
generate_params = [ge.load_params() for ge in engines.generate_engines]
Expand All @@ -136,19 +121,6 @@ def run(
len(config.prefill_slices) + len(config.generate_slices) == 0
)

# Setup Prometheus server
metrics_collector: JetstreamMetricsCollector = None
if metrics_server_config and metrics_server_config.port:
logging.info(
"Starting Prometheus server on port %d", metrics_server_config.port
)
start_http_server(metrics_server_config.port)
metrics_collector = JetstreamMetricsCollector()
else:
logging.info(
"Not starting Prometheus server: --prometheus_port flag not set"
)

prefill_engines = engines.prefill_engines + engines.interleaved_engines
generate_engines = engines.generate_engines + engines.interleaved_engines
prefill_params = prefill_params + shared_params
Expand Down Expand Up @@ -182,7 +154,7 @@ def run(
traceback.print_exc()
os.kill(os.getpid(), signal.SIGKILL)

driver = orchestrator.Driver(
return orchestrator.Driver(
prefill_engines=prefill_engines,
generate_engines=generate_engines,
prefill_params=prefill_params,
Expand All @@ -192,6 +164,56 @@ def run(
metrics_collector=metrics_collector,
is_ray_backend=config.is_ray_backend,
)


def run(
port: int,
config: Type[config_lib.ServerConfig],
devices: Any,
credentials: Any = grpc.insecure_server_credentials(),
threads: int | None = None,
jax_padding: bool = True,
metrics_server_config: config_lib.MetricsServerConfig | None = None,
enable_jax_profiler: bool = False,
jax_profiler_port: int = 9999,
enable_model_warmup: bool = False,
) -> JetStreamServer:
"""Runs a server with a specified config.
Args:
port: Port on which the server will be made available.
config: A ServerConfig to config engine, model, device slices, etc.
devices: Device objects, will be used to get engine with proper slicing.
credentials: Should use grpc credentials by default.
threads: Number of RPC handlers worker threads. This should be at least
equal to the decoding batch size to fully saturate the decoding queue.
jax_padding: The flag to enable JAX padding during tokenization.
metrics_server_config: The config to enable Promethus metric server.
enable_jax_profiler: The flag to enable JAX profiler server.
jax_profiler_port: The port JAX profiler server (default to 9999).
enable_model_warmup: The flag to enable model server warmup with AOT.
Returns:
JetStreamServer that wraps the grpc server and orchestrator driver.
"""
server_start_time = time.time()
logging.info("Kicking off gRPC server.")
# Setup Prometheus server
metrics_collector: JetstreamMetricsCollector = None
if metrics_server_config and metrics_server_config.port:
logging.info(
"Starting Prometheus server on port %d", metrics_server_config.port
)
start_http_server(metrics_server_config.port)
metrics_collector = JetstreamMetricsCollector()
else:
logging.info(
"Not starting Prometheus server: --prometheus_port flag not set"
)

driver = create_driver(
config, devices, jax_padding, metrics_collector, enable_model_warmup
)
# We default threads to the total number of concurrent allowed decodes,
# to make sure we can fully saturate the model. Set default minimum to 64.
threads = threads or max(driver.get_total_concurrent_requests(), 64)
Expand Down
13 changes: 13 additions & 0 deletions jetstream/entrypoints/__init__.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,13 @@
# Copyright 2024 Google LLC
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
32 changes: 32 additions & 0 deletions jetstream/entrypoints/config.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,32 @@
# Copyright 2024 Google LLC
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.

"""Config for JetStream Server (including engine init)."""

from typing import Type

from jetstream.core import config_lib


def get_server_config(
config_str: str,
) -> config_lib.ServerConfig | Type[config_lib.ServerConfig]:
match config_str:
case "InterleavedCPUTestServer":
server_config = config_lib.InterleavedCPUTestServer
case "CPUTestServer":
server_config = config_lib.CPUTestServer
case _:
raise NotImplementedError
return server_config
13 changes: 13 additions & 0 deletions jetstream/entrypoints/http/__init__.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,13 @@
# Copyright 2024 Google LLC
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
132 changes: 132 additions & 0 deletions jetstream/entrypoints/http/api_server.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,132 @@
# Copyright 2024 Google LLC
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.

"""JetStream Http API server."""

import json
import logging
from typing import Sequence
from absl import app as abslapp
from absl import flags
from fastapi import APIRouter, Response
import fastapi
from fastapi.responses import StreamingResponse
from prometheus_client import start_http_server
import uvicorn
from google.protobuf.json_format import Parse

from jetstream.core import config_lib, orchestrator, server_lib
from jetstream.core.metrics.prometheus import JetstreamMetricsCollector
from jetstream.core.proto import jetstream_pb2
from jetstream.entrypoints.config import get_server_config
from jetstream.entrypoints.http.protocol import DecodeRequest
from jetstream.entrypoints.http.utils import proto_to_json_generator

flags.DEFINE_string("host", "0.0.0.0", "server host address")
flags.DEFINE_integer("port", 8080, "http server port")
flags.DEFINE_string(
"config",
"InterleavedCPUTestServer",
"available servers",
)
flags.DEFINE_integer(
"prometheus_port",
9988,
"prometheus_port",
)

llm_orchestrator: orchestrator.LLMOrchestrator

# Define Fast API endpoints (use llm_orchestrator to handle).
router = APIRouter()


@router.get("/")
def root():
"""Root path for Jetstream HTTP Server."""
return Response(
content=json.dumps({"message": "JetStream HTTP Server"}, indent=4),
media_type="application/json",
)


@router.post("/v1/generate")
async def generate(request: DecodeRequest):
proto_request = Parse(request.json(), jetstream_pb2.DecodeRequest())
generator = llm_orchestrator.Decode(proto_request)
return StreamingResponse(
content=proto_to_json_generator(generator), media_type="text/event-stream"
)


@router.get("/v1/health")
async def health() -> Response:
"""Health check."""
response = await llm_orchestrator.HealthCheck(
jetstream_pb2.HealthCheckRequest()
)
return Response(
content=json.dumps({"is_live": str(response.is_live)}, indent=4),
media_type="application/json",
status_code=200,
)


def server(argv: Sequence[str]):
# Init Fast API.
app = fastapi.FastAPI()
app.include_router(router)

# Init LLMOrchestrator which would be the main handler in the api endpoints.
devices = server_lib.get_devices()
print(f"devices: {devices}")
server_config = get_server_config(flags.FLAGS.config)
print(f"server_config: {server_config}")
del argv

metrics_server_config: config_lib.MetricsServerConfig | None = None
# Setup Prometheus server
metrics_collector: JetstreamMetricsCollector = None
if flags.FLAGS.prometheus_port != 0:
metrics_server_config = config_lib.MetricsServerConfig(
port=flags.FLAGS.prometheus_port
)
logging.info(
"Starting Prometheus server on port %d", metrics_server_config.port
)
start_http_server(metrics_server_config.port)
metrics_collector = JetstreamMetricsCollector()
else:
logging.info(
"Not starting Prometheus server: --prometheus_port flag not set"
)

global llm_orchestrator
llm_orchestrator = orchestrator.LLMOrchestrator(
driver=server_lib.create_driver(
config=server_config,
devices=devices,
metrics_collector=metrics_collector,
)
)

# Start uvicorn http server.
uvicorn.run(
app, host=flags.FLAGS.host, port=flags.FLAGS.port, log_level="info"
)


if __name__ == "__main__":
# Run Abseil app w flags parser.
abslapp.run(server)
36 changes: 36 additions & 0 deletions jetstream/entrypoints/http/protocol.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,36 @@
# Copyright 2024 Google LLC
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.

"""Http API server protocol."""

from pydantic import BaseModel # type: ignore


class TextContent(BaseModel):
text: str


class TokenContent(BaseModel):
token_ids: list[int]


class DecodeRequest(BaseModel):
max_tokens: int
text_content: TextContent | None = None
token_content: TokenContent | None = None

# Config to enforce the oneof behavior at runtime.
class Config:
extra = "forbid" # Prevent extra fields.
anystr_strip_whitespace = True
27 changes: 27 additions & 0 deletions jetstream/entrypoints/http/utils.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,27 @@
# Copyright 2024 Google LLC
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.

"""Http API server utilities."""

from google.protobuf.json_format import MessageToJson


async def proto_to_json_generator(proto_generator):
"""Wraps a generator yielding Protocol Buffer messages into a generator
yielding JSON messages.
"""
async for proto_message in proto_generator:
json_string = MessageToJson(proto_message)
yield json_string
Loading

0 comments on commit 1830342

Please sign in to comment.