| | |
| | |
| | |
| | |
| | |
| |
|
| | """ |
| | FastAPI application for the REPL Environment. |
| | |
| | This module creates an HTTP server that exposes the REPLEnvironment |
| | over HTTP and WebSocket endpoints, compatible with EnvClient. |
| | |
| | The server includes llm_query and llm_query_batched support via HuggingFace Inference API, |
| | enabling the Recursive Language Model (RLM) paradigm. |
| | |
| | LLM Token Configuration: |
| | 1. Client can pass `hf_token` in reset() - RECOMMENDED |
| | 2. Server fallback: HF_TOKEN environment variable |
| | |
| | LLM functions are created dynamically in REPLEnvironment.reset() based on the |
| | available token (client or server). |
| | |
| | Usage: |
| | # Development (with auto-reload): |
| | uvicorn server.app:app --reload --host 0.0.0.0 --port 8000 |
| | |
| | # Production: |
| | uvicorn server.app:app --host 0.0.0.0 --port 8000 --workers 4 |
| | |
| | # Or run directly: |
| | uv run --project . server |
| | |
| | Environment Variables: |
| | HF_TOKEN: Fallback HuggingFace API token (client token takes priority) |
| | LLM_MODEL: Model to use for llm_query/llm_query_batched (default: Qwen/Qwen3-Coder-480B-A35B-Instruct) |
| | """ |
| |
|
| | import os |
| |
|
| | |
| | try: |
| | |
| | from openenv.core.env_server.http_server import create_app |
| | from ..models import REPLAction, REPLObservation |
| | from .repl_environment import REPLEnvironment |
| | except ImportError: |
| | |
| | from openenv.core.env_server.http_server import create_app |
| | from models import REPLAction, REPLObservation |
| | from server.repl_environment import REPLEnvironment |
| |
|
| |
|
| | |
| | LLM_MODEL = os.environ.get("LLM_MODEL", "Qwen/Qwen3-Coder-480B-A35B-Instruct") |
| | HF_TOKEN = os.environ.get("HF_TOKEN", None) |
| | |
| |
|
| | |
| | if HF_TOKEN: |
| | print(f"[REPL Server] LLM support ENABLED (server token configured)") |
| | print(f"[REPL Server] Default model: {LLM_MODEL}") |
| | else: |
| | print("[REPL Server] No server HF_TOKEN configured") |
| | print( |
| | "[REPL Server] LLM functions will be enabled if client passes hf_token in reset()" |
| | ) |
| |
|
| | |
| | env_factory = REPLEnvironment |
| |
|
| | |
| | app = create_app(env_factory, REPLAction, REPLObservation, env_name="repl_env") |
| |
|
| |
|
| | def main(): |
| | """ |
| | Entry point for direct execution via uv run or python -m. |
| | |
| | This function enables running the server without Docker: |
| | uv run --project . server |
| | python -m envs.repl_env.server.app |
| | openenv serve repl_env |
| | """ |
| | import uvicorn |
| |
|
| | uvicorn.run(app, host="0.0.0.0", port=8000) |
| |
|
| |
|
| | if __name__ == "__main__": |
| | main() |
| |
|