The Hanging Stream
Difficulty: HARDID: ai-stream-timeout
The Scenario
Your reverse proxy has a 10-second timeout. If no data is sent, it kills the connection with 504 Gateway Timeout.
GPT-4 takes 30 seconds to "think" before generating the first token for complex prompts.
Result: Your users see timeout errors, even though the LLM is working.
The Problem
Your code uses synchronous blocking:
def get_completion(prompt):
response = openai.chat.completions.create(...)
return response.choices[0].message.content # Blocks for 30s
The HTTP request hangs. The gateway times out. Users leave.
The Goal
Refactor to use a Generator that streams tokens as they arrive:
def stream_completion(prompt):
for chunk in openai.chat.completions.create(stream=True):
yield chunk # Send data immediately
Requirements:
- Use
yieldto return chunks incrementally - The test will verify the first chunk is yielded without a full response delay
- Must handle streaming iterators
Note: This is how ChatGPT delivers responses word-by-word in real-time.
solution.py
Loading...
⚠️ Do not include PII or secrets in your code.
SYSTEM_LOGS
5/5
// Waiting for execution trigger...
PREVIEW MODE — SOLVE PREVIOUS MISSIONS TO UNLOCK