File size: 668 Bytes
be14780
dfd3125
72b019e
be14780
24a23b6
be14780
 
f189c3b
b29e1e3
be14780
b29e1e3
 
be14780
 
 
 
 
72b019e
 
be14780
fc124f8
a263ec3
be14780
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
# Use the official vLLM Docker image as the base image
FROM vllm/vllm-openai:latest

# Set working directory
WORKDIR /app

# Copy application files into the container
COPY . /app

# Install additional Python packages if needed
RUN pip install --no-cache-dir -r requirements.txt

# Create a writable cache directory and set the HF_HOME environment variable
RUN mkdir -p /app/.cache && chmod -R 777 /app/.cache
ENV HF_HOME=/app/.cache

# Expose the port used by Gradio (default: 7860)
EXPOSE 7860

# Clear the base image's entrypoint (if needed)
ENTRYPOINT []

# Make the startup script executable
RUN chmod +x start.sh

# Run the startup script
CMD ["bash", "start.sh"]