Text this: PipeMax: Enhancing Offline LLM Inference on Commodity GPU Servers