Speedup bert model inference

What are good and simple ways to make it fast enough on a server?