Submitted by [deleted] t3_10z3qdt in MachineLearning
machineko t1_j8yo6fd wrote
Reply to comment by askingforhelp1111 in [D] Speed up HuggingFace Inference Pipeline by [deleted]
Depends on what models you are using but for most transformers, running on GPUs may be much more efficient than CPUs when you consider $ / M inferences (or inf/$).
Are there specific EC2 instances you have to use or can you deploy on any EC2 instance?
Viewing a single comment thread. View all comments