GPU based algorithm on AWS Lambda

2020-08-10 07:31发布

I have a function which perform some mathematical operations and need a 16gb GPU system, But this function will not be triggered always and rest of time my system will not be in use. I came to know about AWS Lambda. Can I run GPU based algorithm on Lambda?? So that whenever I need GPU, I will get the system on cloud. I need a little description about it.

3条回答
祖国的老花朵
2楼-- · 2020-08-10 07:49

Batch is a good solution for certain types of workload. Another option is GPUs on ECS, which could be used for running frequent tasks utilising GPU.

查看更多
戒情不戒烟
3楼-- · 2020-08-10 07:54

Currently lambda doesn't have GPU.

However, if you just need to do inference; the emulation via CPU works fine on AWS lambda; here is an article that goes into more details:

https://aws.amazon.com/blogs/machine-learning/how-to-deploy-deep-learning-models-with-aws-lambda-and-tensorflow/

查看更多
爷、活的狠高调
4楼-- · 2020-08-10 08:10

You can't specify the runtime environment for AWS Lambda functions, so no, you can't require the presence of a GPU (in fact the physical machines AWS chooses to put into its Lambda pool will almost certainly not have one).

Your best bet would be to run the GPU-requiring function as a Batch job on a compute cluster configured to use p-type instances. The guide here might be helpful.

查看更多
登录 后发表回答