I have a function which perform some mathematical operations and need a 16gb GPU system, But this function will not be triggered always and rest of time my system will not be in use. I came to know about AWS Lambda. Can I run GPU based algorithm on Lambda?? So that whenever I need GPU, I will get the system on cloud. I need a little description about it.
相关问题
- Delete Messages from a Topic in Apache Kafka
- Jackson Deserialization not calling deserialize on
- How to maintain order of key-value in DataFrame sa
- StackExchange API - Deserialize Date in JSON Respo
- Difference between Types.INTEGER and Types.NULL in
Batch is a good solution for certain types of workload. Another option is GPUs on ECS, which could be used for running frequent tasks utilising GPU.
Currently lambda doesn't have GPU.
However, if you just need to do inference; the emulation via CPU works fine on AWS lambda; here is an article that goes into more details:
https://aws.amazon.com/blogs/machine-learning/how-to-deploy-deep-learning-models-with-aws-lambda-and-tensorflow/
You can't specify the runtime environment for AWS Lambda functions, so no, you can't require the presence of a GPU (in fact the physical machines AWS chooses to put into its Lambda pool will almost certainly not have one).
Your best bet would be to run the GPU-requiring function as a Batch job on a compute cluster configured to use p-type instances. The guide here might be helpful.