Running inference as a serverless cloud function
When deploying your model into production, you have to care about configuring and maintaining the runtime environment, scaling, monitoring and more – all tasks, which are more related to DevOps than to ML. In some contexts, you can achieve your goals in a much simpler way by establishing a «serverless» approach. We’ll take a look at the cloud services «AWS Lambda», «Google Cloud Functions», and «Azure Functions» and show how they enable running ML inference.
- Datum
- 04.03.2019
- Uhrzeit
- 19:00 - 19:45
- Konferenz / Veranstaltung
- Berlin Machine Learning Group
- Ort
- Betahaus, Berlin
Nach links wischen, um ggf. weitere Folien zu sehen
Scrollen oder die Pfeiltasten nutzen, um ggf. weitere Folien zu sehen
Scrollen, um ggf. weitere Folien zu sehen
Die Pfeiltasten nutzen, um ggf. weitere Folien zu sehen