|
8 | 8 |
|
9 | 9 | # Serverless containers on AWS
|
10 | 10 |
|
11 |
| -Deploy, manage, and scale containers without managing infrastructure. |
| 11 | +Avoid the complexity of Kubernetes, service meshes, load balancers, queues, and autoscalers. Deploy, manage, and scale containers on a Kubernetes-based serverless platform that is up to 75% less expensive than Lambda. |
12 | 12 |
|
13 | 13 | <br>
|
14 | 14 |
|
15 |
| -## Scale realtime, batch, and async workloads |
| 15 | +## Build production-grade microservices faster |
16 | 16 |
|
17 |
| -**Realtime** - respond to requests in real-time and autoscale based on in-flight request volumes. |
| 17 | +**Autoscaling** - customize autoscaling for realtime, async, and batch workloads. |
18 | 18 |
|
19 |
| -**Batch** - run distributed and fault-tolerant batch processing jobs on-demand. |
| 19 | +**Networking** - get HTTP endpoints that can handle large-scale production traffic. |
20 | 20 |
|
21 |
| -**Async** - process requests asynchronously and autoscale based on request queue length. |
22 |
| - |
23 |
| -<br> |
24 |
| - |
25 |
| -```bash |
26 |
| -$ cortex deploy |
27 |
| - |
28 |
| -creating realtime text-generator |
29 |
| -creating batch image-classifier |
30 |
| -creating async video-analyzer |
31 |
| -``` |
| 21 | +**Observability** - send logs and metrics from every workload to any monitoring tool. |
32 | 22 |
|
33 | 23 | <br>
|
34 | 24 |
|
35 | 25 | ## Allocate CPU, GPU, and memory without limits
|
36 | 26 |
|
37 | 27 | **No resource limits** - allocate as much CPU, GPU, and memory as each workload requires.
|
38 | 28 |
|
39 |
| -**No cold starts** - keep a minimum number of replicas running to ensure that requests are handled in real-time. |
| 29 | +**No cold starts** - keep a minimum number of replicas running to minimize latency. |
40 | 30 |
|
41 | 31 | **No timeouts** - run workloads for as long as you want.
|
42 | 32 |
|
43 | 33 | <br>
|
44 | 34 |
|
45 |
| -```bash |
46 |
| -$ cortex get |
47 |
| - |
48 |
| -WORKLOAD TYPE REPLICAS |
49 |
| -text-generator realtime 32 |
50 |
| -image-classifier batch 64 |
51 |
| -video-analyzer async 16 |
52 |
| -``` |
53 |
| - |
54 |
| -<br> |
55 |
| - |
56 |
| -## Control your AWS spend |
57 |
| - |
58 |
| -**Scale to zero** - optimize the autoscaling behavior of each workload to minimize idle resources. |
| 35 | +## Save up to 75% on your AWS bill |
59 | 36 |
|
60 |
| -**Multi-instance** - run different workloads on different EC2 instances to ensure efficient resource utilization. |
61 |
| - |
62 |
| -**Spot instances** - run workloads on spot instances and fall back to on-demand instances to ensure reliability. |
63 |
| - |
64 |
| -<br> |
| 37 | +**Spot instances** - run workloads on spot and/or reserved instances without affecting reliability. |
65 | 38 |
|
66 |
| -```bash |
67 |
| -$ cortex cluster up |
| 39 | +**Custom instances** - use instances with the optimal price-performance ratio for your workloads. |
68 | 40 |
|
69 |
| -INSTANCE PRICE SPOT SCALE |
70 |
| -c5.xlarge $0.17 yes 0-100 |
71 |
| -g4dn.xlarge $0.53 yes 0-100 |
72 |
| -``` |
| 41 | +**Scale to zero** - scale workloads to zero to minimize idle resources. |
0 commit comments