Workload Placement
Convox provides powerful tools to control where your applications and build processes run within your Kubernetes cluster. By leveraging node group configurations and service placement rules, you can optimize resource usage, improve cost efficiency, and ensure the right workloads run on the right infrastructure.
Overview
Workload placement in Convox is achieved through these key features:
- Custom Node Groups: Define specialized node pools with specific instance types, scaling parameters, labels, and AWS tags.
- Node Selectors: Direct specific services or build processes to appropriate node groups.
- Dedicated Node Pools: Isolate workloads by creating exclusive node groups for particular services.
These capabilities allow for sophisticated infrastructure optimization strategies, such as:
- Separating production services from build processes
- Using cost-effective spot instances for non-critical workloads
- Optimizing instance types for specific workload profiles
- Creating high-performance node groups for specialized services
- Tracking resource usage and costs with AWS tags
Configuration Components
Rack-level Configuration
At the rack level, you can define custom node groups:
additional_node_groups_config
: Creates general-purpose node groupsadditional_build_groups_config
: Creates node groups specifically for build processes
These parameters allow you to specify:
- Instance types
- Disk sizes
- Capacity types (on-demand vs. spot)
- Scaling parameters
- Custom labels for workload targeting
- Unique IDs for node group preservation across updates
- AWS tags for cost allocation and resource organization
Note: These configurations are independent of each other. You can use either one or both depending on your needs. If you only configure additional node groups, builds will continue using the rack's primary build node (if build_node_enabled is set) or the primary rack nodes. If you only configure build node groups, your services will continue running on the standard rack nodes while builds will be isolated according to your build configuration.
Setting Rack Parameters with JSON Files
While you can set configuration directly using a JSON string, most users find it more manageable to use a JSON file, especially for complex configurations.
Using a JSON File for Node Groups
Create a JSON file (e.g., node-groups.json
) with your configuration:
[
{
"id": 101,
"type": "t3.medium",
"capacity_type": "ON_DEMAND",
"min_size": 1,
"max_size": 5,
"label": "critical-services",
"tags": "environment=production,team=frontend"
},
{
"id": 102,
"type": "c5.large",
"capacity_type": "SPOT",
"min_size": 0,
"max_size": 10,
"label": "batch-workers",
"disk": 100,
"tags": "environment=production,team=data,workload=batch"
}
]
Note the use of:
- The
id
field to uniquely identify each node group - The
tags
field to apply AWS resource tags for organization and cost tracking
Then apply the configuration using:
$ convox rack params set additional_node_groups_config=/path/to/node-groups.json -r rackName
Important Note on AWS Rate Limits: When adding or removing multiple node groups, it's recommended to modify no more than three node groups at a time to avoid hitting AWS API rate limits. If you receive a rate limit error during an update simply run the parameter set command again. The operation will resume from where it left off, creating the remaining node groups without duplicating the ones that were already successfully created.
Using a JSON File for Build Node Groups
Similarly, create a JSON file (e.g., build-groups.json
) for build node configuration:
[
{
"id": 201,
"type": "c5.xlarge",
"capacity_type": "SPOT",
"min_size": 0,
"max_size": 3,
"label": "app-build",
"disk": 100,
"tags": "environment=build,team=devops"
}
]
Apply it with:
$ convox rack params set additional_build_groups_config=/path/to/build-groups.json -r rackName
Using a Single JSON String (Alternative Approach)
If you prefer to set configuration directly in the command line without creating a file, you can use a JSON string:
$ convox rack params set 'additional_node_groups_config=[{"id":101,"type":"t3.medium","capacity_type":"ON_DEMAND","min_size":1,"max_size":5,"label":"critical-services","tags":"environment=production,team=frontend"}]' -r rackName
$ convox rack params set 'additional_build_groups_config=[{"id":201,"type":"c5.xlarge","capacity_type":"SPOT","min_size":0,"max_size":3,"label":"app-build","disk":100,"tags":"environment=build,team=devops"}]' -r rackName
This approach is useful for automation scripts or when making quick changes, though it becomes unwieldy for more complex configurations.
Node Group Configuration Options
Each node group configuration supports the following fields:
Field | Required | Description | Default |
---|---|---|---|
id |
No | Unique integer identifier for the node group | Auto-generated |
type |
Yes | The EC2 instance type to use for the node group | |
disk |
No | The disk size in GB for the nodes | Same as main node disk |
capacity_type |
No | Whether to use on-demand or spot instances | ON_DEMAND |
min_size |
No | Minimum number of nodes | 1 |
max_size |
No | Maximum number of nodes | 100 |
label |
No | Custom label value for the node group. Applied as convox.io/label: <label-value> |
None |
tags |
No | Custom AWS tags as comma-separated key-value pairs | None |
dedicated |
No | When true , only services with matching node group labels will be scheduled on these nodes |
false |
ami_id |
No | Custom AMI ID to use | EKS-optimized AMI |
About the id
field
The id
field provides important benefits:
- Preserves node group identity during configuration updates
- Prevents unnecessary recreation of node groups
- Allows for stable references when targeting specific node groups
- Reduces downtime during configuration changes
Without the id
field, Convox generates a random identifier that changes when the configuration is updated, potentially causing unnecessary node group recreation.
App-level Configuration
At the application level, you can control where specific workloads run:
BuildLabels
: Directs build pods to specific node groupsBuildCpu
andBuildMem
: Sets resource requests for build podsnodeSelectorLabels
inconvox.yml
: Directs service pods to specific node groups
Service-level Configuration
In your convox.yml
file, you can specify node selectors for each service:
services:
web:
nodeSelectorLabels:
convox.io/label: app-workers
worker:
nodeSelectorLabels:
convox.io/label: batch-workers
Implementation Examples
Optimizing for Cost and Performance
This example creates a cost-optimized infrastructure with dedicated node pools for different workload types:
-
Rack Configuration:
$ convox rack params set additional_node_groups_config=/path/to/node-groups.json -r production $ convox rack params set additional_build_groups_config=/path/to/build-groups.json -r production
With node groups config:
[ { "id": 101, "type": "t3.medium", "capacity_type": "ON_DEMAND", "min_size": 1, "max_size": 5, "label": "critical-services", "tags": "department=frontend,environment=production,cost-center=web" }, { "id": 102, "type": "c5.large", "capacity_type": "SPOT", "min_size": 0, "max_size": 10, "label": "batch-workers", "tags": "department=data,environment=production,cost-center=analytics" } ]
And build groups config:
[ { "type": "c5.xlarge", "capacity_type": "SPOT", "min_size": 0, "max_size": 5, "label": "app-build", } ]
-
Application Configuration:
$ convox apps params set BuildLabels=convox.io/label=app-build -a myapp $ convox apps params set BuildCpu=1024 BuildMem=4096 -a myapp
-
Service Configuration (in
convox.yml
):services: web: build: . port: 3000 nodeSelectorLabels: convox.io/label: critical-services worker: build: ./worker nodeSelectorLabels: convox.io/label: batch-workers
This configuration creates:
- On-demand nodes for critical services like web frontends
- Spot instance nodes for batch processing workloads
- Separate spot instance nodes optimized for build processes
- AWS tags for accurate cost attribution across teams and environments
Isolating High-Priority Workloads
To create dedicated node groups that exclusively run specific services:
-
Create a JSON file for your node group configuration:
[ { "id": 103, "type": "m5.large", "capacity_type": "ON_DEMAND", "min_size": 2, "max_size": 5, "label": "database-workers", "dedicated": true, "tags": "environment=production,team=datastore,criticality=high" } ]
-
Apply the configuration:
$ convox rack params set additional_node_groups_config=/path/to/dedicated-nodes.json -r production
-
Service Configuration (in
convox.yml
):services: db-processor: build: ./processor nodeSelectorLabels: convox.io/label: database-workers
With dedicated:true
, only services that explicitly select the node group will run on it, ensuring isolation for sensitive workloads.
Flexible Configuration Options
Convox allows you to implement different levels of customization based on your needs:
-
Build Isolation Only: Configure only
additional_build_groups_config
to isolate build processes while keeping services on standard nodes:$ convox rack params set additional_build_groups_config=/path/to/build-groups.json -r production $ convox apps params set BuildLabels=convox.io/label=app-build -a myapp
-
Service Placement Only: Configure only
additional_node_groups_config
to customize service placement while letting builds run on standard nodes:$ convox rack params set additional_node_groups_config=/path/to/node-groups.json -r production
In your
convox.yml
:services: web: nodeSelectorLabels: convox.io/label: critical-services
-
Complete Workload Management: Implement both configurations for full control over placement of both services and build processes.
Best Practices
-
Match Node Resources to Workload Requirements:
- Use compute-optimized instances (c-type) for CPU-intensive workloads
- Use memory-optimized instances (r-type) for memory-intensive workloads
- Use general-purpose instances (m-type or t-type) for balanced workloads
-
Cost Optimization:
- Use spot instances for interruptible workloads like batch processing
- Use on-demand instances for critical production services
- Set appropriate min/max scaling parameters to avoid over-provisioning
- Apply tags to track costs by team, environment, or application
-
Build Process Optimization:
- Configure build nodes with higher CPU and memory for faster builds
- Use spot instances for builds to reduce costs
- Set
min_size: 0
to allow build nodes to scale down when not in use
-
Service Isolation:
- Use the
dedicated
flag for node groups that need strict isolation - Separate services with conflicting resource profiles into different node groups
- Use the
-
Node Group Identity Management:
- Always assign a unique
id
to each node group - Use consistent, meaningful IDs (e.g., 100-199 for production, 200-299 for builds)
- Document your ID allocation to avoid conflicts
- Always assign a unique
-
Tagging Strategy:
- Develop a consistent tagging convention for all node groups
- Include tags for environment, team, cost center, and workload type
- Align tags with your organization's AWS tagging policy
Troubleshooting
Build Failures Due to Node Selection
If builds fail with scheduling errors, verify:
- The build node group exists and has the correct labels
- The
BuildLabels
parameter matches the node group's labels - There are nodes available that match the label criteria
Service Deployment Issues
If services won't deploy, check:
- Node selector labels in
convox.yml
match existing node groups - The referenced node groups have available capacity
- Resource requests in the service definition can be satisfied by the node group
Node Group Scaling
If nodes aren't scaling as expected:
- Verify min/max settings are appropriate
- Check that instance types are available in your region
- Monitor for AWS service quotas that might limit scaling
Node Group Preservation Issues
If node groups are being recreated unexpectedly:
- Ensure each node group has a unique
id
field - Verify that you're not changing immutable fields (like capacity type)
- Check for AWS API rate limits during updates
Conclusion
Effective workload placement is a powerful tool for optimizing your Convox infrastructure. By leveraging custom node groups with preserved identities, service placement rules, and AWS tagging, you can create an infrastructure that balances performance, cost, and isolation requirements for your specific application needs.
For more detailed information, refer to: