« Blazar » : différence entre les versions

De TeriaHowto
Sauter à la navigation Sauter à la recherche
Aucun résumé des modifications
 
(4 versions intermédiaires par le même utilisateur non affichées)
Ligne 3 : Ligne 3 :
== Concept ==
== Concept ==


Imagine this situation : you provides two services, one with high priority, the other one with low priority. During the week, each service runs VMs on multiple node with auto-scaling capability (i.e. another VMs are automatically provisionned if the workload increases). You know that during the week-end, your higher priority service will have more demand and you don't want your lower priority service starving it from resources.<br>
Imagine this situation : you provide two services, one with high priority, the other one with low priority. During the week, each service runs VMs on multiple node with auto-scaling capability (i.e. another VMs are automatically provisionned if the workload increases). You know that during the week-end, your higher priority service will have more demand and you don't want your lower priority service starving it from resources.<br>
Blazar brings the possibility to reserve dedicated compute nodes for your high prority service for a time period, say, the week-end.<br>
Blazar brings the possibility to reserve dedicated compute nodes for your high prority service for a time period, say, the week-end.<br>
Best of all, unreserved compute nodes managed by Blazar can run preemptible VMs, avoiding unused computing power.
Best of all, unreserved compute nodes managed by Blazar can run preemptible VMs, avoiding unused computing power.
Ligne 15 : Ligne 15 :
<syntaxhighlight lang="yaml">
<syntaxhighlight lang="yaml">
enable_blazar: "yes"
enable_blazar: "yes"
kolla_enable_tls_internal: "no"
kolla_enable_tls_internal: "yes"
kolla_enable_tls_backend: "no"
kolla_enable_tls_backend: "yes"
</syntaxhighlight>
</syntaxhighlight>
<br>
 
Note that TLS should be disabled (especially for internal traffic). This is a [https://bugs.launchpad.net/blazar/+bug/2045281 current limitation identified].
If TLS is enabled, on the Ansible deploy host, you have to put a *cafile* option in ''/etc/kolla/config/blazar.conf'' :
<br>
<syntaxhighlight lang="yaml">
Deploy :
[DEFAULT]
<syntaxhighlight lang="bash">
cafile = /etc/pki/ca-trust/source/anchors/kolla-customca-root.crt
kolla-ansible -i <inventory_file> deploy
</syntaxhighlight>
</syntaxhighlight>


=== Activate ''preemptible instances'' in ''nova-scheduler'' ===
=== Activate ''preemptible instances'' and ''BlazarFilter'' in ''nova-scheduler'' ===


On the Ansible deploy host, the file should be ''/etc/kolla/config/nova/nova-scheduler.conf''
On the Ansible deploy host, the file should be ''/etc/kolla/config/nova/nova-scheduler.conf''


<syntaxhighlight lang="yaml">
<syntaxhighlight lang="yaml">
[filter_scheduler]
available_filters = nova.scheduler.filters.all_filters
available_filters = blazarnova.scheduler.filters.blazar_filter.BlazarFilter
enabled_filters = <list_of_other_filters>,BlazarFilter
[blazar:physical:host]
[blazar:physical:host]
allow_preemptibles = true
allow_preemptibles = true
</syntaxhighlight>
=== Deploy ===
<syntaxhighlight lang="bash">
kolla-ansible -i <inventory_file> deploy
</syntaxhighlight>
</syntaxhighlight>



Dernière version du 18 septembre 2024 à 16:47

Let's experiment Blazar, Openstack dedicated service for resource reservation.

Concept

Imagine this situation : you provide two services, one with high priority, the other one with low priority. During the week, each service runs VMs on multiple node with auto-scaling capability (i.e. another VMs are automatically provisionned if the workload increases). You know that during the week-end, your higher priority service will have more demand and you don't want your lower priority service starving it from resources.
Blazar brings the possibility to reserve dedicated compute nodes for your high prority service for a time period, say, the week-end.
Best of all, unreserved compute nodes managed by Blazar can run preemptible VMs, avoiding unused computing power.

Installation / Configuration

Kolla-ansible is used to deploy our private cloud.

Enable Blazar in /etc/kolla/globals.yml

enable_blazar: "yes"
kolla_enable_tls_internal: "yes"
kolla_enable_tls_backend: "yes"

If TLS is enabled, on the Ansible deploy host, you have to put a *cafile* option in /etc/kolla/config/blazar.conf :

[DEFAULT]
cafile = /etc/pki/ca-trust/source/anchors/kolla-customca-root.crt

Activate preemptible instances and BlazarFilter in nova-scheduler

On the Ansible deploy host, the file should be /etc/kolla/config/nova/nova-scheduler.conf

[filter_scheduler]
available_filters = nova.scheduler.filters.all_filters
available_filters = blazarnova.scheduler.filters.blazar_filter.BlazarFilter
enabled_filters = <list_of_other_filters>,BlazarFilter

[blazar:physical:host]
allow_preemptibles = true

Deploy

kolla-ansible -i <inventory_file> deploy

Create a dedicated flavor for preemptible instances

For the sake of simplicity, only two flavors are going to be used :

  • a "normal" (no preemption)
  • a preemptible
# normal flavor
openstack flavor create --id 1 --ram 512 --disk 1 --vcpus 1 g1
# preemptible flavor
openstack flavor create --id 2 --ram 512 --disk 1 --vcpus 1 --property blazar:preemptible=true g1.preempt

Action

Book hosts for blazar

The hosts you want to reserve for blazar must not belong to aggregates.

openstack reservation host create compute1.cloud.ld
openstack reservation host create compute4.cloud.ld
openstack reservation host list
+----+---------------------+-------+-----------+----------+
| id | hypervisor_hostname | vcpus | memory_mb | local_gb |
+----+---------------------+-------+-----------+----------+
| 8  | compute1.cloud.ld   |     2 |      9664 |       52 |
| 4  | compute4.cloud.ld   |     2 |      5632 |       22 |
+----+---------------------+-------+-----------+----------+
visualization of compute resources in Horizon

Create some instances

Creation of 8 instances : one "normal" and seven "preemptible"

openstack server create --image cirros --flavor g1 --network Ext-Net vm1
for i in $(seq 2 8); do
  openstack server create --image cirros --flavor g1.preempt --network Ext-Net vm${i}
done
visualization of instances in Horizon

Notes :

  • Preemptible instances are distributed across compute1 and compute4
  • The normal instance is hosted by a compute node not managed by Blazar

Make a reservation

Reservation of one compute node for 20 minutes in 5 minutes from now. The compute node should have at leat 2 CPUs. The reservation is called lease-1 :

openstack reservation lease create \
--reservation resource_type=physical:host,min=1,max=1,hypervisor_properties='[">=", "$vcpus", "2"]' \
--start-date "$(date --date '+5 min' +"%Y-%m-%d %H:%M")" \
--end-date "$(date --date '+20 min' +"%Y-%m-%d %H:%M")" lease-1
+--------------+-------------------------------------------------------------+
| Field        | Value                                                       |
+--------------+-------------------------------------------------------------+
| created_at   | 2024-01-21 22:00:57                                         |
| degraded     | False                                                       |
| end_date     | 2024-01-21T22:20:00.000000                                  |
| events       | {                                                           |
|              |     "created_at": "2024-01-21 22:00:58",                    |
|              |     "updated_at": null,                                     |
|              |     "id": "1b4700d4-960d-46c0-8022-af743bd778ce",           |
|              |     "lease_id": "7180e34e-2718-47c9-8616-590e6a3585b2",     |
|              |     "event_type": "end_lease",                              |
|              |     "time": "2024-01-21T22:20:00.000000",                   |
|              |     "status": "UNDONE"                                      |
|              | }                                                           |
|              | {                                                           |
|              |     "created_at": "2024-01-21 22:00:58",                    |
|              |     "updated_at": null,                                     |
|              |     "id": "3f8a4928-a360-495d-826d-4b92a5b7a003",           |
|              |     "lease_id": "7180e34e-2718-47c9-8616-590e6a3585b2",     |
|              |     "event_type": "start_lease",                            |
|              |     "time": "2024-01-21T22:05:00.000000",                   |
|              |     "status": "UNDONE"                                      |
|              | }                                                           |
|              | {                                                           |
|              |     "created_at": "2024-01-21 22:00:58",                    |
|              |     "updated_at": null,                                     |
|              |     "id": "564bd93c-e157-4dcb-825a-2258d882a4d2",           |
|              |     "lease_id": "7180e34e-2718-47c9-8616-590e6a3585b2",     |
|              |     "event_type": "before_end_lease",                       |
|              |     "time": "2024-01-21T22:05:00.000000",                   |
|              |     "status": "UNDONE"                                      |
|              | }                                                           |
| id           | 7180e34e-2718-47c9-8616-590e6a3585b2                        |
| name         | lease-1                                                     |
| project_id   | e1bbe999a4a84d91bbf79eaf54adc60d                            |
| reservations | {                                                           |
|              |     "created_at": "2024-01-21 22:00:57",                    |
|              |     "updated_at": "2024-01-21 22:00:58",                    |
|              |     "id": "e379e7c5-b708-4822-80a3-d9e9a0c573a7",           |
|              |     "lease_id": "7180e34e-2718-47c9-8616-590e6a3585b2",     |
|              |     "resource_id": "80c24e3a-beb5-4f72-a566-31a1be252ba5",  |
|              |     "resource_type": "physical:host",                       |
|              |     "status": "pending",                                    |
|              |     "missing_resources": false,                             |
|              |     "resources_changed": false,                             |
|              |     "hypervisor_properties": "[\">=\", \"$vcpus\", \"2\"]", |
|              |     "resource_properties": "",                              |
|              |     "before_end": "default",                                |
|              |     "min": 1,                                               |
|              |     "max": 1                                                |
|              | }                                                           |
| start_date   | 2024-01-21T22:05:00.000000                                  |
| status       | PENDING                                                     |
| trust_id     | 9e5864d9805b4ff1bb452730fef859ab                            |
| updated_at   | 2024-01-21 22:00:58                                         |
| user_id      | e2cde8831daf4285b148499159c4c14e                            |
+--------------+-------------------------------------------------------------+

Note the reservation id for future use : e379e7c5-b708-4822-80a3-d9e9a0c573a7
The reservation is PENDING, i.e. not ACTIVE yet.
After a few minutes, the reservation becomes ACTIVE :

openstack reservation lease show lease-1
+--------------+-------------------------------------------------------------+
| Field        | Value                                                       |
+--------------+-------------------------------------------------------------+
| created_at   | 2024-01-21 22:00:57                                         |
| degraded     | False                                                       |
| end_date     | 2024-01-21T22:20:00.000000                                  |
| events       | {                                                           |
|              |     "created_at": "2024-01-21 22:00:58",                    |
|              |     "updated_at": null,                                     |
|              |     "id": "1b4700d4-960d-46c0-8022-af743bd778ce",           |
|              |     "lease_id": "7180e34e-2718-47c9-8616-590e6a3585b2",     |
|              |     "event_type": "end_lease",                              |
|              |     "time": "2024-01-21T22:20:00.000000",                   |
|              |     "status": "UNDONE"                                      |
|              | }                                                           |
|              | {                                                           |
|              |     "created_at": "2024-01-21 22:00:58",                    |
|              |     "updated_at": "2024-01-21 22:05:11",                    |
|              |     "id": "3f8a4928-a360-495d-826d-4b92a5b7a003",           |
|              |     "lease_id": "7180e34e-2718-47c9-8616-590e6a3585b2",     |
|              |     "event_type": "start_lease",                            |
|              |     "time": "2024-01-21T22:05:00.000000",                   |
|              |     "status": "DONE"                                        |
|              | }                                                           |
|              | {                                                           |
|              |     "created_at": "2024-01-21 22:00:58",                    |
|              |     "updated_at": "2024-01-21 22:05:12",                    |
|              |     "id": "564bd93c-e157-4dcb-825a-2258d882a4d2",           |
|              |     "lease_id": "7180e34e-2718-47c9-8616-590e6a3585b2",     |
|              |     "event_type": "before_end_lease",                       |
|              |     "time": "2024-01-21T22:05:00.000000",                   |
|              |     "status": "DONE"                                        |
|              | }                                                           |
| id           | 7180e34e-2718-47c9-8616-590e6a3585b2                        |
| name         | lease-1                                                     |
| project_id   | e1bbe999a4a84d91bbf79eaf54adc60d                            |
| reservations | {                                                           |
|              |     "created_at": "2024-01-21 22:00:57",                    |
|              |     "updated_at": "2024-01-21 22:05:11",                    |
|              |     "id": "e379e7c5-b708-4822-80a3-d9e9a0c573a7",           |
|              |     "lease_id": "7180e34e-2718-47c9-8616-590e6a3585b2",     |
|              |     "resource_id": "80c24e3a-beb5-4f72-a566-31a1be252ba5",  |
|              |     "resource_type": "physical:host",                       |
|              |     "status": "active",                                     |
|              |     "missing_resources": false,                             |
|              |     "resources_changed": false,                             |
|              |     "hypervisor_properties": "[\">=\", \"$vcpus\", \"2\"]", |
|              |     "resource_properties": "",                              |
|              |     "before_end": "default",                                |
|              |     "min": 1,                                               |
|              |     "max": 1                                                |
|              | }                                                           |
| start_date   | 2024-01-21T22:05:00.000000                                  |
| status       | ACTIVE                                                      |
| trust_id     | 9e5864d9805b4ff1bb452730fef859ab                            |
| updated_at   | 2024-01-21 22:05:11                                         |
| user_id      | e2cde8831daf4285b148499159c4c14e                            |
+--------------+-------------------------------------------------------------+

Let's see the impact on host aggregates with Horizon (compute4.cloud.ld is now reserved and belongs to a special aggregate created for the purpose) :

compute4.cloud.ld is now reserved and belongs to a special aggregate created for the purpose

The preemptibles instances on compute4.cloud.ld are gone :

state of instances in Horizon when the reservation becomes active on compute4.cloud.ld

Let's create an instance with our reservation id :

openstack server create --image cirros --flavor g1 --network Ext-Net --hint reservation=e379e7c5-b708-4822-80a3-d9e9a0c573a7 vm9

And verify the result with Horizon (compute4.cloud.ld is now hosting vm9) :

compute4.cloud.ld is now hosting vm9

At the end of the reservation, lease-1 becomes TERMINATED :

openstack reservation lease show lease-1
+--------------+-------------------------------------------------------------+
| Field        | Value                                                       |
+--------------+-------------------------------------------------------------+
| created_at   | 2024-01-21 22:00:57                                         |
| degraded     | False                                                       |
| end_date     | 2024-01-21T22:20:00.000000                                  |
| events       | {                                                           |
|              |     "created_at": "2024-01-21 22:00:58",                    |
|              |     "updated_at": "2024-01-21 22:20:11",                    |
|              |     "id": "1b4700d4-960d-46c0-8022-af743bd778ce",           |
|              |     "lease_id": "7180e34e-2718-47c9-8616-590e6a3585b2",     |
|              |     "event_type": "end_lease",                              |
|              |     "time": "2024-01-21T22:20:00.000000",                   |
|              |     "status": "DONE"                                        |
|              | }                                                           |
|              | {                                                           |
|              |     "created_at": "2024-01-21 22:00:58",                    |
|              |     "updated_at": "2024-01-21 22:05:11",                    |
|              |     "id": "3f8a4928-a360-495d-826d-4b92a5b7a003",           |
|              |     "lease_id": "7180e34e-2718-47c9-8616-590e6a3585b2",     |
|              |     "event_type": "start_lease",                            |
|              |     "time": "2024-01-21T22:05:00.000000",                   |
|              |     "status": "DONE"                                        |
|              | }                                                           |
|              | {                                                           |
|              |     "created_at": "2024-01-21 22:00:58",                    |
|              |     "updated_at": "2024-01-21 22:05:12",                    |
|              |     "id": "564bd93c-e157-4dcb-825a-2258d882a4d2",           |
|              |     "lease_id": "7180e34e-2718-47c9-8616-590e6a3585b2",     |
|              |     "event_type": "before_end_lease",                       |
|              |     "time": "2024-01-21T22:05:00.000000",                   |
|              |     "status": "DONE"                                        |
|              | }                                                           |
| id           | 7180e34e-2718-47c9-8616-590e6a3585b2                        |
| name         | lease-1                                                     |
| project_id   | e1bbe999a4a84d91bbf79eaf54adc60d                            |
| reservations | {                                                           |
|              |     "created_at": "2024-01-21 22:00:57",                    |
|              |     "updated_at": "2024-01-21 22:20:11",                    |
|              |     "id": "e379e7c5-b708-4822-80a3-d9e9a0c573a7",           |
|              |     "lease_id": "7180e34e-2718-47c9-8616-590e6a3585b2",     |
|              |     "resource_id": "80c24e3a-beb5-4f72-a566-31a1be252ba5",  |
|              |     "resource_type": "physical:host",                       |
|              |     "status": "deleted",                                    |
|              |     "missing_resources": false,                             |
|              |     "resources_changed": false,                             |
|              |     "hypervisor_properties": "[\">=\", \"$vcpus\", \"2\"]", |
|              |     "resource_properties": "",                              |
|              |     "before_end": "default",                                |
|              |     "min": 1,                                               |
|              |     "max": 1                                                |
|              | }                                                           |
| start_date   | 2024-01-21T22:05:00.000000                                  |
| status       | TERMINATED                                                  |
| trust_id     | 9e5864d9805b4ff1bb452730fef859ab                            |
| updated_at   | 2024-01-21 22:20:11                                         |
| user_id      | e2cde8831daf4285b148499159c4c14e                            |
+--------------+-------------------------------------------------------------+

And compute4.cloud.ld returns to the freepool aggregate :

compute4.cloud.ld in freepool aggregate