Common Inventory
This repository implements a common inventory system with eventing.
![License](https://img.shields.io/badge/License-Apache_2.0-blue.svg)
Table of Contents
Development Setup
Prerequisites
Running locally
When running locally, the default settings file is used. By default, this configuration does the following:
- Exposes the inventory API in
localhost
and using port 8000
for http and port 9000
for grpc.
- Sets authentication mechanism to
allow-unauthenticated
, allowing users to be authenticated with their user-agent value.
- Sets authorization mechanism to
allow-all
.
- Configures eventing mechanism to go to stdout.
- Sets database implementation to sqlite3 and the database file to
inventory.db
- Configures log level to
INFO
.
NOTE: You can update the default settings file as required to test different scenarios. Refer to the command line help (make run-help
)
for information on the different parameters.
-
Clone the repository and navigate to the directory.
-
Install the required dependencies
make init
-
Build the project
# when building locally, use the local-build option as FIPS_ENABLED is set by default for builds
make local-build
-
Run the database migration
make migrate
-
Start the development server
make run
Overriding commands in the Makefile
Due to various alternatives to running some images, we accept some arguments to override certains tools
GO binary
Since there are official instructions on how to manage multiple installs
We accept the GO
parameter when running make. e.g.
GO=go1.23.1 make run
or
export GO=go1.23.1
make run
Podman / Docker
We will use podman
if it is installed, and we will fall back to docker
. You can also specify if you want to ensure a particular binary is used
by providing DOCKER
parameter e.g.
DOCKER=docker make api
or
export DOCKER=docker
make api
Note: The podman-compose
provider struggles with compose files that leverage depends_on
as it can't properly handle the dependency graphs. You can fix this issue on Linux by installing the docker-compose-plugin
or also having docker-compose
installed. When installed, podman uses the docker-compose
provider by default instead. The benefit of the docker-compose-plugin
is that it doesn't require the full Docker setup or docker daemon!
Debugging
See DEBUG for instructions on how to debug
Alternatives way of running this service
Kessel Inventory + Kessel Relations
In order to test Kessel Inventory with Kessel Relations, we recommend cloning the Relations API repo locally and leveraging their existing Docker Compose process to spin up the Relations API.
Both Inventory and Relations compose files are configured to use the same docker network (kessel
) to ensure network connectivity between all containers.
For the Inventory Compose deployment:
- A PSK file is provided with the token "1234".
- Default ports in this setup are
8081
for http and 9091
for grpc to not conflict with Relations
- Refer to inventory-api-compose.yaml for additional configuration
To start use:
make inventory-up
To stop use:
make inventory-down
Kessel-Inventory + Kafka
In order to use the kafka configuration, one has to run strimzi and zookeeper.
You can do this by running;
make inventory-up-kafka
Start Kessel Inventory and configuring it to connect to kafka:
eventing:
eventer: kafka
kafka:
bootstrap-servers: "localhost:9092"
# Adapt as required
# security-protocol: "SASL_PLAINTEXT"
# sasl-mechanism: PLAIN
You can use our default config with kafka by running:
INVENTORY_API_CONFIG="./kafka-inventory-api.yaml" make run
Once started, you can watch the messages using kcat (formerly known as kafkacat)
or by exec into the running container like this:
source ./scripts/check_docker_podman.sh
KAFKA_CONTAINER_NAME=$(${DOCKER} ps | grep inventory-api-kafka | awk '{print $1}')
${DOCKER} exec -i -t ${KAFKA_CONTAINER_NAME} /bin/bash
# Once in the container
./bin/kafka-console-consumer.sh --bootstrap-server localhost:9092 --topic kessel-inventory
Manually terminate Kessel inventory and then run the following to stop kafka:
make inventory-down-kafka
Kessel Inventory + Kessel Relations + Keycloak
Similar as above, but instead of running Kafka, this will configure inventory to use a Keycloak service for authentication.
- Sets up a keycloak instance running at port 8084 with myrealm config file.
- Set up a default service account with clientId:
test-svc
. Refer to get-token to learn how to fetch a token.
- Refer to sso-inventory-api.yaml for additional configuration
As before you'll need to run the Relations Compose steps available in the Relations API repo
To start use:
make inventory-up-sso
Once it has started, you will need to fetch a token and use it when making calls to the service.
To get a token use:
make get-token
You can then export an ENV with that value and use in calls such as:
curl -H "Authorization: bearer ${TOKEN}" # ...
To stop use:
make inventory-down-sso
Running in Ephemeral Cluster with Relations API using Bonfire
Instructions to deploy Kessel Inventory in an ephemeral cluster can be found on Kessel docs
API/Proto files Changes
Once there is any change in the proto
files (under (/api/kessel)[./api/kessel]) an update is required.
This command will generate code and an (openapi)[./openapi.yaml] file from the proto files
.
make api
We can run the following command to update if there are expected breaking changes.
make api_breaking
Build Container Images
By default, the quay repository is quay.io/cloudservices/kessel-inventory
. If you wish to use another for testing, set IMAGE value first
export IMAGE=your-quay-repo # if desired
make docker-build-push
Build Container Images (macOS)
This is an alternative to the above command for macOS users, but should work for any arch
export QUAY_REPO_INVENTORY=your-quay-repo # required
podman login quay.io # required, this target assumes you are already logged in
make build-push-minimal
Example Usage
All these examples use the REST API and assume we are running the default local version
adjustments needs to be made to the curl requests if running with different configuration,
such as port, authentication mechanisms, etc.
Note: When testing in Stage, the current schema leveraged by Relations only supports notifications integrations and not any of the infra we have in our API (RHEL hosts, K8s Clusters, etc). Testing with any other resource type will throw errors from Relations API but will still succeed in Inventory API
Health check endpoints
The Kessel Inventory includes health check endpoints for readiness and liveness probes.
Readyz
The readyz endpoint checks if the service is ready to handle requests.
curl http://localhost:8000/api/inventory/v1/readyz
Livez
The livez endpoint checks if the service is alive and functioning correctly.
curl http://localhost:8000/api/inventory/v1/livez
Resource lifecycle
Resources can be added, updated and deleted to our inventory. Right now we support the following resources:
rhel-host
notifications-integration
k8s-cluster
k8s-policy
To add a rhel-host to the inventory:
To hit the REST endpoint use the following curl
command
curl -H "Content-Type: application/json" --data "@data/host.json" http://localhost:8000/api/inventory/v1beta1/resources/rhel-hosts
To hit the gRPC endpoint use the following grpcurl
command
grpcurl -plaintext -d @ localhost:9000 kessel.inventory.v1beta1.resources.KesselRhelHostService.CreateRhelHost < data/host.json
To update it:
To hit the REST endpoint
curl -XPUT -H "Content-Type: application/json" --data "@data/host.json" http://localhost:8000/api/inventory/v1beta1/resources/rhel-hosts
To hit the gRPC endpoint
grpcurl -plaintext -d @ localhost:9000 kessel.inventory.v1beta1.resources.KesselRhelHostService.UpdateRhelHost < data/host.json
and finally, to delete it, note that we use a different file, as the only required information is the reporter data.
To hit the REST endpoint
curl -XDELETE -H "Content-Type: application/json" --data "@data/host-reporter.json" http://localhost:8000/api/inventory/v1beta1/resources/rhel-hosts
To hit the gRPC endpoint
grpcurl -plaintext -d @ localhost:9000 kessel.inventory.v1beta1.resources.KesselRhelHostService.DeleteRhelHost < data/host-reporter.json
To add a notifications integration (useful for testing in stage)
# create the integration (auth is required for stage -- see internal docs)
curl -H "Content-Type: application/json" -H "Authorization: Bearer $TOKEN" -d @data/notifications-integrations.json localhost:8000/api/inventory/v1beta1/resources/notifications-integrations
# delete the integration
curl -X DELETE -H "Content-Type: application/json" -H "Authorization: Bearer $TOKEN" -d @data/notifications-integration-reporter.json localhost:8000/api/inventory/v1beta1/resources/notifications-integrations
### Adding a new relationship (k8s-policy is propagated to k8s-cluster)
To add a `k8s-policy_ispropagatedto-k8s-cluster` relationship, first lets add the related resources `k8s-policy` and `k8s-cluster`.
```shell
curl -H "Content-Type: application/json" --data "@data/k8s-cluster.json" http://localhost:8000/api/inventory/v1beta1/resources/k8s-clusters
curl -H "Content-Type: application/json" --data "@data/k8s-policy.json" http://localhost:8000/api/inventory/v1beta1/resources/k8s-policies
And then you can create the relation:
curl -H "Content-Type: application/json" --data "@data/k8spolicy_ispropagatedto_k8scluster.json" http://localhost:8000/api/inventory/v1beta1/resource-relationships/k8s-policy_is-propagated-to_k8s-cluster
To update it:
curl -X PUT -H "Content-Type: application/json" --data "@data/k8spolicy_ispropagatedto_k8scluster.json" http://localhost:8000/api/inventory/v1beta1/resource-relationships/k8s-policy_is-propagated-to_k8s-cluster
And finally, to delete it, notice that the data file is different this time. We only need the reporter data.
curl -X DELETE -H "Content-Type: application/json" --data "@data/relationship_reporter_data.json" http://localhost:8000/api/inventory/v1beta1/resource-relationships/k8s-policy_is-propagated-to_k8s-cluster
Configuration
Enable integration with Kessel relations
The default development config has this option disabled. You can check Alternatives way of running this service
for configurations that have Kessel relations enabled.
Supposing Kessel relations is running in localhost:9000
, you can enable it by updating the config as follows:
authz:
impl: kessel
kessel:
insecure-client: true
url: localhost:9000
enable-oidc-auth: false
If you want to enable OIDC authentication with SSO, you can use this instead:
authz:
impl: kessel
kessel:
insecure-client: true
url: localhost:9000
enable-oidc-auth: true
sa-client-id: "<service-id>"
sa-client-secret: "<secret>"
sso-token-endpoint: "http://localhost:8084/realms/redhat-external/protocol/openid-connect/token"
Testing
Tests can be run using:
make test
For end-to-test info see here.
Validating FIPS
Inventory API is configured to build with FIPS capable libraries and produce FIPS capaable binaries when running on FIPS enabled clusters.
To validate the current running container is FIPS capable:
# exec or rsh into running pod
# using fips-detect (https://github.com/acardace/fips-detect)
fips-detect /usr/local/bin/inventory-api
# using go tool
go tool nm /usr/local/bin/inventory-api | grep FIPS
Contributing
Follow the steps below to contribute:
- Fork the project
- Create a new branch for your feature
- Submit a pull request