Coder Social home page Coder Social logo

cristianonicolai / kogito-operator Goto Github PK

View Code? Open in Web Editor NEW

This project forked from kiegroup/kogito-operator

0.0 0.0 0.0 35.49 MB

Kubernetes Operator for Kogito

License: Apache License 2.0

Makefile 0.70% Dockerfile 0.05% Shell 4.84% Go 87.54% Gherkin 4.75% Groovy 2.11%

kogito-operator's Introduction

Kogito Operator

Go Report Card CircleCI

The Kogito Operator deploys Kogito Runtimes services from source and all infrastructure requirements for the services, such as persistence with Infinispan and messaging with Apache Kafka. Kogito provides a command-line interface (CLI) that enables you to interact with the Kogito Operator for deployment tasks.

For information about the Kogito Operator architecture and instructions for using the operator and CLI to deploy Kogito services and infrastructures, see the official Kogito Documentation page.

Table of Contents

Created by gh-md-toc

Contributing to the Kogito Operator

Thank you for your interest in contributing to this project!

Any kind of contribution is welcome: code, design ideas, bug reporting, or documentation (including this page).

Trying the Kogito Operator

You can quickly install the operator in your local cluster by executing the following command:

NAMESPACE=mynamespace
VERSION=<current_operator_version>
kubectl apply -n "${NAMESPACE}" -f "https://github.com/kiegroup/kogito-operator/releases/download/${VERSION}/kogito-operator.yaml"

Replace the values above to match your context. The version information you can grab from the releases page.

Alternatively, if you cloned this repo just do:

$ ./hack/install.sh

The script will download the latest version and install the resources for you in the current namespace. You can set the VERSION and NAMESPACE variables before running the script to control which version to install in the given namespace.

Prerequisites

For code contributions, review the following prerequisites:

Kogito Operator environment

The Operator SDK is updated regularly, and the Kogito Operator code typically uses the most recent SDK updates as soon as possible.

If you do not have a preferred IDE, use Visual Studio Code with the vscode-go plugin for Go language tools support.

To use Go modules with VS Code, see Go modules support in VS Code.

To debug Go in your VS code, see Debugging Go code using VS Code.

We check our code with golangci-lint, so it is recommended to add this to your IDE. For adding the golangci-lint with goland, see Go Linter.

For adding the golangci-lint with VScode, install the Go Plugin and enable the linter from the plugins setting.

Kogito Operator unit tests

For information about Operator SDK testing, see Unit testing with the Operator SDK.

In general, the unit tests that are provided with the Kogito Operator are based on that Operator SDK testing resource. You might encounter minor issues as you create specific OpenShift APIs such as BuildConfig and DeploymentConfig that are not listed there. For an example test case with sample API calls, see the kogitobuild_controller_test.go test file.

Kogito Operator collaboration and pull requests

Before you start to work on a new proposed feature or on a fix for a bug, open an issue to discuss your idea or bug report with the maintainers. You can also work on a JIRA issue that has been reported. A developer might already be assigned to address the issue, but you can leave a comment in the JIRA asking if they need some help.

After you update the source with your new proposed feature or bug fix, open a pull request (PR) that meets the following requirements:

  • You have a JIRA associated with the PR.
  • Your PR has the name of the JIRA in the title, for example, [KOGITO-XXX] - Awesome feature that solves it all.
  • The PR solves only the problem described in the JIRA.
  • You have written unit tests for the particular fix or feature.
  • You ran make vet and make test before submitting the PR and everything is working accordingly.
  • You tested the feature on an actual OpenShift cluster.

After you send your PR, a maintainer will review your code and might ask you to make changes and to squash your commits before we can merge.

If you have any questions, contact a Kogito Operator maintainer in the issues page.

Kogito Operator development

Before you begin fixing issues or adding new features to the Kogito Operator, review the previous instructions for contributing to the Kogito Operator repository.

Requirements

Building the Kogito Operator

To build the Kogito Operator, use the following command:

$ make

The output of this command is a ready-to-use Kogito Operator image that you can deploy in any namespace.

Deploying to OpenShift 4.x for development purposes

Prerequisites:

  1. Ensure that you built the Kogito operator first and pushed it to a registry reachable from the OpenShift cluster.

  2. Operator registry tool is installed in your system and available as opm command

Follow the steps below:

  1. Make sure that you defined environment variable BUNDLE_IMG. This variable will be used as an image tag used to store bundle image (containing all the operator metadata).

  2. Make sure that you defined environment variable CATALOG_IMG. This variable will be used as an image tag used to store custom catalog image (containing reference to the custom bundle image).

  3. Make sure that you defined environment variable IMAGE pointing to the custom Kogito operator image tag which you built.

  4. Make sure that you defined environment variable BUILDER defining container runtime engine used to build and push images. Default value is podman.

  5. Update the bundle metadata to point to the custom Kogito operator image by running this command:

$ make update-bundle
  1. Build the bundle image by running this command:
$ make bundle-build
  1. Push the built bundle image by running this command:
$ make bundle-push
  1. Build the catalog image by running this command:
$ make catalog-build
  1. Push the built catalog image by running this command:
$ make catalog-push
  1. Define pushed catalog image as a catalog source for your OpenShift cluster by running this command:
$ cat << EOF | oc apply -f -
apiVersion: operators.coreos.com/v1alpha1
kind: CatalogSource
metadata:
  name: custom-kogito-catalog
  namespace: openshift-marketplace
spec:
  sourceType: grpc
  image: ${CATALOG_IMG}
EOF

After several minutes, the Operator appears under Catalog -> OperatorHub in the OpenShift Web Console. To find the Operator, filter the provider type by custom-kogito-catalog.

Running BDD Tests

REQUIREMENTS:

  • You need to be authenticated to the cluster before running the tests.
  • Native tests need a node with at least 4 GiB of memory available (build resource request).

If you have an OpenShift cluster and admin privileges, you can run BDD tests with the following command:

$ make run-tests [key=value]*

You can set those optional keys:

  • feature is a specific feature you want to run.
    If you define a relative path, this has to be based on the "test" folder as the run is happening there. Default are all enabled features from 'test/features' folder
    Example: feature=features/operator/deploy_quarkus_service.feature

  • tags to run only specific scenarios. It is using tags filtering.
    Scenarios with '@disabled' tag are always ignored.
    Expression can be:

    • "@wip": run all scenarios with wip tag
    • "~@wip": exclude all scenarios with wip tag
    • "@wip && ~@new": run wip scenarios, but exclude new
    • "@wip,@undone": run wip or undone scenarios

    Complete list of supported tags and descriptions can be found in List of test tags

  • concurrent is the number of concurrent tests to be ran.
    Default is 1.

  • timeout sets the timeout in minutes for the overall run.
    Default is 240 minutes.

  • debug to be set to true to activate debug mode.
    Default is false.

  • load_factor sets the tests load factor. Useful for the tests to take into account that the cluster can be overloaded, for example for the calculation of timeouts.
    Default is 1.

  • local to be set to true if running tests in local using either a local or remote cluster. Default is false.

  • ci to be set if running tests with CI. Give CI name.

  • cr_deployment_only to be set if you don't have a CLI built. Default will deploy applications via the CLI.

  • load_default_config sets to true if you want to directly use the default test config (from test/.default_config)

  • container_engine engine used to interact with images and local containers. Default is docker.

  • domain_suffix domain suffix used for exposed services. Ignored when running tests on Openshift.

  • image_cache_mode Use this option to specify whether you want to use image cache for runtime images. Available options are 'always', 'never' or 'if-available'(default).

  • http_retry_nb sets the retry number for all HTTP calls in case it fails (and response code != 500). Default is 3.

  • olm_namespace Set the namespace which is used for cluster scope operators. Default is 'openshift-operators'.

  • operator_image is the Operator image full name.
    Default: operator_image=quay.io/kiegroup/kogito-operator.
  • operator_tag is the Operator image tag.
    Default is the current version.
  • operator_namespaced Set true if you want to run operator namespace scoped. Default is false.
  • operator_installation_source Defines what source is used to install Kogito operator. Available options are olm and yaml. Default is yaml.
  • operator_catalog_image Specifies catalog image containing Kogito operator bundle. Needs to be specified when operator_installation_source is set to olm.
  • deploy_uri set operator deploy folder.
    Default is ./deploy.
  • cli_path set the built CLI path.
    Default is ./build/_output/bin/kogito.
  • services_image_version sets the services (jobs-service, data-index, ...) image version.
  • services_image_namespace sets the services (jobs-service, data-index, ...) image namespace.
  • services_image_registry sets the services (jobs-service, data-index, ...) image registry.
  • data_index_image_tag sets the Kogito Data Index image tag ('services_image_version' is ignored)
  • explainability_image_tag sets the Kogito Explainability image tag ('services_image_version' is ignored)
  • jobs_service_image_tag sets the Kogito Jobs Service image tag ('services_image_version' is ignored)
  • management_console_image_tag sets the Kogito Management Console image tag ('services_image_version' is ignored)
  • task_console_image_tag sets the Kogito Task Console image tag ('services_image_version' is ignored)
  • trusty_image_tag sets the Kogito Trusty image tag ('services_image_version' is ignored)
  • trusty_ui_image_tag sets the Kogito Trusty UI image tag ('services_image_version' is ignored)
  • custom_maven_repo sets a custom Maven repository url for S2I builds, in case your artifacts are in a specific repository. See https://github.com/kiegroup/kogito-images/README.md for more information.
  • maven_mirror is the Maven mirror URL.
    This is helpful when you need to speed up the build time by referring to a closer Maven repository.
  • build_image_registry sets the build image registry.
  • build_image_namespace sets the build image namespace.
  • build_image_name_suffix sets the build image name suffix to append to usual image names.
  • build_image_version sets the build image version
  • build_s2i_image_tag sets the build S2I image full tag.
  • build_runtime_image_tag sets the build Runtime image full tag.
  • disable_maven_native_build_container disables the default Maven native build done in container.
  • runtime_application_image_registry sets the registry for built runtime applications.
  • runtime_application_image_namespace sets the namespace for built runtime applications.
  • runtime_application_image_name_prefix sets the image name prefix to prepend to usual image names for built runtime applications.
  • runtime_application_image_name_suffix sets the image name suffix to append to usual image names for built runtime applications.
  • runtime_application_image_version sets the version for built runtime applications.
  • show_scenarios sets to true to display scenarios which will be executed.
    Default is false.
  • show_steps sets to true to display scenarios and their steps which will be executed.
    Default is false.
  • dry_run sets to true to execute a dry run of the tests, disable crds updates and display the scenarios which will be executed.
    Default is false.
  • keep_namespace sets to true to not delete namespace(s) after scenario run (WARNING: can be resources consuming ...).
    Default is false.
  • namespace_name to specify name of the namespace which will be used for scenario execution (intended for development purposes).
  • local_cluster to be set to true if running tests using a local cluster. Default is false.

Logs will be shown on the Terminal.

To save the test output in a local file for future reference, run the following command:

make run-tests 2>&1 | tee log.out

Running BDD tests with current branch

$ make
$ docker tag quay.io/kiegroup/kogito-operator:2.0.0-snapshot quay.io/{USERNAME}/kogito-operator:2.0.0-snapshot
$ docker push quay.io/{USERNAME}/kogito-operator:2.0.0-snapshot
$ make run-tests operator_image=quay.io/{USERNAME}/kogito-operator

NOTE: Replace {USERNAME} with the username/group you want to push to. Docker needs to be logged in to quay.io and be able to push to your username/group.

Running BDD tests with custom Kogito Build images' version

$ make run-tests build_image_version=<kogito_version>

Running smoke tests

The BDD tests do provide some smoke tests for a quick feedback on basic functionality:

$ make run-smoke-tests [key=value]*

It will run only tests tagged with @smoke. All options from BDD tests do also apply here.

Running performance tests

The BDD tests also provide performance tests. These tests are ignored unless you specifically provide the @performance tag or run:

$ make run-performance-tests [key=value]*

It will run only tests tagged with @performance. All options from BDD tests do also apply here.

NOTE: Performance tests should be run without concurrency.

List of test tags

Tag name Tag meaning
@smoke Smoke tests verifying basic functionality
@performance Performance tests
@travelagency Travel agency tests
@disabled Disabled tests, usually with comment describing reasons
@cli Tests to be executed only using Kogito CLI
@rhpam Tests to be executed for RHPAM Kogito operator
@springboot SpringBoot tests
@quarkus Quarkus tests
@dataindex Tests including DataIndex
@trusty Tests including Trusty
@jobsservice Tests including Jobs service
@managementconsole Tests including Management console
@trustyui Tests including Trusty UI
@binary Tests using Kogito applications built locally and uploaded to OCP as a binary file
@native Tests using native build
@persistence Tests verifying persistence capabilities
@events Tests verifying eventing capabilities
@discovery Tests checking service discovery functionality
@usertasks Tests interacting with user tasks to check authentication/authorization
@security Tests verifying security capabilities
@failover Tests verifying failover capabilities
@resources Tests checking resource requests and limits
@infinispan Tests using the infinispan operator
@kafka Tests using the kafka operator
@keycloak Tests using the keycloak operator
@knative Tests using the Knative functionality

Running the Kogito Operator locally

To run the Kogito Operator locally, change the log level at runtime with the DEBUG environment variable, as shown in the following example:

$ DEBUG=true make run

You can use the following command to vet, format, lint, and test your code:

$ make vet && make test

Remote Debug Kogito Operator using Intellij IDEA

The operator will be deployed over an Kubernetes cluster. In this example we've taken minikube to deploy a Kubernetes cluster locally.

Install Minikube:

For installing the Minikube cluster please follow this tutorial

$ minikube start

Apply Operator manifests:

$ export NAMESPACE=default
$ ./hack/install-manifests.sh

Install Delve:

Delve is a debugger for the Go programming language. For installing Delve please follow this tutorial

Start Operator in remote debug mode:

$ cd cmd/manager
$ export WATCH_NAMESPACE=default
$ dlv debug --headless --listen=:2345 --api-version=2

verify logs on bash console for below line

API server listening at: [::]:2345

Create the Go Remote run/debug configuration:

  1. Click Edit | Run Configurations. Alternatively, click the list of run/debug configurations on the toolbar and select Edit Configurations.
    alt text
  2. In the Run/Debug Configurations dialog, click the Add button (+) and select Go Remote. alt text
  3. In the Host field, type the host IP address (in our case localhost).
  4. In the Port field, type the debugger port that you configured in above dlv command (in our case it's 2345).
  5. Click OK.
    alt text
  6. Put the breakpoint in your code.
  7. From the list of run/debug configurations on the toolbar, select the created Go Remote configuration and click the Debug <configuration_name> button

Running Kogito operator in remote debug on VSCode and GoLand is very similar to above procedure. Please follow these article to the setup remote debugger on VSCode and GoLand

Guide for Kogito Developers

If you made changes in the core/runtimes part of the Kogito and want to test your changes against the operator. Please follow this guide to test your changes.

Releasing Kogito Operator

When releasing the new version of kogito-operator in community-operators. One can use make olm-manifests it would create the manifests in the build/_output/olm/<version> directory in format which the community-operators repo expects. One can then just copy the directory into the kogito-operator and raise the PR.

Note: One needs to create two separate PRs with this folder added in upstream-kogito-operator and community-kogito-operator respectively.

Before raising the PR, make sure the replaces field in the CSV is correct

Kogito Operator tested integrations

Kogito operator integrates with various other technologies and operators. The tested certification matrix is listed below:

Technology Tested version
Infinispan Infinispan operator 2.0.6 (deployed by OLM 2.0.x channel)
Kafka Strimzi 0.23.0 (deployed by OLM stable channel)
Keycloak Keycloak operator 12.0.3 (deployed by OLM alpha channel)
Prometheus Prometheus operator 0.47.0 (deployed by OLM beta channel)
Grafana Grafana operator 3.10.1 (deployed by OLM alpha channel)
MongoDB MongoDB Community Kubernetes Operator 0.2.2

kogito-operator's People

Contributors

0xc1c4da avatar bsig-gh-bot avatar danielezonca avatar duncandoyle avatar errantepiphany avatar gabriel376 avatar gupta-amrit avatar jiripetrlik avatar kaitou786 avatar kevin-mok avatar lcaparelli avatar manaswinidas avatar marianmacik avatar mbiarnes avatar mramendi avatar msmagnanijr avatar r00ta avatar radtriste avatar ricardozanini avatar sgitario avatar spolti avatar sterobin avatar sutaakar avatar tarilabs avatar tchughesiv avatar vaibhavjainwiz avatar xieshenzh avatar

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.