How Can we Configure Camunda Process Engine in Golang? - go

I found links on process engine Configuration
https://docs.camunda.org/manual/7.16/reference/deployment-descriptors/tags/process-engine/
But how to use this in Golang?

You would not use the embedded process engine approach. Instead run a remote process engine and implement service tasks in Go using the external task patter. To get such a remote engine you can either just sign up for a free developer account on Camunda Cloud. If your prefer on premises, then you can use a self-managed deployment (e.g. via docker compose). For both options check here: https://camunda.com/get-started
Once you have an engine and configured credentials (follow https://docs.camunda.io/docs/guides/getting-started/).
The section https://docs.camunda.io/docs/guides/getting-started/implement-service-task/ shows how to configure the service task on the process engine side. On the Go side you can now implement an external worker (instead of zbctl in the example) as described here: https://github.com/camunda/zeebe/tree/main/clients/go

Related

Webhooks for Oracle Cloud Infrastructure - container registry

Looking for a solution to this use case
Docker image is pushed to Oracle Cloud Infrastructure - container registry (OCIR)
Jenkins has a webhook on the OCIR and Jenkins pipeline gets triggered as a new image is available in OCIR
How is it possible to have a webhook or some kind of mechanish for letting Jenkins know there is a new push to OCIR?
This blog post walks you thru how to set up a continuous pipeline that may be able to be used in full or in part to accomplish this
https://blogs.oracle.com/cloud-infrastructure/build-a-continuous-integration-pipeline-using-github,-docker-and-jenkins-on-oracle-cloud-infrastucture
We can listen to the OCI container registry events via Service Connector. You can configure Service Connector to invoke your custom functions on a specific event 'Container Image - Upload' under service name 'Registry'.
You can find a sample illustration below to perform some custom tasks during an image upload to OCI Container Registry.
Ref: https://github.com/RahulMR42/oci-devops-deploy-on-imageupload

Why do some Jelastic providers block Export Environment option

According to Jelastic documentation it is possible to export the Environment configuration and download it so it can be restored in another provider
However I have tried with 2 Jelastic providers and they both have disabled the option for exporting private data.
So exporting/download/upload/import of environment is not possible.
i.e. I was expecting to have a process similar to CPanel backup/restore tool
In fact, another view for the deployment process gives a possibility to get rid of the model of handling the data or configuration on the platform. Try to think a bit differently and using CI/CD approach. The Jelastic provides a platform where something you created, locate on somewhere you're elaborating(VCS or GIT as an example) and based on or depends on the specific stack, already pre-configured like a layer and can be installed(copied) over Jelastic. Don't need to handle the data somewhere in the cloud because you have it locally(means within any VCS) and doing the changes there. Then just do a 'pull' procedure(manually or automatically) on that deployment(test, production, staging) environment you're expecting.
Moreover, you can expect any environments type like a code and perform it creating before deploying the data.
Please, find the articles being described each case:
Deployment Guide
Jelastic Packaging Standard for CI/CD Automation
In case you would like to handle the databases' backups, check this article:
Scheduling Database Backups
Additional FTP add-on can make the copies more easily for each instance:
FTP/FTPS Support in Jelastic

Is there any way to upload Artifacts to the Apache ACE server without Web UI?

We are working on an enterprise system writed by Java. And we use an Apache ACE server to deploy the OSGi bundles, a Jenkins as CI server. When we want to update a bundle, we make a jar file in Eclipse, and upload it to ACE server through Web UI. When we want to release a new version, we must upload all bundles through Web UI. I think that is foolish.
I think there must be a simple way just like when I finish coding, then I can do something just in Elipse to upload the bundle to the ACE server. When we release a version, the Jenkins should also update all of the bundles to ACE server itself.
Certainly, you basically have two options if you want to automate things:
Use the REST based interface to talk to ACE.
Use the shell based interface to script to ACE.
Both are explained on the website, so for more detailed steps refer to:
http://ace.apache.org/docs/rest-api.html
http://ace.apache.org/docs/shell-api.html

Not terminating but stopping cloud build agent

TeamCity supports creating VMs in the cloud for running builds. It can also terminate instance after it is idle for defined period of time.
Is it possible not to terminate but shutdown the instance and to start it again when needed?
If your BA is a VM on Azure then I was able to achieve this by using Azure Automation account (There is a free version).
Basically, you trigger WebHooks for 2 runbooks you import from the gallery - StartAzureV2VM & StopAzureV2VM. The gallery can be found in "Automation Account-> Runbooks -> Browse Gallery"
Then on your TeamCity server you just periodically run some powershell which monitors the build queue on Teamcity and trigger appropriate webhook (with some timeouts etc.).
Here is the script I'm using, feel free to amend it to your needs:
https://gist.github.com/milanio/b300f23883afa9c6288f9365dfb98252

execute dashDB jobs from Workload Scheduler

I have created some stored procedures on my dashDB instance on Bluemix, to manipulate data in tables in the same instance.
I can run these from Data Studio and they work as intended.
Next, I created a process in Workflow Scheduler, which I provisions as a service in the same app, where the dashDB is also a service.
While creating the job step in the process, I noticed a message in the dialog window. I have attached a screen shot here:
http://i.stack.imgur.com/EI2b7.jpg
When I did try to run the process step from Workflow Scheduler, the process failed with a JDBC not found error.
I do realize that the Workflow agent I'm using is hosted on Bluemix, so I am puzzled how I can install the JDBC client there.
Should I be setting up an agent on a local machine outside of bluemix, in a hybrid mode?
Currently there are 2 possibilities:
Open a ticket to ask for a dedicated cloud agent and then download
the JDBC driver on the agent.
Download and install an agent on a VM or on-prem.
It looks like you have the incorrect value for "JDBC jar class path" the correct value is /home/wauser/utils/
I'm not sure why it is required that we put enter this but I was able to get the connection to dashDB working with this change.
JDBC jar class path: /home/wauser/utils/

Resources