The centreon poller does not run after exporting the configuration - poller

I want to connect my poller to start monitoring. Unfortunately when exporting the configuration, the poller does not start running
View no executed poller (NON)
I tried to change the IP of the poller, restart the gorgon services of the two machines, nothing worked.
Anyone have an idea of ​​the solution?

One common oversight on fresh installs is to not disable the firewall and the selinux. From personal experience, it may help to restart all processes in the command line after the first config export attempt from the web interface (Configuration > Pollers > Export Configurations).
On pollers:
systemctl enable centreon centengine
On Central:
systemctl restart php-fpm httpd24-httpd centreon cbd centengine gorgoned
Otherwise, I might need more information. If your issue hasn't been resolved yet, mind sharing:
Your centreon version (from the capture I'd guess Centreon 22.04)
You install medium (from source, packages, OVA/VM)
Your distro
You may find some more clues in the logs:
PHP error logs
For version using PHP 7.2 or 7.3 on centOs 8 or PHP 8
tail -f /var/log/php-fpm/centreon-error.log
For version using PHP 7.3 on centOs 7
tail -f /var/opt/rh/rh-php73/log/php-fpm/centreon-error.log
For version using PHP 7.2 on centOs 7
tail -f /var/opt/rh/rh-php72/log/php-fpm/centreon-error.log
centreon-engine logs (On Central and on Poller)
tail -f /var/log/centreon-engine/centengine.log
centreon-broker logs (On Central)
tail -f /var/log/centreon-broker/central-broker-master.log
centreon gorgone logs (On Central and on Poller)
tail -f /var/log/centreon-gorgone/gorgoned.log

Related

Cannot access Flink dashboard localhost:8081 on windows

I follow the first steps to install Flink.
I can start the cluster without any problem
$ start-cluster.sh
Starting cluster.
Starting standalonesession daemon on host DESKTOP-....
Starting taskexecutor daemon on host DESKTOP-....
But I don't get any status from
$ ps aux | grep flink
I can also not access the dashboard via localhost:8081.
There is an older post having these issues, but the solution didn't work for me, since the described conf files do no longer exist, apparently.
My JAVA_HOME is set as C:\Progra~1\Java\jdk1.8.0_311 to avoid issues with the space in Program Files.
Can you check the logs in the /logs folder? I'm suspecting that C:\Program Files\ could still cause issues because of the space there.
go to download Flink folder and try bash command
$./bin/start-cluster.sh --daemon bootstrap-server localhost:8081
and run code one more
$ ./bin/flink run examples/streaming/WordCount.jar
if you finished run above code which not issue, go to localhost:8081
This still seems to be problematic. I tried to run from Windows Subsystem for Linux (WSL).
I have the following versions: java 11.0.16 and flink 1.15.2.
sudo apt-get update
sudo apt install openjdk-11-jre-headless
export FLINK_HOME=/mnt/c/Projects/Apache/flink-1.15.2
I set the following in flink-conf.yaml
rest.port: 8081
rest.address: localhost
rest.bind-adress: 0.0.0.0
Whereby I changed the bind address for localhost to 0.0.0.0 this seems to have fixed the problem.
$FLINK_HOME/bin/start-cluster.sh
Now I can access the Flink Web Dashboard.

How do I help mitigate log4j via haproxy on Enterprise Linux

Based on this post, haproxy has provided mitigation acls rules that can be used to help fight against log4j attack requests getting proxied to the affected log4j apps.
In reading some of the users comments, It came to my attention that many Enterprise Linux Haproxy system out there might be running an older haproxy version 1.5 which doesn't have the option http-buffer-request directive. This directive is critical to any CVE you are trying to mitigate, even more so if this is your only option until you can upgrade the affected applications.
what is the easiest way to upgrade and to what version?
Answering my own question...
Haproxy 1.6 is needed, however RH provides haproxy 1.8 from the RHSCL repo
Make sure these repos are active on the system and install haproxy1.8
subscription-manager repos --enable=rhel-7-server-optional-rpms --enable=rhel-7-server-rpms --enable=rhel-server-rhscl-7-rpms
yum install -y rh-haproxy18.x86_64
cat the current config into the 1.8 cfg file and run a quick test config
cat /etc/haproxy/haproxy.cfg > /etc/opt/rh/rh-haproxy18/haproxy/haproxy.cfg
/opt/rh/rh-haproxy18/root/usr/sbin/haproxy -c -V -f /etc/opt/rh/rh-haproxy18/haproxy/haproxy.cfg
^ correct any warns or alert errors, In my experience, its mainly directive order or check port strings on your backend services. this is relatively simple to search the net for and correct
nano /etc/opt/rh/rh-haproxy18/haproxy/haproxy.cfg
Add the ACLS Rules, use unique names if using more than one listener/frontend in the cfg so they don't overlap
systemctl disable --now haproxy && systemctl enable --now rh-haproxy18-haproxy
yum remove -y haproxy && yum install -y rh-haproxy18-haproxy-syspaths.x86_64
^ the rh-haproxy18-haproxy-syspaths.x86_64 package replaced the haproxy service name so you do not have to update the systemctl scripts or even keepalived if you are using it to monitor haproxy service
I use this within keepalived which works before and after
killall -0 haproxy
HAProxy have blogged about this topic.
December/2021 – CVE-2021-44228: Log4Shell Remote Code Execution Mitigation

SolrCore is loading running as Windows Service

Logged into Windows Server 2016 as Administrator, I can run Solr from the command line: bin\solr.cmd start -p 8983 -f
I have configured a Solr to run as a Windows Service - running as the same user, with the same command, same startup directory, etc. - however under load, the following error comes back from the upstream application (Sitecore xConnect, though this shouldn't make a difference)
{metadata={error-class=org.apache.solr.common.SolrException,root-error-class=org.apache.solr.common.SolrException},msg=SolrCore is loading,code=503}
To reiterate, everything works fine when Solr is started from the command line, only when it's run as a Windows Service does it error.
Solr version: 6.6.3
Windows version: Server 2016
Environment: AWS (m5.large EC2 instance)
Sitecore compatibility table says to use Solr 6.6.1 with Sitecore, You should still use 6.6.2 as it fixes a bug in Solr 6.6.1 that can affect the installation of SIF. Read here
I recommend you try again with Solr 6.6.2
It turns out that the service was configured to run without the -f flag. So the process would continually stop and re-spawn.

how to update network using composer install

I am developing a nodejs client for a business network. I install the network version 100 . after I make some changes I redeployed it however the changes did not effect at least the nodejs client still logs it is connected to version 100. I increment the version 101 restart the nodejs still got 100 in the log.
what is the procedure to update the network?
deploy command
composer network deploy -a network.bna -p hlfv1 -i PeerAdmin -s anything
oepn connection
this.bizNetworkConnection.connect(this.CONNECTION_PROFILE_NAME, this.businessNetworkIdentifier, participantId, participantPwd)
log network identifier
LOG.info('KYC:', 'businessNetworkDefinition obtained', this.businessNetworkDefinition.getIdentifier());
I think you need composer network update to update the version of an already deployed network.

Elasticsearch windows service not work on Nano server

I am now trying to install ES on windows 2016 Nano server, it works when start ES in command line, but failed to start ES as windows service, maybe due to some changes on Nano server, but I have no hint at all why it didn't work.
Anyone has experience on installing ES on Nano server?
or Where can I get the source code of ES windows service, elasticsearch-windows-x64.exe? It could be helpful to troubleshoot the root cause.
I haven't got a chance to test it on Nano Server 2016, but I've installed ES 5.1.2 + JDK1.8.0_121 on Windows 2016 Server Core successfully and was able to start the service.
Install JDK and set SYSTEM ENV with setx -m JAVA_HOME
"C:\PROGRA~1\Java\jdk1.8.0_121" command. It doesn't work with user level ENV variable such as using set JAVA_HOME=xxx.
unzip ES to c:\elk\elasticsearch
Edit the elasticsearch.yml config file, config the Cluster and Node names and save it
install the ES service C:\elk\elasticsearch\bin>elasticsearch-service.bat install
start the ES service C:\elk\elasticsearch\bin>elasticsearch-service.bat start
Reference:
https://github.com/elastic/elasticsearch/issues/18373
http://blog.technotesdesk.com/setting-a-system-environment-variable-from-command-line-in-windows-server-2012-r2/

Resources