I want to install Apache Jmeter on PKS (pivotal-container-service) for testing micro service on PCF.
I am not able to find any good resources ,did anyone tried and got success?
https://hub.docker.com/r/justb4/jmeter/
https://www.virtuallyghetto.com/2018/03/getting-started-with-vmware-pivotal-container-service-pks-part-1-overview.html
https://network.pivotal.io/products/pcfdev
Looking into Pivotal Container Service main page
Production-ready Kubernetes for the enterprise
So Pivotal Container Service seems to be using Kubernetes under the hood.
The process of installation of JMeter into Kubernetes and/or Pivotal Container Service would be as easy as creating an Image with your own Docker JMeter installation (or use any existing one) and using this image in the PCS.
Related
How can we run the Hazelcast management center on Kubernetes?
What i did so far is - Deployed the hazelcast/management-center Docker image on our Kubernetes cluster. However as seen on the management-center Pod log, it has started on localhost:8080 by default. I did this with a 'Deployment' on K8s pointing to Docker hazelcast/management-center image. Hazelcast Server is running as the Spring Boot application, also other applications are able to connect to this as Hazelcast client.
Question is how can we run/connect the management center to our Hazelcast cluster running on the same namespace in the Kubernetes (with 2 members in the cluster)?
I suggest using the official helm chart, it includes Management Center by default.
If it doesn't work for you, first I would check if there is a k8s Service defined for the deployment, and if yes, if there is a k8s Ingress exposing that service.
Hi I am new to microservices architect and have few questions related to its deployment on production environment.
I am using spring boot microservices and want to deploy them on any open source plantform like kubernetes.
Previously I have all deployments on local servers(conventional hardware). So I don't want to move to any cloud service for now, its part of future planing. I want to have a deployment strategy that should work with servers and in future when I want to move to cloud, effort should be minimum. I need some expert suggestions. I have created a POC of deploying apps to kubernetes with minikube where I am using docker container. But for docker private repositories I need to buy their enterprise services.
And for any cloud platform I also need invest money.
What are the other choices where I can deploy microservices on my conventional hardware.
Caas vs PaaS, CF vs K8s
There are Two ways of doing it. Actually three, the third being not moving to cloud. If moving to cloud, you do it either via OCF or via Docker. for OCF, we have lot of providers including Pivotal. For Docker, you go via Kubernetes. Again there are a lot of providers, AKS, PKS, GCP, etc.
The choice is always yours.
I have a java application running on a single Instance EC2 server. I want to implement autoscaling in the server, but my java application will not work on active-active mode.
So, I have started looking at Hazelcast IMDG open source for Application Scaling Purpose.
I am new to Hazelcast. So, Can anyone give me the idea about How can I implement Hazelcast open source for my Application Scaling and what will be the steps?
You can deploy a Hazelcast cluster on EC2 with Auto Scaling turned on. See here for details: https://github.com/hazelcast/hazelcast-aws/blob/master/README.md
We have a collection of microservices built with Spring Boot, using Spring Cloud Netflix. Up until now, they've been packaged as RPMs and deployed to VMs. Using Eureka has allowed for service registration/discovery (obviously) and our cross-microservice interaction to be done using Spring's RestTemplate with a Virtual IP (VIP), like the following:
http://foo-service/<PATH_TO_RESOURCE>
Client-side load-balancing was another benefit.
Now, we are looking to use Docker and run within Rancher. I'm wondering using Eureka still makes sense in this environment.
Within Rancher, if the Service is named 'foo-service', that name is used as a VIP within the Rancher internal network so the same URL shown above can also work, sans Eureka.
Also, if there are multiple Containers backing a Service, Rancher will round-robin load-balance traffic amongst them.
Plus, it seems Rancher will know about Containers coming and going sooner than Eureka would.
I'm struggling to find a solid reason to keep Eureka.
Not much familiar with Rancher, AFAIK it enables users to deploy a choice of Cattle, Docker Swarm, Apache Mesos or Kubernetes to manage your containers.
So, it finally comes down to whether your infrastructure platform provides service discovery functionality or not (I know Docker swarm and Kubernetes provides Service discovery, not sure about the others); if you get free service discovery out of the box from your platform and if you don't need client side load balancing, eureka is an overkill.
Here is an answer for the question in context of Kubernetes
https://stackoverflow.com/a/40568412/6785908
Quoting the relevant parts
In Kubernetes platform, using Eureka (Or Consul/zookeeper any
other service registries) for service discovery is an overkill; you
can achieve the same (arguably) functionality with Kubernetes Services
(+kube DNS Addon), which will act as a referable IP address and a load
balancer (not client side) for the ephemeral Pods. Read this
[article][1] by Christian Posta. If you want to refer your service by
its name instead of IP address add KubeDNS (A kubernetes add on) to
your cluster.
http://blog.christianposta.com/microservices/netflix-oss-or-kubernetes-how-about-both/
Edit
Since you said,
Within Rancher, if the Service is named 'foo-service', it is used as a
VIP within the Rancher internal network so the same URL shown above
can also work, sans Eureka.
Also, if there are multiple Containers backing a Service, Rancher will
round-robing load-balance traffic amongst them.
So you are getting both Service discovery and the (server side) load balancer from your platform for free. So if you don't have a compelling reason to do client side load balancing, forget about eureka.
I have a use case to monitor the JVM metrics like(heap size,garbage collection etc) when my JAVA application is running .My application is running on EC2.
can you help me how JCONSOLE be used in EC2,or any other open source or will cloud watch help in doing this when my appication is running
Thanks
You can use AWS CloudWatch metrics to have monitoring of the instance. You can use Datadog services to have extra monitoring. You can use this document as a reference for datadog JMX integration.
If not, you can install JMX Directly on your EC2 instance