Is there a connector to textura and SAP? We are implementing s/4hana and we have a department that uses textura.
Simple short yes or no answer along with information if there is a connector.
Related
We are trying to use debezium kafka connector to capture postgres changelog. Based on the online tutorial, we need to start a debezium server, and send a POST http request to debezium api to deploy the connector. We want to capture the kafka connector configuration in a code repository, and have automated debezium based kafka connector deployment. What is debezium best practice for this kind of deployment automation?
I am a junior programmer in banking. I want to make a microservice system that get data from kafka and processes it. after that, save to database and send final data to client app. What technology can i use? I plan to use spring bacth and kafka. Can the technology be implemented in my project or is there a better alternative?
To process data from a Kafka topic I recommend you to use Kafka Streams API, especially Spring Kafka Streams.
Kafka Streams and Spring
And to store the data in a database, you should use a Kafka Sink Connector.
Kafka Connect
This approach is very common and easy if your company has a Kafka ecosystem.
In terms of alternatives, here you will find an interesting comparison:
https://scramjet.org/blog/welcome-to-the-family
3 in 1 serverless
Scramjet takes a slightly different approach - 3 platforms in one.
Both the free product https://hub.scramjet.org/ for installation on your server and the cloud platform are available - currently also free in the beta version https://scramjet.org/#join-beta
I am trying to export data from Kafka to Oracle db. I've searched related questions and web but could not understand that we need a platform (confluent etc.. ) or not. I'd been read the link below but it's not clear enough.
https://docs.confluent.io/3.2.2/connect/connect-jdbc/docs/sink_connector.html
So, what we actually need to export data without 3rd party platform? Thanks in advance.
It's not clear what you mean by "third-party" here
What you linked to is Kafka Connect, which is Apache 2.0 Licensed and open source.
Kafka Connect is a plugin ecosystem, you install connectors individually, written by anyone, or write your own, just like any other Java dependency (i.e. a third-party)
The JDBC connector just happens to be maintained by Confluent. and you can configure the Confluent Hub CLI
to install within any Kafka Connect distribution (or use Kafka Connect Docker images from Confluent)
Alternatively, you use Apache Spark, Flink, Nifi, and many other Kafka Consumer libraries to read data and then start an Oracle transaction per record batch
Or you can explore non-JVM kafka libraries as well and use a language you're more familiar with doing Oracle operations with
I have a Java application on SAP Cloud Platform Cloud Foundry that integrates with SAP S/4HANA Cloud (my company's ERP system) by calling APIs (OData services) in that system. I heard about the SAP S/4HANA Cloud SDK and that it makes such scenarios much easier.
How can I leverage the SAP S/4HANA Cloud SDK? Currently, my code to call SAP S/4HANA looks like this (simplified and joined together) for the scenario of retrieving product master data. I have created the S4Product class myself as representation of the response. The baseUrland authHeader are determined before by talking to the destination service on SAP Cloud Platform.
StringBuilder url = new StringBuilder(baseUrl);
url.append("/sap/opu/odata/sap/API_PRODUCT_SRV/A_Product");
url.append("&$select=Product,CreationDate");
url.append("&$filter=ProductType eq '1'");
url.append("&$top=10");
URL urlObj = new URL(url.toString());
HttpURLConnection connection = (HttpURLConnection) urlObj.openConnection();
connection.setRequestMethod("GET");
connection.setRequestProperty("Content-Type", "application/json");
connection.setRequestProperty("Accept", "application/json");
connection.setRequestProperty("Authorization",authHeader);
connection.setDoInput(true);
final InputStreamReader in = new InputStreamReader(connection.getInputStream());
String response = CharStreams.toString(in);
List<S4Product> result = Arrays.asList(new Gson().fromJson(response, S4Product[].class));
Now I'm asked to do something similar with business partners. How do I do this for the business partner OData service, using the SDK? Do I have to create a new application if I want to use the SDK?
With the Java virtual data model of the SAP S/4HANA Cloud SDK, your code would be replaced by something like the following.
final List<Product> products = new DefaultProductMasterService()
.getAllProduct()
.select(Product.PRODUCT, Product.CREATION_DATE)
.filter(Product.PRODUCT_TYPE.eq("1"))
.top(10)
.execute();
This handles everything you have done before manually, in a fluent and type-safe API. In this case, the class Product is provided by the SAP S/4HANA Cloud SDK, no need to create that yourself. It offers a Java representation of the entity type, with all fields, which we are using to define the select and filter query options.
And for your question about business partners, it would look quite similar to this.
final List<BusinessPartner> businessPartners = new DefaultBusinessPartnerService()
.getAllBusinessPartner()
.select(BusinessPartner.BUSINESS_PARTNER /* more fields ... */)
// example filter
.filter(BusinessPartner.BUSINESS_PARTNER_CATEGORY.eq("1"))
.execute();
BTW, this also covers talking to the destination service and applying authentication headers - you no longer need to do this manually.
You can use the SAP S/4HANA Cloud SDK in any Java project. Just include the dependencies com.sap.cloud.s4hana.cloudplatform:scp-cf (for Cloud Foundry) and com.sap.cloud.s4hana:s4hana-all.
Does Confluent by default provides this JMSSourceConnector for Kafka topic.
Or we need to write custom connector for this?
I dont see any documentation on Confluent page on this.
Currently Confluent doesn't provide source connector for JMS. Please find below link for number of connectors available in Kafka Connect.
http://www.confluent.io/product/connectors/
But developers can develop custom connectors for Kafka Connect. Please see the below URL for more information on how to write custom connectors for Kafka Connect.
http://docs.confluent.io/3.0.1/connect/devguide.html