Skip to main content

Posts

Spring Data JPA Against Big Data Sources

Spring Data JPA Against Big Data Sources Download Hive JDBC Driver https://www.cloudera.com/downloads/connectors/hive/jdbc/2-6-5.html References https://docs.cloudera.com/documentation/other/connectors/hive-jdbc/2-6-9.html https://community.cloudera.com/t5/Community-Articles/Writing-Spring-Boot-Microservices-To-Access-Hive/ta-p/247250 https://shalishvj.wordpress.com/2018/06/02/hive-jdbc-spring-boot-restful-webservice-in-pivotal-cloud-foundry/ https://github.com/tspannhw/hive/blob/master/pom.xml https://github.com/firecodeman/Cloudera-Impala-Hive-JDBC-Example/blob/master/src/main/java/com/cloudera/example/ClouderaHiveJdbcExample.java https://docs.cloudera.com/documentation/other/connectors/hive-jdbc/Cloudera-JDBC-Driver-for-Apache-Hive-Release-Notes.txt https://start.spring.io/

Using the Mm FLaNK Stack for Edge AI

New Features of Apache NiFi 1.13.2

 New Features of Apache NiFi 1.13.2 Check it out :    https://twitter.com/pvillard31/status/1361569608327716867?s=27 Download today :   https://nifi.apache.org/download.html Release Note s:   https://cwiki.apache.org/confluence/display/NIFI/Release+Notes#ReleaseNotes-Version1.13.0 Migration :  https://cwiki.apache.org/confluence/display/NIFI/Migration+Guidance New Features ListenFTP UpdateHiveTable - Hive DDL changes -Hive Update Schema ie Data Drift ie Hive Schema Migration!!!! SampleRecord - different sampling approaches to records ( Interval Sampling,  Probabilistic Sampling,  Reservoir Sampling) CDC Updates Kudu updates AMQP and MQTT Integration Upgrades ConsumeMQTT - readers and writers added HTTP access to NiFi by default is now configured to accept connections to 127.0.0.1/localhost only.  If you want to allow broader access for some reason for HTTP and you understand the security implications you can still control that as always by changing the ' nifi.web.http.host' pr

Drivers to use with NiFi

Cloudera JDBC 2.6.20 driver for Apache Impala.   This release has the following enhancements and fixes: Enhancements and New Features [IMPJ-601] Updated third-party libraries The JDBC 4.1 driver has been updated to use the following libraries: log4j 2.2.1 slf4j 1.7.30 [IMPJ-607] Updated CDP support The driver now supports CDP 7.1 For a list of supported CDP versions, see the Installation and Configuration Guide https://www.cloudera.com/downloads/connectors/impala/jdbc/2-6-20.html https://docs.cloudera.com/documentation/other/connectors/impala-jdbc/2-6-20.html https://www.datainmotion.dev/2021/02/ingest-into-cloud.html https://nifi.apache.org/docs/nifi-docs/components/org.apache.nifi/nifi-dbcp-service-nar/1.13.0/org.apache.nifi.dbcp.DBCPConnectionPool/index.html https://www.datainmotion.dev/2019/10/migrating-apache-flume-flows-to-apache_15.html https://dev.to/tspannhw/read-apache-impala-apache-kudu-tables-and-send-to-apache-kafka-in-bulk-easily-with-apache-nifi-4n3g https://www.datainmo

Ingest Into the Cloud

Ingest Into the Cloud Ingesting Data into Apache Kafka in CDP Public Cloud Describes how to use an Apache NiFi data flow to ingest data into Apache Kafka in CDP Public Cloud. Ingesting Data into Apache Hive in CDP Public Cloud Describes how to use an Apache NiFi data flow to ingest data into Apache Hive in CDP Public Cloud. Ingesting Data into Apache HBase in CDP Public Cloud Describes how to use an Apache NiFi data flow to ingest data into Apache HBase in CDP Public Cloud. Ingesting Data into Apache Kudu in CDP Public Cloud Describes how to use an Apache NiFi data flow to ingest data into Apache Kudu in CDP Public Cloud. Ingesting Data into Amazon S3 Buckets Describes how to use an Apache NiFi data flow to ingest data into Amazon S3 buckets in CDP Public Cloud. Ingesting Data into Azure Data Lake Storage Describes how to use an Apache NiFi data flow to ingest data into Azure Data Lake Storage in CDP Public Cloud. Ingesting Data into Google Cloud Storage Describes how to use an Apache

Using Apache NiFi in OpenShift and Anywhere Else to Act as Your Global Integration Gateway

Using Apache NiFi in OpenShift and Anywhere Else to Act as Your Global Integration Gateway What does it look like? Where Can I Run This Magic Engine: Private Cloud, Public Cloud, Hybrid Cloud, VM, Bare Metal, Single Node, Laptop, Raspberry Pi or anywhere you have a 1GB of RAM and some CPU is a good place to run a powerful graphical integration and dataflow engine.   You can also run MiNiFi C++ or Java agents if you want it even smaller. Sounds Too Powerful and Expensive: Apache NiFi is Open Source and can be run freely anywhere. For What Use Cases: Microservices, Images, Deep Learning and Machine Learning Models, Structured Data, Unstructured Data, NLP, Sentiment Analysis, Semistructured Data, Hive, Hadoop, MongoDB, ElasticSearch, SOLR, ETL/ELT, MySQL CDC, MySQL Insert/Update/Delete/Query, Hosting Unlimited REST Services, Interactive with Websockets, Ingesting Any REST API, Natively Converting JSON/XML/CSV/TSV/Logs/Avro/Parquet, Excel, PDF, Word Documents, Syslog, Kafka, JMS, MQTT, TCP

Automating Starting Services in Apache NiFi and Applying Parameters

Automating Starting Services in Apache NiFi and Applying Parameters Automate all the things!   You can call these commands interactively or script all of them with awesome devops tools.  Andre and Dan can tell you more about that. Enable All NiFi Services on the Canvas By running this three times, I get any stubborn ones or ones that needed something previously running.   This could be put into a loop and check the status before trying again. nifi pg-list nifi pg-status nifi pg-get-services  The NiFi CLI has interactive help available and also some good documentation: https://nifi.apache.org/docs/nifi-docs/html/toolkit-guide.html#nifi_CLI /opt/demo/nifi-toolkit-1.12.1/bin/cli.sh nifi pg-enable-services   -u http://edge2ai-1.dim.local:8080 --processGroupId root   /opt/demo/nifi-toolkit-1.12.1/bin/cli.sh nifi pg-enable-services  -u http://edge2ai-1.dim.local:8080 --processGroupId root   /opt/demo/nifi-toolkit-1.12.1/bin/cli.sh nifi pg-enable-services  -u http://edge2ai-1.dim.local:8080