Skip to main content

Posts

Ingesting Websocket Data for Live Stock Streams with Cloudera Flow Management Powered by Apache NiFi

Ingesting Websocket Data for Live Stock Streams with Cloudera Flow Management Powered by Apache NiFi The stocks I follow have a lot of trades and changes throughout the day, I would like to capture all of this data and make it available to my colleagues.   I will push it to Kafka and make it available via a topic and I may also push it to Slack or Dischord or a webpage or dashboard or Cloudera Visual App dashboard.   We'll see what people request. We will read websockets from wss://ws.finnhub.io?token=YOURTOKEN.   You will need to sign up for a finnhub .io account to get this data.   The API is well documented and very easy to use with Apache NiFi. As updates happen we receive websocket calls and send them to Kafka for use in Flink SQL, Kafka Connect, Spark Streaming, Kafka Streams, Python, .Java Spring Boot Apps, NET Apps and NIFi. Definition of Fields s Symbol. p Last price. t UNIX milliseconds timestamp. v Volume. c List of trade conditions. A comprehensive list of trade condit

Smart Stocks with FLaNK (NiFi, Kafka, Flink SQL)

Smart Stocks with FLaNK (NiFi, Kafka, Flink SQL) I would like to track stocks from IBM and Cloudera frequently during the day using Apache NiFi to read the REST API.   After that I have some Streaming Analytics to perform with Apache Flink SQL and I also want permanent fast storage in Apache Kudu queried with Apache Impala. Let's build that application cloud native in seconds in AWS or Azure. Source Code:   https://github.com/tspannhw/SmartStocks To Script Loading Schemas, Tables, Alerts see scripts/setup.sh: Source Code:   https://github.com/tspannhw/ApacheConAtHome2020 Kafka Topic Kafka Schema Kudu Table Flink Prep Flink SQL Client Run Flink SQL Client Configuration Once our automated admin has built our cloud environment and populated it with the goodness of our app, we can being out continuous sql. If you know your data, build a schema, share to the registry One unique thing we added was a default value  in our Avro schema and making it a logicalType  for timestamp-millis.  Thi