WSO2 CEP get UDP feed - wso2

I have successfully installed wso2 cep engine on Solaris 10 box. As events I'm receiving UDP feed (pipe separated data lines). How can I configure CEP engine to receive the UDP feed and process it?
Thanks

We don't yet have a UDP data receiver but we welcome you to write one :). You can push the events directly to the CEP engine (look at how the current data receivers do it) or use the data publisher API if you want to send it to a remote CEP engine (via thrift .. which would be kind of silly!). If you choose to write one please join architecture#wso2.org and ask for help / suggestions. See http://wso2.org/mail.

Related

Client-Server - How do i filter only dedicated software sending data?

I am writing a server and client code where i want to identify on server side official client software sending data.
The protocol is text based, so anyone can write it's own client - and i want to keep it, but want to group the devices by "Official" / "Non Official" based on the 'fingerprint' of my official client application.
Question:
How can i filter out my clients from other clones sending?
I am thinking to send encrypted message from my client to SERVER using the same key, so once decoded if it's valid, then i know its my app code sending - but before i go this way, any other better ideas ?

Stream Audio Via TCP in UE4

I am trying to build a Virtual Assistant in UE4. I need to somehow send my response from DialogFlow to UE4 for use in the Oculus Lipsync Plugin.
Basically, I have 3 media options for the response:
-16-Bit Linear PCM
-mp3
-Ogg Opus
I have a TCP Server and Client connection set up between a Python Script and UE4, so I can send data to and from easily.
I have my sequencing correct so one script waits for the full Byte Array to be sent via the socket etc.
Basically, I want to send the responses to each Query the user sends to DialogFlow INTO UE4 Via my TCP Socket, and be able to play and access that Audio within UE4.
I need to somehow stream the response from DialogFlow into UE4 as it gives me the responses.
Is what I'm trying to do even possible? I'm just trying to stream in Audio for use within UE4 and I am really struggling to get it working. Very annoying as this is the last piece of the puzzle I need to finish....
Please let me know if you have any advice or help you can offer!

ActiveMq: persistent queue and offline system

I'm a fresh user of ActiveMQ technology, and I have some problem approaching this technology.
I have the following situation:
I have a SW, running in a embedded (offline) ARM device, that archive a set of videos on a upluggable hard disk at run time.
Sometimes (4-5 events a day), I have to associate a alarm event to those videos and to queue the alarm on a persistent queue.
Once a month we have to extract the hard disk and to connect it to another embedded online ARM device, that should notify a ActiveMQ server about the alarms generated by the offline ARM device
And now my question: how can I store the persistent queue on the hard disk, so that the events generated byt the offline ARM device will be available to the online ARM system (the only "connection" between online and offline embedded device is hard disk)?
Please note that I cannot change the way I transmit messages to the online server, since it is a system not developed by my company.
Best regards
Giovanni
It sounds like you want a "store-and-forward" messaging pattern. You could configure the "offline" ActiveMQ broker to attempt to connect to the "online" ActiveMQ broker. The network connector will attempt to connect at configurable intervals and when it is "online" it will begin to send messages automatically.
The slight down side is that the broker will attempt to connect to the remote broker (even when offline), so you'll need to manage log rotation or logging levels to accommodate.
Look for the static:// network connector uri
Network of brokers

can I use pymqi to connect to an IBM MQ multi-instance queue manager

Using PyMQI, which is fine on IBM MQ Single instance queues, but does anyone know if I can pass dual IP Addresses and Ports on the connection string and if the MQ CLient under the hood handles the IBM MW Multi Instance queue management ?
PyMQI sits on top of the underlying MQ libraries. If you are using it with MQ v7.0 or higher then you can specify multiple connection names that are separated by a comma. It will then try each one in order and loop back to the first one if it can not connect to any of them. Some settings related to how long it will retry and how often can be set in the mqclient.ini.
The IBM Knowledge center page "Automatic client reconnection" has good general information on the reconnect options. All of it related to the C/C++ clients applies to PyMQI.

How actually we can monitor WSO2 EMM using CEP or BAM?

I am new with wso2 products. I am using EMM for device and app. But i want to monitor my devices continuously.
As some document suggest, i can done this using CEP in real time and also can use BAM. But i did not found any proper document to install or configure these system together (EMM, CEP, BAM).
My EMM server is already live. I am now tring to analysis the http access log to find out registration failed or other error.
But I actually need to monitor the EMM server in real time.
Do I need to run these servers separately and also have to configure them separately? Is there any way to merge all(EMM,CEP,BAM and even ESS) application and run all of them by running one carbon server?
If I need to run these server separately, then how I can configure CEP/BAM as this will monitor EMM server's API/log files/services etc?
Is there documentation for this? I can not find any document related to CEP/BAM related to EMM.
Can any one please help me for this?
WSO2 BAM product has been replaced by WSO2 DAS (Data Analytics Server). Again WSO2 DAS consist with WSO2 CEP features where as it is capable of doing both real time and batch time processing. Therefore you can DAS for both batch processing and real time processing
Both DAS and CEP deals with event streams where as in DAS streams could be persisted to event tables as data needs to be stored for batch processing [2]. On the other hand both can be integrated with other applications by using Receivers for streams. You can protocols such as thrift, soap, http, mqtt, jms and kafka when creating receivers[3]
For real time processing you need to write Execution Plans [4] which uses Siddhi Query Language [5]. DAS supports Apache spark for batch processing. Both products consist with Analytic dashboard which could be used to visualize real time and persistent data [6].
[1] http://wso2.com/products/data-analytics-server/
[2] https://docs.wso2.com/display/DAS300/Understanding+Event+Streams+and+Event+Tables
[3] https://docs.wso2.com/display/DAS300/Configuring+Event+Receivers
[4] https://docs.wso2.com/display/CEP400/Creating+a+Standalone+Execution+Plan
[5] https://docs.wso2.com/display/CEP400/SiddhiQL+Guide+3.0
[6] https://docs.wso2.com/display/DAS300/Analytics+Dashboard
Yes its possible. You will need to use a data publisher to publish events to DAS.