Using FLUME to store data in Hadoop - hdfs

I have followed all the steps for hadoop installation and Flume from tutorials.
I am a naive in Big Data tools. I am getting the following errors. I dont understand, where the problem is?
I have also read a lot of post on installation, but still I am facing this issue. My ultimate objective is to perform Twitter Sentiment Analysis using R.
17/09/29 02:25:39 INFO node.PollingPropertiesFileConfigurationProvider: Configuration provider starting
17/09/29 02:25:39 INFO node.PollingPropertiesFileConfigurationProvider: Reloading configuration file:/home/shivam/apache-flume-1.6.0-bin/conf/flume.conf
17/09/29 02:25:39 INFO conf.FlumeConfiguration: Processing:HDFS
17/09/29 02:25:39 INFO conf.FlumeConfiguration: Processing:HDFS
17/09/29 02:25:39 INFO conf.FlumeConfiguration: Processing:HDFS
17/09/29 02:25:39 INFO conf.FlumeConfiguration: Processing:HDFS
17/09/29 02:25:39 INFO conf.FlumeConfiguration: Processing:HDFS
17/09/29 02:25:39 INFO conf.FlumeConfiguration: Processing:HDFS
17/09/29 02:25:39 INFO conf.FlumeConfiguration: Added sinks: HDFS Agent: TwitterAgent
17/09/29 02:25:39 INFO conf.FlumeConfiguration: Processing:HDFS
17/09/29 02:25:39 INFO conf.FlumeConfiguration: Processing:HDFS
17/09/29 02:25:39 INFO conf.FlumeConfiguration: Processing:HDFS
17/09/29 02:25:39 INFO conf.FlumeConfiguration: Post-validation flume configuration contains configuration for agents: [TwitterAgent]
17/09/29 02:25:39 INFO node.AbstractConfigurationProvider: Creating channels
17/09/29 02:25:39 INFO channel.DefaultChannelFactory: Creating instance of channel MemChannel type memory
17/09/29 02:25:39 INFO node.AbstractConfigurationProvider: Created channel MemChannel
17/09/29 02:25:39 INFO source.DefaultSourceFactory: Creating instance of source Twitter, type org.apache.flume.source.twitter.TwitterSource
17/09/29 02:25:39 INFO twitter.TwitterSource: Consumer Key: 'fRw12aumIqkAWD6PP5ZHk7vva'
17/09/29 02:25:39 INFO twitter.TwitterSource: Consumer Secret: 'K9K0yL2pwngp3JXEdMGWUOEB7AaGWswXcq72WveRvnD4ZSphNQ'
17/09/29 02:25:39 INFO twitter.TwitterSource: Access Token: '771287280438968320-XnbtNtBt40cs6gUOk6F9bjgmUABM0qG'
17/09/29 02:25:39 INFO twitter.TwitterSource: Access Token Secret: 'afUppGRqcRi2p9fzLhVdYQXkfMEm72xduaWD6uNs3HhKg'
17/09/29 02:25:39 INFO sink.DefaultSinkFactory: Creating instance of sink: HDFS, type: hdfs
17/09/29 02:25:39 INFO node.AbstractConfigurationProvider: Channel MemChannel connected to [Twitter, HDFS]
17/09/29 02:25:39 INFO node.Application: Starting new configuration:{ sourceRunners:{Twitter=EventDrivenSourceRunner: { source:org.apache.flume.source.twitter.TwitterSource{name:Twitter,state:IDLE} }} sinkRunners:{HDFS=SinkRunner: { policy:org.apache.flume.sink.DefaultSinkProcessor#3012a48f counterGroup:{ name:null counters:{} } }} channels:{MemChannel=org.apache.flume.channel.MemoryChannel{name: MemChannel}} }
17/09/29 02:25:39 INFO node.Application: Starting Channel MemChannel
17/09/29 02:25:39 INFO instrumentation.MonitoredCounterGroup: Monitored counter group for type: CHANNEL, name: MemChannel: Successfully registered new MBean.
17/09/29 02:25:39 INFO instrumentation.MonitoredCounterGroup: Component type: CHANNEL, name: MemChannel started
17/09/29 02:25:39 INFO node.Application: Starting Sink HDFS
17/09/29 02:25:39 INFO node.Application: Starting Source Twitter
17/09/29 02:25:39 INFO twitter.TwitterSource: Starting twitter source org.apache.flume.source.twitter.TwitterSource{name:Twitter,state:IDLE} ...
17/09/29 02:25:39 INFO twitter.TwitterSource: Twitter source Twitter started.
17/09/29 02:25:39 INFO twitter4j.TwitterStreamImpl: Establishing connection.
17/09/29 02:25:39 INFO instrumentation.MonitoredCounterGroup: Monitored counter group for type: SINK, name: HDFS: Successfully registered new MBean.
17/09/29 02:25:39 INFO instrumentation.MonitoredCounterGroup: Component type: SINK, name: HDFS started
17/09/29 02:25:42 INFO twitter4j.TwitterStreamImpl: Connection established.
17/09/29 02:25:42 INFO twitter4j.TwitterStreamImpl: Receiving status stream.
17/09/29 02:25:42 INFO hdfs.HDFSDataStream: Serializer = TEXT, UseRawLocalFileSystem = false
17/09/29 02:25:42 INFO hdfs.BucketWriter: Creating hdfs://localhost:9000/user/flume/tweets/FlumeData.1506632142370.tmp
17/09/29 02:25:42 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
17/09/29 02:25:44 WARN hdfs.HDFSEventSink: HDFS IO error
java.net.ConnectException: Call From maverick/127.0.1.1 to localhost:9000 failed on connection exception: java.net.ConnectException: Connection refused; For more details see: http://wiki.apache.org/hadoop/ConnectionRefused
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
at org.apache.hadoop.net.NetUtils.wrapWithMessage(NetUtils.java:792)
at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:732)
at org.apache.hadoop.ipc.Client.call(Client.java:1480)
at org.apache.hadoop.ipc.Client.call(Client.java:1407)
at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:229)
at com.sun.proxy.$Proxy13.create(Unknown Source)
at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.create(ClientNamenodeProtocolTranslatorPB.java:296)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:187)
at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)
at com.sun.proxy.$Proxy14.create(Unknown Source)
at org.apache.hadoop.hdfs.DFSOutputStream.newStreamForCreate(DFSOutputStream.java:1623)
at org.apache.hadoop.hdfs.DFSClient.create(DFSClient.java:1703)
at org.apache.hadoop.hdfs.DFSClient.create(DFSClient.java:1638)
at org.apache.hadoop.hdfs.DistributedFileSystem$7.doCall(DistributedFileSystem.java:448)
at org.apache.hadoop.hdfs.DistributedFileSystem$7.doCall(DistributedFileSystem.java:444)
at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
at org.apache.hadoop.hdfs.DistributedFileSystem.create(DistributedFileSystem.java:444)
at org.apache.hadoop.hdfs.DistributedFileSystem.create(DistributedFileSystem.java:387)
at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:909)
at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:890)
at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:787)
at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:776)
at org.apache.flume.sink.hdfs.HDFSDataStream.doOpen(HDFSDataStream.java:86)
at org.apache.flume.sink.hdfs.HDFSDataStream.open(HDFSDataStream.java:113)
at org.apache.flume.sink.hdfs.BucketWriter$1.call(BucketWriter.java:246)
at org.apache.flume.sink.hdfs.BucketWriter$1.call(BucketWriter.java:235)
at org.apache.flume.sink.hdfs.BucketWriter$9$1.run(BucketWriter.java:679)
at org.apache.flume.auth.SimpleAuthenticator.execute(SimpleAuthenticator.java:50)
at org.apache.flume.sink.hdfs.BucketWriter$9.call(BucketWriter.java:676)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
Caused by: java.net.ConnectException: Connection refused
at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:531)
at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:495)
at org.apache.hadoop.ipc.Client$Connection.setupConnection(Client.java:609)
at org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:707)
at org.apache.hadoop.ipc.Client$Connection.access$2800(Client.java:370)
at org.apache.hadoop.ipc.Client.getConnection(Client.java:1529)
at org.apache.hadoop.ipc.Client.call(Client.java:1446)
... 34 more
17/09/29 02:25:45 INFO twitter.TwitterSource: Processed 100 docs
17/09/29 02:25:45 INFO hdfs.BucketWriter: Creating hdfs://localhost:9000/user/flume/tweets/FlumeData.1506632142371.tmp
17/09/29 02:25:45 WARN hdfs.HDFSEventSink: HDFS IO error
java.net.ConnectException: Call From maverick/127.0.1.1 to localhost:9000 failed on connection exception: java.net.ConnectException: Connection refused; For more details see: http://wiki.apache.org/hadoop/ConnectionRefused
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
17/09/29 02:25:48 INFO twitter.TwitterSource: Processed 200 docs
17/09/29 02:25:50 INFO twitter.TwitterSource: Processed 300 docs
17/09/29 02:25:50 INFO hdfs.BucketWriter: Creating hdfs://localhost:9000/user/flume/tweets/FlumeData.1506632142373.tmp
17/09/29 02:25:50 WARN hdfs.HDFSEventSink: HDFS IO error
java.net.ConnectException: Call From maverick/127.0.1.1 to localhost:9000 failed on connection exception: java.net.ConnectException: Connection refused; For more details see: http://wiki.apache.org/hadoop/ConnectionRefused
Is there any complete solution to it. I can do it again from scratch.

Flume is trying to connect to Hadoop's Namenode, which is supposedly listening at localhost:9000, without success.
This behavior is correct: Hadoop's Namenode usually listens at TCP/8020 or TCP/9000 ports for Inter-Process Communications (IPC) related to Hadoop's File System (HDFS). And it seems, by default, Flume tries to connect to TCP/9000.
Can you confirm you have such process running at localhost and listening at TCP/9000? You can do that with lsof or netstat commands. You should also check your Hadoop configuration in order to check which is the port Hadoop is opening for Namenode's IPC.
Then, there are two options:
You change your Namenode's IPC listening port to 9000. This is done by configuring the fs.default.name property at core-site.xml file.
You configure Flume for connecting to the port you have configured at Hadoop. This is done by configuring the hdfs.path property of your sink as hdfs://127.0.0.1:<your_port>/your/path/.

Related

How to make Pyspark script running on Amazon EMR to recognize boto3 module? It says module not found

Spark version 2.4.5
I have files that need to be processed in an S3 bucket. (s3a://tobeprocessed)
I have a pyspark application that reads files from the S3 bucket and writes output to another S3 bucket (s3://processed).
I intend to run this as a step function in my emr cluster.
I used to following command from my terminal to add a step to my cluster.
aws emr add-steps --cluster-id j-xxxxxx --steps Name=etlapp,Jar=command-runner.jar,Args=[spark-submit,--deploy-mode,cluster,--master,yarn,--conf,spark.yarn.submit.waitAppCompletion=true,s3://bucketname/spark_app.py,s3://bucketname/configuration_file.cfg],ActionOnFailure=CONTINUE
I got an error message like this
STDERR
20/03/10 19:50:46 INFO RMProxy: Connecting to ResourceManager at ip-172-31-27-34.ec2.internal/172.31.27.34:8032
20/03/10 19:50:47 INFO Client: Requesting a new application from cluster with 2 NodeManagers
20/03/10 19:50:47 INFO Client: Verifying our application has not requested more than the maximum memory capability of the cluster (12288 MB per container)
20/03/10 19:50:47 INFO Client: Will allocate AM container, with 2432 MB memory including 384 MB overhead
20/03/10 19:50:47 INFO Client: Setting up container launch context for our AM
20/03/10 19:50:47 INFO Client: Setting up the launch environment for our AM container
20/03/10 19:50:47 INFO Client: Preparing resources for our AM container
20/03/10 19:50:47 WARN Client: Neither spark.yarn.jars nor spark.yarn.archive is set, falling back to uploading libraries under SPARK_HOME.
20/03/10 19:50:49 INFO Client: Uploading resource file:/mnt/tmp/spark-4c4ea7ac-b2bb-4a61-929d-c371d87417ff/__spark_libs__2224504543987850085.zip -> hdfs://ip-172-31-27-34.ec2.internal:8020/user/hadoop/.sparkStaging/application_1583867709817_0003/__spark_libs__2224504543987850085.zip
20/03/10 19:50:50 INFO ClientConfigurationFactory: Set initial getObject socket timeout to 2000 ms.
20/03/10 19:50:50 INFO Client: Uploading resource s3://imdbetlapp/complete_etl.py -> hdfs://ip-172-31-27-34.ec2.internal:8020/user/hadoop/.sparkStaging/application_1583867709817_0003/complete_etl.py
20/03/10 19:50:51 INFO S3NativeFileSystem: Opening 's3://imdbetlapp/complete_etl.py' for reading
20/03/10 19:50:51 INFO Client: Uploading resource file:/usr/lib/spark/python/lib/pyspark.zip -> hdfs://ip-172-31-27-34.ec2.internal:8020/user/hadoop/.sparkStaging/application_1583867709817_0003/pyspark.zip
20/03/10 19:50:51 INFO Client: Uploading resource file:/usr/lib/spark/python/lib/py4j-0.10.7-src.zip -> hdfs://ip-172-31-27-34.ec2.internal:8020/user/hadoop/.sparkStaging/application_1583867709817_0003/py4j-0.10.7-src.zip
20/03/10 19:50:52 INFO Client: Uploading resource file:/mnt/tmp/spark-4c4ea7ac-b2bb-4a61-929d-c371d87417ff/__spark_conf__476112427502500805.zip -> hdfs://ip-172-31-27-34.ec2.internal:8020/user/hadoop/.sparkStaging/application_1583867709817_0003/__spark_conf__.zip
20/03/10 19:50:52 INFO SecurityManager: Changing view acls to: hadoop
20/03/10 19:50:52 INFO SecurityManager: Changing modify acls to: hadoop
20/03/10 19:50:52 INFO SecurityManager: Changing view acls groups to:
20/03/10 19:50:52 INFO SecurityManager: Changing modify acls groups to:
20/03/10 19:50:52 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(hadoop); groups with view permissions: Set(); users with modify permissions: Set(hadoop); groups with modify permissions: Set()
20/03/10 19:50:53 INFO Client: Submitting application application_1583867709817_0003 to ResourceManager
20/03/10 19:50:53 INFO YarnClientImpl: Submitted application application_1583867709817_0003
20/03/10 19:50:54 INFO Client: Application report for application_1583867709817_0003 (state: ACCEPTED)
20/03/10 19:50:54 INFO Client:
client token: N/A
diagnostics: AM container is launched, waiting for AM container to Register with RM
ApplicationMaster host: N/A
ApplicationMaster RPC port: -1
queue: default
start time: 1583869853550
final status: UNDEFINED
tracking URL: http://ip-172-31-27-34.ec2.internal:20888/proxy/application_1583867709817_0003/
user: hadoop
20/03/10 19:50:55 INFO Client: Application report for application_1583867709817_0003 (state: ACCEPTED)
20/03/10 19:50:56 INFO Client: Application report for application_1583867709817_0003 (state: ACCEPTED)
20/03/10 19:50:57 INFO Client: Application report for application_1583867709817_0003 (state: ACCEPTED)
20/03/10 19:50:58 INFO Client: Application report for application_1583867709817_0003 (state: ACCEPTED)
20/03/10 19:50:59 INFO Client: Application report for application_1583867709817_0003 (state: ACCEPTED)
20/03/10 19:51:00 INFO Client: Application report for application_1583867709817_0003 (state: ACCEPTED)
20/03/10 19:51:01 INFO Client: Application report for application_1583867709817_0003 (state: ACCEPTED)
20/03/10 19:51:02 INFO Client: Application report for application_1583867709817_0003 (state: FAILED)
20/03/10 19:51:02 INFO Client:
client token: N/A
diagnostics: Application application_1583867709817_0003 failed 2 times due to AM Container for appattempt_1583867709817_0003_000002 exited with exitCode: 13
Failing this attempt.Diagnostics: Exception from container-launch.
Container id: container_1583867709817_0003_02_000001
Exit code: 13
Stack trace: ExitCodeException exitCode=13:
at org.apache.hadoop.util.Shell.runCommand(Shell.java:972)
at org.apache.hadoop.util.Shell.run(Shell.java:869)
at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:1170)
at org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.launchContainer(DefaultContainerExecutor.java:235)
at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:299)
at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:83)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
Container exited with a non-zero exit code 13
For more detailed output, check the application tracking page: http://ip-172-31-27-34.ec2.internal:8088/cluster/app/application_1583867709817_0003 Then click on links to logs of each attempt.
. Failing the application.
ApplicationMaster host: N/A
ApplicationMaster RPC port: -1
queue: default
start time: 1583869853550
final status: FAILED
tracking URL: http://ip-172-31-27-34.ec2.internal:8088/cluster/app/application_1583867709817_0003
user: hadoop
20/03/10 19:51:02 ERROR Client: Application diagnostics message: Application application_1583867709817_0003 failed 2 times due to AM Container for appattempt_1583867709817_0003_000002 exited with exitCode: 13
Failing this attempt.Diagnostics: Exception from container-launch.
Container id: container_1583867709817_0003_02_000001
Exit code: 13
Stack trace: ExitCodeException exitCode=13:
at org.apache.hadoop.util.Shell.runCommand(Shell.java:972)
at org.apache.hadoop.util.Shell.run(Shell.java:869)
at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:1170)
at org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.launchContainer(DefaultContainerExecutor.java:235)
at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:299)
at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:83)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
Container exited with a non-zero exit code 13
For more detailed output, check the application tracking page: http://ip-172-31-27-34.ec2.internal:8088/cluster/app/application_1583867709817_0003 Then click on links to logs of each attempt.
. Failing the application.
Exception in thread "main" org.apache.spark.SparkException: Application application_1583867709817_0003 finished with failed status
at org.apache.spark.deploy.yarn.Client.run(Client.scala:1149)
at org.apache.spark.deploy.yarn.YarnClusterApplication.start(Client.scala:1526)
at org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:853)
at org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:161)
at org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:184)
at org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:86)
at org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:928)
at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:937)
at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
20/03/10 19:51:02 INFO ShutdownHookManager: Shutdown hook called
20/03/10 19:51:02 INFO ShutdownHookManager: Deleting directory /mnt/tmp/spark-4c4ea7ac-b2bb-4a61-929d-c371d87417ff
20/03/10 19:51:02 INFO ShutdownHookManager: Deleting directory /mnt/tmp/spark-a72b6dba-91bb-46b0-b2c3-893ac3b8581f
Command exiting with ret '1'
When I dug through the container logs, I found the script spitting out an error as follows
boto3 module not found
I ran a bootstrap script that installs boto3 using pip.
I even logged in to the master node and found that boto3 was installed using the command pip list
You have to use a bootstrap script that installs boto3 but you have to be very specific about the version of python being used.
sudo pip-3.6 install boto3
On the console homepage, where you click on Create Cluster, a page appears. On the top, there is an option to "go to advanced options". There you will find the option to "Auto terminate" on "After last step completes"

WSO2 APIM Analytics server setup issue

In order to setup analytical server for API-Manager,i followed exactly same steps as specified in below wso2 documentation.
https://docs.wso2.com/display/AM220/Configuring+APIM+Analytics.
However i am facing below issues while running the API-M And API-Analytic instance.
WSO2 API-M:
[2018-05-08 02:52:05,378] ERROR - DataEndpointConnectionWorker Error while trying to connect to the endpoint. Cannot borrow client
for ssl://localhost:7712
org.wso2.carbon.databridge.agent.exception.DataEndpointAuthenticationException:
Cannot borrow client for ssl://localhost:7712
at org.wso2.carbon.databridge.agent.endpoint.DataEndpointConnectionWorker.connect(DataEndpointConnectionWorker.java:99)
at org.wso2.carbon.databridge.agent.endpoint.DataEndpointConnectionWorker.run(DataEndpointConnectionWorker.java:42)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334)
at java.util.concurrent.FutureTask.run(FutureTask.java:166)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:724) Caused by: org.wso2.carbon.databridge.agent.exception.DataEndpointSecurityException:
Error while trying to connect to ssl://localhost:7712
at org.wso2.carbon.databridge.agent.endpoint.thrift.ThriftSecureClientPoolFactory.createClient(ThriftSecureClientPoolFactory.java:81)
at org.wso2.carbon.databridge.agent.client.AbstractClientPoolFactory.makeObject(AbstractClientPoolFactory.java:39)
at org.apache.commons.pool.impl.GenericKeyedObjectPool.borrowObject(GenericKeyedObjectPool.java:1212)
at org.wso2.carbon.databridge.agent.endpoint.DataEndpointConnectionWorker.connect(DataEndpointConnectionWorker.java:91)
... 7 more Caused by: org.apache.thrift.transport.TTransportException: Could not connect to
localhost on port 7712
at org.apache.thrift.transport.TSSLTransportFactory.createClient(TSSLTransportFactory.java:237)
at org.apache.thrift.transport.TSSLTransportFactory.getClientSocket(TSSLTransportFactory.java:169)
at org.wso2.carbon.databridge.agent.endpoint.thrift.ThriftSecureClientPoolFactory.createClient(ThriftSecureClientPoolFactory.java:63)
... 10 more Caused by: java.net.ConnectException: Connection refused: connect
at java.net.DualStackPlainSocketImpl.connect0(Native Method)
at java.net.DualStackPlainSocketImpl.socketConnect(DualStackPlainSocketImpl.java:79)
at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:339)
at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:200)
at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:182)
at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:172)
at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:392)
at java.net.Socket.connect(Socket.java:579)
at sun.security.ssl.SSLSocketImpl.connect(SSLSocketImpl.java:618)
at sun.security.ssl.SSLSocketImpl.(SSLSocketImpl.java:407)
at sun.security.ssl.SSLSocketFactoryImpl.createSocket(SSLSocketFactoryImpl.java:88)
at org.apache.thrift.transport.TSSLTransportFactory.createClient(TSSLTransportFactory.java:233)
... 12 more [2018-05-08 02:52:34,371] WARN - DataEndpointGroup No receiver is reachable at reconnection, will try
to reconnect every 30 sec [2018-05-08 02:52:35,378] ERROR -
DataEndpointConnectionWorker Error while trying to connect to
ssl://localhost:7712
org.wso2.carbon.databridge.agent.exception.DataEndpointSecurityException:
Error while trying to connect to ssl://localhost:7712
at org.wso2.carbon.databridge.agent.endpoint.thrift.ThriftSecureClientPoolFactory.createClient(ThriftSecureClientPoolFactory.java:81)
at org.wso2.carbon.databridge.agent.client.AbstractClientPoolFactory.makeObject(AbstractClientPoolFactory.java:39)
at org.apache.commons.pool.impl.GenericKeyedObjectPool.borrowObject(GenericKeyedObjectPool.java:1212)
WSO2 API-Analytic:
JAVA_HOME environment variable is set to C:\Program Files\Java\jdk1.7.0_25 CARBON_HOME environment variable is set to
C:\WSO2AM~4\WSO2AM~1\bin.. Loading spark environment variables
[2018-05-08 02:36:02,932] INFO
{org.wso2.carbon.core.internal.CarbonCoreActivator} - Starting WSO2
Carbon... [2018-05-08 02:36:02,936] INFO
{org.wso2.carbon.core.internal.CarbonCoreActivator} - Operating
System : Windows Server 2008 R2 6.1, amd64 [2018-05-08 02:36:02,936]
INFO {org.wso2.carbon.core.internal.CarbonCoreActivator} - Java Home
: C:\Program Files\Java\jdk1.7.0_25\jre [2018-05-08 02:36:02,936]
INFO {org.wso2.carbon.core.internal.CarbonCoreActivator} - Java
Version : 1.7.0_25 [2018-05-08 02:36:02,937] INFO
{org.wso2.carbon.core.internal.CarbonCoreActivator} - Java VM
: Java HotSpot(TM) 64-Bit Server VM 23.25-b01,Oracle Corporation
[2018-05-08 02:36:02,937] INFO
{org.wso2.carbon.core.internal.CarbonCoreActivator} - Carbon Home
: C:\WSO2AM~4\WSO2AM~1\bin.. [2018-05-08 02:36:02,937] INFO
{org.wso2.carbon.core.internal.CarbonCoreActivator} - Java Temp Dir
: C:\WSO2AM~4\WSO2AM~1\bin..\tmp [2018-05-08 02:36:02,937] INFO
{org.wso2.carbon.core.internal.CarbonCoreActivator} - User
: wsoadm_dev_svc, en-US, America/Los_Angeles [2018-05-08 02:36:03,570]
INFO
{org.wso2.carbon.event.output.adapter.kafka.internal.ds.KafkaEventAdapterServiceDS}
- Successfully deployed the Kafka output event adapto [2018-05-08 02:36:03,887] INFO
{org.wso2.carbon.event.template.manager.core.internal.ds.TemplateDeployerServiceTrackerDS}
- Successfully deployed the execution manager [2018-05-08 02:36:15,048] INFO
{org.wso2.carbon.databridge.receiver.binary.internal.BinaryDataReceiver}
- Started Binary SSL Transport on port : 9712 [2018-05-08 02:36:15,050] INFO
{org.wso2.carbon.databridge.receiver.binary.internal.BinaryDataReceiver}
- Started Binary TCP Transport on port : 9612 [2018-05-08 02:36:15,207] INFO
{org.wso2.carbon.databridge.core.internal.DataBridgeDS} -
Successfully deployed Agent Server [2018-05-08 02:36:15,388] INFO
{org.wso2.carbon.registry.core.jdbc.EmbeddedRegistryService} -
Configured Registry in 64ms [2018-05-08 02:36:15,912] INFO
{org.wso2.carbon.registry.core.internal.RegistryCoreServiceComponent}
- Registry Mode : READ-WRITE [2018-05-08 02:36:22,854] INFO {org.wso2.carbon.metrics.impl.util.JmxReporterBuilder} - Creating JMX
reporter for Metrics with domain 'org.wso2.carbon.metrics' [2018-05-08
02:36:22,861] INFO
{org.wso2.carbon.metrics.impl.util.JDBCReporterBuilder} - Creating
JDBC reporter for Metrics with source 'SOADEVV001', data source 'jdbc/
[2018-05-08 02:36:22,862] INFO
{org.wso2.carbon.metrics.impl.reporter.AbstractReporter} - Started
JMX reporter for Metrics [2018-05-08 02:36:22,870] INFO
{org.wso2.carbon.metrics.impl.reporter.AbstractReporter} - Started
JDBC reporter for Metrics [2018-05-08 02:36:26,352] INFO
{org.wso2.carbon.registry.indexing.solr.SolrClient} - Default
Embedded Solr Server Initialized [2018-05-08 02:36:27,524] INFO
{org.wso2.carbon.user.core.internal.UserStoreMgtDSComponent} - Carbon
UserStoreMgtDSComponent activated successfully.
[2018-05-08 02:37:32,134] WARN {org.wso2.carbon.core.init.CarbonServerManager} - Carbon
initialization is delayed due to the following unsatisfied items:
[2018-05-08 02:37:32,498] WARN
{org.wso2.carbon.core.init.CarbonServerManager} - Waiting for
required OSGi Service: org.apache.axis2.engine.AxisObserver
[2018-05-08 02:38:32,135] WARN
{org.wso2.carbon.core.init.CarbonServerManager} - Carbon
initialization is delayed due to the following unsatisfied items:
Note:I am using same Machine for both API-M and API API-Analytic instance.
For APIM Offset is set to 0 and for API-Analytic it is 1.
I am using default H2 DB for for both.
Any help would be appreciated.

WSO2 API Manager 2.1 Analytics - Fails to Start up while connecting to Oracle DB

I am deploying WSO2 API Manager 2.1 and Analytics using the Pattern as specified - https://github.com/wso2/docker-apim/tree/master/docker-compose/pattern-3
Here in all the components - nginx, Publisher, Store, Traffic Manager, Gateway Worker, Gateway Manager, Key Manager and Analytics are deployed as separate docker containers.
When I started these containers, it worked fine and by default it was using the mysql server for storing all the data.
But as per our requirement, we had to use the Oracle DB and hence, we created a user in there with all the required permissions and then run the oracle scripts and finally started all the containers one by one.
In addition for Analytics, we created two separate users for two data sources - WSO2_ANALYTICS_EVENT_STORE_DB and WSO2_ANALYTICS_PROCESSED_DATA_STORE_DB (Didnt run any scripts for these).
And now we have a problem with the Analytics COntainer not able to start and throwing the error -
[2017-07-11 12:53:54,017] INFO {org.wso2.carbon.core.internal.CarbonCoreActivator} - Starting WSO2 Carbon...
[2017-07-11 12:53:54,017] INFO {org.wso2.carbon.core.internal.CarbonCoreActivator} - Operating System : Linux 4.8.0-53-generic, amd64
[2017-07-11 12:53:54,017] INFO {org.wso2.carbon.core.internal.CarbonCoreActivator} - Java Home : /mnt/jdk-7u80/jre
[2017-07-11 12:53:54,017] INFO {org.wso2.carbon.core.internal.CarbonCoreActivator} - Java Version : 1.7.0_80
[2017-07-11 12:53:54,017] INFO {org.wso2.carbon.core.internal.CarbonCoreActivator} - Java VM : Java HotSpot(TM) 64-Bit Server VM 24.80-b11,Oracle Corporation
[2017-07-11 12:53:54,018] INFO {org.wso2.carbon.core.internal.CarbonCoreActivator} - Carbon Home : /mnt/186.12.12.12/wso2am-analytics-2.1.0
[2017-07-11 12:53:54,018] INFO {org.wso2.carbon.core.internal.CarbonCoreActivator} - Java Temp Dir : /mnt/186.12.12.12/wso2am-analytics-2.1.0/tmp
[2017-07-11 12:53:54,018] INFO {org.wso2.carbon.core.internal.CarbonCoreActivator} - User : root, en-US, GMT
[2017-07-11 12:54:34,104] INFO {org.wso2.carbon.core.internal.permission.update.PermissionUpdater} - Permission cache updated for tenant -1234
[2017-07-11 12:54:34,235] INFO {org.wso2.carbon.core.transports.http.HttpsTransportListener} - HTTPS port : 9444
[2017-07-11 12:54:34,235] INFO {org.wso2.carbon.core.transports.http.HttpTransportListener} - HTTP port : 9764
[2017-07-11 12:54:36,547] INFO {org.wso2.carbon.webapp.mgt.TomcatGenericWebappsDeployer} - Deployed webapp: StandardEngine[Catalina].StandardHost[localhost].StandardContext[/analytics].File[/mnt/186.12.12.12/wso2am-analytics-2.1.0/repository/deployment/server/webapps/analytics.war]
[2017-07-11 12:54:36,593] INFO {org.wso2.carbon.webapp.mgt.TomcatGenericWebappsDeployer} - Deployed webapp: StandardEngine[Catalina].StandardHost[localhost].StandardContext[/inputwebsocket].File[/mnt/186.12.12.12/wso2am-analytics-2.1.0/repository/deployment/server/webapps/inputwebsocket.war]
[2017-07-11 12:54:36,622] INFO {org.wso2.carbon.webapp.mgt.TomcatGenericWebappsDeployer} - Deployed webapp: StandardEngine[Catalina].StandardHost[localhost].StandardContext[/outputwebsocket].File[/mnt/186.12.12.12/wso2am-analytics-2.1.0/repository/deployment/server/webapps/outputwebsocket.war]
[
[2017-07-11 12:54:43,116] INFO {org.wso2.carbon.event.processor.core.EventProcessorDeployer} - Execution plan deployment held back and in inactive state : APIMAnalytics-RequestSummarizer-RequestSummarizer-realtime1.siddhiql, Dependency validation exception: Stream org.wso2.apimgt.statistics.requestsPerMinPerKeyStream:1.0.0 does not exist
[2017-07-11 12:54:43,186] INFO {org.wso2.carbon.event.processor.core.EventProcessorDeployer} - Execution plan deployment held back and in inactive state : APIMAnalytics-UnusualIPAccessTemplate-UnusualIPAccessAlert-realtime1.siddhiql, Dependency validation exception: Stream org.wso2.apimgt.statistics.perMinuteRequest:1.0.0 does not exist
[2017-07-11 12:54:43,218] INFO {org.wso2.carbon.databridge.receiver.thrift.ThriftDataReceiver} - Thrift Server started at 0.0.0.0
[2017-07-11 12:54:43,246] INFO {org.wso2.carbon.databridge.receiver.thrift.ThriftDataReceiver} - Thrift SSL port : 7712
[2017-07-11 12:54:43,253] INFO {org.wso2.carbon.databridge.receiver.thrift.ThriftDataReceiver} - Thrift port : 7612
[2017-07-11 12:54:43,277] INFO {org.apache.tomcat.util.net.NioSelectorPool} - Using a shared selector for servlet write/read
[2017-07-11 12:54:43,355] INFO {org.apache.tomcat.util.net.NioSelectorPool} - Using a shared selector for servlet write/read
[2017-07-11 12:54:43,408] INFO {org.wso2.carbon.ntask.core.service.impl.TaskServiceImpl} - Task service starting in STANDALONE mode...
[2017-07-11 12:54:44,030] ERROR {org.wso2.carbon.analytics.dataservice.core.AnalyticsDataServiceComponent} - Error in activating analytics data service: null
java.lang.RuntimeException
at org.wso2.carbon.analytics.datasource.rdbms.RDBMSAnalyticsRecordStore$RDBMSResultSetIterator.next(RDBMSAnalyticsRecordStore.java:881)
at org.wso2.carbon.analytics.datasource.rdbms.RDBMSAnalyticsRecordStore$RDBMSResultSetIterator.hasNext(RDBMSAnalyticsRecordStore.java:843)
at org.apache.commons.collections.IteratorUtils.toList(IteratorUtils.java:848)
at org.apache.commons.collections.IteratorUtils.toList(IteratorUtils.java:825)
at org.wso2.carbon.analytics.datasource.core.util.GenericUtils.listRecords(GenericUtils.java:284)
[2017-07-11 12:54:55,566] INFO {org.wso2.carbon.databridge.core.DataBridge} - user admin connected
[2017-07-11 12:55:05,564] ERROR {org.wso2.carbon.databridge.core.internal.queue.QueueWorker} - Dropping wrongly formatted event sent for -1234
org.wso2.carbon.databridge.core.exception.EventConversionException: Error when converting loganalyzer:1.0.0 of event bundle with events 1
at org.wso2.carbon.databridge.receiver.thrift.converter.ThriftEventConverter.createEventList(ThriftEventConverter.java:181)
at org.wso2.carbon.databridge.receiver.thrift.converter.ThriftEventConverter.toEventList(ThriftEventConverter.java:90)
at org.wso2.carbon.databridge.core.internal.queue.QueueWorker.run(QueueWorker.java:73)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
at java.util.concurrent.FutureTask.run(FutureTask.java:262)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
Caused by: org.wso2.carbon.databridge.core.exception.EventConversionException: No StreamDefinition for streamId loganalyzer:1.0.0 present in cache
Can someone please let me know how to resolve this issue.
You can get the newer version of jar from http://maven.wso2.org/nexus/content/groups/public/org/wso2/carbon/analytics/org.wso2.carbon.analytics.datasource.rdbms/

I am using WSO2 ESB 5.0.0 with JDK 1.8.0_91 in windows 7 64-bit. I want to resolve the errors in the following situations

I am beginner to WSO2 ESB. Previously, I was running ESB server in command prompt using command:
C:/Program Files/wso2-5.0/bin> wso2server.bat --run
It responded successfully and server was running fine.
I stopped the server in command prompt and deployed the "WSO2QuickStartGuideCapp_1.0.0.CAR" file as per example sending a simple message using WSO2-ESB using Eclipse wso2esb-tooling-eclipse-jee-mars-win64-x86_64-5-1.0.0.
However, I am able to get the requests and response.
But, the Server thrown the following errors (Please Scroll Down for the errors) on the eclipse console:
[2016-10-25 10:59:50,591] INFO - DeploymentEngine Deploying Web service: org.wso2.carbon.sts-5.1.1 -
[2016-10-25 10:59:50,701] INFO - DeploymentEngine Deploying Web service: org.wso2.carbon.tryit-4.5.4 -
[2016-10-25 10:59:50,919] INFO - CarbonServerManager Repository : C:/Program Files/wso2esb-5.0.0/repository/deployment/server/
[2016-10-25 10:59:50,997] INFO - TenantLoadingConfig Using tenant lazy loading policy...
[2016-10-25 10:59:51,029] INFO - PermissionUpdater Permission cache updated for tenant -1234
[2016-10-25 10:59:51,091] INFO - RuleEngineConfigDS Successfully registered the Rule Config service
[2016-10-25 10:59:51,216] INFO - ServiceBusInitializer Starting ESB...
[2016-10-25 10:59:51,231] INFO - ServiceBusInitializer Initializing Apache Synapse...
[2016-10-25 10:59:51,247] INFO - SynapseControllerFactory Using Synapse home : C:\Program Files\wso2esb-5.0.0\.
[2016-10-25 10:59:51,247] INFO - SynapseControllerFactory Using synapse.xml location : C:\Program Files\wso2esb-5.0.0\.\.\repository\deployment\server\synapse-configs\default
[2016-10-25 10:59:51,247] INFO - SynapseControllerFactory Using server name : localhost
[2016-10-25 10:59:51,247] INFO - SynapseControllerFactory The timeout handler will run every : 15s
[2016-10-25 10:59:51,263] INFO - Axis2SynapseController Initializing Synapse at : Tue Oct 25 10:59:51 IST 2016
[2016-10-25 10:59:51,278] INFO - RuntimeStatisticCollector Statistics Entry Expiration time set to 136000 milliseconds
[2016-10-25 10:59:51,278] INFO - CarbonSynapseController Loading the mediation configuration from the file system
[2016-10-25 10:59:51,278] INFO - MultiXMLConfigurationBuilder Building synapse configuration from the synapse artifact repository at : .\.\repository/deployment/server/synapse-configs\default
[2016-10-25 10:59:51,372] INFO - XMLConfigurationBuilder Generating the Synapse configuration model by parsing the XML configuration
[2016-10-25 10:59:51,465] INFO - DependencyTracker Sequence : fault was added to the Synapse configuration successfully
[2016-10-25 10:59:51,481] INFO - DependencyTracker Sequence : main was added to the Synapse configuration successfully
[2016-10-25 10:59:51,481] INFO - SynapseConfigurationBuilder Loaded Synapse configuration from the artifact repository at : .\.\repository/deployment/server/synapse-configs\default
[2016-10-25 10:59:51,481] INFO - DependencyTracker Local entry : SERVER_HOST was added to the Synapse configuration successfully
[2016-10-25 10:59:51,481] INFO - DependencyTracker Local entry : SERVER_IP was added to the Synapse configuration successfully
[2016-10-25 10:59:51,481] INFO - Axis2SynapseController Loading mediator extensions...
[2016-10-25 10:59:51,512] INFO - DeploymentInterceptor Deploying Axis2 service: echo {super-tenant}
[2016-10-25 10:59:51,512] INFO - DeploymentEngine Deploying Web service: Echo.aar - file:/C:/Program Files/wso2esb-5.0.0/repository/deployment/server/axis2services/Echo.aar
[2016-10-25 10:59:51,543] INFO - DeploymentInterceptor Deploying Axis2 service: Version {super-tenant}
[2016-10-25 10:59:51,543] INFO - DeploymentEngine Deploying Web service: Version.aar - file:/C:/Program Files/wso2esb-5.0.0/repository/deployment/server/axis2services/Version.aar
[2016-10-25 10:59:51,575] INFO - EventPublisherDeployer Event Publisher deployment held back and in inactive state :MessageFlowConfigurationPublisher.xml, Stream validation exception : Stream org.wso2.esb.analytics.stream.ConfigEntry:1.0.0 does not exist
[2016-10-25 10:59:51,575] INFO - EventPublisherDeployer Event Publisher deployment held back and in inactive state :MessageFlowStatisticsPublisher.xml, Stream validation exception : Stream org.wso2.esb.analytics.stream.FlowEntry:1.0.0 does not exist
[2016-10-25 10:59:51,637] INFO - EventPublisherDeployer Event Publisher undeployed successfully : MessageFlowConfigurationPublisher.xml
[2016-10-25 10:59:52,137] INFO - EventJunction WSO2EventConsumer added to the junction. Stream:org.wso2.esb.analytics.stream.ConfigEntry:1.0.0
[2016-10-25 10:59:52,153] INFO - EventPublisherDeployer Event Publisher configuration successfully deployed and in active state : MessageFlowConfigurationPublisher
[2016-10-25 10:59:52,153] INFO - EventStreamDeployer Stream definition is deployed successfully : org.wso2.esb.analytics.stream.ConfigEntry:1.0.0
[2016-10-25 10:59:52,153] INFO - EventPublisherDeployer Event Publisher undeployed successfully : MessageFlowStatisticsPublisher.xml
[2016-10-25 10:59:52,168] INFO - EventJunction WSO2EventConsumer added to the junction. Stream:org.wso2.esb.analytics.stream.FlowEntry:1.0.0
[2016-10-25 10:59:52,168] INFO - EventPublisherDeployer Event Publisher configuration successfully deployed and in active state : MessageFlowStatisticsPublisher
[2016-10-25 10:59:52,168] INFO - EventStreamDeployer Stream definition is deployed successfully : org.wso2.esb.analytics.stream.FlowEntry:1.0.0
[2016-10-25 10:59:52,168] INFO - Axis2SynapseController Deploying the Synapse service...
[2016-10-25 10:59:52,168] INFO - Axis2SynapseController Deploying Proxy services...
[2016-10-25 10:59:52,168] INFO - Axis2SynapseController Deploying EventSources...
[2016-10-25 10:59:52,200] INFO - ServerManager Server ready for processing...
[2016-10-25 10:59:52,278] INFO - MediationStatisticsComponent JMX mediation statistic publishing enabled for tenant: -1234
[2016-10-25 10:59:52,278] INFO - MediationStatisticsComponent DAS mediation statistic publishing enabled for tenant: -1234
[2016-10-25 10:59:53,307] INFO - PassThroughHttpListener Starting Pass-through HTTP Listener...
[2016-10-25 10:59:53,307] INFO - PassThroughListeningIOReactorManager Pass-through HTTP Listener started on 0.0.0.0:8280
[2016-10-25 10:59:53,307] INFO - PassThroughHttpSSLListener Starting Pass-through HTTPS Listener...
[2016-10-25 10:59:53,307] INFO - PassThroughListeningIOReactorManager Pass-through HTTPS Listener started on 0.0.0.0:8243
[2016-10-25 10:59:53,323] INFO - ApplicationManager Deploying Carbon Application : SampleServicesCompositeApplication_1.0.0.car...
[2016-10-25 10:59:53,822] ERROR - DataEndpointConnectionWorker Error while trying to connect to ssl://localhost:7712
org.wso2.carbon.databridge.agent.exception.DataEndpointSecurityException: Error while trying to connect to ssl://localhost:7712
at org.wso2.carbon.databridge.agent.endpoint.thrift.ThriftSecureClientPoolFactory.createClient(ThriftSecureClientPoolFactory.java:61)
at org.wso2.carbon.databridge.agent.client.AbstractClientPoolFactory.makeObject(AbstractClientPoolFactory.java:39)
at org.apache.commons.pool.impl.GenericKeyedObjectPool.borrowObject(GenericKeyedObjectPool.java:1212)
at org.wso2.carbon.databridge.agent.endpoint.DataEndpointConnectionWorker.connect(DataEndpointConnectionWorker.java:91)
at org.wso2.carbon.databridge.agent.endpoint.DataEndpointConnectionWorker.run(DataEndpointConnectionWorker.java:42)
at java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source)
at java.util.concurrent.FutureTask.run(Unknown Source)
at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
at java.lang.Thread.run(Unknown Source)
Caused by: org.apache.thrift.transport.TTransportException: Could not connect to localhost on port 7712
at org.apache.thrift.transport.TSSLTransportFactory.createClient(TSSLTransportFactory.java:237)
at org.apache.thrift.transport.TSSLTransportFactory.getClientSocket(TSSLTransportFactory.java:169)
at org.wso2.carbon.databridge.agent.endpoint.thrift.ThriftSecureClientPoolFactory.createClient(ThriftSecureClientPoolFactory.java:56)
... 9 more
Caused by: java.net.ConnectException: Connection refused: connect
at java.net.TwoStacksPlainSocketImpl.socketConnect(Native Method)
at java.net.AbstractPlainSocketImpl.doConnect(Unknown Source)
at java.net.AbstractPlainSocketImpl.connectToAddress(Unknown Source)
at java.net.AbstractPlainSocketImpl.connect(Unknown Source)
at java.net.PlainSocketImpl.connect(Unknown Source)
at java.net.SocksSocketImpl.connect(Unknown Source)
at java.net.Socket.connect(Unknown Source)
at sun.security.ssl.SSLSocketImpl.connect(Unknown Source)
at sun.security.ssl.SSLSocketImpl.<init>(Unknown Source)
at sun.security.ssl.SSLSocketFactoryImpl.createSocket(Unknown Source)
at org.apache.thrift.transport.TSSLTransportFactory.createClient(TSSLTransportFactory.java:233)
... 11 more
[2016-10-25 10:59:53,826] ERROR - DataEndpointConnectionWorker Error while trying to connect to the endpoint. Cannot borrow client for ssl://localhost:7712
org.wso2.carbon.databridge.agent.exception.DataEndpointAuthenticationException: Cannot borrow client for ssl://localhost:7712
at org.wso2.carbon.databridge.agent.endpoint.DataEndpointConnectionWorker.connect(DataEndpointConnectionWorker.java:99)
at org.wso2.carbon.databridge.agent.endpoint.DataEndpointConnectionWorker.run(DataEndpointConnectionWorker.java:42)
at java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source)
at java.util.concurrent.FutureTask.run(Unknown Source)
at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
at java.lang.Thread.run(Unknown Source)
Caused by: org.wso2.carbon.databridge.agent.exception.DataEndpointSecurityException: Error while trying to connect to ssl://localhost:7712
at org.wso2.carbon.databridge.agent.endpoint.thrift.ThriftSecureClientPoolFactory.createClient(ThriftSecureClientPoolFactory.java:61)
at org.wso2.carbon.databridge.agent.client.AbstractClientPoolFactory.makeObject(AbstractClientPoolFactory.java:39)
at org.apache.commons.pool.impl.GenericKeyedObjectPool.borrowObject(GenericKeyedObjectPool.java:1212)
at org.wso2.carbon.databridge.agent.endpoint.DataEndpointConnectionWorker.connect(DataEndpointConnectionWorker.java:91)
... 6 more
Caused by: org.apache.thrift.transport.TTransportException: Could not connect to localhost on port 7712
at org.apache.thrift.transport.TSSLTransportFactory.createClient(TSSLTransportFactory.java:237)
at org.apache.thrift.transport.TSSLTransportFactory.getClientSocket(TSSLTransportFactory.java:169)
at org.wso2.carbon.databridge.agent.endpoint.thrift.ThriftSecureClientPoolFactory.createClient(ThriftSecureClientPoolFactory.java:56)
... 9 more
Caused by: java.net.ConnectException: Connection refused: connect
at java.net.TwoStacksPlainSocketImpl.socketConnect(Native Method)
at java.net.AbstractPlainSocketImpl.doConnect(Unknown Source)
at java.net.AbstractPlainSocketImpl.connectToAddress(Unknown Source)
at java.net.AbstractPlainSocketImpl.connect(Unknown Source)
at java.net.PlainSocketImpl.connect(Unknown Source)
at java.net.SocksSocketImpl.connect(Unknown Source)
at java.net.Socket.connect(Unknown Source)
at sun.security.ssl.SSLSocketImpl.connect(Unknown Source)
at sun.security.ssl.SSLSocketImpl.<init>(Unknown Source)
at sun.security.ssl.SSLSocketFactoryImpl.createSocket(Unknown Source)
at org.apache.thrift.transport.TSSLTransportFactory.createClient(TSSLTransportFactory.java:233)
... 11 more
[2016-10-25 10:59:54,949] INFO - API Initializing API: HealthcareAPI
[2016-10-25 10:59:54,952] INFO - DependencyTracker API : HealthcareAPI was added to the Synapse configuration successfully - [ Deployed From Artifact Container: SampleServicesCompositeApplication ]
[2016-10-25 10:59:54,953] INFO - APIDeployer API named 'HealthcareAPI' has been deployed from file : C:\Program Files\wso2esb-5.0.0\tmp\carbonapps\-1234\1477373393323SampleServicesCompositeApplication_1.0.0.car\HealthcareAPI_1.0.0\HealthcareAPI-1.0.0.xml
[2016-10-25 10:59:54,964] INFO - DependencyTracker Endpoint : QueryDoctorEP was added to the Synapse configuration successfully - [ Deployed From Artifact Container: SampleServicesCompositeApplication ]
[2016-10-25 10:59:54,965] INFO - EndpointDeployer Endpoint named 'QueryDoctorEP' has been deployed from file : C:\Program Files\wso2esb-5.0.0\tmp\carbonapps\-1234\1477373393323SampleServicesCompositeApplication_1.0.0.car\QueryDoctorEP_1.0.0\QueryDoctorEP-1.0.0.xml
[2016-10-25 10:59:54,965] INFO - ApplicationManager Successfully Deployed Carbon Application : SampleServicesCompositeApplication_1.0.0 {super-tenant}
[2016-10-25 10:59:54,966] INFO - ApplicationManager Deploying Carbon Application : WSO2QuickStartGuideCapp_1.0.0.car...
[2016-10-25 10:59:55,081] INFO - API Initializing API: HospitalServiceApi
[2016-10-25 10:59:55,081] INFO - DependencyTracker API : HospitalServiceApi was added to the Synapse configuration successfully - [ Deployed From Artifact Container: WSO2QuickStartGuideCapp ]
[2016-10-25 10:59:55,081] INFO - APIDeployer API named 'HospitalServiceApi' has been deployed from file : C:\Program Files\wso2esb-5.0.0\tmp\carbonapps\-1234\1477373394966WSO2QuickStartGuideCapp_1.0.0.car\HospitalServiceApi_1.0.0\HospitalServiceApi-1.0.0.xml
[2016-10-25 10:59:55,083] INFO - DependencyTracker Endpoint : discountEligibility was added to the Synapse configuration successfully - [ Deployed From Artifact Container: WSO2QuickStartGuideCapp ]
[2016-10-25 10:59:55,083] INFO - EndpointDeployer Endpoint named 'discountEligibility' has been deployed from file : C:\Program Files\wso2esb-5.0.0\tmp\carbonapps\-1234\1477373394966WSO2QuickStartGuideCapp_1.0.0.car\discountEligibility_1.0.0\discountEligibility-1.0.0.xml
[2016-10-25 10:59:55,084] INFO - DependencyTracker Endpoint : getPaymentDetails was added to the Synapse configuration successfully - [ Deployed From Artifact Container: WSO2QuickStartGuideCapp ]
[2016-10-25 10:59:55,085] INFO - EndpointDeployer Endpoint named 'getPaymentDetails' has been deployed from file : C:\Program Files\wso2esb-5.0.0\tmp\carbonapps\-1234\1477373394966WSO2QuickStartGuideCapp_1.0.0.car\getPaymentDetails_1.0.0\getPaymentDetails-1.0.0.xml
[2016-10-25 10:59:55,087] INFO - DependencyTracker Endpoint : channellingFee was added to the Synapse configuration successfully - [ Deployed From Artifact Container: WSO2QuickStartGuideCapp ]
[2016-10-25 10:59:55,087] INFO - EndpointDeployer Endpoint named 'channellingFee' has been deployed from file : C:\Program Files\wso2esb-5.0.0\tmp\carbonapps\-1234\1477373394966WSO2QuickStartGuideCapp_1.0.0.car\channellingFee_1.0.0\channellingFee-1.0.0.xml
[2016-10-25 10:59:55,372] INFO - DependencyTracker Sequence : APIInSequence was added to the Synapse configuration successfully - [ Deployed From Artifact Container: WSO2QuickStartGuideCapp ]
[2016-10-25 10:59:55,373] INFO - SequenceDeployer Sequence named 'APIInSequence' has been deployed from file : C:\Program Files\wso2esb-5.0.0\tmp\carbonapps\-1234\1477373394966WSO2QuickStartGuideCapp_1.0.0.car\APIInSequence_1.0.0\APIInSequence-1.0.0.xml
[2016-10-25 10:59:55,378] INFO - DependencyTracker Sequence : faultSequence was added to the Synapse configuration successfully - [ Deployed From Artifact Container: WSO2QuickStartGuideCapp ]
[2016-10-25 10:59:55,378] INFO - SequenceDeployer Sequence named 'faultSequence' has been deployed from file : C:\Program Files\wso2esb-5.0.0\tmp\carbonapps\-1234\1477373394966WSO2QuickStartGuideCapp_1.0.0.car\faultSequence_1.0.0\faultSequence-1.0.0.xml
[2016-10-25 10:59:55,378] INFO - DependencyTracker Endpoint : reserveAppointmentEp was added to the Synapse configuration successfully - [ Deployed From Artifact Container: WSO2QuickStartGuideCapp ]
[2016-10-25 10:59:55,378] INFO - EndpointDeployer Endpoint named 'reserveAppointmentEp' has been deployed from file : C:\Program Files\wso2esb-5.0.0\tmp\carbonapps\-1234\1477373394966WSO2QuickStartGuideCapp_1.0.0.car\reserveAppointmentEp_1.0.0\reserveAppointmentEp-1.0.0.xml
[2016-10-25 10:59:55,378] INFO - DependencyTracker Endpoint : settlePayment was added to the Synapse configuration successfully - [ Deployed From Artifact Container: WSO2QuickStartGuideCapp ]
[2016-10-25 10:59:55,378] INFO - EndpointDeployer Endpoint named 'settlePayment' has been deployed from file : C:\Program Files\wso2esb-5.0.0\tmp\carbonapps\-1234\1477373394966WSO2QuickStartGuideCapp_1.0.0.car\settlePayment_1.0.0\settlePayment-1.0.0.xml
[2016-10-25 10:59:55,378] INFO - ApplicationManager Successfully Deployed Carbon Application : WSO2QuickStartGuideCapp_1.0.0 {super-tenant}
[2016-10-25 10:59:55,378] INFO - NioSelectorPool Using a shared selector for servlet write/read
[2016-10-25 10:59:55,448] INFO - NioSelectorPool Using a shared selector for servlet write/read
[2016-10-25 10:59:55,561] INFO - TaskServiceImpl Task service starting in STANDALONE mode...
[2016-10-25 10:59:55,590] INFO - NTaskTaskManager Initialized task manager. Tenant [-1234]
[2016-10-25 10:59:55,665] INFO - JMXServerManager JMX Service URL : service:jmx:rmi://localhost:11111/jndi/rmi://localhost:9999/jmxrmi
[2016-10-25 10:59:55,665] INFO - StartupFinalizerServiceComponent Server : WSO2 Enterprise Service Bus-5.0.0
[2016-10-25 10:59:55,665] INFO - StartupFinalizerServiceComponent WSO2 Carbon started in 53 sec
[2016-10-25 10:59:55,995] INFO - CarbonUIServiceComponent Mgt Console URL : https://192.168.100.196:9443/carbon/
[2016-10-25 11:00:22,726] WARN - DataEndpointGroup No receiver is reachable at reconnection, will try to reconnect every 30 sec
[2016-10-25 11:00:23,728] ERROR - DataEndpointConnectionWorker Error while trying to connect to ssl://localhost:7712
org.wso2.carbon.databridge.agent.exception.DataEndpointSecurityException: Error while trying to connect to ssl://localhost:7712
at org.wso2.carbon.databridge.agent.endpoint.thrift.ThriftSecureClientPoolFactory.createClient(ThriftSecureClientPoolFactory.java:61)
at org.wso2.carbon.databridge.agent.client.AbstractClientPoolFactory.makeObject(AbstractClientPoolFactory.java:39)
at org.apache.commons.pool.impl.GenericKeyedObjectPool.borrowObject(GenericKeyedObjectPool.java:1212)
at org.wso2.carbon.databridge.agent.endpoint.DataEndpointConnectionWorker.connect(DataEndpointConnectionWorker.java:91)
at org.wso2.carbon.databridge.agent.endpoint.DataEndpointConnectionWorker.run(DataEndpointConnectionWorker.java:42)
at java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source)
at java.util.concurrent.FutureTask.run(Unknown Source)
at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
at java.lang.Thread.run(Unknown Source)
Caused by: org.apache.thrift.transport.TTransportException: Could not connect to localhost on port 7712
at org.apache.thrift.transport.TSSLTransportFactory.createClient(TSSLTransportFactory.java:237)
at org.apache.thrift.transport.TSSLTransportFactory.getClientSocket(TSSLTransportFactory.java:169)
at org.wso2.carbon.databridge.agent.endpoint.thrift.ThriftSecureClientPoolFactory.createClient(ThriftSecureClientPoolFactory.java:56)
... 9 more
Caused by: java.net.ConnectException: Connection refused: connect
at java.net.TwoStacksPlainSocketImpl.socketConnect(Native Method)
at java.net.AbstractPlainSocketImpl.doConnect(Unknown Source)
at java.net.AbstractPlainSocketImpl.connectToAddress(Unknown Source)
at java.net.AbstractPlainSocketImpl.connect(Unknown Source)
at java.net.PlainSocketImpl.connect(Unknown Source)
at java.net.SocksSocketImpl.connect(Unknown Source)
at java.net.Socket.connect(Unknown Source)
at sun.security.ssl.SSLSocketImpl.connect(Unknown Source)
at sun.security.ssl.SSLSocketImpl.<init>(Unknown Source)
at sun.security.ssl.SSLSocketFactoryImpl.createSocket(Unknown Source)
at org.apache.thrift.transport.TSSLTransportFactory.createClient(TSSLTransportFactory.java:233)
... 11 more
[2016-10-25 11:00:23,732] ERROR - DataEndpointConnectionWorker Error while trying to connect to the endpoint. Cannot borrow client for ssl://localhost:7712
org.wso2.carbon.databridge.agent.exception.DataEndpointAuthenticationException: Cannot borrow client for ssl://localhost:7712
at org.wso2.carbon.databridge.agent.endpoint.DataEndpointConnectionWorker.connect(DataEndpointConnectionWorker.java:99)
at org.wso2.carbon.databridge.agent.endpoint.DataEndpointConnectionWorker.run(DataEndpointConnectionWorker.java:42)
at java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source)
at java.util.concurrent.FutureTask.run(Unknown Source)
at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
at java.lang.Thread.run(Unknown Source)
Caused by: org.wso2.carbon.databridge.agent.exception.DataEndpointSecurityException: Error while trying to connect to ssl://localhost:7712
at org.wso2.carbon.databridge.agent.endpoint.thrift.ThriftSecureClientPoolFactory.createClient(ThriftSecureClientPoolFactory.java:61)
at org.wso2.carbon.databridge.agent.client.AbstractClientPoolFactory.makeObject(AbstractClientPoolFactory.java:39)
at org.apache.commons.pool.impl.GenericKeyedObjectPool.borrowObject(GenericKeyedObjectPool.java:1212)
at org.wso2.carbon.databridge.agent.endpoint.DataEndpointConnectionWorker.connect(DataEndpointConnectionWorker.java:91)
... 6 more
Caused by: org.apache.thrift.transport.TTransportException: Could not connect to localhost on port 7712
at org.apache.thrift.transport.TSSLTransportFactory.createClient(TSSLTransportFactory.java:237)
at org.apache.thrift.transport.TSSLTransportFactory.getClientSocket(TSSLTransportFactory.java:169)
at org.wso2.carbon.databridge.agent.endpoint.thrift.ThriftSecureClientPoolFactory.createClient(ThriftSecureClientPoolFactory.java:56)
... 9 more
Caused by: java.net.ConnectException: Connection refused: connect
at java.net.TwoStacksPlainSocketImpl.socketConnect(Native Method)
at java.net.AbstractPlainSocketImpl.doConnect(Unknown Source)
at java.net.AbstractPlainSocketImpl.connectToAddress(Unknown Source)
at java.net.AbstractPlainSocketImpl.connect(Unknown Source)
at java.net.PlainSocketImpl.connect(Unknown Source)
at java.net.SocksSocketImpl.connect(Unknown Source)
at java.net.Socket.connect(Unknown Source)
at sun.security.ssl.SSLSocketImpl.connect(Unknown Source)
at sun.security.ssl.SSLSocketImpl.<init>(Unknown Source)
at sun.security.ssl.SSLSocketFactoryImpl.createSocket(Unknown Source)
at org.apache.thrift.transport.TSSLTransportFactory.createClient(TSSLTransportFactory.java:233)
... 11 more
And, Now while re-running the server on the command prompt also throws the same errors as above. But, yet I am able to login in the carbon management console using:
Username : admin
Password : admin
Please help me resolving the errors specified.
[2016-10-25 10:59:53,822] ERROR - DataEndpointConnectionWorker Error while trying to connect to ssl://localhost:7712
org.wso2.carbon.databridge.agent.exception.DataEndpointSecurityException: Error while trying to connect to ssl://localhost:7712
at org.wso2.carbon.databridge.agent.endpoint.thrift.ThriftSecureClientPoolFactory.createClient(ThriftSecureClientPoolFactory.java:61)
at org.wso2.carbon.databridge.agent.client.AbstractClientPoolFactory.makeObject(AbstractClientPoolFactory.java:39)
at org.apache.commons.pool.impl.GenericKeyedObjectPool.borrowObject(GenericKeyedObjectPool.java:1212)
at org.wso2.carbon.databridge.agent.endpoint.DataEndpointConnectionWorker.connect(DataEndpointConnectionWorker.java:91)
at org.wso2.carbon.databridge.agent.endpoint.DataEndpointConnectionWorker.run(DataEndpointConnectionWorker.java:42)
at java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source)
at java.util.concurrent.FutureTask.run(Unknown Source)
at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
at java.lang.Thread.run(Unknown Source)
Caused by: org.apache.thrift.transport.TTransportException: Could not connect to localhost on port 7712
at org.apache.thrift.transport.TSSLTransportFactory.createClient(TSSLTransportFactory.java:237)
at org.apache.thrift.transport.TSSLTransportFactory.getClientSocket(TSSLTransportFactory.java:169)
at org.wso2.carbon.databridge.agent.endpoint.thrift.ThriftSecureClientPoolFactory.createClient(ThriftSecureClientPoolFactory.java:56)
... 9 more
Caused by: java.net.ConnectException: Connection refused: connect
at java.net.TwoStacksPlainSocketImpl.socketConnect(Native Method)
at java.net.AbstractPlainSocketImpl.doConnect(Unknown Source)
at java.net.AbstractPlainSocketImpl.connectToAddress(Unknown Source)
at java.net.AbstractPlainSocketImpl.connect(Unknown Source)
at java.net.PlainSocketImpl.connect(Unknown Source)
at java.net.SocksSocketImpl.connect(Unknown Source)
at java.net.Socket.connect(Unknown Source)
at sun.security.ssl.SSLSocketImpl.connect(Unknown Source)
at sun.security.ssl.SSLSocketImpl.<init>(Unknown Source)
at sun.security.ssl.SSLSocketFactoryImpl.createSocket(Unknown Source)
at org.apache.thrift.transport.TSSLTransportFactory.createClient(TSSLTransportFactory.java:233)
... 11 more
[2016-10-25 10:59:53,826] ERROR - DataEndpointConnectionWorker Error while trying to connect to the endpoint. Cannot borrow client for ssl://localhost:7712
org.wso2.carbon.databridge.agent.exception.DataEndpointAuthenticationException: Cannot borrow client for ssl://localhost:7712
at org.wso2.carbon.databridge.agent.endpoint.DataEndpointConnectionWorker.connect(DataEndpointConnectionWorker.java:99)
at org.wso2.carbon.databridge.agent.endpoint.DataEndpointConnectionWorker.run(DataEndpointConnectionWorker.java:42)
at java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source)
at java.util.concurrent.FutureTask.run(Unknown Source)
at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
at java.lang.Thread.run(Unknown Source)
Caused by: org.wso2.carbon.databridge.agent.exception.DataEndpointSecurityException: Error while trying to connect to ssl://localhost:7712
at org.wso2.carbon.databridge.agent.endpoint.thrift.ThriftSecureClientPoolFactory.createClient(ThriftSecureClientPoolFactory.java:61)
at org.wso2.carbon.databridge.agent.client.AbstractClientPoolFactory.makeObject(AbstractClientPoolFactory.java:39)
at org.apache.commons.pool.impl.GenericKeyedObjectPool.borrowObject(GenericKeyedObjectPool.java:1212)
at org.wso2.carbon.databridge.agent.endpoint.DataEndpointConnectionWorker.connect(DataEndpointConnectionWorker.java:91)
... 6 more
Caused by: org.apache.thrift.transport.TTransportException: Could not connect to localhost on port 7712
at org.apache.thrift.transport.TSSLTransportFactory.createClient(TSSLTransportFactory.java:237)
at org.apache.thrift.transport.TSSLTransportFactory.getClientSocket(TSSLTransportFactory.java:169)
at org.wso2.carbon.databridge.agent.endpoint.thrift.ThriftSecureClientPoolFactory.createClient(ThriftSecureClientPoolFactory.java:56)
... 9 more
Caused by: java.net.ConnectException: Connection refused: connect
at java.net.TwoStacksPlainSocketImpl.socketConnect(Native Method)
at java.net.AbstractPlainSocketImpl.doConnect(Unknown Source)
at java.net.AbstractPlainSocketImpl.connectToAddress(Unknown Source)
at java.net.AbstractPlainSocketImpl.connect(Unknown Source)
at java.net.PlainSocketImpl.connect(Unknown Source)
at java.net.SocksSocketImpl.connect(Unknown Source)
at java.net.Socket.connect(Unknown Source)
at sun.security.ssl.SSLSocketImpl.connect(Unknown Source)
at sun.security.ssl.SSLSocketImpl.<init>(Unknown Source)
at sun.security.ssl.SSLSocketFactoryImpl.createSocket(Unknown Source)
at org.apache.thrift.transport.TSSLTransportFactory.createClient(TSSLTransportFactory.java:233)
... 11 more
Thanks in Advance
I encountered the same error, and after some investigation I found out that ESB carbon was trying to connect to ssl port 7712 based on default Thrift receiverURL, but it would crash because I had no Thrift server started.
I found that out when I lauched esb analytics to investigate the crash, and it started the Thrift server,which would listen to the same port.
After that, the crash was gone.

WSO2 ESB configuration with SAP adapter to send IDoc

I am trying to configure SAP adapter with WSO2 ESB. I have followed the instructions mentioned on WSO2 website.
https://docs.wso2.com/display/ESB490/SAP+Integration
But when I start the WSO2 ESB, it is giving Exception:
waqasalirazaq#wi3-Inspiron-3521:~/Documents/WSO2/wso2esb-4.8.1/bin$ ./wso2server.sh
JAVA_HOME environment variable is set to /home/waqasalirazaq/jdk1.7.0_51
CARBON_HOME environment variable is set to /home/waqasalirazaq/Documents/WSO2/wso2esb-4.8.1
[2014-11-03 16:49:39,599] INFO - CarbonCoreActivator Starting WSO2 Carbon...
[2014-11-03 16:49:39,603] INFO - CarbonCoreActivator Operating System : Linux 3.2.0-68-generic, amd64
[2014-11-03 16:49:39,604] INFO - CarbonCoreActivator Java Home : /home/waqasalirazaq/jdk1.7.0_51/jre
[2014-11-03 16:49:39,604] INFO - CarbonCoreActivator Java Version : 1.7.0_51
[2014-11-03 16:49:39,604] INFO - CarbonCoreActivator Java VM : Java HotSpot(TM) 64-Bit Server VM 24.51-b03,Oracle Corporation
[2014-11-03 16:49:39,604] INFO - CarbonCoreActivator Carbon Home : /home/waqasalirazaq/Documents/WSO2/wso2esb-4.8.1
[2014-11-03 16:49:39,604] INFO - CarbonCoreActivator Java Temp Dir : /home/waqasalirazaq/Documents/WSO2/wso2esb-4.8.1/tmp
[2014-11-03 16:49:39,605] INFO - CarbonCoreActivator User : waqasalirazaq, en-US, Asia/Karachi
[2014-11-03 16:49:39,702] WARN - ValidationResultPrinter The default keystore (wso2carbon.jks) is currently being used. To maximize security when deploying to a production environment, configure a new keystore with a unique password in the production server profile.
[2014-11-03 16:49:39,714] INFO - AgentHolder Agent created !
[2014-11-03 16:49:39,745] INFO - AgentDS Successfully deployed Agent Client
[2014-11-03 16:49:44,606] INFO - EmbeddedRegistryService Configured Registry in 116ms
[2014-11-03 16:49:44,786] INFO - RegistryCoreServiceComponent Registry Mode : READ-WRITE
[2014-11-03 16:49:45,440] INFO - UserStoreMgtDSComponent Carbon UserStoreMgtDSComponent activated successfully.
[2014-11-03 16:49:52,275] INFO - TaglibUriRule TLD skipped. URI: http://tiles.apache.org/tags-tiles is already defined
[2014-11-03 16:49:53,700] INFO - ClusterBuilder Clustering has been disabled
[2014-11-03 16:49:54,223] INFO - LandingPageWebappDeployer Deployed product landing page webapp: StandardEngine[Catalina].StandardHost[localhost].StandardContext[/home]
[2014-11-03 16:49:54,224] INFO - UserStoreConfigurationDeployer User Store Configuration Deployer initiated.
[2014-11-03 16:49:54,310] INFO - PassThroughHttpSSLSender Initializing Pass-through HTTP/S Sender...
[2014-11-03 16:49:54,343] INFO - ClientConnFactoryBuilder HTTPS Loading Identity Keystore from : repository/resources/security/wso2carbon.jks
[2014-11-03 16:49:54,352] INFO - ClientConnFactoryBuilder HTTPS Loading Trust Keystore from : repository/resources/security/client-truststore.jks
[2014-11-03 16:49:54,406] INFO - PassThroughHttpSSLSender Pass-through HTTPS Sender started...
[2014-11-03 16:49:54,407] INFO - PassThroughHttpSender Initializing Pass-through HTTP/S Sender...
[2014-11-03 16:49:54,412] INFO - PassThroughHttpSender Pass-through HTTP Sender started...
[2014-11-03 16:49:54,416] INFO - JMSSender JMS Sender started
[2014-11-03 16:49:54,418] INFO - JMSSender JMS Transport Sender initialized...
[2014-11-03 16:49:54,419] INFO - SAPTransportSender BAPI Sender started
[2014-11-03 16:49:54,422] FATAL - CarbonServerManager WSO2 Carbon initialization Failed
java.lang.NoSuchMethodError: com.sap.conn.jco.ext.Environment.isServerDataProviderRegistered()Z
at org.wso2.carbon.transports.sap.SAPTransportSender.init(SAPTransportSender.java:84)
at org.apache.axis2.context.ConfigurationContextFactory.initTransportSenders(ConfigurationContextFactory.java:300)
at org.apache.axis2.context.ConfigurationContextFactory.init(ConfigurationContextFactory.java:231)
at org.apache.axis2.context.ConfigurationContextFactory.createConfigurationContext(ConfigurationContextFactory.java:93)
at org.wso2.carbon.core.CarbonConfigurationContextFactory.createNewConfigurationContext(CarbonConfigurationContextFactory.java:65)
at org.wso2.carbon.core.init.CarbonServerManager.initializeCarbon(CarbonServerManager.java:398)
at org.wso2.carbon.core.init.CarbonServerManager.start(CarbonServerManager.java:219)
at org.wso2.carbon.core.internal.CarbonCoreServiceComponent.activate(CarbonCoreServiceComponent.java:77)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at org.eclipse.equinox.internal.ds.model.ServiceComponent.activate(ServiceComponent.java:260)
at org.eclipse.equinox.internal.ds.model.ServiceComponentProp.activate(ServiceComponentProp.java:146)
at org.eclipse.equinox.internal.ds.model.ServiceComponentProp.build(ServiceComponentProp.java:347)
at org.eclipse.equinox.internal.ds.InstanceProcess.buildComponent(InstanceProcess.java:620)
at org.eclipse.equinox.internal.ds.InstanceProcess.buildComponents(InstanceProcess.java:197)
at org.eclipse.equinox.internal.ds.Resolver.getEligible(Resolver.java:343)
at org.eclipse.equinox.internal.ds.SCRManager.serviceChanged(SCRManager.java:222)
at org.eclipse.osgi.internal.serviceregistry.FilteredServiceListener.serviceChanged(FilteredServiceListener.java:107)
at org.eclipse.osgi.framework.internal.core.BundleContextImpl.dispatchEvent(BundleContextImpl.java:861)
at org.eclipse.osgi.framework.eventmgr.EventManager.dispatchEvent(EventManager.java:230)
at org.eclipse.osgi.framework.eventmgr.ListenerQueue.dispatchEventSynchronous(ListenerQueue.java:148)
at org.eclipse.osgi.internal.serviceregistry.ServiceRegistry.publishServiceEventPrivileged(ServiceRegistry.java:819)
at org.eclipse.osgi.internal.serviceregistry.ServiceRegistry.publishServiceEvent(ServiceRegistry.java:771)
at org.eclipse.osgi.internal.serviceregistry.ServiceRegistrationImpl.register(ServiceRegistrationImpl.java:130)
at org.eclipse.osgi.internal.serviceregistry.ServiceRegistry.registerService(ServiceRegistry.java:214)
at org.eclipse.osgi.framework.internal.core.BundleContextImpl.registerService(BundleContextImpl.java:433)
at org.eclipse.equinox.http.servlet.internal.Activator.registerHttpService(Activator.java:81)
at org.eclipse.equinox.http.servlet.internal.Activator.addProxyServlet(Activator.java:60)
at org.eclipse.equinox.http.servlet.internal.ProxyServlet.init(ProxyServlet.java:40)
at org.wso2.carbon.tomcat.ext.servlet.DelegationServlet.init(DelegationServlet.java:38)
at org.apache.catalina.core.StandardWrapper.initServlet(StandardWrapper.java:1267)
at org.apache.catalina.core.StandardWrapper.loadServlet(StandardWrapper.java:1186)
at org.apache.catalina.core.StandardWrapper.load(StandardWrapper.java:1081)
at org.apache.catalina.core.StandardContext.loadOnStartup(StandardContext.java:5027)
at org.apache.catalina.core.StandardContext.startInternal(StandardContext.java:5314)
at org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:150)
at org.apache.catalina.core.ContainerBase$StartChild.call(ContainerBase.java:1559)
at org.apache.catalina.core.ContainerBase$StartChild.call(ContainerBase.java:1549)
at java.util.concurrent.FutureTask.run(FutureTask.java:262)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:744)
I have resolve this issue by placing the Jars (sapjco3.jar & sapidoc3.jar) in /repository/component/lib and JCO native library (sapjco3.so) in /jre/lib/amd64