If there are two versions of dpdk in the system.To make the primary and secondary dpdkproc info application to use the same dpdk version I have set LD_LIBRARY_PATH to location of libdpdk.so and have run the dpdk sample application and dpdk-proc info of same version. Still Version mismatch error occurs.Please suggest some solution for this.
Based on the current update via comments, using the right DPDK version and tools (meson/makefile) the problem is resolved. This is similar to issue faced by #AshwinKumar and solution suggested DPDK proc-info.
Note: always use the same dpdk version for building primary and secodnary and pass same device arguments too.
Related
Good evening.
I'm setting up my game project on a remote computer and am running into a problem with SDL. I'm using SDL 2.0 and I'm trying to figure out why SDL_GetNumVideoDisplays() returns 1 when I have 2 monitors hooked up to my PC. I'm running Ubuntu 14.04 LTS, and both monitors appear to be configured correctly just using the desktop as normal.
Would appreciate tips on where to look to troubleshoot this issue. Thanks.
Update 1 - ./configure for SDL2-2.0.4 source code:
Was looking for some mention of X11 in the configure script and saw the following:
checking for X11/extensions/Xext.h... yes
-- dynamic libX11 -> libX11.so.6
-- dynamic libX11ext -> libXext.so.6
checking for const parameter to XextAddDisplay... yes
checking for XGenericEvent... yes
checking for XkbKeycodeToKeysym in -lX11... yes
checking for X11/Xcursor/Xcursor.h... no
checking for X11/extensions/Xdbe.h... yes
checking for X11/extensions/Xinerama.h... no
checking for X11/extensions/XInput2.h... no
checking for X11/extensions/scrnsaver.h... no
checking for X11/extensions/shape.h... yes
checking for X11/extensions/xf86vmode.h... yes
and
X11 libraries : xdbe xshape xvidmode
My xrandr --version is:
xrandr program version 1.4.1
Server reports RandR version 1.4
xrandr --listactivemonitors doesn't seem to be supported on my install.
Update 2 - Reinstalled OS, completely wiping partition:
Things escalated quickly today when I decided just to take the nuclear option. Fortunately there was nothing on the disk I needed as I had freshly cloned my repo and hadn't made any changes worth committing.
After a day swapping between open and proprietary drivers and tinkering with my xorg.conf, I ended up nuking the disk and installing 16.04 LTS. I installed the SDL2 runtimes from the ubuntu repo, and I installed the SDL2 development source from the tar file on the SDL website. Ran my program again and it worked as expected.
I don't want to call this "solved" as I ended up nuking my OS, but I guess my problem is solved. I just wish I could figure out the information chain on how SDL is informed on how many displays there are.
checking for X11/extensions/Xinerama.h... no
That's probably what torpedoed you. Xinerama is used to query the output count/positions/topology:
The X server's client/server architecture allows the server to expose Xinerama information to the client regardless of whether the Xinerama infrastructure is active. RandR and NVidia's twinview utilize this feature to provide window managers and clients with information about the output layout relative to the framebuffer.
Installing libxinerama-dev & re-configureing probably would have cleared that up.
Do anyone have any idea if the gstreamer-java-1.6.jar downloadable from https://code.google.com/p/gstreamer-java/downloads/list is compatible with gstreamer windows 1.6.1 binaries? I installed both and setup the eclipse environment but somehow it gives me the warning for all(assuming as lot of warnings are on the eclipse console) the gstreamer plugins.
(javaw.exe:9744): GStreamer-WARNING **: Failed to load plugin 'D:\gstreamer\1.0\x86_64\lib\gstreamer-1.0\libgstgio.dll': 'D:\gstreamer\1.0\x86_64\lib\gstreamer-1.0\libgstgio.dll': The specified module could not be found..............
Although I can see everything is fine from command prompt running gstreamer(creating sample pipeline etc).
For any other, it will be helpful that gstreamer-1.6 compatible java bindings are there and available on https://github.com/praxis-live/praxis/tree/master/lib.gst1-java-core/release/modules/ext.
See reply on java-gstreamer forum.
No, it isn't!
The new home of the Java bindings is at https://github.com/gstreamer-java
The old GStreamer-Java only supports GStreamer 0.10. Try
gst1-java-core at https://github.com/gstreamer-java/gst1-java-core
There are no pre-built binaries there, although you can get my one
from Praxis LIVE -
https://github.com/praxis-live/praxis/tree/master/lib.gst1-java-core/release/modules/ext
Click on it, then click view raw and you can download it.
It's still a work in progress, and I haven't tested on Windows yet,
although others may have.
Best wishes,
Neil
I am running a multi-process calculations using MPICH2 from Argonne National Laboratory. The run is on a single machine, so -localonly option of mpiexec is used. Usually everything works.
If the network adapter is enabled when the run starts, and if I disable it during the run, the run fails with error stating:
op_read error on left context: Error = -1
op_read error on parent context: Error = -1
unable to read the cmd header on the left context, Error = -1
.
unable to read the cmd header on the parent context, Error = -1
.
Error posting readv, An existing connection was forcibly closed by the remote host.(10054)
connection to my parent broken, aborting.
state machine failed.
However, if the network adapter is disabled when the run is started, I can enable/disable the adapter as many times as I want, and the run still proceeds to the end.
Is there a way to run mpiexec or modify smpd configuration such that MPICH2 is not using network adapter for inter-process communication for local runs even if the adapter is available when the run starts?
Update
The platform is Windows 7 Professional.
The version of MPICH2 is 1.4.1 (32-bit) which is latest version of MPICH2 that I can find.
As was posted on the mailing list, you're using a very old version of MPICH. You should update to a newer version that does all of this automatically.
UPDATE: As has been pointed out, on Windows, it's not possible to update MPICH past 1.4.1. For anything beyond that, you should use either MS-MPI or Intel MPI.
I am using Fedora-16 (64 bit) Operating System.
I have got registered with CloudFoundry and get download "micro-1.2.0.zip" and
"*VMware-Player-4.0.2-591240.x86_64.bundle*". I have successfully installed the "VMware-Player". Then how can install "micro cloud foundry" correctly?
When extract it and run the "micro.vmx" file, then I have got several warring messages and finally got this error message as,
This kernel requires an x86_64 CPU, but only detected an i686 CPU.
Unable to boot - please use a kernel appropriate for your CPU
If some one know how I can over come this problem or what is the correct way to install "MicroCloudFoundry", please help me. Thanks.
You have to check the processor compatibility as well.
Here is the link where you can check the processor compatibility:
http://downloads.vmware.com/d/details/processor_check_5_5_dt/dCpiQGhkYmRAZQ
I recently downloaded new linux kernel source code and compiled on ubuntu. After my system is not recognizing any usb devices. Is there any changes required to compiling procedure or make files in order to enable USB host?
Please help me to get out of this problem. Thanks.
It is possible for the USB support to be disabled though incorrect or incomplete configuration.
You didn't mention how you configured the kernel though.
A good point to start is the existing configuration for the default Ubuntu kernel. There should be a config file in boot, like /boot/config-3.1.0-1-amd64.
Copy that file to your kernel directory as .config, then use make oldconfig to update the configuration.
When installing the kernel take care to create the corresponding initrd as well.