Final Builder Automation taking too much of time - finalbuilder

I am using final builder to compile all of my projects. It contains the actions like svn update, revert, assembly version switcher and copyright info changer. Previously it was working fine. But suddenly it is taking so much of time to complete each and every single action. 1 hour running is now taking more than 3 to 4 hours. But machine is running fast. Contains 4 GB RAM and i5 Processor.
Could anyone please let me know the solution for this?

Related

Java Mission Control (JMC) 6.0 does not show hot methods when examining a JFR flight recording

After using the Java Flight Recording functionality on a running application, Java's JMC could be used to examine the resulting JFR file to show hot methods, and a percentage of CPU time spent in each method. This was very useful for profiling applications and identifying bottlenecks.
Here is an example of what was previously possible in older versions of JMC:
This screen seems to be missing in JMC 6.0 that is bundled with Java 10. Here is an example of what I see in the Java 10 bundled JMC 6.0:
There seems to be a rudimentary count of method calls, but no Percentage CPU time is listed. There don't appear to be options to add the missing column.
Is there a way to show hot method %CPU time, or has this functionality been removed from JMC 6.0 in JDK10?
The percentage column is not cpu time, but percentage of the total number of method samples.
The same info is displayed as the backdrop in the Count column in JMC 6.0, and I believe if you hover over the column to get the tooltip there might be a percentage number.
I know you are not the first person to miss the percentage column, there is an enhancement request in the JMC Jira for this: https://bugs.openjdk.java.net/browse/JMC-5721
At time of writing JMC 7.1.2 still doesn't give the chance to see the percentage CPU time. I've found great help read jfr files with VisualVM -> Sampler -> Display -> CPU

Extracting Features from VGG

I want to extract features from images in MS COCO dataset using a fine-tuned VGG-19 network.
However, it takes about 6~7 seconds per image, roughly 2 hours per 1k images. (even longer for other fine-tuned models)
There are 120k images in MS COCO dataset, so it'll take at least 10 days.
Is there any way that I can speed up the feature extraction process?
Well, this is not just a command. First you must check whether your GPU is powerful enough to wrestle with deep CNNs. Knowing your GPU model can answer this question.
Second, you have to compile and build Caffe framework with CUDA and GPU-enabled (CPU_Only disabled) in the Makefile.config (or CMakeLists.txt).
Passing all required steps (installing Nvidia Driver, installing CUDA and etc.) you can build caffe for GPU-use. Then by passing the GPU_Device_ID in your command-line you can benefit from speed provided by them.
Follow this link for building Caffe using GPU.
Hope it helps
This ipython notebook example explains the steps to extract features out of any caffe model really well: https://github.com/BVLC/caffe/blob/master/examples/00-classification.ipynb
In pycaffe, you can set gpu mode simply by using caffe.set_mode_gpu().

GNU Parallel host sticky jobs

I am writing a parallel build farm to build C++ cross-platform applications against various platforms / environments. Every time new code is pushed to a git repo, I build and test the latest code against all the platforms.
I've setup parallel to correctly distribute the jobs among several hosts using the --sshlogin option.
I transfer files, collect output and results. It's all working more than fine and I love the tool.
The build time being sometimes quite long for some platforms, I would like the build to be as incremental as possible.
My only issue is that the build is only incremental if the scheduler sends the jobs to the same machine and reuse the artefacts of the previous build on this specific host.
Say I have 3 hosts, I have 1 chance in 3 for the build to be incremental. If a hosts hasn't built this platform in a while, it might take a long time.
Is it possible to gain control over the host a specific input source will run on and only fallback to the other hosts if the host is busy?
Ideally, I would love to see a tag system where I tag input source with a name and tag several hosts with a name, creating pools of jobs and pools of machines specialized into that type of build.
But a very simple implementation where the input sources are distributed in the same order as the order the sshlogins are defined could be a simple & quick fix in my situation.
I tried to find the source code to implement it myself but I only see doc generation when I browse the code on Savannah.
Any ideas?
Thanks,
M
There is currently no support for prioritizing a given argument to a given sshlogin. The source code is at https://savannah.gnu.org/git/?group=parallel
Feel free to join the mailing list and discuss the idea: https://lists.gnu.org/mailman/listinfo/parallel
The only priority in the code is when a job has failed on an sshlogin, then GNU Parallel prefers to retry that job on another sshlogin. Maybe that could be extended?
If a job is marked as having failed -1 time for a given sshlogin, then GNU Parallel ought to prefer to run the job on that sshlogin.
I've been trying to discuss this idea on the mailing list as you suggested but never had any respone in more than 10 days... I guess you must be busy with other things at the moment. So I went along and forked the source code to make the necessary changes and make my solution work.
I pushed it there a week ago:
http://michakfromparis.github.io/gnu-parallel-sticky/
the source code is available on github here:
https://github.com/michaKFromParis/gnu-parallel-sticky
Wasn't exactly easy without any guidance as the source code has a lot of history so I tried to keep the changes surgical to ease merge of your future releases.
I've been using it in production for more than a week now and it works perfectly in my configuration.
It is also compatible with older formats, should be a drop-in replacement for usual parallel uses with extra features on the side.
Would love to get feedback from other users though as it might not be completely dry.
Thanks for sharing the original source code.
Best Regards,
M

Visual Studio Online build times are increasing dramatically

On my development machine (Dell XPS 8700), I can build my solution in 1 minute. Using VSO to build it takes much longer. Also, my build times over the past 10 days are increasing dramatically, even though I've only made minor changes to the source.
Here are the last 4 build times for the line in the log that says "Run MSBuild":
7:11
10:56
14:32
22:33
Does anybody know why this is happening?
With only 60 minutes of included build time per month, this is an issue for me.
Thanks.
We were having some issues with the build service a couple of weeks ago - when you posted the question. Is this problem still occurring?

Neon toolkit and Gate Web Service

I am trying to run any of the services from gate web service, in neon 2.3.
Even Annie that runs so well in gate doesn't run, or better, it stay for indefinite time processing, a thing that should take no more than a couple of seconds. I run wizard, set input directory, leave file pattern as default and set a folder and name for the output ontology, shouldn't it be enough? Shouldn't i get something, even an error?
I think its the location who's giving me problems.
http://safekeeper1.dcs.shef.ac.uk/neon/services/sardine
http://safekeeper1.dcs.shef.ac.uk/neon/services/sprat
http://safekeeper1.dcs.shef.ac.uk/neon/services/annie
http://safekeeper1.dcs.shef.ac.uk/neon/services/termraider
How can i confirm it? Can i run it offline?
Can anyone give me a hand?
Also, i've seen sprat running on gate, on "SPRAT: a tool for automatic semantic pattern-based ontology population"
Can anyone teach me how, and with what versions?
Thx,
Celso Costa