Why does our Coldfusion server require a daily restart? - coldfusion

Our Coldfusion server requires a daily restart using command
-jar jrun.jar -start coldfusion
Not sure why it is stopping daily. Someone please help to resolve this.
We have coldfusion running on a UNIX server.
Server.log has the below:
"Warning","14","04/22/15","16:48:38",,"Template:
/cfide/cf5dev/break/Cage/cage.cfm, Ran: 8 seconds."
"Warning","12","04/22/15","16:50:17",,"Template:
/cfide/cf5dev/break/Cage/cage.cfm, Ran: 22 seconds."
"Warning","14","04/22/15","16:50:20",,"Template:
/cfide/cf5dev/break/Cage/cage.cfm, Ran: 27 seconds."
"Warning","9","04/22/15","16:53:20",,"Template:
/cfide/administrator/logging/downloadlog.cfm, Ran: 2 seconds."
"Warning","9","04/22/15","16:54:23",,"Template:
/cfide/administrator/logging/downloadlog.cfm, Ran: 2 seconds."
Webserver.log
Webserver "Error","13","04/22/15","13:13:52",,"An error occurred while
attempting to establish a connection to the server. The most likely
cause of this problem is that the server is not currently running.
Verify that the server is running and restart it if necessary. Unix
error number 2 occurred: No such file or directory"
Application.log
Wed Apr 22 16:48:26 2015Browser: Mozilla/4.0 (compatible; MSIE
7.0; Windows NT 6.1; WOW64; Trident/4.0; SLCC2; .NET CLR 2.0.50727; .NET CLR 3.5.30729; .NET CLR 3.0.30729; Media Center PC 6.0; .NET4.0C;
.NET4.0E)Remote Address: 10.2.200.119
opt/coldfusion/JRun/logs>ls -lrt
total 2284
-rw-r--r-- 1 cf5dev nobody 0 May 27 2006 default-out.log
-rw-r--r-- 1 cf5dev cfdev 99954 Jan 14 00:45 default-event_5.log
-rw-r--r-- 1 cf5dev cfdev 100017 Feb 6 16:57 default-event_4.log
-rw-r--r-- 1 cf5dev cfdev 99973 Mar 5 00:45 default-event_3.log
-rw-r--r-- 1 cf5dev cfdev 99961 Mar 29 17:01 default-event_2.log
-rw-r--r-- 1 cf5dev cfdev 99981 Apr 22 00:45 default-event_1.log
-rw-r--r-- 1 cf5dev nobody 610627 Apr 23 00:45 default-err.log
-rw-r--r-- 1 cf5dev cfdev 6706 Apr 23 00:45 default-event.log
Default-event.log
04/23 00:45:20 info (JRun) Destroying jcp
04/23 00:45:20 info (JRun) Destroying demo-app
04/23 00:45:20 error (jcp) Socket closed [java.net.SocketException: Socket closed]
java.net.SocketException: Socket closed
at java.net.PlainSocketImpl.socketSetOption(Native Method)
at java.net.PlainSocketImpl.setOption(PlainSocketImpl.java:186)
at java.net.Socket.setTcpNoDelay(Socket.java:377)
at allaire.jrun.jrpp.ProxyService.createRunnable(ProxyService.java:32)
at allaire.jrun.ThreadPool.createRunnable(ThreadPool.java:175)
at allaire.jrun.WorkerThread.run(WorkerThread.java:72)
04/23 00:45:20 info (JRun) Destroying default-app
04/23 00:45:20 info (JRun) FlashGenerator: destroy
04/23 00:45:20 info (JRun) Destroying control
04/23 00:45:20 info (JRun) Destroying monitor
04/23 00:45:20 info (JRun) Destroying logging
04/23 00:45:20 info (JRun) Destroying scheduler
04/23 00:45:28 info (JRun) Loading monitor
04/23 00:45:28 info (JRun) Loading license
04/23 00:45:28 info (license) Enabling unlimited concurrency for Enterprise
04/23 00:45:28 info (JRun) Loading control
04/23 00:45:28 info (control) control listening on 127.0.0.1:22719
04/23 00:45:28 info (JRun) Loading default-app
04/23 00:45:28 info (default-app) Starting default-app: Default User Application - Default application for getting started
04/23 00:45:28 info (default-app) Loading scheduler
04/23 00:45:28 info (default-app) Loading logging
04/23 00:45:28 info (JRun) Loading session
04/23 00:45:28 info (session) Loading persistence storage provider 'file'
04/23 00:45:28 info (JRun) Loading authentication
04/23 00:45:28 info (JRun) Loading propfile
04/23 00:45:28 info (JRun) Loading jsp
04/23 00:45:28 info (JRun) Loading file
04/23 00:45:28 info (JRun) FlashGenerator: init
04/23 00:45:29 info (JRun) default-app Ready
04/23 00:45:29 info (JRun) Loading demo-app
04/23 00:45:29 info (demo-app) Starting demo-app: JRun Demo - JRun Demo
04/23 00:45:29 info (demo-app) Loading scheduler
04/23 00:45:29 info (demo-app) Loading logging
04/23 00:45:29 info (JRun) Loading session
04/23 00:45:29 info (JRun) Loading authentication
04/23 00:45:29 info (JRun) Loading propfile
04/23 00:45:29 info (session) Loading persistence storage provider 'file'
04/23 00:45:29 info (JRun) Loading jsp
04/23 00:45:30 info (JRun) Loading file
04/23 00:45:30 info (JRun) demo-app Ready
04/23 00:45:30 info (JRun) Loading jcp
04/23 00:45:30 info (jcp) jcp listening on 127.0.0.1:22718
04/23 00:45:30 info (JRun) default Ready
Default-err.log
Operating System: SunOS Version 5.10
Java Virtual Machine: mixed mode from Sun Microsystems Inc.
JRun 3.0 3.02a.11569 Starting default...
Current Locale: en
Loading scheduler
Loading logging
Operating System: SunOS Version 5.10
Java Virtual Machine: mixed mode from Sun Microsystems Inc.
JRun 3.0 3.02a.11569 Starting default...
Current Locale: en
Loading scheduler
Loading logging
Operating System: SunOS Version 5.10
Java Virtual Machine: mixed mode from Sun Microsystems Inc.
JRun 3.0 3.02a.11569 Starting default...
Current Locale: en
Loading scheduler
Loading logging
Operating System: SunOS Version 5.10
Java Virtual Machine: mixed mode from Sun Microsystems Inc.
JRun 3.0 3.02a.11569 Starting default...
Current Locale: en
Loading scheduler
Loading logging
Operating System: SunOS Version 5.10
Java Virtual Machine: mixed mode from Sun Microsystems Inc.
JRun 3.0 3.02a.11569 Starting default...
Current Locale: en
Loading scheduler
Loading logging
Which of the log files will give the exact information about this stopping issue?
There is no log file named as coldfusion-out.log

Related

Not able to access | Jenkins portal | After EC-2 instance restart

Issue : Not able to access Jenkins portal after EC-2 instance restart.
Debug steps taken: restarted jenkins service, added jdk path in jenkins.
Getting Output as,
This site can’t be reached
● jenkins.service - Jenkins Continuous Integration Server
Loaded: loaded (/usr/lib/systemd/system/jenkins.service; enabled;
vendor preset: disabled)
Active: active (running) since Wed 2022-05-11 05:15:30 UTC; 28min ago
Main PID: 2073 (java)
Tasks: 33 (limit: 4690)
Memory: 311.1M
CGroup: /system.slice/jenkins.service
└─2073 /usr/bin/java -Djava.awt.headless=true -jar
/usr/share/java/jenkins.war --webroot=/var/cache/jenkins/war --
httpPort=8080
May 11 05:15:30 ip jenkins[2073]: 2022-05-11 05:15:30.671+0000 [id=41]
INFO hudson.model.AsyncPeriodicWork#lambda$doRun$1: Finished
Download >
May 11 05:15:30 ip jenkins[2073]: 2022-05-11 05:15:30.857+0000 [id=26]
INFO jenkins.InitReactorRunner$1#onAttained: Completed
initialization
May 11 05:15:30 ip jenkins[2073]: 2022-05-11 05:15:30.913+0000 [id=20]
INFO hudson.lifecycle.Lifecycle#onReady: Jenkins is fully up and
runni>
May 11 05:15:30 ip systemd[1]: Started Jenkins Continuous Integration
Server.
May 11 05:16:02 ip jenkins[2073]: 2022-05-11 05:16:02.724+0000 [id=56]
INFO hudson.model.AsyncPeriodicWork#lambda$doRun$1: Started EC2
alive >
May 11 05:16:02 ip jenkins[2073]: 2022-05-11 05:16:02.725+0000 [id=56]
INFO hudson.model.AsyncPeriodicWork#lambda$doRun$1: Finished EC2
alive>
May 11 05:26:02 ip jenkins[2073]: 2022-05-11 05:26:02.724+0000 [id=57]
INFO hudson.model.AsyncPeriodicWork#lambda$doRun$1: Started EC2
alive >
May 11 05:26:02 ip jenkins[2073]: 2022-05-11 05:26:02.725+0000 [id=57]
INFO hudson.model.AsyncPeriodicWork#lambda$doRun$1: Finished EC2
alive>
May 11 05:36:02 ip jenkins[2073]: 2022-05-11 05:36:02.724+0000 [id=58]
INFO hudson.model.AsyncPeriodicWork#lambda$doRun$1: Started EC2
alive >
May 11 05:36:02 ip jenkins[2073]: 2022-05-11 05:36:02.725+0000 [id=58]
INFO hudson.model.AsyncPeriodicWork#lambda$doRun$1: Finished EC2
alive>

Not able to save crash dump using kdump

I have a VPS server on amazon's AWS Lightsail service.
I've been testing kdump using the following two commands (to trigger an automatic kernel crash):
# echo 1 > /proc/sys/kernel/sysrq
# echo c > /proc/sysrq-trigger
The problem is that the system crashed and rebooted, but there's no dump saved.
Here is a list of checking I've done:
[centos#server crash]$ systemctl status kdump
● kdump.service - Crash recovery kernel arming
Loaded: loaded (/usr/lib/systemd/system/kdump.service; enabled; vendor preset: enabled)
Active: active (exited) since Mon 2019-03-18 07:43:34 UTC; 5 days ago
Process: 4119 ExecStart=/usr/bin/kdumpctl start (code=exited, status=0/SUCCESS)
Main PID: 4119 (code=exited, status=0/SUCCESS)
CGroup: /system.slice/kdump.service
Mar 18 07:43:32 ip-.ap-northeast-1.compute.internal systemd[1]: Starting Crash recovery kernel arming...
Mar 18 07:43:34 ip-.ap-northeast-1.compute.internal kdumpctl[4119]: kexec: loaded kdump kernel
Mar 18 07:43:34 ip-.ap-northeast-1.compute.internal kdumpctl[4119]: Starting kdump: [OK]
Mar 18 07:43:34 ip-.ap-northeast-1.compute.internal systemd[1]: Started Crash recovery kernel arming.
[centos#server crash]$ dmesg | grep Reserving
[ 0.000000] Reserving 256MB of memory at 368MB for crashkernel (System RAM: 2047MB)
[centos#server crash]$ cat /proc/cmdline
BOOT_IMAGE=/boot/vmlinuz-3.10.0-957.1.3.el7.x86_64 root=UUID=f41e390f-835b-4223-a9bb-9b45984ddf8d ro console=tty0 crashkernel=256M console=ttyS0,115200
[centos#server crash]$ grep -v ^# /etc/kdump.conf
path /var/crash
core_collector makedumpfile -l --message-level 1 -d 31
default reboot
There's no log of the crash in the /var/log/messages indicating any error there might be. So I wonder what I might have missed. Or an AWS Lightsail VPS is not capable of saving a kdump at all...?

AWS BeanStalk doesn't tend to connect to AWS RDS

I think I am almost there.
I created an instance of AWS BeanStalk and added an oracle DB instance to it.
When I found the log, I saw the driver was loaded but it keeps saying that URL is
invalid.
Here are my RDS info and log message.
[RDS Info]
Endpoint = aa1c9autjaqoufk.c2k1ch01futy.ap-northeast-2.rds.amazonaws.com
Port = 1521
Public Access = yes
[System Log]
25-Jun-2018 02:42:56.759 INFO [main] org.apache.coyote.AbstractProtocol.init Initializing ProtocolHandler ["http-nio-8080"]
25-Jun-2018 02:42:56.787 INFO [main] org.apache.tomcat.util.net.NioSelectorPool.getSharedSelector Using a shared selector for servlet write/read
25-Jun-2018 02:42:56.796 INFO [main] org.apache.coyote.AbstractProtocol.init Initializing ProtocolHandler ["ajp-nio-8009"]
25-Jun-2018 02:42:56.799 INFO [main] org.apache.tomcat.util.net.NioSelectorPool.getSharedSelector Using a shared selector for servlet write/read
25-Jun-2018 02:42:56.800 INFO [main] org.apache.catalina.startup.Catalina.load Initialization processed in 1366 ms
25-Jun-2018 02:42:56.842 INFO [main] org.apache.catalina.core.StandardService.startInternal Starting service Catalina
25-Jun-2018 02:42:56.848 INFO [main] org.apache.catalina.core.StandardEngine.startInternal Starting Servlet Engine: Apache Tomcat/8.0.50
25-Jun-2018 02:42:56.872 INFO [localhost-startStop-1] org.apache.catalina.startup.HostConfig.deployDirectory Deploying web application directory /var/lib/tomcat8/webapps/ROOT
25-Jun-2018 02:42:58.613 INFO [localhost-startStop-1] org.apache.jasper.servlet.TldScanner.scanJars At least one JAR was scanned for TLDs yet contained no TLDs. Enable debug logging for this logger for a complete list of JARs that were scanned but no TLDs were found in them. Skipping unneeded JARs during scanning can improve startup time and JSP compilation time.
25-Jun-2018 02:42:58.689 INFO [localhost-startStop-1] org.apache.catalina.startup.HostConfig.deployDirectory Deployment of web application directory /var/lib/tomcat8/webapps/ROOT has finished in 1,817 ms
25-Jun-2018 02:42:58.693 INFO [main] org.apache.coyote.AbstractProtocol.start Starting ProtocolHandler ["http-nio-8080"]
25-Jun-2018 02:42:58.720 INFO [main] org.apache.coyote.AbstractProtocol.start Starting ProtocolHandler ["ajp-nio-8009"]
25-Jun-2018 02:42:58.736 INFO [main] org.apache.catalina.startup.Catalina.start Server startup in 1935 ms
Loading driver...
Driver loaded!
jdbc:oracle:oci://aa1c9autjaqoufk.c2k1ch01futy.ap-northeast-2.rds.amazonaws.com:1521/ebdb?user=username&password=password
SQLException: Invalid Oracle URL specified
SQLState: 99999
VendorError: 17067
Closing the connection.
SQLException: Invalid Oracle URL specified
SQLState: 99999
VendorError: 17067
Closing the connection.
I included ojdbc8 drvier in my web project library and made a build.
Is this about driver? What am I doing wrong?
Message clearly says your URL is incorrect,
It should be something like below.
//step1 load the driver class
Class.forName("oracle.jdbc.driver.OracleDriver");
//step2 create the connection object
Connection con=DriverManager.getConnection(
"jdbc:oracle:thin:#aa1c9autjaqoufk.c2k1ch01futy.ap-northeast-2.rds.amazonaws.com:1521:edb","username","password");
`

I can connect to AWS RDS via sqldeveloper but can't by Java application

It is so werid that I can connect to AWS RDS with sqldeveloper but can't with my java application(java source code or jsp)
When I try to access to RDS, there are errors like:
coyote.AbstractProtocol.init Initializing ProtocolHandler ["http-nio-8080"]
26-Jun-2018 04:24:33.203 INFO [main] org.apache.tomcat.util.net.NioSelectorPool.getSharedSelector Using a shared selector for servlet write/read
26-Jun-2018 04:24:33.212 INFO [main] org.apache.coyote.AbstractProtocol.init Initializing ProtocolHandler ["ajp-nio-8009"]
26-Jun-2018 04:24:33.215 INFO [main] org.apache.tomcat.util.net.NioSelectorPool.getSharedSelector Using a shared selector for servlet write/read
26-Jun-2018 04:24:33.219 INFO [main] org.apache.catalina.startup.Catalina.load Initialization processed in 1387 ms
26-Jun-2018 04:24:33.265 INFO [main] org.apache.catalina.core.StandardService.startInternal Starting service Catalina
26-Jun-2018 04:24:33.266 INFO [main] org.apache.catalina.core.StandardEngine.startInternal Starting Servlet Engine: Apache Tomcat/8.0.50
26-Jun-2018 04:24:33.286 INFO [localhost-startStop-1] org.apache.catalina.startup.HostConfig.deployDirectory Deploying web application directory /var/lib/tomcat8/webapps/ROOT
26-Jun-2018 04:24:35.020 INFO [localhost-startStop-1] org.apache.jasper.servlet.TldScanner.scanJars At least one JAR was scanned for TLDs yet contained no TLDs. Enable debug logging for this logger for a complete list of JARs that were scanned but no TLDs were found in them. Skipping unneeded JARs during scanning can improve startup time and JSP compilation time.
26-Jun-2018 04:24:35.097 INFO [localhost-startStop-1] org.apache.catalina.startup.HostConfig.deployDirectory Deployment of web application directory /var/lib/tomcat8/webapps/ROOT has finished in 1,811 ms
26-Jun-2018 04:24:35.100 INFO [main] org.apache.coyote.AbstractProtocol.start Starting ProtocolHandler ["http-nio-8080"]
26-Jun-2018 04:24:35.106 INFO [main] org.apache.coyote.AbstractProtocol.start Starting ProtocolHandler ["ajp-nio-8009"]
26-Jun-2018 04:24:35.108 INFO [main] org.apache.catalina.startup.Catalina.start Server startup in 1888 ms
Loading driver...
Driver loaded!
jdbc:oracle:thin://IP:1521/ORCL?user=username&password=password
SQLException: Invalid Oracle URL specified
SQLState: 99999
VendorError: 17067
Closing the connection.
SQLException: Invalid Oracle URL specified
SQLState: 99999
VendorError: 17067
Closing the connection.
But the URL is just the same value as I tried with sqldeveloper.
Is there anything wrong?
Please enlighten me since I've been suffering for this about a week! :(
I'm not sure how your application is set up, but I'm using Maven & Spring Boot and I got it working like this:
I mainly followed this guide, ignoring the .sql files, thymeleaf UI, "model.addAttribute("cities", cities);" part, and the html file:
https://zetcode.com/springboot/postgresql/
My application.properties file looks like this
postgres.comment.aa=https://zetcode.com/springboot/postgresql/
spring.main.banner-mode=off
logging.level.org.springframework=ERROR
spring.jpa.hibernate.ddl-auto=none
spring.datasource.initialization-mode=always
spring.datasource.platform=postgres
spring.datasource.url=jdbc:postgresql://your-rds-url-here.us-east-1.rds.amazonaws.com:yourDbPortHere/postgres
spring.datasource.username=postgres
spring.datasource.password=<your db password here>
spring.jpa.properties.hibernate.jdbc.lob.non_contextual_creation=true
If you have custom schemas, you can append "?currentSchema=users" to the url:
spring.datasource.url=jdbc:postgresql://your-rds-url-here.us-east-1.rds.amazonaws.com:yourDbPortHere/postgres?currentSchema=users
Thanks to this SO answer for the schema:
Is it possible to specify the schema when connecting to postgres with JDBC?
These other couple links also helped
https://turreta.com/2015/03/01/how-to-specify-a-default-schema-when-connecting-to-postgresql-using-jdbc/
https://doc.cuba-platform.com/manual-latest/db_schema_connection.html

Error when running 'vagrant up' -- Too many open files - getcwd (Errno::EMFILE)

I have a very basic Vagrantfile, and when I call 'vagrant up' for the first time (after destroying the box) I get the error below just about every time. Any ideas why? I'm at a loss. I'm on OSX.
Vagrantfile:
# -*- mode: ruby -*-
# vi: set ft=ruby :
Vagrant.configure("2") do |config|
config.vm.hostname = "myapp-berkshelf"
config.vm.box = "precise64_provisioned"
config.vm.box_url = "http://dl.dropbox.com/u/1537815/precise64.box"
config.omnibus.chef_version = :latest
config.vm.network :private_network, ip: "33.33.33.10"
config.vm.provision :chef_solo do |chef|
chef.cookbooks_path = ["cookbooks", "my-cookbooks"]
chef.add_recipe "nginx"
end
config.berkshelf.enabled = true
config.vm.provision :chef_solo do |chef|
chef.json = {
:mysql => {
:server_root_password => 'rootpass',
:server_debian_password => 'debpass',
:server_repl_password => 'replpass'
}
}
chef.run_list = [
"recipe[myapp::default]"
]
end
end
Backtrace:
INFO global: Vagrant version: 1.3.1
...<VagrantPlugins::Omnibus::Action::InstallChef:0x00000108507c98>
INFO warden: Calling action: #<Berkshelf::Vagrant::Action::Install:0x00000108507c20>
INFO interface: warn: This version of the Berkshelf plugin has not been fully tested on this version of Vagrant.
[Berkshelf] This version of the Berkshelf plugin has not been fully tested on this version of Vagrant.
INFO interface: warn: You should check for a newer version of vagrant-berkshelf.
[Berkshelf] You should check for a newer version of vagrant-berkshelf.
INFO interface: warn: If you encounter any errors with this version, please report them at https://github.com/RiotGames/vagrant-berkshelf/issues
[Berkshelf] If you encounter any errors with this version, please report them at https://github.com/RiotGames/vagrant-berkshelf/issues
INFO interface: warn: You can also join the discussion in #berkshelf on Freenode.
[Berkshelf] You can also join the discussion in #berkshelf on Freenode.
INFO interface: info: Updating Vagrant's berkshelf: '/Users/macm/.berkshelf/default/vagrant/berkshelf-20130913-24059-1ex15i6-default'
[Berkshelf] Updating Vagrant's berkshelf: '/Users/macm/.berkshelf/default/vagrant/berkshelf-20130913-24059-1ex15i6-default'
INFO interface: info: Using application_ruby (2.1.0)
[Berkshelf] Using application_ruby (2.1.0)
INFO interface: info: Using postgresql (3.0.4)
[Berkshelf] Using postgresql (3.0.4)
INFO interface: info: Using apt (2.1.1)
[Berkshelf] Using apt (2.1.1)
INFO interface: info: Using user (0.3.0)
[Berkshelf] Using user (0.3.0)
INFO interface: info: Using ruby_build (0.8.0)
[Berkshelf] Using ruby_build (0.8.0)
INFO interface: info: Using nginx (1.8.0)
[Berkshelf] Using nginx (1.8.0)
INFO interface: info: Using application (3.0.0)
[Berkshelf] Using application (3.0.0)
INFO interface: info: Using unicorn (1.3.0)
[Berkshelf] Using unicorn (1.3.0)
INFO interface: info: Using apache2 (1.7.0)
[Berkshelf] Using apache2 (1.7.0)
INFO interface: info: Using passenger_apache2 (2.0.4)
[Berkshelf] Using passenger_apache2 (2.0.4)
INFO interface: info: Using build-essential (1.4.2)
[Berkshelf] Using build-essential (1.4.2)
INFO interface: info: Using runit (1.2.0)
[Berkshelf] Using runit (1.2.0)
INFO interface: info: Using yum (2.3.2)
[Berkshelf] Using yum (2.3.2)
INFO interface: info: Using openssl (1.1.0)
[Berkshelf] Using openssl (1.1.0)
INFO interface: info: Using ohai (1.1.12)
[Berkshelf] Using ohai (1.1.12)
INFO warden: Calling action: #<Berkshelf::Vagrant::Action::Upload:0x00000108507bf8>
INFO warden: Calling action: #<Vagrant::Action::Builtin::EnvSet:0x00000108507bd0>
INFO warden: Calling action: #<VagrantPlugins::ProviderVirtualBox::Action::PrepareForwardedPortCollisionParams:0x00000108507ba8>
INFO subprocess: Starting process: ["/usr/bin/VBoxManage", "list", "vms"]
INFO subprocess: Starting process: ["/usr/bin/VBoxManage", "showvminfo", "72dd1e43-2966-4b2a-96ea-e25df6ee7d5e", "--machinereadable"]
INFO warden: Calling action: #<Vagrant::Action::Builtin::HandleForwardedPortCollisions:0x00000108507b80>
INFO handle_port_collisions: Detecting any forwarded port collisions...
INFO warden: Calling action: #<VagrantPlugins::ProviderVirtualBox::Action::PruneNFSExports:0x00000100b1ae08>
INFO subprocess: Starting process: ["/usr/bin/VBoxManage", "list", "vms"]
INFO warden: Calling action: #<Vagrant::Action::Builtin::NFS:0x00000100b1ade0>
INFO warden: Calling action: #<VagrantPlugins::ProviderVirtualBox::Action::PrepareNFSSettings:0x000001084eea68>
INFO warden: Calling action: #<VagrantPlugins::ProviderVirtualBox::Action::ClearSharedFolders:0x000001084d9f50>
INFO subprocess: Starting process: ["/usr/bin/VBoxManage", "showvminfo", "f31f231f-b99d-45ac-91cf-0929308714e9", "--machinereadable"]
INFO warden: Calling action: #<VagrantPlugins::ProviderVirtualBox::Action::ShareFolders:0x000001084d9f28>
INFO interface: info: Creating shared folders metadata...
[default] Creating shared folders metadata...
INFO subprocess: Starting process: ["/usr/bin/VBoxManage", "sharedfolder", "add", "f31f231f-b99d-45ac-91cf-0929308714e9", "--name", "/vagrant", "--hostpath", "/Users/macm/Documents/Development/chef-repo/my-cookbooks/measureful"]
INFO subprocess: Starting process: ["/usr/bin/VBoxManage", "setextradata", "f31f231f-b99d-45ac-91cf-0929308714e9", "VBoxInternal2/SharedFoldersEnableSymlinksCreate//vagrant", "1"]
INFO subprocess: Starting process: ["/usr/bin/VBoxManage", "sharedfolder", "add", "f31f231f-b99d-45ac-91cf-0929308714e9", "--name", "/tmp/vagrant-chef-1/chef-solo-1/cookbooks", "--hostpath", "/Users/macm/.berkshelf/default/vagrant/berkshelf-20130913-24059-1ex15i6-default"]
INFO subprocess: Starting process: ["/usr/bin/VBoxManage", "setextradata", "f31f231f-b99d-45ac-91cf-0929308714e9", "VBoxInternal2/SharedFoldersEnableSymlinksCreate//tmp/vagrant-chef-1/chef-solo-1/cookbooks", "1"]
INFO warden: Calling action: #<VagrantPlugins::ProviderVirtualBox::Action::ClearNetworkInterfaces:0x00000106ba6650>
INFO interface: info: Clearing any previously set network interfaces...
[default] Clearing any previously set network interfaces...
INFO subprocess: Starting process: ["/usr/bin/VBoxManage", "modifyvm", "f31f231f-b99d-45ac-91cf-0929308714e9", "--nic2", "none", "--nic3", "none", "--nic4", "none", "--nic5", "none", "--nic6", "none", "--nic7", "none", "--nic8", "none"]
INFO warden: Calling action: #<VagrantPlugins::ProviderVirtualBox::Action::Network:0x00000106ba6628>
INFO network: Determining network adapters required for high-level configuration...
INFO network: -- Slot 2: hostonly
INFO network: Determining adapters and compiling network configuration...
INFO network: Network slot 1. Type: nat.
INFO network: Network slot 2. Type: hostonly.
INFO subprocess: Starting process: ["/usr/bin/VBoxManage", "list", "bridgedifs"]
INFO network: Searching for matching hostonly network: 33.33.33.10
INFO subprocess: Starting process: ["/usr/bin/VBoxManage", "list", "dhcpservers"]
INFO subprocess: Starting process: ["/usr/bin/VBoxManage", "list", "hostonlyifs"]
INFO network: Enabling adapters...
INFO interface: info: Preparing network interfaces based on configuration...
[default] Preparing network interfaces based on configuration...
INFO subprocess: Starting process: ["/usr/bin/VBoxManage", "modifyvm", "f31f231f-b99d-45ac-91cf-0929308714e9", "--nic1", "nat", "--nic2", "hostonly", "--hostonlyadapter2", "vboxnet0"]
INFO warden: Calling action: #<VagrantPlugins::ProviderVirtualBox::Action::ForwardPorts:0x00000100b07da8>
INFO interface: info: Forwarding ports...
[default] Forwarding ports...
INFO subprocess: Starting process: ["/usr/bin/VBoxManage", "showvminfo", "f31f231f-b99d-45ac-91cf-0929308714e9", "--machinereadable"]
INFO interface: info: -- 22 => 2222 (adapter 1)
[default] -- 22 => 2222 (adapter 1)
INFO subprocess: Starting process: ["/usr/bin/VBoxManage", "modifyvm", "f31f231f-b99d-45ac-91cf-0929308714e9", "--natpf1", "ssh,tcp,127.0.0.1,2222,,22"]
INFO warden: Calling action: #<Vagrant::Action::Builtin::SetHostname:0x00000100b07d80>
INFO warden: Calling action: #<VagrantPlugins::ProviderVirtualBox::Action::SaneDefaults:0x00000100b07d58>
INFO subprocess: Starting process: ["/usr/bin/VBoxManage", "storagectl", "f31f231f-b99d-45ac-91cf-0929308714e9", "--name", "SATA Controller", "--hostiocache", "on"]
INFO sanedefaults: Enabling the Host I/O cache on the SATA controller... (exit status = 0)
INFO subprocess: Starting process: ["/usr/bin/VBoxManage", "modifyvm", "f31f231f-b99d-45ac-91cf-0929308714e9", "--rtcuseutc", "on"]
INFO sanedefaults: Enabling rtcuseutc... (exit status = 0)
INFO sanedefaults: Automatically figuring out whether to enable/disable NAT DNS proxy...
INFO subprocess: Starting process: ["/usr/bin/VBoxManage", "modifyvm", "f31f231f-b99d-45ac-91cf-0929308714e9", "--natdnsproxy1", "on"]
INFO sanedefaults: Enable the NAT DNS proxy on adapter 1... (exit status = 0)
INFO warden: Calling action: #<VagrantPlugins::ProviderVirtualBox::Action::Customize:0x00000100c3ab08>
INFO warden: Calling action: #<VagrantPlugins::ProviderVirtualBox::Action::Boot:0x00000100c3aae0>
INFO interface: info: Booting VM...
[default] Booting VM...
INFO subprocess: Starting process: ["/usr/bin/VBoxManage", "startvm", "f31f231f-b99d-45ac-91cf-0929308714e9", "--type", "headless"]
INFO warden: Calling action: #<VagrantPlugins::ProviderVirtualBox::Action::Customize:0x00000100c3aab8>
INFO warden: Calling action: #<Vagrant::Action::Builtin::WaitForCommunicator:0x00000100c3aa90>
INFO subprocess: Starting process: ["/usr/bin/VBoxManage", "showvminfo", "f31f231f-b99d-45ac-91cf-0929308714e9", "--machinereadable"]
INFO interface: info: Waiting for machine to boot. This may take a few minutes...
[default] Waiting for machine to boot. This may take a few minutes...
INFO subprocess: Starting process: ["/usr/bin/VBoxManage", "showvminfo", "f31f231f-b99d-45ac-91cf-0929308714e9", "--machinereadable"]
INFO ssh: Attempting SSH connnection...
INFO ssh: Attempting to connect to SSH...
INFO ssh: - Host: 127.0.0.1
INFO ssh: - Port: 2222
INFO ssh: - Username: vagrant
INFO ssh: - Key Path: /Users/macm/.vagrant.d/insecure_private_key
INFO subprocess: Starting process: ["/usr/bin/VBoxManage", "showvminfo", "f31f231f-b99d-45ac-91cf-0929308714e9", "--machinereadable"]
ERROR warden: Error occurred: Timed out while waiting for the machine to boot. This means that
Vagrant was unable to communicate with the guest machine within
the configured ("config.ssh.timeout" value) time period. This can
mean a number of things.
If you're using a custom box, make sure that networking is properly
working and you're able to connect to the machine. It is a common
problem that networking isn't setup properly in these boxes.
Verify that authentication configurations are also setup properly,
as well.
If the box appears to be booting properly, you may want to increase
the timeout ("config.ssh.timeout") value.
INFO warden: Beginning recovery process...
INFO warden: Recovery complete.
ERROR warden: Error occurred: Timed out while waiting for the machine to boot. This means that
Vagrant was unable to communicate with the guest machine within
the configured ("config.ssh.timeout" value) time period. This can
mean a number of things.
If you're using a custom box, make sure that networking is properly
working and you're able to connect to the machine. It is a common
problem that networking isn't setup properly in these boxes.
Verify that authentication configurations are also setup properly,
as well.
If the box appears to be booting properly, you may want to increase
the timeout ("config.ssh.timeout") value.
INFO warden: Beginning recovery process...
INFO warden: Calling recover: #<Vagrant::Action::Builtin::Call:0x0000010137f490>
INFO warden: Beginning recovery process...
INFO warden: Recovery complete.
INFO warden: Recovery complete.
ERROR warden: Error occurred: Timed out while waiting for the machine to boot. This means that
Vagrant was unable to communicate with the guest machine within
the configured ("config.ssh.timeout" value) time period. This can
mean a number of things.
If you're using a custom box, make sure that networking is properly
working and you're able to connect to the machine. It is a common
problem that networking isn't setup properly in these boxes.
Verify that authentication configurations are also setup properly,
as well.
If the box appears to be booting properly, you may want to increase
the timeout ("config.ssh.timeout") value.
INFO warden: Beginning recovery process...
INFO warden: Calling recover: #<Vagrant::Action::Builtin::Call:0x00000106b2e790>
INFO warden: Beginning recovery process...
INFO warden: Recovery complete.
INFO warden: Recovery complete.
ERROR warden: Error occurred: Timed out while waiting for the machine to boot. This means that
Vagrant was unable to communicate with the guest machine within
the configured ("config.ssh.timeout" value) time period. This can
mean a number of things.
If you're using a custom box, make sure that networking is properly
working and you're able to connect to the machine. It is a common
problem that networking isn't setup properly in these boxes.
Verify that authentication configurations are also setup properly,
as well.
If the box appears to be booting properly, you may want to increase
the timeout ("config.ssh.timeout") value.
INFO warden: Beginning recovery process...
INFO warden: Calling recover: #<Vagrant::Action::Builtin::Call:0x000001024b4558>
INFO warden: Beginning recovery process...
INFO warden: Recovery complete.
INFO warden: Calling recover: #<Vagrant::Action::Builtin::Call:0x000001024b45f8>
INFO warden: Beginning recovery process...
INFO warden: Calling recover: #<VagrantPlugins::ProviderVirtualBox::Action::Import:0x000001075b8738>
ERROR warden: Error occurred: Too many open files - getcwd
INFO warden: Beginning recovery process...
INFO warden: Calling recover: #<Vagrant::Action::Builtin::Call:0x000001024b4558>
INFO warden: Beginning recovery process...
INFO warden: Recovery complete.
INFO warden: Calling recover: #<Vagrant::Action::Builtin::Call:0x000001024b45f8>
INFO warden: Beginning recovery process...
INFO warden: Calling recover: #<VagrantPlugins::ProviderVirtualBox::Action::Import:0x000001075b8738>
ERROR warden: Error occurred: Too many open files - getcwd
INFO warden: Beginning recovery process...
INFO warden: Calling recover: #<Vagrant::Action::Builtin::Call:0x000001024b4558>
INFO warden: Beginning recovery process...
INFO warden: Recovery complete.
INFO warden: Calling recover: #<Vagrant::Action::Builtin::Call:0x000001024b45f8>
INFO warden: Beginning recovery process...
INFO warden: Calling recover: #<VagrantPlugins::ProviderVirtualBox::Action::Import:0x000001075b8738>
ERROR warden: Error occurred: Too many open files - getcwd
INFO warden: Beginning recovery process...
INFO warden: Calling recover: #<Vagrant::Action::Builtin::Call:0x000001024b4558>
INFO warden: Beginning recovery process...
INFO warden: Recovery complete.
INFO warden: Calling recover: #<Vagrant::Action::Builtin::Call:0x000001024b45f8>
INFO warden: Beginning recovery process...
INFO warden: Calling recover: #<VagrantPlugins::ProviderVirtualBox::Action::Import:0x000001075b8738>
INFO subprocess: Starting process: ["/usr/bin/VBoxManage", "showvminfo", "f31f231f-b99d-45ac-91cf-0929308714e9", "--machinereadable"]
INFO runner: Preparing hooks for middleware sequence...
INFO runner: 2 hooks defined.
INFO runner: Running action: #<Vagrant::Action::Builder:0x000001026cea50>
INFO warden: Calling action: #<VagrantPlugins::ProviderVirtualBox::Action::CheckVirtualbox:0x00000106977d98>
INFO base: VBoxManage path: VBoxManage
INFO subprocess: Starting process: ["/usr/bin/VBoxManage", "--version"]
INFO meta: Using VirtualBox driver: VagrantPlugins::ProviderVirtualBox::Driver::Version_4_2
INFO base: VBoxManage path: VBoxManage
INFO subprocess: Starting process: ["/usr/bin/VBoxManage", "list", "hostonlyifs"]
INFO warden: Calling action: #<Vagrant::Action::Builtin::Call:0x00000106977e38>
INFO runner: Preparing hooks for middleware sequence...
INFO runner: 2 hooks defined.
INFO runner: Running action: #<Vagrant::Action::Builder:0x000001086aeb28>
INFO warden: Calling action: #<VagrantPlugins::ProviderVirtualBox::Action::Created:0x0000010358a990>
INFO subprocess: Starting process: ["/usr/bin/VBoxManage", "showvminfo", "f31f231f-b99d-45ac-91cf-0929308714e9", "--machinereadable"]
INFO runner: Preparing hooks for middleware sequence...
INFO runner: 2 hooks defined.
INFO runner: Running action: #<Vagrant::Action::Warden:0x0000010712a5e0>
INFO warden: Calling action: #<Vagrant::Action::Builtin::Call:0x0000010712a540>
INFO runner: Preparing hooks for middleware sequence...
INFO runner: 2 hooks defined.
INFO runner: Running action: #<Vagrant::Action::Builder:0x00000106acbd48>
INFO warden: Calling action: #<Vagrant::Action::Builtin::DestroyConfirm:0x00000106acc608>
INFO runner: Preparing hooks for middleware sequence...
INFO runner: 2 hooks defined.
INFO runner: Running action: #<Vagrant::Action::Warden:0x00000100d29dc0>
INFO warden: Calling action: #<Vagrant::Action::Builtin::EnvSet:0x00000100d29d20>
INFO warden: Calling action: #<Berkshelf::Vagrant::Action::SetUI:0x00000100d29cf8>
INFO warden: Calling action: #<Berkshelf::Vagrant::Action::LoadShelf:0x00000100d29cd0>
INFO warden: Calling action: #<Berkshelf::Vagrant::Action::ConfigureChef:0x00000100d29ca8>
INFO warden: Calling action: #<Vagrant::Action::Builtin::ConfigValidate:0x00000100d29c58>
INFO warden: Calling action: #<VagrantPlugins::ProviderVirtualBox::Action::CheckAccessible:0x00000100d29c30>
INFO subprocess: Starting process: ["/usr/bin/VBoxManage", "showvminfo", "f31f231f-b99d-45ac-91cf-0929308714e9", "--machinereadable"]
INFO warden: Calling action: #<Vagrant::Action::Builtin::EnvSet:0x00000100d29c08>
INFO warden: Calling action: #<VagrantPlugins::ProviderVirtualBox::Action::CheckVirtualbox:0x00000100d29be0>
INFO base: VBoxManage path: VBoxManage
INFO subprocess: Starting process: ["/usr/bin/VBoxManage", "--version"]
INFO meta: Using VirtualBox driver: VagrantPlugins::ProviderVirtualBox::Driver::Version_4_2
INFO base: VBoxManage path: VBoxManage
INFO subprocess: Starting process: ["/usr/bin/VBoxManage", "list", "hostonlyifs"]
INFO warden: Calling action: #<Vagrant::Action::Builtin::Call:0x00000100d29bb8>
INFO runner: Preparing hooks for middleware sequence...
INFO runner: 2 hooks defined.
INFO runner: Running action: #<Vagrant::Action::Builder:0x00000108368090>
INFO warden: Calling action: #<VagrantPlugins::ProviderVirtualBox::Action::Created:0x0000010836d2c0>
INFO subprocess: Starting process: ["/usr/bin/VBoxManage", "showvminfo", "f31f231f-b99d-45ac-91cf-0929308714e9", "--machinereadable"]
INFO runner: Preparing hooks for middleware sequence...
INFO runner: 2 hooks defined.
INFO runner: Running action: #<Vagrant::Action::Warden:0x00000100df0218>
INFO warden: Calling action: #<VagrantPlugins::ProviderVirtualBox::Action::CheckAccessible:0x00000100df01a0>
INFO subprocess: Starting process: ["/usr/bin/VBoxManage", "showvminfo", "f31f231f-b99d-45ac-91cf-0929308714e9", "--machinereadable"]
INFO warden: Calling action: #<VagrantPlugins::ProviderVirtualBox::Action::DiscardState:0x00000100df0178>
INFO subprocess: Starting process: ["/usr/bin/VBoxManage", "showvminfo", "f31f231f-b99d-45ac-91cf-0929308714e9", "--machinereadable"]
INFO warden: Calling action: #<Vagrant::Action::Builtin::Call:0x00000100df0150>
INFO runner: Preparing hooks for middleware sequence...
INFO runner: 2 hooks defined.
INFO runner: Running action: #<Vagrant::Action::Builder:0x000001074e9640>
INFO warden: Calling action: #<VagrantPlugins::ProviderVirtualBox::Action::IsPaused:0x000001074ee5a0>
INFO subprocess: Starting process: ["/usr/bin/VBoxManage", "showvminfo", "f31f231f-b99d-45ac-91cf-0929308714e9", "--machinereadable"]
INFO runner: Preparing hooks for middleware sequence...
INFO runner: 2 hooks defined.
INFO runner: Running action: #<Vagrant::Action::Warden:0x00000108380e38>
INFO warden: Calling action: #<Vagrant::Action::Builtin::Call:0x00000100df0100>
INFO runner: Preparing hooks for middleware sequence...
INFO runner: 2 hooks defined.
INFO runner: Running action: #<Vagrant::Action::Builder:0x00000106b405d0>
INFO warden: Calling action: #<Vagrant::Action::Builtin::GracefulHalt:0x00000108390b80>
INFO runner: Preparing hooks for middleware sequence...
INFO runner: 2 hooks defined.
INFO runner: Running action: #<Vagrant::Action::Warden:0x0000010681ec80>
INFO warden: Calling action: #<VagrantPlugins::ProviderVirtualBox::Action::ForcedHalt:0x0000010681ec08>
INFO subprocess: Starting process: ["/usr/bin/VBoxManage", "showvminfo", "f31f231f-b99d-45ac-91cf-0929308714e9", "--machinereadable"]
INFO interface: info: Forcing shutdown of VM...
[default] Forcing shutdown of VM...
INFO subprocess: Starting process: ["/usr/bin/VBoxManage", "controlvm", "f31f231f-b99d-45ac-91cf-0929308714e9", "poweroff"]
INFO warden: Calling action: #<VagrantPlugins::ProviderVirtualBox::Action::PruneNFSExports:0x00000100d29b40>
INFO subprocess: Starting process: ["/usr/bin/VBoxManage", "list", "vms"]
INFO warden: Calling action: #<VagrantPlugins::ProviderVirtualBox::Action::Destroy:0x00000100d29b18>
INFO interface: info: Destroying VM and associated drives...
[default] Destroying VM and associated drives...
INFO subprocess: Starting process: ["/usr/bin/VBoxManage", "unregistervm", "f31f231f-b99d-45ac-91cf-0929308714e9", "--delete"]
INFO machine: New machine ID: nil
INFO base: VBoxManage path: VBoxManage
INFO subprocess: Starting process: ["/usr/bin/VBoxManage", "--version"]
INFO meta: Using VirtualBox driver: VagrantPlugins::ProviderVirtualBox::Driver::Version_4_2
INFO base: VBoxManage path: VBoxManage
INFO warden: Calling action: #<VagrantPlugins::ProviderVirtualBox::Action::CleanMachineFolder:0x00000100d29af0>
INFO subprocess: Starting process: ["/usr/bin/VBoxManage", "list", "systemproperties"]
INFO warden: Calling action: #<VagrantPlugins::ProviderVirtualBox::Action::DestroyUnusedNetworkInterfaces:0x00000100d29ac8>
INFO warden: Recovery complete.
INFO warden: Calling recover: #<Vagrant::Action::Builtin::Call:0x000001024b4710>
INFO warden: Beginning recovery process...
INFO warden: Recovery complete.
INFO environment: Running hook: environment_unload
INFO runner: Preparing hooks for middleware sequence...
INFO runner: 1 hooks defined.
INFO runner: Running action: #<Vagrant::Action::Builder:0x00000107155c40>
/Applications/Vagrant/embedded/gems/gems/vagrant-1.3.1/lib/vagrant/util/subprocess.rb:44:in `pwd': Too many open files - getcwd (Errno::EMFILE)
from /Applications/Vagrant/embedded/gems/gems/vagrant-1.3.1/lib/vagrant/util/subprocess.rb:44:in `execute'
from /Applications/Vagrant/embedded/gems/gems/vagrant-1.3.1/lib/vagrant/util/subprocess.rb:24:in `execute'
from /Applications/Vagrant/embedded/gems/gems/vagrant-1.3.1/plugins/providers/virtualbox/driver/base.rb:363:in `block in raw'
from /Applications/Vagrant/embedded/gems/gems/vagrant-1.3.1/lib/vagrant/util/busy.rb:19:in `busy'
from /Applications/Vagrant/embedded/gems/gems/vagrant-1.3.1/plugins/providers/virtualbox/driver/base.rb:362:in `raw'
from /Applications/Vagrant/embedded/gems/gems/vagrant-1.3.1/plugins/providers/virtualbox/driver/base.rb:292:in `block in execute'
from /Applications/Vagrant/embedded/gems/gems/vagrant-1.3.1/lib/vagrant/util/retryable.rb:17:in `retryable'
from /Applications/Vagrant/embedded/gems/gems/vagrant-1.3.1/plugins/providers/virtualbox/driver/base.rb:290:in `execute'
from /Applications/Vagrant/embedded/gems/gems/vagrant-1.3.1/plugins/providers/virtualbox/driver/version_4_2.rb:385:in `read_state'
from /Applications/Vagrant/embedded/gems/gems/vagrant-1.3.1/plugins/providers/virtualbox/provider.rb:70:in `state'
from /Applications/Vagrant/embedded/gems/gems/vagrant-1.3.1/plugins/providers/virtualbox/action/import.rb:36:in `recover'
from /Applications/Vagrant/embedded/gems/gems/vagrant-1.3.1/lib/vagrant/action/warden.rb:66:in `block in recover'
from /Applications/Vagrant/embedded/gems/gems/vagrant-1.3.1/lib/vagrant/action/warden.rb:63:in `each'
from /Applications/Vagrant/embedded/gems/gems/vagrant-1.3.1/lib/vagrant/action/warden.rb:63:in `recover'
from /Applications/Vagrant/embedded/gems/gems/vagrant-1.3.1/lib/vagrant/action/builtin/call.rb:62:in `recover'
from /Applications/Vagrant/embedded/gems/gems/vagrant-1.3.1/lib/vagrant/action/warden.rb:66:in `block in recover'
from /Applications/Vagrant/embedded/gems/gems/vagrant-1.3.1/lib/vagrant/action/warden.rb:63:in `each'
from /Applications/Vagrant/embedded/gems/gems/vagrant-1.3.1/lib/vagrant/action/warden.rb:63:in `recover'
from /Applications/Vagrant/embedded/gems/gems/vagrant-1.3.1/lib/vagrant/action/warden.rb:52:in `rescue in call'
from /Applications/Vagrant/embedded/gems/gems/vagrant-1.3.1/lib/vagrant/action/warden.rb:28:in `call'
from /Users/macm/.vagrant.d/gems/gems/vagrant-berkshelf-1.3.3/lib/berkshelf/vagrant/action/set_ui.rb:12:in `call'
from /Applications/Vagrant/embedded/gems/gems/vagrant-1.3.1/lib/vagrant/action/warden.rb:34:in `call'
from /Applications/Vagrant/embedded/gems/gems/vagrant-1.3.1/lib/vagrant/action/builtin/env_set.rb:19:in `call'
from /Applications/Vagrant/embedded/gems/gems/vagrant-1.3.1/lib/vagrant/action/warden.rb:34:in `call'
from /Applications/Vagrant/embedded/gems/gems/vagrant-1.3.1/lib/vagrant/action/builtin/call.rb:57:in `call'
from /Applications/Vagrant/embedded/gems/gems/vagrant-1.3.1/lib/vagrant/action/warden.rb:34:in `call'
from /Applications/Vagrant/embedded/gems/gems/vagrant-1.3.1/plugins/providers/virtualbox/action/check_virtualbox.rb:17:in `call'
from /Applications/Vagrant/embedded/gems/gems/vagrant-1.3.1/lib/vagrant/action/warden.rb:34:in `call'
from /Applications/Vagrant/embedded/gems/gems/vagrant-1.3.1/lib/vagrant/action/builder.rb:116:in `call'
from /Applications/Vagrant/embedded/gems/gems/vagrant-1.3.1/lib/vagrant/action/runner.rb:61:in `block in run'
from /Applications/Vagrant/embedded/gems/gems/vagrant-1.3.1/lib/vagrant/util/busy.rb:19:in `busy'
from /Applications/Vagrant/embedded/gems/gems/vagrant-1.3.1/lib/vagrant/action/runner.rb:61:in `run'
from /Applications/Vagrant/embedded/gems/gems/vagrant-1.3.1/lib/vagrant/machine.rb:147:in `action'
from /Applications/Vagrant/embedded/gems/gems/vagrant-1.3.1/lib/vagrant/batch_action.rb:63:in `block (2 levels) in run'
If you can't rebuild the Vagrant box, you can have Vagrant set the ulimit before it runs chef-solo.
Update your Vagrantfile to include:
config.vm.provision :shell, :inline => "ulimit -n 4048"
Make sure this occurs before the line:
config.vm.provision :chef_solo do |chef|
...
You may need to tweak the number, but this four times the default, so it should do the trick.
[Update] I have had issues with this working on different boxes. If you don't have time to make a new box, you can temporarily do this:
vagrant up
vagrant ssh
sudo su
ulimit -n 4048
exit
vagrant reload
vagrant provision
This is setting ulimit as root, reloading the virtual machine and rerunning Chef. The moment you run vagrant destroy, you will have to do this again.
INFO runner: Running action: #<Vagrant::Action::Builder:0x00000107155c40>
/Applications/Vagrant/embedded/gems/gems/vagrant-1.3.1/lib/vagrant/util/subprocess.rb:44:in `pwd': Too many open files - getcwd (Errno::EMFILE)
The above error message indicated that you've hit the max number of open files limit in the Terminal Session.
Run ulimit -a and see the open files value, by default (OS X 10.7 is 2560, Ubuntu 1024). You may need to set it to a much larger value and then run vagrant up again. Hopefully it'll work.
The easiest work around that I've found is to always run vagrant up in a fresh terminal window.