Is it possible to add ODBC service to existing CF 2016 install? - coldfusion

We are migrating some code from CF 10 to CF 2016 virtual machines that needs to connect to a couple Access databases. I installed the necessary drivers and setup odbc datasources in Windows but have found that our CF 2016 VM was setup without the ODBC service and I have not been able to find a clear way to add it.
When I go to add/remove programs the only option is to uninstall CF and when I run a CF 2016 installer it will not let me go through the "Server configuration" process because an existing install already exists. Is there a standard approach for adding sub-components to a CF server that were not chosen on first install?

Looks like I found my answer after digging around for a while. The migration wizard needs to be run again to add the ODBC service but there could be various complications with this depending on permissions and other factors. I was not able to get this to work after a brief period of time so I am just going to wrap up my transition to using python for our limited Access needs but I do believe this to be the answer.
https://community.adobe.com/t5/ColdFusion/ColdFusion-11-ODBC-service/td-p/6207226
Here are the basic steps. See link above for various troubleshooting info from Charlie Arehart and others.
Navigate to adminconfig.xml at C:\ColdFusion11\cfusion\lib\ and open it with text editor (say notepad).
Change the value from “false” to “true” in runmigrationwizard
Change the value from “false” to “true” in odbc
Save the file and restart ColdFusion Service.
After restarting the service, access the cf admin url and you will get the migration wizard. Follow the onscreen wizard to continue.
I came upon the answer by way of a thread about getting Access data sources setup in CF11 and CF2016 which Charlie Arehart contributed to and linked over to info about the ODBC service setup.

Related

How to disable windows server 2016 automatic updates on GCE

We have multiple windows server 2016 running, will automatic updates on. I want to change the Windows Update mode to Manual.
The problem is when instance is created, out software installation fails many times due to running updates installation. I tried to set instance metadata 'disable-agent-updates' to TRUE. But it doesn't seem to be working.
Please help on how to modify windows update mode on creation and start of instance.
The quick answer is you cannot with Google provided Windows images.
You can create a Windows VM, configure as you desire and then create a new image that you use to launch new instances. This will accomplish your goal with some effort. This is standard practice for companies that manage/control patches and updates.
Note: I do not recommend disabling Windows Update from installing security patches on startup unless you understand the implications.

Installing Sitecore9.2 in AWS

Can anyone provide me answer to below query?
I wanted to install Sitecore9.2 on AWS, does the installation process requires SQL VMs?
or Can someone point me to right article to this
Thanks in advance.
From my own experience Sitecore XM can use AWS RDS for the Database. If that is a good idea you must know yourself. For the installation, the Sitecore 9 installation uses contained database that may broke the SIF installation, you can turn it on in AWS RDS or use normal database user account but you need a workaround. like first installing on SQL server and migrate to RDS, or installing manual without SIF. or adjust SIF
For more information see:
https://jeroen-de-groot.com/2018/07/19/deploying-sitecore-9-in-aws-rds/
https://sitecore.stackexchange.com/questions/11047/sitecore-9-installation-using-sql-active-directory-user/11063
https://sitecore.stackexchange.com/questions/13859/why-do-we-require-contained-database-for-sitecore-9
https://sheenumalhi.wordpress.com/2019/02/19/sitecore-9-with-aws-rds/

VS2017 Data Lake tools cannot submit U-SQL script to Azure

Using Visual Studio 2017 with Data Lake Tools v. 2.3.1000.1, I am unable to submit a U-SQL job directly to Azure. I only have the option to submit locally:
This is the case even though I am connected to Azure through the "Server Explorer" tool window, from which I can access my U-SQL databases, view jobs that were previously submitted to Azure, etc.:
Using Visual Studio 2015, I have no such issue:
Am I forgetting a setting or a property somewhere, or is this perhaps a bug in Data Lake Tools for VS2017?
Do you still have the issue?
I sometimes see this happen if I open an existing solution and the tabs are already open, and I have not yet logged in. I then log in and the drop down menu of the open window will not be refreshed.
I close and reopen the script and it normally shows up.
Another reason, that I think will be addressed soon, could be that you have a filter on your cloud view explorer on which subscriptions you expose. If you hid the subscription there with your ADLA account, you will not see it in the pull down.
In any case, please let us know if you still have the issue and none of the two suggestions help.

Connecting to Google Cloud Spanner from DBVisualizer

I've created a test cloud spanner instance and database have have been attempting to connect to it through DBVisualizer.
I have authenticated using the gcloud auth command, and have the driver set up within DBVisualizer.
The connection string I'm using is:
jdbc:cloudspanner://;Project=testapp;Instance=test-instance;Database=test-spanner;PvtKeyPath=/Users/userhome/.config/gcloud/application_default_credentials.json
However, when I try to connect I get the following error:
[Simba][SpannerJDBCDriver](100004) Failed to connect to Spanner: No NameResolverProviders found via ServiceLoader, including for DNS. This is probably due to a broken build. If using ProGuard, check your configuration
Is there anyway to get a connection from a DB Management Tool such as DB Visualizer?
I found a solution on MacOS at least. Copy the CloudSpannerJDBC42.jar and google-cloud-spanner-0.9.4-beta.jar to DBvisualizers lib folder. In the case of MacOS the location is:
/Applications/DbVisualizer.app/Contents/java/app/lib
Restart DBVisualizer and then you can connect.
I don't think DBVisualizer supports Cloud Spanner right now. See their documentation: https://www.dbvis.com/features/
As the product is still pretty new publicly, we'll hopefully be seeing more 3rd party support in the coming months.
I've run into similar problems with the driver supplied by Google, so I decided to develop my own. The driver has both a 'thin' version and a 'fat' version. The thin version is intended as a dependency to be included in Java applications you develop yourself. The thick version can be used for standalone purposes, such as these kind of connections. The thick version (and other) can be found here: https://github.com/olavloite/spanner-jdbc/releases
More information about the whole driver can be found on my GitHub page.
The driver does work with DBVisualizer. Follow these steps to set it up:
Download the driver and place it in your JRE/lib/ext directory (this is necessary because of dynamic loading of services done by the underlying Google Cloudspanner API). Make sure you place it in the lib/ext directory of the JRE you are actually using with DBVisualizer.
Open DBVisualizer and open Driver Manager. Click on Create a new Driver.
Give it the name Cloudspanner
URL format is jdbc:cloudspanner://localhost;Project=projectId;Instance=instanceId;Database=databaseName;PvtKeyPath=key_file
Driver class is automatically selected.
Close the Driver Manager and make a new connection using the new driver.

Publishing to ProGet and I can't see any packages

I'm trying to set up a NuGet server using ProGet and am hitting a brick wall when publishing a package and it doesn't appear in the feed. The package is written to disk and works in other NuGet feeds. Other packages also don't appear in the ProGet feed so I'm pretty certain that the package is fine and that the problem lies with ProGet.
I'm using the community version of ProGet, but I don't see why that would affect anything.
Any ideas are most welcome!
The ProGet service is responsible for indexing packages, so if it's not running, packages could be uploaded but not displayed in any feeds. Here are the common troubleshooting steps for this scenario:
Verify that the ProGet Windows service (INEDOPROGETSVC) is running.
Ensure that the user account hosting the ProGet service has access to the feed storage path. Since it is NETWORK SERVICE by default, it would not see your mapped drives, and may not have access to the UNC path where the packages are stored.
Try running the ProGet service interactively, i.e. stop the ProGet Windows service and run ProGet.Service.exe manually as a console application to see any live output. Remember to restart the ProGet service when you close the console application.
Check for feed indexing errors to see if there was a problem indexing a particular packages. I know in much older versions that single "poisoned" packages (bad .nuspec file, invalid directory structure, etc.) could halt the indexing altogether.
Thanks John Rasch, found that your first point gave me what i needed to look in the right direction.
I could not find the (INEDOPROGETSVC) service but I did find (ProGet Servie) - I restarted this and then refreshed the my feed and all the missing packages showed up.
My version of ProGet is v3.3.12
Thanks John.