How to point to a different database in oracle APEX v 5.x? - oracle-apex

I have Oracle APEX configured on my laptop pointing to Oracle express DB on my laptop also.
I want to point to a different database on another server (specifically Oracle eBusiness suite database). How could this be achieved?

Does all the data live in the other database? Or does most of the data live in your local database and you just need to pull a bit of data from the other database?
If you are building applications that interact primarily with the data in the Oracle eBusiness Suite database, you'd realistically want to install APEX (if it is not already installed there) in the Oracle eBusiness Suite database and build your APEX application there. If you are building applications that interact primarily with data in your local database and you just need to pull a bit of data from the eBusiness Suite database, you can create a database link in your local database that connects to the remote database and reference objects over the database link.

Related

connect oracle apex application to multiple databases

is there any simple way to connect an oracle apex application to another database other than apex admin database (like using WebLogic data source).
I read this but there are a lot of problem with that.
Run these commands
``
java -jar ords.war setup --database AML2
Enter number for [1] Basic [2] TNS [3] Custom URL [1]:3
Enter the Custom JDBC URL:sys/Syspass//192.168.1.1:1521/AML2
java -jar ords.war map-url --type base-path --workspace-id NEW_DB /NEW_DB AML2
``
And get success message after executing each command but when create and execute new workspace and app in the specified path (NEW_DB) get this error
The connection pool named: aml2_pu does not exist
how to resolve this problem or is there any simple or clear way to define connection in a file or WebLogic data source.
thanks for your attention
An APEX installation is in the database, it doesn't "connect to the database", there also isn't something like an "admin database" - as such, it cannot connect to other databases. Your apex application and the data for that application live in the same database instance.
However, if you want to work with data from other databases in your application you can of course do so. The "old way" is using database links, however, nowadays connecting over rest is the way to go.
The documentation you are referencing is for having ORDS pointing to multiple databases. It is possible to have a single ORDS installation point to multiple APEX instances - in a previous job we used to have dev/uat/int instance on the same ORDS instance. Each of those databases have their own apex installation.

Update SQLite database on disk

My Django application (a PoC, not a final product) with a backend library uses a SQLite database - read only. The SQLite database is part of the repo and deployed to Heroku. This is working fine.
I have the requirement to allow updates to this database via the Django admin interface. This is not a Django managed database, so from Django's point of view just a binary file.
I could allow for a FileField to handle this, overwriting the database; I guess this would work in a self-managed server, but I am on Heroku and have the constraints imposed by Disk Backed Storage. My SQLite is not my webapp database, but limitations apply the same: I can not write to the webapp's filesystem and get any guarantee the new data will be visible by the running webapp.
I can think of alternatives, all with drawbacks:
Put the SQLite database in another server (a "media" server), and access it remotely: this will severely impact performance. Besides, accessing SQLite databases over the network does not seem easy.
Create a deploy script for the customer to upload the database via the usual deploy mechanisms. Since the customer is not technically fit, and I can not provide direct support, this is unfeasible.
Move out of Heroku to a self-managed server, so I can implement this quick-and-dirty upload without complications.
Do you have another suggestion?
PythonAnywhere.com
deploy your app and you can easily access all of your files and update them and your Sqlite3 database is going to be updated in real time without losing data.
herokuapp.com erase your Sqlite3 database every 24 hours that's why it's not preferred for Sqlite3 having web apps

Microsoft CRM Dynamics GP Using database after CRM License Expires

my company wants to stop using GP but would like to retain all the historical data (for reporting) from the Great Plains 2010 system . Can someone please advise if I can just take a backup of the Great Plains database after the license expires and run reports in SSRS from the old database? Please advise what the best approach should be.
If your reports currently run correctly in SSRS you are good to go - there is no SSRS dependency on Dynamics GP (just the SQL database.)
You can easily test this scenario. Create a new SQL Server Instance. Take a backup of your existing GP company database(s) and restore them to the new SQL Server instance (configure security the same as your production instance.) Now, create copies of your reports and give them a new data source pointing to the new SQL Server instance. Verify that all of the copied reports run successfully.
This is the equivelent to not using GP any longer as there will be no application pointing at the restored databases - just plain old SSRS reports reading from a SQL database.

How to profile embed firebird database

Did I make wrong to use firebird database I don't know. It has lot's of good futures but I can't figure out why my query (stored procedure) didn't work.
Is there any profiler/monitoring tool for firebird?
Firebird database is working stand alone so it is embeded db. And It doesn't allow to connect with 2 users. If there is a profiler I wonder how it will connect while I'm executing my queries.
IBExpert and Database Worbench have stored procedure debugger
There is also many monitoring tools http://www.firebirdfaq.org/faq95/
I advice you to install server version if you want to have more than 2 users

Options for maintaining MySQL databases for a django development team

What are some options to avoid the latency of pointing local django development servers to a remote MySQL database?
If developers use local MySQL databases to avoid the latency, what are some useful tools to sync schema updates of the remote db with the local db and avoid manually creating, downloading, and loading dumps?
Thanks!
One possibility is to configure the remote MySQL database to replicate to the developers local machine - assuming you have control of the remote database's configuration.
See the MySQL docs for replication notes. Using MySQL replication the remote node would be the Master and the developer machines would be Slaves. The main advantage of this approach is your developer machines would always remain synchronized to the Master database. One possible disadvantage (depending on the number of developer machines you are slaving) is a degradation in the remote database's performance due to extra load introduced by replication.
It sounds like you want to do schema migrations. Basically it's a way to log schema changes so that you can update and even roll back along with your source changes (if you change a model you also check in a new migration that has up and down commands). While this will likely become an official feature at some point, there are several third-party solutions to choose from. It's really a personal preference, here are some popular ones:
South
Django Evolution
dmigrations
I use a combination of South for schema migrations, and storing JSON fixtures (or SQL dumps) of useful test data in the VCS repo for the project. Works pretty seamlessly.