Load data to teradata from sas - sas

I am trying to load data to teradata from sas. The data should be ordered by a specific field.
There is an error because of the order by clause.
Does anyone know how I can do that?
Trying to upload ordered data to teradata

Related

Unable to load data into BigQuery US Dataset, but can load into East-Asia Dataset

So the issue is, I am unable to load data to GCP BigQuery in dataset 'dw' located in US location. However I am able to load the data in East-Asia location.
I am trying to load data to partitioned tables in dataset 'dw' (US location) using NiFi ingestion tool but no error and no data loaded. I even tried inserting manually from BigQuery editor, unfortunately no error and no data inserted into dw.aes_mapdata2.
However, I am able to load data to dataset TEST.aes_mapdata2_copy which is in location "asia-east1".
Any ideas on what I am doing wrong?
We figured it out. We had to create new tables. Seems we messed up the settings with the original tables we created.

Connecting Looker to SAS

Does anyone know if you can connect Looker directly to a SAS table/server? or would you use a work-around to upload the SAS data to a SQL server, then connect Looker to that?
The SAS dataset (analogous to a table) is in a proprietary, binary format. If you have SAS Access to OleDb or ODBC, I would upload the table from SAS and not bother using a SAS format.
You must have SAS Access to upload the dataset from SAS, however, otherwise you are facing how to read a SAS dataset outside of SAS issue. Let us know what you have, in terms of software, so we can help guide a solution.

Loading sas files into netezza on Aginity

I'm trying to load data with over 50 million records into Netezza using Aginity. The problem is that my data table is a sas format with extension .sas7bdat that I open on SAS Enterprise Guide. So currently I have a script that transforms the sas files into csv file that I can load into Netezza.
Is there a way that I can load these sas files directly into Netezza without having to convert it to csv? Does Aginity provide an easy method of doing this? Any help would be great, thanks!
I am not sure how can do it Netezza using Aginity but it very simple and easy in SAS enetrprise guideAs You said you have SAS eneterprise guide, please check whether you have SAS/ACCESS to Netezza. To check whether you have an SAS/ACCESS to Netezza then you can use below code in SAS enterprise guide
proc setinit; run;
If you have SAS/Access to Netezza, which probaly you may have it. To code this in SAS Enterprise guide is very easy and you can easily emulate in SAS eneterprise, by using examples shown in the link.
http://support.sas.com/documentation/cdl/en/acreldb/63647/HTML/default/viewer.htm#a003181092.htm
Byusing libname statement in your SAS enterprise guide and can use proc append or proc sql and also use bulk load as as shown in above link.

Open a .accdb file in sas enterprise miner

I have to perform data prediction using SAS Enterprise Miner. SAS only has an option to load a SAS table but the problem is i have a database with two tables that i have to use for the prediction. How do i get the two tables from the database into SAS Enterprise Miner.
I have not used Access or Miner in 10 years. However, I would suggest something like:
Does Access have the concept of a view? If so, then create the view in Access and then just use that.
or
Write a stand alone piece of code that pulls the tables from Access and does the join in SAS. I believe you can do this in Miner. If not, just do it in Enterprise Guide and save the table to a location you can pick it up with Miner.
or
Create a SAS view from the Access database that does the join and use that in Miner. Same as above, just save a view instead of a table -- allows you to update the Access database without having to recreate the table.
After looking into this problem I finally decided to use base SAS to read the database and extract the individual tables, save them into a SAS library, in that way they will be saved on the local drive then i could use them in Enterprise Miner.
/*Create a new library to store converted files*/
libname db 'C:\\Users\\Documents\\Data Sources';
/*import the first table from the database into the newly created library*/
proc import out = db.Table1
datatable = 'Table1'
dbms= ACCESS Replace;
database= "C:\\Users\\Documents\\Data Sources\\DBName.mdb";
usedate=yes;
scantime=no;
dbsaslabel=none;
run;
/*import the second table from the database into the newly created library*/
proc import out = db.Table2
datatable = 'Table2'
dbms= ACCESS Replace;
database= "C:\\Users\\Documents\\Data Sources\\DBName.mdb";
usedate=yes;
scantime=no;
dbsaslabel=none;
run;

Teradata Prepare object error

I am trying to access the SAS table which I made outside of Teradata Passthrough in working space for query run. Now it gives me an error. My question is how to access the table not in teradata inside the passthrough
proc sql;
connect to teradata (user="&user_id.#LDAP" password="&TERADATA_PASS" server='ABC'
connection=global database="GTY");
select * from connection to teradata(
select * from mm)
;
quit;
mm is not in teradata but made in working directory.
You probably can't access that in passthrough directly. You either need to run your query using libname access to the Teradata, or you need to put the information you need into a macro variable or text file that could be included in the passthrough query. In passthrough you can only access what you could access in an interactive Teradata session - so unless you have SAS defined as an ODBC or such provider for Teradata, it's a no go.
Typically what I do in this case is first try to execute the entire process through libname access, and if that fails (either because of execution time or because of a need for passthrough-only elements like stored procedures) then I use libname access to load the table to a table inside the RDBMS (Teradata here). Then it's available in your passthrough session for use (as a native Teradata table).