Power BI + Snowflake SSO - powerbi

Snowflake
PowerBI
Is it possible to use different warehouse in snowflake for data refresh in Power BI than warehouse used for dashboard development?
I'd like to have separate warehouse for data refresh in order to be able to track snowflake expenses.
Let's say I'm using SF_TUTS_WH as a warehouse for development, but I'd like to use PBI_WH as warehouse for data refresh in PowerBI. Is it possible to achieve this or do I have to change snowflake warehouse in PowerBI file before I'll load it to PBI online?
FYI: https://www.snowflake.com/blog/using-sso-between-power-bi-and-snowflake/

One option which will work for you is to have the ware house name paramterized for the reports. The value of the parameter can be updated from the service based on the requirements.Please check out the below video link to see how the parameters can be updated in the power bi service.
https://youtu.be/MJVSu47iMmo
Hope this helps
Thanks ,
Prasanth

Related

Is there any REST API to refresh specific table in a Power BI Dataset?

Like a refresh data menu in Power BI desktop. I have exported these datasets from SQL Server database by running stored procedure query.
Unfortunatly not.
The most specific API to refresh data is "Refresh Dataset In Group" but it will update the full dateset.
https://learn.microsoft.com/en-us/rest/api/power-bi/datasets/refresh-dataset-in-group
If your workspace is assigned to a Power BI Premium Capacity or Premium Per User you can use the XMLA endpoint programmatically to refresh just one table in a dataset.

Power BI Desktop.. Automatic Refresh and Published

I am rather new to Power BI, but I think I have a rather simple question.
Power BI is connected to Google BigQuery through import. Within PowerBI I make some transformations to create my visualisations. These visualisations create a report.
I publish the report and can then find it on https://app.powerbi.com/ in my Power BI environment.
The data in Google BigQuery is updated daily. Can I update the report online automatically without me having to press refresh in Power BI desktop? As the data gets updated automatically and I will not be changing the visualisations daily, I would think this is achievable.
If the answer is yes, how might I achieve this in the most simplest of terms?
Yes, you can. It is called scheduled refresh. After publishing your report to Power BI Service, go to Settings -> Datasets, select your dataset in the list, enter the credentials under Data source credentials section (if needed), and then configure the settings under Scheduled refresh section:

How to deploy Power BI reports and connect them to a single Power BI Dataset

As far as I know, deploying a Power BI report from Power BI Desktop results in two items, the report itself and the dataset. When deploying a new report using the same dataset, will deploy the new report and a second copy of the same dataset in Power BI Service. That is not what I wanted. To not confuse end users and other, I want only an unique dataset deployed.
I want to make use of Azure Devops deploying to Power BI Service in a Dev, Test and Prod way. The dataset will be an azure analysis services data model, but the principle should be the same. I need to reduce the dataset to be exactly one and all reports must relate to that data model. I have heard of a Rest API or powershell scripting that can come to a rescue here.
So if any of you have done this or know of a good article that describes how to do this, I would be grateful.
Regards Geir
The best option is to separate the Power BI report in the frontend and the backend. You create a file purely for the dataset if you are importing, no reports created on it. You can then create the reports, using the service connection to the dataset, or with Power BI desktop, in the connection to Power BI Dataset option. Both will use 'Live Connection' mode, so you cannot add any other data sources to the model, for example bring in a CSV file or SQL database.
If you are connecting to an Azure Analysis Service data model, you can use this approach, however as it is only a connection only, not a full fat dataset, it should not be an issue to have copies of the dataset, as it is just the connection. Having copies of the dataset is only an issue if you are importing data, then it is best to move things to data flows, and use the same front/back end method, and the planning around the scheduling of the dataflows then datasets
You can use the REST API to move reports and the datasets that they connect to, and move items to new workspaces. If you have Power BI Premium that has a life cycle tool to move items between dev/test/live workspaces
If you create a report in desktop and choose 'Power BI Dataset' as live connection to work over it - when you upload the report to the same workspace, it will only upload the report and connect to the same dataset
https://radacad.com/power-bi-shared-datasets-what-is-it-how-does-it-work-and-why-should-you-care#:~:text=A%20shared%20dataset%20is%20a%20dataset%20that%20shared%20between%20multiple,tenant%20in%20Power%20BI%20environment.

Import Azure Time Series Insights data into Power BI

We were implementing the Azure Time Series insights(TSI) for the data streaming through IoT hub and wanting to connect to Power BI for advanced visualization. There is a provision to build a custom query to connect to Power BI from Azure TSI but i guess there is a limitation of fetching a dataset for specific time frame and only for the time series ID selected for the view.
Is there a way we can get the latest data in Power BI as and when the data refreshes in Azure TSI and for additional number of time series ID which were not selected while building the query.
Cheers,
Amit
If you generate a query from TSI to feed into Power BI, you can manually edit it to include the TSIDs you want to see. If you set the "isSearchSpanRelative" parameter to true, the data should refresh in Power BI as it comes into TSI.

Combine several pbix models into a single repository

I have a collection of .pbix models that follow a similar structure, ie, have the same tables and relationships.
It is too complex to combine them all into a single .pbix.
Is there a way to upload all these tables into a single repository, like PBI Service dataflows or a data warehouse, or something similar.
And then get the data back to PBI Desktop and perform DAX calculations, visualizations and report.
Any suggestions/ ideas?
Thank you so much for helping!
You can publish them to Power BI Service, and then create separate reports, but using these published datasets as a data source.
See Connect to datasets in the Power BI service from Power BI Desktop.
After publishing your "model" reports to Power BI Online, start making a new blank report, but instead of getting the data from files/databases/etc., choose Power BI service as a data source and select the previously published dataset. After that, you can publish your report the same way, but in this case you can share one dataset (your model) between multiple reports.