Where can I get the Power BI Dashboard in a Day data - powerbi

There is a in depth tutorial that leads you through building a dashboard. I can't get started because the first thing is to connect to a cloud resource that doesn't(?) exist.
Any help would be great.

You can find the files here.
Two notes about these files:
1) These are from June and might be outdated if you are looking at newer instructions.
2) To get around GitHub's size cap, I had to break a couple of .csv files up:
AU Sales.csv, which should have no effect on following along with the instructions.
bi_salesFact.csv, which will cause you to treat the US Sales data like the international data by importing it using the folder option instead of the file option

Related

How do I find ID to download ImageNet Subset?

I am new to ImageNet and would like to download full sized images of one of the subsets/synsets however I have found it incredibly difficult to actually find what subsets are available and where to find the ID code so I can download this.
All previous answers (from only 7 months ago) contain links which are now all invalid. Some seem to imply there is some sort of algorithm to making up an ID as it is linked to wordnet??
Essentially I would like a dataset of plastic or plastic waste or ideally marine debris. Any help on how to get the relevant ImageNet ID or suggestions on other datasets would be much much appreciated!!
I used this repo to achieve what you're looking for. Follow the following steps:
Create an account on Imagenet website
Once you get the permission, download the list of WordNet IDs for your task
Once you've the .txt file containing the WordNet IDs, you are all set to run main.py
As per your need, you can adjust the number of images per class
By default ImageNet images are automatically resized into 224x224. To remove that resizing, or implement other types of preprocessing, simply modify the code in line #40
Source: Refer this medium article for more details.
You can find all the 1000 classes of ImageNet here.
EDIT:
Above method doesn't work post March 2021. As per this update:
The new website is simpler; we removed tangential or outdated functions to focus on the core use case—enabling users to download the data, including the full ImageNet dataset and the ImageNet Large Scale Visual Recognition Challenge (ILSVRC).
So with this, to parse and search imagenet now you may have to use nltk.
More recently, the organizers hosted a Kaggle challenge based on the original dataset with additional labels for object detection. To download the dataset you need to register a Kaggle account and join this challenge. Please note that by doing so, you agree to abide by the competition rules.
Please be aware that this file is very large (168 GB) and the download will take anywhere from minutes to days depending on your network connection.
Install the Kaggle CLI and set up credentials as per this guideline.
pip install kaggle
Then run these:
kaggle competitions download -c imagenet-object-localization-challenge
unzip imagenet-object-localization-challenge.zip -d <YOUR_FOLDER>
Additionally to understand ImageNet hierarchy refer this.

Power Bi Desktop Error Message: Each original name must belong to the set of all names

I have a query that imports two sheets from an xlsx file, combines the data, unpivots it, and does other cleanup steps. For the resulting fact table I also defined relations to 3 further dim tables.
Then I decided to refactor the query in the PowerQuery editor, including splitting the query up into 3 steps and disabling load of the the first two. (To play it safe I also stored my pbix w/o applying the steps.)
But then after a Close & Apply I got the following cryptic error message:
Something went wrong.
Each original name must belong to the set of all names.
PBIDesktop version is 2.96.701.0 64-bit (August 2021)
Searching the web I found people suggesting that the error might be related to specific PBIDesktop versions and that moving back to a previous version would solve the issue.
For me it was easier than that, though a bit mysterious:
In PowerQuery I enabled load for all sub-queries,
let it apply all my steps and the error message was gone.
Then back in PowerQuery I could disable load for the sub-queries w/o triggering the error again!
I hope that this might help someone somewhen.

Power BI - "Expression.Error: The column of the table wasn't found

This error pops up randomly when I import data to Power BI using Supermetrics API about a day after when trying to refresh.
For example, using Adobe Analytics data, when I try to refresh my dashboard a day later, I sometimes get this error, saying
When I click on "Go to error" it takes me to Changed type step, but I see that the code remains the same as I first imported it.
Going to Navigation step, I see that the data has changed, but I don't know why, since I haven't changed anything since I pulled data to Power BI.
What it used to look like:
vs what it looks like now
I have tried seeking help from Microsoft community, but no use, since most topics suggest changes to the code, which I haven't done.
This is 100% an issue in import, and as this issue has only just come up now it is due to a change in source format.
Looking at the breakdown of your imported data, it seems as though you are accessing a web source? This makes me think that the html make-up of that webpage has changed.
Unfortunately, the easiest way of dealing with something like this, is to rebuild the query from the ground up. If you are still struggling, let me know and I will have a look into it further.
Possibly, the 'Date' column is not arriving from the data feed (Supermetrics API) during your refresh. When you come across this issue, read the data feed in other Power BI window or check in the source so that you can confirm, the 'Date' column is missing during a particular point in time.
Thanks.
For me, I found that when my data source changed in terms of column header name, this error was thrown. You probably already sorted it out by now, but any future readers, I would double check the source to make sure no changes were made that might affect this, and then if the source is unchanged in it's version history, then rebuild the inquiry again. It sucks if this is due to a PBI issue, but these two methods should solve 99% of the instances of this error.

How can you view the code for the example Community Visualizations in Google Data Studio?

I am trying to make a Sankey chart (alluvial diagram) in Data Studio. I have found the "Community Visualizations" page and I can see the sankey diagram is one of the examples (https://developers.google.com/datastudio/visualization/). However, when I try to look in the bucket public-community-viz-showcase-reports, despite these supposedly being public it says I don't have the appropriate permissions to view them. I want to view the code used to generate the showcase report so I can modify it for my own purposes (I need to add color coding of the flows and multiple columns). Is it possible to do this?
Some of the files were uploaded to the Community Visualizations repository. The Sankey one was marked as experimental and deleted in this commit. The reason can be due to updating it to the new version but we aware of that if using it in production. Anyway, you can still browse through the repository history to find older files containing the original code. For example:
Sankey folder
index.js
Note that it also contains instructions on how to build the visualizations with the new changes you apply to the code.
By the way, even if you don't have storage.objects.list to run $ gsutil ls gs://public-community-viz-showcase-reports/sankey you do have storage.objects.get and can retrieve individual files of course. The problem in doing it that way is that files are minified to improve performance and not really readable.
As an example, an excerpt of index.js:
$ gsutil cat gs://public-community-viz-showcase-reports/sankey/index.js | head -c 500
is the following:
!function(t){var n={};function e(r){if(n[r])return n[r].exports;var i=n[r]={i:r,l:!1,exports:{}};return t[r].call(i.exports,i,i.exports,e),i.l=!0,i.exports}e.m=t,e.c=n,e.d=function(t,n,r){e.o(t,n)||Object.defineProperty(t,n,{enumerable:!0,get:r})},e.r=function(t){"undefined"!=typeof Symbol&&Symbol.toStringTag&&Object.defineProperty(t,Symbol.toStringTag,{value:"Module"}),Object.defineProperty(t,"__esModule",{value:!0})},e.t=function(t,n){if(1&n&&(t=e(t)),8&n)return t;if(4&n&&"object"==typeof t&&t
The files for several of the example community visualizations now lives in the experimental-visualizations repository.

How to access PSSE working case variables by Python a code

I am a transmission planning engineer and trying to automate the execution of PSSE 100 times or more at one go through a Python code. I already runs, change loads, reruns psse and write bus based summary report to *.csv file. What I really want to do is select the first active power load variable of a PSSE case and increase it by 1 MW. Then run psse, write results to a csv file. Change the selected load back to its original value and move on to the next active load to do the same again and again until I have done same for all load busses.
This will help me to calculate transmission loss factors for entire network with one go.
Thanks
#dsmtlk, if you're experienced in Python, you can readily find the information you need in the PSSE API Manual located in your PSSE program folder (mine is in C:\Program Files (x86)\PTI\PSSE33\DOCS). The API routines for getting bus data are in section 8.6. The routine for changing bus data—viz., psspy.load_data_4()—is in section 2.21.
If you're new to Python, here are a couple links I found helpful when I first started:
https://docs.python.org/2/tutorial/
http://www.tutorialspoint.com/python/