This question is answered with "it contains assets for the preview canvas", which isn't enough information.
Does the preview assets folder give me any additional power over the preview canvas? If so, how can I utilize it?
Does the preview assets folder give me any additional power over the preview canvas? If so, how can I utilize it?
Preview Assets as it is seen below just by default registered development time only catalog of resources.
So you can store there any images, colors, files, ie any resources, which can be used in Preview Canvas only, for testing purpose. In example to not download one from internet, cloud, or fetch from database. Because Preview is for fast UI-only look & test, so data source is not important, so to test & tune UI you don't need to fetch external data but use locally stored test data.
You can add/name any other development time asset/folder in there as well.
Related
I am new to Google BigQuery.
I am now logged into bq console, but the resource window for tables and datasets is too small to navigate and is pinned into bottom left, and its not resizable, please help on how can i unpin the window to navigate it.
It looks like this.
Looks like what you want is something like below
There are many way to accomplish this. I will present two of them
Customize existing/available UI to fit your own specific needs using so called Bookmarklet
Bookmarklets are saved and used as normal bookmarks. As such, they are simple "one-click" tools which add functionality to the browser.
There is a wide usage of bookmarklets - one of which is to modify the appearance of a web page within the browser - which is exactly your use case.
You can create bookmarklet that will switch visibility of elements above the data navigator as in above example
if your needs are more sophisticated and you are looking for more perks than just hiding elements of UI - you can look into third party IDE for BigQuery
I can recommend one to check out - Goliath - the part of Potens.io Suite for BigQuery. You will find there everything you would expected from professional big data IDE tool.
It is free to use and is available on GCP Marketplace
Disclosure - I am part and lead of Potens.io Team (which is also clearly stated in my SO Profile)
Did you mean you want to unpin a project that has "Pin icon" in the screenshot?
If yes, You can unpin a project to clcik "UNPIN PROJECT" right side of bigquery console.
You have to click project(e.g bigquery-public-data) first to meet "UNPIN PROJECT".
I've looked for this across the web a few times, and I feel like this hasn't been asked exactly, or I may just be getting bogged down with the wrong syntax. Hoping to get an easy answer here (yes, you can't get this, is an acceptable answer).
The variations from the base CentOS image are listed here: Link to GCP
However, they don't actually provide a download for this image. I'm trying to get a local VM running in VMWare with this image.
I feel as though they'd provide this to their clients to make it easier to prepare for use of their product, but I'm not finding it anywhere.
If anyone could toss me a link to a pre-configured CentOS ISO with the minor changes, I'd definitely take that as an alternative. I'm just not confident in my skills with Linux enough to configure the firewall properly :)
GCP doesn't support Google-provied images for exporting. However, they support exporting images for custom images.
I don't have any experience about image exporting, but I think this works.
Create custom images
You can create custom images based on your GCE VM instance.
Go navigation -> Compute engine -> images page.
You can create custom image via disk or snapshot in this page.
Select one and create a custom image.
Export your image
After creating custom image successfully, Go custom image page and click "export" on upper side.
Select export format and GCS destination. then click export.
Now you have an image in the Google Cloud storage.
Download image file and import to your local VM machine.
I would like to add a Google Trends chart for a specific search term to my Google Data Studio report, but Trends is not an option in the Data Source list. I wasn't able to find an option to embed JavaScript either. Is it possible to add a Trends chart to a report in Data Studio? Thanks!
I am posting this workaround as it seems no similar solution has been provided since.
You can actually do this, using a small workaround:
Create the graph you want to embed using Google Trends.
Click the "embed" icon in the upper right corner of the graph, and copy the JS-code (for either desktop or mobile device)
Create a simple empty HTML-file using notepad or similar text editor. (including , , as per common standard). Place it in an empty folder on your hard drive.
Paste the Google Trends-embed code into the section of your HTML-file.
Go to https://app.netlify.com/drop and upload the whole folder (including your .html-file). Copy the direct link provided by Netlify. (note: Any other form of public hosting should work fine, this is just my personal preference)
In Google Data Studio, click "URL embed" and paste your direct link.
Voila!
(Note: As this is a direct graph link and not a data feed, it, unfortunately, won't let you filter or change settings. but if configured wisely before copying the embed code, should do the trick for any time range, year-on-year or similar needs.)
Hope this helps someone :)
You can use supermetrics.com that has a google trends (free) datasource and then import a common sheet into your dashboard, the only problem is that you wont be able to change the date range, meaning its only "one way"
Unfortunately, the Google Trends data connector has stopped working in Supermetrics. They use an unofficial Google API that has been faulty lately.
The connector was removed Dec 2018.
I was looking for an answer to this question through Google and the site but did not find the right answer.
The system opencart attribute whether it is possible to set values in an image?
version opencart - 2.1.0.2 (rs.1)
more:
I wanted to download the products via the file .xls. To do this, I installed the module export / import .xls. structure file:
Here, a column with a red background is an attribute, it must be the values of the image. This file shows the image link. Now, how can I upload the product up correctly with this file?
this is a readme.txt :
Export/Import Tool (V2.19) for OpenCart 2.0.x
=============================================
The Import/Export Tool allows the admin user to do a bulk export
of categories, products, options and attributes to an Excel spreadsheet file.
The spreadsheet file can be edited offline and then be re-imported to the OpenCart database.
New features include:
The Import can be incremental, that is, data is updated or added to the OpenCart server.
Or it can be non-incremental which means all old data is first deleted before the Import.
The Export can be limited to certain data ranges only for products and categories.
Multiple languages are now supported, too.
Requirements and Limitations
============================
Memory requirements can be quite high.
Not every shared web hosting account supports a high process memory usage.
Therefore, if you use a basic shared web hosting account,
no more than a few thousand products can be exported or imported at a time.
Use a more dedicated web hosting account if a higher number of products
are to be processed in one go. Or export and import multiple times in smaller batches.
Installation
============
In the OpenCart admin backend, do the following steps:
Step 1)
Go to Extensions > Extension Installer
Step 2)
Upload the opencart-2-0-0-x-export-import-multilingual-2-xx.ocmod.zip
Step 3)
Go to Extensions > Modifications
You should see any entry for this Export/Import tool
Step 4)
Click on the Refresh button (top right of the page)
Step 5)
Go to System > Users > User Group > Edit Administrator
Step 6)
Set access and modify permissions for 'tool/export_import'
That's it!
If during the install you get an error saying "Could not connect as ......"
while uploading this zipped extension via the Extension Installer,
you probably have the FTP support disabled from your hosting.
In that case you may try the following OpenCart Extension Installer fix first:
<http://www.opencart.com/index.php?route=extension/extension/info&extension_id=18892>
Further help and customized versions
====================================
This tool has been successfully tested for a standard OpenCart 2.0.x.
Don't use other Opencart versions with this module.
If you need a customized version of the Export/Import Tool,
let us know and we can create one for a charge.
You can contact us at <http://www.mhccorp.com>
I am working on a SP2013 project for a customer, and I need to set up a working environment for development, testing and production. Let's assume for the sake of simplicity that the work consists only of various customizations (lists, libraries, apps, themes etc.) and no code.
My setup is as follows:
The production environment is on some servers on the customer site
The test environment is set up in Azure
The development environment is on a virtual machine on my PC
Now, let's assume everything is set up correctly on each environment, and I want to be able to support the following tasks:
I do customizations on my dev environment, and want to deploy this in test for others to test, prefereably with existing data
After testing and QA, I want to deploy from test to production. This must of course only affect customizations, not existing data
Every now and then I would like to take a snapshot of the production environment and move it to test, so that the deployment of a new feature from development can be done as realistic as possible
I want to perform these tasks as smoothly and efficiently as possible, especially when deploying from dev to test which is done often. Deploying from test to production will not be done that often, and hence some more manual work will be tolerated.
I know of a few mechanisms that might be relevant:
Content deployment
Cross site publishing
Content database backup/restore
Save site as template, export wsp and import
(Last resort) Manually set up each customization by hand
Could some of you experienced SharePoint devs/admins make some recommendations as to which mechanism to use in which situation, when to not use it etc.? Are there other methods that should be mentioned? Remember that the three environments reside in separated physical environments, which will probably make a fully automated solution difficult. Would it make it easier if I set up the test environment on the customer site (i.e. part of the same farm)?
Another option depending on your specific customisations might be a third party tool. There are a number of them out there. ShareGate is one I have personally been using for migration work and seems very simple and effective for moving content around quickly between environments. Attunity Repliweb for SharePoint is another that might be worth looking at for the sort of development specific release work that you require.
As for native options, I am still finding my way as well but here are my suggestions :
Where possible I have used Visual Studio to create solution packages containing features to deploy pieces of functionality. A branding solution package for example might include several features that deploy your custom master pages, theme / look files, common JavaScript libraries and images.
Feature deployment makes it easy for you to deploy or remove functionality between environments and to reuse functionality between sites. Additionally you can add your Visual Studio solutions to a source control system such as VS Online or GitHub.
For one off sites I have created a dev site, configured it then used the built in SharePoint backup and restore to deploy it to prod. Subsequent changes have been created in dev and then manually applied to test and prod. Depending on the customisations this has been quite time consuming. You might combine this with a tool such as ShareGate to automate the deployment of individual artifacts such as a customized list from one environment to another.
For moving content around I have been using a combination of ShareGate for things such as documents alongside Boost Solutions Excel Import for handling list data. This allows me to export large amounts of list data to excel and easily reimport it into a new list which might be a copy that I have added new functionality to in preparation for replacing the old prod list or perhaps dev / test lists that I am populating before doing a full site backup to restore to production.
Good luck and hopefully some of these suggestions are useful to you ! I will be following this question as I am also interested to hear of better methods / habits for managing the SharePoint development cycle.
I finished setting up a development environment for a SharePoint 2013 production environment that I maintain. The last step was to move my production content to my development environment. I had to dig around a bit to find the PowerShell etc. Rather than go through that again next time, I decided to write a blog about it, so that I’d have all the steps in one place.
The first step is to back up the content database that you want to restore to development. To do this open up SQL Server Management Studio, right-click on the database you want to back up, hover over tasks, and select backup. You will be presented with the Back Up Database window. Make sure that your backup type is set to full, give the backup a name or stick with the default, and note or change the destination.
You can skip these steps if you have scheduled backups running and are able to access the backup drive. In that case just go grab a copy of the most recent full backup and copy it to your development SQL Server.
The next step is to restore the database to development. To do this open up SQL Server Management Studio in your development environment, right-click on the Databases folder, and select Restore Database. When presented with the Restore Database window, click on the Device Radio Button and click the ellipsis next to the text box. This will bring up the Select backup devices window. From there click Add, locate your backup file and click OK, click OK again to be returned to the Restore Database Window, and finally from there click OK. Now your database has been restored, and you are ready to add it to SharePoint.
If you don’t already have one with content in it that you don’t care about in it, create a new Web Application…
https://sharepointv15.wordpress.com/2012/07/24/create-a-web-application-in-sharepoint-2013/
Don’t worry about creating a site collection.
Now go to Central Admin and click on Manage content databases under Application Management.
rsd1
Make sure that the correct Web Application is selected. If it is not click on the drop down arrow next to the Web Application name, click change web application and select the correct Web Application in the window that you are presented with.
rsd2
Next click on the Content Database name
On the Manage Content Database Settings screen scroll down, click on the Remove Content Database check box, click OK on the warning pop up and click OK at the bottom of the screen.
rsd3
Now you’ll need to open up the SharePoint 2013 Management Shell as an administrator. To do this click on your start menu, click all programs, click on the Sharepoint 2013 folder, right-click the SharePoint 2013 Management Shell and select Run as Administrator.
From here you will run the Mount-SPContentDatabase cmdlet
Mount-SPContentDatabase “MyDatabase” -DatabaseServer “MyServer” -WebApplication http://sitename
Click below for details on this cmdlt…
http://technet.microsoft.com/en-us/library/ff607581.aspx
At this point you should be able to navigate to the web application URL and see the Site Collection that lives in the database you just mounted.
Note: This will work in SharePoint 2010 or SharePoint 2013. However, the database must be the same version of SharePoint that the farm you are trying to mount to is. If it is a lower version it will automatically try to upgrade it, so keep that in mind.
follow the below link.
https://sharepointv15.wordpress.com/2013/03/21/moving-content-between-environments/