I am looking to export an entire sitecore page. Ideally, upon exporting the links would move from relative to absolutely, but that's not necessary. I want to download the html, css, images, everything from one of my pages. Any help would be greatly appreciated.
Assuming that you want to download one page only, you should be able to just use Save option in your browser. Most of the modern browsers supports all of your requirements. Sitecore web page is like any other html page.
If you want to download the whole site, you can try to use http://www.httrack.com/, wget or Firefox plugin as described here.
If none of those are enough, try to search for save entire site with images and css in your favourite search engine - there are plenty of other possibilites.
If I only have to locally save a page hosted in our Sitecore (really doesn't matter what's hosting it) I use wget.
This seems to work quite well:
wget.exe -E -H -k -K -p --no-check-certificate <your url>
You can change the parameters to recursively spider the whole site, if you do so make sure to check with your security department to not set of any alarms.
Related
Currently I am using Hugo on GitHub Pages with ease but I have a few problems in which you could possibly help me to solve. Thanks in advance if you take your time to care a bit about my question, I highly appreciate that. I am trying to be as short as possible.
My short introductory page is located here
My blog (which needs fixes) is located here
Both sites are set up via the Hugo Quick Start, therefore I used the submodule way of getting the themes instead of cloning.
Both sites are deployed into separate repositories. I have asked in other channels what would be the best and cleanest form of maintaining these sites, then some said, GitHub Project Pages is nice.
To reach this feature, I have set publishDir to docs in both projects
Set the master branch with /docs folder option at GitHub Pages in both cases.
The content of my introductory page loads fine while - by default - the CSS of my blog does not really want to load. (The browser wants to search it at the root - where nothing is located - instead of /Blog).
I have avoided the problem by building my Hugo site with hugo -t blog and it works fine, you can check on the link above.
When I tried to link this to Netlify, it did not work. Unfortunately, it did not build with the default version of Hugo they provide, but I was able to fix this with the creation of netlify.toml where I set 0.36 for Hugo’s version.
So after that at Netlify I had to provide a build command. There hugo should go, but it did not work. CSSs are not loaded. Unfortunately, providing hugo -t blog as build command did not help either.
What I would like to acheive:
To be able to build my site without parameters just as in the other case. :)
Repositories of the projects:
Homepage
Blog
Thanks,
Armand
The -t parameter instructs Hugo to use the specified theme. You can instruct Hugo to use the same theme via "theme" setting in your config file instead of the -t command line parameter.
I am using Joomla for my website when I search in google for SAPBuddy I always get search result.
Can some one help me I tried to add my side in Google webmaster, but the result is same.
sapbuddy.com/index.php/component/users/?view=reset
Check you've followed these steps.
Create a sitemap. You can use an online tool for this, or a Joomla extension
If you used an online tool to create your sitemap, upload the site the xml file to your server. If you're using an extension, follow their direction. When finished, you should be able to open it with your web browser, e.g. www.domain.com/sitemap.xml . Check if the indexed pages looke good and copy this URL
In webmaster tooks, add your domain then register this sitemap, pasting in your URL
After a short period, check back on Webmaster tools. It will show which pages have been indexed and if there were any errors.
Good luck!
I'm using fossil to manage some home projects and keeping notes in the wiki. After running like this for a few months, I'd like to at least try to use embedded documentation; mainly so as to be able easily to go back to previous versions.
I've studied the website page about managing project documentation which confirms that this is a technique I want to follow up, but I can't make out how to do it.
I've cut-and-pasted one of my wiki pages and added it to my fossil repo, but I can't work out where it should go in the directory structure to be accessible as described on the above page.
I've tried in a few places none of which worked. The document is currently %fossil-root%\doc\foo.wiki, (I'm on Windows), where %fossil-root% is the directory holding _ _FOSSIL__ (slighly mangled filename because of markdown), but having started a server with fossil ui, when I point my browser at http://localhost:8080/doc/foo.wiki, fossil presents me with a nicely formatted page saying it can't find index.html. I created /doc/index.html to see what would happen, but it made no difference.
Please can someone help me out, and/or point me to an example repository containing embedded documentation or another "how-to" document.
If your document is located in %fossil-root%\doc\foo.wiki, you can access it at the following URL:
http://localhost:8080/doc/trunk/doc/foo.wiki
This URL breaks down as follows:
http://localhost:8080 is the root URL to access Fossil when you run fossil ui
/doc signals that you want to access embedded documentation
/trunk indicates the checkin containing the documentation you wish to access
/doc/foo.wiki is the path of the document inside the repository
Instead of trunk, you can also specify a tag, or a branch name, or even a hexadecimal checkin identifier.
In the URL you were using, http://localhost:8080/doc/foo.wiki, foo.wiki is interpreted as the checkin name, and no document path is specified, which logically means Fossil won't find anything.
As for an example repository containing embedded documentation, the homepage of the Fossil website itself is a prime example:
https://www.fossil-scm.org/index.html/doc/trunk/www/index.wiki
where
https://www.fossil-scm.org/index.html is Fossil's root URL
/doc indicates a request for embedded documentation
/trunk indicates we want to fetch files from the trunk
/www/ is the path to the requested file inside the repository
index.wiki is the name of the file inside the repository.
So, in the 'trunk' branch of the repository, the file www/index.wiki contains the home page of the Fossil website.
You simply need to put the documentation under the %fossil-root%\www\ directory (or any other directory under version control) in your repository and then you can, for example, add the following line to your header's mainmenu section to link to it:
html "<a href='$home/doc/trunk/www/foo.wiki'>Documentation</a>\n"
As I said, it can be any directory under version control. To test this, pick any file in the repository, let's say a README file at the top level, and go to http://localhost:8080/doc/trunk/README. You should see the README file load up in your browser in a raw text format. By putting wiki or html files under a particular directory such as www you make it easy to organize the files that you specifically want rendered as documentation, which makes it easier to link to them.
http://www.fossil-scm.org/index.html/doc/trunk/www/embeddeddoc.wiki
After fossil 1.33, just prepare your document in the repository.
If the wiki file is put in
/doc/index.wiki
And use web browser to setup -> Admin -> Configuration.
There is a "Index Page" field, fill in your main index.html.
For example:
/doc/trunk/doc/index.wiki
Or if you just want the released version:
/doc/<version>/doc/index.wiki
everyone:
Recently,I meet a problem when I'm using Django to devlope my site.
In my project, one requirement is to view file online.
That is to say, one can browses files of formats:pdf, jpeg,doc and so on
directly in browsers such as chrome or firefox and so on.
I know this can be done by flash,FlexPaper is one solution. But in django,
not in flash,is there any other solutions? Anyone knows that? If not, I can
only use flexpaper and maybe site can't be accessed via apple devices
Many thanks!!
I think, you can use https://github.com/NetAngels/django-webodt (documentation) with option inline=True:
render_to_response(..., inline=True)
There are plenty of non-flash file browser plugins out there: Elfinder has a Django connector, and you can probably get an integration with CKfinder.
The django packages page also has a list of django file browsers link.
I want to make my website available offline even if the user clears the cache and cookies. Is is possible? Also I am dealing with database. Is is possible to handle databases offline?
A user could store a local copy of a single webpage using Chrome (right click save-as) and it will store all resources (images, css, js) required to fully load the page offline. Other browsers will have similar options.
You can use wget to mirror a whole website for offline browsing.
wget --mirror --convert-links --html-extension -p http://www.example.com/
of course neither of these options will handle database driven elements of your site/page.
If you want to mock a database or dynamic elements of a page offline then Google Gears is probably the closest to what you are looking for but I think it was deprecated by Google last year.
If your users have modern browsers, try HTML5 Application Cache.
References:
Overview - http://www.html5rocks.com/en/features/offline
Demo - https://jonathanstark.com/labs/app-cache-7/
Tutorial - https://www.html5rocks.com/en/tutorials/appcache/beginner/
Article - http://grinninggecko.com/developing-cross-platform-html5-offline-app-1/
Summary: Click me, I'm the newish thing that browsers now support!
I clicked some of the links found in other answers, and all tools mentioned are deprecated or will/should be soon.
Later when I wasn't connected to the internet, I opened a site operated by Google (either Google Docs or YouTube, I sadly forgot since then) and went to view the page source, as I was curious to see other answers in action. I found something called ORIGIN-TRIAL in the manifest file.
After a quick Google search, I found this, which brought me to this, which somehow brought me to the last link:
https://developers.google.com/web/fundamentals/primers/service-workers
In conclusion, use Service Workers now. If you're curious if it now works with all browsers, don't worry. All popular browsers should support it as seen here.
No, if your databases are housed online. then you need a internet connection for the PHP/ASP (whatever you're using to deal with DBs) to connect/communicate to the DB's
For storing data locally and accessing them offline take a look at Gears and Web Storage.
The main problem is what degree of functionality you want to provide with your website. It always requires some work on the client (user) side to "store" aka. save your website offline. You would have to store all your functionality in one page that the user stores (be it a Flash movie or some Javascript-Code).
You can use simple command to download whole website locally with all links working properly.
wget -rk 'http://www.website.com'
For https url you need to add one more property like below :
wget -rk --no-check-certificate 'https://www.website.com'