Komodo Edit and SSH Private Keyfile in Pageant for Amazon EC2 - amazon-web-services

I would like to use Komodo Edit to edit files on my Ubuntu Amazon EC2 instance (running Django).
According to this Old Nabble from 2008 (oh boy) Komodo Edit should support SSH authentication via Pageant keyfiles.
So, I imported my .pem keyfile in PuTTYGen, converted it to .ppk (no password) and loaded it into Pageant. I am able to use PuTTY just fine to SSH into my instance.
I can also use Notepad++'s reasonable NppFTP with the AWS instance by adding the server and using the original .pem file directly (NppFTP doesn't seem to use pageant.)
However, I would like to use Komodo Edit, so I loaded up Komodo Edit, went to Edit -> Preferences --> Servers, and put in my Public DNS address (ec2-174-129-xxx-xxx.compute-2.amazonaws.com) and my username which was required ('ubuntu').
When attempting to connect, however, I get a "Javascript Application Error: ''" (a seemingly empty error) from Komodo Edit. I can't find any sort of logs or console to watch the handshake (Notepad's NppFTP plugin had a nice one.)
Obviously I can just use NppFTP but I would like to get this feature working. Any ideas?

Use putty 0.60. According to this site there's an incompatibility between the versions you are using.

Related

'Could not get reply from fzputtygen' error when adding private key to FileZilla

I'm trying to access a DigitalOcean droplet through FileZilla. I created a keypair using 'ssh-keygen' and provided DigitalOcean with the public key. When I open FileZilla -> Edit -> Settings -> SFTP, I click 'Add key file...' and select my private key (starting '-----BEGIN OPENSSH PRIVATE KEY-----'). I then get a popup telling me the format isn't supported and asks me if i'd like to convert the file, to which I say yes and enter the password for the file which validates. I provide a new filename in the same directory and click save, which give me the 'Could not get reply from fzputtgen' error message.
I've tried re-downloading and installing FileZilla with no success but really have no idea where to go as the error is so vague. Hopefully somebody here has experience with this issue and can shed some light as to how I might go about fixing it. I'm sure I could create a keypair using PuTTY and skip this step, but it would be nice to have an explanation for why this is failing.
It seems 3.55 has the issue but works OK on 3.56.2.
Please refer to answer dated 27 Sep 2021 here: https://forum.filezilla-project.org/viewtopic.php?t=54091
It seems that the error "Could not get reply from fzputtygen" while import RSA key generated using ssh-keygen is a new bug in Mac computers, as reported at https://trac.filezilla-project.org/ticket/12494#no1.
For this specific issue, keep track (at above url) until this still open case gets solved.

Postman Monitoring request error "Error: NETERR: getaddrinfo ENOTFOUND localhost"

I am trying to figure out how to get monitoring to work in Postman.
I have written tests on the desktop client for get/create/put and everything works fine. I'm using a localhost address and port 5004 which is the port for the API.
http://127.0.0.1:5004/bookings
I have tried to change the proxy in setting to localhost and port 5004, I have tried to change it to 127.0.0.1:5004, I have tried to disable SSL on the desktop client. I am running the monitoring on using the desktop client from the browser, that doesn't work either.
I have also checked if my etc/host file contains 127.0.0.1 localhost and it does.
Not sure what else I can try, I would appreciate any help. :)
Accessible APIs:
Monitors require all URLs to be publicly available on the internet as
they run in the Postman cloud. A monitor cannot directly access your
localhost or run requests behind a firewall. However, to overcome this
issue, static IPs are available on Postman Business and Enterprise
plans.
https://learning.postman.com/docs/designing-and-developing-your-api/monitoring-your-api/intro-monitors/
you cannot use the monitor for inhouse and localhost websites, You could upgrade to enterprise or business plans and see if that helps
The issue is maybe you configured the environment variable and passed the correct value in the URL also
But while running the collection test class
didn't select the correct environment
Select the environment configured to run that collection, shown in attached screenshot
I faced this problem and the issue was my DNS address. After changing the DNS server it was solved.
I had the same problem, I had space between ip and :port =>0.0.0.0b:1111 in my env. I deleted the space and I can connect. => 0.0.0.0:1111.
I faced the same issue and it got solved by removing the env variable
what you should try is sign out, sign in again. make sure the environment variables are not empty.
also, try using the feature console.log(get environment variable name) it would be helpful.
make sure to click on persist all in the environment variable.
The key for me was to click on "persist all" in the environment variable
read the github
some comments which helped me resolve the issue

Google: Permission denied to generate login hint for target domain NOT on localhost

I am trying to create a Google sign-in and getting the error:
Permission denied to generate login hint for target domain
Before you mark this a duplicate, this is not the same as the question asked at Google sign in website Error : Permission denied to generate login hint for target domain because in that case the questioner was on localhost, whereas I am getting this error on the server.
Specifically, I have included the url of the server in the Authorized Javascript Origins, as in the following image:
and when I get the error, the request shows that the same url was sent, as in the following image:
Is there something else I should be putting in my Restrictions page? Is there any way to figure out what is going on here? Is there a log at the developer console that can tell me what is happening?
Okay, I figured this out. I was using an IP address (as in "http://175.132.64.120") for the redirect uri, as this was a test site on the live server, and Google only accepts actual urls (as in "http://mycompany.com" or "http://localhost") as redirect uris.
Which, you know, THEY COULD HAVE SAID SOMEWHERE IN THE DOCUMENTATION, but whatever.
I know this is an old question, but it's the first result when you look for the problem via Google, so I'll share my solution with you guys.
When deploying Google OAuth service in a private network, namely some IP that can't be accessed via the Internet, you should use a magic DNS service, like xip.io that will give you an URL that your browser will resolve to your internal IP. You see, Google needs to be able to reach your authorized origin via your browser, that's why setting localhost works if you're serving it on your computer, but it won't work when you're deploying outside the Internet, as in a VPN, intranet, or with a tunnel.
So, the steps:
get your IP address, the one you're deploying at and it's not a public domain, let's say it's 10.0.0.1 as an example.
add http://10.0.0.1.xip.io to your Authorized Javascript Origins on the Google Developer Console.
open your site by visiting http://10.0.0.1.xip.io
clear your cache for the site, if necessary.
Log in with Google, and voilĂ .
I got to this solution using this answer in another question.
If you are using http://127.0.0.1/projects/testplateform, change it into http://localhost/projects/testplateform, it will work just fine.
If you testing in your machine (locally). then dont use the IP address (i.e. http://127.0.0.1:8888) in the Client ID configuration , but use the local host instead and it should work
Example: http://localhost:8888
To allow ip address to be used as valid javascript origin, first add an entry in your /etc/hosts file
10.0.0.1 mydevserver.com
and then add this domain mydeveserver.com in Authorized Javascript Origins. If you are using some nonstandard port, then specify it with your domain in Authorized Javascript Origins.
Note: Remove your cache and it will work.
Just ran across this same issue on an external test server, without a DNS entry yet. If you have permission on your local machine just edit your /etc/hosts file:
175.132.64.120 www.jimboweb.com
And use use http://www.jimboweb.com as an authorized domain.
I have a server in private net, ip 172.16.X.X
The problem was solved with app port ssh-forwarding to my localhost port.
Now I am able to use deployed app with google oauth browsing to localhost.
ssh -N -L8081:localhost:8080 ${user}#${host}
I also add localhost:8081 to "Authorized URI redirect" and "Authorized JavaScript sources" in console.developers.google.com:
google developers console
After battling with it for a few hours, I found out that my config in the Google Cloud console was all correct and similar to the answers provided. Due to caching issues or something, I had to recreate a OAuth Client ID and then it suddenly started working.
Its a pretty old issue, but I encountered it and there wasn't any helpful resource, as such I am posting my solution.
For me the issue was when I hosted my web-app locally, a using google-auth for logging in.
The URL I was trying to hit was :- http://127.0.0.1:8000/master
I just changed from IP to http://localhost:8000/master/
And it worked. I was able to log in to the website using Google Auth.
Hope this helps someone someday.
install xampp and run apache server,
put your files (index and co) in a folder in the xampp dir (c:\xampp\htdocs\yourfolder).
Type this in your browser url - http://localhost/yourfolder/index.html

Key length error logging into store on GREG 5.0 using SSO and custom Cert

We have been implementing GREG5.0 and using default configurations everything works fine. Once we replace the default localhost certificate in the wso2cabon.jks keystore with our own we receive "java.security.SignatureException: Signature length not correct: got 256 but was expecting 128" when we log into Store or Publisher using SSO.
We have removed the default keypair from wso2carbon.jks and added our own certificate. The password for our keystore and certificate are the same. We have updated all the configuration files per the wso2 carbon 4.4 documentation. We have updated JavaHome with local_policy.jar and us_export_policy.jar in order to allow for the longer key length.
The administrator console works great with no issues. If we change the login method of store or publisher to "basic" then it works fine. When we have the login method set to "SSO" we end up sitting on a blank page at this location https://servername/store/acs. We have the same result in the browser if we are running as a windows server or in console mode but, if we are running as a windows service then we have no error and no indication of what happened. If we are running in console mode then I get the error mentioned above spit out in the console.
I also noticed this behavior on Identity Server 5.0 when accessing dashboard.
We are running on windows.
Is there another location in WSO2 that I need to update to accomodate an increased key length?
Joe
The location I missed updating was the IdentityAlias in repository/deployment/server/jaggeryapps/store/config/store.json repository/deployment/server/jaggeryapps/publisher/config/publisher.json. Once I updated that value to match the alias of the keypair I was using in wso2carbon.jks that appeared as though it solved the keylength error and created another problem.
So now it was giving me a NullPointerException. I had provided the alias of our keypair but that was not the same as the alias for our certificate exported from our keypair that we loaded in client-truststore.jks. So I decided to set both alias' so they would match. With that change I was finally able to successfully able to access the store and publisher.
After some further testing it did not care what my keypair alias was as long as the value in IdentityAlias matched the alias of my certificate loaded in client-truststore.jks.
Hope this helps someone.
Joe

Coldfusion 11 Installation - invalid credentials

I just installed ColdFusion 11 on a Windows 2008 server. After installation, I got the message to proceed with configuration using the link
http://foo.com:80/CFIDE/administrator/index.cfm
where foo.com is 127.0.0.1
However, it keeps rejecting the userid and password I provided during installation for administrator.
Any help is highly appreciated.
This blog entry should help you: Troubleshooting ColdFusion Admin Login Issues.
You can also try the following quick workarounds:
Try an alternate browser (issue primarily occurs with IE).
Try alternative address
http://127.0.0.1:port_no/CFIDE/administrator/index.cfm or
http://localhost:port_no/CFIDE/administrator/index.cfm or
http://ip:port_no/CFIDE/administrator/index.cfm.
Try the passwordreset.bat and then try with the same ip or loopbak ip
address.
Enable internal/built port for ColdFusion and then try
http://localhost:8500/CFIDE/administrator/index.cfm, where 8500 is
the internal port for ColdFusion.
Note: Enabling internal/built port would require a ColdFusion Application Server restart.
This is usually due to reasons posted here - https://forums.adobe.com/message/6502639#6502639
The issue is happening because of mime type (.air) entry specified in the web.config of the webroot.
Can you try the following steps:-
Navigate to the webroot of your site.
Take a backup of web.config and open it with text editor (notepad etc)
Comment the entry for mimeType extension of ".air" like <!--<mimeMap fileExtension=".air"
mimeType="application/vnd.adobe.air-application-installer-package+zip"
/>-->
Save the file and restart IIS.
Launch CF Admin and try to login.
Commenting out the .air type in site specific web.config file has resolved the issue.