SSH connection problem after installing SSL in Google Cloud VM - ssl

I installed SSL Let's Encrypt on Google Cloud VM and now I can't access SSH connection. I can't even access it via https://console.cloud.google.com. I checked the Firewall rules, and the documentation, but I couldn't find what caused this.

These are the possible reasons why the VM instance have been inaccessible via SSH :
If there are no issues with the firewall rules, you can check if port 22 is open inside the VM instance. You can try using an online port checker to verify it.
After installing Let's Encrypt, your VM instance might have exhausted it's resources causing the service for port 22 to be terminated. I suggest you try rebooting/restarting the VM instance.
Upgrade the machine type of your VM instance incase you encounter the issue re-occurs.

Related

google colab ssh connection timesout

I am trying to follow the numerous tutorials and gists such as:
https://gist.github.com/creotiv/d091515703672ec0bf1a6271336806f0
https://stackoverflow.com/questions/48459804/how-can-i-ssh-to-google-colaboratory-vm/53252985#53252985
When I run the steps, it seems like everything went fine (I get the root password), but I do see this:
invoke-rc.d: could not determine current runlevel
invoke-rc.d: policy-rc.d denied execution of start.
but unfortunately, after all the steps when I do the following on my local machine:
ssh -p17057 root#0.tcp.ngrok.io
I get:
ssh: connect to host 0.tcp.ngrok.io port 17057: Connection timed out
I am on vanilla Debian Buster - any pointers to why this is happening would be incredibly useful to debug
thank you.
I also had trouble ssh'ing into the ngrok tcp tunnel. I was using my local laptop as the access point to the colab VM. What I did was fire up an EC2 instance on AWS and use that instead. Also, I used ssh reverse tunnel and dropped the need for ngrok altogether since the ec2 machine already had a public IP. Check my answer here: https://stackoverflow.com/a/63186681/4126114

Bitnami WildFly 10.0.1 on Google Compute Engine can't access management interface

I have recently launched the Bitnami Wildfly on Google Compute Engine. I have done all the configuration and everythings is working fine except for the management interface which give an error message:
The management interface could not be loaded.
Authentication required.
I know i have to create a tunnel via SSH with Putty, I did that and I am able to access the server terminal. The Tunnel i created is to allow connections via localhost:9999 from the remote port 9990.
I have also tried editing standalone.xml and changed the interface to accept connection on , that too did not work.
Kindly assist.
Solved: It appears that when configuring the tunnel use 127.0.0.1:9990 instead of localhost:9990, I guess this is because the standalone.xml is configured to accept connections on 127.0.0.1.

ServiceControl doesn't seem to do anything

I have ServiceControl setup on it's own VM and configured to use SQL Transport which I have pointed at the SQL database that's currently being used for NServiceBus. I've opened up port 33333 in the firewall on both the VM and in Azure NSG. I also installed OpenSSH and opened those ports.
On my local machine, I've opened a tunnel to the ServiceControl VM and forwarded port 33333. The tunnel opens without issue. I am running ServiceInsight on my local machine and have it connect to localhost:33333 which is forwarded to the remote VM. It connects without any reported error.
That's it. There is no data displayed in ServiceInsight. No endpoints in the endpoint explorer or anything.
Did I miss something? I'm not sure how to troubleshoot this.

Cannot connect to VM in bluemix UK Area via SSH

Has anybody tried out virtual machines in the UK area of bluemix?
I am able to start a vm but get an timeout when i try so connect to the vm via ssh.
I used the std Debian image that can be choosen on setup time and injected an ssh key for connecting. The security group I used was allow_all.
When trying to ping or to connect via ssh directly or the openstack cli the connection times out.
regards
Johannes
There was a bug in the setup of the vm, so I actually had no chance to access it.

SSH over VPN connection

We have an AWS EC2 server that we've configured to be only accessible (via SSH) from within our office network. Obviously this isn't ideal for remote arrangements where someone has to connect to the EC2 instance and is working remotely outside the office such as during a business trip.
I've managed to set-up a VPN through PPTP and can connect to the office network (I have two local IP's one from wlan0 and one from ppp0) regardless of anywhere I am. However, when I SSH to the EC2 instance, it's still rejecting me most likely because it sees that I'm still trying to ssh from outside the network.
I think I'm missing something very important. What do you think, what am I missing to to get the ssh to use the vpn connection rather than one in wlan/eth?
I'm starting to think this isn't possible so kindly let me know. My other option is to ssh to a machine within the office network and then use that machine to ssh to the EC2 instance but I've been hesistant to do that as it seems excessive.
Find your gateway through which you want to connect to your ec2-instance.
For Linux systems, route -n will show you the routing table. Now add the route
sudo route add -host gw
Just having a VPN to your office does not mean all traffic from your machine will go through the VPN - that depends on how you configure the VPN.
You can SSH from your machine to an office machine, and from there ssh to EC2 - that should work.
Alternatively, configure your system to route traffic for the ec2 instance through the VPN connection - doing this depends on your specific VPN configuration.