`solana airdrop 10 <pub-key>` Error: connection timed out - solana

I am running:
solana airdrop 10 7Fbdg5vecamm8MLxDw7bPN2xENaBE7fP65tKW935BEhq
Where 7Fbdg5vecamm8MLxDw7bPN2xENaBE7fP65tKW935BEhq is my public key.
However, I'm getting the following output:
Requesting airdrop of 10 SOL from 109.230.214.107:9900
Error: connection timed out
My config is:
$ solana config get
Config File: /Users/patrick/.config/solana/cli/config.yml
RPC URL: https://api.devnet.solana.com
WebSocket URL: wss://api.devnet.solana.com/ (computed)
Keypair Path: /Users/patrick/.config/solana/id.json

It looks like you did everything right, so it could be a few things:
You were requesting too much SOL all at once. Typically, you should airdrop 1 at a time.
The network was bogged down at the time. This can happen, especially on devnet
The Solana tool suite is out of date. Be sure to update to the latest version, which is 1.7.10 at the time of writing, using solana-install init 1.7.10
One of those should do the trick!

Related

Can't connect to Solana devnet/testnet

I've been trying to connect to solana devnet/testnet through cli and python solana SDK, but both without success, getting below error from CLI:
"Error: RPC request error: cluster version query failed: error sending request for url (https://api.devnet.solana.com/): operation timed out"
and from python SDK:
"requests.exceptions.ConnectionError: HTTPSConnectionPool(host='api.devnet.solana.com', port=443): Max retries exceeded with url: / (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f492ffa4490>: Failed to establish a new connection: [Errno 101] Network is unreachable'))"
I CAN connect to localhost using the same method with CLI. Strangely, I can't even find similar issues by googling, so I had to ask here by myself.
Any help will be much appreciated.

Why my pc is not conecting to services? (SSL)

All starts one day that my pc command line programs stop working, everyone has an error related to SSL, for example:
CURL: "failed to verify the legitimacy of the server ca bundle"
NPM: "UNABLE_TO_VERIFY_LEAF_SIGNATURE"
MONGODB: MongoError: connection 0 to xxx.mlab.com:xxx timed out
I am using the same code with mongo on an ec2 instance working ok, but local is not, i am using:
Windows 10
CURL 7.58
NPM 3.10
NODE 6.11.2
MONGODB 2.2.32
Any ideas?

What might cause the Kubernetes API server to fail to write the client CA configmap?

I'm experiencing that the Kubernetes API server fails to start during cluster bootstrapping with the following error log, apparently due to being unable to initialize its "client CA configmap":
E1029 14:35:56.211083 5 client_ca_hook.go:78] Timeout: request did not complete within allowed duration
F1029 14:35:56.211121 5 hooks.go:126] PostStartHook “ca-registration” failed: unable to initialize client CA configmap: timed out waiting for the condition
It seems to happen here in the Kubernetes source code. What might cause this error?
See the full log here.
Update: It seems that my etcd cluster isn't accessible from master nodes, even though the same command works from etcd member machines:
$ sudo ETCDCTL_API=3 etcdctl --cacert=/opt/tectonic/tls/etcd-client-ca.crt \
--cert=/opt/tectonic/tls/etcd-client.crt --key=/opt/tectonic/tls/etcd-client.key \
--endpoints=https://coreos-testing-etcd-0.socialfoodie.club:2379 \
endpoint health
https://coreos-testing-etcd-0.socialfoodie.club:2379 is unhealthy: failed to connect: grpc: timed out when dialing
Error: unhealthy cluster
I found out that despite the cryptic error message in the API server, the cause is that it can't write to the etcd cluster. The reason was that the API server was configured with a different client certificate authority than what the etcd cluster was using, due to a timing issue wrt. copying certificates in my Terraform cluster setup. I figured out that the CA was the problem by using curl to contact the etcd cluster instead of etcdctl, as it gave a clear error message.
Thanks to #johnharris85 for suggesting etcd connectivity being an issue!

Cannot access Gitlab repo from browser or from terminal

I cannot access Gitlab from the browser.
I am getting the following message:
This site can’t be reached
gitlab.com took too long to respond.
Try:
Checking the connection
Checking the proxy and the firewall
Running Windows Network Diagnostics
ERR_TIMED_OUT
I am not able to use git pull origin master and git push -u origin master from the terminal.
It is showing the following error message:
**fatal: unable to access "https://gitlab.com/myproject.git/" : OpenSSL SSL_connect: SS_ERROR_SYSCALL in connection to gitlab.com:443**
Help me to fix the isssue,
Thanks.
Check your DNS settings, can you reach other sites? This doesn't seem to be a gitlab specific problem.
Also, your remote URL seems quite weird. It's usually gitlab.com/user/repo.git
Try other sites, do they work?
Try acquiring the IP Address from gitlab using an online whois service. Does opening this IP work?
Adjust your origin:
git remote rm origin
git remote add origin "https://username#gitlab.com/username/password.git"
Try this and report back if something worked.

Cert already in hash table exception

I am using chef dk version 12 and i have done basic setup and uploaded many cookbooks , currently i am using remote_directory in my default.rb
What i have observed is whenever there are too many files /hierarchy in the directory the upload fails with the below exception :-
ERROR: SSL Validation failure connecting to host: xyz.com - SSL_write: cert already in hash table
ERROR: Could not establish a secure connection to the server.
Use `knife ssl check` to troubleshoot your SSL configuration.
If your Chef Server uses a self-signed certificate, you can use
`knife ssl fetch` to make knife trust the server's certificates.
Original Exception: OpenSSL::SSL::SSLError: SSL_write: cert already in hash table
As mentioned earlier connection to server isnt a problem it happens only when there are too many files/the hierarchy is more .
Can you please suggest what i can do? I have tried searching online for solutions but failed to get a solution
I have checked the question here but it doesnt solve my problem
Chef uses embedded ruby and openssl for people not working with chef
Some updates on suggestion of tensibai,
The exceptions have changed since adding the option of --concurrency 1 ,
Initially i had received,
INFO: HTTP Request Returned 403 Forbidden:ERROR: Failed to upload filepath\file (7a81e65b51f0d514ec645da49de6417d) to example.com:443/bookshelf/… 3088476d373416dfbaf187590b5d5687210a75&Expires=1435139052&Signature=SP/70MZP4C2U‌​dUd9%2B5Ct1jEV1EQ%3D : 403 "Forbidden" <?xml version="1.0" encoding="UTF-8"?><Error><Code>AccessDenied</Code><Message>Access Denied</Message>
Then yesterday it has changed to
INFO: HTTP Request Returned 413 Request Entity Too Large: error
ERROR: Request Entity Too Large
Response: JSON must be no more than 1000000 bytes.
Should i decrease the number of files or is there any other option?
Knife --version results in Chef: 12.3.0
Should i decrease the number of files or is there any other option?
Ususally the files inside a cookbook are not intended to be too large and too numerous, if you got a lot of files to ditribute it's a sign you should change the way you distribute thoose files.
One option could be to make a tarball, but this makes harder to manage the deleted files.
Another option if you're on an internal chef-server is to follow the advice here and change the client_max_body_size 2M; value for nginx but I can't guarantee it will work.
I had same error and i ran chef-server-ctl reconfigure on chef server then tried uploading cookbook again and all started working fine again

Resources