My git
client repeatedly fails with the following error after trying to clone the repository for some time.
What could be the issue here?
Note: I have registered my SSH key with the GIT hosting provider
Receiving objects: 13% (1309/10065), 796.00 KiB | 6 KiB/s
fatal: The remote end hung up unexpectedly
This question is related to
git
I found my problem to be with the .netrc file, if so for you too then you can do the following:
Open your .netrc file and edit it to include github credentials.
Type nano ~/netrc
or gedit ~/netrc
Then include the following: *machine github.com
login username
password SECRET
machine api.github.com
login username
password SECRET*
You can include your raw password there but for security purposes, generate an auth token here github token and paste it in place of your password.
Hope this helps someone
This solved my problem:
git clone --depth=20 https://repo.git -b master
Increasing postBuffer size and maxRequestBuffer will help you in this problem. Just follow the steps.
steps:
1 .Open terminal or Git Bash and with "cd" go to the location where you wanted to clone repo.
2.Set compression to 0
git config --global core.compression 0
3.Set postBuffer size
git config --global http.postBuffer 1048576000
4.Set maxRequestBuffer size
git config --global http.maxRequestBuffer 100M
5.Now start clone
git clone <repo url>
6.Wait till clone get complete.
Thank you. Happy Coding !!!
Based on this answer, I tried following (with https url):
git clone --depth 25 url-here
git fetch --depth 50
git fetch --depth 100
git fetch --depth 200
...and so on
git fetch --unshallow
- and it's done.The process obviously takes much more time, but in my case setting http.postBuffer
and core.compression
didn't help.
UPD: I found out that fetching via ssh works for any repo size (discovered accidentally), done with git clone <ssh url>
, given you have created ssh keys. Once repo is fetched, I change remote address using git remote set-url <https url to repo>
It might be because of commits' size that are being pushed.. Breakdown the number of commits by the following steps:
git log -5
See the last 5 commits and you would know which ones are not pushed to remote. Select a commit id and push all commits up to that id:
git push <remote_name> <commit_id>:<branch_name>
NOTE: I just checked my commit which could have the biggest size; first pushed up till then. The trick worked.!!
The only thing that worked for me was to clone the repo using the HTTPS link instead of the SSH link.
None of the suggested solutions worked for me when cloning a repository via ssh
. However, I was able to clone using https
, then later changed the remote to ssh
via:
git remote set-url origin [email protected]:USERNAME/REPOSITORY.git
It may be as simple as a server problem. If using GitHub, check https://twitter.com/githubstatus. I saw this for the first time just now and discovered GitHub's having a wobble. A few minutes later it worked again fine.
Obs.: Changing http.postBuffer
might also require to set up the Nginx configuration file for gitlab to accept larger body sizes for the client, by tuning the value of client_max_body_size.
However, there is a workaround if you have access to the Gitlab machine or to a machine in its network, and that is by making use of git bundle
.
git bundle create my-repo.bundle --all
git clone my-repo.bundle
git remote set-url origin "path/to/your/repo.git"
git push
All the best!
I had a similar problem, but with a bamboo job. Bamboo was failing doing a local clone (local but over an SSH proxy) of a cached repository, I deleted the cache and after that it worked, but any time it tries to clone from the local cache there is a failure. Seems like a problem with bamboo's version of the SSH proxy not git per se.
I have the same error while using BitBucket. What I did was remove https from the URL of my repo and set the URL using HTTP
.
git remote set-url origin http://[email protected]/mj/pt.git
I had the same issue and it was related with a bad internet connection, so after try with some git configs, i've just disconnected from my network and connected again and it works!.
It seems that after connection lost (or the action that fires this situation), git is stuck.
I hope that it could be a help for someone more here.
Best,
This worked for me, setting up Googles nameserver because no standard nameserver was specified, followed by restarting networking:
sudo echo "dns-nameservers 8.8.8.8" >> /etc/network/interfaces && sudo ifdown venet0:0 && sudo ifup venet0:0
Same error with Bitbucket. Fixed by
git config --global http.postBuffer 500M
git config --global http.maxRequestBuffer 100M
git config --global core.compression 0
Wasted a few hours trying some of these solutions but eventually traced this to a corporate IPS (Instrusion Protection System) dropping the connection after a certain amount of data is transferred.
I got solution after using below command:
git repack -a -f -d --window=250 --depth=250
in /etc/resolv.conf
add the line to the end of the file
options single-request
I was facing this issue when cloning data (via HTTP) from remote git repo hosted on AWS EC2 instance managed by elastic beanstalk. The cloning itself was also done on AWS EC2 instance.
I tried all aforementioned solutions and their combinations:
http.postBuffer
http.maxrequestbuffer
git clone
and then git fetch --unshallow
- see fatal: early EOF fatal: index-pack failedpackedGitLimit
et al, see here: fatal: early EOF fatal: index-pack failedclient_max_body_size
to both big value and 0 (unlimited); setting proxy_request_buffering off;
options single-request
in /etc/resolv.confgit clone
After all of this, I was still facing the same issue over and over again, until I found that issue is in Elastic Load Balancer (ELB) cutting the connection. After accessing the EC2 instance (the one hosting git repo) directly instead of going through ELB I've finally managed to clone git repo! I'm still not sure which of ELB (timeout) parameters is responsible for this, so I still have to do some research.
UPDATE
It seems that changing Connection Draining policy for AWS Elastic Load Balancer by raising timeout from 20 seconds to 300 seconds resolved this issue for us.
The relation between the git clone
errors and "connection draining" is strange and not obvious to us. It might be that connection draining timeout change caused some internal changes in ELB configuration that fixed the issue with premature connection closing.
This is the related question on AWS forum (no answer yet): https://forums.aws.amazon.com/thread.jspa?threadID=258572
SOLVED WITH WIFI Router Setting :
I got same issue when I am in wifi with Settings PPPoE(auto login by wifi router).
Git download speed is very slow 15kb.
packet_write_wait: Connection to 17.121.133.16 port 22: Broken pipe fatal: The remote end hung up unexpectedly fatal: early EOF fatal: index-pack failed
Solution : 1. Changed setting to Dynamic IP, reboot wifi router. 2. From web browser login to Internet service provider portal (do not configure PPPoE , auto login from the wifi router).
After changing Git download speed is 1.7MiB.
I faced with this problem using git in Kubuntu. I've also noticed overall instability in networking and found a solution.
in /etc/resolv.conf add the line to the end of the file
options single-request
This fixed delays before every domain name resolution and git started to work like a charm after this.
For shared bandwidth try to clone when load is less. Otherwise, try with a high speed connection. If still does not work, please use below command,
git config --global http.postBuffer 2048M
git config --global http.maxRequestBuffer 1024M
git config --global core.compression 9
git config --global ssh.postBuffer 2048M
git config --global ssh.maxRequestBuffer 1024M
git config --global pack.windowMemory 256m
git config --global pack.packSizeLimit 256m
And try to clone again. You might need to change those settings according to your available memory size.
The http.postBuffer trick did not work for me. However:
For others experiencing this problem, it may be an issue with GnuTLS. If you set Verbose mode, you may see the underlying error look something along the lines of the code below.
Unfortunately, my only solution so far is to use SSH.
I've seen a solution posted elsewhere to compile Git with OpenSSL instead of GnuTLS. There is an active bug report for the issue here.
GIT_CURL_VERBOSE=1 git clone https://github.com/django/django.git
Cloning into 'django'...
* Couldn't find host github.com in the .netrc file; using defaults
* About to connect() to github.com port 443 (#0)
* Trying 192.30.252.131... * Connected to github.com (192.30.252.131) port 443 (#0)
* found 153 certificates in /etc/ssl/certs/ca-certificates.crt
* server certificate verification OK
* common name: github.com (matched)
* server certificate expiration date OK
* server certificate activation date OK
* certificate public key: RSA
* certificate version: #3
* subject:
* start date: Mon, 10 Jun 2013 00:00:00 GMT
* expire date: Wed, 02 Sep 2015 12:00:00 GMT
* issuer: C=US,O=DigiCert Inc,OU=www.digicert.com,CN=DigiCert High Assurance EV CA-1
* compression: NULL
* cipher: ARCFOUR-128
* MAC: SHA1
> GET /django/django.git/info/refs?service=git-upload-pack HTTP/1.1
User-Agent: git/1.8.4
Host: github.com
Accept: */*
Accept-Encoding: gzip
Pragma: no-cache
< HTTP/1.1 200 OK
< Server: GitHub.com
< Date: Thu, 10 Oct 2013 03:28:14 GMT
< Content-Type: application/x-git-upload-pack-advertisement
< Transfer-Encoding: chunked
< Expires: Fri, 01 Jan 1980 00:00:00 GMT
< Pragma: no-cache
< Cache-Control: no-cache, max-age=0, must-revalidate
< Vary: Accept-Encoding
<
* Connection #0 to host github.com left intact
* Couldn't find host github.com in the .netrc file; using defaults
* About to connect() to github.com port 443 (#0)
* Trying 192.30.252.131... * connected
* found 153 certificates in /etc/ssl/certs/ca-certificates.crt
* SSL re-using session ID
* server certificate verification OK
* common name: github.com (matched)
* server certificate expiration date OK
* server certificate activation date OK
* certificate public key: RSA
* certificate version: #3
* subject:
* start date: Mon, 10 Jun 2013 00:00:00 GMT
* expire date: Wed, 02 Sep 2015 12:00:00 GMT
* issuer: C=US,O=DigiCert Inc,OU=www.digicert.com,CN=DigiCert High Assurance EV CA-1
* compression: NULL
* cipher: ARCFOUR-128
* MAC: SHA1
> POST /django/django.git/git-upload-pack HTTP/1.1
User-Agent: git/1.8.4
Host: github.com
Accept-Encoding: gzip
Content-Type: application/x-git-upload-pack-request
Accept: application/x-git-upload-pack-result
Content-Encoding: gzip
Content-Length: 2299
* upload completely sent off: 2299out of 2299 bytes
< HTTP/1.1 200 OK
< Server: GitHub.com
< Date: Thu, 10 Oct 2013 03:28:15 GMT
< Content-Type: application/x-git-upload-pack-result
< Transfer-Encoding: chunked
< Expires: Fri, 01 Jan 1980 00:00:00 GMT
< Pragma: no-cache
< Cache-Control: no-cache, max-age=0, must-revalidate
< Vary: Accept-Encoding
<
remote: Counting objects: 232015, done.
remote: Compressing objects: 100% (65437/65437), done.
* GnuTLS recv error (-9): A TLS packet with unexpected length was received.
* Closing connection #0
error: RPC failed; result=56, HTTP code = 200
fatal: The remote end hung up unexpectedly
fatal: early EOF
fatal: index-pack failed
On MacOSX High Sierra the solution for me was:
brew install git-lfs
and my repository was cloned without any errors.
I was doing git push from my OS X El Capitan Mac. I was getting same error, I tried everything to fix, what I found on google/stackoverflow. As far as version is concerned I am using fairly latest version of github which is 2.7.4. I have create a project in my local system, and I wanted this to be public in my github account. Project size was not around 8MB. I noticed that when I was pushing some files of size around 1.5MB, it was pushing properly, but with large size failed for me, with same error,
Only option I had was to push changes in chunk of MB. Now I have pushed all changes. This is workaround for me until I get fix for this solution.
So you can also try pushing change in multiple commit. Or if you have multiple folder you can push changes by each folder (if folder size is not big).
Hope this will help you to continuous working on project.
If you are using https and you are getting the error.
I used https instead of http and it solved my problem
git config --global https.postBuffer 524288000
use ssh
instead of http
, it's not a good answer for this question but at least it works for me
I also had the same problem.The reason for this problem is as Kurtis's descriptions about GNUTLS.
If you have the same reason and your system is Ubuntu, you can solve this problem by installing the latest version of git from ppa:git-core/ppa
.The commands are as below.
sudo add-apt-repository ppa:git-core/ppa
sudo apt-get update
sudo apt-get git
I got the same issue, I fixed this with trial and error method. I changed the core.compression value until it works.
I started with "git config --global core.compression 1" after 3 attempts
"git config --global core.compression 4" worked for me.
Well, I wanted to push a 219 MB solution, but I had no luck with
git config --global http.postBuffer 524288000
And what's the point of having a 525 MB post buffer anyway? it's silly. So I looked at the git error below:
Total 993 (delta 230), reused 0 (delta 0)
POST git-receive-pack (5173245 bytes)
error: fatal: RPC failed; curl 56 SSL read: error:00000000:lib(0):func(0):reason(0), errno 10054
So git want's to post 5 MB, then I made the post buffer 6 MB, and it works
git config --global http.postBuffer 6291456
I had the same issue and was searching the solution around web. I found that our corporate routing in IPv6 is not maintained.
I turn off the (IPv6 option on the ethernet port in Windows 10) and there is no issue.
This is due the internet connectivity issue, i faced the same issue. I did a shallow copy of code using
git clone --depth 1 //FORKLOCATION
Later unshallowed the clone using
git fetch --unshallow
Faced same issue, try to merge with another branch and take a pull from them. It works for me same.
The tricks above did not help me, as the repo was larger than the max push size allowed at github. What did work was a recommendation from https://github.com/git-lfs/git-lfs/issues/3758 which suggested pushing a bit at a time:
If your branch has a long history, you can try pushing a smaller number of commits at a time (say, 2000) with something like this:
git rev-list --reverse master | ruby -ne 'i ||= 0; i += 1; puts $_ if i % 2000 == 0' | xargs -I{} git push origin +{}:refs/heads/master
That will walk through the history of master, pushing objects 2000 at a time. (You can, of course, substitute a different branch in both places if you like.) When that's done, you should be able to push master one final time, and things should be up to date. If 2000 is too many and you hit the problem again, you can adjust the number so it's smaller.
I had to remove the branch flag for the git clone
command.
It worked for me after I tried it.
git config --global http.postBuffer 2048M
git config --global http.maxRequestBuffer 1024M
git config --global core.compression 9
git config --global ssh.postBuffer 2048M
git config --global ssh.maxRequestBuffer 1024M
git config --global pack.windowMemory 256m
git config --global pack.packSizeLimit 256m
Thank you for all
Check your internet speed. Also check the following commands:
$ git config --global http.postBuffer 2M
$ git pull origin master
Source: Stackoverflow.com