lfs | Docker configuration for building Linux From Scratch system | Continuous Deployment library
kandi X-RAY | lfs Summary
kandi X-RAY | lfs Summary
This repository contains docker configuration to build bootable iso image with Linux From Scratch 8.2.
Support
Quality
Security
License
Reuse
Top functions reviewed by kandi - BETA
Currently covering the most popular Java, JavaScript and Python libraries. See a Sample of lfs
lfs Key Features
lfs Examples and Code Snippets
Community Discussions
Trending Discussions on lfs
QUESTION
I'm running gitlab-ce on-prem with min.io as a local S3 service. CI/CD caching is working, and basic connectivity with the S3-compatible minio is good. (Versions: gitlab-ce:13.9.2-ce.0
, gitlab-runner:v13.9.0
, and minio/minio:latest
currently c253244b6fb0
.)
Is there additional configuration to differentiate between job-artifacts and pipeline-artifacts and storing them in on-prem S3-compatible object storage?
In my test repo, the "build" stage builds a sparse R package. When I was using local in-gitlab job artifacts, it succeeds and moves on to the "test" and "deploy" stages, no problems. (And that works with S3-stored cache, though that configuration is solely within gitlab-runner
.) Now that I've configured minio as a local S3-compatible object storage for artifacts, though, it fails.
ANSWER
Answered 2021-Jun-14 at 18:30The answer is to bypass the empty-string test; the underlying protocol does not support region-less configuration, nor is there a configuration option to support it.
The trick is able to work because the use of 'endpoint'
causes the 'region'
to be ignored. With that, setting the region to something and forcing the endpoint allows it to work:
QUESTION
I am trying to access the CREMA-D Dataset. I tried the following two ways in order to access the files in the dataset :
- I cloned it by typing the following command :
ANSWER
Answered 2021-Jun-02 at 16:04The main issue was due to the files actually being file pointers themselves. When I installed git-lfs
and cloned the repository via that, it all worked.
QUESTION
I want to create QR Code from user data. I have using below library for creating QR Code.
...ANSWER
Answered 2021-Jun-01 at 18:56The base64 value you provided in your question is malformed, my recommendation is to not use Json Serialization for this Api response effort.
Try using Convert.ToBase64String
- Ensure the method returns
string
- and the jQuery request accepts/expects
text
fordataType
response.
QUESTION
I have a local git repository where sync'ed with remote GitHub repository. On a feature branch, I needed to add/commit an large binary .pt file (236Mb) and then push it to remote origin in GitHub. Initially I added the file normally (git add), committed the file (git commit)and then tried to push (git push). The push to GitHub failed due to size of file and suggesting to use git-lfs. Following this error, my colleague pushed a .gitattributes file to remote master branch on GitHub with this content:
...ANSWER
Answered 2021-May-30 at 01:20It looks like you made some commits that have the big file, then you added some more commits in which you've replaced the big file with the LFS-indirection file. What this means is that you need to remove the commits that have the big files.
Remember that when you use Git-LFS, you're using two external web sites to store your data. First, you send the big files to some LFS storage site. These big files are not (yet) committed. Then, with the real files safely stored elsewhere, you create and push commits that store only the information needed to retrieve the big files from the LFS storage site. So the commits, in the form of a Git repository, exist somewhere (e.g., on GitHub) but don't have any big files in any of them. Instead, they have small files that have instructions that say don't use this file, use a big file you get with the LFS big-file-swapper-replacer-trick, here's how to get the big file. The LFS wrappers around Git intercept Git's attempt to put the small replacement file into your working tree, and put the big file there instead.
(Note that this means that anyone who clones your repository with regular Git gets, and sees, these weird small indirection files. They must set up Git-LFS themselves so that they can have the same wrapper program intercept the attempt to put the small files out for viewing or editing.)
But if you do this in the wrong order, you first commit the big files. Then you send the big files to a second site, then you remove the big files and put in the small "here's how to get the real file" files, and make another commit that contains these small files. You still have the commit with the big files in it!
Git is designed to add new commits, and not to take old commits away. This makes it very hard to remove any commit that has a big file in it. To do that automatically requires specialized tools. Fortunately, GitHub already have instructions and tools for you, to tell you how to do this. See them here. The short version is to use git lfs migrate
, if you have a modern Git-LFS. See also this other GitHub page.
QUESTION
I’m using git large files in one of my github private repositories and all was just good, but when I copy and paste the link for the raw data in an unlogged browser it was possible to see it. Is there a way to solve this problem?
In git large files page there is a topic saying “Keep the same access controls and permissions for large files as the rest of your Git repository when working with a remote host like GitHub.”
Is that true? Is there a way to configure the repo to hidden my information? I’ve tried to understand one of glfs tutorials but couldn’t find the ansewr.
...ANSWER
Answered 2021-May-26 at 01:23The link you've posted contains a token in the URL. That token exists to make it possible to view the raw URL even though the URL is on a different domain, and it contains credentials to permit someone to view that file.
Normally, the access to Git LFS files is restricted to the same access as the rest of your repository. Only parties who can read the files in your repository can get a valid raw data link like you've gotten, so as long as you don't distribute links to files in your private repository, you should be fine.
The token in the URL is specific to a file and a user, and usually it is sufficient to change your password to expire all the links.
QUESTION
I am trying to push a unity project to github. The project is 10GB in size but once gitignore is factored the size is more around 3GB which shouldn't be too big for github. I also have a .gitattributes file for use with git lfs (which I am pretty sure is configured right).
Each time I try to push to github, it hangs like so: look at last line. It will just stay like this forever. Seems like its stuck on something.
I looked up my issue and people have been doing things like changing the post buffer using "git config --global http.postBuffer 524288000" and it seems to work for them, but it never has for me. Any help would be appreciated.
...ANSWER
Answered 2021-May-25 at 02:31Ok, my bad on this one. My unity project is simply too big. I've settled for simply including the scripts for now in my GitHub repo as they house the majority of my work that I would be interested in sharing. I have misunderstood the usage of GitHub here, housing entire projects is something more for a site like itch.io as opposed to GitHub, which is more for source code
QUESTION
I have access to a GitLab repository that I need to migrate into GitHub. I need to be able to pull changes from the GitLab repository and push those to GitHub, while still working on the GitHub repository in a different new branch.
So, I pulled the repo from GitLab, added my GitHub remote upstream, created a new branch and tried to push the new branch to GitHub.
The problem is, they pushed a node_modules.zip
(in a subfolder) on one of the earliest commits into the GitLab upstream and this file is over 200MB in size.
I know that I need Git LFS to support those files, but since the files are already in the history of the GitLab upstream, I cannot migrate to Git LFS without losing the upstream, I think. I was thinking about just getting rid about those files, because I don't need need files that I can get with npm i
real quick.
Since I don't need the master branch on GitHub at all, I was hoping to rewrite the history on my new branch, but I have problems to figure out how this can be done.
The very important part is that I am able to pull new commits from GitLab, merge them to my own branch and push my own branch to GitHub.
Is this possible? If so, how?
...ANSWER
Answered 2021-May-20 at 06:31History, in any Git repository, is nothing more or less than the set of commits in that repository (as found by starting from some branch names, tag names, and/or other names and working backwards, the way Git does in general). Commits themselves hold full snapshots of every file, plus metadata; the metadata hold the hash IDs of previous commits; and the resulting chains of commits form Merkle trees which guarantee the validity of the commits.
The problem is, they pushed a
node_modules.zip
(in a subfolder) on one of the earliest commits into the GitLab upstream and this file is over 200MB in size.
Which of course exceeds the GitHub maximum size. So this commit cannot be sent to GitHub.
Since I don't need the master branch at all, I was hoping to rewrite the history on my new branch, but I have problems to figure out how this can be done.
The rewrite is not difficult (use a tool such as filter-branch, filter-repo, or the BFG). But once the rewrite is done, this is a completely new chain of commits: a different Merkle tree. It is a different history. It can be used with itself, but if it is combined with the original history, what you now have is a tree that requires both histories to be valid and complete.
The very important part is that I am able to pull new commits from GitLab, merge them to my own branch and push my own branch to GitHub.
The short version of all of this is "you can't".
The longer version is that you can, but only by maintaining a parallel history and never combining your history with their history. This is messy, ugly, painful, and difficult or impossible to automate. There are as far as I know no tools for doing this. You use one repository (e.g., on your own machine) that has both histories in it, and another (on GitHub) that has only your rewritten history. When they have new commits, you cherry-pick them to your own history. How you keep track of which commits are new-to-your-history is up to you: this is where a tool would come in handy. Such tools might exist, or you could write one. A barely-adequate one is probably not hard to write, but a good one would be hard.
QUESTION
Trying to access git-bash
prepared git repo with canonical git
, and I'm getting:
ANSWER
Answered 2021-Mar-30 at 00:58In general, the http.sslBackend
option is only usable on Windows. Most Linux distros don't offer it as an option, since they don't compile with multiple TLS libraries.
The proper solution is to remove all of the http.sslBackend
options:
QUESTION
I am trying to push my repository that has a lot of .png
, .mp4
, .h5
files, that I can ignore.
ANSWER
Answered 2021-May-12 at 01:29A good real-world analogy is tricky here, but suppose you have a cement delivery service. You delivered, to some customer, 1000 cubic yards of cement, which have now set. Your customer calls you up and says they did not want that much cement: they only wanted ten cubic yards. So you have your trucks deliver them another 10 yards of cement. Did you improve their situation, or just make it worse?
Now, the reason for the above analogy is simple enough: Git commits only ever add new stuff to the repository. If you put a giant file into some commit, that giant file is in that commit forever. If someone didn't want that giant file, and you make a new commit that lacks the giant file, well, now you have two commits: one with the giant file, and one without. The second commit depends on, and therefore requires, the first commit.
No matter what else you do, if you keep adding on new commits, you're just making everything worse. What you need is not more cement (er, commits). You need some kind of jackhammer and cement-removal service (something to remove commits from your repository).
That's not all, though! While we already see that listing a file in .gitignore
has no effect on any existing commit—no existing commit can ever be changed, at all; you just have to get them hauled away and stop using them entirely to get rid of them—it also has no good effect on new commits, at least not yet. If a file is tracked, it has no effect on that file in future commits either.
So:
Your first job is to remove some commit(s). We generally do this with
git reset
. Thegit reset
command is destructive! Be very careful here: you probably should work on a clone of your repository, rather than the original.Then, having removed the bad commit(s), you may need to explicitly remove the large files, so that they are no longer tracked. A file is tracked, in Git, if the file is in Git's index. How do you know if a file is in Git's index? Well, you know because it's tracked. This is kind of a circular problem, as you can see. :-) But if you definitely want it not-tracked, you can run:
QUESTION
i'm using git lfs for putting large file in github. today i recieve this email:
...Git LFS has been disabled on your personal account because you’ve exceeded your data plan by at least 150%. Please purchase additional data packs to cover your bandwidth and storage usage:
https://github.com/account/billing/data/upgrade i don't want to purchase so i deleted all of my files from github to reduce the size. so now there is no file in github. and now i want to push a small file to github with the following command:
ANSWER
Answered 2021-May-06 at 07:39You are still getting the quota error most likely because you still have commits in one or more branches containing very large files. Keep in mind that your Git history is a snapshot of everything which has ever happened. Deleting the files from the current branch and pushing just means you won't add more very large content. The content already in the history would still be there.
To get an idea on how to get started removing the large files from you repo, see this helpful Stack Overflow question:
How to remove/delete a large file from commit history in Git repository?
Community Discussions, Code Snippets contain sources that include Stack Exchange Network
Vulnerabilities
No vulnerabilities reported
Install lfs
Support
Reuse Trending Solutions
Find, review, and download reusable Libraries, Code Snippets, Cloud APIs from over 650 million Knowledge Items
Find more librariesStay Updated
Subscribe to our newsletter for trending solutions and developer bootcamps
Share this Page