-
-
Notifications
You must be signed in to change notification settings - Fork 9
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Feature: Repository size restrictions #12
Comments
Could we instead only take the first 100 files or similar? |
That should be simpler. Right. We can start with 1000 files. As I've tried the number before and worked fine. |
Also, skip super-large files. |
An issue with this would be, since we're getting the repository as a zip, we still need to fetch the whole repo and use only the first 1000 files. |
This is a similar performance bottleneck I encountered on designing the
|
Hey. The https://api.github.com/repos/open-sauced/ai endpoint provides us with the size of the size. We can introduce a check. 100 MB is what I was thinking. We can work the number. We're ignoring files that can't be read as UTF-8 like multi-media files and a few other extensions. Lines 102 to 109 in b183145
Lines 136 to 141 in b183145
A worker pool could be great. We'll have to monitor how the deployment performs once the feature is out to decide on the changes. |
Type of feature
🍕 Feature
Current behavior
Presently, GitHub repositories are not subject to any size restrictions, which can potentially lead to potential exploits and create performance bottlenecks.
Suggested solution
To address this we'll need to implement a cap on the repository size and total file count before indexing the repository.
Additional context
No response
Code of Conduct
Contributing Docs
The text was updated successfully, but these errors were encountered: