- Feb 24, 2020
-
-
GitLab Bot authored
-
- Feb 19, 2020
-
-
GitLab Bot authored
-
- Feb 11, 2020
-
-
GitLab Bot authored
-
- Jan 30, 2020
-
-
GitLab Bot authored
-
- Jan 10, 2020
-
-
GitLab Bot authored
-
- Dec 19, 2019
-
-
GitLab Bot authored
-
- Dec 10, 2019
-
-
GitLab Bot authored
-
- Nov 25, 2019
-
-
Nick Thomas authored
-
- Oct 11, 2019
-
-
GitLab Bot authored
-
- Oct 02, 2019
-
-
GitLab Bot authored
-
- Sep 25, 2019
-
-
GitLab Bot authored
-
- Sep 01, 2019
-
-
Nick Thomas authored
-
- Aug 28, 2019
-
-
Jacob Vosmaer (GitLab) authored
-
- Aug 23, 2019
-
-
Jan Provaznik authored
-
- Aug 09, 2019
-
-
Stan Hu authored
This sanitizes some log messages to be consistent with CE. Full list of changes: https://gitlab.com/gitlab-org/gitlab-workhorse/blob/master/CHANGELOG
-
- Apr 24, 2019
-
-
Nick Thomas authored
-
- Apr 11, 2019
-
-
- Apr 10, 2019
-
-
Nick Thomas authored
-
- Apr 04, 2019
-
-
We're moving from using terminology related to terminals when we refer to Websockets connections in Workhorse. It's more appropiate a concept like channel.
-
- Apr 02, 2019
-
-
Patrick Bajao authored
-
Patrick Bajao authored
Add `GetArchiveRequest` to git-archive params. Modifies `Git::Repository#archive_metadata` to append `path` to `ArchivePrefix` so it'll not hit the cache of repository archive when it already exists.
-
Jan Provaznik authored
Adds a rake task which can be used for removing EXIF data from existing uploads.
-
- Feb 13, 2019
-
-
Nick Thomas authored
-
- Feb 06, 2019
-
-
Nick Thomas authored
-
- Feb 04, 2019
-
-
Nick Thomas authored
-
- Jan 31, 2019
-
-
Nick Thomas authored
LFS uploads are handled in concert by workhorse and rails. In normal use, workhorse: * Authorizes the request with rails (upload_authorize) * Handles the upload of the file to a tempfile - disk or object storage * Validates the file size and contents * Hands off to rails to complete the upload (upload_finalize) In `upload_finalize`, the LFS object is linked to the project. As LFS objects are deduplicated across all projects, it may already exist. If not, the temporary file is copied to the correct place, and will be used by all future LFS objects with the same OID. Workhorse uses the Content-Type of the request to decide to follow this routine, as the URLs are ambiguous. If the Content-Type is anything but "application/octet-stream", the request is proxied directly to rails, on the assumption that this is a normal file edit request. If it's an actual LFS request with a different content-type, however, it is routed to the Rails `upload_finalize` action, which treats it as an LFS upload just as it would a workhorse-modified request. The outcome is that users can upload LFS objects that don't match the declared size or OID. They can also create links to LFS objects they don't really own, allowing them to read the contents of files if they know just the size or OID. We can close this hole by requiring requests to `upload_finalize` to be sourced from Workhorse. The mechanism to do this already exists.
-
- Jan 22, 2019
-
-
Nick Thomas authored
LFS uploads are handled in concert by workhorse and rails. In normal use, workhorse: * Authorizes the request with rails (upload_authorize) * Handles the upload of the file to a tempfile - disk or object storage * Validates the file size and contents * Hands off to rails to complete the upload (upload_finalize) In `upload_finalize`, the LFS object is linked to the project. As LFS objects are deduplicated across all projects, it may already exist. If not, the temporary file is copied to the correct place, and will be used by all future LFS objects with the same OID. Workhorse uses the Content-Type of the request to decide to follow this routine, as the URLs are ambiguous. If the Content-Type is anything but "application/octet-stream", the request is proxied directly to rails, on the assumption that this is a normal file edit request. If it's an actual LFS request with a different content-type, however, it is routed to the Rails `upload_finalize` action, which treats it as an LFS upload just as it would a workhorse-modified request. The outcome is that users can upload LFS objects that don't match the declared size or OID. They can also create links to LFS objects they don't really own, allowing them to read the contents of files if they know just the size or OID. We can close this hole by requiring requests to `upload_finalize` to be sourced from Workhorse. The mechanism to do this already exists.
-
-
- Dec 11, 2018
-
-
Nick Thomas authored
-
- Dec 10, 2018
-
-
Andrew Newdigate authored
-
- Dec 06, 2018
-
-
- Dec 04, 2018
-
-
Stan Hu authored
-
- Nov 30, 2018
-
-
Nick Thomas authored
-
- Nov 28, 2018
-
- Nov 23, 2018
-
-
Mark Chao authored
-
- Nov 14, 2018
-
-
Stan Hu authored
This version now supports redis[s]:// URL schemes.
-
- Nov 07, 2018
-
-
Andrew Newdigate authored
-
- Oct 01, 2018
-
-
Alejandro Rodríguez authored
-
- Sep 27, 2018
-
-
Stan Hu authored
-
- Sep 06, 2018
-
-
Stan Hu authored
There were several issues: 1. With Google Cloud Storage, we can't override the Content-Type with Response-Content-Type once it is set. Setting the value to `application/octet-stream` doesn't buy us anything. GCS defaults to `application/octet-stream`, and AWS uses `binary/octet-stream`. Just remove this `Content-Type` when we upload new files. 2. CarrierWave and fog-google need to support query parameters: https://github.com/fog/fog-google/pull/409/files, https://github.com/carrierwaveuploader/carrierwave/pull/2332/files. CarrierWave has been monkey-patched until an official release. 3. Workhorse also needs to remove the Content-Type header in the request (https://gitlab.com/gitlab-org/gitlab-workhorse/blob/ef80978ff89e628c8eeb66556720e30587d3deb6/internal/objectstore/object.go#L66), or we'll get a 403 error when uploading due to signed URLs not matching the headers. Upgrading to Workhorse 6.1.0 for https://gitlab.com/gitlab-org/gitlab-workhorse/merge_requests/297 will make Workhorse use the headers that are used by Rails. Closes #49957
-