This document guides you through adding another package management system support to GitLab.
See already supported package types in Packages documentation
Since GitLab packages' UI is pretty generic, it is possible to add basic new package system support with solely backend changes. This guide is superficial and does not cover the way the code should be written. However, you can find a good example by looking at the following merge requests:
- npm registry support.
- Maven repository.
- Composer repository for PHP dependencies.
- Terraform modules registry.
- Instance-level endpoint for Maven repository.
The existing database model requires the following:
- Every package belongs to a project.
- Every package file belongs to a package.
- A package can have one or more package files.
- The package model is based on storing information about the package and its version.
Package systems work with GitLab via API. For example
implements API endpoints to work with npm clients. So, the first thing to do is to
add a new
lib/api/your_name_packages.rb file with API endpoints that are
necessary to make the package system client to work. Usually that means having
- GET package information.
- GET package file content.
- PUT upload package.
Since the packages belong to a project, it's expected to have project-level endpoint (remote) for uploading and downloading them. For example:
GET https://gitlab.com/api/v4/projects/<your_project_id>/packages/npm/ PUT https://gitlab.com/api/v4/projects/<your_project_id>/packages/npm/
Group-level and instance-level endpoints are good to have but are optional.
Packages are scoped within various levels of access, which is generally configured by setting your remote. A remote endpoint may be set at the project level, meaning when installing packages, only packages belonging to that project are visible. Alternatively, a group-level endpoint may be used to allow visibility to all packages within a given group. Lastly, an instance-level endpoint can be used to allow visibility to all packages within an entire GitLab instance.
Using group and project level endpoints allows for more flexibility in package naming, however, more remotes have to be managed. Using instance level endpoints requires stricter naming conventions.
The current state of existing package registries availability is:
|Repository Type||Project Level||Group Level||Instance Level|
|Conan||Yes||No - open issue||Yes|
|npm||No - open issue||Yes||No - open issue|
|NuGet||Yes||Yes||No - open issue|
|Go||Yes||No - open issue||No - open-issue|
npm is currently a hybrid of the instance level and group level.
It is using the top-level group or namespace as the defining portion of the name
NOTE: Composer package naming scope is Instance Level.
To avoid name conflict for instance-level endpoints you must define a package naming convention that gives a way to identify the project that the package belongs to. This generally involves using the project ID or full project path in the package name. See Conan's naming convention as an example.
For group and project-level endpoints, naming can be less constrained and it is up to the group and project members to be certain that there is no conflict between two package names. However, the system should prevent a user from reusing an existing name within a given scope.
Otherwise, naming should follow the package manager's naming conventions and include a validation in the
model for that package type.
Services and finders
Logic for performing tasks such as creating package or package file records or finding packages should not live within the API file, but should live in services and finders. Existing services and finders should be used or extended when possible to keep the common package logic grouped as much as possible.
GitLab has a
packages section in its configuration file (
It applies to all package systems supported by GitLab. Usually you don't need
to add anything there.
Packages can be configured to use object storage, therefore your code must support it.
The way new package systems are integrated in GitLab is using an MVC. Therefore, the first iteration should support the bare minimum user actions:
- Uploading a package
- Pulling a package
- Required actions
Required actions are all the additional requests that GitLab needs to handle so the corresponding package manager CLI can work properly. It could be a search feature or an endpoint providing meta information about a package. For example:
- For NuGet, the search request was implemented during the first MVC iteration, to support Visual Studio.
- For npm, there is a metadata endpoint used by
npmto get the tarball URL.
There are usually 2 phases for the MVC:
Keep iterations small
When implementing a new package manager, it is tempting to create one large merge request containing all of the necessary endpoints and services necessary to support basic usage. Instead, put the API endpoints behind a feature flag and submit each endpoint or behavior (download, upload, etc) in a different merge request to shorten the review process.
During this phase, the idea is to collect as much information as possible about the API used by the package system. Here some aspects that can be useful to include:
- Authentication: What authentication mechanisms are available (OAuth, Basic Authorization, other). Keep in mind that GitLab users often want to use their Personal Access Tokens. Although not needed for the MVC first iteration, the CI job tokens have to be supported at some point in the future.
- Requests: Which requests are needed to have a working MVC. Ideally, produce a list of all the requests needed for the MVC (including required actions). Further investigation could provide an example for each request with the request and the response bodies.
- Upload: Carefully analyze how the upload process works. This is likely the most complex request to implement. A detailed analysis is desired here as uploads can be encoded in different ways (body or multipart) and can even be in a totally different format (for example, a JSON structure where the package file is a Base64 value of a particular field). These different encodings lead to slightly different implementations on GitLab and GitLab Workhorse. For more detailed information, review file uploads.
- Endpoints: Suggest a list of endpoint URLs to implement in GitLab.
Split work: Suggest a list of changes to do to incrementally build the MVC.
This gives a good idea of how much work there is to be done. Here is an example
list that would need to be adapted on a case by case basis:
- Empty file structure (API file, base service for this package)
- Authentication system for "logging in" to the package manager
- Identify metadata and create applicable tables
- Workhorse route for object storage direct upload
- Endpoints required for upload/publish
- Endpoints required for install/download
- Endpoints required for required actions
The analysis usually takes a full milestone to complete, though it's not impossible to start the implementation in the same milestone.
In particular, the upload request can have some requirements in the GitLab Workhorse project. This project has a different release cycle than the rails backend. It's strongly recommended that you open an issue there as soon as the upload request analysis is done. This way GitLab Workhorse is already ready when the upload request is implemented on the rails backend.
The implementation of the different Merge Requests varies between different package system integrations. Contributors should take into account some important aspects of the implementation phase.
The MVC must support Personal Access Tokens right from the start. We currently support two options for these tokens: OAuth and Basic Access.
OAuth authentication is already supported. You can see an example in the npm API.
Basic Access authentication
support is done by overriding a specific function in the API helpers, like
this example in the Conan API.
For this authentication mechanism, keep in mind that some clients can send an unauthenticated
request first, wait for the 401 Unauthorized response with the
field, then send an updated (authenticated) request. This case is more involved as
GitLab needs to handle the 401 Unauthorized response. The NuGet API
supports this case.
There are project and group level permissions for
authorize the requesting user
against the project or group before continuing.
Database and handling metadata
The current database model allows you to store a name and a version for each package.
Every time you upload a new package, you can either create a new record of
or add files to existing record.
PackageFile should be able to store all file-related
information like the file
If there is specific data necessary to be stored for only one package system support,
consider creating a separate metadata model. See
Packages::Maven::Metadatum model as an example for package specific data, and
Packages::Conan::FileMetadatum model as an example for package file specific data.
If there is package specific behavior for a given package manager, add those methods to the metadata models and delegate from the package model.
Note that the existing package UI only displays information within the
tables. If the data stored in the metadata tables need to be displayed, a ~frontend change is required.
File uploads should be handled by GitLab Workhorse using object accelerated uploads. What this means is that the workhorse proxy that checks all incoming requests to GitLab intercept the upload request, upload the file, and forward a request to the main GitLab codebase only containing the metadata and file location rather than the file itself. An overview of this process can be found in the development documentation.
In terms of code, this means a route must be added to the GitLab Workhorse project for each upload endpoint being added (instance, group, project). This merge request demonstrates adding an instance-level endpoint for Conan to workhorse. You can also see the Maven project level endpoint implemented in the same file.
Once the route has been added, you must add an additional
/authorize version of the upload endpoint to your API file.
Here is an example
of the additional endpoint added for Maven. The
/authorize endpoint verifies and authorizes the request from workhorse,
then the normal upload endpoint is implemented below, consuming the metadata that workhorse provides in order to
create the package record. Workhorse provides a variety of file metadata such as type, size, and different checksum formats.
For testing purposes, you may want to enable object storage in your local development environment.
Rate Limits on GitLab.com
Package manager clients can make rapid requests that exceed the
GitLab.com standard API rate limits.
This results in a
429 Too Many Requests error.
We have opened a set of paths to allow higher rate limits. Unless it is not possible, new package managers should follow these conventions so they can take advantage of the expanded package rate limit.
These route prefixes guarantee a higher rate limit:
/api/v4/packages/ /api/v4/projects/:project_id/packages/ /api/v4/groups/:group_id/-/packages/
While working on the MVC, contributors might find features that are not mandatory for the MVC but can provide a better user experience. It's generally a good idea to keep an eye on those and open issues.
Here are some examples
- Endpoints required for search
- Front end updates to display additional package information and metadata
- Limits on file sizes
- Tracking for metrics
- Read more metadata fields from the package to make it available to the front end. For example, it's usual to be able to tag a package. Those tags can be read and saved by backend and then displayed on the packages UI.
- Endpoints for the upper levels of the remote hierarchy. This step might need to create a naming convention
This documentation is just guidelines on how to implement a package manager to match the existing structure and logic already present within GitLab. While the structure is intended to be extendable and flexible enough to allow for any given package manager, if there is good reason to stray due to the constraints or needs of a given package manager, then it should be raised and discussed within the implementation issue or merge request to work towards the most efficient outcome.