No description
Find a file
2020-03-20 13:02:27 +01:00
cmd Merge remote-tracking branch 'origin/master' 2020-03-20 12:30:59 +01:00
extras Initial 2014-10-16 20:01:43 +02:00
server handle obscure Errors properly 2020-03-20 13:02:27 +01:00
.bowerrc several small improvements 2014-10-16 20:48:07 +02:00
.dockerignore update dockerfile (#157) 2018-09-22 10:55:02 +02:00
.gitignore Add Goland Exceptions 2019-03-19 13:57:34 +01:00
.jshintrc Initial 2014-10-16 20:01:43 +02:00
.travis.yml Min go 1.13 required 2020-03-15 19:46:22 +01:00
Dockerfile Min go 1.13 required 2020-03-15 19:46:22 +01:00
examples.md Create examples.md and update README.md 2019-08-23 20:27:27 +00:00
fuzzit.sh FIX BUILD 2019-10-07 22:18:36 +02:00
go.mod handle obscure Errors properly 2020-03-20 13:02:27 +01:00
go.sum Update Dependencies to Release Version and Interface Compliance 2020-03-20 12:41:05 +01:00
LICENSE Fixes (#149) 2018-08-13 08:56:08 +02:00
main.go Major rewrite 2017-03-22 18:09:21 +01:00
manifest.json Major rewrite 2017-03-22 18:09:21 +01:00
README.md Fix Typo's 2020-03-06 15:43:14 +01:00
Vagrantfile Initial 2014-10-16 20:01:43 +02:00

transfer.sh Gitter Go Report Card Docker pulls Build Status Fuzzit Status

Easy and fast file sharing from the command-line. This code contains the server with everything you need to create your own instance.

Transfer.sh currently supports the s3 (Amazon S3), gdrive (Google Drive) providers, and local file system (local).

Disclaimer

This project repository has no relation with the service at https://transfer.sh that's managed by https://storj.io. So far we cannot address any issue related to the service at https://transfer.sh.

Usage

Upload:

$ curl --upload-file ./hello.txt https://transfer.sh/hello.txt

Encrypt & upload:

$ cat /tmp/hello.txt|gpg -ac -o-|curl -X PUT --upload-file "-" https://transfer.sh/test.txt

Download & decrypt:

$ curl https://transfer.sh/1lDau/test.txt|gpg -o- > /tmp/hello.txt

Upload to virustotal:

$ curl -X PUT --upload-file nhgbhhj https://transfer.sh/test.txt/virustotal

Deleting

$ curl -X DELETE <X-Url-Delete Response Header URL>

Request Headers

Max-Downloads

$ curl --upload-file ./hello.txt https://transfer.sh/hello.txt -H "Max-Downloads: 1" # Limit the number of downloads

Max-Days

$ curl --upload-file ./hello.txt https://transfer.sh/hello.txt -H "Max-Days: 1" # Set the number of days before deletion

Response Headers

X-Url-Delete

The URL used to request the deletion of a file. Returned as a response header.

curl -sD - --upload-file ./hello https://transfer.sh/hello.txt | grep 'X-Url-Delete'
X-Url-Delete: https://transfer.sh/hello.txt/BAYh0/hello.txt/PDw0NHPcqU

Examples

See good usage examples on examples.md

Create direct download link:

https://transfer.sh/1lDau/test.txt --> https://transfer.sh/get/1lDau/test.txt

Inline file:

https://transfer.sh/1lDau/test.txt --> https://transfer.sh/inline/1lDau/test.txt

Usage

Parameter Description Value Env
listener port to use for http (:80)
profile-listener port to use for profiler (:6060)
force-https redirect to https false
tls-listener port to use for https (:443)
tls-listener-only flag to enable tls listener only
tls-cert-file path to tls certificate
tls-private-key path to tls private key
http-auth-user user for basic http auth on upload
http-auth-pass pass for basic http auth on upload
ip-whitelist comma separated list of ips allowed to connect to the service
ip-blacklist comma separated list of ips not allowed to connect to the service
temp-path path to temp folder system temp
web-path path to static web files (for development or custom front end)
proxy-path path prefix when service is run behind a proxy
ga-key google analytics key for the front end
uservoice-key user voice key for the front end
provider which storage provider to use (s3, storj, gdrive or local)
aws-access-key aws access key AWS_ACCESS_KEY
aws-secret-key aws access key AWS_SECRET_KEY
bucket aws bucket BUCKET
s3-endpoint Custom S3 endpoint.
s3-region region of the s3 bucket eu-west-1 S3_REGION
s3-no-multipart disables s3 multipart upload false
s3-path-style Forces path style URLs, required for Minio. false
storj-access Access for the project STORJ_ACCESS
storj-bucket Bucket to use within the project STORJ_BUCKET
basedir path storage for local/gdrive provider
gdrive-client-json-filepath path to oauth client json config for gdrive provider
gdrive-local-config-path path to store local transfer.sh config cache for gdrive provider
gdrive-chunk-size chunk size for gdrive upload in megabytes, must be lower than available memory (8 MB)
lets-encrypt-hosts hosts to use for lets encrypt certificates (comma seperated)
log path to log file

If you want to use TLS using lets encrypt certificates, set lets-encrypt-hosts to your domain, set tls-listener to :443 and enable force-https.

If you want to use TLS using your own certificates, set tls-listener to :443, force-https, tls-cert=file and tls-private-key.

Development

Switched to GO111MODULE

go run main.go --provider=local --listener :8080 --temp-path=/tmp/ --basedir=/tmp/

Build

If on go < 1.11

go get -u -v ./...
go build -o transfersh main.go

Docker

For easy deployment, we've created a Docker container.

docker run --publish 8080:8080 dutchcoders/transfer.sh:latest --provider local --basedir /tmp/

S3 Usage

For the usage with a AWS S3 Bucket, you just need to specify the following options:

  • provider
  • aws-access-key
  • aws-secret-key
  • bucket
  • s3-region

If you specify the s3-region, you don't need to set the endpoint URL since the correct endpoint will used automatically.

Custom S3 providers

To use a custom non-AWS S3 provider, you need to specify the endpoint as defined from your cloud provider.

Storj Network Provider

To use the Storj Network as storage provider you need to specify the following flags:

  • provider --provider storj
  • storj-access (either via flag or environment variable STORJ_ACCESS)
  • storj-bucket (either via flag or environment variable STORJ_BUCKET)

Creating Bucket and Scope

In preparation you need to create a scope (or copy it from the uplink configuration) and a bucket.

To get started, download the latest uplink from the release page: https://github.com/storj/storj/releases

After extracting, execute uplink setup. The Wizard asks for Satellite to use, the API Key (which you can retrieve via the Satellite UI), as well as an Encryption Key. Once the uplink is setup create the bucket using the following schema: uplink mb sj://<BUCKET> where is your desired name.

Afterwards you can copy the SCOPE out of the configuration file of the uplink and then start the startup of the transfer.sh endpoint. For enhanced security its recommended to provide both the scope and the bucket name as ENV Variables.

Example:

export STORJ_BUCKET=transfersh
export STORJ_ACCESS=<SCOPE>
transfer.sh --provider storj

Google Drive Usage

For the usage with Google drive, you need to specify the following options:

  • provider
  • gdrive-client-json-filepath
  • gdrive-local-config-path
  • basedir

Creating Gdrive Client Json

You need to create a Oauth Client id from console.cloud.google.com download the file and place into a safe directory

Usage example

go run main.go --provider gdrive --basedir /tmp/ --gdrive-client-json-filepath /[credential_dir] --gdrive-local-config-path [directory_to_save_config]

Contributions

Contributions are welcome.

Creators

Remco Verhoef

Uvis Grinfelds

Maintainer

Andrea Spacca

Code and documentation copyright 2011-2018 Remco Verhoef. Code released under the MIT license.