mirror of
https://github.com/sigoden/dufs.git
synced 2026-04-09 09:09:03 +03:00
Compare commits
83 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
58a46f7c3a | ||
|
|
ef757281b3 | ||
|
|
de0614816a | ||
|
|
81d2c49e3f | ||
|
|
ee21894452 | ||
|
|
0ac0c048ec | ||
|
|
17063454d3 | ||
|
|
af347f9cf0 | ||
|
|
006e03ed30 | ||
|
|
77f86a4c60 | ||
|
|
a66f95b39f | ||
|
|
52506bc01f | ||
|
|
270cc0cba2 | ||
|
|
5988442d5c | ||
|
|
3873f4794a | ||
|
|
cd84dff87f | ||
|
|
8590f3e841 | ||
|
|
44a4ddf973 | ||
|
|
37800f630d | ||
|
|
5c850256f4 | ||
|
|
0cec573579 | ||
|
|
073b098111 | ||
|
|
6ff8b29b69 | ||
|
|
7584fe3d08 | ||
|
|
653cd167d0 | ||
|
|
ab29e39148 | ||
|
|
f8d6859354 | ||
|
|
130435c387 | ||
|
|
afdfde01f0 | ||
|
|
ae97c714d6 | ||
|
|
c352dab470 | ||
|
|
743db47f90 | ||
|
|
a476c15a09 | ||
|
|
0d74fa3ec5 | ||
|
|
b83cc6938b | ||
|
|
a187b14885 | ||
|
|
d3de3db0d9 | ||
|
|
80ac9afe68 | ||
|
|
4ef07737e1 | ||
|
|
5782c5f413 | ||
|
|
8b4cab1e69 | ||
|
|
70300b133c | ||
|
|
7ea4bb808d | ||
|
|
6766e0d437 | ||
|
|
53c9bc8bea | ||
|
|
60df3b473c | ||
|
|
6510ae8be9 | ||
|
|
9545fb6e37 | ||
|
|
0fd0f11298 | ||
|
|
46aa8fcc02 | ||
|
|
09bb738866 | ||
|
|
3612ef10d1 | ||
|
|
7ac2039a36 | ||
|
|
7f83de765a | ||
|
|
9b3779b13a | ||
|
|
11a52f29c4 | ||
|
|
10204c723f | ||
|
|
204421643d | ||
|
|
d9706d75ef | ||
|
|
40df0bd2f9 | ||
|
|
a53411b4d6 | ||
|
|
609017b2f5 | ||
|
|
7dc0b0e218 | ||
|
|
6be36b8e51 | ||
|
|
8be545d3da | ||
|
|
4f3a8d275b | ||
|
|
9c412f4276 | ||
|
|
27c269d6a0 | ||
|
|
57b4a74279 | ||
|
|
1112b936b8 | ||
|
|
033d37c4d4 | ||
|
|
577eea5fa4 | ||
|
|
d22be95dda | ||
|
|
8d7c1fbf53 | ||
|
|
4622c48120 | ||
|
|
f8ea41638f | ||
|
|
2890b3929d | ||
|
|
f5c0aefd8e | ||
|
|
8a1e7674df | ||
|
|
3c6206849f | ||
|
|
652f836c23 | ||
|
|
fb5b50f059 | ||
|
|
e43554b795 |
7
.github/ISSUE_TEMPLATE/bug_report.md
vendored
7
.github/ISSUE_TEMPLATE/bug_report.md
vendored
@@ -9,9 +9,10 @@ about: Create a report to help us improve
|
|||||||
|
|
||||||
**Log**
|
**Log**
|
||||||
|
|
||||||
If applicable, add logs to help explain your problem.
|
The dufs log is crucial for locating the problem, so please do not omit it.
|
||||||
|
|
||||||
**Environment:**
|
**Environment:**
|
||||||
- Dufs version:
|
- Dufs version:
|
||||||
- Browser/Webdav Info:
|
- Browser/Webdav info:
|
||||||
- OS Info:
|
- OS info:
|
||||||
|
- Proxy server: e.g. nginx, cloudflare
|
||||||
|
|||||||
10
.github/workflows/ci.yaml
vendored
10
.github/workflows/ci.yaml
vendored
@@ -29,16 +29,12 @@ jobs:
|
|||||||
RUSTFLAGS: --deny warnings
|
RUSTFLAGS: --deny warnings
|
||||||
|
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v2
|
- uses: actions/checkout@v3
|
||||||
|
|
||||||
- name: Install Rust Toolchain Components
|
- name: Install Rust Toolchain Components
|
||||||
uses: actions-rs/toolchain@v1
|
uses: dtolnay/rust-toolchain@stable
|
||||||
with:
|
|
||||||
components: clippy, rustfmt
|
|
||||||
override: true
|
|
||||||
toolchain: stable
|
|
||||||
|
|
||||||
- uses: Swatinem/rust-cache@v1
|
- uses: Swatinem/rust-cache@v2
|
||||||
|
|
||||||
- name: Test
|
- name: Test
|
||||||
run: cargo test --all
|
run: cargo test --all
|
||||||
|
|||||||
84
.github/workflows/release.yaml
vendored
84
.github/workflows/release.yaml
vendored
@@ -27,7 +27,7 @@ jobs:
|
|||||||
- target: aarch64-pc-windows-msvc
|
- target: aarch64-pc-windows-msvc
|
||||||
os: windows-latest
|
os: windows-latest
|
||||||
use-cross: true
|
use-cross: true
|
||||||
cargo-flags: "--no-default-features"
|
cargo-flags: ""
|
||||||
- target: x86_64-apple-darwin
|
- target: x86_64-apple-darwin
|
||||||
os: macos-latest
|
os: macos-latest
|
||||||
cargo-flags: ""
|
cargo-flags: ""
|
||||||
@@ -71,31 +71,43 @@ jobs:
|
|||||||
use-cross: true
|
use-cross: true
|
||||||
cargo-flags: "--no-default-features"
|
cargo-flags: "--no-default-features"
|
||||||
runs-on: ${{matrix.os}}
|
runs-on: ${{matrix.os}}
|
||||||
|
env:
|
||||||
|
BUILD_CMD: cargo
|
||||||
|
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v2
|
- uses: actions/checkout@v3
|
||||||
|
|
||||||
- name: Check Tag
|
- name: Check Tag
|
||||||
id: check-tag
|
id: check-tag
|
||||||
shell: bash
|
shell: bash
|
||||||
run: |
|
run: |
|
||||||
tag=${GITHUB_REF##*/}
|
ver=${GITHUB_REF##*/}
|
||||||
echo "::set-output name=version::$tag"
|
echo "version=$ver" >> $GITHUB_OUTPUT
|
||||||
if [[ "$tag" =~ [0-9]+.[0-9]+.[0-9]+$ ]]; then
|
if [[ "$ver" =~ [0-9]+.[0-9]+.[0-9]+$ ]]; then
|
||||||
echo "::set-output name=rc::false"
|
echo "rc=false" >> $GITHUB_OUTPUT
|
||||||
else
|
else
|
||||||
echo "::set-output name=rc::true"
|
echo "rc=true" >> $GITHUB_OUTPUT
|
||||||
fi
|
fi
|
||||||
|
|
||||||
|
|
||||||
- name: Install Rust Toolchain Components
|
- name: Install Rust Toolchain Components
|
||||||
uses: actions-rs/toolchain@v1
|
uses: dtolnay/rust-toolchain@stable
|
||||||
with:
|
with:
|
||||||
override: true
|
targets: ${{ matrix.target }}
|
||||||
target: ${{ matrix.target }}
|
# Since rust 1.72, mips platforms are tier 3
|
||||||
toolchain: stable
|
toolchain: 1.71
|
||||||
profile: minimal # minimal component installation (ie, no documentation)
|
|
||||||
|
- name: Install cross
|
||||||
|
if: matrix.use-cross
|
||||||
|
uses: taiki-e/install-action@v2
|
||||||
|
with:
|
||||||
|
tool: cross
|
||||||
|
|
||||||
|
- name: Overwrite build command env variable
|
||||||
|
if: matrix.use-cross
|
||||||
|
shell: bash
|
||||||
|
run: echo "BUILD_CMD=cross" >> $GITHUB_ENV
|
||||||
|
|
||||||
- name: Show Version Information (Rust, cargo, GCC)
|
- name: Show Version Information (Rust, cargo, GCC)
|
||||||
shell: bash
|
shell: bash
|
||||||
run: |
|
run: |
|
||||||
@@ -107,11 +119,8 @@ jobs:
|
|||||||
rustc -V
|
rustc -V
|
||||||
|
|
||||||
- name: Build
|
- name: Build
|
||||||
uses: actions-rs/cargo@v1
|
shell: bash
|
||||||
with:
|
run: $BUILD_CMD build --locked --release --target=${{ matrix.target }} ${{ matrix.cargo-flags }}
|
||||||
use-cross: ${{ matrix.use-cross }}
|
|
||||||
command: build
|
|
||||||
args: --locked --release --target=${{ matrix.target }} ${{ matrix.cargo-flags }}
|
|
||||||
|
|
||||||
- name: Build Archive
|
- name: Build Archive
|
||||||
shell: bash
|
shell: bash
|
||||||
@@ -123,8 +132,7 @@ jobs:
|
|||||||
set -euxo pipefail
|
set -euxo pipefail
|
||||||
|
|
||||||
bin=${GITHUB_REPOSITORY##*/}
|
bin=${GITHUB_REPOSITORY##*/}
|
||||||
src=`pwd`
|
dist_dir=`pwd`/dist
|
||||||
dist=$src/dist
|
|
||||||
name=$bin-$version-$target
|
name=$bin-$version-$target
|
||||||
executable=target/$target/release/$bin
|
executable=target/$target/release/$bin
|
||||||
|
|
||||||
@@ -132,22 +140,22 @@ jobs:
|
|||||||
executable=$executable.exe
|
executable=$executable.exe
|
||||||
fi
|
fi
|
||||||
|
|
||||||
mkdir $dist
|
mkdir $dist_dir
|
||||||
cp $executable $dist
|
cp $executable $dist_dir
|
||||||
cd $dist
|
cd $dist_dir
|
||||||
|
|
||||||
if [[ "$RUNNER_OS" == "Windows" ]]; then
|
if [[ "$RUNNER_OS" == "Windows" ]]; then
|
||||||
archive=$dist/$name.zip
|
archive=$dist_dir/$name.zip
|
||||||
7z a $archive *
|
7z a $archive *
|
||||||
echo "::set-output name=archive::`pwd -W`/$name.zip"
|
echo "archive=dist/$name.zip" >> $GITHUB_OUTPUT
|
||||||
else
|
else
|
||||||
archive=$dist/$name.tar.gz
|
archive=$dist_dir/$name.tar.gz
|
||||||
tar czf $archive *
|
tar -czf $archive *
|
||||||
echo "::set-output name=archive::$archive"
|
echo "archive=dist/$name.tar.gz" >> $GITHUB_OUTPUT
|
||||||
fi
|
fi
|
||||||
|
|
||||||
- name: Publish Archive
|
- name: Publish Archive
|
||||||
uses: softprops/action-gh-release@v0.1.5
|
uses: softprops/action-gh-release@v1
|
||||||
if: ${{ startsWith(github.ref, 'refs/tags/') }}
|
if: ${{ startsWith(github.ref, 'refs/tags/') }}
|
||||||
with:
|
with:
|
||||||
draft: false
|
draft: false
|
||||||
@@ -163,16 +171,16 @@ jobs:
|
|||||||
needs: release
|
needs: release
|
||||||
steps:
|
steps:
|
||||||
- name: Set up QEMU
|
- name: Set up QEMU
|
||||||
uses: docker/setup-qemu-action@v1
|
uses: docker/setup-qemu-action@v2
|
||||||
- name: Set up Docker Buildx
|
- name: Set up Docker Buildx
|
||||||
uses: docker/setup-buildx-action@v1
|
uses: docker/setup-buildx-action@v2
|
||||||
- name: Login to DockerHub
|
- name: Login to DockerHub
|
||||||
uses: docker/login-action@v1
|
uses: docker/login-action@v2
|
||||||
with:
|
with:
|
||||||
username: ${{ secrets.DOCKERHUB_USERNAME }}
|
username: ${{ secrets.DOCKERHUB_USERNAME }}
|
||||||
password: ${{ secrets.DOCKERHUB_TOKEN }}
|
password: ${{ secrets.DOCKERHUB_TOKEN }}
|
||||||
- name: Build and push
|
- name: Build and push
|
||||||
uses: docker/build-push-action@v2
|
uses: docker/build-push-action@v4
|
||||||
with:
|
with:
|
||||||
build-args: |
|
build-args: |
|
||||||
REPO=${{ github.repository }}
|
REPO=${{ github.repository }}
|
||||||
@@ -184,20 +192,18 @@ jobs:
|
|||||||
linux/arm/v7
|
linux/arm/v7
|
||||||
push: ${{ needs.release.outputs.rc == 'false' }}
|
push: ${{ needs.release.outputs.rc == 'false' }}
|
||||||
tags: ${{ github.repository }}:latest, ${{ github.repository }}:${{ github.ref_name }}
|
tags: ${{ github.repository }}:latest, ${{ github.repository }}:${{ github.ref_name }}
|
||||||
|
|
||||||
publish-crate:
|
publish-crate:
|
||||||
name: Publish to crates.io
|
name: Publish to crates.io
|
||||||
if: ${{ needs.release.outputs.rc == 'false' }}
|
if: ${{ needs.release.outputs.rc == 'false' }}
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
needs: release
|
needs: release
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v2
|
- uses: actions/checkout@v3
|
||||||
- uses: actions-rs/toolchain@v1
|
|
||||||
with:
|
|
||||||
profile: minimal
|
|
||||||
toolchain: stable
|
|
||||||
- name: Publish
|
|
||||||
|
|
||||||
|
- uses: dtolnay/rust-toolchain@stable
|
||||||
|
|
||||||
|
- name: Publish
|
||||||
env:
|
env:
|
||||||
CARGO_REGISTRY_TOKEN: ${{ secrets.CRATES_IO_API_TOKEN }}
|
CARGO_REGISTRY_TOKEN: ${{ secrets.CRATES_IO_API_TOKEN }}
|
||||||
run: cargo publish
|
run: cargo publish
|
||||||
127
CHANGELOG.md
127
CHANGELOG.md
@@ -2,6 +2,133 @@
|
|||||||
|
|
||||||
All notable changes to this project will be documented in this file.
|
All notable changes to this project will be documented in this file.
|
||||||
|
|
||||||
|
## [0.39.0] - 2024-01-11
|
||||||
|
|
||||||
|
### Bug Fixes
|
||||||
|
|
||||||
|
- Upload more than 100 files in directory ([#317](https://github.com/sigoden/dufs/issues/317))
|
||||||
|
- Auth precedence ([#325](https://github.com/sigoden/dufs/issues/325))
|
||||||
|
- Serve files with names containing newline char ([#328](https://github.com/sigoden/dufs/issues/328))
|
||||||
|
- Corrupted zip when downloading large folders ([#337](https://github.com/sigoden/dufs/issues/337))
|
||||||
|
|
||||||
|
### Features
|
||||||
|
|
||||||
|
- Empty search `?q=` list all paths ([#311](https://github.com/sigoden/dufs/issues/311))
|
||||||
|
- Add `--compress` option ([#319](https://github.com/sigoden/dufs/issues/319))
|
||||||
|
- Upgrade to hyper 1.0 ([#321](https://github.com/sigoden/dufs/issues/321))
|
||||||
|
- Auth supports forbidden permissions ([#329](https://github.com/sigoden/dufs/issues/329))
|
||||||
|
- Supports resumable uploads ([#343](https://github.com/sigoden/dufs/issues/343))
|
||||||
|
|
||||||
|
### Refactor
|
||||||
|
|
||||||
|
- Change the format of www-authenticate ([#312](https://github.com/sigoden/dufs/issues/312))
|
||||||
|
- Change the value name of `--config` ([#313](https://github.com/sigoden/dufs/issues/313))
|
||||||
|
- Optimize http range parsing and handling ([#323](https://github.com/sigoden/dufs/issues/323))
|
||||||
|
- Propfind with auth no need to list all ([#344](https://github.com/sigoden/dufs/issues/344))
|
||||||
|
|
||||||
|
## [0.38.0] - 2023-11-28
|
||||||
|
|
||||||
|
### Bug Fixes
|
||||||
|
|
||||||
|
- Unable to start if config file omit bind/port fields ([#294](https://github.com/sigoden/dufs/issues/294))
|
||||||
|
|
||||||
|
### Features
|
||||||
|
|
||||||
|
- Password can contain `:` `@` `|` ([#297](https://github.com/sigoden/dufs/issues/297))
|
||||||
|
- Deprecate the use of `|` to separate auth rules ([#298](https://github.com/sigoden/dufs/issues/298))
|
||||||
|
- More flexible config values ([#299](https://github.com/sigoden/dufs/issues/299))
|
||||||
|
- Ui supports view file ([#301](https://github.com/sigoden/dufs/issues/301))
|
||||||
|
|
||||||
|
### Refactor
|
||||||
|
|
||||||
|
- Take improvements from the edge browser ([#289](https://github.com/sigoden/dufs/issues/289))
|
||||||
|
- Ui change the cursor for upload-btn to a pointer ([#291](https://github.com/sigoden/dufs/issues/291))
|
||||||
|
- Ui improve uploading progress ([#296](https://github.com/sigoden/dufs/issues/296))
|
||||||
|
|
||||||
|
## [0.37.1] - 2023-11-08
|
||||||
|
|
||||||
|
### Bug Fixes
|
||||||
|
|
||||||
|
- Use DUFS_CONFIG to specify the config file path ([#286](https://github.com/sigoden/dufs/issues/286)
|
||||||
|
|
||||||
|
## [0.37.0] - 2023-11-08
|
||||||
|
|
||||||
|
### Bug Fixes
|
||||||
|
|
||||||
|
- Sort path ignore case ([#264](https://github.com/sigoden/dufs/issues/264))
|
||||||
|
- Ui show user-name next to the user-icon ([#278](https://github.com/sigoden/dufs/issues/278))
|
||||||
|
- Auto delete half-uploaded files ([#280](https://github.com/sigoden/dufs/issues/280))
|
||||||
|
|
||||||
|
### Features
|
||||||
|
|
||||||
|
- Deprecate `--auth-method`, as both options are available ([#279](https://github.com/sigoden/dufs/issues/279))
|
||||||
|
- Support config file with `--config` option ([#281](https://github.com/sigoden/dufs/issues/281))
|
||||||
|
- Support hashed password ([#283](https://github.com/sigoden/dufs/issues/283))
|
||||||
|
|
||||||
|
### Refactor
|
||||||
|
|
||||||
|
- Remove one clone on `assets_prefix` ([#270](https://github.com/sigoden/dufs/issues/270))
|
||||||
|
- Optimize tests
|
||||||
|
- Improve code quanity ([#282](https://github.com/sigoden/dufs/issues/282))
|
||||||
|
|
||||||
|
## [0.36.0] - 2023-08-24
|
||||||
|
|
||||||
|
### Bug Fixes
|
||||||
|
|
||||||
|
- Ui readonly if no write perm ([#258](https://github.com/sigoden/dufs/issues/258))
|
||||||
|
|
||||||
|
### Testing
|
||||||
|
|
||||||
|
- Remove dependency on native tls ([#255](https://github.com/sigoden/dufs/issues/255))
|
||||||
|
|
||||||
|
## [0.35.0] - 2023-08-14
|
||||||
|
|
||||||
|
### Bug Fixes
|
||||||
|
|
||||||
|
- Search should ignore entry path ([#235](https://github.com/sigoden/dufs/issues/235))
|
||||||
|
- Typo __ASSERTS_PREFIX__ ([#252](https://github.com/sigoden/dufs/issues/252))
|
||||||
|
|
||||||
|
### Features
|
||||||
|
|
||||||
|
- Sort by type first, then sort by name/mtime/size ([#241](https://github.com/sigoden/dufs/issues/241))
|
||||||
|
|
||||||
|
## [0.34.2] - 2023-06-05
|
||||||
|
|
||||||
|
### Bug Fixes
|
||||||
|
|
||||||
|
- Ui refresh page after login ([#230](https://github.com/sigoden/dufs/issues/230))
|
||||||
|
- Webdav only see public folder even logging in ([#231](https://github.com/sigoden/dufs/issues/231))
|
||||||
|
|
||||||
|
## [0.34.1] - 2023-06-02
|
||||||
|
|
||||||
|
### Bug Fixes
|
||||||
|
|
||||||
|
- Auth logic ([#224](https://github.com/sigoden/dufs/issues/224))
|
||||||
|
- Allow all cors headers and methods ([#225](https://github.com/sigoden/dufs/issues/225))
|
||||||
|
|
||||||
|
### Refactor
|
||||||
|
|
||||||
|
- Ui checkAuth ([#226](https://github.com/sigoden/dufs/issues/226))
|
||||||
|
|
||||||
|
## [0.34.0] - 2023-06-01
|
||||||
|
|
||||||
|
### Bug Fixes
|
||||||
|
|
||||||
|
- URL-encoded filename when downloading in safari ([#203](https://github.com/sigoden/dufs/issues/203))
|
||||||
|
- Ui path table show move action ([#219](https://github.com/sigoden/dufs/issues/219))
|
||||||
|
- Ui set default max uploading to 1 ([#220](https://github.com/sigoden/dufs/issues/220))
|
||||||
|
|
||||||
|
### Features
|
||||||
|
|
||||||
|
- Webui editing support multiple encodings ([#197](https://github.com/sigoden/dufs/issues/197))
|
||||||
|
- Add timestamp metadata to generated zip file ([#204](https://github.com/sigoden/dufs/issues/204))
|
||||||
|
- Show precise file size with decimal ([#210](https://github.com/sigoden/dufs/issues/210))
|
||||||
|
- [**breaking**] New auth ([#218](https://github.com/sigoden/dufs/issues/218))
|
||||||
|
|
||||||
|
### Refactor
|
||||||
|
|
||||||
|
- Cli positional rename root => SERVE_PATH([#215](https://github.com/sigoden/dufs/issues/215))
|
||||||
|
|
||||||
## [0.33.0] - 2023-03-17
|
## [0.33.0] - 2023-03-17
|
||||||
|
|
||||||
### Bug Fixes
|
### Bug Fixes
|
||||||
|
|||||||
1642
Cargo.lock
generated
1642
Cargo.lock
generated
File diff suppressed because it is too large
Load Diff
53
Cargo.toml
53
Cargo.toml
@@ -1,6 +1,6 @@
|
|||||||
[package]
|
[package]
|
||||||
name = "dufs"
|
name = "dufs"
|
||||||
version = "0.33.0"
|
version = "0.39.0"
|
||||||
edition = "2021"
|
edition = "2021"
|
||||||
authors = ["sigoden <sigoden@gmail.com>"]
|
authors = ["sigoden <sigoden@gmail.com>"]
|
||||||
description = "Dufs is a distinctive utility file server"
|
description = "Dufs is a distinctive utility file server"
|
||||||
@@ -13,55 +13,64 @@ keywords = ["static", "file", "server", "webdav", "cli"]
|
|||||||
[dependencies]
|
[dependencies]
|
||||||
clap = { version = "4", features = ["wrap_help", "env"] }
|
clap = { version = "4", features = ["wrap_help", "env"] }
|
||||||
clap_complete = "4"
|
clap_complete = "4"
|
||||||
chrono = "0.4"
|
chrono = { version = "0.4", default-features = false, features = ["clock"] }
|
||||||
tokio = { version = "1", features = ["rt-multi-thread", "macros", "fs", "io-util", "signal"]}
|
tokio = { version = "1", features = ["rt-multi-thread", "macros", "fs", "io-util", "signal"]}
|
||||||
tokio-util = { version = "0.7", features = ["io-util"] }
|
tokio-util = { version = "0.7", features = ["io-util", "compat"] }
|
||||||
hyper = { version = "0.14", features = ["http1", "server", "tcp", "stream"] }
|
hyper = { version = "1.0", features = ["http1", "server"] }
|
||||||
percent-encoding = "2.1"
|
percent-encoding = "2.3"
|
||||||
serde = { version = "1", features = ["derive"] }
|
serde = { version = "1", features = ["derive"] }
|
||||||
serde_json = "1"
|
serde_json = "1"
|
||||||
futures = "0.3"
|
futures-util = { version = "0.3", default-features = false, features = ["alloc"] }
|
||||||
base64 = "0.21"
|
async_zip = { version = "0.0.16", default-features = false, features = ["deflate", "bzip2", "xz", "chrono", "tokio"] }
|
||||||
async_zip = { version = "0.0.12", default-features = false, features = ["deflate"] }
|
headers = "0.4"
|
||||||
headers = "0.3"
|
|
||||||
mime_guess = "2.0"
|
mime_guess = "2.0"
|
||||||
if-addrs = "0.10.1"
|
if-addrs = "0.11"
|
||||||
rustls = { version = "0.20", default-features = false, features = ["tls12"], optional = true }
|
rustls-pemfile = { version = "2.0", optional = true }
|
||||||
rustls-pemfile = { version = "1", optional = true }
|
tokio-rustls = { version = "0.25", optional = true }
|
||||||
tokio-rustls = { version = "0.23", optional = true }
|
|
||||||
md5 = "0.7"
|
md5 = "0.7"
|
||||||
lazy_static = "1.4"
|
lazy_static = "1.4"
|
||||||
uuid = { version = "1.1", features = ["v4", "fast-rng"] }
|
uuid = { version = "1.4", features = ["v4", "fast-rng"] }
|
||||||
urlencoding = "2.1"
|
urlencoding = "2.1"
|
||||||
xml-rs = "0.8"
|
xml-rs = "0.8"
|
||||||
log = "0.4"
|
log = "0.4"
|
||||||
socket2 = "0.5"
|
socket2 = "0.5"
|
||||||
async-stream = "0.3"
|
async-stream = "0.3"
|
||||||
walkdir = "2.3"
|
walkdir = "2.3"
|
||||||
form_urlencoded = "1.0"
|
form_urlencoded = "1.2"
|
||||||
alphanumeric-sort = "1.4"
|
alphanumeric-sort = "1.4"
|
||||||
content_inspector = "0.2"
|
content_inspector = "0.2"
|
||||||
anyhow = "1.0"
|
anyhow = "1.0"
|
||||||
chardetng = "0.1"
|
chardetng = "0.1"
|
||||||
glob = "0.3.1"
|
glob = "0.3"
|
||||||
|
indexmap = "2.0"
|
||||||
|
serde_yaml = "0.9"
|
||||||
|
sha-crypt = "0.5"
|
||||||
|
base64 = "0.21"
|
||||||
|
smart-default = "0.7"
|
||||||
|
rustls-pki-types = "1.0"
|
||||||
|
hyper-util = { version = "0.1", features = ["server-auto", "tokio"] }
|
||||||
|
http-body-util = "0.1"
|
||||||
|
bytes = "1.5"
|
||||||
|
pin-project-lite = "0.2"
|
||||||
|
|
||||||
[features]
|
[features]
|
||||||
default = ["tls"]
|
default = ["tls"]
|
||||||
tls = ["rustls", "rustls-pemfile", "tokio-rustls"]
|
tls = ["rustls-pemfile", "tokio-rustls"]
|
||||||
|
|
||||||
[dev-dependencies]
|
[dev-dependencies]
|
||||||
assert_cmd = "2"
|
assert_cmd = "2"
|
||||||
reqwest = { version = "0.11", features = ["blocking", "multipart", "rustls-tls"], default-features = false }
|
reqwest = { version = "0.11", features = ["blocking", "multipart", "rustls-tls"], default-features = false }
|
||||||
assert_fs = "1"
|
assert_fs = "1"
|
||||||
port_check = "0.1"
|
port_check = "0.1"
|
||||||
rstest = "0.16"
|
rstest = "0.18"
|
||||||
regex = "1"
|
regex = "1"
|
||||||
url = "2"
|
url = "2"
|
||||||
diqwest = { version = "1", features = ["blocking"] }
|
diqwest = { version = "2.0", features = ["blocking"], default-features = false }
|
||||||
predicates = "3"
|
predicates = "3"
|
||||||
indexmap = "1.9"
|
|
||||||
|
|
||||||
[profile.release]
|
[profile.release]
|
||||||
|
opt-level = 3
|
||||||
lto = true
|
lto = true
|
||||||
strip = true
|
codegen-units = 1
|
||||||
opt-level = "z"
|
panic = "abort"
|
||||||
|
strip = "symbols"
|
||||||
|
|||||||
173
README.md
173
README.md
@@ -13,8 +13,8 @@ Dufs is a distinctive utility file server that supports static serving, uploadin
|
|||||||
- Download folder as zip file
|
- Download folder as zip file
|
||||||
- Upload files and folders (Drag & Drop)
|
- Upload files and folders (Drag & Drop)
|
||||||
- Create/Edit/Search files
|
- Create/Edit/Search files
|
||||||
- Partial responses (Parallel/Resume download)
|
- Resumable/partial uploads/downloads
|
||||||
- Path level access control
|
- Access control
|
||||||
- Support https
|
- Support https
|
||||||
- Support webdav
|
- Support webdav
|
||||||
- Easy to use with curl
|
- Easy to use with curl
|
||||||
@@ -48,18 +48,18 @@ Download from [Github Releases](https://github.com/sigoden/dufs/releases), unzip
|
|||||||
```
|
```
|
||||||
Dufs is a distinctive utility file server - https://github.com/sigoden/dufs
|
Dufs is a distinctive utility file server - https://github.com/sigoden/dufs
|
||||||
|
|
||||||
Usage: dufs [OPTIONS] [root]
|
Usage: dufs [OPTIONS] [serve-path]
|
||||||
|
|
||||||
Arguments:
|
Arguments:
|
||||||
[root] Specific path to serve [default: .]
|
[serve-path] Specific path to serve [default: .]
|
||||||
|
|
||||||
Options:
|
Options:
|
||||||
|
-c, --config <file> Specify configuration file
|
||||||
-b, --bind <addrs> Specify bind address or unix socket
|
-b, --bind <addrs> Specify bind address or unix socket
|
||||||
-p, --port <port> Specify port to listen on [default: 5000]
|
-p, --port <port> Specify port to listen on [default: 5000]
|
||||||
--path-prefix <path> Specify a path prefix
|
--path-prefix <path> Specify a path prefix
|
||||||
--hidden <value> Hide paths from directory listings, separated by `,`
|
--hidden <value> Hide paths from directory listings, e.g. tmp,*.log,*.lock
|
||||||
-a, --auth <rules> Add auth for path
|
-a, --auth <rules> Add auth roles, e.g. user:pass@/dir1:rw,/dir2
|
||||||
--auth-method <value> Select auth method [default: digest] [possible values: basic, digest]
|
|
||||||
-A, --allow-all Allow all operations
|
-A, --allow-all Allow all operations
|
||||||
--allow-upload Allow upload files/folders
|
--allow-upload Allow upload files/folders
|
||||||
--allow-delete Allow delete files/folders
|
--allow-delete Allow delete files/folders
|
||||||
@@ -70,18 +70,19 @@ Options:
|
|||||||
--render-index Serve index.html when requesting a directory, returns 404 if not found index.html
|
--render-index Serve index.html when requesting a directory, returns 404 if not found index.html
|
||||||
--render-try-index Serve index.html when requesting a directory, returns directory listing if not found index.html
|
--render-try-index Serve index.html when requesting a directory, returns directory listing if not found index.html
|
||||||
--render-spa Serve SPA(Single Page Application)
|
--render-spa Serve SPA(Single Page Application)
|
||||||
--assets <path> Use custom assets to override builtin assets
|
--assets <path> Set the path to the assets directory for overriding the built-in assets
|
||||||
|
--log-format <format> Customize http log format
|
||||||
|
--compress <level> Set zip compress level [default: low] [possible values: none, low, medium, high]
|
||||||
|
--completions <shell> Print shell completion script for <shell> [possible values: bash, elvish, fish, powershell, zsh]
|
||||||
--tls-cert <path> Path to an SSL/TLS certificate to serve with HTTPS
|
--tls-cert <path> Path to an SSL/TLS certificate to serve with HTTPS
|
||||||
--tls-key <path> Path to the SSL/TLS certificate's private key
|
--tls-key <path> Path to the SSL/TLS certificate's private key
|
||||||
--log-format <format> Customize http log format
|
-h, --help Print help
|
||||||
--completions <shell> Print shell completion script for <shell> [possible values: bash, elvish, fish, powershell, zsh]
|
-V, --version Print version
|
||||||
-h, --help Print help information
|
|
||||||
-V, --version Print version information
|
|
||||||
```
|
```
|
||||||
|
|
||||||
## Examples
|
## Examples
|
||||||
|
|
||||||
Serve current working directory in readonly mode
|
Serve current working directory in read-only mode
|
||||||
|
|
||||||
```
|
```
|
||||||
dufs
|
dufs
|
||||||
@@ -126,7 +127,7 @@ dufs --render-index
|
|||||||
Require username/password
|
Require username/password
|
||||||
|
|
||||||
```
|
```
|
||||||
dufs -a /@admin:123
|
dufs -a admin:123@/:rw
|
||||||
```
|
```
|
||||||
|
|
||||||
Listen on specific host:ip
|
Listen on specific host:ip
|
||||||
@@ -186,16 +187,30 @@ curl -X MOVE https://127.0.0.1:5000/path -H "Destination: https://127.0.0.1:5000
|
|||||||
List/search directory contents
|
List/search directory contents
|
||||||
|
|
||||||
```
|
```
|
||||||
curl http://127.0.0.1:5000?simple # output names only, just like `ls -1`
|
curl http://127.0.0.1:5000?q=Dockerfile # search for files, similar to `find -name Dockerfile`
|
||||||
|
curl http://127.0.0.1:5000?simple # output names only, similar to `ls -1`
|
||||||
curl http://127.0.0.1:5000?json # output paths in json format
|
curl http://127.0.0.1:5000?json # output paths in json format
|
||||||
curl http://127.0.0.1:5000?q=Dockerfile&simple # search for files, just like `find -name Dockerfile`
|
|
||||||
```
|
```
|
||||||
|
|
||||||
With authorization
|
With authorization
|
||||||
|
|
||||||
```
|
```
|
||||||
curl --user user:pass --digest http://192.168.8.10:5000/file # digest auth
|
curl http://127.0.0.1:5000/file --user user:pass # basic auth
|
||||||
curl --user user:pass http://192.168.8.10:5000/file # basic auth
|
curl http://127.0.0.1:5000/file --user user:pass --digest # digest auth
|
||||||
|
```
|
||||||
|
|
||||||
|
Resumable downloads
|
||||||
|
|
||||||
|
```
|
||||||
|
curl -C- -o file http://127.0.0.1:5000/file
|
||||||
|
```
|
||||||
|
|
||||||
|
Resumable uploads
|
||||||
|
|
||||||
|
```
|
||||||
|
upload_offset=$(curl -I -s http://127.0.0.1:5000/file | tr -d '\r' | sed -n 's/content-length: //p')
|
||||||
|
dd skip=$upload_offset if=file status=none ibs=1 | \
|
||||||
|
curl -X PATCH -H "X-Update-Range: append" --data-binary @- http://127.0.0.1:5000/file
|
||||||
```
|
```
|
||||||
|
|
||||||
<details>
|
<details>
|
||||||
@@ -203,43 +218,47 @@ curl --user user:pass http://192.168.8.10:5000/file # basic auth
|
|||||||
|
|
||||||
### Access Control
|
### Access Control
|
||||||
|
|
||||||
Dufs supports path level access control. You can control who can do what on which path with `--auth`/`-a`.
|
Dufs supports account based access control. You can control who can do what on which path with `--auth`/`-a`.
|
||||||
|
|
||||||
```
|
```
|
||||||
dufs -a <path>@<readwrite>
|
dufs -a admin:admin@/:rw -a guest:guest@/
|
||||||
dufs -a <path>@<readwrite>@<readonly>
|
dufs -a user:pass@/:rw,/dir1,/dir2:- -a @/
|
||||||
dufs -a <path>@<readwrite>@*
|
|
||||||
```
|
```
|
||||||
|
|
||||||
- `<path>`: Protected url path
|
1. Use `@` to separate the account and paths. No account means anonymous user.
|
||||||
- `<readwrite>`: Account with readwrite permissions. If dufs is run with `dufs --allow-all`, the permissions are upload/delete/search/view/download. If dufs is run with `dufs --allow-upload`, the permissions are upload/view/download.
|
2. Use `:` to separate the username and password of the account.
|
||||||
- `<readonly>`: Account with readonly permissions. The permissions are search/view/download if dufs allow search, otherwise view/download..
|
3. Use `,` to separate paths.
|
||||||
|
4. Use path suffix `:rw`, `:ro`, `:-` to set permissions: `read-write`, `read-only`, `forbidden`. `:ro` can be omitted.
|
||||||
|
|
||||||
|
- `-a admin:admin@/:rw`: `admin` has complete permissions for all paths.
|
||||||
|
- `-a guest:guest@/`: `guest` has read-only permissions for all paths.
|
||||||
|
- `-a user:pass@/:rw,/dir1,/dir2:-`: `user` has read-write permissions for `/*`, has read-only permissions for `/dir1/*`, but is fordden for `/dir2/*`.
|
||||||
|
- `-a @/`: All paths is publicly accessible, everyone can view/download it.
|
||||||
|
|
||||||
|
> There are no restrictions on using ':' and '@' characters in a password. For example, `user:pa:ss@1@/:rw` is valid, the password is `pa:ss@1`.
|
||||||
|
|
||||||
|
#### Hashed Password
|
||||||
|
|
||||||
|
DUFS supports the use of sha-512 hashed password.
|
||||||
|
|
||||||
|
Create hashed password
|
||||||
|
|
||||||
```
|
```
|
||||||
dufs -A -a /@admin:admin
|
$ mkpasswd -m sha-512 -s
|
||||||
|
Password: 123456
|
||||||
|
$6$qCAVUG7yn7t/hH4d$BWm8r5MoDywNmDP/J3V2S2a6flmKHC1IpblfoqZfuK.LtLBZ0KFXP9QIfJP8RqL8MCw4isdheoAMTuwOz.pAO/
|
||||||
```
|
```
|
||||||
`admin` has all permissions for all paths.
|
|
||||||
|
|
||||||
|
Use hashed password
|
||||||
```
|
```
|
||||||
dufs -A -a /@admin:admin@guest:guest
|
dufs -a 'admin:$6$qCAVUG7yn7t/hH4d$BWm8r5MoDywNmDP/J3V2S2a6flmKHC1IpblfoqZfuK.LtLBZ0KFXP9QIfJP8RqL8MCw4isdheoAMTuwOz.pAO/@/:rw'
|
||||||
```
|
```
|
||||||
`guest` has readonly permissions for all paths.
|
|
||||||
|
|
||||||
```
|
Two important things for hashed passwords:
|
||||||
dufs -A -a /@admin:admin@*
|
|
||||||
```
|
|
||||||
All paths is public, everyone can view/download it.
|
|
||||||
|
|
||||||
```
|
1. Dufs only supports sha-512 hashed passwords, so ensure that the password string always starts with `$6$`.
|
||||||
dufs -A -a /@admin:admin -a /user1@user1:pass1 -a /user2@pass2:user2
|
2. Digest authentication does not function properly with hashed passwords.
|
||||||
```
|
|
||||||
`user1` has all permissions for `/user1*` path.
|
|
||||||
`user2` has all permissions for `/user2*` path.
|
|
||||||
|
|
||||||
```
|
|
||||||
dufs -a /@admin:admin
|
|
||||||
```
|
|
||||||
Since dufs only allows viewing/downloading, `admin` can only view/download files.
|
|
||||||
|
|
||||||
### Hide Paths
|
### Hide Paths
|
||||||
|
|
||||||
@@ -252,9 +271,10 @@ dufs --hidden .git,.DS_Store,tmp
|
|||||||
> The glob used in --hidden only matches file and directory names, not paths. So `--hidden dir1/file` is invalid.
|
> The glob used in --hidden only matches file and directory names, not paths. So `--hidden dir1/file` is invalid.
|
||||||
|
|
||||||
```sh
|
```sh
|
||||||
dufs --hidden '.*' # hidden dotfiles
|
dufs --hidden '.*' # hidden dotfiles
|
||||||
dufs --hidden '*/' # hidden all folders
|
dufs --hidden '*/' # hidden all folders
|
||||||
dufs --hidden '*.log,*.lock' # hidden by exts
|
dufs --hidden '*.log,*.lock' # hidden by exts
|
||||||
|
dufs --hidden '*.log' --hidden '*.lock'
|
||||||
```
|
```
|
||||||
|
|
||||||
### Log Format
|
### Log Format
|
||||||
@@ -302,7 +322,66 @@ dufs --log-format '$remote_addr $remote_user "$request" $status' -a /@admin:admi
|
|||||||
|
|
||||||
All options can be set using environment variables prefixed with `DUFS_`.
|
All options can be set using environment variables prefixed with `DUFS_`.
|
||||||
|
|
||||||
`dufs --port 8080 --allow-all` is equal to `DUFS_PORT=8080 DUFS_ALLOW_ALL=true dufs`.
|
```
|
||||||
|
[serve-path] DUFS_SERVE_PATH="."
|
||||||
|
--config <file> DUFS_CONFIG=config.yaml
|
||||||
|
-b, --bind <addrs> DUFS_BIND=0.0.0.0
|
||||||
|
-p, --port <port> DUFS_PORT=5000
|
||||||
|
--path-prefix <path> DUFS_PATH_PREFIX=/static
|
||||||
|
--hidden <value> DUFS_HIDDEN=tmp,*.log,*.lock
|
||||||
|
-a, --auth <rules> DUFS_AUTH="admin:admin@/:rw|@/"
|
||||||
|
-A, --allow-all DUFS_ALLOW_ALL=true
|
||||||
|
--allow-upload DUFS_ALLOW_UPLOAD=true
|
||||||
|
--allow-delete DUFS_ALLOW_DELETE=true
|
||||||
|
--allow-search DUFS_ALLOW_SEARCH=true
|
||||||
|
--allow-symlink DUFS_ALLOW_SYMLINK=true
|
||||||
|
--allow-archive DUFS_ALLOW_ARCHIVE=true
|
||||||
|
--enable-cors DUFS_ENABLE_CORS=true
|
||||||
|
--render-index DUFS_RENDER_INDEX=true
|
||||||
|
--render-try-index DUFS_RENDER_TRY_INDEX=true
|
||||||
|
--render-spa DUFS_RENDER_SPA=true
|
||||||
|
--assets <path> DUFS_ASSETS=/assets
|
||||||
|
--log-format <format> DUFS_LOG_FORMAT=""
|
||||||
|
--compress <compress> DUFS_COMPRESS="low"
|
||||||
|
--tls-cert <path> DUFS_TLS_CERT=cert.pem
|
||||||
|
--tls-key <path> DUFS_TLS_KEY=key.pem
|
||||||
|
```
|
||||||
|
|
||||||
|
## Configuration File
|
||||||
|
|
||||||
|
You can specify and use the configuration file by selecting the option `--config <path-to-config.yaml>`.
|
||||||
|
|
||||||
|
The following are the configuration items:
|
||||||
|
|
||||||
|
```yaml
|
||||||
|
serve-path: '.'
|
||||||
|
bind: 0.0.0.0
|
||||||
|
port: 5000
|
||||||
|
path-prefix: /dufs
|
||||||
|
hidden:
|
||||||
|
- tmp
|
||||||
|
- '*.log'
|
||||||
|
- '*.lock'
|
||||||
|
auth:
|
||||||
|
- admin:admin@/:rw
|
||||||
|
- user:pass@/src:rw,/share
|
||||||
|
- '@/' # According to the YAML spec, quoting is required.
|
||||||
|
allow-all: false
|
||||||
|
allow-upload: true
|
||||||
|
allow-delete: true
|
||||||
|
allow-search: true
|
||||||
|
allow-symlink: true
|
||||||
|
allow-archive: true
|
||||||
|
enable-cors: true
|
||||||
|
render-index: true
|
||||||
|
render-try-index: true
|
||||||
|
render-spa: true
|
||||||
|
assets: ./assets/
|
||||||
|
log-format: '$remote_addr "$request" $status $http_user_agent'
|
||||||
|
compress: low
|
||||||
|
tls-cert: tests/data/cert.pem
|
||||||
|
tls-key: tests/data/key_pkcs1.pem
|
||||||
|
```
|
||||||
|
|
||||||
### Customize UI
|
### Customize UI
|
||||||
|
|
||||||
@@ -317,7 +396,7 @@ Your assets folder must contains a `index.html` file.
|
|||||||
`index.html` can use the following placeholder variables to retrieve internal data.
|
`index.html` can use the following placeholder variables to retrieve internal data.
|
||||||
|
|
||||||
- `__INDEX_DATA__`: directory listing data
|
- `__INDEX_DATA__`: directory listing data
|
||||||
- `__ASSERTS_PREFIX__`: assets url prefix
|
- `__ASSETS_PREFIX__`: assets url prefix
|
||||||
|
|
||||||
</details>
|
</details>
|
||||||
|
|
||||||
|
|||||||
@@ -73,6 +73,10 @@ body {
|
|||||||
display: none;
|
display: none;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
.upload-file label {
|
||||||
|
cursor: pointer;
|
||||||
|
}
|
||||||
|
|
||||||
.searchbar {
|
.searchbar {
|
||||||
display: flex;
|
display: flex;
|
||||||
flex-wrap: nowrap;
|
flex-wrap: nowrap;
|
||||||
@@ -103,11 +107,6 @@ body {
|
|||||||
cursor: pointer;
|
cursor: pointer;
|
||||||
}
|
}
|
||||||
|
|
||||||
.upload-status span {
|
|
||||||
width: 70px;
|
|
||||||
display: inline-block;
|
|
||||||
}
|
|
||||||
|
|
||||||
.main {
|
.main {
|
||||||
padding: 3.3em 1em 0;
|
padding: 3.3em 1em 0;
|
||||||
}
|
}
|
||||||
@@ -134,6 +133,10 @@ body {
|
|||||||
padding-left: 0.6em;
|
padding-left: 0.6em;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
.cell-status span {
|
||||||
|
display: inline-block;
|
||||||
|
}
|
||||||
|
|
||||||
.paths-table thead a {
|
.paths-table thead a {
|
||||||
color: unset;
|
color: unset;
|
||||||
text-decoration: none;
|
text-decoration: none;
|
||||||
@@ -148,7 +151,7 @@ body {
|
|||||||
}
|
}
|
||||||
|
|
||||||
.paths-table .cell-actions {
|
.paths-table .cell-actions {
|
||||||
width: 75px;
|
width: 90px;
|
||||||
display: flex;
|
display: flex;
|
||||||
padding-left: 0.6em;
|
padding-left: 0.6em;
|
||||||
}
|
}
|
||||||
@@ -208,22 +211,37 @@ body {
|
|||||||
height: calc(100vh - 5rem);
|
height: calc(100vh - 5rem);
|
||||||
border: 1px solid #ced4da;
|
border: 1px solid #ced4da;
|
||||||
outline: none;
|
outline: none;
|
||||||
|
padding: 5px;
|
||||||
}
|
}
|
||||||
|
|
||||||
.toolbox2 {
|
.toolbox-right {
|
||||||
margin-left: auto;
|
margin-left: auto;
|
||||||
margin-right: 2em;
|
margin-right: 2em;
|
||||||
}
|
}
|
||||||
|
|
||||||
.save-btn {
|
.save-btn {
|
||||||
cursor: pointer;
|
cursor: pointer;
|
||||||
|
-webkit-user-select: none;
|
||||||
user-select: none;
|
user-select: none;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
.user-btn {
|
||||||
|
display: flex;
|
||||||
|
align-items: center;
|
||||||
|
}
|
||||||
|
|
||||||
|
.user-name {
|
||||||
|
padding-left: 3px;
|
||||||
|
}
|
||||||
|
|
||||||
.not-editable {
|
.not-editable {
|
||||||
font-style: italic;
|
font-style: italic;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
.retry-btn {
|
||||||
|
cursor: pointer;
|
||||||
|
}
|
||||||
|
|
||||||
@media (min-width: 768px) {
|
@media (min-width: 768px) {
|
||||||
.path a {
|
.path a {
|
||||||
min-width: 400px;
|
min-width: 400px;
|
||||||
|
|||||||
@@ -1,15 +1,15 @@
|
|||||||
<!DOCTYPE html>
|
<!DOCTYPE html>
|
||||||
<html>
|
<html lang="en-US">
|
||||||
|
|
||||||
<head>
|
<head>
|
||||||
<meta charset="utf-8" />
|
<meta charset="utf-8" />
|
||||||
<meta name="viewport" content="width=device-width" />
|
<meta name="viewport" content="width=device-width" />
|
||||||
<link rel="icon" type="image/x-icon" href="__ASSERTS_PREFIX__favicon.ico">
|
<link rel="icon" type="image/x-icon" href="__ASSETS_PREFIX__favicon.ico">
|
||||||
<link rel="stylesheet" href="__ASSERTS_PREFIX__index.css">
|
<link rel="stylesheet" href="__ASSETS_PREFIX__index.css">
|
||||||
<script>
|
<script>
|
||||||
DATA = __INDEX_DATA__
|
DATA = __INDEX_DATA__
|
||||||
</script>
|
</script>
|
||||||
<script src="__ASSERTS_PREFIX__index.js"></script>
|
<script src="__ASSETS_PREFIX__index.js"></script>
|
||||||
</head>
|
</head>
|
||||||
|
|
||||||
<body>
|
<body>
|
||||||
@@ -48,7 +48,7 @@
|
|||||||
d="M7.646 1.146a.5.5 0 0 1 .708 0l3 3a.5.5 0 0 1-.708.708L8.5 2.707V11.5a.5.5 0 0 1-1 0V2.707L5.354 4.854a.5.5 0 1 1-.708-.708l3-3z" />
|
d="M7.646 1.146a.5.5 0 0 1 .708 0l3 3a.5.5 0 0 1-.708.708L8.5 2.707V11.5a.5.5 0 0 1-1 0V2.707L5.354 4.854a.5.5 0 1 1-.708-.708l3-3z" />
|
||||||
</svg>
|
</svg>
|
||||||
</label>
|
</label>
|
||||||
<input type="file" id="file" name="file" multiple>
|
<input type="file" id="file" title="Upload files" name="file" multiple>
|
||||||
</div>
|
</div>
|
||||||
<div class="control new-folder hidden" title="New folder">
|
<div class="control new-folder hidden" title="New folder">
|
||||||
<svg width="16" height="16" viewBox="0 0 16 16">
|
<svg width="16" height="16" viewBox="0 0 16 16">
|
||||||
@@ -74,10 +74,10 @@
|
|||||||
d="M11.742 10.344a6.5 6.5 0 1 0-1.397 1.398h-.001c.03.04.062.078.098.115l3.85 3.85a1 1 0 0 0 1.415-1.414l-3.85-3.85a1.007 1.007 0 0 0-.115-.1zM12 6.5a5.5 5.5 0 1 1-11 0 5.5 5.5 0 0 1 11 0z" />
|
d="M11.742 10.344a6.5 6.5 0 1 0-1.397 1.398h-.001c.03.04.062.078.098.115l3.85 3.85a1 1 0 0 0 1.415-1.414l-3.85-3.85a1.007 1.007 0 0 0-.115-.1zM12 6.5a5.5 5.5 0 1 1-11 0 5.5 5.5 0 0 1 11 0z" />
|
||||||
</svg>
|
</svg>
|
||||||
</div>
|
</div>
|
||||||
<input id="search" name="q" type="text" maxlength="128" autocomplete="off" tabindex="1">
|
<input id="search" title="Searching for folders or files" name="q" type="text" maxlength="128" autocomplete="off" tabindex="1">
|
||||||
<input type="submit" hidden />
|
<input type="submit" hidden />
|
||||||
</form>
|
</form>
|
||||||
<div class="toolbox2">
|
<div class="toolbox-right">
|
||||||
<div class="login-btn hidden" title="Login for upload/move/delete/edit permissions">
|
<div class="login-btn hidden" title="Login for upload/move/delete/edit permissions">
|
||||||
<svg width="16" height="16" viewBox="0 0 16 16">
|
<svg width="16" height="16" viewBox="0 0 16 16">
|
||||||
<path fill-rule="evenodd"
|
<path fill-rule="evenodd"
|
||||||
@@ -91,6 +91,7 @@
|
|||||||
<path
|
<path
|
||||||
d="M8 8a3 3 0 1 0 0-6 3 3 0 0 0 0 6Zm2-3a2 2 0 1 1-4 0 2 2 0 0 1 4 0Zm4 8c0 1-1 1-1 1H3s-1 0-1-1 1-4 6-4 6 3 6 4Zm-1-.004c-.001-.246-.154-.986-.832-1.664C11.516 10.68 10.289 10 8 10c-2.29 0-3.516.68-4.168 1.332-.678.678-.83 1.418-.832 1.664h10Z" />
|
d="M8 8a3 3 0 1 0 0-6 3 3 0 0 0 0 6Zm2-3a2 2 0 1 1-4 0 2 2 0 0 1 4 0Zm4 8c0 1-1 1-1 1H3s-1 0-1-1 1-4 6-4 6 3 6 4Zm-1-.004c-.001-.246-.154-.986-.832-1.664C11.516 10.68 10.289 10 8 10c-2.29 0-3.516.68-4.168 1.332-.678.678-.83 1.418-.832 1.664h10Z" />
|
||||||
</svg>
|
</svg>
|
||||||
|
<span class="user-name"></span>
|
||||||
</div>
|
</div>
|
||||||
<div class="save-btn hidden" title="Save file">
|
<div class="save-btn hidden" title="Save file">
|
||||||
<svg viewBox="0 0 1024 1024" width="24" height="24">
|
<svg viewBox="0 0 1024 1024" width="24" height="24">
|
||||||
@@ -121,7 +122,7 @@
|
|||||||
</div>
|
</div>
|
||||||
<div class="editor-page hidden">
|
<div class="editor-page hidden">
|
||||||
<div class="not-editable hidden"></div>
|
<div class="not-editable hidden"></div>
|
||||||
<textarea class="editor hidden" cols="10"></textarea>
|
<textarea id="editor" class="editor hidden" aria-label="Editor" cols="10"></textarea>
|
||||||
</div>
|
</div>
|
||||||
</div>
|
</div>
|
||||||
<script>
|
<script>
|
||||||
|
|||||||
301
assets/index.js
301
assets/index.js
@@ -10,7 +10,7 @@
|
|||||||
* @typedef {object} DATA
|
* @typedef {object} DATA
|
||||||
* @property {string} href
|
* @property {string} href
|
||||||
* @property {string} uri_prefix
|
* @property {string} uri_prefix
|
||||||
* @property {"Index" | "Edit"} kind
|
* @property {"Index" | "Edit" | "View"} kind
|
||||||
* @property {PathItem[]} paths
|
* @property {PathItem[]} paths
|
||||||
* @property {boolean} allow_upload
|
* @property {boolean} allow_upload
|
||||||
* @property {boolean} allow_delete
|
* @property {boolean} allow_delete
|
||||||
@@ -22,6 +22,8 @@
|
|||||||
* @property {string} editable
|
* @property {string} editable
|
||||||
*/
|
*/
|
||||||
|
|
||||||
|
var DUFS_MAX_UPLOADINGS = 1;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* @type {DATA} DATA
|
* @type {DATA} DATA
|
||||||
*/
|
*/
|
||||||
@@ -53,9 +55,15 @@ const ICONS = {
|
|||||||
download: `<svg width="16" height="16" viewBox="0 0 16 16"><path d="M.5 9.9a.5.5 0 0 1 .5.5v2.5a1 1 0 0 0 1 1h12a1 1 0 0 0 1-1v-2.5a.5.5 0 0 1 1 0v2.5a2 2 0 0 1-2 2H2a2 2 0 0 1-2-2v-2.5a.5.5 0 0 1 .5-.5z"/><path d="M7.646 11.854a.5.5 0 0 0 .708 0l3-3a.5.5 0 0 0-.708-.708L8.5 10.293V1.5a.5.5 0 0 0-1 0v8.793L5.354 8.146a.5.5 0 1 0-.708.708l3 3z"/></svg>`,
|
download: `<svg width="16" height="16" viewBox="0 0 16 16"><path d="M.5 9.9a.5.5 0 0 1 .5.5v2.5a1 1 0 0 0 1 1h12a1 1 0 0 0 1-1v-2.5a.5.5 0 0 1 1 0v2.5a2 2 0 0 1-2 2H2a2 2 0 0 1-2-2v-2.5a.5.5 0 0 1 .5-.5z"/><path d="M7.646 11.854a.5.5 0 0 0 .708 0l3-3a.5.5 0 0 0-.708-.708L8.5 10.293V1.5a.5.5 0 0 0-1 0v8.793L5.354 8.146a.5.5 0 1 0-.708.708l3 3z"/></svg>`,
|
||||||
move: `<svg width="16" height="16" viewBox="0 0 16 16"><path fill-rule="evenodd" d="M1.5 1.5A.5.5 0 0 0 1 2v4.8a2.5 2.5 0 0 0 2.5 2.5h9.793l-3.347 3.346a.5.5 0 0 0 .708.708l4.2-4.2a.5.5 0 0 0 0-.708l-4-4a.5.5 0 0 0-.708.708L13.293 8.3H3.5A1.5 1.5 0 0 1 2 6.8V2a.5.5 0 0 0-.5-.5z"/></svg>`,
|
move: `<svg width="16" height="16" viewBox="0 0 16 16"><path fill-rule="evenodd" d="M1.5 1.5A.5.5 0 0 0 1 2v4.8a2.5 2.5 0 0 0 2.5 2.5h9.793l-3.347 3.346a.5.5 0 0 0 .708.708l4.2-4.2a.5.5 0 0 0 0-.708l-4-4a.5.5 0 0 0-.708.708L13.293 8.3H3.5A1.5 1.5 0 0 1 2 6.8V2a.5.5 0 0 0-.5-.5z"/></svg>`,
|
||||||
edit: `<svg width="16" height="16" viewBox="0 0 16 16"><path d="M12.146.146a.5.5 0 0 1 .708 0l3 3a.5.5 0 0 1 0 .708l-10 10a.5.5 0 0 1-.168.11l-5 2a.5.5 0 0 1-.65-.65l2-5a.5.5 0 0 1 .11-.168l10-10zM11.207 2.5 13.5 4.793 14.793 3.5 12.5 1.207 11.207 2.5zm1.586 3L10.5 3.207 4 9.707V10h.5a.5.5 0 0 1 .5.5v.5h.5a.5.5 0 0 1 .5.5v.5h.293l6.5-6.5zm-9.761 5.175-.106.106-1.528 3.821 3.821-1.528.106-.106A.5.5 0 0 1 5 12.5V12h-.5a.5.5 0 0 1-.5-.5V11h-.5a.5.5 0 0 1-.468-.325z"/></svg>`,
|
edit: `<svg width="16" height="16" viewBox="0 0 16 16"><path d="M12.146.146a.5.5 0 0 1 .708 0l3 3a.5.5 0 0 1 0 .708l-10 10a.5.5 0 0 1-.168.11l-5 2a.5.5 0 0 1-.65-.65l2-5a.5.5 0 0 1 .11-.168l10-10zM11.207 2.5 13.5 4.793 14.793 3.5 12.5 1.207 11.207 2.5zm1.586 3L10.5 3.207 4 9.707V10h.5a.5.5 0 0 1 .5.5v.5h.5a.5.5 0 0 1 .5.5v.5h.293l6.5-6.5zm-9.761 5.175-.106.106-1.528 3.821 3.821-1.528.106-.106A.5.5 0 0 1 5 12.5V12h-.5a.5.5 0 0 1-.5-.5V11h-.5a.5.5 0 0 1-.468-.325z"/></svg>`,
|
||||||
delete: `<svg width="16" height="16" fill="currentColor"viewBox="0 0 16 16"><path d="M6.854 7.146a.5.5 0 1 0-.708.708L7.293 9l-1.147 1.146a.5.5 0 0 0 .708.708L8 9.707l1.146 1.147a.5.5 0 0 0 .708-.708L8.707 9l1.147-1.146a.5.5 0 0 0-.708-.708L8 8.293 6.854 7.146z"/><path d="M14 14V4.5L9.5 0H4a2 2 0 0 0-2 2v12a2 2 0 0 0 2 2h8a2 2 0 0 0 2-2zM9.5 3A1.5 1.5 0 0 0 11 4.5h2V14a1 1 0 0 1-1 1H4a1 1 0 0 1-1-1V2a1 1 0 0 1 1-1h5.5v2z"/></svg>`,
|
delete: `<svg width="16" height="16" viewBox="0 0 16 16"><path d="M6.854 7.146a.5.5 0 1 0-.708.708L7.293 9l-1.147 1.146a.5.5 0 0 0 .708.708L8 9.707l1.146 1.147a.5.5 0 0 0 .708-.708L8.707 9l1.147-1.146a.5.5 0 0 0-.708-.708L8 8.293 6.854 7.146z"/><path d="M14 14V4.5L9.5 0H4a2 2 0 0 0-2 2v12a2 2 0 0 0 2 2h8a2 2 0 0 0 2-2zM9.5 3A1.5 1.5 0 0 0 11 4.5h2V14a1 1 0 0 1-1 1H4a1 1 0 0 1-1-1V2a1 1 0 0 1 1-1h5.5v2z"/></svg>`,
|
||||||
|
view: `<svg width="16" height="16" viewBox="0 0 16 16"><path d="M4 0a2 2 0 0 0-2 2v12a2 2 0 0 0 2 2h8a2 2 0 0 0 2-2V2a2 2 0 0 0-2-2zm0 1h8a1 1 0 0 1 1 1v12a1 1 0 0 1-1 1H4a1 1 0 0 1-1-1V2a1 1 0 0 1 1-1"/></svg>`,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @type Map<string, Uploader>
|
||||||
|
*/
|
||||||
|
const failUploaders = new Map();
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* @type Element
|
* @type Element
|
||||||
*/
|
*/
|
||||||
@@ -84,6 +92,10 @@ let $editor;
|
|||||||
* @type Element
|
* @type Element
|
||||||
*/
|
*/
|
||||||
let $userBtn;
|
let $userBtn;
|
||||||
|
/**
|
||||||
|
* @type Element
|
||||||
|
*/
|
||||||
|
let $userName;
|
||||||
|
|
||||||
function ready() {
|
function ready() {
|
||||||
$pathsTable = document.querySelector(".paths-table")
|
$pathsTable = document.querySelector(".paths-table")
|
||||||
@@ -93,6 +105,7 @@ function ready() {
|
|||||||
$emptyFolder = document.querySelector(".empty-folder");
|
$emptyFolder = document.querySelector(".empty-folder");
|
||||||
$editor = document.querySelector(".editor");
|
$editor = document.querySelector(".editor");
|
||||||
$userBtn = document.querySelector(".user-btn");
|
$userBtn = document.querySelector(".user-btn");
|
||||||
|
$userName = document.querySelector(".user-name");
|
||||||
|
|
||||||
addBreadcrumb(DATA.href, DATA.uri_prefix);
|
addBreadcrumb(DATA.href, DATA.uri_prefix);
|
||||||
|
|
||||||
@@ -103,35 +116,41 @@ function ready() {
|
|||||||
setupIndexPage();
|
setupIndexPage();
|
||||||
|
|
||||||
} else if (DATA.kind == "Edit") {
|
} else if (DATA.kind == "Edit") {
|
||||||
document.title = `Edit of ${DATA.href} - Dufs`;
|
document.title = `Edit ${DATA.href} - Dufs`;
|
||||||
document.querySelector(".editor-page").classList.remove("hidden");;
|
document.querySelector(".editor-page").classList.remove("hidden");;
|
||||||
|
|
||||||
setupEditPage();
|
setupEditorPage();
|
||||||
|
} else if (DATA.kind == "View") {
|
||||||
|
document.title = `View ${DATA.href} - Dufs`;
|
||||||
|
document.querySelector(".editor-page").classList.remove("hidden");;
|
||||||
|
|
||||||
|
setupEditorPage();
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
class Uploader {
|
class Uploader {
|
||||||
/**
|
/**
|
||||||
*
|
*
|
||||||
* @param {File} file
|
* @param {File} file
|
||||||
* @param {string[]} dirs
|
* @param {string[]} pathParts
|
||||||
*/
|
*/
|
||||||
constructor(file, dirs) {
|
constructor(file, pathParts) {
|
||||||
/**
|
/**
|
||||||
* @type Element
|
* @type Element
|
||||||
*/
|
*/
|
||||||
this.$uploadStatus = null
|
this.$uploadStatus = null
|
||||||
this.uploaded = 0;
|
this.uploaded = 0;
|
||||||
|
this.uploadOffset = 0;
|
||||||
this.lastUptime = 0;
|
this.lastUptime = 0;
|
||||||
this.name = [...dirs, file.name].join("/");
|
this.name = [...pathParts, file.name].join("/");
|
||||||
this.idx = Uploader.globalIdx++;
|
this.idx = Uploader.globalIdx++;
|
||||||
this.file = file;
|
this.file = file;
|
||||||
|
this.url = newUrl(this.name);
|
||||||
}
|
}
|
||||||
|
|
||||||
upload() {
|
upload() {
|
||||||
const { idx, name } = this;
|
const { idx, name, url } = this;
|
||||||
const url = newUrl(name);
|
|
||||||
const encodedName = encodedStr(name);
|
const encodedName = encodedStr(name);
|
||||||
$uploadersTable.insertAdjacentHTML("beforeend", `
|
$uploadersTable.insertAdjacentHTML("beforeend", `
|
||||||
<tr id="upload${idx}" class="uploader">
|
<tr id="upload${idx}" class="uploader">
|
||||||
@@ -147,14 +166,25 @@ class Uploader {
|
|||||||
$emptyFolder.classList.add("hidden");
|
$emptyFolder.classList.add("hidden");
|
||||||
this.$uploadStatus = document.getElementById(`uploadStatus${idx}`);
|
this.$uploadStatus = document.getElementById(`uploadStatus${idx}`);
|
||||||
this.$uploadStatus.innerHTML = '-';
|
this.$uploadStatus.innerHTML = '-';
|
||||||
|
this.$uploadStatus.addEventListener("click", e => {
|
||||||
|
const nodeId = e.target.id;
|
||||||
|
const matches = /^retry(\d+)$/.exec(nodeId);
|
||||||
|
if (matches) {
|
||||||
|
const id = parseInt(matches[1]);
|
||||||
|
let uploader = failUploaders.get(id);
|
||||||
|
if (uploader) uploader.retry();
|
||||||
|
}
|
||||||
|
});
|
||||||
Uploader.queues.push(this);
|
Uploader.queues.push(this);
|
||||||
Uploader.runQueue();
|
Uploader.runQueue();
|
||||||
}
|
}
|
||||||
|
|
||||||
ajax() {
|
ajax() {
|
||||||
Uploader.runnings += 1;
|
const { url } = this;
|
||||||
const url = newUrl(this.name);
|
|
||||||
|
this.uploaded = 0;
|
||||||
this.lastUptime = Date.now();
|
this.lastUptime = Date.now();
|
||||||
|
|
||||||
const ajax = new XMLHttpRequest();
|
const ajax = new XMLHttpRequest();
|
||||||
ajax.upload.addEventListener("progress", e => this.progress(e), false);
|
ajax.upload.addEventListener("progress", e => this.progress(e), false);
|
||||||
ajax.addEventListener("readystatechange", () => {
|
ajax.addEventListener("readystatechange", () => {
|
||||||
@@ -162,38 +192,65 @@ class Uploader {
|
|||||||
if (ajax.status >= 200 && ajax.status < 300) {
|
if (ajax.status >= 200 && ajax.status < 300) {
|
||||||
this.complete();
|
this.complete();
|
||||||
} else {
|
} else {
|
||||||
this.fail();
|
if (ajax.status != 0) {
|
||||||
|
this.fail(`${ajax.status} ${ajax.statusText}`);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
})
|
})
|
||||||
ajax.addEventListener("error", () => this.fail(), false);
|
ajax.addEventListener("error", () => this.fail(), false);
|
||||||
ajax.addEventListener("abort", () => this.fail(), false);
|
ajax.addEventListener("abort", () => this.fail(), false);
|
||||||
ajax.open("PUT", url);
|
if (this.uploadOffset > 0) {
|
||||||
ajax.send(this.file);
|
ajax.open("PATCH", url);
|
||||||
|
ajax.setRequestHeader("X-Update-Range", "append");
|
||||||
|
ajax.send(this.file.slice(this.uploadOffset));
|
||||||
|
} else {
|
||||||
|
ajax.open("PUT", url);
|
||||||
|
ajax.send(this.file);
|
||||||
|
// setTimeout(() => ajax.abort(), 3000);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
async retry() {
|
||||||
|
const { url } = this;
|
||||||
|
let res = await fetch(url, {
|
||||||
|
method: "HEAD",
|
||||||
|
});
|
||||||
|
let uploadOffset = 0;
|
||||||
|
if (res.status == 200) {
|
||||||
|
let value = res.headers.get("content-length");
|
||||||
|
uploadOffset = parseInt(value) || 0;
|
||||||
|
}
|
||||||
|
this.uploadOffset = uploadOffset;
|
||||||
|
this.ajax()
|
||||||
|
}
|
||||||
|
|
||||||
progress(event) {
|
progress(event) {
|
||||||
const now = Date.now();
|
const now = Date.now();
|
||||||
const speed = (event.loaded - this.uploaded) / (now - this.lastUptime) * 1000;
|
const speed = (event.loaded - this.uploaded) / (now - this.lastUptime) * 1000;
|
||||||
const [speedValue, speedUnit] = formatSize(speed);
|
const [speedValue, speedUnit] = formatSize(speed);
|
||||||
const speedText = `${speedValue}${speedUnit.toLowerCase()}/s`;
|
const speedText = `${speedValue} ${speedUnit}/s`;
|
||||||
const progress = formatPercent((event.loaded / event.total) * 100);
|
const progress = formatPercent(((event.loaded + this.uploadOffset) / this.file.size) * 100);
|
||||||
const duration = formatDuration((event.total - event.loaded) / speed)
|
const duration = formatDuration((event.total - event.loaded) / speed)
|
||||||
this.$uploadStatus.innerHTML = `<span>${speedText}</span><span>${progress}</span><span>${duration}</span>`;
|
this.$uploadStatus.innerHTML = `<span style="width: 80px;">${speedText}</span><span>${progress} ${duration}</span>`;
|
||||||
this.uploaded = event.loaded;
|
this.uploaded = event.loaded;
|
||||||
this.lastUptime = now;
|
this.lastUptime = now;
|
||||||
}
|
}
|
||||||
|
|
||||||
complete() {
|
complete() {
|
||||||
this.$uploadStatus.innerHTML = `✓`;
|
const $uploadStatusNew = this.$uploadStatus.cloneNode(true);
|
||||||
Uploader.runnings -= 1;
|
$uploadStatusNew.innerHTML = `✓`;
|
||||||
|
this.$uploadStatus.parentNode.replaceChild($uploadStatusNew, this.$uploadStatus);
|
||||||
|
this.$uploadStatus = null;
|
||||||
|
failUploaders.delete(this.idx);
|
||||||
|
Uploader.runnings--;
|
||||||
Uploader.runQueue();
|
Uploader.runQueue();
|
||||||
}
|
}
|
||||||
|
|
||||||
fail() {
|
fail(reason = "") {
|
||||||
this.$uploadStatus.innerHTML = `✗`;
|
this.$uploadStatus.innerHTML = `<span style="width: 20px;" title="${reason}">✗</span><span class="retry-btn" id="retry${this.idx}" title="Retry">↻</span>`;
|
||||||
Uploader.runnings -= 1;
|
failUploaders.set(this.idx, this);
|
||||||
|
Uploader.runnings--;
|
||||||
Uploader.runQueue();
|
Uploader.runQueue();
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -211,19 +268,24 @@ Uploader.queues = [];
|
|||||||
|
|
||||||
|
|
||||||
Uploader.runQueue = async () => {
|
Uploader.runQueue = async () => {
|
||||||
if (Uploader.runnings > 2) return;
|
if (Uploader.runnings >= DUFS_MAX_UPLOADINGS) return;
|
||||||
|
if (Uploader.queues.length == 0) return;
|
||||||
|
Uploader.runnings++;
|
||||||
let uploader = Uploader.queues.shift();
|
let uploader = Uploader.queues.shift();
|
||||||
if (!uploader) return;
|
|
||||||
if (!Uploader.auth) {
|
if (!Uploader.auth) {
|
||||||
Uploader.auth = true;
|
Uploader.auth = true;
|
||||||
await login();
|
try {
|
||||||
|
await checkAuth()
|
||||||
|
} catch {
|
||||||
|
Uploader.auth = false;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
uploader.ajax();
|
uploader.ajax();
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Add breadcrumb
|
* Add breadcrumb
|
||||||
* @param {string} href
|
* @param {string} href
|
||||||
* @param {string} uri_prefix
|
* @param {string} uri_prefix
|
||||||
*/
|
*/
|
||||||
function addBreadcrumb(href, uri_prefix) {
|
function addBreadcrumb(href, uri_prefix) {
|
||||||
@@ -246,7 +308,7 @@ function addBreadcrumb(href, uri_prefix) {
|
|||||||
}
|
}
|
||||||
const encodedName = encodedStr(name);
|
const encodedName = encodedStr(name);
|
||||||
if (i === 0) {
|
if (i === 0) {
|
||||||
$breadcrumb.insertAdjacentHTML("beforeend", `<a href="${path}"><svg width="16" height="16" viewBox="0 0 16 16"><path d="M6.5 14.5v-3.505c0-.245.25-.495.5-.495h2c.25 0 .5.25.5.5v3.5a.5.5 0 0 0 .5.5h4a.5.5 0 0 0 .5-.5v-7a.5.5 0 0 0-.146-.354L13 5.793V2.5a.5.5 0 0 0-.5-.5h-1a.5.5 0 0 0-.5.5v1.293L8.354 1.146a.5.5 0 0 0-.708 0l-6 6A.5.5 0 0 0 1.5 7.5v7a.5.5 0 0 0 .5.5h4a.5.5 0 0 0 .5-.5z"/></svg></a>`);
|
$breadcrumb.insertAdjacentHTML("beforeend", `<a href="${path}" title="Root"><svg width="16" height="16" viewBox="0 0 16 16"><path d="M6.5 14.5v-3.505c0-.245.25-.495.5-.495h2c.25 0 .5.25.5.5v3.5a.5.5 0 0 0 .5.5h4a.5.5 0 0 0 .5-.5v-7a.5.5 0 0 0-.146-.354L13 5.793V2.5a.5.5 0 0 0-.5-.5h-1a.5.5 0 0 0-.5.5v1.293L8.354 1.146a.5.5 0 0 0-.708 0l-6 6A.5.5 0 0 0 1.5 7.5v7a.5.5 0 0 0 .5.5h4a.5.5 0 0 0 .5-.5z"/></svg></a>`);
|
||||||
} else if (i === len - 1) {
|
} else if (i === len - 1) {
|
||||||
$breadcrumb.insertAdjacentHTML("beforeend", `<b>${encodedName}</b>`);
|
$breadcrumb.insertAdjacentHTML("beforeend", `<b>${encodedName}</b>`);
|
||||||
} else {
|
} else {
|
||||||
@@ -262,7 +324,7 @@ function setupIndexPage() {
|
|||||||
if (DATA.allow_archive) {
|
if (DATA.allow_archive) {
|
||||||
const $download = document.querySelector(".download");
|
const $download = document.querySelector(".download");
|
||||||
$download.href = baseUrl() + "?zip";
|
$download.href = baseUrl() + "?zip";
|
||||||
$download.title = "Download folder as div .zip file";
|
$download.title = "Download folder as a .zip file";
|
||||||
$download.classList.remove("hidden");
|
$download.classList.remove("hidden");
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -310,13 +372,13 @@ function renderPathsTableHead() {
|
|||||||
<tr>
|
<tr>
|
||||||
${headerItems.map(item => {
|
${headerItems.map(item => {
|
||||||
let svg = `<svg width="12" height="12" viewBox="0 0 16 16"><path fill-rule="evenodd" d="M11.5 15a.5.5 0 0 0 .5-.5V2.707l3.146 3.147a.5.5 0 0 0 .708-.708l-4-4a.5.5 0 0 0-.708 0l-4 4a.5.5 0 1 0 .708.708L11 2.707V14.5a.5.5 0 0 0 .5.5zm-7-14a.5.5 0 0 1 .5.5v11.793l3.146-3.147a.5.5 0 0 1 .708.708l-4 4a.5.5 0 0 1-.708 0l-4-4a.5.5 0 0 1 .708-.708L4 13.293V1.5a.5.5 0 0 1 .5-.5z"/></svg>`;
|
let svg = `<svg width="12" height="12" viewBox="0 0 16 16"><path fill-rule="evenodd" d="M11.5 15a.5.5 0 0 0 .5-.5V2.707l3.146 3.147a.5.5 0 0 0 .708-.708l-4-4a.5.5 0 0 0-.708 0l-4 4a.5.5 0 1 0 .708.708L11 2.707V14.5a.5.5 0 0 0 .5.5zm-7-14a.5.5 0 0 1 .5.5v11.793l3.146-3.147a.5.5 0 0 1 .708.708l-4 4a.5.5 0 0 1-.708 0l-4-4a.5.5 0 0 1 .708-.708L4 13.293V1.5a.5.5 0 0 1 .5-.5z"/></svg>`;
|
||||||
let order = "asc";
|
let order = "desc";
|
||||||
if (PARAMS.sort === item.name) {
|
if (PARAMS.sort === item.name) {
|
||||||
if (PARAMS.order === "asc") {
|
if (PARAMS.order === "desc") {
|
||||||
order = "desc";
|
order = "asc";
|
||||||
svg = `<svg width="12" height="12" viewBox="0 0 16 16"><path fill-rule="evenodd" d="M8 15a.5.5 0 0 0 .5-.5V2.707l3.146 3.147a.5.5 0 0 0 .708-.708l-4-4a.5.5 0 0 0-.708 0l-4 4a.5.5 0 1 0 .708.708L7.5 2.707V14.5a.5.5 0 0 0 .5.5z"/></svg>`
|
|
||||||
} else {
|
|
||||||
svg = `<svg width="12" height="12" viewBox="0 0 16 16"><path fill-rule="evenodd" d="M8 1a.5.5 0 0 1 .5.5v11.793l3.146-3.147a.5.5 0 0 1 .708.708l-4 4a.5.5 0 0 1-.708 0l-4-4a.5.5 0 0 1 .708-.708L7.5 13.293V1.5A.5.5 0 0 1 8 1z"/></svg>`
|
svg = `<svg width="12" height="12" viewBox="0 0 16 16"><path fill-rule="evenodd" d="M8 1a.5.5 0 0 1 .5.5v11.793l3.146-3.147a.5.5 0 0 1 .708.708l-4 4a.5.5 0 0 1-.708 0l-4-4a.5.5 0 0 1 .708-.708L7.5 13.293V1.5A.5.5 0 0 1 8 1z"/></svg>`
|
||||||
|
} else {
|
||||||
|
svg = `<svg width="12" height="12" viewBox="0 0 16 16"><path fill-rule="evenodd" d="M8 15a.5.5 0 0 0 .5-.5V2.707l3.146 3.147a.5.5 0 0 0 .708-.708l-4-4a.5.5 0 0 0-.708 0l-4 4a.5.5 0 1 0 .708.708L7.5 2.707V14.5a.5.5 0 0 0 .5.5z"/></svg>`
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
const qs = new URLSearchParams({ ...PARAMS, order, sort: item.name }).toString();
|
const qs = new URLSearchParams({ ...PARAMS, order, sort: item.name }).toString();
|
||||||
@@ -348,8 +410,8 @@ function renderPathsTableBody() {
|
|||||||
|
|
||||||
/**
|
/**
|
||||||
* Add pathitem
|
* Add pathitem
|
||||||
* @param {PathItem} file
|
* @param {PathItem} file
|
||||||
* @param {number} index
|
* @param {number} index
|
||||||
*/
|
*/
|
||||||
function addPath(file, index) {
|
function addPath(file, index) {
|
||||||
const encodedName = encodedStr(file.name);
|
const encodedName = encodedStr(file.name);
|
||||||
@@ -358,6 +420,7 @@ function addPath(file, index) {
|
|||||||
let actionDownload = "";
|
let actionDownload = "";
|
||||||
let actionMove = "";
|
let actionMove = "";
|
||||||
let actionEdit = "";
|
let actionEdit = "";
|
||||||
|
let actionView = "";
|
||||||
let isDir = file.path_type.endsWith("Dir");
|
let isDir = file.path_type.endsWith("Dir");
|
||||||
if (isDir) {
|
if (isDir) {
|
||||||
url += "/";
|
url += "/";
|
||||||
@@ -375,21 +438,24 @@ function addPath(file, index) {
|
|||||||
}
|
}
|
||||||
if (DATA.allow_delete) {
|
if (DATA.allow_delete) {
|
||||||
if (DATA.allow_upload) {
|
if (DATA.allow_upload) {
|
||||||
if (isDir) {
|
actionMove = `<div onclick="movePath(${index})" class="action-btn" id="moveBtn${index}" title="Move to new path">${ICONS.move}</div>`;
|
||||||
actionMove = `<div onclick="movePath(${index})" class="action-btn" id="moveBtn${index}" title="Move to new path">${ICONS.move}</div>`;
|
if (!isDir) {
|
||||||
} else {
|
|
||||||
actionEdit = `<a class="action-btn" title="Edit file" target="_blank" href="${url}?edit">${ICONS.edit}</a>`;
|
actionEdit = `<a class="action-btn" title="Edit file" target="_blank" href="${url}?edit">${ICONS.edit}</a>`;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
actionDelete = `
|
actionDelete = `
|
||||||
<div onclick="deletePath(${index})" class="action-btn" id="deleteBtn${index}" title="Delete">${ICONS.delete}</div>`;
|
<div onclick="deletePath(${index})" class="action-btn" id="deleteBtn${index}" title="Delete">${ICONS.delete}</div>`;
|
||||||
}
|
}
|
||||||
|
if (!actionEdit && !isDir) {
|
||||||
|
actionView = `<a class="action-btn" title="View file" target="_blank" href="${url}?view">${ICONS.view}</a>`;
|
||||||
|
}
|
||||||
let actionCell = `
|
let actionCell = `
|
||||||
<td class="cell-actions">
|
<td class="cell-actions">
|
||||||
${actionDownload}
|
${actionDownload}
|
||||||
|
${actionView}
|
||||||
${actionMove}
|
${actionMove}
|
||||||
${actionEdit}
|
|
||||||
${actionDelete}
|
${actionDelete}
|
||||||
|
${actionEdit}
|
||||||
</td>`
|
</td>`
|
||||||
|
|
||||||
$pathsTableBody.insertAdjacentHTML("beforeend", `
|
$pathsTableBody.insertAdjacentHTML("beforeend", `
|
||||||
@@ -433,11 +499,18 @@ function setupDropzone() {
|
|||||||
function setupAuth() {
|
function setupAuth() {
|
||||||
if (DATA.user) {
|
if (DATA.user) {
|
||||||
$userBtn.classList.remove("hidden");
|
$userBtn.classList.remove("hidden");
|
||||||
$userBtn.title = DATA.user;
|
$userName.textContent = DATA.user;
|
||||||
} else {
|
} else {
|
||||||
const $loginBtn = document.querySelector(".login-btn");
|
const $loginBtn = document.querySelector(".login-btn");
|
||||||
$loginBtn.classList.remove("hidden");
|
$loginBtn.classList.remove("hidden");
|
||||||
$loginBtn.addEventListener("click", () => login(true));
|
$loginBtn.addEventListener("click", async () => {
|
||||||
|
try {
|
||||||
|
await checkAuth()
|
||||||
|
location.reload();
|
||||||
|
} catch (err) {
|
||||||
|
alert(err.message);
|
||||||
|
}
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -487,32 +560,40 @@ function setupNewFile() {
|
|||||||
});
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
async function setupEditPage() {
|
async function setupEditorPage() {
|
||||||
const url = baseUrl();
|
const url = baseUrl();
|
||||||
|
|
||||||
const $download = document.querySelector(".download");
|
const $download = document.querySelector(".download");
|
||||||
$download.classList.remove("hidden");
|
$download.classList.remove("hidden");
|
||||||
$download.href = url;
|
$download.href = url;
|
||||||
|
|
||||||
const $moveFile = document.querySelector(".move-file");
|
if (DATA.kind == "Edit") {
|
||||||
$moveFile.classList.remove("hidden");
|
const $moveFile = document.querySelector(".move-file");
|
||||||
$moveFile.addEventListener("click", async () => {
|
$moveFile.classList.remove("hidden");
|
||||||
const query = location.href.slice(url.length);
|
$moveFile.addEventListener("click", async () => {
|
||||||
const newFileUrl = await doMovePath(url);
|
const query = location.href.slice(url.length);
|
||||||
if (newFileUrl) {
|
const newFileUrl = await doMovePath(url);
|
||||||
location.href = newFileUrl + query;
|
if (newFileUrl) {
|
||||||
}
|
location.href = newFileUrl + query;
|
||||||
});
|
}
|
||||||
|
|
||||||
const $deleteFile = document.querySelector(".delete-file");
|
|
||||||
$deleteFile.classList.remove("hidden");
|
|
||||||
$deleteFile.addEventListener("click", async () => {
|
|
||||||
const url = baseUrl();
|
|
||||||
const name = baseName(url);
|
|
||||||
await doDeletePath(name, url, () => {
|
|
||||||
location.href = location.href.split("/").slice(0, -1).join("/");
|
|
||||||
});
|
});
|
||||||
})
|
|
||||||
|
const $deleteFile = document.querySelector(".delete-file");
|
||||||
|
$deleteFile.classList.remove("hidden");
|
||||||
|
$deleteFile.addEventListener("click", async () => {
|
||||||
|
const url = baseUrl();
|
||||||
|
const name = baseName(url);
|
||||||
|
await doDeletePath(name, url, () => {
|
||||||
|
location.href = location.href.split("/").slice(0, -1).join("/");
|
||||||
|
});
|
||||||
|
})
|
||||||
|
|
||||||
|
const $saveBtn = document.querySelector(".save-btn");
|
||||||
|
$saveBtn.classList.remove("hidden");
|
||||||
|
$saveBtn.addEventListener("click", saveChange);
|
||||||
|
} else if (DATA.kind == "View") {
|
||||||
|
$editor.readonly = true;
|
||||||
|
}
|
||||||
|
|
||||||
if (!DATA.editable) {
|
if (!DATA.editable) {
|
||||||
const $notEditable = document.querySelector(".not-editable");
|
const $notEditable = document.querySelector(".not-editable");
|
||||||
@@ -522,21 +603,24 @@ async function setupEditPage() {
|
|||||||
$notEditable.insertAdjacentHTML("afterend", `<iframe src="${url}" sandbox width="100%" height="${window.innerHeight - 100}px"></iframe>`)
|
$notEditable.insertAdjacentHTML("afterend", `<iframe src="${url}" sandbox width="100%" height="${window.innerHeight - 100}px"></iframe>`)
|
||||||
} else {
|
} else {
|
||||||
$notEditable.classList.remove("hidden");
|
$notEditable.classList.remove("hidden");
|
||||||
$notEditable.textContent = "Cannot edit because it is too large or binary.";
|
$notEditable.textContent = "Cannot edit because file is too large or binary.";
|
||||||
}
|
}
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
const $saveBtn = document.querySelector(".save-btn");
|
|
||||||
$saveBtn.classList.remove("hidden");
|
|
||||||
$saveBtn.addEventListener("click", saveChange);
|
|
||||||
|
|
||||||
$editor.classList.remove("hidden");
|
$editor.classList.remove("hidden");
|
||||||
try {
|
try {
|
||||||
const res = await fetch(baseUrl());
|
const res = await fetch(baseUrl());
|
||||||
await assertResOK(res);
|
await assertResOK(res);
|
||||||
const text = await res.text();
|
const encoding = getEncoding(res.headers.get("content-type"));
|
||||||
$editor.value = text;
|
if (encoding === "utf-8") {
|
||||||
|
$editor.value = await res.text();
|
||||||
|
} else {
|
||||||
|
const bytes = await res.arrayBuffer();
|
||||||
|
const dataView = new DataView(bytes)
|
||||||
|
const decoder = new TextDecoder(encoding)
|
||||||
|
$editor.value = decoder.decode(dataView);
|
||||||
|
}
|
||||||
} catch (err) {
|
} catch (err) {
|
||||||
alert(`Failed get file, ${err.message}`);
|
alert(`Failed get file, ${err.message}`);
|
||||||
}
|
}
|
||||||
@@ -544,8 +628,8 @@ async function setupEditPage() {
|
|||||||
|
|
||||||
/**
|
/**
|
||||||
* Delete path
|
* Delete path
|
||||||
* @param {number} index
|
* @param {number} index
|
||||||
* @returns
|
* @returns
|
||||||
*/
|
*/
|
||||||
async function deletePath(index) {
|
async function deletePath(index) {
|
||||||
const file = DATA.paths[index];
|
const file = DATA.paths[index];
|
||||||
@@ -564,7 +648,7 @@ async function deletePath(index) {
|
|||||||
async function doDeletePath(name, url, cb) {
|
async function doDeletePath(name, url, cb) {
|
||||||
if (!confirm(`Delete \`${name}\`?`)) return;
|
if (!confirm(`Delete \`${name}\`?`)) return;
|
||||||
try {
|
try {
|
||||||
await login();
|
await checkAuth();
|
||||||
const res = await fetch(url, {
|
const res = await fetch(url, {
|
||||||
method: "DELETE",
|
method: "DELETE",
|
||||||
});
|
});
|
||||||
@@ -577,8 +661,8 @@ async function doDeletePath(name, url, cb) {
|
|||||||
|
|
||||||
/**
|
/**
|
||||||
* Move path
|
* Move path
|
||||||
* @param {number} index
|
* @param {number} index
|
||||||
* @returns
|
* @returns
|
||||||
*/
|
*/
|
||||||
async function movePath(index) {
|
async function movePath(index) {
|
||||||
const file = DATA.paths[index];
|
const file = DATA.paths[index];
|
||||||
@@ -604,12 +688,12 @@ async function doMovePath(fileUrl) {
|
|||||||
const newFileUrl = fileUrlObj.origin + prefix + newPath.split("/").map(encodeURIComponent).join("/");
|
const newFileUrl = fileUrlObj.origin + prefix + newPath.split("/").map(encodeURIComponent).join("/");
|
||||||
|
|
||||||
try {
|
try {
|
||||||
await login();
|
await checkAuth();
|
||||||
const res1 = await fetch(newFileUrl, {
|
const res1 = await fetch(newFileUrl, {
|
||||||
method: "HEAD",
|
method: "HEAD",
|
||||||
});
|
});
|
||||||
if (res1.status === 200) {
|
if (res1.status === 200) {
|
||||||
if (!confirm("Override exsis file?")) {
|
if (!confirm("Override existing file?")) {
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -636,39 +720,31 @@ async function saveChange() {
|
|||||||
method: "PUT",
|
method: "PUT",
|
||||||
body: $editor.value,
|
body: $editor.value,
|
||||||
});
|
});
|
||||||
|
location.reload();
|
||||||
} catch (err) {
|
} catch (err) {
|
||||||
alert(`Failed to save file, ${err.message}`);
|
alert(`Failed to save file, ${err.message}`);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
async function login(alert = false) {
|
async function checkAuth() {
|
||||||
if (!DATA.auth) return;
|
if (!DATA.auth) return;
|
||||||
try {
|
const res = await fetch(baseUrl(), {
|
||||||
const res = await fetch(baseUrl(), {
|
method: "WRITEABLE",
|
||||||
method: "WRITEABLE",
|
});
|
||||||
});
|
await assertResOK(res);
|
||||||
await assertResOK(res);
|
document.querySelector(".login-btn").classList.add("hidden");
|
||||||
document.querySelector(".login-btn").classList.add("hidden");
|
$userBtn.classList.remove("hidden");
|
||||||
$userBtn.classList.remove("hidden");
|
$userName.textContent = "";
|
||||||
$userBtn.title = "";
|
|
||||||
} catch (err) {
|
|
||||||
let message = `Cannot login, ${err.message}`;
|
|
||||||
if (alert) {
|
|
||||||
alert(message);
|
|
||||||
} else {
|
|
||||||
throw new Error(message);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Create a folder
|
* Create a folder
|
||||||
* @param {string} name
|
* @param {string} name
|
||||||
*/
|
*/
|
||||||
async function createFolder(name) {
|
async function createFolder(name) {
|
||||||
const url = newUrl(name);
|
const url = newUrl(name);
|
||||||
try {
|
try {
|
||||||
await login();
|
await checkAuth();
|
||||||
const res = await fetch(url, {
|
const res = await fetch(url, {
|
||||||
method: "MKCOL",
|
method: "MKCOL",
|
||||||
});
|
});
|
||||||
@@ -682,7 +758,7 @@ async function createFolder(name) {
|
|||||||
async function createFile(name) {
|
async function createFile(name) {
|
||||||
const url = newUrl(name);
|
const url = newUrl(name);
|
||||||
try {
|
try {
|
||||||
await login();
|
await checkAuth();
|
||||||
const res = await fetch(url, {
|
const res = await fetch(url, {
|
||||||
method: "PUT",
|
method: "PUT",
|
||||||
body: "",
|
body: "",
|
||||||
@@ -701,8 +777,16 @@ async function addFileEntries(entries, dirs) {
|
|||||||
new Uploader(file, dirs).upload();
|
new Uploader(file, dirs).upload();
|
||||||
});
|
});
|
||||||
} else if (entry.isDirectory) {
|
} else if (entry.isDirectory) {
|
||||||
const dirReader = entry.createReader()
|
const dirReader = entry.createReader();
|
||||||
dirReader.readEntries(entries => addFileEntries(entries, [...dirs, entry.name]));
|
|
||||||
|
const successCallback = entries => {
|
||||||
|
if (entries.length > 0) {
|
||||||
|
addFileEntries(entries, [...dirs, entry.name]);
|
||||||
|
dirReader.readEntries(successCallback);
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
dirReader.readEntries(successCallback);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -762,11 +846,15 @@ function padZero(value, size) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
function formatSize(size) {
|
function formatSize(size) {
|
||||||
if (size == null) return []
|
if (size == null) return [0, "B"]
|
||||||
const sizes = ['B', 'KB', 'MB', 'GB', 'TB'];
|
const sizes = ['B', 'KB', 'MB', 'GB', 'TB'];
|
||||||
if (size == 0) return [0, "B"];
|
if (size == 0) return [0, "B"];
|
||||||
const i = parseInt(Math.floor(Math.log(size) / Math.log(1024)));
|
const i = parseInt(Math.floor(Math.log(size) / Math.log(1024)));
|
||||||
return [Math.round(size / Math.pow(1024, i), 2), sizes[i]];
|
ratio = 1
|
||||||
|
if (i >= 3) {
|
||||||
|
ratio = 100
|
||||||
|
}
|
||||||
|
return [Math.round(size * ratio / Math.pow(1024, i), 2) / ratio, sizes[i]];
|
||||||
}
|
}
|
||||||
|
|
||||||
function formatDuration(seconds) {
|
function formatDuration(seconds) {
|
||||||
@@ -793,6 +881,17 @@ function encodedStr(rawStr) {
|
|||||||
|
|
||||||
async function assertResOK(res) {
|
async function assertResOK(res) {
|
||||||
if (!(res.status >= 200 && res.status < 300)) {
|
if (!(res.status >= 200 && res.status < 300)) {
|
||||||
throw new Error(await res.text())
|
throw new Error(await res.text() || `Invalid status ${res.status}`);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
function getEncoding(contentType) {
|
||||||
|
const charset = contentType?.split(";")[1];
|
||||||
|
if (/charset/i.test(charset)) {
|
||||||
|
let encoding = charset.split("=")[1];
|
||||||
|
if (encoding) {
|
||||||
|
return encoding.toLowerCase()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return 'utf-8'
|
||||||
|
}
|
||||||
|
|||||||
632
src/args.rs
632
src/args.rs
@@ -1,18 +1,16 @@
|
|||||||
use anyhow::{bail, Context, Result};
|
use anyhow::{bail, Context, Result};
|
||||||
use clap::builder::PossibleValuesParser;
|
use async_zip::Compression;
|
||||||
use clap::{value_parser, Arg, ArgAction, ArgMatches, Command};
|
use clap::builder::{PossibleValue, PossibleValuesParser};
|
||||||
|
use clap::{value_parser, Arg, ArgAction, ArgMatches, Command, ValueEnum};
|
||||||
use clap_complete::{generate, Generator, Shell};
|
use clap_complete::{generate, Generator, Shell};
|
||||||
#[cfg(feature = "tls")]
|
use serde::{Deserialize, Deserializer};
|
||||||
use rustls::{Certificate, PrivateKey};
|
use smart_default::SmartDefault;
|
||||||
use std::env;
|
use std::env;
|
||||||
use std::net::IpAddr;
|
use std::net::IpAddr;
|
||||||
use std::path::{Path, PathBuf};
|
use std::path::{Path, PathBuf};
|
||||||
|
|
||||||
use crate::auth::AccessControl;
|
use crate::auth::AccessControl;
|
||||||
use crate::auth::AuthMethod;
|
use crate::http_logger::HttpLogger;
|
||||||
use crate::log_http::{LogHttp, DEFAULT_LOG_FORMAT};
|
|
||||||
#[cfg(feature = "tls")]
|
|
||||||
use crate::tls::{load_certs, load_private_key};
|
|
||||||
use crate::utils::encode_uri;
|
use crate::utils::encode_uri;
|
||||||
|
|
||||||
pub fn build_cli() -> Command {
|
pub fn build_cli() -> Command {
|
||||||
@@ -25,12 +23,21 @@ pub fn build_cli() -> Command {
|
|||||||
env!("CARGO_PKG_REPOSITORY")
|
env!("CARGO_PKG_REPOSITORY")
|
||||||
))
|
))
|
||||||
.arg(
|
.arg(
|
||||||
Arg::new("root")
|
Arg::new("serve-path")
|
||||||
.env("DUFS_ROOT")
|
.env("DUFS_SERVE_PATH")
|
||||||
.hide_env(true)
|
.hide_env(true)
|
||||||
.default_value(".")
|
|
||||||
.value_parser(value_parser!(PathBuf))
|
.value_parser(value_parser!(PathBuf))
|
||||||
.help("Specific path to serve"),
|
.help("Specific path to serve [default: .]"),
|
||||||
|
)
|
||||||
|
.arg(
|
||||||
|
Arg::new("config")
|
||||||
|
.env("DUFS_CONFIG")
|
||||||
|
.hide_env(true)
|
||||||
|
.short('c')
|
||||||
|
.long("config")
|
||||||
|
.value_parser(value_parser!(PathBuf))
|
||||||
|
.help("Specify configuration file")
|
||||||
|
.value_name("file"),
|
||||||
)
|
)
|
||||||
.arg(
|
.arg(
|
||||||
Arg::new("bind")
|
Arg::new("bind")
|
||||||
@@ -49,9 +56,8 @@ pub fn build_cli() -> Command {
|
|||||||
.hide_env(true)
|
.hide_env(true)
|
||||||
.short('p')
|
.short('p')
|
||||||
.long("port")
|
.long("port")
|
||||||
.default_value("5000")
|
|
||||||
.value_parser(value_parser!(u16))
|
.value_parser(value_parser!(u16))
|
||||||
.help("Specify port to listen on")
|
.help("Specify port to listen on [default: 5000]")
|
||||||
.value_name("port"),
|
.value_name("port"),
|
||||||
)
|
)
|
||||||
.arg(
|
.arg(
|
||||||
@@ -67,7 +73,9 @@ pub fn build_cli() -> Command {
|
|||||||
.env("DUFS_HIDDEN")
|
.env("DUFS_HIDDEN")
|
||||||
.hide_env(true)
|
.hide_env(true)
|
||||||
.long("hidden")
|
.long("hidden")
|
||||||
.help("Hide paths from directory listings, separated by `,`")
|
.action(ArgAction::Append)
|
||||||
|
.value_delimiter(',')
|
||||||
|
.help("Hide paths from directory listings, e.g. tmp,*.log,*.lock")
|
||||||
.value_name("value"),
|
.value_name("value"),
|
||||||
)
|
)
|
||||||
.arg(
|
.arg(
|
||||||
@@ -76,13 +84,13 @@ pub fn build_cli() -> Command {
|
|||||||
.hide_env(true)
|
.hide_env(true)
|
||||||
.short('a')
|
.short('a')
|
||||||
.long("auth")
|
.long("auth")
|
||||||
.help("Add auth for path")
|
.help("Add auth roles, e.g. user:pass@/dir1:rw,/dir2")
|
||||||
.action(ArgAction::Append)
|
.action(ArgAction::Append)
|
||||||
.value_delimiter(',')
|
|
||||||
.value_name("rules"),
|
.value_name("rules"),
|
||||||
)
|
)
|
||||||
.arg(
|
.arg(
|
||||||
Arg::new("auth-method")
|
Arg::new("auth-method")
|
||||||
|
.hide(true)
|
||||||
.env("DUFS_AUTH_METHOD")
|
.env("DUFS_AUTH_METHOD")
|
||||||
.hide_env(true)
|
.hide_env(true)
|
||||||
.long("auth-method")
|
.long("auth-method")
|
||||||
@@ -177,9 +185,33 @@ pub fn build_cli() -> Command {
|
|||||||
.env("DUFS_ASSETS")
|
.env("DUFS_ASSETS")
|
||||||
.hide_env(true)
|
.hide_env(true)
|
||||||
.long("assets")
|
.long("assets")
|
||||||
.help("Use custom assets to override builtin assets")
|
.help("Set the path to the assets directory for overriding the built-in assets")
|
||||||
.value_parser(value_parser!(PathBuf))
|
.value_parser(value_parser!(PathBuf))
|
||||||
.value_name("path")
|
.value_name("path")
|
||||||
|
)
|
||||||
|
.arg(
|
||||||
|
Arg::new("log-format")
|
||||||
|
.env("DUFS_LOG_FORMAT")
|
||||||
|
.hide_env(true)
|
||||||
|
.long("log-format")
|
||||||
|
.value_name("format")
|
||||||
|
.help("Customize http log format"),
|
||||||
|
)
|
||||||
|
.arg(
|
||||||
|
Arg::new("compress")
|
||||||
|
.env("DUFS_COMPRESS")
|
||||||
|
.hide_env(true)
|
||||||
|
.value_parser(clap::builder::EnumValueParser::<Compress>::new())
|
||||||
|
.long("compress")
|
||||||
|
.value_name("level")
|
||||||
|
.help("Set zip compress level [default: low]")
|
||||||
|
)
|
||||||
|
.arg(
|
||||||
|
Arg::new("completions")
|
||||||
|
.long("completions")
|
||||||
|
.value_name("shell")
|
||||||
|
.value_parser(value_parser!(Shell))
|
||||||
|
.help("Print shell completion script for <shell>"),
|
||||||
);
|
);
|
||||||
|
|
||||||
#[cfg(feature = "tls")]
|
#[cfg(feature = "tls")]
|
||||||
@@ -203,38 +235,38 @@ pub fn build_cli() -> Command {
|
|||||||
.help("Path to the SSL/TLS certificate's private key"),
|
.help("Path to the SSL/TLS certificate's private key"),
|
||||||
);
|
);
|
||||||
|
|
||||||
app.arg(
|
app
|
||||||
Arg::new("log-format")
|
|
||||||
.env("DUFS_LOG_FORMAT")
|
|
||||||
.hide_env(true)
|
|
||||||
.long("log-format")
|
|
||||||
.value_name("format")
|
|
||||||
.help("Customize http log format"),
|
|
||||||
)
|
|
||||||
.arg(
|
|
||||||
Arg::new("completions")
|
|
||||||
.long("completions")
|
|
||||||
.value_name("shell")
|
|
||||||
.value_parser(value_parser!(Shell))
|
|
||||||
.help("Print shell completion script for <shell>"),
|
|
||||||
)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn print_completions<G: Generator>(gen: G, cmd: &mut Command) {
|
pub fn print_completions<G: Generator>(gen: G, cmd: &mut Command) {
|
||||||
generate(gen, cmd, cmd.get_name().to_string(), &mut std::io::stdout());
|
generate(gen, cmd, cmd.get_name().to_string(), &mut std::io::stdout());
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug)]
|
#[derive(Debug, Deserialize, SmartDefault, PartialEq)]
|
||||||
|
#[serde(default)]
|
||||||
|
#[serde(rename_all = "kebab-case")]
|
||||||
pub struct Args {
|
pub struct Args {
|
||||||
|
#[serde(default = "default_serve_path")]
|
||||||
|
#[default(default_serve_path())]
|
||||||
|
pub serve_path: PathBuf,
|
||||||
|
#[serde(deserialize_with = "deserialize_bind_addrs")]
|
||||||
|
#[serde(rename = "bind")]
|
||||||
|
#[serde(default = "default_addrs")]
|
||||||
|
#[default(default_addrs())]
|
||||||
pub addrs: Vec<BindAddr>,
|
pub addrs: Vec<BindAddr>,
|
||||||
|
#[serde(default = "default_port")]
|
||||||
|
#[default(default_port())]
|
||||||
pub port: u16,
|
pub port: u16,
|
||||||
pub path: PathBuf,
|
#[serde(skip)]
|
||||||
pub path_is_file: bool,
|
pub path_is_file: bool,
|
||||||
pub path_prefix: String,
|
pub path_prefix: String,
|
||||||
|
#[serde(skip)]
|
||||||
pub uri_prefix: String,
|
pub uri_prefix: String,
|
||||||
|
#[serde(deserialize_with = "deserialize_string_or_vec")]
|
||||||
pub hidden: Vec<String>,
|
pub hidden: Vec<String>,
|
||||||
pub auth_method: AuthMethod,
|
#[serde(deserialize_with = "deserialize_access_control")]
|
||||||
pub auth: AccessControl,
|
pub auth: AccessControl,
|
||||||
|
pub allow_all: bool,
|
||||||
pub allow_upload: bool,
|
pub allow_upload: bool,
|
||||||
pub allow_delete: bool,
|
pub allow_delete: bool,
|
||||||
pub allow_search: bool,
|
pub allow_search: bool,
|
||||||
@@ -244,12 +276,13 @@ pub struct Args {
|
|||||||
pub render_spa: bool,
|
pub render_spa: bool,
|
||||||
pub render_try_index: bool,
|
pub render_try_index: bool,
|
||||||
pub enable_cors: bool,
|
pub enable_cors: bool,
|
||||||
pub assets_path: Option<PathBuf>,
|
pub assets: Option<PathBuf>,
|
||||||
pub log_http: LogHttp,
|
#[serde(deserialize_with = "deserialize_log_http")]
|
||||||
#[cfg(feature = "tls")]
|
#[serde(rename = "log-format")]
|
||||||
pub tls: Option<(Vec<Certificate>, PrivateKey)>,
|
pub http_logger: HttpLogger,
|
||||||
#[cfg(not(feature = "tls"))]
|
pub compress: Compress,
|
||||||
pub tls: Option<()>,
|
pub tls_cert: Option<PathBuf>,
|
||||||
|
pub tls_key: Option<PathBuf>,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Args {
|
impl Args {
|
||||||
@@ -258,95 +291,168 @@ impl Args {
|
|||||||
/// If a parsing error occurred, exit the process and print out informative
|
/// If a parsing error occurred, exit the process and print out informative
|
||||||
/// error message to user.
|
/// error message to user.
|
||||||
pub fn parse(matches: ArgMatches) -> Result<Args> {
|
pub fn parse(matches: ArgMatches) -> Result<Args> {
|
||||||
let port = *matches.get_one::<u16>("port").unwrap();
|
let mut args = Self::default();
|
||||||
let addrs = matches
|
|
||||||
.get_many::<String>("bind")
|
if let Some(config_path) = matches.get_one::<PathBuf>("config") {
|
||||||
.map(|bind| bind.map(|v| v.as_str()).collect())
|
let contents = std::fs::read_to_string(config_path)
|
||||||
.unwrap_or_else(|| vec!["0.0.0.0", "::"]);
|
.with_context(|| format!("Failed to read config at {}", config_path.display()))?;
|
||||||
let addrs: Vec<BindAddr> = Args::parse_addrs(&addrs)?;
|
args = serde_yaml::from_str(&contents)
|
||||||
let path = Args::parse_path(matches.get_one::<PathBuf>("root").unwrap())?;
|
.with_context(|| format!("Failed to load config at {}", config_path.display()))?;
|
||||||
let path_is_file = path.metadata()?.is_file();
|
}
|
||||||
let path_prefix = matches
|
|
||||||
.get_one::<String>("path-prefix")
|
if let Some(path) = matches.get_one::<PathBuf>("serve-path") {
|
||||||
.map(|v| v.trim_matches('/').to_owned())
|
args.serve_path = path.clone()
|
||||||
.unwrap_or_default();
|
}
|
||||||
let uri_prefix = if path_prefix.is_empty() {
|
|
||||||
|
args.serve_path = Self::sanitize_path(args.serve_path)?;
|
||||||
|
|
||||||
|
if let Some(port) = matches.get_one::<u16>("port") {
|
||||||
|
args.port = *port
|
||||||
|
}
|
||||||
|
|
||||||
|
if let Some(addrs) = matches.get_many::<String>("bind") {
|
||||||
|
let addrs: Vec<_> = addrs.map(|v| v.as_str()).collect();
|
||||||
|
args.addrs = BindAddr::parse_addrs(&addrs)?;
|
||||||
|
}
|
||||||
|
|
||||||
|
args.path_is_file = args.serve_path.metadata()?.is_file();
|
||||||
|
if let Some(path_prefix) = matches.get_one::<String>("path-prefix") {
|
||||||
|
args.path_prefix = path_prefix.clone();
|
||||||
|
}
|
||||||
|
args.path_prefix = args.path_prefix.trim_matches('/').to_string();
|
||||||
|
|
||||||
|
args.uri_prefix = if args.path_prefix.is_empty() {
|
||||||
"/".to_owned()
|
"/".to_owned()
|
||||||
} else {
|
} else {
|
||||||
format!("/{}/", &encode_uri(&path_prefix))
|
format!("/{}/", &encode_uri(&args.path_prefix))
|
||||||
};
|
|
||||||
let hidden: Vec<String> = matches
|
|
||||||
.get_one::<String>("hidden")
|
|
||||||
.map(|v| v.split(',').map(|x| x.to_string()).collect())
|
|
||||||
.unwrap_or_default();
|
|
||||||
let enable_cors = matches.get_flag("enable-cors");
|
|
||||||
let auth: Vec<&str> = matches
|
|
||||||
.get_many::<String>("auth")
|
|
||||||
.map(|auth| auth.map(|v| v.as_str()).collect())
|
|
||||||
.unwrap_or_default();
|
|
||||||
let auth_method = match matches.get_one::<String>("auth-method").unwrap().as_str() {
|
|
||||||
"basic" => AuthMethod::Basic,
|
|
||||||
_ => AuthMethod::Digest,
|
|
||||||
};
|
|
||||||
let auth = AccessControl::new(&auth, &uri_prefix)?;
|
|
||||||
let allow_upload = matches.get_flag("allow-all") || matches.get_flag("allow-upload");
|
|
||||||
let allow_delete = matches.get_flag("allow-all") || matches.get_flag("allow-delete");
|
|
||||||
let allow_search = matches.get_flag("allow-all") || matches.get_flag("allow-search");
|
|
||||||
let allow_symlink = matches.get_flag("allow-all") || matches.get_flag("allow-symlink");
|
|
||||||
let allow_archive = matches.get_flag("allow-all") || matches.get_flag("allow-archive");
|
|
||||||
let render_index = matches.get_flag("render-index");
|
|
||||||
let render_try_index = matches.get_flag("render-try-index");
|
|
||||||
let render_spa = matches.get_flag("render-spa");
|
|
||||||
#[cfg(feature = "tls")]
|
|
||||||
let tls = match (
|
|
||||||
matches.get_one::<PathBuf>("tls-cert"),
|
|
||||||
matches.get_one::<PathBuf>("tls-key"),
|
|
||||||
) {
|
|
||||||
(Some(certs_file), Some(key_file)) => {
|
|
||||||
let certs = load_certs(certs_file)?;
|
|
||||||
let key = load_private_key(key_file)?;
|
|
||||||
Some((certs, key))
|
|
||||||
}
|
|
||||||
_ => None,
|
|
||||||
};
|
|
||||||
#[cfg(not(feature = "tls"))]
|
|
||||||
let tls = None;
|
|
||||||
let log_http: LogHttp = matches
|
|
||||||
.get_one::<String>("log-format")
|
|
||||||
.map(|v| v.as_str())
|
|
||||||
.unwrap_or(DEFAULT_LOG_FORMAT)
|
|
||||||
.parse()?;
|
|
||||||
let assets_path = match matches.get_one::<PathBuf>("assets") {
|
|
||||||
Some(v) => Some(Args::parse_assets_path(v)?),
|
|
||||||
None => None,
|
|
||||||
};
|
};
|
||||||
|
|
||||||
Ok(Args {
|
if let Some(hidden) = matches.get_many::<String>("hidden") {
|
||||||
addrs,
|
args.hidden = hidden.cloned().collect();
|
||||||
port,
|
} else {
|
||||||
path,
|
let mut hidden = vec![];
|
||||||
path_is_file,
|
std::mem::swap(&mut args.hidden, &mut hidden);
|
||||||
path_prefix,
|
args.hidden = hidden
|
||||||
uri_prefix,
|
.into_iter()
|
||||||
hidden,
|
.flat_map(|v| v.split(',').map(|v| v.to_string()).collect::<Vec<String>>())
|
||||||
auth_method,
|
.collect();
|
||||||
auth,
|
}
|
||||||
enable_cors,
|
|
||||||
allow_delete,
|
if !args.enable_cors {
|
||||||
allow_upload,
|
args.enable_cors = matches.get_flag("enable-cors");
|
||||||
allow_search,
|
}
|
||||||
allow_symlink,
|
|
||||||
allow_archive,
|
if let Some(rules) = matches.get_many::<String>("auth") {
|
||||||
render_index,
|
let rules: Vec<_> = rules.map(|v| v.as_str()).collect();
|
||||||
render_try_index,
|
args.auth = AccessControl::new(&rules)?;
|
||||||
render_spa,
|
}
|
||||||
tls,
|
|
||||||
log_http,
|
if !args.allow_all {
|
||||||
assets_path,
|
args.allow_all = matches.get_flag("allow-all");
|
||||||
})
|
}
|
||||||
|
|
||||||
|
let allow_all = args.allow_all;
|
||||||
|
|
||||||
|
if !args.allow_upload {
|
||||||
|
args.allow_upload = allow_all || matches.get_flag("allow-upload");
|
||||||
|
}
|
||||||
|
if !args.allow_delete {
|
||||||
|
args.allow_delete = allow_all || matches.get_flag("allow-delete");
|
||||||
|
}
|
||||||
|
if !args.allow_search {
|
||||||
|
args.allow_search = allow_all || matches.get_flag("allow-search");
|
||||||
|
}
|
||||||
|
if !args.allow_symlink {
|
||||||
|
args.allow_symlink = allow_all || matches.get_flag("allow-symlink");
|
||||||
|
}
|
||||||
|
if !args.allow_archive {
|
||||||
|
args.allow_archive = allow_all || matches.get_flag("allow-archive");
|
||||||
|
}
|
||||||
|
if !args.render_index {
|
||||||
|
args.render_index = matches.get_flag("render-index");
|
||||||
|
}
|
||||||
|
|
||||||
|
if !args.render_try_index {
|
||||||
|
args.render_try_index = matches.get_flag("render-try-index");
|
||||||
|
}
|
||||||
|
|
||||||
|
if !args.render_spa {
|
||||||
|
args.render_spa = matches.get_flag("render-spa");
|
||||||
|
}
|
||||||
|
|
||||||
|
if let Some(assets_path) = matches.get_one::<PathBuf>("assets") {
|
||||||
|
args.assets = Some(assets_path.clone());
|
||||||
|
}
|
||||||
|
|
||||||
|
if let Some(assets_path) = &args.assets {
|
||||||
|
args.assets = Some(Args::sanitize_assets_path(assets_path)?);
|
||||||
|
}
|
||||||
|
|
||||||
|
if let Some(log_format) = matches.get_one::<String>("log-format") {
|
||||||
|
args.http_logger = log_format.parse()?;
|
||||||
|
}
|
||||||
|
|
||||||
|
if let Some(compress) = matches.get_one::<Compress>("compress") {
|
||||||
|
args.compress = *compress;
|
||||||
|
}
|
||||||
|
|
||||||
|
#[cfg(feature = "tls")]
|
||||||
|
{
|
||||||
|
if let Some(tls_cert) = matches.get_one::<PathBuf>("tls-cert") {
|
||||||
|
args.tls_cert = Some(tls_cert.clone())
|
||||||
|
}
|
||||||
|
|
||||||
|
if let Some(tls_key) = matches.get_one::<PathBuf>("tls-key") {
|
||||||
|
args.tls_key = Some(tls_key.clone())
|
||||||
|
}
|
||||||
|
|
||||||
|
match (&args.tls_cert, &args.tls_key) {
|
||||||
|
(Some(_), Some(_)) => {}
|
||||||
|
(Some(_), _) => bail!("No tls-key set"),
|
||||||
|
(_, Some(_)) => bail!("No tls-cert set"),
|
||||||
|
(None, None) => {}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
#[cfg(not(feature = "tls"))]
|
||||||
|
{
|
||||||
|
args.tls_cert = None;
|
||||||
|
args.tls_key = None;
|
||||||
|
}
|
||||||
|
|
||||||
|
Ok(args)
|
||||||
}
|
}
|
||||||
|
|
||||||
fn parse_addrs(addrs: &[&str]) -> Result<Vec<BindAddr>> {
|
fn sanitize_path<P: AsRef<Path>>(path: P) -> Result<PathBuf> {
|
||||||
|
let path = path.as_ref();
|
||||||
|
if !path.exists() {
|
||||||
|
bail!("Path `{}` doesn't exist", path.display());
|
||||||
|
}
|
||||||
|
|
||||||
|
env::current_dir()
|
||||||
|
.and_then(|mut p| {
|
||||||
|
p.push(path); // If path is absolute, it replaces the current path.
|
||||||
|
std::fs::canonicalize(p)
|
||||||
|
})
|
||||||
|
.with_context(|| format!("Failed to access path `{}`", path.display()))
|
||||||
|
}
|
||||||
|
|
||||||
|
fn sanitize_assets_path<P: AsRef<Path>>(path: P) -> Result<PathBuf> {
|
||||||
|
let path = Self::sanitize_path(path)?;
|
||||||
|
if !path.join("index.html").exists() {
|
||||||
|
bail!("Path `{}` doesn't contains index.html", path.display());
|
||||||
|
}
|
||||||
|
Ok(path)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, Clone, PartialEq, Eq, PartialOrd, Ord)]
|
||||||
|
pub enum BindAddr {
|
||||||
|
Address(IpAddr),
|
||||||
|
Path(PathBuf),
|
||||||
|
}
|
||||||
|
|
||||||
|
impl BindAddr {
|
||||||
|
fn parse_addrs(addrs: &[&str]) -> Result<Vec<Self>> {
|
||||||
let mut bind_addrs = vec![];
|
let mut bind_addrs = vec![];
|
||||||
let mut invalid_addrs = vec![];
|
let mut invalid_addrs = vec![];
|
||||||
for addr in addrs {
|
for addr in addrs {
|
||||||
@@ -368,32 +474,262 @@ impl Args {
|
|||||||
}
|
}
|
||||||
Ok(bind_addrs)
|
Ok(bind_addrs)
|
||||||
}
|
}
|
||||||
|
}
|
||||||
|
|
||||||
fn parse_path<P: AsRef<Path>>(path: P) -> Result<PathBuf> {
|
#[derive(Debug, Clone, Copy, PartialEq, Deserialize)]
|
||||||
let path = path.as_ref();
|
#[serde(rename_all = "lowercase")]
|
||||||
if !path.exists() {
|
pub enum Compress {
|
||||||
bail!("Path `{}` doesn't exist", path.display());
|
None,
|
||||||
}
|
Low,
|
||||||
|
Medium,
|
||||||
|
High,
|
||||||
|
}
|
||||||
|
|
||||||
env::current_dir()
|
impl Default for Compress {
|
||||||
.and_then(|mut p| {
|
fn default() -> Self {
|
||||||
p.push(path); // If path is absolute, it replaces the current path.
|
Self::Low
|
||||||
std::fs::canonicalize(p)
|
|
||||||
})
|
|
||||||
.with_context(|| format!("Failed to access path `{}`", path.display()))
|
|
||||||
}
|
|
||||||
|
|
||||||
fn parse_assets_path<P: AsRef<Path>>(path: P) -> Result<PathBuf> {
|
|
||||||
let path = Self::parse_path(path)?;
|
|
||||||
if !path.join("index.html").exists() {
|
|
||||||
bail!("Path `{}` doesn't contains index.html", path.display());
|
|
||||||
}
|
|
||||||
Ok(path)
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug, Clone, PartialEq, Eq, PartialOrd, Ord)]
|
impl ValueEnum for Compress {
|
||||||
pub enum BindAddr {
|
fn value_variants<'a>() -> &'a [Self] {
|
||||||
Address(IpAddr),
|
&[Self::None, Self::Low, Self::Medium, Self::High]
|
||||||
Path(PathBuf),
|
}
|
||||||
|
|
||||||
|
fn to_possible_value(&self) -> Option<clap::builder::PossibleValue> {
|
||||||
|
Some(match self {
|
||||||
|
Compress::None => PossibleValue::new("none"),
|
||||||
|
Compress::Low => PossibleValue::new("low"),
|
||||||
|
Compress::Medium => PossibleValue::new("medium"),
|
||||||
|
Compress::High => PossibleValue::new("high"),
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl Compress {
|
||||||
|
pub fn to_compression(self) -> Compression {
|
||||||
|
match self {
|
||||||
|
Compress::None => Compression::Stored,
|
||||||
|
Compress::Low => Compression::Deflate,
|
||||||
|
Compress::Medium => Compression::Bz,
|
||||||
|
Compress::High => Compression::Xz,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn deserialize_bind_addrs<'de, D>(deserializer: D) -> Result<Vec<BindAddr>, D::Error>
|
||||||
|
where
|
||||||
|
D: Deserializer<'de>,
|
||||||
|
{
|
||||||
|
struct StringOrVec;
|
||||||
|
|
||||||
|
impl<'de> serde::de::Visitor<'de> for StringOrVec {
|
||||||
|
type Value = Vec<BindAddr>;
|
||||||
|
|
||||||
|
fn expecting(&self, formatter: &mut std::fmt::Formatter) -> std::fmt::Result {
|
||||||
|
formatter.write_str("string or list of strings")
|
||||||
|
}
|
||||||
|
|
||||||
|
fn visit_str<E>(self, s: &str) -> Result<Self::Value, E>
|
||||||
|
where
|
||||||
|
E: serde::de::Error,
|
||||||
|
{
|
||||||
|
BindAddr::parse_addrs(&[s]).map_err(serde::de::Error::custom)
|
||||||
|
}
|
||||||
|
|
||||||
|
fn visit_seq<S>(self, seq: S) -> Result<Self::Value, S::Error>
|
||||||
|
where
|
||||||
|
S: serde::de::SeqAccess<'de>,
|
||||||
|
{
|
||||||
|
let addrs: Vec<&'de str> =
|
||||||
|
Deserialize::deserialize(serde::de::value::SeqAccessDeserializer::new(seq))?;
|
||||||
|
BindAddr::parse_addrs(&addrs).map_err(serde::de::Error::custom)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
deserializer.deserialize_any(StringOrVec)
|
||||||
|
}
|
||||||
|
|
||||||
|
fn deserialize_string_or_vec<'de, D>(deserializer: D) -> Result<Vec<String>, D::Error>
|
||||||
|
where
|
||||||
|
D: Deserializer<'de>,
|
||||||
|
{
|
||||||
|
struct StringOrVec;
|
||||||
|
|
||||||
|
impl<'de> serde::de::Visitor<'de> for StringOrVec {
|
||||||
|
type Value = Vec<String>;
|
||||||
|
|
||||||
|
fn expecting(&self, formatter: &mut std::fmt::Formatter) -> std::fmt::Result {
|
||||||
|
formatter.write_str("string or list of strings")
|
||||||
|
}
|
||||||
|
|
||||||
|
fn visit_str<E>(self, s: &str) -> Result<Self::Value, E>
|
||||||
|
where
|
||||||
|
E: serde::de::Error,
|
||||||
|
{
|
||||||
|
Ok(vec![s.to_owned()])
|
||||||
|
}
|
||||||
|
|
||||||
|
fn visit_seq<S>(self, seq: S) -> Result<Self::Value, S::Error>
|
||||||
|
where
|
||||||
|
S: serde::de::SeqAccess<'de>,
|
||||||
|
{
|
||||||
|
Deserialize::deserialize(serde::de::value::SeqAccessDeserializer::new(seq))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
deserializer.deserialize_any(StringOrVec)
|
||||||
|
}
|
||||||
|
|
||||||
|
fn deserialize_access_control<'de, D>(deserializer: D) -> Result<AccessControl, D::Error>
|
||||||
|
where
|
||||||
|
D: Deserializer<'de>,
|
||||||
|
{
|
||||||
|
let rules: Vec<&str> = Vec::deserialize(deserializer)?;
|
||||||
|
AccessControl::new(&rules).map_err(serde::de::Error::custom)
|
||||||
|
}
|
||||||
|
|
||||||
|
fn deserialize_log_http<'de, D>(deserializer: D) -> Result<HttpLogger, D::Error>
|
||||||
|
where
|
||||||
|
D: Deserializer<'de>,
|
||||||
|
{
|
||||||
|
let value: String = Deserialize::deserialize(deserializer)?;
|
||||||
|
value.parse().map_err(serde::de::Error::custom)
|
||||||
|
}
|
||||||
|
|
||||||
|
fn default_serve_path() -> PathBuf {
|
||||||
|
PathBuf::from(".")
|
||||||
|
}
|
||||||
|
|
||||||
|
fn default_addrs() -> Vec<BindAddr> {
|
||||||
|
BindAddr::parse_addrs(&["0.0.0.0", "::"]).unwrap()
|
||||||
|
}
|
||||||
|
|
||||||
|
fn default_port() -> u16 {
|
||||||
|
5000
|
||||||
|
}
|
||||||
|
|
||||||
|
#[cfg(test)]
|
||||||
|
mod tests {
|
||||||
|
use super::*;
|
||||||
|
|
||||||
|
use assert_fs::prelude::*;
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_default() {
|
||||||
|
let cli = build_cli();
|
||||||
|
let matches = cli.try_get_matches_from(vec![""]).unwrap();
|
||||||
|
let args = Args::parse(matches).unwrap();
|
||||||
|
let cwd = Args::sanitize_path(std::env::current_dir().unwrap()).unwrap();
|
||||||
|
assert_eq!(args.serve_path, cwd);
|
||||||
|
assert_eq!(args.port, default_port());
|
||||||
|
assert_eq!(args.addrs, default_addrs());
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_args_from_cli1() {
|
||||||
|
let tmpdir = assert_fs::TempDir::new().unwrap();
|
||||||
|
let cli = build_cli();
|
||||||
|
let matches = cli
|
||||||
|
.try_get_matches_from(vec![
|
||||||
|
"",
|
||||||
|
"--hidden",
|
||||||
|
"tmp,*.log,*.lock",
|
||||||
|
&tmpdir.to_string_lossy(),
|
||||||
|
])
|
||||||
|
.unwrap();
|
||||||
|
let args = Args::parse(matches).unwrap();
|
||||||
|
assert_eq!(args.serve_path, Args::sanitize_path(&tmpdir).unwrap());
|
||||||
|
assert_eq!(args.hidden, ["tmp", "*.log", "*.lock"]);
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_args_from_cli2() {
|
||||||
|
let cli = build_cli();
|
||||||
|
let matches = cli
|
||||||
|
.try_get_matches_from(vec![
|
||||||
|
"", "--hidden", "tmp", "--hidden", "*.log", "--hidden", "*.lock",
|
||||||
|
])
|
||||||
|
.unwrap();
|
||||||
|
let args = Args::parse(matches).unwrap();
|
||||||
|
assert_eq!(args.hidden, ["tmp", "*.log", "*.lock"]);
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_args_from_empty_config_file() {
|
||||||
|
let tmpdir = assert_fs::TempDir::new().unwrap();
|
||||||
|
let config_file = tmpdir.child("config.yaml");
|
||||||
|
config_file.write_str("").unwrap();
|
||||||
|
|
||||||
|
let cli = build_cli();
|
||||||
|
let matches = cli
|
||||||
|
.try_get_matches_from(vec!["", "-c", &config_file.to_string_lossy()])
|
||||||
|
.unwrap();
|
||||||
|
let args = Args::parse(matches).unwrap();
|
||||||
|
let cwd = Args::sanitize_path(std::env::current_dir().unwrap()).unwrap();
|
||||||
|
assert_eq!(args.serve_path, cwd);
|
||||||
|
assert_eq!(args.port, default_port());
|
||||||
|
assert_eq!(args.addrs, default_addrs());
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_args_from_config_file1() {
|
||||||
|
let tmpdir = assert_fs::TempDir::new().unwrap();
|
||||||
|
let config_file = tmpdir.child("config.yaml");
|
||||||
|
let contents = format!(
|
||||||
|
r#"
|
||||||
|
serve-path: {}
|
||||||
|
bind: 0.0.0.0
|
||||||
|
port: 3000
|
||||||
|
allow-upload: true
|
||||||
|
hidden: tmp,*.log,*.lock
|
||||||
|
"#,
|
||||||
|
tmpdir.display()
|
||||||
|
);
|
||||||
|
config_file.write_str(&contents).unwrap();
|
||||||
|
|
||||||
|
let cli = build_cli();
|
||||||
|
let matches = cli
|
||||||
|
.try_get_matches_from(vec!["", "-c", &config_file.to_string_lossy()])
|
||||||
|
.unwrap();
|
||||||
|
let args = Args::parse(matches).unwrap();
|
||||||
|
assert_eq!(args.serve_path, Args::sanitize_path(&tmpdir).unwrap());
|
||||||
|
assert_eq!(
|
||||||
|
args.addrs,
|
||||||
|
vec![BindAddr::Address("0.0.0.0".parse().unwrap())]
|
||||||
|
);
|
||||||
|
assert_eq!(args.hidden, ["tmp", "*.log", "*.lock"]);
|
||||||
|
assert_eq!(args.port, 3000);
|
||||||
|
assert!(args.allow_upload);
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_args_from_config_file2() {
|
||||||
|
let tmpdir = assert_fs::TempDir::new().unwrap();
|
||||||
|
let config_file = tmpdir.child("config.yaml");
|
||||||
|
let contents = r#"
|
||||||
|
bind:
|
||||||
|
- 127.0.0.1
|
||||||
|
- 192.168.8.10
|
||||||
|
hidden:
|
||||||
|
- tmp
|
||||||
|
- '*.log'
|
||||||
|
- '*.lock'
|
||||||
|
"#;
|
||||||
|
config_file.write_str(contents).unwrap();
|
||||||
|
|
||||||
|
let cli = build_cli();
|
||||||
|
let matches = cli
|
||||||
|
.try_get_matches_from(vec!["", "-c", &config_file.to_string_lossy()])
|
||||||
|
.unwrap();
|
||||||
|
let args = Args::parse(matches).unwrap();
|
||||||
|
assert_eq!(
|
||||||
|
args.addrs,
|
||||||
|
vec![
|
||||||
|
BindAddr::Address("127.0.0.1".parse().unwrap()),
|
||||||
|
BindAddr::Address("192.168.8.10".parse().unwrap())
|
||||||
|
]
|
||||||
|
);
|
||||||
|
assert_eq!(args.hidden, ["tmp", "*.log", "*.lock"]);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
725
src/auth.rs
725
src/auth.rs
@@ -1,16 +1,20 @@
|
|||||||
|
use crate::{args::Args, server::Response, utils::unix_now};
|
||||||
|
|
||||||
use anyhow::{anyhow, bail, Result};
|
use anyhow::{anyhow, bail, Result};
|
||||||
use base64::{engine::general_purpose, Engine as _};
|
use base64::{engine::general_purpose, Engine as _};
|
||||||
use headers::HeaderValue;
|
use headers::HeaderValue;
|
||||||
use hyper::Method;
|
use hyper::{header::WWW_AUTHENTICATE, Method};
|
||||||
|
use indexmap::IndexMap;
|
||||||
use lazy_static::lazy_static;
|
use lazy_static::lazy_static;
|
||||||
use md5::Context;
|
use md5::Context;
|
||||||
use std::collections::HashMap;
|
use std::{
|
||||||
|
collections::HashMap,
|
||||||
|
path::{Path, PathBuf},
|
||||||
|
};
|
||||||
use uuid::Uuid;
|
use uuid::Uuid;
|
||||||
|
|
||||||
use crate::utils::{encode_uri, unix_now};
|
|
||||||
|
|
||||||
const REALM: &str = "DUFS";
|
const REALM: &str = "DUFS";
|
||||||
const DIGEST_AUTH_TIMEOUT: u32 = 86400;
|
const DIGEST_AUTH_TIMEOUT: u32 = 604800; // 7 days
|
||||||
|
|
||||||
lazy_static! {
|
lazy_static! {
|
||||||
static ref NONCESTARTHASH: Context = {
|
static ref NONCESTARTHASH: Context = {
|
||||||
@@ -21,57 +25,74 @@ lazy_static! {
|
|||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug)]
|
#[derive(Debug, Clone, PartialEq)]
|
||||||
pub struct AccessControl {
|
pub struct AccessControl {
|
||||||
rules: HashMap<String, PathControl>,
|
use_hashed_password: bool,
|
||||||
|
users: IndexMap<String, (String, AccessPaths)>,
|
||||||
|
anony: Option<AccessPaths>,
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug)]
|
impl Default for AccessControl {
|
||||||
pub struct PathControl {
|
fn default() -> Self {
|
||||||
readwrite: Account,
|
AccessControl {
|
||||||
readonly: Option<Account>,
|
use_hashed_password: false,
|
||||||
share: bool,
|
anony: Some(AccessPaths::new(AccessPerm::ReadWrite)),
|
||||||
|
users: IndexMap::new(),
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl AccessControl {
|
impl AccessControl {
|
||||||
pub fn new(raw_rules: &[&str], uri_prefix: &str) -> Result<Self> {
|
pub fn new(raw_rules: &[&str]) -> Result<Self> {
|
||||||
let mut rules = HashMap::default();
|
|
||||||
if raw_rules.is_empty() {
|
if raw_rules.is_empty() {
|
||||||
return Ok(Self { rules });
|
return Ok(Default::default());
|
||||||
}
|
}
|
||||||
for rule in raw_rules {
|
let new_raw_rules = split_rules(raw_rules);
|
||||||
let parts: Vec<&str> = rule.split('@').collect();
|
let mut use_hashed_password = false;
|
||||||
let create_err = || anyhow!("Invalid auth `{rule}`");
|
let mut annoy_paths = None;
|
||||||
match parts.as_slice() {
|
let mut account_paths_pairs = vec![];
|
||||||
[path, readwrite] => {
|
for rule in &new_raw_rules {
|
||||||
let control = PathControl {
|
let (account, paths) =
|
||||||
readwrite: Account::new(readwrite).ok_or_else(create_err)?,
|
split_account_paths(rule).ok_or_else(|| anyhow!("Invalid auth `{rule}`"))?;
|
||||||
readonly: None,
|
if account.is_empty() {
|
||||||
share: false,
|
if annoy_paths.is_some() {
|
||||||
};
|
bail!("Invalid auth, no duplicate anonymous rules");
|
||||||
rules.insert(sanitize_path(path, uri_prefix), control);
|
|
||||||
}
|
}
|
||||||
[path, readwrite, readonly] => {
|
annoy_paths = Some(paths)
|
||||||
let (readonly, share) = if *readonly == "*" {
|
} else if let Some((user, pass)) = account.split_once(':') {
|
||||||
(None, true)
|
if user.is_empty() || pass.is_empty() {
|
||||||
} else {
|
bail!("Invalid auth `{rule}`");
|
||||||
(Some(Account::new(readonly).ok_or_else(create_err)?), false)
|
|
||||||
};
|
|
||||||
let control = PathControl {
|
|
||||||
readwrite: Account::new(readwrite).ok_or_else(create_err)?,
|
|
||||||
readonly,
|
|
||||||
share,
|
|
||||||
};
|
|
||||||
rules.insert(sanitize_path(path, uri_prefix), control);
|
|
||||||
}
|
}
|
||||||
_ => return Err(create_err()),
|
account_paths_pairs.push((user, pass, paths));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
Ok(Self { rules })
|
let mut anony = None;
|
||||||
|
if let Some(paths) = annoy_paths {
|
||||||
|
let mut access_paths = AccessPaths::default();
|
||||||
|
access_paths.merge(paths);
|
||||||
|
anony = Some(access_paths);
|
||||||
|
}
|
||||||
|
let mut users = IndexMap::new();
|
||||||
|
for (user, pass, paths) in account_paths_pairs.into_iter() {
|
||||||
|
let mut access_paths = anony.clone().unwrap_or_default();
|
||||||
|
access_paths
|
||||||
|
.merge(paths)
|
||||||
|
.ok_or_else(|| anyhow!("Invalid auth `{user}:{pass}@{paths}"))?;
|
||||||
|
if pass.starts_with("$6$") {
|
||||||
|
use_hashed_password = true;
|
||||||
|
}
|
||||||
|
users.insert(user.to_string(), (pass.to_string(), access_paths));
|
||||||
|
}
|
||||||
|
|
||||||
|
Ok(Self {
|
||||||
|
use_hashed_password,
|
||||||
|
users,
|
||||||
|
anony,
|
||||||
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn valid(&self) -> bool {
|
pub fn exist(&self) -> bool {
|
||||||
!self.rules.is_empty()
|
!self.users.is_empty()
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn guard(
|
pub fn guard(
|
||||||
@@ -79,254 +100,310 @@ impl AccessControl {
|
|||||||
path: &str,
|
path: &str,
|
||||||
method: &Method,
|
method: &Method,
|
||||||
authorization: Option<&HeaderValue>,
|
authorization: Option<&HeaderValue>,
|
||||||
auth_method: AuthMethod,
|
) -> (Option<String>, Option<AccessPaths>) {
|
||||||
) -> GuardType {
|
if let Some(authorization) = authorization {
|
||||||
if self.rules.is_empty() {
|
if let Some(user) = get_auth_user(authorization) {
|
||||||
return GuardType::ReadWrite;
|
if let Some((pass, paths)) = self.users.get(&user) {
|
||||||
|
if method == Method::OPTIONS {
|
||||||
|
return (Some(user), Some(AccessPaths::new(AccessPerm::ReadOnly)));
|
||||||
|
}
|
||||||
|
if check_auth(authorization, method.as_str(), &user, pass).is_some() {
|
||||||
|
return (Some(user), paths.find(path, !is_readonly_method(method)));
|
||||||
|
} else {
|
||||||
|
return (None, None);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if method == Method::OPTIONS {
|
if method == Method::OPTIONS {
|
||||||
return GuardType::ReadOnly;
|
return (None, Some(AccessPaths::new(AccessPerm::ReadOnly)));
|
||||||
}
|
}
|
||||||
|
|
||||||
let mut controls = vec![];
|
if let Some(paths) = self.anony.as_ref() {
|
||||||
for path in walk_path(path) {
|
return (None, paths.find(path, !is_readonly_method(method)));
|
||||||
if let Some(control) = self.rules.get(path) {
|
|
||||||
controls.push(control);
|
|
||||||
if let Some(authorization) = authorization {
|
|
||||||
let Account { user, pass } = &control.readwrite;
|
|
||||||
if auth_method
|
|
||||||
.validate(authorization, method.as_str(), user, pass)
|
|
||||||
.is_some()
|
|
||||||
{
|
|
||||||
return GuardType::ReadWrite;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
if is_readonly_method(method) {
|
|
||||||
for control in controls.into_iter() {
|
(None, None)
|
||||||
if control.share {
|
|
||||||
return GuardType::ReadOnly;
|
|
||||||
}
|
|
||||||
if let Some(authorization) = authorization {
|
|
||||||
if let Some(Account { user, pass }) = &control.readonly {
|
|
||||||
if auth_method
|
|
||||||
.validate(authorization, method.as_str(), user, pass)
|
|
||||||
.is_some()
|
|
||||||
{
|
|
||||||
return GuardType::ReadOnly;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
GuardType::Reject
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug, Clone, Copy, PartialEq, Eq, PartialOrd, Ord)]
|
#[derive(Debug, Default, Clone, PartialEq, Eq)]
|
||||||
pub enum GuardType {
|
pub struct AccessPaths {
|
||||||
Reject,
|
perm: AccessPerm,
|
||||||
ReadWrite,
|
children: IndexMap<String, AccessPaths>,
|
||||||
ReadOnly,
|
|
||||||
}
|
}
|
||||||
|
|
||||||
impl GuardType {
|
impl AccessPaths {
|
||||||
pub fn is_reject(&self) -> bool {
|
pub fn new(perm: AccessPerm) -> Self {
|
||||||
*self == GuardType::Reject
|
Self {
|
||||||
|
perm,
|
||||||
|
..Default::default()
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
|
||||||
|
|
||||||
fn sanitize_path(path: &str, uri_prefix: &str) -> String {
|
pub fn perm(&self) -> AccessPerm {
|
||||||
let new_path = match (uri_prefix, path) {
|
self.perm
|
||||||
("/", "/") => "/".into(),
|
}
|
||||||
(_, "/") => uri_prefix.trim_end_matches('/').into(),
|
|
||||||
_ => format!("{}{}", uri_prefix, path.trim_matches('/')),
|
|
||||||
};
|
|
||||||
encode_uri(&new_path)
|
|
||||||
}
|
|
||||||
|
|
||||||
fn walk_path(path: &str) -> impl Iterator<Item = &str> {
|
pub fn set_perm(&mut self, perm: AccessPerm) {
|
||||||
let mut idx = 0;
|
if !perm.inherit() {
|
||||||
path.split('/').enumerate().map(move |(i, part)| {
|
self.perm = perm;
|
||||||
let end = if i == 0 { 1 } else { idx + part.len() + i };
|
}
|
||||||
let value = &path[..end];
|
}
|
||||||
idx += part.len();
|
|
||||||
value
|
|
||||||
})
|
|
||||||
}
|
|
||||||
|
|
||||||
fn is_readonly_method(method: &Method) -> bool {
|
pub fn merge(&mut self, paths: &str) -> Option<()> {
|
||||||
method == Method::GET
|
for item in paths.trim_matches(',').split(',') {
|
||||||
|| method == Method::OPTIONS
|
let (path, perm) = match item.split_once(':') {
|
||||||
|| method == Method::HEAD
|
None => (item, AccessPerm::ReadOnly),
|
||||||
|| method.as_str() == "PROPFIND"
|
Some((path, "ro")) => (path, AccessPerm::ReadOnly),
|
||||||
}
|
Some((path, "rw")) => (path, AccessPerm::ReadWrite),
|
||||||
|
Some((path, "-")) => (path, AccessPerm::Forbidden),
|
||||||
|
_ => return None,
|
||||||
|
};
|
||||||
|
self.add(path, perm);
|
||||||
|
}
|
||||||
|
Some(())
|
||||||
|
}
|
||||||
|
|
||||||
#[derive(Debug, Clone)]
|
fn add(&mut self, path: &str, perm: AccessPerm) {
|
||||||
struct Account {
|
let path = path.trim_matches('/');
|
||||||
user: String,
|
if path.is_empty() {
|
||||||
pass: String,
|
self.set_perm(perm);
|
||||||
}
|
} else {
|
||||||
|
let parts: Vec<&str> = path.split('/').collect();
|
||||||
|
self.add_impl(&parts, perm);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
impl Account {
|
fn add_impl(&mut self, parts: &[&str], perm: AccessPerm) {
|
||||||
fn new(data: &str) -> Option<Self> {
|
let parts_len = parts.len();
|
||||||
let p: Vec<&str> = data.trim().split(':').collect();
|
if parts_len == 0 {
|
||||||
if p.len() != 2 {
|
self.set_perm(perm);
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
let child = self.children.entry(parts[0].to_string()).or_default();
|
||||||
|
child.add_impl(&parts[1..], perm)
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn find(&self, path: &str, writable: bool) -> Option<AccessPaths> {
|
||||||
|
let parts: Vec<&str> = path
|
||||||
|
.trim_matches('/')
|
||||||
|
.split('/')
|
||||||
|
.filter(|v| !v.is_empty())
|
||||||
|
.collect();
|
||||||
|
let target = self.find_impl(&parts, self.perm)?;
|
||||||
|
if target.perm().forbidden() {
|
||||||
return None;
|
return None;
|
||||||
}
|
}
|
||||||
let user = p[0];
|
if writable && !target.perm().readwrite() {
|
||||||
let pass = p[1];
|
return None;
|
||||||
let mut h = Context::new();
|
}
|
||||||
h.consume(format!("{user}:{REALM}:{pass}").as_bytes());
|
Some(target)
|
||||||
Some(Account {
|
|
||||||
user: user.to_owned(),
|
|
||||||
pass: format!("{:x}", h.compute()),
|
|
||||||
})
|
|
||||||
}
|
}
|
||||||
}
|
|
||||||
|
|
||||||
#[derive(Debug, Clone)]
|
fn find_impl(&self, parts: &[&str], perm: AccessPerm) -> Option<AccessPaths> {
|
||||||
pub enum AuthMethod {
|
let perm = if !self.perm.inherit() {
|
||||||
Basic,
|
self.perm
|
||||||
Digest,
|
} else {
|
||||||
}
|
perm
|
||||||
|
};
|
||||||
impl AuthMethod {
|
if parts.is_empty() {
|
||||||
pub fn www_auth(&self, stale: bool) -> Result<String> {
|
if perm.inherit() {
|
||||||
match self {
|
return Some(self.clone());
|
||||||
AuthMethod::Basic => Ok(format!("Basic realm=\"{REALM}\"")),
|
} else {
|
||||||
AuthMethod::Digest => {
|
return Some(AccessPaths::new(perm));
|
||||||
let str_stale = if stale { "stale=true," } else { "" };
|
|
||||||
Ok(format!(
|
|
||||||
"Digest realm=\"{}\",nonce=\"{}\",{}qop=\"auth\"",
|
|
||||||
REALM,
|
|
||||||
create_nonce()?,
|
|
||||||
str_stale
|
|
||||||
))
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
let child = match self.children.get(parts[0]) {
|
||||||
pub fn get_user(&self, authorization: &HeaderValue) -> Option<String> {
|
Some(v) => v,
|
||||||
match self {
|
None => {
|
||||||
AuthMethod::Basic => {
|
if perm.inherit() {
|
||||||
let value: Vec<u8> = general_purpose::STANDARD
|
|
||||||
.decode(strip_prefix(authorization.as_bytes(), b"Basic ")?)
|
|
||||||
.ok()?;
|
|
||||||
let parts: Vec<&str> = std::str::from_utf8(&value).ok()?.split(':').collect();
|
|
||||||
Some(parts[0].to_string())
|
|
||||||
}
|
|
||||||
AuthMethod::Digest => {
|
|
||||||
let digest_value = strip_prefix(authorization.as_bytes(), b"Digest ")?;
|
|
||||||
let digest_map = to_headermap(digest_value).ok()?;
|
|
||||||
digest_map
|
|
||||||
.get(b"username".as_ref())
|
|
||||||
.and_then(|b| std::str::from_utf8(b).ok())
|
|
||||||
.map(|v| v.to_string())
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
pub fn validate(
|
|
||||||
&self,
|
|
||||||
authorization: &HeaderValue,
|
|
||||||
method: &str,
|
|
||||||
auth_user: &str,
|
|
||||||
auth_pass: &str,
|
|
||||||
) -> Option<()> {
|
|
||||||
match self {
|
|
||||||
AuthMethod::Basic => {
|
|
||||||
let basic_value: Vec<u8> = general_purpose::STANDARD
|
|
||||||
.decode(strip_prefix(authorization.as_bytes(), b"Basic ")?)
|
|
||||||
.ok()?;
|
|
||||||
let parts: Vec<&str> = std::str::from_utf8(&basic_value).ok()?.split(':').collect();
|
|
||||||
|
|
||||||
if parts[0] != auth_user {
|
|
||||||
return None;
|
return None;
|
||||||
|
} else {
|
||||||
|
return Some(AccessPaths::new(perm));
|
||||||
}
|
}
|
||||||
|
|
||||||
let mut h = Context::new();
|
|
||||||
h.consume(format!("{}:{}:{}", parts[0], REALM, parts[1]).as_bytes());
|
|
||||||
|
|
||||||
let http_pass = format!("{:x}", h.compute());
|
|
||||||
|
|
||||||
if http_pass == auth_pass {
|
|
||||||
return Some(());
|
|
||||||
}
|
|
||||||
|
|
||||||
None
|
|
||||||
}
|
}
|
||||||
AuthMethod::Digest => {
|
};
|
||||||
let digest_value = strip_prefix(authorization.as_bytes(), b"Digest ")?;
|
child.find_impl(&parts[1..], perm)
|
||||||
let digest_map = to_headermap(digest_value).ok()?;
|
}
|
||||||
if let (Some(username), Some(nonce), Some(user_response)) = (
|
|
||||||
digest_map
|
pub fn child_names(&self) -> Vec<&String> {
|
||||||
.get(b"username".as_ref())
|
self.children.keys().collect()
|
||||||
.and_then(|b| std::str::from_utf8(b).ok()),
|
}
|
||||||
digest_map.get(b"nonce".as_ref()),
|
|
||||||
digest_map.get(b"response".as_ref()),
|
pub fn child_paths(&self, base: &Path) -> Vec<PathBuf> {
|
||||||
) {
|
if !self.perm().inherit() {
|
||||||
match validate_nonce(nonce) {
|
return vec![base.to_path_buf()];
|
||||||
Ok(true) => {}
|
}
|
||||||
_ => return None,
|
let mut output = vec![];
|
||||||
}
|
self.child_paths_impl(&mut output, base);
|
||||||
if auth_user != username {
|
output
|
||||||
return None;
|
}
|
||||||
}
|
|
||||||
let mut ha = Context::new();
|
fn child_paths_impl(&self, output: &mut Vec<PathBuf>, base: &Path) {
|
||||||
ha.consume(method);
|
for (name, child) in self.children.iter() {
|
||||||
ha.consume(b":");
|
let base = base.join(name);
|
||||||
if let Some(uri) = digest_map.get(b"uri".as_ref()) {
|
if child.perm().inherit() {
|
||||||
ha.consume(uri);
|
child.child_paths_impl(output, &base);
|
||||||
}
|
} else {
|
||||||
let ha = format!("{:x}", ha.compute());
|
output.push(base)
|
||||||
let mut correct_response = None;
|
|
||||||
if let Some(qop) = digest_map.get(b"qop".as_ref()) {
|
|
||||||
if qop == &b"auth".as_ref() || qop == &b"auth-int".as_ref() {
|
|
||||||
correct_response = Some({
|
|
||||||
let mut c = Context::new();
|
|
||||||
c.consume(auth_pass);
|
|
||||||
c.consume(b":");
|
|
||||||
c.consume(nonce);
|
|
||||||
c.consume(b":");
|
|
||||||
if let Some(nc) = digest_map.get(b"nc".as_ref()) {
|
|
||||||
c.consume(nc);
|
|
||||||
}
|
|
||||||
c.consume(b":");
|
|
||||||
if let Some(cnonce) = digest_map.get(b"cnonce".as_ref()) {
|
|
||||||
c.consume(cnonce);
|
|
||||||
}
|
|
||||||
c.consume(b":");
|
|
||||||
c.consume(qop);
|
|
||||||
c.consume(b":");
|
|
||||||
c.consume(&*ha);
|
|
||||||
format!("{:x}", c.compute())
|
|
||||||
});
|
|
||||||
}
|
|
||||||
}
|
|
||||||
let correct_response = match correct_response {
|
|
||||||
Some(r) => r,
|
|
||||||
None => {
|
|
||||||
let mut c = Context::new();
|
|
||||||
c.consume(auth_pass);
|
|
||||||
c.consume(b":");
|
|
||||||
c.consume(nonce);
|
|
||||||
c.consume(b":");
|
|
||||||
c.consume(&*ha);
|
|
||||||
format!("{:x}", c.compute())
|
|
||||||
}
|
|
||||||
};
|
|
||||||
if correct_response.as_bytes() == *user_response {
|
|
||||||
// grant access
|
|
||||||
return Some(());
|
|
||||||
}
|
|
||||||
}
|
|
||||||
None
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, Clone, Copy, PartialEq, Eq, PartialOrd, Ord, Default)]
|
||||||
|
pub enum AccessPerm {
|
||||||
|
#[default]
|
||||||
|
Inherit,
|
||||||
|
ReadOnly,
|
||||||
|
ReadWrite,
|
||||||
|
Forbidden,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl AccessPerm {
|
||||||
|
pub fn inherit(&self) -> bool {
|
||||||
|
self == &AccessPerm::Inherit
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn readwrite(&self) -> bool {
|
||||||
|
self == &AccessPerm::ReadWrite
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn forbidden(&self) -> bool {
|
||||||
|
self == &AccessPerm::Forbidden
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn www_authenticate(res: &mut Response, args: &Args) -> Result<()> {
|
||||||
|
if args.auth.use_hashed_password {
|
||||||
|
let basic = HeaderValue::from_str(&format!("Basic realm=\"{}\"", REALM))?;
|
||||||
|
res.headers_mut().insert(WWW_AUTHENTICATE, basic);
|
||||||
|
} else {
|
||||||
|
let nonce = create_nonce()?;
|
||||||
|
let digest = HeaderValue::from_str(&format!(
|
||||||
|
"Digest realm=\"{}\", nonce=\"{}\", qop=\"auth\"",
|
||||||
|
REALM, nonce
|
||||||
|
))?;
|
||||||
|
let basic = HeaderValue::from_str(&format!("Basic realm=\"{}\"", REALM))?;
|
||||||
|
res.headers_mut().append(WWW_AUTHENTICATE, digest);
|
||||||
|
res.headers_mut().append(WWW_AUTHENTICATE, basic);
|
||||||
|
}
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn get_auth_user(authorization: &HeaderValue) -> Option<String> {
|
||||||
|
if let Some(value) = strip_prefix(authorization.as_bytes(), b"Basic ") {
|
||||||
|
let value: Vec<u8> = general_purpose::STANDARD.decode(value).ok()?;
|
||||||
|
let parts: Vec<&str> = std::str::from_utf8(&value).ok()?.split(':').collect();
|
||||||
|
Some(parts[0].to_string())
|
||||||
|
} else if let Some(value) = strip_prefix(authorization.as_bytes(), b"Digest ") {
|
||||||
|
let digest_map = to_headermap(value).ok()?;
|
||||||
|
let username = digest_map.get(b"username".as_ref())?;
|
||||||
|
std::str::from_utf8(username).map(|v| v.to_string()).ok()
|
||||||
|
} else {
|
||||||
|
None
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn check_auth(
|
||||||
|
authorization: &HeaderValue,
|
||||||
|
method: &str,
|
||||||
|
auth_user: &str,
|
||||||
|
auth_pass: &str,
|
||||||
|
) -> Option<()> {
|
||||||
|
if let Some(value) = strip_prefix(authorization.as_bytes(), b"Basic ") {
|
||||||
|
let value: Vec<u8> = general_purpose::STANDARD.decode(value).ok()?;
|
||||||
|
let parts: Vec<&str> = std::str::from_utf8(&value).ok()?.split(':').collect();
|
||||||
|
|
||||||
|
if parts[0] != auth_user {
|
||||||
|
return None;
|
||||||
|
}
|
||||||
|
|
||||||
|
if auth_pass.starts_with("$6$") {
|
||||||
|
if let Ok(()) = sha_crypt::sha512_check(parts[1], auth_pass) {
|
||||||
|
return Some(());
|
||||||
|
}
|
||||||
|
} else if parts[1] == auth_pass {
|
||||||
|
return Some(());
|
||||||
|
}
|
||||||
|
|
||||||
|
None
|
||||||
|
} else if let Some(value) = strip_prefix(authorization.as_bytes(), b"Digest ") {
|
||||||
|
let digest_map = to_headermap(value).ok()?;
|
||||||
|
if let (Some(username), Some(nonce), Some(user_response)) = (
|
||||||
|
digest_map
|
||||||
|
.get(b"username".as_ref())
|
||||||
|
.and_then(|b| std::str::from_utf8(b).ok()),
|
||||||
|
digest_map.get(b"nonce".as_ref()),
|
||||||
|
digest_map.get(b"response".as_ref()),
|
||||||
|
) {
|
||||||
|
match validate_nonce(nonce) {
|
||||||
|
Ok(true) => {}
|
||||||
|
_ => return None,
|
||||||
|
}
|
||||||
|
if auth_user != username {
|
||||||
|
return None;
|
||||||
|
}
|
||||||
|
|
||||||
|
let mut h = Context::new();
|
||||||
|
h.consume(format!("{}:{}:{}", auth_user, REALM, auth_pass).as_bytes());
|
||||||
|
let auth_pass = format!("{:x}", h.compute());
|
||||||
|
|
||||||
|
let mut ha = Context::new();
|
||||||
|
ha.consume(method);
|
||||||
|
ha.consume(b":");
|
||||||
|
if let Some(uri) = digest_map.get(b"uri".as_ref()) {
|
||||||
|
ha.consume(uri);
|
||||||
|
}
|
||||||
|
let ha = format!("{:x}", ha.compute());
|
||||||
|
let mut correct_response = None;
|
||||||
|
if let Some(qop) = digest_map.get(b"qop".as_ref()) {
|
||||||
|
if qop == &b"auth".as_ref() || qop == &b"auth-int".as_ref() {
|
||||||
|
correct_response = Some({
|
||||||
|
let mut c = Context::new();
|
||||||
|
c.consume(&auth_pass);
|
||||||
|
c.consume(b":");
|
||||||
|
c.consume(nonce);
|
||||||
|
c.consume(b":");
|
||||||
|
if let Some(nc) = digest_map.get(b"nc".as_ref()) {
|
||||||
|
c.consume(nc);
|
||||||
|
}
|
||||||
|
c.consume(b":");
|
||||||
|
if let Some(cnonce) = digest_map.get(b"cnonce".as_ref()) {
|
||||||
|
c.consume(cnonce);
|
||||||
|
}
|
||||||
|
c.consume(b":");
|
||||||
|
c.consume(qop);
|
||||||
|
c.consume(b":");
|
||||||
|
c.consume(&*ha);
|
||||||
|
format!("{:x}", c.compute())
|
||||||
|
});
|
||||||
|
}
|
||||||
|
}
|
||||||
|
let correct_response = match correct_response {
|
||||||
|
Some(r) => r,
|
||||||
|
None => {
|
||||||
|
let mut c = Context::new();
|
||||||
|
c.consume(&auth_pass);
|
||||||
|
c.consume(b":");
|
||||||
|
c.consume(nonce);
|
||||||
|
c.consume(b":");
|
||||||
|
c.consume(&*ha);
|
||||||
|
format!("{:x}", c.compute())
|
||||||
|
}
|
||||||
|
};
|
||||||
|
if correct_response.as_bytes() == *user_response {
|
||||||
|
return Some(());
|
||||||
|
}
|
||||||
|
}
|
||||||
|
None
|
||||||
|
} else {
|
||||||
|
None
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
/// Check if a nonce is still valid.
|
/// Check if a nonce is still valid.
|
||||||
/// Return an error if it was never valid
|
/// Return an error if it was never valid
|
||||||
fn validate_nonce(nonce: &[u8]) -> Result<bool> {
|
fn validate_nonce(nonce: &[u8]) -> Result<bool> {
|
||||||
@@ -355,6 +432,13 @@ fn validate_nonce(nonce: &[u8]) -> Result<bool> {
|
|||||||
bail!("invalid nonce");
|
bail!("invalid nonce");
|
||||||
}
|
}
|
||||||
|
|
||||||
|
fn is_readonly_method(method: &Method) -> bool {
|
||||||
|
method == Method::GET
|
||||||
|
|| method == Method::OPTIONS
|
||||||
|
|| method == Method::HEAD
|
||||||
|
|| method.as_str() == "PROPFIND"
|
||||||
|
}
|
||||||
|
|
||||||
fn strip_prefix<'a>(search: &'a [u8], prefix: &[u8]) -> Option<&'a [u8]> {
|
fn strip_prefix<'a>(search: &'a [u8], prefix: &[u8]) -> Option<&'a [u8]> {
|
||||||
let l = prefix.len();
|
let l = prefix.len();
|
||||||
if search.len() < l {
|
if search.len() < l {
|
||||||
@@ -417,3 +501,126 @@ fn create_nonce() -> Result<String> {
|
|||||||
let n = format!("{:08x}{:032x}", secs, h.compute());
|
let n = format!("{:08x}{:032x}", secs, h.compute());
|
||||||
Ok(n[..34].to_string())
|
Ok(n[..34].to_string())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
fn split_account_paths(s: &str) -> Option<(&str, &str)> {
|
||||||
|
let i = s.find("@/")?;
|
||||||
|
Some((&s[0..i], &s[i + 1..]))
|
||||||
|
}
|
||||||
|
|
||||||
|
fn split_rules(rules: &[&str]) -> Vec<String> {
|
||||||
|
let mut output = vec![];
|
||||||
|
for rule in rules {
|
||||||
|
let parts: Vec<&str> = rule.split('|').collect();
|
||||||
|
let mut rules_list = vec![];
|
||||||
|
let mut concated_part = String::new();
|
||||||
|
for (i, part) in parts.iter().enumerate() {
|
||||||
|
if part.contains("@/") {
|
||||||
|
concated_part.push_str(part);
|
||||||
|
let mut concated_part_tmp = String::new();
|
||||||
|
std::mem::swap(&mut concated_part_tmp, &mut concated_part);
|
||||||
|
rules_list.push(concated_part_tmp);
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
concated_part.push_str(part);
|
||||||
|
if i < parts.len() - 1 {
|
||||||
|
concated_part.push('|');
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if !concated_part.is_empty() {
|
||||||
|
rules_list.push(concated_part)
|
||||||
|
}
|
||||||
|
output.extend(rules_list);
|
||||||
|
}
|
||||||
|
output
|
||||||
|
}
|
||||||
|
|
||||||
|
#[cfg(test)]
|
||||||
|
mod tests {
|
||||||
|
use super::*;
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_split_account_paths() {
|
||||||
|
assert_eq!(
|
||||||
|
split_account_paths("user:pass@/:rw"),
|
||||||
|
Some(("user:pass", "/:rw"))
|
||||||
|
);
|
||||||
|
assert_eq!(
|
||||||
|
split_account_paths("user:pass@@/:rw"),
|
||||||
|
Some(("user:pass@", "/:rw"))
|
||||||
|
);
|
||||||
|
assert_eq!(
|
||||||
|
split_account_paths("user:pass@1@/:rw"),
|
||||||
|
Some(("user:pass@1", "/:rw"))
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_compact_split_rules() {
|
||||||
|
assert_eq!(
|
||||||
|
split_rules(&["user1:pass1@/:rw|user2:pass2@/:rw"]),
|
||||||
|
["user1:pass1@/:rw", "user2:pass2@/:rw"]
|
||||||
|
);
|
||||||
|
assert_eq!(
|
||||||
|
split_rules(&["user1:pa|ss1@/:rw|user2:pa|ss2@/:rw"]),
|
||||||
|
["user1:pa|ss1@/:rw", "user2:pa|ss2@/:rw"]
|
||||||
|
);
|
||||||
|
assert_eq!(
|
||||||
|
split_rules(&["user1:pa|ss1@/:rw|@/"]),
|
||||||
|
["user1:pa|ss1@/:rw", "@/"]
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_access_paths() {
|
||||||
|
let mut paths = AccessPaths::default();
|
||||||
|
paths.add("/dir1", AccessPerm::ReadWrite);
|
||||||
|
paths.add("/dir2/dir21", AccessPerm::ReadWrite);
|
||||||
|
paths.add("/dir2/dir21/dir211", AccessPerm::ReadOnly);
|
||||||
|
paths.add("/dir2/dir21/dir212", AccessPerm::Forbidden);
|
||||||
|
paths.add("/dir2/dir22", AccessPerm::ReadOnly);
|
||||||
|
paths.add("/dir2/dir22/dir221", AccessPerm::ReadWrite);
|
||||||
|
paths.add("/dir2/dir23/dir231", AccessPerm::ReadWrite);
|
||||||
|
assert_eq!(
|
||||||
|
paths.child_paths(Path::new("/tmp")),
|
||||||
|
[
|
||||||
|
"/tmp/dir1",
|
||||||
|
"/tmp/dir2/dir21",
|
||||||
|
"/tmp/dir2/dir22",
|
||||||
|
"/tmp/dir2/dir23/dir231",
|
||||||
|
]
|
||||||
|
.iter()
|
||||||
|
.map(PathBuf::from)
|
||||||
|
.collect::<Vec<_>>()
|
||||||
|
);
|
||||||
|
assert_eq!(
|
||||||
|
paths
|
||||||
|
.find("dir2", false)
|
||||||
|
.map(|v| v.child_paths(Path::new("/tmp/dir2"))),
|
||||||
|
Some(
|
||||||
|
[
|
||||||
|
"/tmp/dir2/dir21",
|
||||||
|
"/tmp/dir2/dir22",
|
||||||
|
"/tmp/dir2/dir23/dir231"
|
||||||
|
]
|
||||||
|
.iter()
|
||||||
|
.map(PathBuf::from)
|
||||||
|
.collect::<Vec<_>>()
|
||||||
|
)
|
||||||
|
);
|
||||||
|
assert_eq!(paths.find("dir2", true), None);
|
||||||
|
assert_eq!(
|
||||||
|
paths.find("dir1/file", true),
|
||||||
|
Some(AccessPaths::new(AccessPerm::ReadWrite))
|
||||||
|
);
|
||||||
|
assert_eq!(
|
||||||
|
paths.find("dir2/dir21/file", true),
|
||||||
|
Some(AccessPaths::new(AccessPerm::ReadWrite))
|
||||||
|
);
|
||||||
|
assert_eq!(
|
||||||
|
paths.find("dir2/dir21/dir211/file", false),
|
||||||
|
Some(AccessPaths::new(AccessPerm::ReadOnly))
|
||||||
|
);
|
||||||
|
assert_eq!(paths.find("dir2/dir21/dir211/file", true), None);
|
||||||
|
assert_eq!(paths.find("dir2/dir21/dir212", false), None);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|||||||
@@ -1,23 +1,29 @@
|
|||||||
use std::{collections::HashMap, str::FromStr, sync::Arc};
|
use std::{collections::HashMap, str::FromStr};
|
||||||
|
|
||||||
use crate::{args::Args, server::Request};
|
use crate::{auth::get_auth_user, server::Request};
|
||||||
|
|
||||||
pub const DEFAULT_LOG_FORMAT: &str = r#"$remote_addr "$request" $status"#;
|
pub const DEFAULT_LOG_FORMAT: &str = r#"$remote_addr "$request" $status"#;
|
||||||
|
|
||||||
#[derive(Debug)]
|
#[derive(Debug, Clone, PartialEq)]
|
||||||
pub struct LogHttp {
|
pub struct HttpLogger {
|
||||||
elements: Vec<LogElement>,
|
elements: Vec<LogElement>,
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug)]
|
impl Default for HttpLogger {
|
||||||
|
fn default() -> Self {
|
||||||
|
DEFAULT_LOG_FORMAT.parse().unwrap()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, Clone, PartialEq)]
|
||||||
enum LogElement {
|
enum LogElement {
|
||||||
Variable(String),
|
Variable(String),
|
||||||
Header(String),
|
Header(String),
|
||||||
Literal(String),
|
Literal(String),
|
||||||
}
|
}
|
||||||
|
|
||||||
impl LogHttp {
|
impl HttpLogger {
|
||||||
pub fn data(&self, req: &Request, args: &Arc<Args>) -> HashMap<String, String> {
|
pub fn data(&self, req: &Request) -> HashMap<String, String> {
|
||||||
let mut data = HashMap::default();
|
let mut data = HashMap::default();
|
||||||
for element in self.elements.iter() {
|
for element in self.elements.iter() {
|
||||||
match element {
|
match element {
|
||||||
@@ -26,10 +32,8 @@ impl LogHttp {
|
|||||||
data.insert(name.to_string(), format!("{} {}", req.method(), req.uri()));
|
data.insert(name.to_string(), format!("{} {}", req.method(), req.uri()));
|
||||||
}
|
}
|
||||||
"remote_user" => {
|
"remote_user" => {
|
||||||
if let Some(user) = req
|
if let Some(user) =
|
||||||
.headers()
|
req.headers().get("authorization").and_then(get_auth_user)
|
||||||
.get("authorization")
|
|
||||||
.and_then(|v| args.auth_method.get_user(v))
|
|
||||||
{
|
{
|
||||||
data.insert(name.to_string(), user);
|
data.insert(name.to_string(), user);
|
||||||
}
|
}
|
||||||
@@ -66,7 +70,7 @@ impl LogHttp {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl FromStr for LogHttp {
|
impl FromStr for HttpLogger {
|
||||||
type Err = anyhow::Error;
|
type Err = anyhow::Error;
|
||||||
fn from_str(s: &str) -> Result<Self, Self::Err> {
|
fn from_str(s: &str) -> Result<Self, Self::Err> {
|
||||||
let mut elements = vec![];
|
let mut elements = vec![];
|
||||||
105
src/http_utils.rs
Normal file
105
src/http_utils.rs
Normal file
@@ -0,0 +1,105 @@
|
|||||||
|
use bytes::{Bytes, BytesMut};
|
||||||
|
use futures_util::Stream;
|
||||||
|
use http_body_util::{combinators::BoxBody, BodyExt, Full};
|
||||||
|
use hyper::body::{Body, Incoming};
|
||||||
|
use std::{
|
||||||
|
pin::Pin,
|
||||||
|
task::{Context, Poll},
|
||||||
|
};
|
||||||
|
use tokio::io::AsyncRead;
|
||||||
|
use tokio_util::io::poll_read_buf;
|
||||||
|
|
||||||
|
#[derive(Debug)]
|
||||||
|
pub struct IncomingStream {
|
||||||
|
inner: Incoming,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl IncomingStream {
|
||||||
|
pub fn new(inner: Incoming) -> Self {
|
||||||
|
Self { inner }
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl Stream for IncomingStream {
|
||||||
|
type Item = Result<Bytes, anyhow::Error>;
|
||||||
|
|
||||||
|
#[inline]
|
||||||
|
fn poll_next(mut self: Pin<&mut Self>, cx: &mut Context<'_>) -> Poll<Option<Self::Item>> {
|
||||||
|
loop {
|
||||||
|
match futures_util::ready!(Pin::new(&mut self.inner).poll_frame(cx)?) {
|
||||||
|
Some(frame) => match frame.into_data() {
|
||||||
|
Ok(data) => return Poll::Ready(Some(Ok(data))),
|
||||||
|
Err(_frame) => {}
|
||||||
|
},
|
||||||
|
None => return Poll::Ready(None),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pin_project_lite::pin_project! {
|
||||||
|
pub struct LengthLimitedStream<R> {
|
||||||
|
#[pin]
|
||||||
|
reader: Option<R>,
|
||||||
|
remaining: usize,
|
||||||
|
buf: BytesMut,
|
||||||
|
capacity: usize,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<R> LengthLimitedStream<R> {
|
||||||
|
pub fn new(reader: R, limit: usize) -> Self {
|
||||||
|
Self {
|
||||||
|
reader: Some(reader),
|
||||||
|
remaining: limit,
|
||||||
|
buf: BytesMut::new(),
|
||||||
|
capacity: 4096,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<R: AsyncRead> Stream for LengthLimitedStream<R> {
|
||||||
|
type Item = std::io::Result<Bytes>;
|
||||||
|
fn poll_next(mut self: Pin<&mut Self>, cx: &mut Context<'_>) -> Poll<Option<Self::Item>> {
|
||||||
|
let mut this = self.as_mut().project();
|
||||||
|
|
||||||
|
if *this.remaining == 0 {
|
||||||
|
self.project().reader.set(None);
|
||||||
|
return Poll::Ready(None);
|
||||||
|
}
|
||||||
|
|
||||||
|
let reader = match this.reader.as_pin_mut() {
|
||||||
|
Some(r) => r,
|
||||||
|
None => return Poll::Ready(None),
|
||||||
|
};
|
||||||
|
|
||||||
|
if this.buf.capacity() == 0 {
|
||||||
|
this.buf.reserve(*this.capacity);
|
||||||
|
}
|
||||||
|
|
||||||
|
match poll_read_buf(reader, cx, &mut this.buf) {
|
||||||
|
Poll::Pending => Poll::Pending,
|
||||||
|
Poll::Ready(Err(err)) => {
|
||||||
|
self.project().reader.set(None);
|
||||||
|
Poll::Ready(Some(Err(err)))
|
||||||
|
}
|
||||||
|
Poll::Ready(Ok(0)) => {
|
||||||
|
self.project().reader.set(None);
|
||||||
|
Poll::Ready(None)
|
||||||
|
}
|
||||||
|
Poll::Ready(Ok(_)) => {
|
||||||
|
let mut chunk = this.buf.split();
|
||||||
|
let chunk_size = (*this.remaining).min(chunk.len());
|
||||||
|
chunk.truncate(chunk_size);
|
||||||
|
*this.remaining -= chunk_size;
|
||||||
|
Poll::Ready(Some(Ok(chunk.freeze())))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn body_full(content: impl Into<hyper::body::Bytes>) -> BoxBody<Bytes, anyhow::Error> {
|
||||||
|
Full::new(content.into())
|
||||||
|
.map_err(anyhow::Error::new)
|
||||||
|
.boxed()
|
||||||
|
}
|
||||||
188
src/main.rs
188
src/main.rs
@@ -1,38 +1,37 @@
|
|||||||
mod args;
|
mod args;
|
||||||
mod auth;
|
mod auth;
|
||||||
mod log_http;
|
mod http_logger;
|
||||||
|
mod http_utils;
|
||||||
mod logger;
|
mod logger;
|
||||||
mod server;
|
mod server;
|
||||||
mod streamer;
|
|
||||||
#[cfg(feature = "tls")]
|
|
||||||
mod tls;
|
|
||||||
#[cfg(unix)]
|
|
||||||
mod unix;
|
|
||||||
mod utils;
|
mod utils;
|
||||||
|
|
||||||
#[macro_use]
|
#[macro_use]
|
||||||
extern crate log;
|
extern crate log;
|
||||||
|
|
||||||
use crate::args::{build_cli, print_completions, Args};
|
use crate::args::{build_cli, print_completions, Args};
|
||||||
use crate::server::{Request, Server};
|
use crate::server::Server;
|
||||||
#[cfg(feature = "tls")]
|
#[cfg(feature = "tls")]
|
||||||
use crate::tls::{TlsAcceptor, TlsStream};
|
use crate::utils::{load_certs, load_private_key};
|
||||||
|
|
||||||
use anyhow::{anyhow, Context, Result};
|
use anyhow::{anyhow, Context, Result};
|
||||||
use std::net::{IpAddr, SocketAddr, TcpListener as StdTcpListener};
|
|
||||||
use std::sync::atomic::{AtomicBool, Ordering};
|
|
||||||
use std::sync::Arc;
|
|
||||||
|
|
||||||
use args::BindAddr;
|
use args::BindAddr;
|
||||||
use clap_complete::Shell;
|
use clap_complete::Shell;
|
||||||
use futures::future::join_all;
|
use futures_util::future::join_all;
|
||||||
use tokio::net::TcpListener;
|
|
||||||
use tokio::task::JoinHandle;
|
|
||||||
|
|
||||||
use hyper::server::conn::{AddrIncoming, AddrStream};
|
use hyper::{body::Incoming, service::service_fn, Request};
|
||||||
use hyper::service::{make_service_fn, service_fn};
|
use hyper_util::{
|
||||||
|
rt::{TokioExecutor, TokioIo},
|
||||||
|
server::conn::auto::Builder,
|
||||||
|
};
|
||||||
|
use std::net::{IpAddr, SocketAddr, TcpListener as StdTcpListener};
|
||||||
|
use std::sync::{
|
||||||
|
atomic::{AtomicBool, Ordering},
|
||||||
|
Arc,
|
||||||
|
};
|
||||||
|
use tokio::{net::TcpListener, task::JoinHandle};
|
||||||
#[cfg(feature = "tls")]
|
#[cfg(feature = "tls")]
|
||||||
use rustls::ServerConfig;
|
use tokio_rustls::{rustls::ServerConfig, TlsAcceptor};
|
||||||
|
|
||||||
#[tokio::main]
|
#[tokio::main]
|
||||||
async fn main() -> Result<()> {
|
async fn main() -> Result<()> {
|
||||||
@@ -45,10 +44,10 @@ async fn main() -> Result<()> {
|
|||||||
return Ok(());
|
return Ok(());
|
||||||
}
|
}
|
||||||
let args = Args::parse(matches)?;
|
let args = Args::parse(matches)?;
|
||||||
let args = Arc::new(args);
|
|
||||||
let running = Arc::new(AtomicBool::new(true));
|
let running = Arc::new(AtomicBool::new(true));
|
||||||
let handles = serve(args.clone(), running.clone())?;
|
let listening = print_listening(&args)?;
|
||||||
print_listening(args)?;
|
let handles = serve(args, running.clone())?;
|
||||||
|
println!("{listening}");
|
||||||
|
|
||||||
tokio::select! {
|
tokio::select! {
|
||||||
ret = join_all(handles) => {
|
ret = join_all(handles) => {
|
||||||
@@ -66,58 +65,66 @@ async fn main() -> Result<()> {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
fn serve(
|
fn serve(args: Args, running: Arc<AtomicBool>) -> Result<Vec<JoinHandle<()>>> {
|
||||||
args: Arc<Args>,
|
let addrs = args.addrs.clone();
|
||||||
running: Arc<AtomicBool>,
|
|
||||||
) -> Result<Vec<JoinHandle<Result<(), hyper::Error>>>> {
|
|
||||||
let inner = Arc::new(Server::init(args.clone(), running)?);
|
|
||||||
let mut handles = vec![];
|
|
||||||
let port = args.port;
|
let port = args.port;
|
||||||
for bind_addr in args.addrs.iter() {
|
let tls_config = (args.tls_cert.clone(), args.tls_key.clone());
|
||||||
let inner = inner.clone();
|
let server_handle = Arc::new(Server::init(args, running)?);
|
||||||
let serve_func = move |remote_addr: Option<SocketAddr>| {
|
let mut handles = vec![];
|
||||||
let inner = inner.clone();
|
for bind_addr in addrs.iter() {
|
||||||
async move {
|
let server_handle = server_handle.clone();
|
||||||
Ok::<_, hyper::Error>(service_fn(move |req: Request| {
|
|
||||||
let inner = inner.clone();
|
|
||||||
inner.call(req, remote_addr)
|
|
||||||
}))
|
|
||||||
}
|
|
||||||
};
|
|
||||||
match bind_addr {
|
match bind_addr {
|
||||||
BindAddr::Address(ip) => {
|
BindAddr::Address(ip) => {
|
||||||
let incoming = create_addr_incoming(SocketAddr::new(*ip, port))
|
let listener = create_listener(SocketAddr::new(*ip, port))
|
||||||
.with_context(|| format!("Failed to bind `{ip}:{port}`"))?;
|
.with_context(|| format!("Failed to bind `{ip}:{port}`"))?;
|
||||||
match args.tls.as_ref() {
|
|
||||||
|
match &tls_config {
|
||||||
#[cfg(feature = "tls")]
|
#[cfg(feature = "tls")]
|
||||||
Some((certs, key)) => {
|
(Some(cert_file), Some(key_file)) => {
|
||||||
let config = ServerConfig::builder()
|
let certs = load_certs(cert_file)?;
|
||||||
.with_safe_defaults()
|
let key = load_private_key(key_file)?;
|
||||||
|
let mut config = ServerConfig::builder()
|
||||||
.with_no_client_auth()
|
.with_no_client_auth()
|
||||||
.with_single_cert(certs.clone(), key.clone())?;
|
.with_single_cert(certs, key)?;
|
||||||
|
config.alpn_protocols = vec![b"h2".to_vec(), b"http/1.1".to_vec()];
|
||||||
let config = Arc::new(config);
|
let config = Arc::new(config);
|
||||||
let accepter = TlsAcceptor::new(config.clone(), incoming);
|
let tls_accepter = TlsAcceptor::from(config);
|
||||||
let new_service = make_service_fn(move |socket: &TlsStream| {
|
|
||||||
let remote_addr = socket.remote_addr();
|
let handle = tokio::spawn(async move {
|
||||||
serve_func(Some(remote_addr))
|
loop {
|
||||||
|
let (cnx, addr) = listener.accept().await.unwrap();
|
||||||
|
let Ok(stream) = tls_accepter.accept(cnx).await else {
|
||||||
|
warn!("During cls handshake connection from {}", addr);
|
||||||
|
continue;
|
||||||
|
};
|
||||||
|
let stream = TokioIo::new(stream);
|
||||||
|
tokio::spawn(handle_stream(
|
||||||
|
server_handle.clone(),
|
||||||
|
stream,
|
||||||
|
Some(addr),
|
||||||
|
));
|
||||||
|
}
|
||||||
});
|
});
|
||||||
let server =
|
|
||||||
tokio::spawn(hyper::Server::builder(accepter).serve(new_service));
|
handles.push(handle);
|
||||||
handles.push(server);
|
|
||||||
}
|
}
|
||||||
#[cfg(not(feature = "tls"))]
|
(None, None) => {
|
||||||
Some(_) => {
|
let handle = tokio::spawn(async move {
|
||||||
|
loop {
|
||||||
|
let (cnx, addr) = listener.accept().await.unwrap();
|
||||||
|
let stream = TokioIo::new(cnx);
|
||||||
|
tokio::spawn(handle_stream(
|
||||||
|
server_handle.clone(),
|
||||||
|
stream,
|
||||||
|
Some(addr),
|
||||||
|
));
|
||||||
|
}
|
||||||
|
});
|
||||||
|
handles.push(handle);
|
||||||
|
}
|
||||||
|
_ => {
|
||||||
unreachable!()
|
unreachable!()
|
||||||
}
|
}
|
||||||
None => {
|
|
||||||
let new_service = make_service_fn(move |socket: &AddrStream| {
|
|
||||||
let remote_addr = socket.remote_addr();
|
|
||||||
serve_func(Some(remote_addr))
|
|
||||||
});
|
|
||||||
let server =
|
|
||||||
tokio::spawn(hyper::Server::builder(incoming).serve(new_service));
|
|
||||||
handles.push(server);
|
|
||||||
}
|
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
BindAddr::Path(path) => {
|
BindAddr::Path(path) => {
|
||||||
@@ -128,10 +135,15 @@ fn serve(
|
|||||||
{
|
{
|
||||||
let listener = tokio::net::UnixListener::bind(path)
|
let listener = tokio::net::UnixListener::bind(path)
|
||||||
.with_context(|| format!("Failed to bind `{}`", path.display()))?;
|
.with_context(|| format!("Failed to bind `{}`", path.display()))?;
|
||||||
let acceptor = unix::UnixAcceptor::from_listener(listener);
|
let handle = tokio::spawn(async move {
|
||||||
let new_service = make_service_fn(move |_| serve_func(None));
|
loop {
|
||||||
let server = tokio::spawn(hyper::Server::builder(acceptor).serve(new_service));
|
let (cnx, _) = listener.accept().await.unwrap();
|
||||||
handles.push(server);
|
let stream = TokioIo::new(cnx);
|
||||||
|
tokio::spawn(handle_stream(server_handle.clone(), stream, None));
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
|
handles.push(handle);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -139,7 +151,30 @@ fn serve(
|
|||||||
Ok(handles)
|
Ok(handles)
|
||||||
}
|
}
|
||||||
|
|
||||||
fn create_addr_incoming(addr: SocketAddr) -> Result<AddrIncoming> {
|
async fn handle_stream<T>(handle: Arc<Server>, stream: TokioIo<T>, addr: Option<SocketAddr>)
|
||||||
|
where
|
||||||
|
T: tokio::io::AsyncRead + tokio::io::AsyncWrite + Unpin + Send + 'static,
|
||||||
|
{
|
||||||
|
let hyper_service =
|
||||||
|
service_fn(move |request: Request<Incoming>| handle.clone().call(request, addr));
|
||||||
|
|
||||||
|
let ret = Builder::new(TokioExecutor::new())
|
||||||
|
.serve_connection_with_upgrades(stream, hyper_service)
|
||||||
|
.await;
|
||||||
|
|
||||||
|
if let Err(err) = ret {
|
||||||
|
let scope = match addr {
|
||||||
|
Some(addr) => format!(" from {}", addr),
|
||||||
|
None => String::new(),
|
||||||
|
};
|
||||||
|
match err.downcast_ref::<std::io::Error>() {
|
||||||
|
Some(err) if err.kind() == std::io::ErrorKind::UnexpectedEof => {}
|
||||||
|
_ => warn!("Serving connection{}: {}", scope, err),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn create_listener(addr: SocketAddr) -> Result<TcpListener> {
|
||||||
use socket2::{Domain, Protocol, Socket, Type};
|
use socket2::{Domain, Protocol, Socket, Type};
|
||||||
let socket = Socket::new(Domain::for_address(addr), Type::STREAM, Some(Protocol::TCP))?;
|
let socket = Socket::new(Domain::for_address(addr), Type::STREAM, Some(Protocol::TCP))?;
|
||||||
if addr.is_ipv6() {
|
if addr.is_ipv6() {
|
||||||
@@ -150,11 +185,12 @@ fn create_addr_incoming(addr: SocketAddr) -> Result<AddrIncoming> {
|
|||||||
socket.listen(1024 /* Default backlog */)?;
|
socket.listen(1024 /* Default backlog */)?;
|
||||||
let std_listener = StdTcpListener::from(socket);
|
let std_listener = StdTcpListener::from(socket);
|
||||||
std_listener.set_nonblocking(true)?;
|
std_listener.set_nonblocking(true)?;
|
||||||
let incoming = AddrIncoming::from_listener(TcpListener::from_std(std_listener)?)?;
|
let listener = TcpListener::from_std(std_listener)?;
|
||||||
Ok(incoming)
|
Ok(listener)
|
||||||
}
|
}
|
||||||
|
|
||||||
fn print_listening(args: Arc<Args>) -> Result<()> {
|
fn print_listening(args: &Args) -> Result<String> {
|
||||||
|
let mut output = String::new();
|
||||||
let mut bind_addrs = vec![];
|
let mut bind_addrs = vec![];
|
||||||
let (mut ipv4, mut ipv6) = (false, false);
|
let (mut ipv4, mut ipv6) = (false, false);
|
||||||
for bind_addr in args.addrs.iter() {
|
for bind_addr in args.addrs.iter() {
|
||||||
@@ -195,7 +231,11 @@ fn print_listening(args: Arc<Args>) -> Result<()> {
|
|||||||
IpAddr::V4(_) => format!("{}:{}", addr, args.port),
|
IpAddr::V4(_) => format!("{}:{}", addr, args.port),
|
||||||
IpAddr::V6(_) => format!("[{}]:{}", addr, args.port),
|
IpAddr::V6(_) => format!("[{}]:{}", addr, args.port),
|
||||||
};
|
};
|
||||||
let protocol = if args.tls.is_some() { "https" } else { "http" };
|
let protocol = if args.tls_cert.is_some() {
|
||||||
|
"https"
|
||||||
|
} else {
|
||||||
|
"http"
|
||||||
|
};
|
||||||
format!("{}://{}{}", protocol, addr, args.uri_prefix)
|
format!("{}://{}{}", protocol, addr, args.uri_prefix)
|
||||||
}
|
}
|
||||||
BindAddr::Path(path) => path.display().to_string(),
|
BindAddr::Path(path) => path.display().to_string(),
|
||||||
@@ -203,17 +243,17 @@ fn print_listening(args: Arc<Args>) -> Result<()> {
|
|||||||
.collect::<Vec<_>>();
|
.collect::<Vec<_>>();
|
||||||
|
|
||||||
if urls.len() == 1 {
|
if urls.len() == 1 {
|
||||||
println!("Listening on {}", urls[0]);
|
output.push_str(&format!("Listening on {}", urls[0]))
|
||||||
} else {
|
} else {
|
||||||
let info = urls
|
let info = urls
|
||||||
.iter()
|
.iter()
|
||||||
.map(|v| format!(" {v}"))
|
.map(|v| format!(" {v}"))
|
||||||
.collect::<Vec<String>>()
|
.collect::<Vec<String>>()
|
||||||
.join("\n");
|
.join("\n");
|
||||||
println!("Listening on:\n{info}\n");
|
output.push_str(&format!("Listening on:\n{info}\n"))
|
||||||
}
|
}
|
||||||
|
|
||||||
Ok(())
|
Ok(output)
|
||||||
}
|
}
|
||||||
|
|
||||||
async fn shutdown_signal() {
|
async fn shutdown_signal() {
|
||||||
|
|||||||
885
src/server.rs
885
src/server.rs
File diff suppressed because it is too large
Load Diff
@@ -1,68 +0,0 @@
|
|||||||
use async_stream::stream;
|
|
||||||
use futures::{Stream, StreamExt};
|
|
||||||
use std::io::Error;
|
|
||||||
use std::pin::Pin;
|
|
||||||
use tokio::io::{AsyncRead, AsyncReadExt};
|
|
||||||
|
|
||||||
pub struct Streamer<R>
|
|
||||||
where
|
|
||||||
R: AsyncRead + Unpin + Send + 'static,
|
|
||||||
{
|
|
||||||
reader: R,
|
|
||||||
buf_size: usize,
|
|
||||||
}
|
|
||||||
|
|
||||||
impl<R> Streamer<R>
|
|
||||||
where
|
|
||||||
R: AsyncRead + Unpin + Send + 'static,
|
|
||||||
{
|
|
||||||
#[inline]
|
|
||||||
pub fn new(reader: R, buf_size: usize) -> Self {
|
|
||||||
Self { reader, buf_size }
|
|
||||||
}
|
|
||||||
pub fn into_stream(
|
|
||||||
mut self,
|
|
||||||
) -> Pin<Box<impl ?Sized + Stream<Item = Result<Vec<u8>, Error>> + 'static>> {
|
|
||||||
let stream = stream! {
|
|
||||||
loop {
|
|
||||||
let mut buf = vec![0; self.buf_size];
|
|
||||||
let r = self.reader.read(&mut buf).await?;
|
|
||||||
if r == 0 {
|
|
||||||
break
|
|
||||||
}
|
|
||||||
buf.truncate(r);
|
|
||||||
yield Ok(buf);
|
|
||||||
}
|
|
||||||
};
|
|
||||||
stream.boxed()
|
|
||||||
}
|
|
||||||
// allow truncation as truncated remaining is always less than buf_size: usize
|
|
||||||
pub fn into_stream_sized(
|
|
||||||
mut self,
|
|
||||||
max_length: u64,
|
|
||||||
) -> Pin<Box<impl ?Sized + Stream<Item = Result<Vec<u8>, Error>> + 'static>> {
|
|
||||||
let stream = stream! {
|
|
||||||
let mut remaining = max_length;
|
|
||||||
loop {
|
|
||||||
if remaining == 0 {
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
let bs = if remaining >= self.buf_size as u64 {
|
|
||||||
self.buf_size
|
|
||||||
} else {
|
|
||||||
remaining as usize
|
|
||||||
};
|
|
||||||
let mut buf = vec![0; bs];
|
|
||||||
let r = self.reader.read(&mut buf).await?;
|
|
||||||
if r == 0 {
|
|
||||||
break;
|
|
||||||
} else {
|
|
||||||
buf.truncate(r);
|
|
||||||
yield Ok(buf);
|
|
||||||
}
|
|
||||||
remaining -= r as u64;
|
|
||||||
}
|
|
||||||
};
|
|
||||||
stream.boxed()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
161
src/tls.rs
161
src/tls.rs
@@ -1,161 +0,0 @@
|
|||||||
use anyhow::{anyhow, bail, Context as AnyhowContext, Result};
|
|
||||||
use core::task::{Context, Poll};
|
|
||||||
use futures::ready;
|
|
||||||
use hyper::server::accept::Accept;
|
|
||||||
use hyper::server::conn::{AddrIncoming, AddrStream};
|
|
||||||
use rustls::{Certificate, PrivateKey};
|
|
||||||
use std::future::Future;
|
|
||||||
use std::net::SocketAddr;
|
|
||||||
use std::path::Path;
|
|
||||||
use std::pin::Pin;
|
|
||||||
use std::sync::Arc;
|
|
||||||
use std::{fs, io};
|
|
||||||
use tokio::io::{AsyncRead, AsyncWrite, ReadBuf};
|
|
||||||
use tokio_rustls::rustls::ServerConfig;
|
|
||||||
|
|
||||||
enum State {
|
|
||||||
Handshaking(tokio_rustls::Accept<AddrStream>),
|
|
||||||
Streaming(tokio_rustls::server::TlsStream<AddrStream>),
|
|
||||||
}
|
|
||||||
|
|
||||||
// tokio_rustls::server::TlsStream doesn't expose constructor methods,
|
|
||||||
// so we have to TlsAcceptor::accept and handshake to have access to it
|
|
||||||
// TlsStream implements AsyncRead/AsyncWrite handshaking tokio_rustls::Accept first
|
|
||||||
pub struct TlsStream {
|
|
||||||
state: State,
|
|
||||||
remote_addr: SocketAddr,
|
|
||||||
}
|
|
||||||
|
|
||||||
impl TlsStream {
|
|
||||||
fn new(stream: AddrStream, config: Arc<ServerConfig>) -> TlsStream {
|
|
||||||
let remote_addr = stream.remote_addr();
|
|
||||||
let accept = tokio_rustls::TlsAcceptor::from(config).accept(stream);
|
|
||||||
TlsStream {
|
|
||||||
state: State::Handshaking(accept),
|
|
||||||
remote_addr,
|
|
||||||
}
|
|
||||||
}
|
|
||||||
pub fn remote_addr(&self) -> SocketAddr {
|
|
||||||
self.remote_addr
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
impl AsyncRead for TlsStream {
|
|
||||||
fn poll_read(
|
|
||||||
self: Pin<&mut Self>,
|
|
||||||
cx: &mut Context,
|
|
||||||
buf: &mut ReadBuf,
|
|
||||||
) -> Poll<io::Result<()>> {
|
|
||||||
let pin = self.get_mut();
|
|
||||||
match pin.state {
|
|
||||||
State::Handshaking(ref mut accept) => match ready!(Pin::new(accept).poll(cx)) {
|
|
||||||
Ok(mut stream) => {
|
|
||||||
let result = Pin::new(&mut stream).poll_read(cx, buf);
|
|
||||||
pin.state = State::Streaming(stream);
|
|
||||||
result
|
|
||||||
}
|
|
||||||
Err(err) => Poll::Ready(Err(err)),
|
|
||||||
},
|
|
||||||
State::Streaming(ref mut stream) => Pin::new(stream).poll_read(cx, buf),
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
impl AsyncWrite for TlsStream {
|
|
||||||
fn poll_write(
|
|
||||||
self: Pin<&mut Self>,
|
|
||||||
cx: &mut Context<'_>,
|
|
||||||
buf: &[u8],
|
|
||||||
) -> Poll<io::Result<usize>> {
|
|
||||||
let pin = self.get_mut();
|
|
||||||
match pin.state {
|
|
||||||
State::Handshaking(ref mut accept) => match ready!(Pin::new(accept).poll(cx)) {
|
|
||||||
Ok(mut stream) => {
|
|
||||||
let result = Pin::new(&mut stream).poll_write(cx, buf);
|
|
||||||
pin.state = State::Streaming(stream);
|
|
||||||
result
|
|
||||||
}
|
|
||||||
Err(err) => Poll::Ready(Err(err)),
|
|
||||||
},
|
|
||||||
State::Streaming(ref mut stream) => Pin::new(stream).poll_write(cx, buf),
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
fn poll_flush(mut self: Pin<&mut Self>, cx: &mut Context<'_>) -> Poll<io::Result<()>> {
|
|
||||||
match self.state {
|
|
||||||
State::Handshaking(_) => Poll::Ready(Ok(())),
|
|
||||||
State::Streaming(ref mut stream) => Pin::new(stream).poll_flush(cx),
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
fn poll_shutdown(mut self: Pin<&mut Self>, cx: &mut Context<'_>) -> Poll<io::Result<()>> {
|
|
||||||
match self.state {
|
|
||||||
State::Handshaking(_) => Poll::Ready(Ok(())),
|
|
||||||
State::Streaming(ref mut stream) => Pin::new(stream).poll_shutdown(cx),
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
pub struct TlsAcceptor {
|
|
||||||
config: Arc<ServerConfig>,
|
|
||||||
incoming: AddrIncoming,
|
|
||||||
}
|
|
||||||
|
|
||||||
impl TlsAcceptor {
|
|
||||||
pub fn new(config: Arc<ServerConfig>, incoming: AddrIncoming) -> TlsAcceptor {
|
|
||||||
TlsAcceptor { config, incoming }
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
impl Accept for TlsAcceptor {
|
|
||||||
type Conn = TlsStream;
|
|
||||||
type Error = io::Error;
|
|
||||||
|
|
||||||
fn poll_accept(
|
|
||||||
self: Pin<&mut Self>,
|
|
||||||
cx: &mut Context<'_>,
|
|
||||||
) -> Poll<Option<Result<Self::Conn, Self::Error>>> {
|
|
||||||
let pin = self.get_mut();
|
|
||||||
match ready!(Pin::new(&mut pin.incoming).poll_accept(cx)) {
|
|
||||||
Some(Ok(sock)) => Poll::Ready(Some(Ok(TlsStream::new(sock, pin.config.clone())))),
|
|
||||||
Some(Err(e)) => Poll::Ready(Some(Err(e))),
|
|
||||||
None => Poll::Ready(None),
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Load public certificate from file.
|
|
||||||
pub fn load_certs<T: AsRef<Path>>(filename: T) -> Result<Vec<Certificate>> {
|
|
||||||
// Open certificate file.
|
|
||||||
let cert_file = fs::File::open(filename.as_ref())
|
|
||||||
.with_context(|| format!("Failed to access `{}`", filename.as_ref().display()))?;
|
|
||||||
let mut reader = io::BufReader::new(cert_file);
|
|
||||||
|
|
||||||
// Load and return certificate.
|
|
||||||
let certs = rustls_pemfile::certs(&mut reader).with_context(|| "Failed to load certificate")?;
|
|
||||||
if certs.is_empty() {
|
|
||||||
bail!("No supported certificate in file");
|
|
||||||
}
|
|
||||||
Ok(certs.into_iter().map(Certificate).collect())
|
|
||||||
}
|
|
||||||
|
|
||||||
// Load private key from file.
|
|
||||||
pub fn load_private_key<T: AsRef<Path>>(filename: T) -> Result<PrivateKey> {
|
|
||||||
let key_file = fs::File::open(filename.as_ref())
|
|
||||||
.with_context(|| format!("Failed to access `{}`", filename.as_ref().display()))?;
|
|
||||||
let mut reader = io::BufReader::new(key_file);
|
|
||||||
|
|
||||||
// Load and return a single private key.
|
|
||||||
let keys = rustls_pemfile::read_all(&mut reader)
|
|
||||||
.with_context(|| "There was a problem with reading private key")?
|
|
||||||
.into_iter()
|
|
||||||
.find_map(|item| match item {
|
|
||||||
rustls_pemfile::Item::RSAKey(key)
|
|
||||||
| rustls_pemfile::Item::PKCS8Key(key)
|
|
||||||
| rustls_pemfile::Item::ECKey(key) => Some(key),
|
|
||||||
_ => None,
|
|
||||||
})
|
|
||||||
.ok_or_else(|| anyhow!("No supported private key in file"))?;
|
|
||||||
|
|
||||||
Ok(PrivateKey(keys))
|
|
||||||
}
|
|
||||||
31
src/unix.rs
31
src/unix.rs
@@ -1,31 +0,0 @@
|
|||||||
use hyper::server::accept::Accept;
|
|
||||||
use tokio::net::UnixListener;
|
|
||||||
|
|
||||||
use std::pin::Pin;
|
|
||||||
use std::task::{Context, Poll};
|
|
||||||
|
|
||||||
pub struct UnixAcceptor {
|
|
||||||
inner: UnixListener,
|
|
||||||
}
|
|
||||||
|
|
||||||
impl UnixAcceptor {
|
|
||||||
pub fn from_listener(listener: UnixListener) -> Self {
|
|
||||||
Self { inner: listener }
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
impl Accept for UnixAcceptor {
|
|
||||||
type Conn = tokio::net::UnixStream;
|
|
||||||
type Error = std::io::Error;
|
|
||||||
|
|
||||||
fn poll_accept(
|
|
||||||
self: Pin<&mut Self>,
|
|
||||||
cx: &mut Context<'_>,
|
|
||||||
) -> Poll<Option<Result<Self::Conn, Self::Error>>> {
|
|
||||||
match self.inner.poll_accept(cx) {
|
|
||||||
Poll::Pending => Poll::Pending,
|
|
||||||
Poll::Ready(Ok((socket, _addr))) => Poll::Ready(Some(Ok(socket))),
|
|
||||||
Poll::Ready(Err(err)) => Poll::Ready(Some(Err(err))),
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
151
src/utils.rs
151
src/utils.rs
@@ -1,4 +1,7 @@
|
|||||||
use anyhow::{anyhow, Context, Result};
|
use anyhow::{anyhow, Context, Result};
|
||||||
|
use chrono::{DateTime, Utc};
|
||||||
|
#[cfg(feature = "tls")]
|
||||||
|
use rustls_pki_types::{CertificateDer, PrivateKeyDer};
|
||||||
use std::{
|
use std::{
|
||||||
borrow::Cow,
|
borrow::Cow,
|
||||||
path::Path,
|
path::Path,
|
||||||
@@ -28,6 +31,21 @@ pub fn get_file_name(path: &Path) -> &str {
|
|||||||
.unwrap_or_default()
|
.unwrap_or_default()
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[cfg(unix)]
|
||||||
|
pub async fn get_file_mtime_and_mode(path: &Path) -> Result<(DateTime<Utc>, u16)> {
|
||||||
|
use std::os::unix::prelude::MetadataExt;
|
||||||
|
let meta = tokio::fs::metadata(path).await?;
|
||||||
|
let datetime: DateTime<Utc> = meta.modified()?.into();
|
||||||
|
Ok((datetime, meta.mode() as u16))
|
||||||
|
}
|
||||||
|
|
||||||
|
#[cfg(not(unix))]
|
||||||
|
pub async fn get_file_mtime_and_mode(path: &Path) -> Result<(DateTime<Utc>, u16)> {
|
||||||
|
let meta = tokio::fs::metadata(&path).await?;
|
||||||
|
let datetime: DateTime<Utc> = meta.modified()?.into();
|
||||||
|
Ok((datetime, 0o644))
|
||||||
|
}
|
||||||
|
|
||||||
pub fn try_get_file_name(path: &Path) -> Result<&str> {
|
pub fn try_get_file_name(path: &Path) -> Result<&str> {
|
||||||
path.file_name()
|
path.file_name()
|
||||||
.and_then(|v| v.to_str())
|
.and_then(|v| v.to_str())
|
||||||
@@ -42,26 +60,115 @@ pub fn glob(pattern: &str, target: &str) -> bool {
|
|||||||
pat.matches(target)
|
pat.matches(target)
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
// Load public certificate from file.
|
||||||
fn test_glob_key() {
|
#[cfg(feature = "tls")]
|
||||||
assert!(glob("", ""));
|
pub fn load_certs<T: AsRef<Path>>(filename: T) -> Result<Vec<CertificateDer<'static>>> {
|
||||||
assert!(glob(".*", ".git"));
|
// Open certificate file.
|
||||||
assert!(glob("abc", "abc"));
|
let cert_file = std::fs::File::open(filename.as_ref())
|
||||||
assert!(glob("a*c", "abc"));
|
.with_context(|| format!("Failed to access `{}`", filename.as_ref().display()))?;
|
||||||
assert!(glob("a?c", "abc"));
|
let mut reader = std::io::BufReader::new(cert_file);
|
||||||
assert!(glob("a*c", "abbc"));
|
|
||||||
assert!(glob("*c", "abc"));
|
// Load and return certificate.
|
||||||
assert!(glob("a*", "abc"));
|
let mut certs = vec![];
|
||||||
assert!(glob("?c", "bc"));
|
for cert in rustls_pemfile::certs(&mut reader) {
|
||||||
assert!(glob("a?", "ab"));
|
let cert = cert.with_context(|| "Failed to load certificate")?;
|
||||||
assert!(!glob("abc", "adc"));
|
certs.push(cert)
|
||||||
assert!(!glob("abc", "abcd"));
|
}
|
||||||
assert!(!glob("a?c", "abbc"));
|
if certs.is_empty() {
|
||||||
assert!(!glob("*.log", "log"));
|
anyhow::bail!("No supported certificate in file");
|
||||||
assert!(glob("*.abc-cba", "xyz.abc-cba"));
|
}
|
||||||
assert!(glob("*.abc-cba", "123.xyz.abc-cba"));
|
Ok(certs)
|
||||||
assert!(glob("*.log", ".log"));
|
}
|
||||||
assert!(glob("*.log", "a.log"));
|
|
||||||
assert!(glob("*/", "abc/"));
|
// Load private key from file.
|
||||||
assert!(!glob("*/", "abc"));
|
#[cfg(feature = "tls")]
|
||||||
|
pub fn load_private_key<T: AsRef<Path>>(filename: T) -> Result<PrivateKeyDer<'static>> {
|
||||||
|
let key_file = std::fs::File::open(filename.as_ref())
|
||||||
|
.with_context(|| format!("Failed to access `{}`", filename.as_ref().display()))?;
|
||||||
|
let mut reader = std::io::BufReader::new(key_file);
|
||||||
|
|
||||||
|
// Load and return a single private key.
|
||||||
|
for key in rustls_pemfile::read_all(&mut reader) {
|
||||||
|
let key = key.with_context(|| "There was a problem with reading private key")?;
|
||||||
|
match key {
|
||||||
|
rustls_pemfile::Item::Pkcs1Key(key) => return Ok(PrivateKeyDer::Pkcs1(key)),
|
||||||
|
rustls_pemfile::Item::Pkcs8Key(key) => return Ok(PrivateKeyDer::Pkcs8(key)),
|
||||||
|
rustls_pemfile::Item::Sec1Key(key) => return Ok(PrivateKeyDer::Sec1(key)),
|
||||||
|
_ => {}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
anyhow::bail!("No supported private key in file");
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn parse_range(range: &str, size: u64) -> Option<(u64, u64)> {
|
||||||
|
let (unit, range) = range.split_once('=')?;
|
||||||
|
if unit != "bytes" || range.contains(',') {
|
||||||
|
return None;
|
||||||
|
}
|
||||||
|
let (start, end) = range.split_once('-')?;
|
||||||
|
if start.is_empty() {
|
||||||
|
let offset = end.parse::<u64>().ok()?;
|
||||||
|
if offset <= size {
|
||||||
|
Some((size - offset, size - 1))
|
||||||
|
} else {
|
||||||
|
None
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
let start = start.parse::<u64>().ok()?;
|
||||||
|
if start < size {
|
||||||
|
if end.is_empty() {
|
||||||
|
Some((start, size - 1))
|
||||||
|
} else {
|
||||||
|
let end = end.parse::<u64>().ok()?;
|
||||||
|
if end < size {
|
||||||
|
Some((start, end))
|
||||||
|
} else {
|
||||||
|
None
|
||||||
|
}
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
None
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[cfg(test)]
|
||||||
|
mod tests {
|
||||||
|
use super::*;
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_glob_key() {
|
||||||
|
assert!(glob("", ""));
|
||||||
|
assert!(glob(".*", ".git"));
|
||||||
|
assert!(glob("abc", "abc"));
|
||||||
|
assert!(glob("a*c", "abc"));
|
||||||
|
assert!(glob("a?c", "abc"));
|
||||||
|
assert!(glob("a*c", "abbc"));
|
||||||
|
assert!(glob("*c", "abc"));
|
||||||
|
assert!(glob("a*", "abc"));
|
||||||
|
assert!(glob("?c", "bc"));
|
||||||
|
assert!(glob("a?", "ab"));
|
||||||
|
assert!(!glob("abc", "adc"));
|
||||||
|
assert!(!glob("abc", "abcd"));
|
||||||
|
assert!(!glob("a?c", "abbc"));
|
||||||
|
assert!(!glob("*.log", "log"));
|
||||||
|
assert!(glob("*.abc-cba", "xyz.abc-cba"));
|
||||||
|
assert!(glob("*.abc-cba", "123.xyz.abc-cba"));
|
||||||
|
assert!(glob("*.log", ".log"));
|
||||||
|
assert!(glob("*.log", "a.log"));
|
||||||
|
assert!(glob("*/", "abc/"));
|
||||||
|
assert!(!glob("*/", "abc"));
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_parse_range() {
|
||||||
|
assert_eq!(parse_range("bytes=0-499", 500), Some((0, 499)));
|
||||||
|
assert_eq!(parse_range("bytes=0-", 500), Some((0, 499)));
|
||||||
|
assert_eq!(parse_range("bytes=299-", 500), Some((299, 499)));
|
||||||
|
assert_eq!(parse_range("bytes=-500", 500), Some((0, 499)));
|
||||||
|
assert_eq!(parse_range("bytes=-300", 500), Some((200, 499)));
|
||||||
|
assert_eq!(parse_range("bytes=500-", 500), None);
|
||||||
|
assert_eq!(parse_range("bytes=-501", 500), None);
|
||||||
|
assert_eq!(parse_range("bytes=0-500", 500), None);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -32,7 +32,7 @@ fn asset_js(server: TestServer) -> Result<(), Error> {
|
|||||||
assert_eq!(resp.status(), 200);
|
assert_eq!(resp.status(), 200);
|
||||||
assert_eq!(
|
assert_eq!(
|
||||||
resp.headers().get("content-type").unwrap(),
|
resp.headers().get("content-type").unwrap(),
|
||||||
"application/javascript"
|
"application/javascript; charset=UTF-8"
|
||||||
);
|
);
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
@@ -46,7 +46,10 @@ fn asset_css(server: TestServer) -> Result<(), Error> {
|
|||||||
);
|
);
|
||||||
let resp = reqwest::blocking::get(url)?;
|
let resp = reqwest::blocking::get(url)?;
|
||||||
assert_eq!(resp.status(), 200);
|
assert_eq!(resp.status(), 200);
|
||||||
assert_eq!(resp.headers().get("content-type").unwrap(), "text/css");
|
assert_eq!(
|
||||||
|
resp.headers().get("content-type").unwrap(),
|
||||||
|
"text/css; charset=UTF-8"
|
||||||
|
);
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -90,7 +93,7 @@ fn asset_js_with_prefix(
|
|||||||
assert_eq!(resp.status(), 200);
|
assert_eq!(resp.status(), 200);
|
||||||
assert_eq!(
|
assert_eq!(
|
||||||
resp.headers().get("content-type").unwrap(),
|
resp.headers().get("content-type").unwrap(),
|
||||||
"application/javascript"
|
"application/javascript; charset=UTF-8"
|
||||||
);
|
);
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|||||||
200
tests/auth.rs
200
tests/auth.rs
@@ -3,13 +3,22 @@ mod utils;
|
|||||||
|
|
||||||
use diqwest::blocking::WithDigestAuth;
|
use diqwest::blocking::WithDigestAuth;
|
||||||
use fixtures::{server, Error, TestServer};
|
use fixtures::{server, Error, TestServer};
|
||||||
|
use indexmap::IndexSet;
|
||||||
use rstest::rstest;
|
use rstest::rstest;
|
||||||
|
|
||||||
#[rstest]
|
#[rstest]
|
||||||
fn no_auth(#[with(&["--auth", "/@user:pass", "-A"])] server: TestServer) -> Result<(), Error> {
|
fn no_auth(#[with(&["--auth", "user:pass@/:rw", "-A"])] server: TestServer) -> Result<(), Error> {
|
||||||
let resp = reqwest::blocking::get(server.url())?;
|
let resp = reqwest::blocking::get(server.url())?;
|
||||||
assert_eq!(resp.status(), 401);
|
assert_eq!(resp.status(), 401);
|
||||||
assert!(resp.headers().contains_key("www-authenticate"));
|
let values: Vec<&str> = resp
|
||||||
|
.headers()
|
||||||
|
.get_all("www-authenticate")
|
||||||
|
.iter()
|
||||||
|
.map(|v| v.to_str().unwrap())
|
||||||
|
.collect();
|
||||||
|
assert!(values[0].starts_with("Digest"));
|
||||||
|
assert!(values[1].starts_with("Basic"));
|
||||||
|
|
||||||
let url = format!("{}file1", server.url());
|
let url = format!("{}file1", server.url());
|
||||||
let resp = fetch!(b"PUT", &url).body(b"abc".to_vec()).send()?;
|
let resp = fetch!(b"PUT", &url).body(b"abc".to_vec()).send()?;
|
||||||
assert_eq!(resp.status(), 401);
|
assert_eq!(resp.status(), 401);
|
||||||
@@ -17,7 +26,49 @@ fn no_auth(#[with(&["--auth", "/@user:pass", "-A"])] server: TestServer) -> Resu
|
|||||||
}
|
}
|
||||||
|
|
||||||
#[rstest]
|
#[rstest]
|
||||||
fn auth(#[with(&["--auth", "/@user:pass", "-A"])] server: TestServer) -> Result<(), Error> {
|
#[case(server(&["--auth", "user:pass@/:rw", "-A"]), "user", "pass")]
|
||||||
|
#[case(server(&["--auth", "user:pa:ss@1@/:rw", "-A"]), "user", "pa:ss@1")]
|
||||||
|
fn auth(#[case] server: TestServer, #[case] user: &str, #[case] pass: &str) -> Result<(), Error> {
|
||||||
|
let url = format!("{}file1", server.url());
|
||||||
|
let resp = fetch!(b"PUT", &url).body(b"abc".to_vec()).send()?;
|
||||||
|
assert_eq!(resp.status(), 401);
|
||||||
|
let resp = fetch!(b"PUT", &url)
|
||||||
|
.body(b"abc".to_vec())
|
||||||
|
.send_with_digest_auth(user, pass)?;
|
||||||
|
assert_eq!(resp.status(), 201);
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
const HASHED_PASSWORD_AUTH: &str = "user:$6$gQxZwKyWn/ZmWEA2$4uV7KKMnSUnET2BtWTj/9T5.Jq3h/MdkOlnIl5hdlTxDZ4MZKmJ.kl6C.NL9xnNPqC4lVHC1vuI0E5cLpTJX81@/:rw"; // user:pass
|
||||||
|
|
||||||
|
#[rstest]
|
||||||
|
fn auth_hashed_password(
|
||||||
|
#[with(&["--auth", HASHED_PASSWORD_AUTH, "-A"])] server: TestServer,
|
||||||
|
) -> Result<(), Error> {
|
||||||
|
let url = format!("{}file1", server.url());
|
||||||
|
let resp = fetch!(b"PUT", &url).body(b"abc".to_vec()).send()?;
|
||||||
|
assert_eq!(resp.status(), 401);
|
||||||
|
if let Err(err) = fetch!(b"PUT", &url)
|
||||||
|
.body(b"abc".to_vec())
|
||||||
|
.send_with_digest_auth("user", "pass")
|
||||||
|
{
|
||||||
|
assert_eq!(
|
||||||
|
format!("{err:?}"),
|
||||||
|
r#"DigestAuth(MissingRequired("realm", "Basic realm=\"DUFS\""))"#
|
||||||
|
);
|
||||||
|
}
|
||||||
|
let resp = fetch!(b"PUT", &url)
|
||||||
|
.body(b"abc".to_vec())
|
||||||
|
.basic_auth("user", Some("pass"))
|
||||||
|
.send()?;
|
||||||
|
assert_eq!(resp.status(), 201);
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
#[rstest]
|
||||||
|
fn auth_and_public(
|
||||||
|
#[with(&["-a", "user:pass@/:rw", "-a", "@/", "-A"])] server: TestServer,
|
||||||
|
) -> Result<(), Error> {
|
||||||
let url = format!("{}file1", server.url());
|
let url = format!("{}file1", server.url());
|
||||||
let resp = fetch!(b"PUT", &url).body(b"abc".to_vec()).send()?;
|
let resp = fetch!(b"PUT", &url).body(b"abc".to_vec()).send()?;
|
||||||
assert_eq!(resp.status(), 401);
|
assert_eq!(resp.status(), 401);
|
||||||
@@ -25,11 +76,14 @@ fn auth(#[with(&["--auth", "/@user:pass", "-A"])] server: TestServer) -> Result<
|
|||||||
.body(b"abc".to_vec())
|
.body(b"abc".to_vec())
|
||||||
.send_with_digest_auth("user", "pass")?;
|
.send_with_digest_auth("user", "pass")?;
|
||||||
assert_eq!(resp.status(), 201);
|
assert_eq!(resp.status(), 201);
|
||||||
|
let resp = fetch!(b"GET", &url).send()?;
|
||||||
|
assert_eq!(resp.status(), 200);
|
||||||
|
assert_eq!(resp.text()?, "abc");
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
#[rstest]
|
#[rstest]
|
||||||
fn auth_skip(#[with(&["--auth", "/@user:pass@*"])] server: TestServer) -> Result<(), Error> {
|
fn auth_skip(#[with(&["--auth", "@/"])] server: TestServer) -> Result<(), Error> {
|
||||||
let resp = reqwest::blocking::get(server.url())?;
|
let resp = reqwest::blocking::get(server.url())?;
|
||||||
assert_eq!(resp.status(), 200);
|
assert_eq!(resp.status(), 200);
|
||||||
Ok(())
|
Ok(())
|
||||||
@@ -37,7 +91,7 @@ fn auth_skip(#[with(&["--auth", "/@user:pass@*"])] server: TestServer) -> Result
|
|||||||
|
|
||||||
#[rstest]
|
#[rstest]
|
||||||
fn auth_skip_on_options_method(
|
fn auth_skip_on_options_method(
|
||||||
#[with(&["--auth", "/@user:pass"])] server: TestServer,
|
#[with(&["--auth", "user:pass@/:rw"])] server: TestServer,
|
||||||
) -> Result<(), Error> {
|
) -> Result<(), Error> {
|
||||||
let url = format!("{}index.html", server.url());
|
let url = format!("{}index.html", server.url());
|
||||||
let resp = fetch!(b"OPTIONS", &url).send()?;
|
let resp = fetch!(b"OPTIONS", &url).send()?;
|
||||||
@@ -47,13 +101,27 @@ fn auth_skip_on_options_method(
|
|||||||
|
|
||||||
#[rstest]
|
#[rstest]
|
||||||
fn auth_check(
|
fn auth_check(
|
||||||
#[with(&["--auth", "/@user:pass@user2:pass2", "-A"])] server: TestServer,
|
#[with(&["--auth", "user:pass@/:rw", "--auth", "user2:pass2@/", "-A"])] server: TestServer,
|
||||||
) -> Result<(), Error> {
|
) -> Result<(), Error> {
|
||||||
let url = format!("{}index.html", server.url());
|
let url = format!("{}index.html", server.url());
|
||||||
let resp = fetch!(b"WRITEABLE", &url).send()?;
|
let resp = fetch!(b"WRITEABLE", &url).send()?;
|
||||||
assert_eq!(resp.status(), 401);
|
assert_eq!(resp.status(), 401);
|
||||||
let resp = fetch!(b"WRITEABLE", &url).send_with_digest_auth("user2", "pass2")?;
|
let resp = fetch!(b"WRITEABLE", &url).send_with_digest_auth("user2", "pass2")?;
|
||||||
|
assert_eq!(resp.status(), 403);
|
||||||
|
let resp = fetch!(b"WRITEABLE", &url).send_with_digest_auth("user", "pass")?;
|
||||||
|
assert_eq!(resp.status(), 200);
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
#[rstest]
|
||||||
|
fn auth_compact_rules(
|
||||||
|
#[with(&["--auth", "user:pass@/:rw|user2:pass2@/", "-A"])] server: TestServer,
|
||||||
|
) -> Result<(), Error> {
|
||||||
|
let url = format!("{}index.html", server.url());
|
||||||
|
let resp = fetch!(b"WRITEABLE", &url).send()?;
|
||||||
assert_eq!(resp.status(), 401);
|
assert_eq!(resp.status(), 401);
|
||||||
|
let resp = fetch!(b"WRITEABLE", &url).send_with_digest_auth("user2", "pass2")?;
|
||||||
|
assert_eq!(resp.status(), 403);
|
||||||
let resp = fetch!(b"WRITEABLE", &url).send_with_digest_auth("user", "pass")?;
|
let resp = fetch!(b"WRITEABLE", &url).send_with_digest_auth("user", "pass")?;
|
||||||
assert_eq!(resp.status(), 200);
|
assert_eq!(resp.status(), 200);
|
||||||
Ok(())
|
Ok(())
|
||||||
@@ -61,7 +129,7 @@ fn auth_check(
|
|||||||
|
|
||||||
#[rstest]
|
#[rstest]
|
||||||
fn auth_readonly(
|
fn auth_readonly(
|
||||||
#[with(&["--auth", "/@user:pass@user2:pass2", "-A"])] server: TestServer,
|
#[with(&["--auth", "user:pass@/:rw", "--auth", "user2:pass2@/", "-A"])] server: TestServer,
|
||||||
) -> Result<(), Error> {
|
) -> Result<(), Error> {
|
||||||
let url = format!("{}index.html", server.url());
|
let url = format!("{}index.html", server.url());
|
||||||
let resp = fetch!(b"GET", &url).send()?;
|
let resp = fetch!(b"GET", &url).send()?;
|
||||||
@@ -72,13 +140,30 @@ fn auth_readonly(
|
|||||||
let resp = fetch!(b"PUT", &url)
|
let resp = fetch!(b"PUT", &url)
|
||||||
.body(b"abc".to_vec())
|
.body(b"abc".to_vec())
|
||||||
.send_with_digest_auth("user2", "pass2")?;
|
.send_with_digest_auth("user2", "pass2")?;
|
||||||
assert_eq!(resp.status(), 401);
|
assert_eq!(resp.status(), 403);
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
#[rstest]
|
||||||
|
fn auth_forbidden(
|
||||||
|
#[with(&["--auth", "user:pass@/:rw,/dir1:-", "-A"])] server: TestServer,
|
||||||
|
) -> Result<(), Error> {
|
||||||
|
let url = format!("{}file1", server.url());
|
||||||
|
let resp = fetch!(b"PUT", &url)
|
||||||
|
.body(b"abc".to_vec())
|
||||||
|
.send_with_digest_auth("user", "pass")?;
|
||||||
|
assert_eq!(resp.status(), 201);
|
||||||
|
let url = format!("{}dir1/file1", server.url());
|
||||||
|
let resp = fetch!(b"PUT", &url)
|
||||||
|
.body(b"abc".to_vec())
|
||||||
|
.send_with_digest_auth("user", "pass")?;
|
||||||
|
assert_eq!(resp.status(), 403);
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
#[rstest]
|
#[rstest]
|
||||||
fn auth_nest(
|
fn auth_nest(
|
||||||
#[with(&["--auth", "/@user:pass@user2:pass2", "--auth", "/dir1@user3:pass3", "-A"])]
|
#[with(&["--auth", "user:pass@/:rw", "--auth", "user2:pass2@/", "--auth", "user3:pass3@/dir1:rw", "-A"])]
|
||||||
server: TestServer,
|
server: TestServer,
|
||||||
) -> Result<(), Error> {
|
) -> Result<(), Error> {
|
||||||
let url = format!("{}dir1/file1", server.url());
|
let url = format!("{}dir1/file1", server.url());
|
||||||
@@ -97,7 +182,8 @@ fn auth_nest(
|
|||||||
|
|
||||||
#[rstest]
|
#[rstest]
|
||||||
fn auth_nest_share(
|
fn auth_nest_share(
|
||||||
#[with(&["--auth", "/@user:pass@*", "--auth", "/dir1@user3:pass3", "-A"])] server: TestServer,
|
#[with(&["--auth", "@/", "--auth", "user:pass@/:rw", "--auth", "user3:pass3@/dir1:rw", "-A"])]
|
||||||
|
server: TestServer,
|
||||||
) -> Result<(), Error> {
|
) -> Result<(), Error> {
|
||||||
let url = format!("{}index.html", server.url());
|
let url = format!("{}index.html", server.url());
|
||||||
let resp = fetch!(b"GET", &url).send()?;
|
let resp = fetch!(b"GET", &url).send()?;
|
||||||
@@ -106,8 +192,8 @@ fn auth_nest_share(
|
|||||||
}
|
}
|
||||||
|
|
||||||
#[rstest]
|
#[rstest]
|
||||||
#[case(server(&["--auth", "/@user:pass", "--auth-method", "basic", "-A"]), "user", "pass")]
|
#[case(server(&["--auth", "user:pass@/:rw", "-A"]), "user", "pass")]
|
||||||
#[case(server(&["--auth", "/@u1:p1", "--auth-method", "basic", "-A"]), "u1", "p1")]
|
#[case(server(&["--auth", "u1:p1@/:rw", "-A"]), "u1", "p1")]
|
||||||
fn auth_basic(
|
fn auth_basic(
|
||||||
#[case] server: TestServer,
|
#[case] server: TestServer,
|
||||||
#[case] user: &str,
|
#[case] user: &str,
|
||||||
@@ -126,7 +212,8 @@ fn auth_basic(
|
|||||||
|
|
||||||
#[rstest]
|
#[rstest]
|
||||||
fn auth_webdav_move(
|
fn auth_webdav_move(
|
||||||
#[with(&["--auth", "/@user:pass@*", "--auth", "/dir1@user3:pass3", "-A"])] server: TestServer,
|
#[with(&["--auth", "user:pass@/:rw", "--auth", "user3:pass3@/dir1:rw", "-A"])]
|
||||||
|
server: TestServer,
|
||||||
) -> Result<(), Error> {
|
) -> Result<(), Error> {
|
||||||
let origin_url = format!("{}dir1/test.html", server.url());
|
let origin_url = format!("{}dir1/test.html", server.url());
|
||||||
let new_url = format!("{}test2.html", server.url());
|
let new_url = format!("{}test2.html", server.url());
|
||||||
@@ -139,7 +226,8 @@ fn auth_webdav_move(
|
|||||||
|
|
||||||
#[rstest]
|
#[rstest]
|
||||||
fn auth_webdav_copy(
|
fn auth_webdav_copy(
|
||||||
#[with(&["--auth", "/@user:pass@*", "--auth", "/dir1@user3:pass3", "-A"])] server: TestServer,
|
#[with(&["--auth", "user:pass@/:rw", "--auth", "user3:pass3@/dir1:rw", "-A"])]
|
||||||
|
server: TestServer,
|
||||||
) -> Result<(), Error> {
|
) -> Result<(), Error> {
|
||||||
let origin_url = format!("{}dir1/test.html", server.url());
|
let origin_url = format!("{}dir1/test.html", server.url());
|
||||||
let new_url = format!("{}test2.html", server.url());
|
let new_url = format!("{}test2.html", server.url());
|
||||||
@@ -152,7 +240,7 @@ fn auth_webdav_copy(
|
|||||||
|
|
||||||
#[rstest]
|
#[rstest]
|
||||||
fn auth_path_prefix(
|
fn auth_path_prefix(
|
||||||
#[with(&["--auth", "/@user:pass", "--path-prefix", "xyz", "-A"])] server: TestServer,
|
#[with(&["--auth", "user:pass@/:rw", "--path-prefix", "xyz", "-A"])] server: TestServer,
|
||||||
) -> Result<(), Error> {
|
) -> Result<(), Error> {
|
||||||
let url = format!("{}xyz/index.html", server.url());
|
let url = format!("{}xyz/index.html", server.url());
|
||||||
let resp = fetch!(b"GET", &url).send()?;
|
let resp = fetch!(b"GET", &url).send()?;
|
||||||
@@ -161,3 +249,85 @@ fn auth_path_prefix(
|
|||||||
assert_eq!(resp.status(), 200);
|
assert_eq!(resp.status(), 200);
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[rstest]
|
||||||
|
fn auth_partial_index(
|
||||||
|
#[with(&["--auth", "user:pass@/dir1:rw,/dir2:rw", "-A"])] server: TestServer,
|
||||||
|
) -> Result<(), Error> {
|
||||||
|
let resp = fetch!(b"GET", server.url()).send_with_digest_auth("user", "pass")?;
|
||||||
|
assert_eq!(resp.status(), 200);
|
||||||
|
let paths = utils::retrieve_index_paths(&resp.text()?);
|
||||||
|
assert_eq!(paths, IndexSet::from(["dir1/".into(), "dir2/".into()]));
|
||||||
|
let resp = fetch!(b"GET", format!("{}?q={}", server.url(), "test.html"))
|
||||||
|
.send_with_digest_auth("user", "pass")?;
|
||||||
|
assert_eq!(resp.status(), 200);
|
||||||
|
let paths = utils::retrieve_index_paths(&resp.text()?);
|
||||||
|
assert_eq!(
|
||||||
|
paths,
|
||||||
|
IndexSet::from(["dir1/test.html".into(), "dir2/test.html".into()])
|
||||||
|
);
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
#[rstest]
|
||||||
|
fn no_auth_propfind_dir(
|
||||||
|
#[with(&["--auth", "admin:admin@/:rw", "--auth", "@/dir-assets", "-A"])] server: TestServer,
|
||||||
|
) -> Result<(), Error> {
|
||||||
|
let resp = fetch!(b"PROPFIND", server.url()).send()?;
|
||||||
|
assert_eq!(resp.status(), 207);
|
||||||
|
let body = resp.text()?;
|
||||||
|
assert!(body.contains("<D:href>/dir-assets/</D:href>"));
|
||||||
|
assert!(body.contains("<D:href>/dir1/</D:href>"));
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
#[rstest]
|
||||||
|
fn auth_propfind_dir(
|
||||||
|
#[with(&["--auth", "admin:admin@/:rw", "--auth", "user:pass@/dir-assets", "-A"])]
|
||||||
|
server: TestServer,
|
||||||
|
) -> Result<(), Error> {
|
||||||
|
let resp = fetch!(b"PROPFIND", server.url()).send_with_digest_auth("user", "pass")?;
|
||||||
|
assert_eq!(resp.status(), 207);
|
||||||
|
let body = resp.text()?;
|
||||||
|
assert!(body.contains("<D:href>/dir-assets/</D:href>"));
|
||||||
|
assert!(!body.contains("<D:href>/dir1/</D:href>"));
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
#[rstest]
|
||||||
|
fn auth_data(
|
||||||
|
#[with(&["-a", "user:pass@/:rw", "-a", "@/", "-A"])] server: TestServer,
|
||||||
|
) -> Result<(), Error> {
|
||||||
|
let resp = reqwest::blocking::get(server.url())?;
|
||||||
|
let content = resp.text()?;
|
||||||
|
let json = utils::retrive_json(&content).unwrap();
|
||||||
|
assert_eq!(json["allow_delete"], serde_json::Value::Bool(false));
|
||||||
|
assert_eq!(json["allow_upload"], serde_json::Value::Bool(false));
|
||||||
|
let resp = fetch!(b"GET", server.url())
|
||||||
|
.basic_auth("user", Some("pass"))
|
||||||
|
.send()?;
|
||||||
|
let content = resp.text()?;
|
||||||
|
let json = utils::retrive_json(&content).unwrap();
|
||||||
|
assert_eq!(json["allow_delete"], serde_json::Value::Bool(true));
|
||||||
|
assert_eq!(json["allow_upload"], serde_json::Value::Bool(true));
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
#[rstest]
|
||||||
|
fn auth_precedence(
|
||||||
|
#[with(&["--auth", "user:pass@/dir1:rw,/dir1/test.txt", "-A"])] server: TestServer,
|
||||||
|
) -> Result<(), Error> {
|
||||||
|
let url = format!("{}dir1/test.txt", server.url());
|
||||||
|
let resp = fetch!(b"PUT", &url)
|
||||||
|
.body(b"abc".to_vec())
|
||||||
|
.send_with_digest_auth("user", "pass")?;
|
||||||
|
assert_eq!(resp.status(), 403);
|
||||||
|
|
||||||
|
let url = format!("{}dir1/file1", server.url());
|
||||||
|
let resp = fetch!(b"PUT", &url)
|
||||||
|
.body(b"abc".to_vec())
|
||||||
|
.send_with_digest_auth("user", "pass")?;
|
||||||
|
assert_eq!(resp.status(), 201);
|
||||||
|
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|||||||
@@ -76,9 +76,7 @@ fn validate_printed_urls(tmpdir: TempDir, port: u16, #[case] args: &[&str]) -> R
|
|||||||
.collect::<Vec<_>>();
|
.collect::<Vec<_>>();
|
||||||
|
|
||||||
assert!(!urls.is_empty());
|
assert!(!urls.is_empty());
|
||||||
for url in urls {
|
reqwest::blocking::get(urls[0])?.error_for_status()?;
|
||||||
reqwest::blocking::get(url)?.error_for_status()?;
|
|
||||||
}
|
|
||||||
|
|
||||||
child.kill()?;
|
child.kill()?;
|
||||||
|
|
||||||
|
|||||||
56
tests/config.rs
Normal file
56
tests/config.rs
Normal file
@@ -0,0 +1,56 @@
|
|||||||
|
mod fixtures;
|
||||||
|
mod utils;
|
||||||
|
|
||||||
|
use assert_cmd::prelude::*;
|
||||||
|
use assert_fs::TempDir;
|
||||||
|
use diqwest::blocking::WithDigestAuth;
|
||||||
|
use fixtures::{port, tmpdir, wait_for_port, Error};
|
||||||
|
use rstest::rstest;
|
||||||
|
use std::path::PathBuf;
|
||||||
|
use std::process::{Command, Stdio};
|
||||||
|
|
||||||
|
#[rstest]
|
||||||
|
fn use_config_file(tmpdir: TempDir, port: u16) -> Result<(), Error> {
|
||||||
|
let config_path = get_config_path().display().to_string();
|
||||||
|
let mut child = Command::cargo_bin("dufs")?
|
||||||
|
.arg(tmpdir.path())
|
||||||
|
.arg("-p")
|
||||||
|
.arg(port.to_string())
|
||||||
|
.args(["--config", &config_path])
|
||||||
|
.stdout(Stdio::piped())
|
||||||
|
.spawn()?;
|
||||||
|
|
||||||
|
wait_for_port(port);
|
||||||
|
|
||||||
|
let url = format!("http://localhost:{port}/dufs/index.html");
|
||||||
|
let resp = fetch!(b"GET", &url).send()?;
|
||||||
|
assert_eq!(resp.status(), 401);
|
||||||
|
|
||||||
|
let url = format!("http://localhost:{port}/dufs/index.html");
|
||||||
|
let resp = fetch!(b"GET", &url).send_with_digest_auth("user", "pass")?;
|
||||||
|
assert_eq!(resp.text()?, "This is index.html");
|
||||||
|
|
||||||
|
let url = format!("http://localhost:{port}/dufs?simple");
|
||||||
|
let resp = fetch!(b"GET", &url).send_with_digest_auth("user", "pass")?;
|
||||||
|
let text: String = resp.text().unwrap();
|
||||||
|
assert!(text.split('\n').any(|c| c == "dir1/"));
|
||||||
|
assert!(!text.split('\n').any(|c| c == "dir3/"));
|
||||||
|
assert!(!text.split('\n').any(|c| c == "test.txt"));
|
||||||
|
|
||||||
|
let url = format!("http://localhost:{port}/dufs/dir1/upload.txt");
|
||||||
|
let resp = fetch!(b"PUT", &url)
|
||||||
|
.body("Hello")
|
||||||
|
.send_with_digest_auth("user", "pass")?;
|
||||||
|
assert_eq!(resp.status(), 201);
|
||||||
|
|
||||||
|
child.kill()?;
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
fn get_config_path() -> PathBuf {
|
||||||
|
let mut path = std::env::current_dir().expect("Failed to get current directory");
|
||||||
|
path.push("tests");
|
||||||
|
path.push("data");
|
||||||
|
path.push("config.yaml");
|
||||||
|
path
|
||||||
|
}
|
||||||
@@ -19,15 +19,15 @@ fn cors(#[with(&["--enable-cors"])] server: TestServer) -> Result<(), Error> {
|
|||||||
);
|
);
|
||||||
assert_eq!(
|
assert_eq!(
|
||||||
resp.headers().get("access-control-allow-methods").unwrap(),
|
resp.headers().get("access-control-allow-methods").unwrap(),
|
||||||
"GET,HEAD,PUT,OPTIONS,DELETE,PROPFIND,COPY,MOVE"
|
"*"
|
||||||
);
|
);
|
||||||
assert_eq!(
|
assert_eq!(
|
||||||
resp.headers().get("access-control-allow-headers").unwrap(),
|
resp.headers().get("access-control-allow-headers").unwrap(),
|
||||||
"Authorization,Destination,Range,Content-Type"
|
"Authorization,*"
|
||||||
);
|
);
|
||||||
assert_eq!(
|
assert_eq!(
|
||||||
resp.headers().get("access-control-expose-headers").unwrap(),
|
resp.headers().get("access-control-expose-headers").unwrap(),
|
||||||
"WWW-Authenticate,Content-Range,Accept-Ranges,Content-Disposition"
|
"Authorization,*"
|
||||||
);
|
);
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|||||||
9
tests/data/config.yaml
Normal file
9
tests/data/config.yaml
Normal file
@@ -0,0 +1,9 @@
|
|||||||
|
bind:
|
||||||
|
- 0.0.0.0
|
||||||
|
path-prefix: dufs
|
||||||
|
hidden:
|
||||||
|
- dir3
|
||||||
|
- test.txt
|
||||||
|
auth:
|
||||||
|
- user:pass@/:rw
|
||||||
|
allow-upload: true
|
||||||
@@ -16,7 +16,14 @@ pub const BIN_FILE: &str = "😀.bin";
|
|||||||
|
|
||||||
/// File names for testing purpose
|
/// File names for testing purpose
|
||||||
#[allow(dead_code)]
|
#[allow(dead_code)]
|
||||||
pub static FILES: &[&str] = &["test.txt", "test.html", "index.html", BIN_FILE];
|
pub static FILES: &[&str] = &[
|
||||||
|
"test.txt",
|
||||||
|
"test.html",
|
||||||
|
"index.html",
|
||||||
|
#[cfg(not(target_os = "windows"))]
|
||||||
|
"file\n1.txt",
|
||||||
|
BIN_FILE,
|
||||||
|
];
|
||||||
|
|
||||||
/// Directory names for testing directory don't exist
|
/// Directory names for testing directory don't exist
|
||||||
#[allow(dead_code)]
|
#[allow(dead_code)]
|
||||||
@@ -46,7 +53,7 @@ pub fn tmpdir() -> TempDir {
|
|||||||
let tmpdir = assert_fs::TempDir::new().expect("Couldn't create a temp dir for tests");
|
let tmpdir = assert_fs::TempDir::new().expect("Couldn't create a temp dir for tests");
|
||||||
for file in FILES {
|
for file in FILES {
|
||||||
if *file == BIN_FILE {
|
if *file == BIN_FILE {
|
||||||
tmpdir.child(file).write_binary(b"bin\0\0123").unwrap();
|
tmpdir.child(file).write_binary(b"bin\0\x00123").unwrap();
|
||||||
} else {
|
} else {
|
||||||
tmpdir
|
tmpdir
|
||||||
.child(file)
|
.child(file)
|
||||||
@@ -58,7 +65,7 @@ pub fn tmpdir() -> TempDir {
|
|||||||
if *directory == DIR_ASSETS {
|
if *directory == DIR_ASSETS {
|
||||||
tmpdir
|
tmpdir
|
||||||
.child(format!("{}{}", directory, "index.html"))
|
.child(format!("{}{}", directory, "index.html"))
|
||||||
.write_str("__ASSERTS_PREFIX__index.js;DATA = __INDEX_DATA__")
|
.write_str("__ASSETS_PREFIX__index.js;DATA = __INDEX_DATA__")
|
||||||
.unwrap();
|
.unwrap();
|
||||||
} else {
|
} else {
|
||||||
for file in FILES {
|
for file in FILES {
|
||||||
@@ -68,7 +75,7 @@ pub fn tmpdir() -> TempDir {
|
|||||||
if *file == BIN_FILE {
|
if *file == BIN_FILE {
|
||||||
tmpdir
|
tmpdir
|
||||||
.child(format!("{directory}{file}"))
|
.child(format!("{directory}{file}"))
|
||||||
.write_binary(b"bin\0\0123")
|
.write_binary(b"bin\0\x00123")
|
||||||
.unwrap();
|
.unwrap();
|
||||||
} else {
|
} else {
|
||||||
tmpdir
|
tmpdir
|
||||||
|
|||||||
@@ -40,7 +40,12 @@ fn head_dir_404(server: TestServer) -> Result<(), Error> {
|
|||||||
}
|
}
|
||||||
|
|
||||||
#[rstest]
|
#[rstest]
|
||||||
fn get_dir_zip(#[with(&["-A"])] server: TestServer) -> Result<(), Error> {
|
#[case(server(&["--allow-archive"] as &[&str]))]
|
||||||
|
#[case(server(&["--allow-archive", "--compress", "none"]))]
|
||||||
|
#[case(server(&["--allow-archive", "--compress", "low"]))]
|
||||||
|
#[case(server(&["--allow-archive", "--compress", "medium"]))]
|
||||||
|
#[case(server(&["--allow-archive", "--compress", "high"]))]
|
||||||
|
fn get_dir_zip(#[case] server: TestServer) -> Result<(), Error> {
|
||||||
let resp = reqwest::blocking::get(format!("{}?zip", server.url()))?;
|
let resp = reqwest::blocking::get(format!("{}?zip", server.url()))?;
|
||||||
assert_eq!(resp.status(), 200);
|
assert_eq!(resp.status(), 200);
|
||||||
assert_eq!(
|
assert_eq!(
|
||||||
@@ -123,6 +128,15 @@ fn get_dir_search3(#[with(&["-A"])] server: TestServer) -> Result<(), Error> {
|
|||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[rstest]
|
||||||
|
fn get_dir_search4(#[with(&["-A"])] server: TestServer) -> Result<(), Error> {
|
||||||
|
let resp = reqwest::blocking::get(format!("{}dir1?q=dir1&simple", server.url()))?;
|
||||||
|
assert_eq!(resp.status(), 200);
|
||||||
|
let text = resp.text().unwrap();
|
||||||
|
assert!(text.is_empty());
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
#[rstest]
|
#[rstest]
|
||||||
fn head_dir_search(#[with(&["-A"])] server: TestServer) -> Result<(), Error> {
|
fn head_dir_search(#[with(&["-A"])] server: TestServer) -> Result<(), Error> {
|
||||||
let resp = fetch!(b"HEAD", format!("{}?q={}", server.url(), "test.html")).send()?;
|
let resp = fetch!(b"HEAD", format!("{}?q={}", server.url(), "test.html")).send()?;
|
||||||
@@ -138,9 +152,7 @@ fn head_dir_search(#[with(&["-A"])] server: TestServer) -> Result<(), Error> {
|
|||||||
#[rstest]
|
#[rstest]
|
||||||
fn empty_search(#[with(&["-A"])] server: TestServer) -> Result<(), Error> {
|
fn empty_search(#[with(&["-A"])] server: TestServer) -> Result<(), Error> {
|
||||||
let resp = reqwest::blocking::get(format!("{}?q=", server.url()))?;
|
let resp = reqwest::blocking::get(format!("{}?q=", server.url()))?;
|
||||||
assert_eq!(resp.status(), 200);
|
assert_resp_paths!(resp);
|
||||||
let paths = utils::retrieve_index_paths(&resp.text()?);
|
|
||||||
assert!(paths.is_empty());
|
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -184,6 +196,29 @@ fn get_file_404(server: TestServer) -> Result<(), Error> {
|
|||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[rstest]
|
||||||
|
fn get_file_emoji_path(server: TestServer) -> Result<(), Error> {
|
||||||
|
let resp = reqwest::blocking::get(format!("{}{BIN_FILE}", server.url()))?;
|
||||||
|
assert_eq!(resp.status(), 200);
|
||||||
|
assert_eq!(
|
||||||
|
resp.headers().get("content-disposition").unwrap(),
|
||||||
|
"inline; filename=\"😀.bin\"; filename*=UTF-8''%F0%9F%98%80.bin"
|
||||||
|
);
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
#[cfg(not(target_os = "windows"))]
|
||||||
|
#[rstest]
|
||||||
|
fn get_file_newline_path(server: TestServer) -> Result<(), Error> {
|
||||||
|
let resp = reqwest::blocking::get(format!("{}file%0A1.txt", server.url()))?;
|
||||||
|
assert_eq!(resp.status(), 200);
|
||||||
|
assert_eq!(
|
||||||
|
resp.headers().get("content-disposition").unwrap(),
|
||||||
|
"inline; filename=\"file 1.txt\""
|
||||||
|
);
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
#[rstest]
|
#[rstest]
|
||||||
fn get_file_edit(server: TestServer) -> Result<(), Error> {
|
fn get_file_edit(server: TestServer) -> Result<(), Error> {
|
||||||
let resp = fetch!(b"GET", format!("{}index.html?edit", server.url())).send()?;
|
let resp = fetch!(b"GET", format!("{}index.html?edit", server.url())).send()?;
|
||||||
@@ -215,9 +250,12 @@ fn options_dir(server: TestServer) -> Result<(), Error> {
|
|||||||
assert_eq!(resp.status(), 200);
|
assert_eq!(resp.status(), 200);
|
||||||
assert_eq!(
|
assert_eq!(
|
||||||
resp.headers().get("allow").unwrap(),
|
resp.headers().get("allow").unwrap(),
|
||||||
"GET,HEAD,PUT,OPTIONS,DELETE,PROPFIND,COPY,MOVE"
|
"GET,HEAD,PUT,OPTIONS,DELETE,PATCH,PROPFIND,COPY,MOVE"
|
||||||
|
);
|
||||||
|
assert_eq!(
|
||||||
|
resp.headers().get("dav").unwrap(),
|
||||||
|
"1, 2, 3, sabredav-partialupdate"
|
||||||
);
|
);
|
||||||
assert_eq!(resp.headers().get("dav").unwrap(), "1,2");
|
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -295,3 +333,19 @@ fn get_file_content_type(server: TestServer) -> Result<(), Error> {
|
|||||||
);
|
);
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[rstest]
|
||||||
|
fn resumable_upload(#[with(&["--allow-upload"])] server: TestServer) -> Result<(), Error> {
|
||||||
|
let url = format!("{}file1", server.url());
|
||||||
|
let resp = fetch!(b"PUT", &url).body(b"abc".to_vec()).send()?;
|
||||||
|
assert_eq!(resp.status(), 201);
|
||||||
|
let resp = fetch!(b"PATCH", &url)
|
||||||
|
.header("X-Update-Range", "append")
|
||||||
|
.body(b"123".to_vec())
|
||||||
|
.send()?;
|
||||||
|
assert_eq!(resp.status(), 204);
|
||||||
|
let resp = reqwest::blocking::get(url)?;
|
||||||
|
assert_eq!(resp.status(), 200);
|
||||||
|
assert_eq!(resp.text().unwrap(), "abc123");
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|||||||
@@ -11,8 +11,8 @@ use std::io::Read;
|
|||||||
use std::process::{Command, Stdio};
|
use std::process::{Command, Stdio};
|
||||||
|
|
||||||
#[rstest]
|
#[rstest]
|
||||||
#[case(&["-a", "/@user:pass", "--log-format", "$remote_user"], false)]
|
#[case(&["-a", "user:pass@/:rw", "--log-format", "$remote_user"], false)]
|
||||||
#[case(&["-a", "/@user:pass", "--log-format", "$remote_user", "--auth-method", "basic"], true)]
|
#[case(&["-a", "user:pass@/:rw", "--log-format", "$remote_user"], true)]
|
||||||
fn log_remote_user(
|
fn log_remote_user(
|
||||||
tmpdir: TempDir,
|
tmpdir: TempDir,
|
||||||
port: u16,
|
port: u16,
|
||||||
@@ -41,7 +41,7 @@ fn log_remote_user(
|
|||||||
|
|
||||||
assert_eq!(resp.status(), 200);
|
assert_eq!(resp.status(), 200);
|
||||||
|
|
||||||
let mut buf = [0; 1000];
|
let mut buf = [0; 2048];
|
||||||
let buf_len = stdout.read(&mut buf)?;
|
let buf_len = stdout.read(&mut buf)?;
|
||||||
let output = std::str::from_utf8(&buf[0..buf_len])?;
|
let output = std::str::from_utf8(&buf[0..buf_len])?;
|
||||||
|
|
||||||
@@ -69,10 +69,12 @@ fn no_log(tmpdir: TempDir, port: u16, #[case] args: &[&str]) -> Result<(), Error
|
|||||||
let resp = fetch!(b"GET", &format!("http://localhost:{port}")).send()?;
|
let resp = fetch!(b"GET", &format!("http://localhost:{port}")).send()?;
|
||||||
assert_eq!(resp.status(), 200);
|
assert_eq!(resp.status(), 200);
|
||||||
|
|
||||||
let mut buf = [0; 1000];
|
let mut buf = [0; 2048];
|
||||||
let buf_len = stdout.read(&mut buf)?;
|
let buf_len = stdout.read(&mut buf)?;
|
||||||
let output = std::str::from_utf8(&buf[0..buf_len])?;
|
let output = std::str::from_utf8(&buf[0..buf_len])?;
|
||||||
|
|
||||||
assert_eq!(output.lines().last().unwrap(), "");
|
assert_eq!(output.lines().last().unwrap(), "");
|
||||||
|
|
||||||
|
child.kill()?;
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
@@ -2,7 +2,7 @@ mod fixtures;
|
|||||||
mod utils;
|
mod utils;
|
||||||
|
|
||||||
use fixtures::{server, Error, TestServer};
|
use fixtures::{server, Error, TestServer};
|
||||||
use headers::HeaderValue;
|
use reqwest::header::HeaderValue;
|
||||||
use rstest::rstest;
|
use rstest::rstest;
|
||||||
|
|
||||||
#[rstest]
|
#[rstest]
|
||||||
@@ -23,14 +23,10 @@ fn get_file_range_beyond(server: TestServer) -> Result<(), Error> {
|
|||||||
let resp = fetch!(b"GET", format!("{}index.html", server.url()))
|
let resp = fetch!(b"GET", format!("{}index.html", server.url()))
|
||||||
.header("range", HeaderValue::from_static("bytes=12-20"))
|
.header("range", HeaderValue::from_static("bytes=12-20"))
|
||||||
.send()?;
|
.send()?;
|
||||||
assert_eq!(resp.status(), 206);
|
assert_eq!(resp.status(), 416);
|
||||||
assert_eq!(
|
assert_eq!(resp.headers().get("content-range").unwrap(), "bytes */18");
|
||||||
resp.headers().get("content-range").unwrap(),
|
|
||||||
"bytes 12-17/18"
|
|
||||||
);
|
|
||||||
assert_eq!(resp.headers().get("accept-ranges").unwrap(), "bytes");
|
assert_eq!(resp.headers().get("accept-ranges").unwrap(), "bytes");
|
||||||
assert_eq!(resp.headers().get("content-length").unwrap(), "6");
|
assert_eq!(resp.headers().get("content-length").unwrap(), "0");
|
||||||
assert_eq!(resp.text()?, "x.html");
|
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -53,7 +53,7 @@ fn path_prefix_single_file(tmpdir: TempDir, port: u16, #[case] file: &str) -> Re
|
|||||||
let resp = reqwest::blocking::get(format!("http://localhost:{port}/xyz/index.html"))?;
|
let resp = reqwest::blocking::get(format!("http://localhost:{port}/xyz/index.html"))?;
|
||||||
assert_eq!(resp.text()?, "This is index.html");
|
assert_eq!(resp.text()?, "This is index.html");
|
||||||
let resp = reqwest::blocking::get(format!("http://localhost:{port}"))?;
|
let resp = reqwest::blocking::get(format!("http://localhost:{port}"))?;
|
||||||
assert_eq!(resp.status(), 404);
|
assert_eq!(resp.status(), 403);
|
||||||
|
|
||||||
child.kill()?;
|
child.kill()?;
|
||||||
Ok(())
|
Ok(())
|
||||||
|
|||||||
22
tests/tls.rs
22
tests/tls.rs
@@ -7,6 +7,8 @@ use predicates::str::contains;
|
|||||||
use reqwest::blocking::ClientBuilder;
|
use reqwest::blocking::ClientBuilder;
|
||||||
use rstest::rstest;
|
use rstest::rstest;
|
||||||
|
|
||||||
|
use crate::fixtures::port;
|
||||||
|
|
||||||
/// Can start the server with TLS and receive encrypted responses.
|
/// Can start the server with TLS and receive encrypted responses.
|
||||||
#[rstest]
|
#[rstest]
|
||||||
#[case(server(&[
|
#[case(server(&[
|
||||||
@@ -33,8 +35,16 @@ fn tls_works(#[case] server: TestServer) -> Result<(), Error> {
|
|||||||
/// Wrong path for cert throws error.
|
/// Wrong path for cert throws error.
|
||||||
#[rstest]
|
#[rstest]
|
||||||
fn wrong_path_cert() -> Result<(), Error> {
|
fn wrong_path_cert() -> Result<(), Error> {
|
||||||
|
let port = port().to_string();
|
||||||
Command::cargo_bin("dufs")?
|
Command::cargo_bin("dufs")?
|
||||||
.args(["--tls-cert", "wrong", "--tls-key", "tests/data/key.pem"])
|
.args([
|
||||||
|
"--tls-cert",
|
||||||
|
"wrong",
|
||||||
|
"--tls-key",
|
||||||
|
"tests/data/key.pem",
|
||||||
|
"--port",
|
||||||
|
&port,
|
||||||
|
])
|
||||||
.assert()
|
.assert()
|
||||||
.failure()
|
.failure()
|
||||||
.stderr(contains("Failed to access `wrong`"));
|
.stderr(contains("Failed to access `wrong`"));
|
||||||
@@ -45,8 +55,16 @@ fn wrong_path_cert() -> Result<(), Error> {
|
|||||||
/// Wrong paths for key throws errors.
|
/// Wrong paths for key throws errors.
|
||||||
#[rstest]
|
#[rstest]
|
||||||
fn wrong_path_key() -> Result<(), Error> {
|
fn wrong_path_key() -> Result<(), Error> {
|
||||||
|
let port = port().to_string();
|
||||||
Command::cargo_bin("dufs")?
|
Command::cargo_bin("dufs")?
|
||||||
.args(["--tls-cert", "tests/data/cert.pem", "--tls-key", "wrong"])
|
.args([
|
||||||
|
"--tls-cert",
|
||||||
|
"tests/data/cert.pem",
|
||||||
|
"--tls-key",
|
||||||
|
"wrong",
|
||||||
|
"--port",
|
||||||
|
&port,
|
||||||
|
])
|
||||||
.assert()
|
.assert()
|
||||||
.failure()
|
.failure()
|
||||||
.stderr(contains("Failed to access `wrong`"));
|
.stderr(contains("Failed to access `wrong`"));
|
||||||
|
|||||||
@@ -20,7 +20,7 @@ macro_rules! assert_resp_paths {
|
|||||||
#[macro_export]
|
#[macro_export]
|
||||||
macro_rules! fetch {
|
macro_rules! fetch {
|
||||||
($method:literal, $url:expr) => {
|
($method:literal, $url:expr) => {
|
||||||
reqwest::blocking::Client::new().request(hyper::Method::from_bytes($method)?, $url)
|
reqwest::blocking::Client::new().request(reqwest::Method::from_bytes($method)?, $url)
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -59,7 +59,8 @@ pub fn encode_uri(v: &str) -> String {
|
|||||||
parts.join("/")
|
parts.join("/")
|
||||||
}
|
}
|
||||||
|
|
||||||
fn retrive_json(content: &str) -> Option<Value> {
|
#[allow(dead_code)]
|
||||||
|
pub fn retrive_json(content: &str) -> Option<Value> {
|
||||||
let lines: Vec<&str> = content.lines().collect();
|
let lines: Vec<&str> = content.lines().collect();
|
||||||
let line = lines.iter().find(|v| v.contains("DATA ="))?;
|
let line = lines.iter().find(|v| v.contains("DATA ="))?;
|
||||||
let line_col = line.find("DATA =").unwrap() + 6;
|
let line_col = line.find("DATA =").unwrap() + 6;
|
||||||
|
|||||||
Reference in New Issue
Block a user