mirror of
https://github.com/sigoden/dufs.git
synced 2026-04-09 09:09:03 +03:00
Compare commits
37 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
6510ae8be9 | ||
|
|
9545fb6e37 | ||
|
|
0fd0f11298 | ||
|
|
46aa8fcc02 | ||
|
|
09bb738866 | ||
|
|
3612ef10d1 | ||
|
|
7ac2039a36 | ||
|
|
7f83de765a | ||
|
|
9b3779b13a | ||
|
|
11a52f29c4 | ||
|
|
10204c723f | ||
|
|
204421643d | ||
|
|
d9706d75ef | ||
|
|
40df0bd2f9 | ||
|
|
a53411b4d6 | ||
|
|
609017b2f5 | ||
|
|
7dc0b0e218 | ||
|
|
6be36b8e51 | ||
|
|
8be545d3da | ||
|
|
4f3a8d275b | ||
|
|
9c412f4276 | ||
|
|
27c269d6a0 | ||
|
|
57b4a74279 | ||
|
|
1112b936b8 | ||
|
|
033d37c4d4 | ||
|
|
577eea5fa4 | ||
|
|
d22be95dda | ||
|
|
8d7c1fbf53 | ||
|
|
4622c48120 | ||
|
|
f8ea41638f | ||
|
|
2890b3929d | ||
|
|
f5c0aefd8e | ||
|
|
8a1e7674df | ||
|
|
3c6206849f | ||
|
|
652f836c23 | ||
|
|
fb5b50f059 | ||
|
|
e43554b795 |
10
.github/workflows/ci.yaml
vendored
10
.github/workflows/ci.yaml
vendored
@@ -29,16 +29,12 @@ jobs:
|
||||
RUSTFLAGS: --deny warnings
|
||||
|
||||
steps:
|
||||
- uses: actions/checkout@v2
|
||||
- uses: actions/checkout@v3
|
||||
|
||||
- name: Install Rust Toolchain Components
|
||||
uses: actions-rs/toolchain@v1
|
||||
with:
|
||||
components: clippy, rustfmt
|
||||
override: true
|
||||
toolchain: stable
|
||||
uses: dtolnay/rust-toolchain@stable
|
||||
|
||||
- uses: Swatinem/rust-cache@v1
|
||||
- uses: Swatinem/rust-cache@v2
|
||||
|
||||
- name: Test
|
||||
run: cargo test --all
|
||||
|
||||
78
.github/workflows/release.yaml
vendored
78
.github/workflows/release.yaml
vendored
@@ -71,30 +71,40 @@ jobs:
|
||||
use-cross: true
|
||||
cargo-flags: "--no-default-features"
|
||||
runs-on: ${{matrix.os}}
|
||||
env:
|
||||
BUILD_CMD: cargo
|
||||
|
||||
steps:
|
||||
- uses: actions/checkout@v2
|
||||
- uses: actions/checkout@v3
|
||||
|
||||
- name: Check Tag
|
||||
id: check-tag
|
||||
shell: bash
|
||||
run: |
|
||||
tag=${GITHUB_REF##*/}
|
||||
echo "::set-output name=version::$tag"
|
||||
if [[ "$tag" =~ [0-9]+.[0-9]+.[0-9]+$ ]]; then
|
||||
echo "::set-output name=rc::false"
|
||||
ver=${GITHUB_REF##*/}
|
||||
echo "version=$ver" >> $GITHUB_OUTPUT
|
||||
if [[ "$ver" =~ [0-9]+.[0-9]+.[0-9]+$ ]]; then
|
||||
echo "rc=false" >> $GITHUB_OUTPUT
|
||||
else
|
||||
echo "::set-output name=rc::true"
|
||||
echo "rc=true" >> $GITHUB_OUTPUT
|
||||
fi
|
||||
|
||||
|
||||
- name: Install Rust Toolchain Components
|
||||
uses: actions-rs/toolchain@v1
|
||||
uses: dtolnay/rust-toolchain@stable
|
||||
with:
|
||||
override: true
|
||||
target: ${{ matrix.target }}
|
||||
toolchain: stable
|
||||
profile: minimal # minimal component installation (ie, no documentation)
|
||||
targets: ${{ matrix.target }}
|
||||
|
||||
- name: Install cross
|
||||
if: matrix.use-cross
|
||||
uses: taiki-e/install-action@v2
|
||||
with:
|
||||
tool: cross
|
||||
|
||||
- name: Overwrite build command env variable
|
||||
if: matrix.use-cross
|
||||
shell: bash
|
||||
run: echo "BUILD_CMD=cross" >> $GITHUB_ENV
|
||||
|
||||
- name: Show Version Information (Rust, cargo, GCC)
|
||||
shell: bash
|
||||
@@ -107,11 +117,8 @@ jobs:
|
||||
rustc -V
|
||||
|
||||
- name: Build
|
||||
uses: actions-rs/cargo@v1
|
||||
with:
|
||||
use-cross: ${{ matrix.use-cross }}
|
||||
command: build
|
||||
args: --locked --release --target=${{ matrix.target }} ${{ matrix.cargo-flags }}
|
||||
shell: bash
|
||||
run: $BUILD_CMD build --locked --release --target=${{ matrix.target }} ${{ matrix.cargo-flags }}
|
||||
|
||||
- name: Build Archive
|
||||
shell: bash
|
||||
@@ -123,8 +130,7 @@ jobs:
|
||||
set -euxo pipefail
|
||||
|
||||
bin=${GITHUB_REPOSITORY##*/}
|
||||
src=`pwd`
|
||||
dist=$src/dist
|
||||
dist_dir=`pwd`/dist
|
||||
name=$bin-$version-$target
|
||||
executable=target/$target/release/$bin
|
||||
|
||||
@@ -132,22 +138,22 @@ jobs:
|
||||
executable=$executable.exe
|
||||
fi
|
||||
|
||||
mkdir $dist
|
||||
cp $executable $dist
|
||||
cd $dist
|
||||
mkdir $dist_dir
|
||||
cp $executable $dist_dir
|
||||
cd $dist_dir
|
||||
|
||||
if [[ "$RUNNER_OS" == "Windows" ]]; then
|
||||
archive=$dist/$name.zip
|
||||
archive=$dist_dir/$name.zip
|
||||
7z a $archive *
|
||||
echo "::set-output name=archive::`pwd -W`/$name.zip"
|
||||
echo "archive=dist/$name.zip" >> $GITHUB_OUTPUT
|
||||
else
|
||||
archive=$dist/$name.tar.gz
|
||||
tar czf $archive *
|
||||
echo "::set-output name=archive::$archive"
|
||||
archive=$dist_dir/$name.tar.gz
|
||||
tar -czf $archive *
|
||||
echo "archive=dist/$name.tar.gz" >> $GITHUB_OUTPUT
|
||||
fi
|
||||
|
||||
- name: Publish Archive
|
||||
uses: softprops/action-gh-release@v0.1.5
|
||||
uses: softprops/action-gh-release@v1
|
||||
if: ${{ startsWith(github.ref, 'refs/tags/') }}
|
||||
with:
|
||||
draft: false
|
||||
@@ -163,16 +169,16 @@ jobs:
|
||||
needs: release
|
||||
steps:
|
||||
- name: Set up QEMU
|
||||
uses: docker/setup-qemu-action@v1
|
||||
uses: docker/setup-qemu-action@v2
|
||||
- name: Set up Docker Buildx
|
||||
uses: docker/setup-buildx-action@v1
|
||||
uses: docker/setup-buildx-action@v2
|
||||
- name: Login to DockerHub
|
||||
uses: docker/login-action@v1
|
||||
uses: docker/login-action@v2
|
||||
with:
|
||||
username: ${{ secrets.DOCKERHUB_USERNAME }}
|
||||
password: ${{ secrets.DOCKERHUB_TOKEN }}
|
||||
- name: Build and push
|
||||
uses: docker/build-push-action@v2
|
||||
uses: docker/build-push-action@v4
|
||||
with:
|
||||
build-args: |
|
||||
REPO=${{ github.repository }}
|
||||
@@ -191,13 +197,11 @@ jobs:
|
||||
runs-on: ubuntu-latest
|
||||
needs: release
|
||||
steps:
|
||||
- uses: actions/checkout@v2
|
||||
- uses: actions-rs/toolchain@v1
|
||||
with:
|
||||
profile: minimal
|
||||
toolchain: stable
|
||||
- name: Publish
|
||||
- uses: actions/checkout@v3
|
||||
|
||||
- uses: dtolnay/rust-toolchain@stable
|
||||
|
||||
- name: Publish
|
||||
env:
|
||||
CARGO_REGISTRY_TOKEN: ${{ secrets.CRATES_IO_API_TOKEN }}
|
||||
run: cargo publish
|
||||
58
CHANGELOG.md
58
CHANGELOG.md
@@ -2,6 +2,64 @@
|
||||
|
||||
All notable changes to this project will be documented in this file.
|
||||
|
||||
## [0.36.0] - 2023-08-24
|
||||
|
||||
### Bug Fixes
|
||||
|
||||
- Ui readonly if no write perm ([#258](https://github.com/sigoden/dufs/issues/258))
|
||||
|
||||
### Testing
|
||||
|
||||
- Remove dependency on native tls ([#255](https://github.com/sigoden/dufs/issues/255))
|
||||
|
||||
## [0.35.0] - 2023-08-14
|
||||
|
||||
### Bug Fixes
|
||||
|
||||
- Search should ignore entry path ([#235](https://github.com/sigoden/dufs/issues/235))
|
||||
- Typo __ASSERTS_PREFIX__ ([#252](https://github.com/sigoden/dufs/issues/252))
|
||||
|
||||
### Features
|
||||
|
||||
- Sort by type first, then sort by name/mtime/size ([#241](https://github.com/sigoden/dufs/issues/241))
|
||||
|
||||
## [0.34.2] - 2023-06-05
|
||||
|
||||
### Bug Fixes
|
||||
|
||||
- Ui refresh page after login ([#230](https://github.com/sigoden/dufs/issues/230))
|
||||
- Webdav only see public folder even logging in ([#231](https://github.com/sigoden/dufs/issues/231))
|
||||
|
||||
## [0.34.1] - 2023-06-02
|
||||
|
||||
### Bug Fixes
|
||||
|
||||
- Auth logic ([#224](https://github.com/sigoden/dufs/issues/224))
|
||||
- Allow all cors headers and methods ([#225](https://github.com/sigoden/dufs/issues/225))
|
||||
|
||||
### Refactor
|
||||
|
||||
- Ui checkAuth ([#226](https://github.com/sigoden/dufs/issues/226))
|
||||
|
||||
## [0.34.0] - 2023-06-01
|
||||
|
||||
### Bug Fixes
|
||||
|
||||
- URL-encoded filename when downloading in safari ([#203](https://github.com/sigoden/dufs/issues/203))
|
||||
- Ui path table show move action ([#219](https://github.com/sigoden/dufs/issues/219))
|
||||
- Ui set default max uploading to 1 ([#220](https://github.com/sigoden/dufs/issues/220))
|
||||
|
||||
### Features
|
||||
|
||||
- Webui editing support multiple encodings ([#197](https://github.com/sigoden/dufs/issues/197))
|
||||
- Add timestamp metadata to generated zip file ([#204](https://github.com/sigoden/dufs/issues/204))
|
||||
- Show precise file size with decimal ([#210](https://github.com/sigoden/dufs/issues/210))
|
||||
- [**breaking**] New auth ([#218](https://github.com/sigoden/dufs/issues/218))
|
||||
|
||||
### Refactor
|
||||
|
||||
- Cli positional rename root => SERVE_PATH([#215](https://github.com/sigoden/dufs/issues/215))
|
||||
|
||||
## [0.33.0] - 2023-03-17
|
||||
|
||||
### Bug Fixes
|
||||
|
||||
1192
Cargo.lock
generated
1192
Cargo.lock
generated
File diff suppressed because it is too large
Load Diff
30
Cargo.toml
30
Cargo.toml
@@ -1,6 +1,6 @@
|
||||
[package]
|
||||
name = "dufs"
|
||||
version = "0.33.0"
|
||||
version = "0.36.0"
|
||||
edition = "2021"
|
||||
authors = ["sigoden <sigoden@gmail.com>"]
|
||||
description = "Dufs is a distinctive utility file server"
|
||||
@@ -13,37 +13,38 @@ keywords = ["static", "file", "server", "webdav", "cli"]
|
||||
[dependencies]
|
||||
clap = { version = "4", features = ["wrap_help", "env"] }
|
||||
clap_complete = "4"
|
||||
chrono = "0.4"
|
||||
chrono = { version = "0.4", default-features = false, features = ["clock"] }
|
||||
tokio = { version = "1", features = ["rt-multi-thread", "macros", "fs", "io-util", "signal"]}
|
||||
tokio-util = { version = "0.7", features = ["io-util"] }
|
||||
tokio-util = { version = "0.7", features = ["io-util", "compat"] }
|
||||
hyper = { version = "0.14", features = ["http1", "server", "tcp", "stream"] }
|
||||
percent-encoding = "2.1"
|
||||
percent-encoding = "2.3"
|
||||
serde = { version = "1", features = ["derive"] }
|
||||
serde_json = "1"
|
||||
futures = "0.3"
|
||||
base64 = "0.21"
|
||||
async_zip = { version = "0.0.12", default-features = false, features = ["deflate"] }
|
||||
async_zip = { version = "0.0.15", default-features = false, features = ["deflate", "chrono", "tokio"] }
|
||||
headers = "0.3"
|
||||
mime_guess = "2.0"
|
||||
if-addrs = "0.10.1"
|
||||
rustls = { version = "0.20", default-features = false, features = ["tls12"], optional = true }
|
||||
rustls = { version = "0.21", default-features = false, features = ["tls12"], optional = true }
|
||||
rustls-pemfile = { version = "1", optional = true }
|
||||
tokio-rustls = { version = "0.23", optional = true }
|
||||
tokio-rustls = { version = "0.24", optional = true }
|
||||
md5 = "0.7"
|
||||
lazy_static = "1.4"
|
||||
uuid = { version = "1.1", features = ["v4", "fast-rng"] }
|
||||
uuid = { version = "1.4", features = ["v4", "fast-rng"] }
|
||||
urlencoding = "2.1"
|
||||
xml-rs = "0.8"
|
||||
log = "0.4"
|
||||
socket2 = "0.5"
|
||||
async-stream = "0.3"
|
||||
walkdir = "2.3"
|
||||
form_urlencoded = "1.0"
|
||||
form_urlencoded = "1.2"
|
||||
alphanumeric-sort = "1.4"
|
||||
content_inspector = "0.2"
|
||||
anyhow = "1.0"
|
||||
chardetng = "0.1"
|
||||
glob = "0.3.1"
|
||||
indexmap = "2.0"
|
||||
|
||||
[features]
|
||||
default = ["tls"]
|
||||
@@ -54,14 +55,15 @@ assert_cmd = "2"
|
||||
reqwest = { version = "0.11", features = ["blocking", "multipart", "rustls-tls"], default-features = false }
|
||||
assert_fs = "1"
|
||||
port_check = "0.1"
|
||||
rstest = "0.16"
|
||||
rstest = "0.18"
|
||||
regex = "1"
|
||||
url = "2"
|
||||
diqwest = { version = "1", features = ["blocking"] }
|
||||
diqwest = { version = "1", features = ["blocking", "rustls-tls"], default-features = false }
|
||||
predicates = "3"
|
||||
indexmap = "1.9"
|
||||
|
||||
[profile.release]
|
||||
opt-level = 3
|
||||
lto = true
|
||||
strip = true
|
||||
opt-level = "z"
|
||||
codegen-units = 1
|
||||
panic = "abort"
|
||||
strip = "symbols"
|
||||
|
||||
73
README.md
73
README.md
@@ -14,7 +14,7 @@ Dufs is a distinctive utility file server that supports static serving, uploadin
|
||||
- Upload files and folders (Drag & Drop)
|
||||
- Create/Edit/Search files
|
||||
- Partial responses (Parallel/Resume download)
|
||||
- Path level access control
|
||||
- Access control
|
||||
- Support https
|
||||
- Support webdav
|
||||
- Easy to use with curl
|
||||
@@ -48,17 +48,17 @@ Download from [Github Releases](https://github.com/sigoden/dufs/releases), unzip
|
||||
```
|
||||
Dufs is a distinctive utility file server - https://github.com/sigoden/dufs
|
||||
|
||||
Usage: dufs [OPTIONS] [root]
|
||||
Usage: dufs [OPTIONS] [serve_path]
|
||||
|
||||
Arguments:
|
||||
[root] Specific path to serve [default: .]
|
||||
[serve_path] Specific path to serve [default: .]
|
||||
|
||||
Options:
|
||||
-b, --bind <addrs> Specify bind address or unix socket
|
||||
-p, --port <port> Specify port to listen on [default: 5000]
|
||||
--path-prefix <path> Specify a path prefix
|
||||
--hidden <value> Hide paths from directory listings, separated by `,`
|
||||
-a, --auth <rules> Add auth for path
|
||||
-a, --auth <rules> Add auth role
|
||||
--auth-method <value> Select auth method [default: digest] [possible values: basic, digest]
|
||||
-A, --allow-all Allow all operations
|
||||
--allow-upload Allow upload files/folders
|
||||
@@ -75,8 +75,8 @@ Options:
|
||||
--tls-key <path> Path to the SSL/TLS certificate's private key
|
||||
--log-format <format> Customize http log format
|
||||
--completions <shell> Print shell completion script for <shell> [possible values: bash, elvish, fish, powershell, zsh]
|
||||
-h, --help Print help information
|
||||
-V, --version Print version information
|
||||
-h, --help Print help
|
||||
-V, --version Print version
|
||||
```
|
||||
|
||||
## Examples
|
||||
@@ -126,7 +126,7 @@ dufs --render-index
|
||||
Require username/password
|
||||
|
||||
```
|
||||
dufs -a /@admin:123
|
||||
dufs -a admin:123@/:rw
|
||||
```
|
||||
|
||||
Listen on specific host:ip
|
||||
@@ -203,41 +203,46 @@ curl --user user:pass http://192.168.8.10:5000/file # basic auth
|
||||
|
||||
### Access Control
|
||||
|
||||
Dufs supports path level access control. You can control who can do what on which path with `--auth`/`-a`.
|
||||
Dufs supports account based access control. You can control who can do what on which path with `--auth`/`-a`.
|
||||
|
||||
```
|
||||
dufs -a <path>@<readwrite>
|
||||
dufs -a <path>@<readwrite>@<readonly>
|
||||
dufs -a <path>@<readwrite>@*
|
||||
dufs -a [user:pass]@path[:rw][,path[:rw]...][|...]
|
||||
```
|
||||
|
||||
- `<path>`: Protected url path
|
||||
- `<readwrite>`: Account with readwrite permissions. If dufs is run with `dufs --allow-all`, the permissions are upload/delete/search/view/download. If dufs is run with `dufs --allow-upload`, the permissions are upload/view/download.
|
||||
- `<readonly>`: Account with readonly permissions. The permissions are search/view/download if dufs allow search, otherwise view/download..
|
||||
1. Multiple rules are separated by "|"
|
||||
2. User and pass are the account name and password, if omitted, it is an anonymous user
|
||||
3. One rule can set multiple paths, separated by ","
|
||||
4. Add `:rw` after the path to indicate that the path has read and write permissions, otherwise the path has readonly permissions.
|
||||
|
||||
```
|
||||
dufs -A -a /@admin:admin
|
||||
dufs -A -a admin:admin@/:rw
|
||||
```
|
||||
`admin` has all permissions for all paths.
|
||||
|
||||
```
|
||||
dufs -A -a /@admin:admin@guest:guest
|
||||
dufs -A -a admin:admin@/:rw -a guest:guest@/
|
||||
```
|
||||
`guest` has readonly permissions for all paths.
|
||||
|
||||
```
|
||||
dufs -A -a /@admin:admin@*
|
||||
dufs -A -a admin:admin@/:rw -a @/
|
||||
```
|
||||
All paths is public, everyone can view/download it.
|
||||
|
||||
```
|
||||
dufs -A -a /@admin:admin -a /user1@user1:pass1 -a /user2@pass2:user2
|
||||
dufs -A -a admin:admin@/:rw -a user1:pass1@/user1:rw -a user2:pass2@/user2
|
||||
dufs -A -a "admin:admin@/:rw|user1:pass1@/user1:rw|user2:pass2@/user2"
|
||||
```
|
||||
`user1` has all permissions for `/user1*` path.
|
||||
`user2` has all permissions for `/user2*` path.
|
||||
`user1` has all permissions for `/user1/*` path.
|
||||
`user2` has all permissions for `/user2/*` path.
|
||||
|
||||
```
|
||||
dufs -a /@admin:admin
|
||||
dufs -A -a user:pass@/dir1:rw,/dir2:rw,dir3
|
||||
```
|
||||
`user` has all permissions for `/dir1/*` and `/dir2/*`, has readonly permissions for `/dir3/`.
|
||||
|
||||
```
|
||||
dufs -a admin:admin@/
|
||||
```
|
||||
Since dufs only allows viewing/downloading, `admin` can only view/download files.
|
||||
|
||||
@@ -302,7 +307,29 @@ dufs --log-format '$remote_addr $remote_user "$request" $status' -a /@admin:admi
|
||||
|
||||
All options can be set using environment variables prefixed with `DUFS_`.
|
||||
|
||||
`dufs --port 8080 --allow-all` is equal to `DUFS_PORT=8080 DUFS_ALLOW_ALL=true dufs`.
|
||||
```
|
||||
[SERVE_PATH] DUFS_SERVE_PATH=/dir
|
||||
-b, --bind <addrs> DUFS_BIND=0.0.0.0
|
||||
-p, --port <port> DUFS_PORT=5000
|
||||
--path-prefix <path> DUFS_PATH_PREFIX=/path
|
||||
--hidden <value> DUFS_HIDDEN=*.log
|
||||
-a, --auth <rules> DUFS_AUTH="admin:admin@/:rw|@/"
|
||||
--auth-method <value> DUFS_AUTH_METHOD=basic
|
||||
-A, --allow-all DUFS_ALLOW_ALL=true
|
||||
--allow-upload DUFS_ALLOW_UPLOAD=true
|
||||
--allow-delete DUFS_ALLOW_DELETE=true
|
||||
--allow-search DUFS_ALLOW_SEARCH=true
|
||||
--allow-symlink DUFS_ALLOW_SYMLINK=true
|
||||
--allow-archive DUFS_ALLOW_ARCHIVE=true
|
||||
--enable-cors DUFS_ENABLE_CORS=true
|
||||
--render-index DUFS_RENDER_INDEX=true
|
||||
--render-try-index DUFS_RENDER_TRY_INDEX=true
|
||||
--render-spa DUFS_RENDER_SPA=true
|
||||
--assets <path> DUFS_ASSETS=/assets
|
||||
--tls-cert <path> DUFS_TLS_CERT=cert.pem
|
||||
--tls-key <path> DUFS_TLS_KEY=key.pem
|
||||
--log-format <format> DUFS_LOG_FORMAT=""
|
||||
```
|
||||
|
||||
### Customize UI
|
||||
|
||||
@@ -317,7 +344,7 @@ Your assets folder must contains a `index.html` file.
|
||||
`index.html` can use the following placeholder variables to retrieve internal data.
|
||||
|
||||
- `__INDEX_DATA__`: directory listing data
|
||||
- `__ASSERTS_PREFIX__`: assets url prefix
|
||||
- `__ASSETS_PREFIX__`: assets url prefix
|
||||
|
||||
</details>
|
||||
|
||||
|
||||
@@ -148,7 +148,7 @@ body {
|
||||
}
|
||||
|
||||
.paths-table .cell-actions {
|
||||
width: 75px;
|
||||
width: 90px;
|
||||
display: flex;
|
||||
padding-left: 0.6em;
|
||||
}
|
||||
|
||||
@@ -4,12 +4,12 @@
|
||||
<head>
|
||||
<meta charset="utf-8" />
|
||||
<meta name="viewport" content="width=device-width" />
|
||||
<link rel="icon" type="image/x-icon" href="__ASSERTS_PREFIX__favicon.ico">
|
||||
<link rel="stylesheet" href="__ASSERTS_PREFIX__index.css">
|
||||
<link rel="icon" type="image/x-icon" href="__ASSETS_PREFIX__favicon.ico">
|
||||
<link rel="stylesheet" href="__ASSETS_PREFIX__index.css">
|
||||
<script>
|
||||
DATA = __INDEX_DATA__
|
||||
</script>
|
||||
<script src="__ASSERTS_PREFIX__index.js"></script>
|
||||
<script src="__ASSETS_PREFIX__index.js"></script>
|
||||
</head>
|
||||
|
||||
<body>
|
||||
|
||||
@@ -22,6 +22,8 @@
|
||||
* @property {string} editable
|
||||
*/
|
||||
|
||||
var DUFS_MAX_UPLOADINGS = 1;
|
||||
|
||||
/**
|
||||
* @type {DATA} DATA
|
||||
*/
|
||||
@@ -103,7 +105,7 @@ function ready() {
|
||||
setupIndexPage();
|
||||
|
||||
} else if (DATA.kind == "Edit") {
|
||||
document.title = `Edit of ${DATA.href} - Dufs`;
|
||||
document.title = `Edit ${DATA.href} - Dufs`;
|
||||
document.querySelector(".editor-page").classList.remove("hidden");;
|
||||
|
||||
setupEditPage();
|
||||
@@ -152,7 +154,6 @@ class Uploader {
|
||||
}
|
||||
|
||||
ajax() {
|
||||
Uploader.runnings += 1;
|
||||
const url = newUrl(this.name);
|
||||
this.lastUptime = Date.now();
|
||||
const ajax = new XMLHttpRequest();
|
||||
@@ -187,13 +188,13 @@ class Uploader {
|
||||
|
||||
complete() {
|
||||
this.$uploadStatus.innerHTML = `✓`;
|
||||
Uploader.runnings -= 1;
|
||||
Uploader.runnings--;
|
||||
Uploader.runQueue();
|
||||
}
|
||||
|
||||
fail() {
|
||||
this.$uploadStatus.innerHTML = `✗`;
|
||||
Uploader.runnings -= 1;
|
||||
Uploader.runnings--;
|
||||
Uploader.runQueue();
|
||||
}
|
||||
}
|
||||
@@ -211,12 +212,17 @@ Uploader.queues = [];
|
||||
|
||||
|
||||
Uploader.runQueue = async () => {
|
||||
if (Uploader.runnings > 2) return;
|
||||
if (Uploader.runnings >= DUFS_MAX_UPLOADINGS) return;
|
||||
if (Uploader.queues.length == 0) return;
|
||||
Uploader.runnings++;
|
||||
let uploader = Uploader.queues.shift();
|
||||
if (!uploader) return;
|
||||
if (!Uploader.auth) {
|
||||
Uploader.auth = true;
|
||||
await login();
|
||||
try {
|
||||
await checkAuth()
|
||||
} catch {
|
||||
Uploader.auth = false;
|
||||
}
|
||||
}
|
||||
uploader.ajax();
|
||||
}
|
||||
@@ -262,7 +268,7 @@ function setupIndexPage() {
|
||||
if (DATA.allow_archive) {
|
||||
const $download = document.querySelector(".download");
|
||||
$download.href = baseUrl() + "?zip";
|
||||
$download.title = "Download folder as div .zip file";
|
||||
$download.title = "Download folder as a .zip file";
|
||||
$download.classList.remove("hidden");
|
||||
}
|
||||
|
||||
@@ -375,9 +381,8 @@ function addPath(file, index) {
|
||||
}
|
||||
if (DATA.allow_delete) {
|
||||
if (DATA.allow_upload) {
|
||||
if (isDir) {
|
||||
actionMove = `<div onclick="movePath(${index})" class="action-btn" id="moveBtn${index}" title="Move to new path">${ICONS.move}</div>`;
|
||||
} else {
|
||||
if (!isDir) {
|
||||
actionEdit = `<a class="action-btn" title="Edit file" target="_blank" href="${url}?edit">${ICONS.edit}</a>`;
|
||||
}
|
||||
}
|
||||
@@ -388,8 +393,8 @@ function addPath(file, index) {
|
||||
<td class="cell-actions">
|
||||
${actionDownload}
|
||||
${actionMove}
|
||||
${actionEdit}
|
||||
${actionDelete}
|
||||
${actionEdit}
|
||||
</td>`
|
||||
|
||||
$pathsTableBody.insertAdjacentHTML("beforeend", `
|
||||
@@ -437,7 +442,14 @@ function setupAuth() {
|
||||
} else {
|
||||
const $loginBtn = document.querySelector(".login-btn");
|
||||
$loginBtn.classList.remove("hidden");
|
||||
$loginBtn.addEventListener("click", () => login(true));
|
||||
$loginBtn.addEventListener("click", async () => {
|
||||
try {
|
||||
await checkAuth()
|
||||
location.reload();
|
||||
} catch (err) {
|
||||
alert(err.message);
|
||||
}
|
||||
});
|
||||
}
|
||||
}
|
||||
|
||||
@@ -522,7 +534,7 @@ async function setupEditPage() {
|
||||
$notEditable.insertAdjacentHTML("afterend", `<iframe src="${url}" sandbox width="100%" height="${window.innerHeight - 100}px"></iframe>`)
|
||||
} else {
|
||||
$notEditable.classList.remove("hidden");
|
||||
$notEditable.textContent = "Cannot edit because it is too large or binary.";
|
||||
$notEditable.textContent = "Cannot edit because file is too large or binary.";
|
||||
}
|
||||
return;
|
||||
}
|
||||
@@ -535,8 +547,15 @@ async function setupEditPage() {
|
||||
try {
|
||||
const res = await fetch(baseUrl());
|
||||
await assertResOK(res);
|
||||
const text = await res.text();
|
||||
$editor.value = text;
|
||||
const encoding = getEncoding(res.headers.get("content-type"));
|
||||
if (encoding === "utf-8") {
|
||||
$editor.value = await res.text();
|
||||
} else {
|
||||
const bytes = await res.arrayBuffer();
|
||||
const dataView = new DataView(bytes)
|
||||
const decoder = new TextDecoder(encoding)
|
||||
$editor.value = decoder.decode(dataView);
|
||||
}
|
||||
} catch (err) {
|
||||
alert(`Failed get file, ${err.message}`);
|
||||
}
|
||||
@@ -564,7 +583,7 @@ async function deletePath(index) {
|
||||
async function doDeletePath(name, url, cb) {
|
||||
if (!confirm(`Delete \`${name}\`?`)) return;
|
||||
try {
|
||||
await login();
|
||||
await checkAuth();
|
||||
const res = await fetch(url, {
|
||||
method: "DELETE",
|
||||
});
|
||||
@@ -604,12 +623,12 @@ async function doMovePath(fileUrl) {
|
||||
const newFileUrl = fileUrlObj.origin + prefix + newPath.split("/").map(encodeURIComponent).join("/");
|
||||
|
||||
try {
|
||||
await login();
|
||||
await checkAuth();
|
||||
const res1 = await fetch(newFileUrl, {
|
||||
method: "HEAD",
|
||||
});
|
||||
if (res1.status === 200) {
|
||||
if (!confirm("Override exsis file?")) {
|
||||
if (!confirm("Override existing file?")) {
|
||||
return;
|
||||
}
|
||||
}
|
||||
@@ -636,14 +655,14 @@ async function saveChange() {
|
||||
method: "PUT",
|
||||
body: $editor.value,
|
||||
});
|
||||
location.reload();
|
||||
} catch (err) {
|
||||
alert(`Failed to save file, ${err.message}`);
|
||||
}
|
||||
}
|
||||
|
||||
async function login(alert = false) {
|
||||
async function checkAuth() {
|
||||
if (!DATA.auth) return;
|
||||
try {
|
||||
const res = await fetch(baseUrl(), {
|
||||
method: "WRITEABLE",
|
||||
});
|
||||
@@ -651,14 +670,6 @@ async function login(alert = false) {
|
||||
document.querySelector(".login-btn").classList.add("hidden");
|
||||
$userBtn.classList.remove("hidden");
|
||||
$userBtn.title = "";
|
||||
} catch (err) {
|
||||
let message = `Cannot login, ${err.message}`;
|
||||
if (alert) {
|
||||
alert(message);
|
||||
} else {
|
||||
throw new Error(message);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
@@ -668,7 +679,7 @@ async function login(alert = false) {
|
||||
async function createFolder(name) {
|
||||
const url = newUrl(name);
|
||||
try {
|
||||
await login();
|
||||
await checkAuth();
|
||||
const res = await fetch(url, {
|
||||
method: "MKCOL",
|
||||
});
|
||||
@@ -682,7 +693,7 @@ async function createFolder(name) {
|
||||
async function createFile(name) {
|
||||
const url = newUrl(name);
|
||||
try {
|
||||
await login();
|
||||
await checkAuth();
|
||||
const res = await fetch(url, {
|
||||
method: "PUT",
|
||||
body: "",
|
||||
@@ -766,7 +777,11 @@ function formatSize(size) {
|
||||
const sizes = ['B', 'KB', 'MB', 'GB', 'TB'];
|
||||
if (size == 0) return [0, "B"];
|
||||
const i = parseInt(Math.floor(Math.log(size) / Math.log(1024)));
|
||||
return [Math.round(size / Math.pow(1024, i), 2), sizes[i]];
|
||||
ratio = 1
|
||||
if (i >= 3) {
|
||||
ratio = 100
|
||||
}
|
||||
return [Math.round(size * ratio / Math.pow(1024, i), 2) / ratio, sizes[i]];
|
||||
}
|
||||
|
||||
function formatDuration(seconds) {
|
||||
@@ -793,6 +808,17 @@ function encodedStr(rawStr) {
|
||||
|
||||
async function assertResOK(res) {
|
||||
if (!(res.status >= 200 && res.status < 300)) {
|
||||
throw new Error(await res.text())
|
||||
throw new Error(await res.text() || `Invalid status ${res.status}`);
|
||||
}
|
||||
}
|
||||
|
||||
function getEncoding(contentType) {
|
||||
const charset = contentType?.split(";")[1];
|
||||
if (/charset/i.test(charset)) {
|
||||
let encoding = charset.split("=")[1];
|
||||
if (encoding) {
|
||||
return encoding.toLowerCase()
|
||||
}
|
||||
}
|
||||
return 'utf-8'
|
||||
}
|
||||
|
||||
12
src/args.rs
12
src/args.rs
@@ -25,8 +25,8 @@ pub fn build_cli() -> Command {
|
||||
env!("CARGO_PKG_REPOSITORY")
|
||||
))
|
||||
.arg(
|
||||
Arg::new("root")
|
||||
.env("DUFS_ROOT")
|
||||
Arg::new("serve_path")
|
||||
.env("DUFS_SERVE_PATH")
|
||||
.hide_env(true)
|
||||
.default_value(".")
|
||||
.value_parser(value_parser!(PathBuf))
|
||||
@@ -76,9 +76,9 @@ pub fn build_cli() -> Command {
|
||||
.hide_env(true)
|
||||
.short('a')
|
||||
.long("auth")
|
||||
.help("Add auth for path")
|
||||
.help("Add auth role")
|
||||
.action(ArgAction::Append)
|
||||
.value_delimiter(',')
|
||||
.value_delimiter('|')
|
||||
.value_name("rules"),
|
||||
)
|
||||
.arg(
|
||||
@@ -264,7 +264,7 @@ impl Args {
|
||||
.map(|bind| bind.map(|v| v.as_str()).collect())
|
||||
.unwrap_or_else(|| vec!["0.0.0.0", "::"]);
|
||||
let addrs: Vec<BindAddr> = Args::parse_addrs(&addrs)?;
|
||||
let path = Args::parse_path(matches.get_one::<PathBuf>("root").unwrap())?;
|
||||
let path = Args::parse_path(matches.get_one::<PathBuf>("serve_path").unwrap())?;
|
||||
let path_is_file = path.metadata()?.is_file();
|
||||
let path_prefix = matches
|
||||
.get_one::<String>("path-prefix")
|
||||
@@ -288,7 +288,7 @@ impl Args {
|
||||
"basic" => AuthMethod::Basic,
|
||||
_ => AuthMethod::Digest,
|
||||
};
|
||||
let auth = AccessControl::new(&auth, &uri_prefix)?;
|
||||
let auth = AccessControl::new(&auth)?;
|
||||
let allow_upload = matches.get_flag("allow-all") || matches.get_flag("allow-upload");
|
||||
let allow_delete = matches.get_flag("allow-all") || matches.get_flag("allow-delete");
|
||||
let allow_search = matches.get_flag("allow-all") || matches.get_flag("allow-search");
|
||||
|
||||
403
src/auth.rs
403
src/auth.rs
@@ -2,12 +2,16 @@ use anyhow::{anyhow, bail, Result};
|
||||
use base64::{engine::general_purpose, Engine as _};
|
||||
use headers::HeaderValue;
|
||||
use hyper::Method;
|
||||
use indexmap::IndexMap;
|
||||
use lazy_static::lazy_static;
|
||||
use md5::Context;
|
||||
use std::collections::HashMap;
|
||||
use std::{
|
||||
collections::HashMap,
|
||||
path::{Path, PathBuf},
|
||||
};
|
||||
use uuid::Uuid;
|
||||
|
||||
use crate::utils::{encode_uri, unix_now};
|
||||
use crate::utils::unix_now;
|
||||
|
||||
const REALM: &str = "DUFS";
|
||||
const DIGEST_AUTH_TIMEOUT: u32 = 86400;
|
||||
@@ -21,57 +25,63 @@ lazy_static! {
|
||||
};
|
||||
}
|
||||
|
||||
#[derive(Debug)]
|
||||
#[derive(Debug, Default)]
|
||||
pub struct AccessControl {
|
||||
rules: HashMap<String, PathControl>,
|
||||
}
|
||||
|
||||
#[derive(Debug)]
|
||||
pub struct PathControl {
|
||||
readwrite: Account,
|
||||
readonly: Option<Account>,
|
||||
share: bool,
|
||||
users: IndexMap<String, (String, AccessPaths)>,
|
||||
anony: Option<AccessPaths>,
|
||||
}
|
||||
|
||||
impl AccessControl {
|
||||
pub fn new(raw_rules: &[&str], uri_prefix: &str) -> Result<Self> {
|
||||
let mut rules = HashMap::default();
|
||||
pub fn new(raw_rules: &[&str]) -> Result<Self> {
|
||||
if raw_rules.is_empty() {
|
||||
return Ok(Self { rules });
|
||||
}
|
||||
for rule in raw_rules {
|
||||
let parts: Vec<&str> = rule.split('@').collect();
|
||||
let create_err = || anyhow!("Invalid auth `{rule}`");
|
||||
match parts.as_slice() {
|
||||
[path, readwrite] => {
|
||||
let control = PathControl {
|
||||
readwrite: Account::new(readwrite).ok_or_else(create_err)?,
|
||||
readonly: None,
|
||||
share: false,
|
||||
};
|
||||
rules.insert(sanitize_path(path, uri_prefix), control);
|
||||
}
|
||||
[path, readwrite, readonly] => {
|
||||
let (readonly, share) = if *readonly == "*" {
|
||||
(None, true)
|
||||
} else {
|
||||
(Some(Account::new(readonly).ok_or_else(create_err)?), false)
|
||||
};
|
||||
let control = PathControl {
|
||||
readwrite: Account::new(readwrite).ok_or_else(create_err)?,
|
||||
readonly,
|
||||
share,
|
||||
};
|
||||
rules.insert(sanitize_path(path, uri_prefix), control);
|
||||
}
|
||||
_ => return Err(create_err()),
|
||||
}
|
||||
}
|
||||
Ok(Self { rules })
|
||||
return Ok(AccessControl {
|
||||
anony: Some(AccessPaths::new(AccessPerm::ReadWrite)),
|
||||
users: IndexMap::new(),
|
||||
});
|
||||
}
|
||||
|
||||
pub fn valid(&self) -> bool {
|
||||
!self.rules.is_empty()
|
||||
let create_err = |v: &str| anyhow!("Invalid auth `{v}`");
|
||||
let mut anony = None;
|
||||
let mut anony_paths = vec![];
|
||||
let mut users = IndexMap::new();
|
||||
for rule in raw_rules {
|
||||
let (user, list) = rule.split_once('@').ok_or_else(|| create_err(rule))?;
|
||||
if user.is_empty() && anony.is_some() {
|
||||
bail!("Invalid auth, duplicate anonymous rules");
|
||||
}
|
||||
let mut paths = AccessPaths::default();
|
||||
for value in list.trim_matches(',').split(',') {
|
||||
let (path, perm) = match value.split_once(':') {
|
||||
None => (value, AccessPerm::ReadOnly),
|
||||
Some((path, "rw")) => (path, AccessPerm::ReadWrite),
|
||||
_ => return Err(create_err(rule)),
|
||||
};
|
||||
if user.is_empty() {
|
||||
anony_paths.push((path, perm));
|
||||
}
|
||||
paths.add(path, perm);
|
||||
}
|
||||
if user.is_empty() {
|
||||
anony = Some(paths);
|
||||
} else if let Some((user, pass)) = user.split_once(':') {
|
||||
if user.is_empty() || pass.is_empty() {
|
||||
return Err(create_err(rule));
|
||||
}
|
||||
users.insert(user.to_string(), (pass.to_string(), paths));
|
||||
} else {
|
||||
return Err(create_err(rule));
|
||||
}
|
||||
}
|
||||
for (path, perm) in anony_paths {
|
||||
for (_, (_, paths)) in users.iter_mut() {
|
||||
paths.add(path, perm)
|
||||
}
|
||||
}
|
||||
Ok(Self { users, anony })
|
||||
}
|
||||
|
||||
pub fn exist(&self) -> bool {
|
||||
!self.users.is_empty()
|
||||
}
|
||||
|
||||
pub fn guard(
|
||||
@@ -80,81 +90,157 @@ impl AccessControl {
|
||||
method: &Method,
|
||||
authorization: Option<&HeaderValue>,
|
||||
auth_method: AuthMethod,
|
||||
) -> GuardType {
|
||||
if self.rules.is_empty() {
|
||||
return GuardType::ReadWrite;
|
||||
) -> (Option<String>, Option<AccessPaths>) {
|
||||
if let Some(authorization) = authorization {
|
||||
if let Some(user) = auth_method.get_user(authorization) {
|
||||
if let Some((pass, paths)) = self.users.get(&user) {
|
||||
if method == Method::OPTIONS {
|
||||
return (Some(user), Some(AccessPaths::new(AccessPerm::ReadOnly)));
|
||||
}
|
||||
if auth_method
|
||||
.check(authorization, method.as_str(), &user, pass)
|
||||
.is_some()
|
||||
{
|
||||
return (Some(user), paths.find(path, !is_readonly_method(method)));
|
||||
} else {
|
||||
return (None, None);
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
if method == Method::OPTIONS {
|
||||
return GuardType::ReadOnly;
|
||||
return (None, Some(AccessPaths::new(AccessPerm::ReadOnly)));
|
||||
}
|
||||
|
||||
let mut controls = vec![];
|
||||
for path in walk_path(path) {
|
||||
if let Some(control) = self.rules.get(path) {
|
||||
controls.push(control);
|
||||
if let Some(authorization) = authorization {
|
||||
let Account { user, pass } = &control.readwrite;
|
||||
if auth_method
|
||||
.validate(authorization, method.as_str(), user, pass)
|
||||
.is_some()
|
||||
{
|
||||
return GuardType::ReadWrite;
|
||||
if let Some(paths) = self.anony.as_ref() {
|
||||
return (None, paths.find(path, !is_readonly_method(method)));
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
if is_readonly_method(method) {
|
||||
for control in controls.into_iter() {
|
||||
if control.share {
|
||||
return GuardType::ReadOnly;
|
||||
}
|
||||
if let Some(authorization) = authorization {
|
||||
if let Some(Account { user, pass }) = &control.readonly {
|
||||
if auth_method
|
||||
.validate(authorization, method.as_str(), user, pass)
|
||||
.is_some()
|
||||
{
|
||||
return GuardType::ReadOnly;
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
GuardType::Reject
|
||||
|
||||
(None, None)
|
||||
}
|
||||
}
|
||||
|
||||
#[derive(Debug, Clone, Copy, PartialEq, Eq, PartialOrd, Ord)]
|
||||
pub enum GuardType {
|
||||
Reject,
|
||||
ReadWrite,
|
||||
ReadOnly,
|
||||
#[derive(Debug, Default, Clone, PartialEq, Eq)]
|
||||
pub struct AccessPaths {
|
||||
perm: AccessPerm,
|
||||
children: IndexMap<String, AccessPaths>,
|
||||
}
|
||||
|
||||
impl GuardType {
|
||||
pub fn is_reject(&self) -> bool {
|
||||
*self == GuardType::Reject
|
||||
impl AccessPaths {
|
||||
pub fn new(perm: AccessPerm) -> Self {
|
||||
Self {
|
||||
perm,
|
||||
..Default::default()
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
fn sanitize_path(path: &str, uri_prefix: &str) -> String {
|
||||
let new_path = match (uri_prefix, path) {
|
||||
("/", "/") => "/".into(),
|
||||
(_, "/") => uri_prefix.trim_end_matches('/').into(),
|
||||
_ => format!("{}{}", uri_prefix, path.trim_matches('/')),
|
||||
pub fn perm(&self) -> AccessPerm {
|
||||
self.perm
|
||||
}
|
||||
|
||||
fn set_perm(&mut self, perm: AccessPerm) {
|
||||
if self.perm < perm {
|
||||
self.perm = perm
|
||||
}
|
||||
}
|
||||
|
||||
pub fn add(&mut self, path: &str, perm: AccessPerm) {
|
||||
let path = path.trim_matches('/');
|
||||
if path.is_empty() {
|
||||
self.set_perm(perm);
|
||||
} else {
|
||||
let parts: Vec<&str> = path.split('/').collect();
|
||||
self.add_impl(&parts, perm);
|
||||
}
|
||||
}
|
||||
|
||||
fn add_impl(&mut self, parts: &[&str], perm: AccessPerm) {
|
||||
let parts_len = parts.len();
|
||||
if parts_len == 0 {
|
||||
self.set_perm(perm);
|
||||
return;
|
||||
}
|
||||
let child = self.children.entry(parts[0].to_string()).or_default();
|
||||
child.add_impl(&parts[1..], perm)
|
||||
}
|
||||
|
||||
pub fn find(&self, path: &str, writable: bool) -> Option<AccessPaths> {
|
||||
let parts: Vec<&str> = path
|
||||
.trim_matches('/')
|
||||
.split('/')
|
||||
.filter(|v| !v.is_empty())
|
||||
.collect();
|
||||
let target = self.find_impl(&parts, self.perm)?;
|
||||
if writable && !target.perm().readwrite() {
|
||||
return None;
|
||||
}
|
||||
Some(target)
|
||||
}
|
||||
|
||||
fn find_impl(&self, parts: &[&str], perm: AccessPerm) -> Option<AccessPaths> {
|
||||
let perm = self.perm.max(perm);
|
||||
if parts.is_empty() {
|
||||
if perm.indexonly() {
|
||||
return Some(self.clone());
|
||||
} else {
|
||||
return Some(AccessPaths::new(perm));
|
||||
}
|
||||
}
|
||||
let child = match self.children.get(parts[0]) {
|
||||
Some(v) => v,
|
||||
None => {
|
||||
if perm.indexonly() {
|
||||
return None;
|
||||
} else {
|
||||
return Some(AccessPaths::new(perm));
|
||||
}
|
||||
}
|
||||
};
|
||||
encode_uri(&new_path)
|
||||
child.find_impl(&parts[1..], perm)
|
||||
}
|
||||
|
||||
pub fn child_paths(&self) -> Vec<&String> {
|
||||
self.children.keys().collect()
|
||||
}
|
||||
|
||||
pub fn leaf_paths(&self, base: &Path) -> Vec<PathBuf> {
|
||||
if !self.perm().indexonly() {
|
||||
return vec![base.to_path_buf()];
|
||||
}
|
||||
let mut output = vec![];
|
||||
self.leaf_paths_impl(&mut output, base);
|
||||
output
|
||||
}
|
||||
|
||||
fn leaf_paths_impl(&self, output: &mut Vec<PathBuf>, base: &Path) {
|
||||
for (name, child) in self.children.iter() {
|
||||
let base = base.join(name);
|
||||
if child.perm().indexonly() {
|
||||
child.leaf_paths_impl(output, &base);
|
||||
} else {
|
||||
output.push(base)
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
fn walk_path(path: &str) -> impl Iterator<Item = &str> {
|
||||
let mut idx = 0;
|
||||
path.split('/').enumerate().map(move |(i, part)| {
|
||||
let end = if i == 0 { 1 } else { idx + part.len() + i };
|
||||
let value = &path[..end];
|
||||
idx += part.len();
|
||||
value
|
||||
})
|
||||
#[derive(Debug, Clone, Copy, PartialEq, Eq, PartialOrd, Ord, Default)]
|
||||
pub enum AccessPerm {
|
||||
#[default]
|
||||
IndexOnly,
|
||||
ReadOnly,
|
||||
ReadWrite,
|
||||
}
|
||||
|
||||
impl AccessPerm {
|
||||
pub fn readwrite(&self) -> bool {
|
||||
self == &AccessPerm::ReadWrite
|
||||
}
|
||||
|
||||
pub fn indexonly(&self) -> bool {
|
||||
self == &AccessPerm::IndexOnly
|
||||
}
|
||||
}
|
||||
|
||||
fn is_readonly_method(method: &Method) -> bool {
|
||||
@@ -164,29 +250,6 @@ fn is_readonly_method(method: &Method) -> bool {
|
||||
|| method.as_str() == "PROPFIND"
|
||||
}
|
||||
|
||||
#[derive(Debug, Clone)]
|
||||
struct Account {
|
||||
user: String,
|
||||
pass: String,
|
||||
}
|
||||
|
||||
impl Account {
|
||||
fn new(data: &str) -> Option<Self> {
|
||||
let p: Vec<&str> = data.trim().split(':').collect();
|
||||
if p.len() != 2 {
|
||||
return None;
|
||||
}
|
||||
let user = p[0];
|
||||
let pass = p[1];
|
||||
let mut h = Context::new();
|
||||
h.consume(format!("{user}:{REALM}:{pass}").as_bytes());
|
||||
Some(Account {
|
||||
user: user.to_owned(),
|
||||
pass: format!("{:x}", h.compute()),
|
||||
})
|
||||
}
|
||||
}
|
||||
|
||||
#[derive(Debug, Clone)]
|
||||
pub enum AuthMethod {
|
||||
Basic,
|
||||
@@ -194,20 +257,17 @@ pub enum AuthMethod {
|
||||
}
|
||||
|
||||
impl AuthMethod {
|
||||
pub fn www_auth(&self, stale: bool) -> Result<String> {
|
||||
pub fn www_auth(&self) -> Result<String> {
|
||||
match self {
|
||||
AuthMethod::Basic => Ok(format!("Basic realm=\"{REALM}\"")),
|
||||
AuthMethod::Digest => {
|
||||
let str_stale = if stale { "stale=true," } else { "" };
|
||||
Ok(format!(
|
||||
"Digest realm=\"{}\",nonce=\"{}\",{}qop=\"auth\"",
|
||||
AuthMethod::Digest => Ok(format!(
|
||||
"Digest realm=\"{}\",nonce=\"{}\",qop=\"auth\"",
|
||||
REALM,
|
||||
create_nonce()?,
|
||||
str_stale
|
||||
))
|
||||
}
|
||||
)),
|
||||
}
|
||||
}
|
||||
|
||||
pub fn get_user(&self, authorization: &HeaderValue) -> Option<String> {
|
||||
match self {
|
||||
AuthMethod::Basic => {
|
||||
@@ -227,7 +287,8 @@ impl AuthMethod {
|
||||
}
|
||||
}
|
||||
}
|
||||
pub fn validate(
|
||||
|
||||
fn check(
|
||||
&self,
|
||||
authorization: &HeaderValue,
|
||||
method: &str,
|
||||
@@ -245,12 +306,7 @@ impl AuthMethod {
|
||||
return None;
|
||||
}
|
||||
|
||||
let mut h = Context::new();
|
||||
h.consume(format!("{}:{}:{}", parts[0], REALM, parts[1]).as_bytes());
|
||||
|
||||
let http_pass = format!("{:x}", h.compute());
|
||||
|
||||
if http_pass == auth_pass {
|
||||
if parts[1] == auth_pass {
|
||||
return Some(());
|
||||
}
|
||||
|
||||
@@ -273,6 +329,11 @@ impl AuthMethod {
|
||||
if auth_user != username {
|
||||
return None;
|
||||
}
|
||||
|
||||
let mut h = Context::new();
|
||||
h.consume(format!("{}:{}:{}", auth_user, REALM, auth_pass).as_bytes());
|
||||
let auth_pass = format!("{:x}", h.compute());
|
||||
|
||||
let mut ha = Context::new();
|
||||
ha.consume(method);
|
||||
ha.consume(b":");
|
||||
@@ -285,7 +346,7 @@ impl AuthMethod {
|
||||
if qop == &b"auth".as_ref() || qop == &b"auth-int".as_ref() {
|
||||
correct_response = Some({
|
||||
let mut c = Context::new();
|
||||
c.consume(auth_pass);
|
||||
c.consume(&auth_pass);
|
||||
c.consume(b":");
|
||||
c.consume(nonce);
|
||||
c.consume(b":");
|
||||
@@ -308,7 +369,7 @@ impl AuthMethod {
|
||||
Some(r) => r,
|
||||
None => {
|
||||
let mut c = Context::new();
|
||||
c.consume(auth_pass);
|
||||
c.consume(&auth_pass);
|
||||
c.consume(b":");
|
||||
c.consume(nonce);
|
||||
c.consume(b":");
|
||||
@@ -317,7 +378,6 @@ impl AuthMethod {
|
||||
}
|
||||
};
|
||||
if correct_response.as_bytes() == *user_response {
|
||||
// grant access
|
||||
return Some(());
|
||||
}
|
||||
}
|
||||
@@ -417,3 +477,54 @@ fn create_nonce() -> Result<String> {
|
||||
let n = format!("{:08x}{:032x}", secs, h.compute());
|
||||
Ok(n[..34].to_string())
|
||||
}
|
||||
|
||||
#[cfg(test)]
|
||||
mod tests {
|
||||
use super::*;
|
||||
|
||||
#[test]
|
||||
fn test_access_paths() {
|
||||
let mut paths = AccessPaths::default();
|
||||
paths.add("/dir1", AccessPerm::ReadWrite);
|
||||
paths.add("/dir2/dir1", AccessPerm::ReadWrite);
|
||||
paths.add("/dir2/dir2", AccessPerm::ReadOnly);
|
||||
paths.add("/dir2/dir3/dir1", AccessPerm::ReadWrite);
|
||||
assert_eq!(
|
||||
paths.leaf_paths(Path::new("/tmp")),
|
||||
[
|
||||
"/tmp/dir1",
|
||||
"/tmp/dir2/dir1",
|
||||
"/tmp/dir2/dir2",
|
||||
"/tmp/dir2/dir3/dir1"
|
||||
]
|
||||
.iter()
|
||||
.map(PathBuf::from)
|
||||
.collect::<Vec<_>>()
|
||||
);
|
||||
assert_eq!(
|
||||
paths
|
||||
.find("dir2", false)
|
||||
.map(|v| v.leaf_paths(Path::new("/tmp/dir2"))),
|
||||
Some(
|
||||
["/tmp/dir2/dir1", "/tmp/dir2/dir2", "/tmp/dir2/dir3/dir1"]
|
||||
.iter()
|
||||
.map(PathBuf::from)
|
||||
.collect::<Vec<_>>()
|
||||
)
|
||||
);
|
||||
assert_eq!(paths.find("dir2", true), None);
|
||||
assert!(paths.find("dir1/file", true).is_some());
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_access_paths_perm() {
|
||||
let mut paths = AccessPaths::default();
|
||||
assert_eq!(paths.perm(), AccessPerm::IndexOnly);
|
||||
paths.set_perm(AccessPerm::ReadOnly);
|
||||
assert_eq!(paths.perm(), AccessPerm::ReadOnly);
|
||||
paths.set_perm(AccessPerm::ReadWrite);
|
||||
assert_eq!(paths.perm(), AccessPerm::ReadWrite);
|
||||
paths.set_perm(AccessPerm::ReadOnly);
|
||||
assert_eq!(paths.perm(), AccessPerm::ReadWrite);
|
||||
}
|
||||
}
|
||||
|
||||
373
src/server.rs
373
src/server.rs
@@ -1,18 +1,23 @@
|
||||
#![allow(clippy::too_many_arguments)]
|
||||
|
||||
use crate::auth::{AccessPaths, AccessPerm};
|
||||
use crate::streamer::Streamer;
|
||||
use crate::utils::{decode_uri, encode_uri, get_file_name, glob, try_get_file_name};
|
||||
use crate::utils::{
|
||||
decode_uri, encode_uri, get_file_mtime_and_mode, get_file_name, glob, try_get_file_name,
|
||||
};
|
||||
use crate::Args;
|
||||
use anyhow::{anyhow, Result};
|
||||
use walkdir::WalkDir;
|
||||
use xml::escape::escape_str_pcdata;
|
||||
|
||||
use async_zip::write::ZipFileWriter;
|
||||
use async_zip::{Compression, ZipEntryBuilder};
|
||||
use async_zip::tokio::write::ZipFileWriter;
|
||||
use async_zip::{Compression, ZipDateTime, ZipEntryBuilder};
|
||||
use chrono::{LocalResult, TimeZone, Utc};
|
||||
use futures::TryStreamExt;
|
||||
use headers::{
|
||||
AcceptRanges, AccessControlAllowCredentials, AccessControlAllowOrigin, Connection,
|
||||
ContentLength, ContentType, ETag, HeaderMap, HeaderMapExt, IfModifiedSince, IfNoneMatch,
|
||||
IfRange, LastModified, Range,
|
||||
AcceptRanges, AccessControlAllowCredentials, AccessControlAllowOrigin, ContentLength,
|
||||
ContentType, ETag, HeaderMap, HeaderMapExt, IfModifiedSince, IfNoneMatch, IfRange,
|
||||
LastModified, Range,
|
||||
};
|
||||
use hyper::header::{
|
||||
HeaderValue, AUTHORIZATION, CONTENT_DISPOSITION, CONTENT_LENGTH, CONTENT_RANGE, CONTENT_TYPE,
|
||||
@@ -21,17 +26,19 @@ use hyper::header::{
|
||||
use hyper::{Body, Method, StatusCode, Uri};
|
||||
use serde::Serialize;
|
||||
use std::borrow::Cow;
|
||||
use std::cmp::Ordering;
|
||||
use std::collections::HashMap;
|
||||
use std::fs::Metadata;
|
||||
use std::io::SeekFrom;
|
||||
use std::net::SocketAddr;
|
||||
use std::path::{Path, PathBuf};
|
||||
use std::sync::atomic::{AtomicBool, Ordering};
|
||||
use std::sync::atomic::{self, AtomicBool};
|
||||
use std::sync::Arc;
|
||||
use std::time::SystemTime;
|
||||
use tokio::fs::File;
|
||||
use tokio::io::{AsyncReadExt, AsyncSeekExt, AsyncWrite};
|
||||
use tokio::{fs, io};
|
||||
use tokio_util::compat::FuturesAsyncWriteCompatExt;
|
||||
use tokio_util::io::StreamReader;
|
||||
use uuid::Uuid;
|
||||
|
||||
@@ -134,16 +141,32 @@ impl Server {
|
||||
}
|
||||
|
||||
let authorization = headers.get(AUTHORIZATION);
|
||||
let guard_type = self.args.auth.guard(
|
||||
req_path,
|
||||
let relative_path = match self.resolve_path(req_path) {
|
||||
Some(v) => v,
|
||||
None => {
|
||||
status_forbid(&mut res);
|
||||
return Ok(res);
|
||||
}
|
||||
};
|
||||
|
||||
let guard = self.args.auth.guard(
|
||||
&relative_path,
|
||||
&method,
|
||||
authorization,
|
||||
self.args.auth_method.clone(),
|
||||
);
|
||||
if guard_type.is_reject() {
|
||||
|
||||
let (user, access_paths) = match guard {
|
||||
(None, None) => {
|
||||
self.auth_reject(&mut res)?;
|
||||
return Ok(res);
|
||||
}
|
||||
(Some(_), None) => {
|
||||
status_forbid(&mut res);
|
||||
return Ok(res);
|
||||
}
|
||||
(x, Some(y)) => (x, y),
|
||||
};
|
||||
|
||||
let query = req.uri().query().unwrap_or_default();
|
||||
let query_params: HashMap<String, String> = form_urlencoded::parse(query.as_bytes())
|
||||
@@ -169,8 +192,7 @@ impl Server {
|
||||
}
|
||||
return Ok(res);
|
||||
}
|
||||
|
||||
let path = match self.extract_path(req_path) {
|
||||
let path = match self.join_path(&relative_path) {
|
||||
Some(v) => v,
|
||||
None => {
|
||||
status_forbid(&mut res);
|
||||
@@ -207,31 +229,38 @@ impl Server {
|
||||
status_not_found(&mut res);
|
||||
return Ok(res);
|
||||
}
|
||||
self.handle_zip_dir(path, head_only, &mut res).await?;
|
||||
self.handle_zip_dir(path, head_only, access_paths, &mut res)
|
||||
.await?;
|
||||
} else if allow_search && query_params.contains_key("q") {
|
||||
let user = self.retrieve_user(authorization);
|
||||
self.handle_search_dir(path, &query_params, head_only, user, &mut res)
|
||||
self.handle_search_dir(
|
||||
path,
|
||||
&query_params,
|
||||
head_only,
|
||||
user,
|
||||
access_paths,
|
||||
&mut res,
|
||||
)
|
||||
.await?;
|
||||
} else {
|
||||
let user = self.retrieve_user(authorization);
|
||||
self.handle_render_index(
|
||||
path,
|
||||
&query_params,
|
||||
headers,
|
||||
head_only,
|
||||
user,
|
||||
access_paths,
|
||||
&mut res,
|
||||
)
|
||||
.await?;
|
||||
}
|
||||
} else if render_index || render_spa {
|
||||
let user = self.retrieve_user(authorization);
|
||||
self.handle_render_index(
|
||||
path,
|
||||
&query_params,
|
||||
headers,
|
||||
head_only,
|
||||
user,
|
||||
access_paths,
|
||||
&mut res,
|
||||
)
|
||||
.await?;
|
||||
@@ -240,19 +269,32 @@ impl Server {
|
||||
status_not_found(&mut res);
|
||||
return Ok(res);
|
||||
}
|
||||
self.handle_zip_dir(path, head_only, &mut res).await?;
|
||||
self.handle_zip_dir(path, head_only, access_paths, &mut res)
|
||||
.await?;
|
||||
} else if allow_search && query_params.contains_key("q") {
|
||||
let user = self.retrieve_user(authorization);
|
||||
self.handle_search_dir(path, &query_params, head_only, user, &mut res)
|
||||
self.handle_search_dir(
|
||||
path,
|
||||
&query_params,
|
||||
head_only,
|
||||
user,
|
||||
access_paths,
|
||||
&mut res,
|
||||
)
|
||||
.await?;
|
||||
} else {
|
||||
let user = self.retrieve_user(authorization);
|
||||
self.handle_ls_dir(path, true, &query_params, head_only, user, &mut res)
|
||||
self.handle_ls_dir(
|
||||
path,
|
||||
true,
|
||||
&query_params,
|
||||
head_only,
|
||||
user,
|
||||
access_paths,
|
||||
&mut res,
|
||||
)
|
||||
.await?;
|
||||
}
|
||||
} else if is_file {
|
||||
if query_params.contains_key("edit") {
|
||||
let user = self.retrieve_user(authorization);
|
||||
self.handle_edit_file(path, head_only, user, &mut res)
|
||||
.await?;
|
||||
} else {
|
||||
@@ -263,8 +305,15 @@ impl Server {
|
||||
self.handle_render_spa(path, headers, head_only, &mut res)
|
||||
.await?;
|
||||
} else if allow_upload && req_path.ends_with('/') {
|
||||
let user = self.retrieve_user(authorization);
|
||||
self.handle_ls_dir(path, false, &query_params, head_only, user, &mut res)
|
||||
self.handle_ls_dir(
|
||||
path,
|
||||
false,
|
||||
&query_params,
|
||||
head_only,
|
||||
user,
|
||||
access_paths,
|
||||
&mut res,
|
||||
)
|
||||
.await?;
|
||||
} else {
|
||||
status_not_found(&mut res);
|
||||
@@ -292,7 +341,14 @@ impl Server {
|
||||
method => match method.as_str() {
|
||||
"PROPFIND" => {
|
||||
if is_dir {
|
||||
self.handle_propfind_dir(path, headers, &mut res).await?;
|
||||
let access_paths = if access_paths.perm().indexonly() {
|
||||
// see https://github.com/sigoden/dufs/issues/229
|
||||
AccessPaths::new(AccessPerm::ReadOnly)
|
||||
} else {
|
||||
access_paths
|
||||
};
|
||||
self.handle_propfind_dir(path, headers, access_paths, &mut res)
|
||||
.await?;
|
||||
} else if is_file {
|
||||
self.handle_propfind_file(path, &mut res).await?;
|
||||
} else {
|
||||
@@ -399,11 +455,12 @@ impl Server {
|
||||
query_params: &HashMap<String, String>,
|
||||
head_only: bool,
|
||||
user: Option<String>,
|
||||
access_paths: AccessPaths,
|
||||
res: &mut Response,
|
||||
) -> Result<()> {
|
||||
let mut paths = vec![];
|
||||
if exist {
|
||||
paths = match self.list_dir(path, path).await {
|
||||
paths = match self.list_dir(path, path, access_paths.clone()).await {
|
||||
Ok(paths) => paths,
|
||||
Err(_) => {
|
||||
status_forbid(res);
|
||||
@@ -411,7 +468,16 @@ impl Server {
|
||||
}
|
||||
}
|
||||
};
|
||||
self.send_index(path, paths, exist, query_params, head_only, user, res)
|
||||
self.send_index(
|
||||
path,
|
||||
paths,
|
||||
exist,
|
||||
query_params,
|
||||
head_only,
|
||||
user,
|
||||
access_paths,
|
||||
res,
|
||||
)
|
||||
}
|
||||
|
||||
async fn handle_search_dir(
|
||||
@@ -420,6 +486,7 @@ impl Server {
|
||||
query_params: &HashMap<String, String>,
|
||||
head_only: bool,
|
||||
user: Option<String>,
|
||||
access_paths: AccessPaths,
|
||||
res: &mut Response,
|
||||
) -> Result<()> {
|
||||
let mut paths: Vec<PathItem> = vec![];
|
||||
@@ -432,11 +499,14 @@ impl Server {
|
||||
let hidden = Arc::new(self.args.hidden.to_vec());
|
||||
let hidden = hidden.clone();
|
||||
let running = self.running.clone();
|
||||
let access_paths = access_paths.clone();
|
||||
let search_paths = tokio::task::spawn_blocking(move || {
|
||||
let mut it = WalkDir::new(&path_buf).into_iter();
|
||||
let mut paths: Vec<PathBuf> = vec![];
|
||||
for dir in access_paths.leaf_paths(&path_buf) {
|
||||
let mut it = WalkDir::new(&dir).into_iter();
|
||||
it.next();
|
||||
while let Some(Ok(entry)) = it.next() {
|
||||
if !running.load(Ordering::SeqCst) {
|
||||
if !running.load(atomic::Ordering::SeqCst) {
|
||||
break;
|
||||
}
|
||||
let entry_path = entry.path();
|
||||
@@ -464,6 +534,7 @@ impl Server {
|
||||
}
|
||||
paths.push(entry_path.to_path_buf());
|
||||
}
|
||||
}
|
||||
paths
|
||||
})
|
||||
.await?;
|
||||
@@ -473,19 +544,28 @@ impl Server {
|
||||
}
|
||||
}
|
||||
}
|
||||
self.send_index(path, paths, true, query_params, head_only, user, res)
|
||||
self.send_index(
|
||||
path,
|
||||
paths,
|
||||
true,
|
||||
query_params,
|
||||
head_only,
|
||||
user,
|
||||
access_paths,
|
||||
res,
|
||||
)
|
||||
}
|
||||
|
||||
async fn handle_zip_dir(&self, path: &Path, head_only: bool, res: &mut Response) -> Result<()> {
|
||||
async fn handle_zip_dir(
|
||||
&self,
|
||||
path: &Path,
|
||||
head_only: bool,
|
||||
access_paths: AccessPaths,
|
||||
res: &mut Response,
|
||||
) -> Result<()> {
|
||||
let (mut writer, reader) = tokio::io::duplex(BUF_SIZE);
|
||||
let filename = try_get_file_name(path)?;
|
||||
res.headers_mut().insert(
|
||||
CONTENT_DISPOSITION,
|
||||
HeaderValue::from_str(&format!(
|
||||
"attachment; filename=\"{}.zip\"",
|
||||
encode_uri(filename),
|
||||
))?,
|
||||
);
|
||||
set_content_diposition(res, false, &format!("{}.zip", filename))?;
|
||||
res.headers_mut()
|
||||
.insert("content-type", HeaderValue::from_static("application/zip"));
|
||||
if head_only {
|
||||
@@ -495,7 +575,7 @@ impl Server {
|
||||
let hidden = self.args.hidden.clone();
|
||||
let running = self.running.clone();
|
||||
tokio::spawn(async move {
|
||||
if let Err(e) = zip_dir(&mut writer, &path, &hidden, running).await {
|
||||
if let Err(e) = zip_dir(&mut writer, &path, access_paths, &hidden, running).await {
|
||||
error!("Failed to zip {}, {}", path.display(), e);
|
||||
}
|
||||
});
|
||||
@@ -511,6 +591,7 @@ impl Server {
|
||||
headers: &HeaderMap<HeaderValue>,
|
||||
head_only: bool,
|
||||
user: Option<String>,
|
||||
access_paths: AccessPaths,
|
||||
res: &mut Response,
|
||||
) -> Result<()> {
|
||||
let index_path = path.join(INDEX_NAME);
|
||||
@@ -523,7 +604,7 @@ impl Server {
|
||||
self.handle_send_file(&index_path, headers, head_only, res)
|
||||
.await?;
|
||||
} else if self.args.render_try_index {
|
||||
self.handle_ls_dir(path, true, query_params, head_only, user, res)
|
||||
self.handle_ls_dir(path, true, query_params, head_only, user, access_paths, res)
|
||||
.await?;
|
||||
} else {
|
||||
status_not_found(res)
|
||||
@@ -644,10 +725,7 @@ impl Server {
|
||||
);
|
||||
|
||||
let filename = try_get_file_name(path)?;
|
||||
res.headers_mut().insert(
|
||||
CONTENT_DISPOSITION,
|
||||
HeaderValue::from_str(&format!("inline; filename=\"{}\"", encode_uri(filename),))?,
|
||||
);
|
||||
set_content_diposition(res, true, filename)?;
|
||||
|
||||
res.headers_mut().typed_insert(AcceptRanges::bytes());
|
||||
|
||||
@@ -708,7 +786,7 @@ impl Server {
|
||||
uri_prefix: self.args.uri_prefix.clone(),
|
||||
allow_upload: self.args.allow_upload,
|
||||
allow_delete: self.args.allow_delete,
|
||||
auth: self.args.auth.valid(),
|
||||
auth: self.args.auth.exist(),
|
||||
user,
|
||||
editable,
|
||||
};
|
||||
@@ -716,7 +794,7 @@ impl Server {
|
||||
.typed_insert(ContentType::from(mime_guess::mime::TEXT_HTML_UTF_8));
|
||||
let output = self
|
||||
.html
|
||||
.replace("__ASSERTS_PREFIX__", &self.assets_prefix)
|
||||
.replace("__ASSETS_PREFIX__", &self.assets_prefix)
|
||||
.replace("__INDEX_DATA__", &serde_json::to_string(&data)?);
|
||||
res.headers_mut()
|
||||
.typed_insert(ContentLength(output.as_bytes().len() as u64));
|
||||
@@ -731,6 +809,7 @@ impl Server {
|
||||
&self,
|
||||
path: &Path,
|
||||
headers: &HeaderMap<HeaderValue>,
|
||||
access_paths: AccessPaths,
|
||||
res: &mut Response,
|
||||
) -> Result<()> {
|
||||
let depth: u32 = match headers.get("depth") {
|
||||
@@ -748,7 +827,7 @@ impl Server {
|
||||
None => vec![],
|
||||
};
|
||||
if depth != 0 {
|
||||
match self.list_dir(path, &self.args.path).await {
|
||||
match self.list_dir(path, &self.args.path, access_paths).await {
|
||||
Ok(child) => paths.extend(child),
|
||||
Err(_) => {
|
||||
status_forbid(res);
|
||||
@@ -868,17 +947,33 @@ impl Server {
|
||||
query_params: &HashMap<String, String>,
|
||||
head_only: bool,
|
||||
user: Option<String>,
|
||||
access_paths: AccessPaths,
|
||||
res: &mut Response,
|
||||
) -> Result<()> {
|
||||
if let Some(sort) = query_params.get("sort") {
|
||||
if sort == "name" {
|
||||
paths.sort_by(|v1, v2| {
|
||||
alphanumeric_sort::compare_str(v1.name.to_lowercase(), v2.name.to_lowercase())
|
||||
paths.sort_by(|v1, v2| match v1.path_type.cmp(&v2.path_type) {
|
||||
Ordering::Equal => {
|
||||
alphanumeric_sort::compare_str(v1.name.clone(), v2.name.clone())
|
||||
}
|
||||
v => v,
|
||||
})
|
||||
} else if sort == "mtime" {
|
||||
paths.sort_by(|v1, v2| v1.mtime.cmp(&v2.mtime))
|
||||
paths.sort_by(|v1, v2| match v1.path_type.cmp(&v2.path_type) {
|
||||
Ordering::Equal => v1.mtime.cmp(&v2.mtime),
|
||||
v => v,
|
||||
})
|
||||
} else if sort == "size" {
|
||||
paths.sort_by(|v1, v2| v1.size.unwrap_or(0).cmp(&v2.size.unwrap_or(0)))
|
||||
paths.sort_by(|v1, v2| match v1.path_type.cmp(&v2.path_type) {
|
||||
Ordering::Equal => {
|
||||
if v1.is_dir() {
|
||||
alphanumeric_sort::compare_str(v1.name.clone(), v2.name.clone())
|
||||
} else {
|
||||
v1.size.unwrap_or(0).cmp(&v2.size.unwrap_or(0))
|
||||
}
|
||||
}
|
||||
v => v,
|
||||
})
|
||||
}
|
||||
if query_params
|
||||
.get("order")
|
||||
@@ -913,16 +1008,17 @@ impl Server {
|
||||
return Ok(());
|
||||
}
|
||||
let href = format!("/{}", normalize_path(path.strip_prefix(&self.args.path)?));
|
||||
let readwrite = access_paths.perm().readwrite();
|
||||
let data = IndexData {
|
||||
kind: DataKind::Index,
|
||||
href,
|
||||
uri_prefix: self.args.uri_prefix.clone(),
|
||||
allow_upload: self.args.allow_upload,
|
||||
allow_delete: self.args.allow_delete,
|
||||
allow_upload: self.args.allow_upload && readwrite,
|
||||
allow_delete: self.args.allow_delete && readwrite,
|
||||
allow_search: self.args.allow_search,
|
||||
allow_archive: self.args.allow_archive,
|
||||
dir_exists: exist,
|
||||
auth: self.args.auth.valid(),
|
||||
auth: self.args.auth.exist(),
|
||||
user,
|
||||
paths,
|
||||
};
|
||||
@@ -934,7 +1030,7 @@ impl Server {
|
||||
res.headers_mut()
|
||||
.typed_insert(ContentType::from(mime_guess::mime::TEXT_HTML_UTF_8));
|
||||
self.html
|
||||
.replace("__ASSERTS_PREFIX__", &self.assets_prefix)
|
||||
.replace("__ASSETS_PREFIX__", &self.assets_prefix)
|
||||
.replace("__INDEX_DATA__", &serde_json::to_string(&data)?)
|
||||
};
|
||||
res.headers_mut()
|
||||
@@ -947,10 +1043,10 @@ impl Server {
|
||||
}
|
||||
|
||||
fn auth_reject(&self, res: &mut Response) -> Result<()> {
|
||||
let value = self.args.auth_method.www_auth(false)?;
|
||||
let value = self.args.auth_method.www_auth()?;
|
||||
set_webdav_headers(res);
|
||||
res.headers_mut().typed_insert(Connection::close());
|
||||
res.headers_mut().insert(WWW_AUTHENTICATE, value.parse()?);
|
||||
// set 401 to make the browser pop up the login box
|
||||
*res.status_mut() = StatusCode::UNAUTHORIZED;
|
||||
Ok(())
|
||||
}
|
||||
@@ -972,20 +1068,32 @@ impl Server {
|
||||
return None;
|
||||
}
|
||||
};
|
||||
|
||||
let relative_path = match self.resolve_path(&dest_path) {
|
||||
Some(v) => v,
|
||||
None => {
|
||||
*res.status_mut() = StatusCode::BAD_REQUEST;
|
||||
return None;
|
||||
}
|
||||
};
|
||||
|
||||
let authorization = headers.get(AUTHORIZATION);
|
||||
let guard_type = self.args.auth.guard(
|
||||
&dest_path,
|
||||
let guard = self.args.auth.guard(
|
||||
&relative_path,
|
||||
req.method(),
|
||||
authorization,
|
||||
self.args.auth_method.clone(),
|
||||
);
|
||||
if guard_type.is_reject() {
|
||||
*res.status_mut() = StatusCode::FORBIDDEN;
|
||||
*res.body_mut() = Body::from("Forbidden");
|
||||
|
||||
match guard {
|
||||
(_, Some(_)) => {}
|
||||
_ => {
|
||||
status_forbid(res);
|
||||
return None;
|
||||
}
|
||||
};
|
||||
|
||||
let dest = match self.extract_path(&dest_path) {
|
||||
let dest = match self.join_path(&relative_path) {
|
||||
Some(dest) => dest,
|
||||
None => {
|
||||
*res.status_mut() = StatusCode::BAD_REQUEST;
|
||||
@@ -1002,49 +1110,61 @@ impl Server {
|
||||
Some(uri.path().to_string())
|
||||
}
|
||||
|
||||
fn extract_path(&self, path: &str) -> Option<PathBuf> {
|
||||
let mut slash_stripped_path = path;
|
||||
while let Some(p) = slash_stripped_path.strip_prefix('/') {
|
||||
slash_stripped_path = p
|
||||
fn resolve_path(&self, path: &str) -> Option<String> {
|
||||
let path = path.trim_matches('/');
|
||||
let path = decode_uri(path)?;
|
||||
let prefix = self.args.path_prefix.as_str();
|
||||
if prefix == "/" {
|
||||
return Some(path.to_string());
|
||||
}
|
||||
let decoded_path = decode_uri(slash_stripped_path)?;
|
||||
let slashes_switched = if cfg!(windows) {
|
||||
decoded_path.replace('/', "\\")
|
||||
} else {
|
||||
decoded_path.into_owned()
|
||||
};
|
||||
let stripped_path = match self.strip_path_prefix(&slashes_switched) {
|
||||
Some(path) => path,
|
||||
None => return None,
|
||||
};
|
||||
Some(self.args.path.join(stripped_path))
|
||||
path.strip_prefix(prefix.trim_start_matches('/'))
|
||||
.map(|v| v.trim_matches('/').to_string())
|
||||
}
|
||||
|
||||
fn strip_path_prefix<'a, P: AsRef<Path>>(&self, path: &'a P) -> Option<&'a Path> {
|
||||
let path = path.as_ref();
|
||||
if self.args.path_prefix.is_empty() {
|
||||
Some(path)
|
||||
} else {
|
||||
path.strip_prefix(&self.args.path_prefix).ok()
|
||||
fn join_path(&self, path: &str) -> Option<PathBuf> {
|
||||
if path.is_empty() {
|
||||
return Some(self.args.path.clone());
|
||||
}
|
||||
let path = if cfg!(windows) {
|
||||
path.replace('/', "\\")
|
||||
} else {
|
||||
path.to_string()
|
||||
};
|
||||
Some(self.args.path.join(path))
|
||||
}
|
||||
|
||||
async fn list_dir(&self, entry_path: &Path, base_path: &Path) -> Result<Vec<PathItem>> {
|
||||
async fn list_dir(
|
||||
&self,
|
||||
entry_path: &Path,
|
||||
base_path: &Path,
|
||||
access_paths: AccessPaths,
|
||||
) -> Result<Vec<PathItem>> {
|
||||
let mut paths: Vec<PathItem> = vec![];
|
||||
if access_paths.perm().indexonly() {
|
||||
for name in access_paths.child_paths() {
|
||||
let entry_path = entry_path.join(name);
|
||||
self.add_pathitem(&mut paths, base_path, &entry_path).await;
|
||||
}
|
||||
} else {
|
||||
let mut rd = fs::read_dir(entry_path).await?;
|
||||
while let Ok(Some(entry)) = rd.next_entry().await {
|
||||
let entry_path = entry.path();
|
||||
let base_name = get_file_name(&entry_path);
|
||||
if let Ok(Some(item)) = self.to_pathitem(entry_path.as_path(), base_path).await {
|
||||
if is_hidden(&self.args.hidden, base_name, item.is_dir()) {
|
||||
continue;
|
||||
}
|
||||
paths.push(item);
|
||||
self.add_pathitem(&mut paths, base_path, &entry_path).await;
|
||||
}
|
||||
}
|
||||
Ok(paths)
|
||||
}
|
||||
|
||||
async fn add_pathitem(&self, paths: &mut Vec<PathItem>, base_path: &Path, entry_path: &Path) {
|
||||
let base_name = get_file_name(entry_path);
|
||||
if let Ok(Some(item)) = self.to_pathitem(entry_path, base_path).await {
|
||||
if is_hidden(&self.args.hidden, base_name, item.is_dir()) {
|
||||
return;
|
||||
}
|
||||
paths.push(item);
|
||||
}
|
||||
}
|
||||
|
||||
async fn to_pathitem<P: AsRef<Path>>(&self, path: P, base_path: P) -> Result<Option<PathItem>> {
|
||||
let path = path.as_ref();
|
||||
let (meta, meta2) = tokio::join!(fs::metadata(&path), fs::symlink_metadata(&path));
|
||||
@@ -1074,10 +1194,6 @@ impl Server {
|
||||
size,
|
||||
}))
|
||||
}
|
||||
|
||||
fn retrieve_user(&self, authorization: Option<&HeaderValue>) -> Option<String> {
|
||||
self.args.auth_method.get_user(authorization?)
|
||||
}
|
||||
}
|
||||
|
||||
#[derive(Debug, Serialize)]
|
||||
@@ -1175,7 +1291,7 @@ impl PathItem {
|
||||
}
|
||||
}
|
||||
|
||||
#[derive(Debug, Serialize, Eq, PartialEq, Ord, PartialOrd)]
|
||||
#[derive(Debug, Serialize, Eq, PartialEq)]
|
||||
enum PathType {
|
||||
Dir,
|
||||
SymlinkDir,
|
||||
@@ -1183,6 +1299,24 @@ enum PathType {
|
||||
SymlinkFile,
|
||||
}
|
||||
|
||||
impl Ord for PathType {
|
||||
fn cmp(&self, other: &Self) -> Ordering {
|
||||
let to_value = |t: &Self| -> u8 {
|
||||
if matches!(t, Self::Dir | Self::SymlinkDir) {
|
||||
0
|
||||
} else {
|
||||
1
|
||||
}
|
||||
};
|
||||
to_value(self).cmp(&to_value(other))
|
||||
}
|
||||
}
|
||||
impl PartialOrd for PathType {
|
||||
fn partial_cmp(&self, other: &Self) -> Option<Ordering> {
|
||||
Some(self.cmp(other))
|
||||
}
|
||||
}
|
||||
|
||||
fn to_timestamp(time: &SystemTime) -> u64 {
|
||||
time.duration_since(SystemTime::UNIX_EPOCH)
|
||||
.unwrap_or_default()
|
||||
@@ -1214,17 +1348,15 @@ fn add_cors(res: &mut Response) {
|
||||
.typed_insert(AccessControlAllowCredentials);
|
||||
res.headers_mut().insert(
|
||||
"Access-Control-Allow-Methods",
|
||||
HeaderValue::from_static("GET,HEAD,PUT,OPTIONS,DELETE,PROPFIND,COPY,MOVE"),
|
||||
HeaderValue::from_static("*"),
|
||||
);
|
||||
res.headers_mut().insert(
|
||||
"Access-Control-Allow-Headers",
|
||||
HeaderValue::from_static("Authorization,Destination,Range,Content-Type"),
|
||||
HeaderValue::from_static("Authorization,*"),
|
||||
);
|
||||
res.headers_mut().insert(
|
||||
"Access-Control-Expose-Headers",
|
||||
HeaderValue::from_static(
|
||||
"WWW-Authenticate,Content-Range,Accept-Ranges,Content-Disposition",
|
||||
),
|
||||
HeaderValue::from_static("Authorization,*"),
|
||||
);
|
||||
}
|
||||
|
||||
@@ -1245,18 +1377,21 @@ fn res_multistatus(res: &mut Response, content: &str) {
|
||||
async fn zip_dir<W: AsyncWrite + Unpin>(
|
||||
writer: &mut W,
|
||||
dir: &Path,
|
||||
access_paths: AccessPaths,
|
||||
hidden: &[String],
|
||||
running: Arc<AtomicBool>,
|
||||
) -> Result<()> {
|
||||
let mut writer = ZipFileWriter::new(writer);
|
||||
let mut writer = ZipFileWriter::with_tokio(writer);
|
||||
let hidden = Arc::new(hidden.to_vec());
|
||||
let hidden = hidden.clone();
|
||||
let dir_path_buf = dir.to_path_buf();
|
||||
let dir_clone = dir.to_path_buf();
|
||||
let zip_paths = tokio::task::spawn_blocking(move || {
|
||||
let mut it = WalkDir::new(&dir_path_buf).into_iter();
|
||||
let mut paths: Vec<PathBuf> = vec![];
|
||||
for dir in access_paths.leaf_paths(&dir_clone) {
|
||||
let mut it = WalkDir::new(&dir).into_iter();
|
||||
it.next();
|
||||
while let Some(Ok(entry)) = it.next() {
|
||||
if !running.load(Ordering::SeqCst) {
|
||||
if !running.load(atomic::Ordering::SeqCst) {
|
||||
break;
|
||||
}
|
||||
let entry_path = entry.path();
|
||||
@@ -1287,6 +1422,7 @@ async fn zip_dir<W: AsyncWrite + Unpin>(
|
||||
}
|
||||
paths.push(entry_path.to_path_buf());
|
||||
}
|
||||
}
|
||||
paths
|
||||
})
|
||||
.await?;
|
||||
@@ -1295,12 +1431,14 @@ async fn zip_dir<W: AsyncWrite + Unpin>(
|
||||
Some(v) => v,
|
||||
None => continue,
|
||||
};
|
||||
let builder =
|
||||
ZipEntryBuilder::new(filename.into(), Compression::Deflate).unix_permissions(0o644);
|
||||
let (datetime, mode) = get_file_mtime_and_mode(&zip_path).await?;
|
||||
let builder = ZipEntryBuilder::new(filename.into(), Compression::Deflate)
|
||||
.unix_permissions(mode)
|
||||
.last_modification_date(ZipDateTime::from_chrono(&datetime));
|
||||
let mut file = File::open(&zip_path).await?;
|
||||
let mut file_writer = writer.write_entry_stream(builder).await?;
|
||||
let mut file_writer = writer.write_entry_stream(builder).await?.compat_write();
|
||||
io::copy(&mut file, &mut file_writer).await?;
|
||||
file_writer.close().await?;
|
||||
file_writer.into_inner().close().await?;
|
||||
}
|
||||
writer.close().await?;
|
||||
Ok(())
|
||||
@@ -1359,6 +1497,21 @@ fn status_no_content(res: &mut Response) {
|
||||
*res.status_mut() = StatusCode::NO_CONTENT;
|
||||
}
|
||||
|
||||
fn set_content_diposition(res: &mut Response, inline: bool, filename: &str) -> Result<()> {
|
||||
let kind = if inline { "inline" } else { "attachment" };
|
||||
let value = if filename.is_ascii() {
|
||||
HeaderValue::from_str(&format!("{kind}; filename=\"{}\"", filename,))?
|
||||
} else {
|
||||
HeaderValue::from_str(&format!(
|
||||
"{kind}; filename=\"{}\"; filename*=UTF-8''{}",
|
||||
filename,
|
||||
encode_uri(filename),
|
||||
))?
|
||||
};
|
||||
res.headers_mut().insert(CONTENT_DISPOSITION, value);
|
||||
Ok(())
|
||||
}
|
||||
|
||||
fn is_hidden(hidden: &[String], file_name: &str, is_dir_type: bool) -> bool {
|
||||
hidden.iter().any(|v| {
|
||||
if is_dir_type {
|
||||
|
||||
16
src/utils.rs
16
src/utils.rs
@@ -1,4 +1,5 @@
|
||||
use anyhow::{anyhow, Context, Result};
|
||||
use chrono::{DateTime, Utc};
|
||||
use std::{
|
||||
borrow::Cow,
|
||||
path::Path,
|
||||
@@ -28,6 +29,21 @@ pub fn get_file_name(path: &Path) -> &str {
|
||||
.unwrap_or_default()
|
||||
}
|
||||
|
||||
#[cfg(unix)]
|
||||
pub async fn get_file_mtime_and_mode(path: &Path) -> Result<(DateTime<Utc>, u16)> {
|
||||
use std::os::unix::prelude::MetadataExt;
|
||||
let meta = tokio::fs::metadata(path).await?;
|
||||
let datetime: DateTime<Utc> = meta.modified()?.into();
|
||||
Ok((datetime, meta.mode() as u16))
|
||||
}
|
||||
|
||||
#[cfg(not(unix))]
|
||||
pub async fn get_file_mtime_and_mode(path: &Path) -> Result<(DateTime<Utc>, u16)> {
|
||||
let meta = tokio::fs::metadata(&path).await?;
|
||||
let datetime: DateTime<Utc> = meta.modified()?.into();
|
||||
Ok((datetime, 0o644))
|
||||
}
|
||||
|
||||
pub fn try_get_file_name(path: &Path) -> Result<&str> {
|
||||
path.file_name()
|
||||
.and_then(|v| v.to_str())
|
||||
|
||||
101
tests/auth.rs
101
tests/auth.rs
@@ -3,10 +3,11 @@ mod utils;
|
||||
|
||||
use diqwest::blocking::WithDigestAuth;
|
||||
use fixtures::{server, Error, TestServer};
|
||||
use indexmap::IndexSet;
|
||||
use rstest::rstest;
|
||||
|
||||
#[rstest]
|
||||
fn no_auth(#[with(&["--auth", "/@user:pass", "-A"])] server: TestServer) -> Result<(), Error> {
|
||||
fn no_auth(#[with(&["--auth", "user:pass@/:rw", "-A"])] server: TestServer) -> Result<(), Error> {
|
||||
let resp = reqwest::blocking::get(server.url())?;
|
||||
assert_eq!(resp.status(), 401);
|
||||
assert!(resp.headers().contains_key("www-authenticate"));
|
||||
@@ -17,7 +18,7 @@ fn no_auth(#[with(&["--auth", "/@user:pass", "-A"])] server: TestServer) -> Resu
|
||||
}
|
||||
|
||||
#[rstest]
|
||||
fn auth(#[with(&["--auth", "/@user:pass", "-A"])] server: TestServer) -> Result<(), Error> {
|
||||
fn auth(#[with(&["--auth", "user:pass@/:rw", "-A"])] server: TestServer) -> Result<(), Error> {
|
||||
let url = format!("{}file1", server.url());
|
||||
let resp = fetch!(b"PUT", &url).body(b"abc".to_vec()).send()?;
|
||||
assert_eq!(resp.status(), 401);
|
||||
@@ -29,7 +30,24 @@ fn auth(#[with(&["--auth", "/@user:pass", "-A"])] server: TestServer) -> Result<
|
||||
}
|
||||
|
||||
#[rstest]
|
||||
fn auth_skip(#[with(&["--auth", "/@user:pass@*"])] server: TestServer) -> Result<(), Error> {
|
||||
fn auth_and_public(
|
||||
#[with(&["--auth", "user:pass@/:rw|@/", "-A"])] server: TestServer,
|
||||
) -> Result<(), Error> {
|
||||
let url = format!("{}file1", server.url());
|
||||
let resp = fetch!(b"PUT", &url).body(b"abc".to_vec()).send()?;
|
||||
assert_eq!(resp.status(), 401);
|
||||
let resp = fetch!(b"PUT", &url)
|
||||
.body(b"abc".to_vec())
|
||||
.send_with_digest_auth("user", "pass")?;
|
||||
assert_eq!(resp.status(), 201);
|
||||
let resp = fetch!(b"GET", &url).send()?;
|
||||
assert_eq!(resp.status(), 200);
|
||||
assert_eq!(resp.text()?, "abc");
|
||||
Ok(())
|
||||
}
|
||||
|
||||
#[rstest]
|
||||
fn auth_skip(#[with(&["--auth", "@/"])] server: TestServer) -> Result<(), Error> {
|
||||
let resp = reqwest::blocking::get(server.url())?;
|
||||
assert_eq!(resp.status(), 200);
|
||||
Ok(())
|
||||
@@ -37,7 +55,7 @@ fn auth_skip(#[with(&["--auth", "/@user:pass@*"])] server: TestServer) -> Result
|
||||
|
||||
#[rstest]
|
||||
fn auth_skip_on_options_method(
|
||||
#[with(&["--auth", "/@user:pass"])] server: TestServer,
|
||||
#[with(&["--auth", "user:pass@/:rw"])] server: TestServer,
|
||||
) -> Result<(), Error> {
|
||||
let url = format!("{}index.html", server.url());
|
||||
let resp = fetch!(b"OPTIONS", &url).send()?;
|
||||
@@ -47,13 +65,13 @@ fn auth_skip_on_options_method(
|
||||
|
||||
#[rstest]
|
||||
fn auth_check(
|
||||
#[with(&["--auth", "/@user:pass@user2:pass2", "-A"])] server: TestServer,
|
||||
#[with(&["--auth", "user:pass@/:rw|user2:pass2@/", "-A"])] server: TestServer,
|
||||
) -> Result<(), Error> {
|
||||
let url = format!("{}index.html", server.url());
|
||||
let resp = fetch!(b"WRITEABLE", &url).send()?;
|
||||
assert_eq!(resp.status(), 401);
|
||||
let resp = fetch!(b"WRITEABLE", &url).send_with_digest_auth("user2", "pass2")?;
|
||||
assert_eq!(resp.status(), 401);
|
||||
assert_eq!(resp.status(), 403);
|
||||
let resp = fetch!(b"WRITEABLE", &url).send_with_digest_auth("user", "pass")?;
|
||||
assert_eq!(resp.status(), 200);
|
||||
Ok(())
|
||||
@@ -61,7 +79,7 @@ fn auth_check(
|
||||
|
||||
#[rstest]
|
||||
fn auth_readonly(
|
||||
#[with(&["--auth", "/@user:pass@user2:pass2", "-A"])] server: TestServer,
|
||||
#[with(&["--auth", "user:pass@/:rw|user2:pass2@/", "-A"])] server: TestServer,
|
||||
) -> Result<(), Error> {
|
||||
let url = format!("{}index.html", server.url());
|
||||
let resp = fetch!(b"GET", &url).send()?;
|
||||
@@ -72,13 +90,13 @@ fn auth_readonly(
|
||||
let resp = fetch!(b"PUT", &url)
|
||||
.body(b"abc".to_vec())
|
||||
.send_with_digest_auth("user2", "pass2")?;
|
||||
assert_eq!(resp.status(), 401);
|
||||
assert_eq!(resp.status(), 403);
|
||||
Ok(())
|
||||
}
|
||||
|
||||
#[rstest]
|
||||
fn auth_nest(
|
||||
#[with(&["--auth", "/@user:pass@user2:pass2", "--auth", "/dir1@user3:pass3", "-A"])]
|
||||
#[with(&["--auth", "user:pass@/:rw|user2:pass2@/", "--auth", "user3:pass3@/dir1:rw", "-A"])]
|
||||
server: TestServer,
|
||||
) -> Result<(), Error> {
|
||||
let url = format!("{}dir1/file1", server.url());
|
||||
@@ -97,7 +115,8 @@ fn auth_nest(
|
||||
|
||||
#[rstest]
|
||||
fn auth_nest_share(
|
||||
#[with(&["--auth", "/@user:pass@*", "--auth", "/dir1@user3:pass3", "-A"])] server: TestServer,
|
||||
#[with(&["--auth", "@/", "--auth", "user:pass@/:rw", "--auth", "user3:pass3@/dir1:rw", "-A"])]
|
||||
server: TestServer,
|
||||
) -> Result<(), Error> {
|
||||
let url = format!("{}index.html", server.url());
|
||||
let resp = fetch!(b"GET", &url).send()?;
|
||||
@@ -106,8 +125,8 @@ fn auth_nest_share(
|
||||
}
|
||||
|
||||
#[rstest]
|
||||
#[case(server(&["--auth", "/@user:pass", "--auth-method", "basic", "-A"]), "user", "pass")]
|
||||
#[case(server(&["--auth", "/@u1:p1", "--auth-method", "basic", "-A"]), "u1", "p1")]
|
||||
#[case(server(&["--auth", "user:pass@/:rw", "--auth-method", "basic", "-A"]), "user", "pass")]
|
||||
#[case(server(&["--auth", "u1:p1@/:rw", "--auth-method", "basic", "-A"]), "u1", "p1")]
|
||||
fn auth_basic(
|
||||
#[case] server: TestServer,
|
||||
#[case] user: &str,
|
||||
@@ -126,7 +145,8 @@ fn auth_basic(
|
||||
|
||||
#[rstest]
|
||||
fn auth_webdav_move(
|
||||
#[with(&["--auth", "/@user:pass@*", "--auth", "/dir1@user3:pass3", "-A"])] server: TestServer,
|
||||
#[with(&["--auth", "user:pass@/:rw", "--auth", "user3:pass3@/dir1:rw", "-A"])]
|
||||
server: TestServer,
|
||||
) -> Result<(), Error> {
|
||||
let origin_url = format!("{}dir1/test.html", server.url());
|
||||
let new_url = format!("{}test2.html", server.url());
|
||||
@@ -139,7 +159,8 @@ fn auth_webdav_move(
|
||||
|
||||
#[rstest]
|
||||
fn auth_webdav_copy(
|
||||
#[with(&["--auth", "/@user:pass@*", "--auth", "/dir1@user3:pass3", "-A"])] server: TestServer,
|
||||
#[with(&["--auth", "user:pass@/:rw", "--auth", "user3:pass3@/dir1:rw", "-A"])]
|
||||
server: TestServer,
|
||||
) -> Result<(), Error> {
|
||||
let origin_url = format!("{}dir1/test.html", server.url());
|
||||
let new_url = format!("{}test2.html", server.url());
|
||||
@@ -152,7 +173,7 @@ fn auth_webdav_copy(
|
||||
|
||||
#[rstest]
|
||||
fn auth_path_prefix(
|
||||
#[with(&["--auth", "/@user:pass", "--path-prefix", "xyz", "-A"])] server: TestServer,
|
||||
#[with(&["--auth", "user:pass@/:rw", "--path-prefix", "xyz", "-A"])] server: TestServer,
|
||||
) -> Result<(), Error> {
|
||||
let url = format!("{}xyz/index.html", server.url());
|
||||
let resp = fetch!(b"GET", &url).send()?;
|
||||
@@ -161,3 +182,53 @@ fn auth_path_prefix(
|
||||
assert_eq!(resp.status(), 200);
|
||||
Ok(())
|
||||
}
|
||||
|
||||
#[rstest]
|
||||
fn auth_partial_index(
|
||||
#[with(&["--auth", "user:pass@/dir1:rw,/dir2:rw", "-A"])] server: TestServer,
|
||||
) -> Result<(), Error> {
|
||||
let resp = fetch!(b"GET", server.url()).send_with_digest_auth("user", "pass")?;
|
||||
assert_eq!(resp.status(), 200);
|
||||
let paths = utils::retrieve_index_paths(&resp.text()?);
|
||||
assert_eq!(paths, IndexSet::from(["dir1/".into(), "dir2/".into()]));
|
||||
let resp = fetch!(b"GET", format!("{}?q={}", server.url(), "test.html"))
|
||||
.send_with_digest_auth("user", "pass")?;
|
||||
assert_eq!(resp.status(), 200);
|
||||
let paths = utils::retrieve_index_paths(&resp.text()?);
|
||||
assert_eq!(
|
||||
paths,
|
||||
IndexSet::from(["dir1/test.html".into(), "dir2/test.html".into()])
|
||||
);
|
||||
Ok(())
|
||||
}
|
||||
|
||||
#[rstest]
|
||||
fn no_auth_propfind_dir(
|
||||
#[with(&["--auth", "user:pass@/:rw", "--auth", "@/dir-assets", "-A"])] server: TestServer,
|
||||
) -> Result<(), Error> {
|
||||
let resp = fetch!(b"PROPFIND", server.url()).send()?;
|
||||
assert_eq!(resp.status(), 207);
|
||||
let body = resp.text()?;
|
||||
assert!(body.contains("<D:href>/dir-assets/</D:href>"));
|
||||
assert!(body.contains("<D:href>/dir1/</D:href>"));
|
||||
Ok(())
|
||||
}
|
||||
|
||||
#[rstest]
|
||||
fn auth_data(
|
||||
#[with(&["--auth", "user:pass@/:rw|@/", "-A", "--auth-method", "basic"])] server: TestServer,
|
||||
) -> Result<(), Error> {
|
||||
let resp = reqwest::blocking::get(server.url())?;
|
||||
let content = resp.text()?;
|
||||
let json = utils::retrive_json(&content).unwrap();
|
||||
assert_eq!(json["allow_delete"], serde_json::Value::Bool(false));
|
||||
assert_eq!(json["allow_upload"], serde_json::Value::Bool(false));
|
||||
let resp = fetch!(b"GET", server.url())
|
||||
.basic_auth("user", Some("pass"))
|
||||
.send()?;
|
||||
let content = resp.text()?;
|
||||
let json = utils::retrive_json(&content).unwrap();
|
||||
assert_eq!(json["allow_delete"], serde_json::Value::Bool(true));
|
||||
assert_eq!(json["allow_upload"], serde_json::Value::Bool(true));
|
||||
Ok(())
|
||||
}
|
||||
|
||||
@@ -19,15 +19,15 @@ fn cors(#[with(&["--enable-cors"])] server: TestServer) -> Result<(), Error> {
|
||||
);
|
||||
assert_eq!(
|
||||
resp.headers().get("access-control-allow-methods").unwrap(),
|
||||
"GET,HEAD,PUT,OPTIONS,DELETE,PROPFIND,COPY,MOVE"
|
||||
"*"
|
||||
);
|
||||
assert_eq!(
|
||||
resp.headers().get("access-control-allow-headers").unwrap(),
|
||||
"Authorization,Destination,Range,Content-Type"
|
||||
"Authorization,*"
|
||||
);
|
||||
assert_eq!(
|
||||
resp.headers().get("access-control-expose-headers").unwrap(),
|
||||
"WWW-Authenticate,Content-Range,Accept-Ranges,Content-Disposition"
|
||||
"Authorization,*"
|
||||
);
|
||||
Ok(())
|
||||
}
|
||||
|
||||
@@ -46,7 +46,7 @@ pub fn tmpdir() -> TempDir {
|
||||
let tmpdir = assert_fs::TempDir::new().expect("Couldn't create a temp dir for tests");
|
||||
for file in FILES {
|
||||
if *file == BIN_FILE {
|
||||
tmpdir.child(file).write_binary(b"bin\0\0123").unwrap();
|
||||
tmpdir.child(file).write_binary(b"bin\0\x00123").unwrap();
|
||||
} else {
|
||||
tmpdir
|
||||
.child(file)
|
||||
@@ -58,7 +58,7 @@ pub fn tmpdir() -> TempDir {
|
||||
if *directory == DIR_ASSETS {
|
||||
tmpdir
|
||||
.child(format!("{}{}", directory, "index.html"))
|
||||
.write_str("__ASSERTS_PREFIX__index.js;DATA = __INDEX_DATA__")
|
||||
.write_str("__ASSETS_PREFIX__index.js;DATA = __INDEX_DATA__")
|
||||
.unwrap();
|
||||
} else {
|
||||
for file in FILES {
|
||||
@@ -68,7 +68,7 @@ pub fn tmpdir() -> TempDir {
|
||||
if *file == BIN_FILE {
|
||||
tmpdir
|
||||
.child(format!("{directory}{file}"))
|
||||
.write_binary(b"bin\0\0123")
|
||||
.write_binary(b"bin\0\x00123")
|
||||
.unwrap();
|
||||
} else {
|
||||
tmpdir
|
||||
|
||||
@@ -123,6 +123,15 @@ fn get_dir_search3(#[with(&["-A"])] server: TestServer) -> Result<(), Error> {
|
||||
Ok(())
|
||||
}
|
||||
|
||||
#[rstest]
|
||||
fn get_dir_search4(#[with(&["-A"])] server: TestServer) -> Result<(), Error> {
|
||||
let resp = reqwest::blocking::get(format!("{}dir1?q=dir1&simple", server.url()))?;
|
||||
assert_eq!(resp.status(), 200);
|
||||
let text = resp.text().unwrap();
|
||||
assert!(text.is_empty());
|
||||
Ok(())
|
||||
}
|
||||
|
||||
#[rstest]
|
||||
fn head_dir_search(#[with(&["-A"])] server: TestServer) -> Result<(), Error> {
|
||||
let resp = fetch!(b"HEAD", format!("{}?q={}", server.url(), "test.html")).send()?;
|
||||
@@ -184,6 +193,17 @@ fn get_file_404(server: TestServer) -> Result<(), Error> {
|
||||
Ok(())
|
||||
}
|
||||
|
||||
#[rstest]
|
||||
fn get_file_emoji_path(server: TestServer) -> Result<(), Error> {
|
||||
let resp = reqwest::blocking::get(format!("{}{BIN_FILE}", server.url()))?;
|
||||
assert_eq!(resp.status(), 200);
|
||||
assert_eq!(
|
||||
resp.headers().get("content-disposition").unwrap(),
|
||||
"inline; filename=\"😀.bin\"; filename*=UTF-8''%F0%9F%98%80.bin"
|
||||
);
|
||||
Ok(())
|
||||
}
|
||||
|
||||
#[rstest]
|
||||
fn get_file_edit(server: TestServer) -> Result<(), Error> {
|
||||
let resp = fetch!(b"GET", format!("{}index.html?edit", server.url())).send()?;
|
||||
|
||||
@@ -11,8 +11,8 @@ use std::io::Read;
|
||||
use std::process::{Command, Stdio};
|
||||
|
||||
#[rstest]
|
||||
#[case(&["-a", "/@user:pass", "--log-format", "$remote_user"], false)]
|
||||
#[case(&["-a", "/@user:pass", "--log-format", "$remote_user", "--auth-method", "basic"], true)]
|
||||
#[case(&["-a", "user:pass@/:rw", "--log-format", "$remote_user"], false)]
|
||||
#[case(&["-a", "user:pass@/:rw", "--log-format", "$remote_user", "--auth-method", "basic"], true)]
|
||||
fn log_remote_user(
|
||||
tmpdir: TempDir,
|
||||
port: u16,
|
||||
|
||||
@@ -53,7 +53,7 @@ fn path_prefix_single_file(tmpdir: TempDir, port: u16, #[case] file: &str) -> Re
|
||||
let resp = reqwest::blocking::get(format!("http://localhost:{port}/xyz/index.html"))?;
|
||||
assert_eq!(resp.text()?, "This is index.html");
|
||||
let resp = reqwest::blocking::get(format!("http://localhost:{port}"))?;
|
||||
assert_eq!(resp.status(), 404);
|
||||
assert_eq!(resp.status(), 403);
|
||||
|
||||
child.kill()?;
|
||||
Ok(())
|
||||
|
||||
@@ -59,7 +59,8 @@ pub fn encode_uri(v: &str) -> String {
|
||||
parts.join("/")
|
||||
}
|
||||
|
||||
fn retrive_json(content: &str) -> Option<Value> {
|
||||
#[allow(dead_code)]
|
||||
pub fn retrive_json(content: &str) -> Option<Value> {
|
||||
let lines: Vec<&str> = content.lines().collect();
|
||||
let line = lines.iter().find(|v| v.contains("DATA ="))?;
|
||||
let line_col = line.find("DATA =").unwrap() + 6;
|
||||
|
||||
Reference in New Issue
Block a user