mirror of
https://github.com/sigoden/dufs.git
synced 2026-04-09 17:13:02 +03:00
Compare commits
2 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
22cf74e3c0 | ||
|
|
6a6ac37be4 |
8
.github/ISSUE_TEMPLATE/bug_report.md
vendored
8
.github/ISSUE_TEMPLATE/bug_report.md
vendored
@@ -7,12 +7,6 @@ about: Create a report to help us improve
|
||||
|
||||
<!-- A clear and concise description of what the bug is. -->
|
||||
|
||||
**Configuration**
|
||||
|
||||
<!-- The dufs command-line arguments or configuration -->
|
||||
|
||||
<!-- If the problems are related to auth/perm, please conceal only the user:pass, but do not hide the entire `auth` configuration. -->
|
||||
|
||||
**Log**
|
||||
|
||||
The dufs log is crucial for locating the problem, so please do not omit it.
|
||||
@@ -21,4 +15,4 @@ The dufs log is crucial for locating the problem, so please do not omit it.
|
||||
- Dufs version:
|
||||
- Browser/Webdav info:
|
||||
- OS info:
|
||||
- Proxy server: e.g. nginx, cloudflare
|
||||
- Proxy server: e.g. nginx, cloudflare
|
||||
|
||||
2
.github/workflows/ci.yaml
vendored
2
.github/workflows/ci.yaml
vendored
@@ -29,7 +29,7 @@ jobs:
|
||||
RUSTFLAGS: --deny warnings
|
||||
|
||||
steps:
|
||||
- uses: actions/checkout@v4
|
||||
- uses: actions/checkout@v3
|
||||
|
||||
- name: Install Rust Toolchain Components
|
||||
uses: dtolnay/rust-toolchain@stable
|
||||
|
||||
42
.github/workflows/release.yaml
vendored
42
.github/workflows/release.yaml
vendored
@@ -54,13 +54,28 @@ jobs:
|
||||
os: ubuntu-latest
|
||||
use-cross: true
|
||||
cargo-flags: ""
|
||||
|
||||
- target: mips-unknown-linux-musl
|
||||
os: ubuntu-latest
|
||||
use-cross: true
|
||||
cargo-flags: "--no-default-features"
|
||||
- target: mipsel-unknown-linux-musl
|
||||
os: ubuntu-latest
|
||||
use-cross: true
|
||||
cargo-flags: "--no-default-features"
|
||||
- target: mips64-unknown-linux-gnuabi64
|
||||
os: ubuntu-latest
|
||||
use-cross: true
|
||||
cargo-flags: "--no-default-features"
|
||||
- target: mips64el-unknown-linux-gnuabi64
|
||||
os: ubuntu-latest
|
||||
use-cross: true
|
||||
cargo-flags: "--no-default-features"
|
||||
runs-on: ${{matrix.os}}
|
||||
env:
|
||||
BUILD_CMD: cargo
|
||||
|
||||
steps:
|
||||
- uses: actions/checkout@v4
|
||||
- uses: actions/checkout@v3
|
||||
|
||||
- name: Check Tag
|
||||
id: check-tag
|
||||
@@ -79,18 +94,20 @@ jobs:
|
||||
uses: dtolnay/rust-toolchain@stable
|
||||
with:
|
||||
targets: ${{ matrix.target }}
|
||||
# Since rust 1.72, mips platforms are tier 3
|
||||
toolchain: 1.71
|
||||
|
||||
- name: Install cross
|
||||
if: matrix.use-cross
|
||||
uses: taiki-e/install-action@v2
|
||||
with:
|
||||
tool: cross
|
||||
|
||||
|
||||
- name: Overwrite build command env variable
|
||||
if: matrix.use-cross
|
||||
shell: bash
|
||||
run: echo "BUILD_CMD=cross" >> $GITHUB_ENV
|
||||
|
||||
|
||||
- name: Show Version Information (Rust, cargo, GCC)
|
||||
shell: bash
|
||||
run: |
|
||||
@@ -138,12 +155,14 @@ jobs:
|
||||
fi
|
||||
|
||||
- name: Publish Archive
|
||||
uses: softprops/action-gh-release@v2
|
||||
uses: softprops/action-gh-release@v1
|
||||
if: ${{ startsWith(github.ref, 'refs/tags/') }}
|
||||
with:
|
||||
draft: false
|
||||
files: ${{ steps.package.outputs.archive }}
|
||||
prerelease: ${{ steps.check-tag.outputs.rc == 'true' }}
|
||||
env:
|
||||
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
|
||||
|
||||
docker:
|
||||
name: Publish to Docker Hub
|
||||
@@ -152,18 +171,17 @@ jobs:
|
||||
needs: release
|
||||
steps:
|
||||
- name: Set up QEMU
|
||||
uses: docker/setup-qemu-action@v3
|
||||
uses: docker/setup-qemu-action@v2
|
||||
- name: Set up Docker Buildx
|
||||
uses: docker/setup-buildx-action@v3
|
||||
uses: docker/setup-buildx-action@v2
|
||||
- name: Login to DockerHub
|
||||
uses: docker/login-action@v3
|
||||
uses: docker/login-action@v2
|
||||
with:
|
||||
username: ${{ github.repository_owner }}
|
||||
username: ${{ secrets.DOCKERHUB_USERNAME }}
|
||||
password: ${{ secrets.DOCKERHUB_TOKEN }}
|
||||
- name: Build and push
|
||||
uses: docker/build-push-action@v5
|
||||
uses: docker/build-push-action@v4
|
||||
with:
|
||||
file: Dockerfile-release
|
||||
build-args: |
|
||||
REPO=${{ github.repository }}
|
||||
VER=${{ github.ref_name }}
|
||||
@@ -181,7 +199,7 @@ jobs:
|
||||
runs-on: ubuntu-latest
|
||||
needs: release
|
||||
steps:
|
||||
- uses: actions/checkout@v4
|
||||
- uses: actions/checkout@v3
|
||||
|
||||
- uses: dtolnay/rust-toolchain@stable
|
||||
|
||||
|
||||
21
CHANGELOG.md
21
CHANGELOG.md
@@ -2,25 +2,6 @@
|
||||
|
||||
All notable changes to this project will be documented in this file.
|
||||
|
||||
## [0.41.0] - 2024-05-22
|
||||
|
||||
### Bug Fixes
|
||||
|
||||
- Timestamp format of getlastmodified in dav xml ([#366](https://github.com/sigoden/dufs/issues/366))
|
||||
- Strange issue that occurs only on Microsoft WebDAV ([#382](https://github.com/sigoden/dufs/issues/382))
|
||||
- Head div overlap main contents when wrap ([#386](https://github.com/sigoden/dufs/issues/386))
|
||||
|
||||
### Features
|
||||
|
||||
- Tls handshake timeout ([#368](https://github.com/sigoden/dufs/issues/368))
|
||||
- Add api to get the hash of a file ([#375](https://github.com/sigoden/dufs/issues/375))
|
||||
- Add log-file option ([#383](https://github.com/sigoden/dufs/issues/383))
|
||||
|
||||
### Refactor
|
||||
|
||||
- Digest_auth related tests ([#372](https://github.com/sigoden/dufs/issues/372))
|
||||
- Add fixed-width numerals to date and size on file list page ([#378](https://github.com/sigoden/dufs/issues/378))
|
||||
|
||||
## [0.40.0] - 2024-02-13
|
||||
|
||||
### Bug Fixes
|
||||
@@ -104,7 +85,7 @@ All notable changes to this project will be documented in this file.
|
||||
|
||||
- Remove one clone on `assets_prefix` ([#270](https://github.com/sigoden/dufs/issues/270))
|
||||
- Optimize tests
|
||||
- Improve code quality ([#282](https://github.com/sigoden/dufs/issues/282))
|
||||
- Improve code quanity ([#282](https://github.com/sigoden/dufs/issues/282))
|
||||
|
||||
## [0.36.0] - 2023-08-24
|
||||
|
||||
|
||||
814
Cargo.lock
generated
814
Cargo.lock
generated
File diff suppressed because it is too large
Load Diff
27
Cargo.toml
27
Cargo.toml
@@ -1,6 +1,6 @@
|
||||
[package]
|
||||
name = "dufs"
|
||||
version = "0.41.0"
|
||||
version = "0.40.0"
|
||||
edition = "2021"
|
||||
authors = ["sigoden <sigoden@gmail.com>"]
|
||||
description = "Dufs is a distinctive utility file server"
|
||||
@@ -11,28 +11,28 @@ categories = ["command-line-utilities", "web-programming::http-server"]
|
||||
keywords = ["static", "file", "server", "webdav", "cli"]
|
||||
|
||||
[dependencies]
|
||||
clap = { version = "4.5", features = ["wrap_help", "env"] }
|
||||
clap_complete = "4.5"
|
||||
clap = { version = "~4.4", features = ["wrap_help", "env"] }
|
||||
clap_complete = "~4.4"
|
||||
chrono = { version = "0.4", default-features = false, features = ["clock"] }
|
||||
tokio = { version = "1", features = ["rt-multi-thread", "macros", "fs", "io-util", "signal"]}
|
||||
tokio-util = { version = "0.7", features = ["io-util", "compat"] }
|
||||
hyper = { version = "1", features = ["http1", "server"] }
|
||||
hyper = { version = "1.0", features = ["http1", "server"] }
|
||||
percent-encoding = "2.3"
|
||||
serde = { version = "1", features = ["derive"] }
|
||||
serde_json = "1"
|
||||
futures-util = { version = "0.3", default-features = false, features = ["alloc"] }
|
||||
async_zip = { version = "0.0.17", default-features = false, features = ["deflate", "bzip2", "xz", "chrono", "tokio"] }
|
||||
async_zip = { version = "0.0.16", default-features = false, features = ["deflate", "bzip2", "xz", "chrono", "tokio"] }
|
||||
headers = "0.4"
|
||||
mime_guess = "2.0"
|
||||
if-addrs = "0.12"
|
||||
if-addrs = "0.11"
|
||||
rustls-pemfile = { version = "2.0", optional = true }
|
||||
tokio-rustls = { version = "0.26", optional = true, default-features = false, features = ["ring", "tls12"]}
|
||||
tokio-rustls = { version = "0.25", optional = true }
|
||||
md5 = "0.7"
|
||||
lazy_static = "1.4"
|
||||
uuid = { version = "1.7", features = ["v4", "fast-rng"] }
|
||||
urlencoding = "2.1"
|
||||
xml-rs = "0.8"
|
||||
log = { version = "0.4", features = ["std"] }
|
||||
log = "0.4"
|
||||
socket2 = "0.5"
|
||||
async-stream = "0.3"
|
||||
walkdir = "2.3"
|
||||
@@ -45,14 +45,13 @@ glob = "0.3"
|
||||
indexmap = "2.2"
|
||||
serde_yaml = "0.9"
|
||||
sha-crypt = "0.5"
|
||||
base64 = "0.22"
|
||||
base64 = "0.21"
|
||||
smart-default = "0.7"
|
||||
rustls-pki-types = "1.2"
|
||||
hyper-util = { version = "0.1", features = ["server-auto", "tokio"] }
|
||||
http-body-util = "0.1"
|
||||
bytes = "1.5"
|
||||
pin-project-lite = "0.2"
|
||||
sha2 = "0.10.8"
|
||||
|
||||
[features]
|
||||
default = ["tls"]
|
||||
@@ -60,14 +59,14 @@ tls = ["rustls-pemfile", "tokio-rustls"]
|
||||
|
||||
[dev-dependencies]
|
||||
assert_cmd = "2"
|
||||
reqwest = { version = "0.12", features = ["blocking", "multipart", "rustls-tls"], default-features = false }
|
||||
reqwest = { version = "0.11", features = ["blocking", "multipart", "rustls-tls"], default-features = false }
|
||||
assert_fs = "1"
|
||||
port_check = "0.2"
|
||||
rstest = "0.19"
|
||||
port_check = "0.1"
|
||||
rstest = "0.18"
|
||||
regex = "1"
|
||||
url = "2"
|
||||
diqwest = { version = "2.0", features = ["blocking"], default-features = false }
|
||||
predicates = "3"
|
||||
digest_auth = "0.3.1"
|
||||
|
||||
[profile.release]
|
||||
opt-level = 3
|
||||
|
||||
23
Dockerfile
23
Dockerfile
@@ -1,12 +1,17 @@
|
||||
FROM --platform=linux/amd64 messense/rust-musl-cross:x86_64-musl AS amd64
|
||||
COPY . .
|
||||
RUN cargo install --path . --root /
|
||||
|
||||
FROM --platform=linux/amd64 messense/rust-musl-cross:aarch64-musl AS arm64
|
||||
COPY . .
|
||||
RUN cargo install --path . --root /
|
||||
|
||||
FROM ${TARGETARCH} AS builder
|
||||
FROM alpine as builder
|
||||
ARG REPO VER TARGETPLATFORM
|
||||
RUN if [ "$TARGETPLATFORM" = "linux/amd64" ]; then \
|
||||
TARGET="x86_64-unknown-linux-musl"; \
|
||||
elif [ "$TARGETPLATFORM" = "linux/arm64" ]; then \
|
||||
TARGET="aarch64-unknown-linux-musl"; \
|
||||
elif [ "$TARGETPLATFORM" = "linux/386" ]; then \
|
||||
TARGET="i686-unknown-linux-musl"; \
|
||||
elif [ "$TARGETPLATFORM" = "linux/arm/v7" ]; then \
|
||||
TARGET="armv7-unknown-linux-musleabihf"; \
|
||||
fi && \
|
||||
wget https://github.com/${REPO}/releases/download/${VER}/dufs-${VER}-${TARGET}.tar.gz && \
|
||||
tar -xf dufs-${VER}-${TARGET}.tar.gz && \
|
||||
mv dufs /bin/
|
||||
|
||||
FROM scratch
|
||||
COPY --from=builder /bin/dufs /bin/dufs
|
||||
|
||||
@@ -1,19 +0,0 @@
|
||||
FROM alpine as builder
|
||||
ARG REPO VER TARGETPLATFORM
|
||||
RUN if [ "$TARGETPLATFORM" = "linux/amd64" ]; then \
|
||||
TARGET="x86_64-unknown-linux-musl"; \
|
||||
elif [ "$TARGETPLATFORM" = "linux/arm64" ]; then \
|
||||
TARGET="aarch64-unknown-linux-musl"; \
|
||||
elif [ "$TARGETPLATFORM" = "linux/386" ]; then \
|
||||
TARGET="i686-unknown-linux-musl"; \
|
||||
elif [ "$TARGETPLATFORM" = "linux/arm/v7" ]; then \
|
||||
TARGET="armv7-unknown-linux-musleabihf"; \
|
||||
fi && \
|
||||
wget https://github.com/${REPO}/releases/download/${VER}/dufs-${VER}-${TARGET}.tar.gz && \
|
||||
tar -xf dufs-${VER}-${TARGET}.tar.gz && \
|
||||
mv dufs /bin/
|
||||
|
||||
FROM scratch
|
||||
COPY --from=builder /bin/dufs /bin/dufs
|
||||
STOPSIGNAL SIGINT
|
||||
ENTRYPOINT ["/bin/dufs"]
|
||||
23
README.md
23
README.md
@@ -2,7 +2,6 @@
|
||||
|
||||
[](https://github.com/sigoden/dufs/actions/workflows/ci.yaml)
|
||||
[](https://crates.io/crates/dufs)
|
||||
[](https://hub.docker.com/r/sigoden/dufs)
|
||||
|
||||
Dufs is a distinctive utility file server that supports static serving, uploading, searching, accessing control, webdav...
|
||||
|
||||
@@ -31,7 +30,7 @@ cargo install dufs
|
||||
### With docker
|
||||
|
||||
```
|
||||
docker run -v `pwd`:/data -p 5000:5000 --rm sigoden/dufs /data -A
|
||||
docker run -v `pwd`:/data -p 5000:5000 --rm -it sigoden/dufs /data -A
|
||||
```
|
||||
|
||||
### With [Homebrew](https://brew.sh)
|
||||
@@ -73,7 +72,6 @@ Options:
|
||||
--render-spa Serve SPA(Single Page Application)
|
||||
--assets <path> Set the path to the assets directory for overriding the built-in assets
|
||||
--log-format <format> Customize http log format
|
||||
--log-file <file> Specify the file to save logs to, other than stdout/stderr
|
||||
--compress <level> Set zip compress level [default: low] [possible values: none, low, medium, high]
|
||||
--completions <shell> Print shell completion script for <shell> [possible values: bash, elvish, fish, powershell, zsh]
|
||||
--tls-cert <path> Path to an SSL/TLS certificate to serve with HTTPS
|
||||
@@ -159,8 +157,7 @@ curl -T path-to-file http://127.0.0.1:5000/new-path/path-to-file
|
||||
|
||||
Download a file
|
||||
```sh
|
||||
curl http://127.0.0.1:5000/path-to-file # download the file
|
||||
curl http://127.0.0.1:5000/path-to-file?hash # retrieve the sha256 hash of the file
|
||||
curl http://127.0.0.1:5000/path-to-file
|
||||
```
|
||||
|
||||
Download a folder as zip file
|
||||
@@ -178,13 +175,13 @@ curl -X DELETE http://127.0.0.1:5000/path-to-file-or-folder
|
||||
Create a directory
|
||||
|
||||
```sh
|
||||
curl -X MKCOL http://127.0.0.1:5000/path-to-folder
|
||||
curl -X MKCOL https://127.0.0.1:5000/path-to-folder
|
||||
```
|
||||
|
||||
Move the file/folder to the new path
|
||||
|
||||
```sh
|
||||
curl -X MOVE http://127.0.0.1:5000/path -H "Destination: http://127.0.0.1:5000/new-path"
|
||||
curl -X MOVE https://127.0.0.1:5000/path -H "Destination: https://127.0.0.1:5000/new-path"
|
||||
```
|
||||
|
||||
List/search directory contents
|
||||
@@ -249,12 +246,12 @@ Create hashed password
|
||||
```
|
||||
$ mkpasswd -m sha-512 -s
|
||||
Password: 123456
|
||||
$6$tWMB51u6Kb2ui3wd$5gVHP92V9kZcMwQeKTjyTRgySsYJu471Jb1I6iHQ8iZ6s07GgCIO69KcPBRuwPE5tDq05xMAzye0NxVKuJdYs/
|
||||
$6$qCAVUG7yn7t/hH4d$BWm8r5MoDywNmDP/J3V2S2a6flmKHC1IpblfoqZfuK.LtLBZ0KFXP9QIfJP8RqL8MCw4isdheoAMTuwOz.pAO/
|
||||
```
|
||||
|
||||
Use hashed password
|
||||
```
|
||||
dufs -a 'admin:$6$tWMB51u6Kb2ui3wd$5gVHP92V9kZcMwQeKTjyTRgySsYJu471Jb1I6iHQ8iZ6s07GgCIO69KcPBRuwPE5tDq05xMAzye0NxVKuJdYs/@/:rw'
|
||||
dufs -a 'admin:$6$qCAVUG7yn7t/hH4d$BWm8r5MoDywNmDP/J3V2S2a6flmKHC1IpblfoqZfuK.LtLBZ0KFXP9QIfJP8RqL8MCw4isdheoAMTuwOz.pAO/@/:rw'
|
||||
```
|
||||
|
||||
Two important things for hashed passwords:
|
||||
@@ -330,7 +327,7 @@ All options can be set using environment variables prefixed with `DUFS_`.
|
||||
--config <file> DUFS_CONFIG=config.yaml
|
||||
-b, --bind <addrs> DUFS_BIND=0.0.0.0
|
||||
-p, --port <port> DUFS_PORT=5000
|
||||
--path-prefix <path> DUFS_PATH_PREFIX=/dufs
|
||||
--path-prefix <path> DUFS_PATH_PREFIX=/static
|
||||
--hidden <value> DUFS_HIDDEN=tmp,*.log,*.lock
|
||||
-a, --auth <rules> DUFS_AUTH="admin:admin@/:rw|@/"
|
||||
-A, --allow-all DUFS_ALLOW_ALL=true
|
||||
@@ -343,10 +340,9 @@ All options can be set using environment variables prefixed with `DUFS_`.
|
||||
--render-index DUFS_RENDER_INDEX=true
|
||||
--render-try-index DUFS_RENDER_TRY_INDEX=true
|
||||
--render-spa DUFS_RENDER_SPA=true
|
||||
--assets <path> DUFS_ASSETS=./assets
|
||||
--assets <path> DUFS_ASSETS=/assets
|
||||
--log-format <format> DUFS_LOG_FORMAT=""
|
||||
--log-file <file> DUFS_LOG_FILE=./dufs.log
|
||||
--compress <compress> DUFS_COMPRESS=low
|
||||
--compress <compress> DUFS_COMPRESS="low"
|
||||
--tls-cert <path> DUFS_TLS_CERT=cert.pem
|
||||
--tls-key <path> DUFS_TLS_KEY=key.pem
|
||||
```
|
||||
@@ -382,7 +378,6 @@ render-try-index: true
|
||||
render-spa: true
|
||||
assets: ./assets/
|
||||
log-format: '$remote_addr "$request" $status $http_user_agent'
|
||||
log-file: ./dufs.log
|
||||
compress: low
|
||||
tls-cert: tests/data/cert.pem
|
||||
tls-key: tests/data/key_pkcs1.pem
|
||||
|
||||
0
assets/favicon.ico
Normal file → Executable file
0
assets/favicon.ico
Normal file → Executable file
|
Before Width: | Height: | Size: 9.1 KiB After Width: | Height: | Size: 9.1 KiB |
@@ -6,7 +6,7 @@ html {
|
||||
|
||||
body {
|
||||
/* prevent premature breadcrumb wrapping on mobile */
|
||||
min-width: 538px;
|
||||
min-width: 500px;
|
||||
margin: 0;
|
||||
}
|
||||
|
||||
@@ -19,15 +19,14 @@ body {
|
||||
flex-wrap: wrap;
|
||||
align-items: center;
|
||||
padding: 0.6em 1em;
|
||||
position: sticky;
|
||||
top: 0;
|
||||
position: fixed;
|
||||
width: 100%;
|
||||
background-color: white;
|
||||
}
|
||||
|
||||
.breadcrumb {
|
||||
font-size: 1.25em;
|
||||
padding-right: 0.6em;
|
||||
word-break: break-all;
|
||||
}
|
||||
|
||||
.breadcrumb>a {
|
||||
@@ -109,7 +108,7 @@ body {
|
||||
}
|
||||
|
||||
.main {
|
||||
padding: 0 1em;
|
||||
padding: 3.3em 1em 0;
|
||||
}
|
||||
|
||||
.empty-folder {
|
||||
@@ -154,20 +153,18 @@ body {
|
||||
.paths-table .cell-actions {
|
||||
width: 90px;
|
||||
display: flex;
|
||||
padding-left: 0.5em;
|
||||
padding-left: 0.6em;
|
||||
}
|
||||
|
||||
.paths-table .cell-mtime {
|
||||
width: 120px;
|
||||
padding-left: 0.5em;
|
||||
font-variant-numeric: tabular-nums;
|
||||
padding-left: 0.6em;
|
||||
}
|
||||
|
||||
.paths-table .cell-size {
|
||||
text-align: right;
|
||||
width: 70px;
|
||||
padding-left: 0.5em;
|
||||
font-variant-numeric: tabular-nums;
|
||||
padding-left: 0.6em;
|
||||
}
|
||||
|
||||
.path svg {
|
||||
@@ -189,7 +186,7 @@ body {
|
||||
display: block;
|
||||
text-decoration: none;
|
||||
max-width: calc(100vw - 375px);
|
||||
min-width: 170px;
|
||||
min-width: 200px;
|
||||
}
|
||||
|
||||
.path a:hover {
|
||||
|
||||
@@ -114,6 +114,7 @@ function ready() {
|
||||
document.querySelector(".index-page").classList.remove("hidden");
|
||||
|
||||
setupIndexPage();
|
||||
|
||||
} else if (DATA.kind == "Edit") {
|
||||
document.title = `Edit ${DATA.href} - Dufs`;
|
||||
document.querySelector(".editor-page").classList.remove("hidden");;
|
||||
@@ -885,12 +886,12 @@ async function assertResOK(res) {
|
||||
}
|
||||
|
||||
function getEncoding(contentType) {
|
||||
const charset = contentType?.split(";")[1];
|
||||
if (/charset/i.test(charset)) {
|
||||
let encoding = charset.split("=")[1];
|
||||
if (encoding) {
|
||||
return encoding.toLowerCase()
|
||||
const charset = contentType?.split(";")[1];
|
||||
if (/charset/i.test(charset)) {
|
||||
let encoding = charset.split("=")[1];
|
||||
if (encoding) {
|
||||
return encoding.toLowerCase()
|
||||
}
|
||||
}
|
||||
}
|
||||
return 'utf-8'
|
||||
return 'utf-8'
|
||||
}
|
||||
|
||||
18
src/args.rs
18
src/args.rs
@@ -197,15 +197,6 @@ pub fn build_cli() -> Command {
|
||||
.value_name("format")
|
||||
.help("Customize http log format"),
|
||||
)
|
||||
.arg(
|
||||
Arg::new("log-file")
|
||||
.env("DUFS_LOG_FILE")
|
||||
.hide_env(true)
|
||||
.long("log-file")
|
||||
.value_name("file")
|
||||
.value_parser(value_parser!(PathBuf))
|
||||
.help("Specify the file to save logs to, other than stdout/stderr"),
|
||||
)
|
||||
.arg(
|
||||
Arg::new("compress")
|
||||
.env("DUFS_COMPRESS")
|
||||
@@ -289,7 +280,6 @@ pub struct Args {
|
||||
#[serde(deserialize_with = "deserialize_log_http")]
|
||||
#[serde(rename = "log-format")]
|
||||
pub http_logger: HttpLogger,
|
||||
pub log_file: Option<PathBuf>,
|
||||
pub compress: Compress,
|
||||
pub tls_cert: Option<PathBuf>,
|
||||
pub tls_key: Option<PathBuf>,
|
||||
@@ -311,7 +301,7 @@ impl Args {
|
||||
}
|
||||
|
||||
if let Some(path) = matches.get_one::<PathBuf>("serve-path") {
|
||||
args.serve_path.clone_from(path)
|
||||
args.serve_path = path.clone()
|
||||
}
|
||||
|
||||
args.serve_path = Self::sanitize_path(args.serve_path)?;
|
||||
@@ -327,7 +317,7 @@ impl Args {
|
||||
|
||||
args.path_is_file = args.serve_path.metadata()?.is_file();
|
||||
if let Some(path_prefix) = matches.get_one::<String>("path-prefix") {
|
||||
args.path_prefix.clone_from(path_prefix)
|
||||
args.path_prefix = path_prefix.clone();
|
||||
}
|
||||
args.path_prefix = args.path_prefix.trim_matches('/').to_string();
|
||||
|
||||
@@ -402,10 +392,6 @@ impl Args {
|
||||
args.http_logger = log_format.parse()?;
|
||||
}
|
||||
|
||||
if let Some(log_file) = matches.get_one::<PathBuf>("log-file") {
|
||||
args.log_file = Some(log_file.clone());
|
||||
}
|
||||
|
||||
if let Some(compress) = matches.get_one::<Compress>("compress") {
|
||||
args.compress = *compress;
|
||||
}
|
||||
|
||||
@@ -100,7 +100,6 @@ impl AccessControl {
|
||||
path: &str,
|
||||
method: &Method,
|
||||
authorization: Option<&HeaderValue>,
|
||||
guard_options: bool,
|
||||
) -> (Option<String>, Option<AccessPaths>) {
|
||||
if let Some(authorization) = authorization {
|
||||
if let Some(user) = get_auth_user(authorization) {
|
||||
@@ -117,7 +116,7 @@ impl AccessControl {
|
||||
return (None, None);
|
||||
}
|
||||
|
||||
if !guard_options && method == Method::OPTIONS {
|
||||
if method == Method::OPTIONS {
|
||||
return (None, Some(AccessPaths::new(AccessPerm::ReadOnly)));
|
||||
}
|
||||
|
||||
|
||||
@@ -1,14 +1,8 @@
|
||||
use anyhow::{Context, Result};
|
||||
use chrono::{Local, SecondsFormat};
|
||||
use log::{Level, LevelFilter, Metadata, Record};
|
||||
use std::fs::{File, OpenOptions};
|
||||
use std::io::Write;
|
||||
use std::path::PathBuf;
|
||||
use std::sync::Mutex;
|
||||
use log::{Level, Metadata, Record};
|
||||
use log::{LevelFilter, SetLoggerError};
|
||||
|
||||
struct SimpleLogger {
|
||||
file: Option<Mutex<File>>,
|
||||
}
|
||||
struct SimpleLogger;
|
||||
|
||||
impl log::Log for SimpleLogger {
|
||||
fn enabled(&self, metadata: &Metadata) -> bool {
|
||||
@@ -18,20 +12,10 @@ impl log::Log for SimpleLogger {
|
||||
fn log(&self, record: &Record) {
|
||||
if self.enabled(record.metadata()) {
|
||||
let timestamp = Local::now().to_rfc3339_opts(SecondsFormat::Secs, true);
|
||||
let text = format!("{} {} - {}", timestamp, record.level(), record.args());
|
||||
match &self.file {
|
||||
Some(file) => {
|
||||
if let Ok(mut file) = file.lock() {
|
||||
let _ = writeln!(file, "{text}");
|
||||
}
|
||||
}
|
||||
None => {
|
||||
if record.level() < Level::Info {
|
||||
eprintln!("{text}");
|
||||
} else {
|
||||
println!("{text}");
|
||||
}
|
||||
}
|
||||
if record.level() < Level::Info {
|
||||
eprintln!("{} {} - {}", timestamp, record.level(), record.args());
|
||||
} else {
|
||||
println!("{} {} - {}", timestamp, record.level(), record.args());
|
||||
}
|
||||
}
|
||||
}
|
||||
@@ -39,23 +23,8 @@ impl log::Log for SimpleLogger {
|
||||
fn flush(&self) {}
|
||||
}
|
||||
|
||||
pub fn init(log_file: Option<PathBuf>) -> Result<()> {
|
||||
let file = match log_file {
|
||||
None => None,
|
||||
Some(log_file) => {
|
||||
let file = OpenOptions::new()
|
||||
.create(true)
|
||||
.append(true)
|
||||
.open(&log_file)
|
||||
.with_context(|| {
|
||||
format!("Failed to open the log file at '{}'", log_file.display())
|
||||
})?;
|
||||
Some(Mutex::new(file))
|
||||
}
|
||||
};
|
||||
let logger = SimpleLogger { file };
|
||||
log::set_boxed_logger(Box::new(logger))
|
||||
.map(|_| log::set_max_level(LevelFilter::Info))
|
||||
.with_context(|| "Failed to init logger")?;
|
||||
Ok(())
|
||||
static LOGGER: SimpleLogger = SimpleLogger;
|
||||
|
||||
pub fn init() -> Result<(), SetLoggerError> {
|
||||
log::set_logger(&LOGGER).map(|()| log::set_max_level(LevelFilter::Info))
|
||||
}
|
||||
|
||||
48
src/main.rs
48
src/main.rs
@@ -29,14 +29,13 @@ use std::sync::{
|
||||
atomic::{AtomicBool, Ordering},
|
||||
Arc,
|
||||
};
|
||||
use std::time::Duration;
|
||||
use tokio::time::timeout;
|
||||
use tokio::{net::TcpListener, task::JoinHandle};
|
||||
#[cfg(feature = "tls")]
|
||||
use tokio_rustls::{rustls::ServerConfig, TlsAcceptor};
|
||||
|
||||
#[tokio::main]
|
||||
async fn main() -> Result<()> {
|
||||
logger::init().map_err(|e| anyhow!("Failed to init logger, {e}"))?;
|
||||
let cmd = build_cli();
|
||||
let matches = cmd.get_matches();
|
||||
if let Some(generator) = matches.get_one::<Shell>("completions") {
|
||||
@@ -45,7 +44,6 @@ async fn main() -> Result<()> {
|
||||
return Ok(());
|
||||
}
|
||||
let mut args = Args::parse(matches)?;
|
||||
logger::init(args.log_file.clone()).map_err(|e| anyhow!("Failed to init logger, {e}"))?;
|
||||
let (new_addrs, print_addrs) = check_addrs(&args)?;
|
||||
args.addrs = new_addrs;
|
||||
let running = Arc::new(AtomicBool::new(true));
|
||||
@@ -93,19 +91,12 @@ fn serve(args: Args, running: Arc<AtomicBool>) -> Result<Vec<JoinHandle<()>>> {
|
||||
config.alpn_protocols = vec![b"h2".to_vec(), b"http/1.1".to_vec()];
|
||||
let config = Arc::new(config);
|
||||
let tls_accepter = TlsAcceptor::from(config);
|
||||
let handshake_timeout = Duration::from_secs(10);
|
||||
|
||||
let handle = tokio::spawn(async move {
|
||||
loop {
|
||||
let Ok((stream, addr)) = listener.accept().await else {
|
||||
continue;
|
||||
};
|
||||
let Some(stream) =
|
||||
timeout(handshake_timeout, tls_accepter.accept(stream))
|
||||
.await
|
||||
.ok()
|
||||
.and_then(|v| v.ok())
|
||||
else {
|
||||
let (cnx, addr) = listener.accept().await.unwrap();
|
||||
let Ok(stream) = tls_accepter.accept(cnx).await else {
|
||||
warn!("During cls handshake connection from {}", addr);
|
||||
continue;
|
||||
};
|
||||
let stream = TokioIo::new(stream);
|
||||
@@ -122,10 +113,8 @@ fn serve(args: Args, running: Arc<AtomicBool>) -> Result<Vec<JoinHandle<()>>> {
|
||||
(None, None) => {
|
||||
let handle = tokio::spawn(async move {
|
||||
loop {
|
||||
let Ok((stream, addr)) = listener.accept().await else {
|
||||
continue;
|
||||
};
|
||||
let stream = TokioIo::new(stream);
|
||||
let (cnx, addr) = listener.accept().await.unwrap();
|
||||
let stream = TokioIo::new(cnx);
|
||||
tokio::spawn(handle_stream(
|
||||
server_handle.clone(),
|
||||
stream,
|
||||
@@ -150,10 +139,8 @@ fn serve(args: Args, running: Arc<AtomicBool>) -> Result<Vec<JoinHandle<()>>> {
|
||||
.with_context(|| format!("Failed to bind `{}`", path.display()))?;
|
||||
let handle = tokio::spawn(async move {
|
||||
loop {
|
||||
let Ok((stream, _addr)) = listener.accept().await else {
|
||||
continue;
|
||||
};
|
||||
let stream = TokioIo::new(stream);
|
||||
let (cnx, _) = listener.accept().await.unwrap();
|
||||
let stream = TokioIo::new(cnx);
|
||||
tokio::spawn(handle_stream(server_handle.clone(), stream, None));
|
||||
}
|
||||
});
|
||||
@@ -173,15 +160,18 @@ where
|
||||
let hyper_service =
|
||||
service_fn(move |request: Request<Incoming>| handle.clone().call(request, addr));
|
||||
|
||||
match Builder::new(TokioExecutor::new())
|
||||
let ret = Builder::new(TokioExecutor::new())
|
||||
.serve_connection_with_upgrades(stream, hyper_service)
|
||||
.await
|
||||
{
|
||||
Ok(()) => {}
|
||||
Err(_err) => {
|
||||
// This error only appears when the client doesn't send a request and terminate the connection.
|
||||
//
|
||||
// If client sends one request then terminate connection whenever, it doesn't appear.
|
||||
.await;
|
||||
|
||||
if let Err(err) = ret {
|
||||
let scope = match addr {
|
||||
Some(addr) => format!(" from {}", addr),
|
||||
None => String::new(),
|
||||
};
|
||||
match err.downcast_ref::<std::io::Error>() {
|
||||
Some(err) if err.kind() == std::io::ErrorKind::UnexpectedEof => {}
|
||||
_ => warn!("Serving connection{}: {}", scope, err),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@@ -23,13 +23,12 @@ use hyper::body::Frame;
|
||||
use hyper::{
|
||||
body::Incoming,
|
||||
header::{
|
||||
HeaderValue, AUTHORIZATION, CONNECTION, CONTENT_DISPOSITION, CONTENT_LENGTH, CONTENT_RANGE,
|
||||
HeaderValue, AUTHORIZATION, CONTENT_DISPOSITION, CONTENT_LENGTH, CONTENT_RANGE,
|
||||
CONTENT_TYPE, RANGE,
|
||||
},
|
||||
Method, StatusCode, Uri,
|
||||
};
|
||||
use serde::Serialize;
|
||||
use sha2::{Digest, Sha256};
|
||||
use std::borrow::Cow;
|
||||
use std::cmp::Ordering;
|
||||
use std::collections::HashMap;
|
||||
@@ -107,18 +106,12 @@ impl Server {
|
||||
let uri = req.uri().clone();
|
||||
let assets_prefix = &self.assets_prefix;
|
||||
let enable_cors = self.args.enable_cors;
|
||||
let is_microsoft_webdav = req
|
||||
.headers()
|
||||
.get("user-agent")
|
||||
.and_then(|v| v.to_str().ok())
|
||||
.map(|v| v.starts_with("Microsoft-WebDAV-MiniRedir/"))
|
||||
.unwrap_or_default();
|
||||
let mut http_log_data = self.args.http_logger.data(&req);
|
||||
if let Some(addr) = addr {
|
||||
http_log_data.insert("remote_addr".to_string(), addr.ip().to_string());
|
||||
}
|
||||
|
||||
let mut res = match self.clone().handle(req, is_microsoft_webdav).await {
|
||||
let mut res = match self.clone().handle(req).await {
|
||||
Ok(res) => {
|
||||
http_log_data.insert("status".to_string(), res.status().as_u16().to_string());
|
||||
if !uri.path().starts_with(assets_prefix) {
|
||||
@@ -138,22 +131,13 @@ impl Server {
|
||||
}
|
||||
};
|
||||
|
||||
if is_microsoft_webdav {
|
||||
// microsoft webdav requires this.
|
||||
res.headers_mut()
|
||||
.insert(CONNECTION, HeaderValue::from_static("close"));
|
||||
}
|
||||
if enable_cors {
|
||||
add_cors(&mut res);
|
||||
}
|
||||
Ok(res)
|
||||
}
|
||||
|
||||
pub async fn handle(
|
||||
self: Arc<Self>,
|
||||
req: Request,
|
||||
is_microsoft_webdav: bool,
|
||||
) -> Result<Response> {
|
||||
pub async fn handle(self: Arc<Self>, req: Request) -> Result<Response> {
|
||||
let mut res = Response::default();
|
||||
|
||||
let req_path = req.uri().path();
|
||||
@@ -177,10 +161,7 @@ impl Server {
|
||||
}
|
||||
|
||||
let authorization = headers.get(AUTHORIZATION);
|
||||
let guard =
|
||||
self.args
|
||||
.auth
|
||||
.guard(&relative_path, &method, authorization, is_microsoft_webdav);
|
||||
let guard = self.args.auth.guard(&relative_path, &method, authorization);
|
||||
|
||||
let (user, access_paths) = match guard {
|
||||
(None, None) => {
|
||||
@@ -326,8 +307,6 @@ impl Server {
|
||||
} else if query_params.contains_key("view") {
|
||||
self.handle_edit_file(path, DataKind::View, head_only, user, &mut res)
|
||||
.await?;
|
||||
} else if query_params.contains_key("hash") {
|
||||
self.handle_hash_file(path, head_only, &mut res).await?;
|
||||
} else {
|
||||
self.handle_send_file(path, headers, head_only, &mut res)
|
||||
.await?;
|
||||
@@ -936,24 +915,6 @@ impl Server {
|
||||
Ok(())
|
||||
}
|
||||
|
||||
async fn handle_hash_file(
|
||||
&self,
|
||||
path: &Path,
|
||||
head_only: bool,
|
||||
res: &mut Response,
|
||||
) -> Result<()> {
|
||||
let output = sha256_file(path).await?;
|
||||
res.headers_mut()
|
||||
.typed_insert(ContentType::from(mime_guess::mime::TEXT_HTML_UTF_8));
|
||||
res.headers_mut()
|
||||
.typed_insert(ContentLength(output.as_bytes().len() as u64));
|
||||
if head_only {
|
||||
return Ok(());
|
||||
}
|
||||
*res.body_mut() = body_full(output);
|
||||
Ok(())
|
||||
}
|
||||
|
||||
async fn handle_propfind_dir(
|
||||
&self,
|
||||
path: &Path,
|
||||
@@ -1222,7 +1183,7 @@ impl Server {
|
||||
let guard = self
|
||||
.args
|
||||
.auth
|
||||
.guard(&dest_path, req.method(), authorization, false);
|
||||
.guard(&dest_path, req.method(), authorization);
|
||||
|
||||
match guard {
|
||||
(_, Some(_)) => {}
|
||||
@@ -1401,7 +1362,7 @@ impl PathItem {
|
||||
|
||||
pub fn to_dav_xml(&self, prefix: &str) -> String {
|
||||
let mtime = match Utc.timestamp_millis_opt(self.mtime as i64) {
|
||||
LocalResult::Single(v) => format!("{}", v.format("%a, %d %b %Y %H:%M:%S GMT")),
|
||||
LocalResult::Single(v) => v.to_rfc2822(),
|
||||
_ => String::new(),
|
||||
};
|
||||
let mut href = encode_uri(&format!("{}{}", prefix, &self.name));
|
||||
@@ -1574,6 +1535,7 @@ async fn zip_dir<W: AsyncWrite + Unpin>(
|
||||
) -> Result<()> {
|
||||
let mut writer = ZipFileWriter::with_tokio(writer);
|
||||
let hidden = Arc::new(hidden.to_vec());
|
||||
let hidden = hidden.clone();
|
||||
let dir_clone = dir.to_path_buf();
|
||||
let zip_paths = tokio::task::spawn_blocking(move || {
|
||||
let mut paths: Vec<PathBuf> = vec![];
|
||||
@@ -1755,20 +1717,3 @@ fn parse_upload_offset(headers: &HeaderMap<HeaderValue>, size: u64) -> Result<Op
|
||||
let (start, _) = parse_range(value, size).ok_or_else(err)?;
|
||||
Ok(Some(start))
|
||||
}
|
||||
|
||||
async fn sha256_file(path: &Path) -> Result<String> {
|
||||
let mut file = fs::File::open(path).await?;
|
||||
let mut hasher = Sha256::new();
|
||||
let mut buffer = [0u8; 8192];
|
||||
|
||||
loop {
|
||||
let bytes_read = file.read(&mut buffer).await?;
|
||||
if bytes_read == 0 {
|
||||
break;
|
||||
}
|
||||
hasher.update(&buffer[..bytes_read]);
|
||||
}
|
||||
|
||||
let result = hasher.finalize();
|
||||
Ok(format!("{:x}", result))
|
||||
}
|
||||
|
||||
@@ -1,8 +1,7 @@
|
||||
mod digest_auth_util;
|
||||
mod fixtures;
|
||||
mod utils;
|
||||
|
||||
use digest_auth_util::send_with_digest_auth;
|
||||
use diqwest::blocking::WithDigestAuth;
|
||||
use fixtures::{server, Error, TestServer};
|
||||
use indexmap::IndexSet;
|
||||
use rstest::rstest;
|
||||
@@ -33,7 +32,9 @@ fn auth(#[case] server: TestServer, #[case] user: &str, #[case] pass: &str) -> R
|
||||
let url = format!("{}file1", server.url());
|
||||
let resp = fetch!(b"PUT", &url).body(b"abc".to_vec()).send()?;
|
||||
assert_eq!(resp.status(), 401);
|
||||
let resp = send_with_digest_auth(fetch!(b"PUT", &url).body(b"abc".to_vec()), user, pass)?;
|
||||
let resp = fetch!(b"PUT", &url)
|
||||
.body(b"abc".to_vec())
|
||||
.send_with_digest_auth(user, pass)?;
|
||||
assert_eq!(resp.status(), 201);
|
||||
Ok(())
|
||||
}
|
||||
@@ -66,12 +67,13 @@ fn auth_hashed_password(
|
||||
let url = format!("{}file1", server.url());
|
||||
let resp = fetch!(b"PUT", &url).body(b"abc".to_vec()).send()?;
|
||||
assert_eq!(resp.status(), 401);
|
||||
if let Err(err) =
|
||||
send_with_digest_auth(fetch!(b"PUT", &url).body(b"abc".to_vec()), "user", "pass")
|
||||
if let Err(err) = fetch!(b"PUT", &url)
|
||||
.body(b"abc".to_vec())
|
||||
.send_with_digest_auth("user", "pass")
|
||||
{
|
||||
assert_eq!(
|
||||
err.to_string(),
|
||||
r#"Missing "realm" in header: Basic realm="DUFS""#
|
||||
format!("{err:?}"),
|
||||
r#"DigestAuth(MissingRequired("realm", "Basic realm=\"DUFS\""))"#
|
||||
);
|
||||
}
|
||||
let resp = fetch!(b"PUT", &url)
|
||||
@@ -89,7 +91,9 @@ fn auth_and_public(
|
||||
let url = format!("{}file1", server.url());
|
||||
let resp = fetch!(b"PUT", &url).body(b"abc".to_vec()).send()?;
|
||||
assert_eq!(resp.status(), 401);
|
||||
let resp = send_with_digest_auth(fetch!(b"PUT", &url).body(b"abc".to_vec()), "user", "pass")?;
|
||||
let resp = fetch!(b"PUT", &url)
|
||||
.body(b"abc".to_vec())
|
||||
.send_with_digest_auth("user", "pass")?;
|
||||
assert_eq!(resp.status(), 201);
|
||||
let resp = fetch!(b"GET", &url).send()?;
|
||||
assert_eq!(resp.status(), 200);
|
||||
@@ -121,9 +125,9 @@ fn auth_check(
|
||||
let url = format!("{}index.html", server.url());
|
||||
let resp = fetch!(b"WRITEABLE", &url).send()?;
|
||||
assert_eq!(resp.status(), 401);
|
||||
let resp = send_with_digest_auth(fetch!(b"WRITEABLE", &url), "user2", "pass2")?;
|
||||
let resp = fetch!(b"WRITEABLE", &url).send_with_digest_auth("user2", "pass2")?;
|
||||
assert_eq!(resp.status(), 403);
|
||||
let resp = send_with_digest_auth(fetch!(b"WRITEABLE", &url), "user", "pass")?;
|
||||
let resp = fetch!(b"WRITEABLE", &url).send_with_digest_auth("user", "pass")?;
|
||||
assert_eq!(resp.status(), 200);
|
||||
Ok(())
|
||||
}
|
||||
@@ -135,9 +139,9 @@ fn auth_compact_rules(
|
||||
let url = format!("{}index.html", server.url());
|
||||
let resp = fetch!(b"WRITEABLE", &url).send()?;
|
||||
assert_eq!(resp.status(), 401);
|
||||
let resp = send_with_digest_auth(fetch!(b"WRITEABLE", &url), "user2", "pass2")?;
|
||||
let resp = fetch!(b"WRITEABLE", &url).send_with_digest_auth("user2", "pass2")?;
|
||||
assert_eq!(resp.status(), 403);
|
||||
let resp = send_with_digest_auth(fetch!(b"WRITEABLE", &url), "user", "pass")?;
|
||||
let resp = fetch!(b"WRITEABLE", &url).send_with_digest_auth("user", "pass")?;
|
||||
assert_eq!(resp.status(), 200);
|
||||
Ok(())
|
||||
}
|
||||
@@ -149,10 +153,12 @@ fn auth_readonly(
|
||||
let url = format!("{}index.html", server.url());
|
||||
let resp = fetch!(b"GET", &url).send()?;
|
||||
assert_eq!(resp.status(), 401);
|
||||
let resp = send_with_digest_auth(fetch!(b"GET", &url), "user2", "pass2")?;
|
||||
let resp = fetch!(b"GET", &url).send_with_digest_auth("user2", "pass2")?;
|
||||
assert_eq!(resp.status(), 200);
|
||||
let url = format!("{}file1", server.url());
|
||||
let resp = send_with_digest_auth(fetch!(b"PUT", &url).body(b"abc".to_vec()), "user2", "pass2")?;
|
||||
let resp = fetch!(b"PUT", &url)
|
||||
.body(b"abc".to_vec())
|
||||
.send_with_digest_auth("user2", "pass2")?;
|
||||
assert_eq!(resp.status(), 403);
|
||||
Ok(())
|
||||
}
|
||||
@@ -165,9 +171,13 @@ fn auth_nest(
|
||||
let url = format!("{}dir1/file1", server.url());
|
||||
let resp = fetch!(b"PUT", &url).body(b"abc".to_vec()).send()?;
|
||||
assert_eq!(resp.status(), 401);
|
||||
let resp = send_with_digest_auth(fetch!(b"PUT", &url).body(b"abc".to_vec()), "user3", "pass3")?;
|
||||
let resp = fetch!(b"PUT", &url)
|
||||
.body(b"abc".to_vec())
|
||||
.send_with_digest_auth("user3", "pass3")?;
|
||||
assert_eq!(resp.status(), 201);
|
||||
let resp = send_with_digest_auth(fetch!(b"PUT", &url).body(b"abc".to_vec()), "user", "pass")?;
|
||||
let resp = fetch!(b"PUT", &url)
|
||||
.body(b"abc".to_vec())
|
||||
.send_with_digest_auth("user", "pass")?;
|
||||
assert_eq!(resp.status(), 201);
|
||||
Ok(())
|
||||
}
|
||||
@@ -209,11 +219,9 @@ fn auth_webdav_move(
|
||||
) -> Result<(), Error> {
|
||||
let origin_url = format!("{}dir1/test.html", server.url());
|
||||
let new_url = format!("{}test2.html", server.url());
|
||||
let resp = send_with_digest_auth(
|
||||
fetch!(b"MOVE", &origin_url).header("Destination", &new_url),
|
||||
"user3",
|
||||
"pass3",
|
||||
)?;
|
||||
let resp = fetch!(b"MOVE", &origin_url)
|
||||
.header("Destination", &new_url)
|
||||
.send_with_digest_auth("user3", "pass3")?;
|
||||
assert_eq!(resp.status(), 403);
|
||||
Ok(())
|
||||
}
|
||||
@@ -225,11 +233,9 @@ fn auth_webdav_copy(
|
||||
) -> Result<(), Error> {
|
||||
let origin_url = format!("{}dir1/test.html", server.url());
|
||||
let new_url = format!("{}test2.html", server.url());
|
||||
let resp = send_with_digest_auth(
|
||||
fetch!(b"COPY", &origin_url).header("Destination", &new_url),
|
||||
"user3",
|
||||
"pass3",
|
||||
)?;
|
||||
let resp = fetch!(b"COPY", &origin_url)
|
||||
.header("Destination", &new_url)
|
||||
.send_with_digest_auth("user3", "pass3")?;
|
||||
assert_eq!(resp.status(), 403);
|
||||
Ok(())
|
||||
}
|
||||
@@ -241,7 +247,7 @@ fn auth_path_prefix(
|
||||
let url = format!("{}xyz/index.html", server.url());
|
||||
let resp = fetch!(b"GET", &url).send()?;
|
||||
assert_eq!(resp.status(), 401);
|
||||
let resp = send_with_digest_auth(fetch!(b"GET", &url), "user", "pass")?;
|
||||
let resp = fetch!(b"GET", &url).send_with_digest_auth("user", "pass")?;
|
||||
assert_eq!(resp.status(), 200);
|
||||
Ok(())
|
||||
}
|
||||
@@ -250,15 +256,12 @@ fn auth_path_prefix(
|
||||
fn auth_partial_index(
|
||||
#[with(&["--auth", "user:pass@/dir1:rw,/dir2:rw", "-A"])] server: TestServer,
|
||||
) -> Result<(), Error> {
|
||||
let resp = send_with_digest_auth(fetch!(b"GET", server.url()), "user", "pass")?;
|
||||
let resp = fetch!(b"GET", server.url()).send_with_digest_auth("user", "pass")?;
|
||||
assert_eq!(resp.status(), 200);
|
||||
let paths = utils::retrieve_index_paths(&resp.text()?);
|
||||
assert_eq!(paths, IndexSet::from(["dir1/".into(), "dir2/".into()]));
|
||||
let resp = send_with_digest_auth(
|
||||
fetch!(b"GET", format!("{}?q={}", server.url(), "test.html")),
|
||||
"user",
|
||||
"pass",
|
||||
)?;
|
||||
let resp = fetch!(b"GET", format!("{}?q={}", server.url(), "test.html"))
|
||||
.send_with_digest_auth("user", "pass")?;
|
||||
assert_eq!(resp.status(), 200);
|
||||
let paths = utils::retrieve_index_paths(&resp.text()?);
|
||||
assert_eq!(
|
||||
@@ -285,7 +288,7 @@ fn auth_propfind_dir(
|
||||
#[with(&["--auth", "admin:admin@/:rw", "--auth", "user:pass@/dir-assets", "-A"])]
|
||||
server: TestServer,
|
||||
) -> Result<(), Error> {
|
||||
let resp = send_with_digest_auth(fetch!(b"PROPFIND", server.url()), "user", "pass")?;
|
||||
let resp = fetch!(b"PROPFIND", server.url()).send_with_digest_auth("user", "pass")?;
|
||||
assert_eq!(resp.status(), 207);
|
||||
let body = resp.text()?;
|
||||
assert!(body.contains("<D:href>/dir-assets/</D:href>"));
|
||||
@@ -299,14 +302,14 @@ fn auth_data(
|
||||
) -> Result<(), Error> {
|
||||
let resp = reqwest::blocking::get(server.url())?;
|
||||
let content = resp.text()?;
|
||||
let json = utils::retrieve_json(&content).unwrap();
|
||||
let json = utils::retrive_json(&content).unwrap();
|
||||
assert_eq!(json["allow_delete"], serde_json::Value::Bool(false));
|
||||
assert_eq!(json["allow_upload"], serde_json::Value::Bool(false));
|
||||
let resp = fetch!(b"GET", server.url())
|
||||
.basic_auth("user", Some("pass"))
|
||||
.send()?;
|
||||
let content = resp.text()?;
|
||||
let json = utils::retrieve_json(&content).unwrap();
|
||||
let json = utils::retrive_json(&content).unwrap();
|
||||
assert_eq!(json["allow_delete"], serde_json::Value::Bool(true));
|
||||
assert_eq!(json["allow_upload"], serde_json::Value::Bool(true));
|
||||
Ok(())
|
||||
@@ -317,11 +320,15 @@ fn auth_precedence(
|
||||
#[with(&["--auth", "user:pass@/dir1:rw,/dir1/test.txt", "-A"])] server: TestServer,
|
||||
) -> Result<(), Error> {
|
||||
let url = format!("{}dir1/test.txt", server.url());
|
||||
let resp = send_with_digest_auth(fetch!(b"PUT", &url).body(b"abc".to_vec()), "user", "pass")?;
|
||||
let resp = fetch!(b"PUT", &url)
|
||||
.body(b"abc".to_vec())
|
||||
.send_with_digest_auth("user", "pass")?;
|
||||
assert_eq!(resp.status(), 403);
|
||||
|
||||
let url = format!("{}dir1/file1", server.url());
|
||||
let resp = send_with_digest_auth(fetch!(b"PUT", &url).body(b"abc".to_vec()), "user", "pass")?;
|
||||
let resp = fetch!(b"PUT", &url)
|
||||
.body(b"abc".to_vec())
|
||||
.send_with_digest_auth("user", "pass")?;
|
||||
assert_eq!(resp.status(), 201);
|
||||
|
||||
Ok(())
|
||||
|
||||
@@ -1,10 +1,9 @@
|
||||
mod digest_auth_util;
|
||||
mod fixtures;
|
||||
mod utils;
|
||||
|
||||
use assert_cmd::prelude::*;
|
||||
use assert_fs::TempDir;
|
||||
use digest_auth_util::send_with_digest_auth;
|
||||
use diqwest::blocking::WithDigestAuth;
|
||||
use fixtures::{port, tmpdir, wait_for_port, Error};
|
||||
use rstest::rstest;
|
||||
use std::path::PathBuf;
|
||||
@@ -28,18 +27,20 @@ fn use_config_file(tmpdir: TempDir, port: u16) -> Result<(), Error> {
|
||||
assert_eq!(resp.status(), 401);
|
||||
|
||||
let url = format!("http://localhost:{port}/dufs/index.html");
|
||||
let resp = send_with_digest_auth(fetch!(b"GET", &url), "user", "pass")?;
|
||||
let resp = fetch!(b"GET", &url).send_with_digest_auth("user", "pass")?;
|
||||
assert_eq!(resp.text()?, "This is index.html");
|
||||
|
||||
let url = format!("http://localhost:{port}/dufs?simple");
|
||||
let resp = send_with_digest_auth(fetch!(b"GET", &url), "user", "pass")?;
|
||||
let resp = fetch!(b"GET", &url).send_with_digest_auth("user", "pass")?;
|
||||
let text: String = resp.text().unwrap();
|
||||
assert!(text.split('\n').any(|c| c == "dir1/"));
|
||||
assert!(!text.split('\n').any(|c| c == "dir3/"));
|
||||
assert!(!text.split('\n').any(|c| c == "test.txt"));
|
||||
|
||||
let url = format!("http://localhost:{port}/dufs/dir1/upload.txt");
|
||||
let resp = send_with_digest_auth(fetch!(b"PUT", &url).body("Hello"), "user", "pass")?;
|
||||
let resp = fetch!(b"PUT", &url)
|
||||
.body("Hello")
|
||||
.send_with_digest_auth("user", "pass")?;
|
||||
assert_eq!(resp.status(), 201);
|
||||
|
||||
child.kill()?;
|
||||
|
||||
@@ -1,91 +0,0 @@
|
||||
/// Refs https://github.dev/maoertel/diqwest/blob/main/src/blocking.rs
|
||||
use anyhow::{anyhow, Result};
|
||||
use digest_auth::{AuthContext, AuthorizationHeader, HttpMethod};
|
||||
use hyper::{header::AUTHORIZATION, HeaderMap, StatusCode};
|
||||
use reqwest::blocking::{RequestBuilder, Response};
|
||||
use url::Position;
|
||||
|
||||
pub fn send_with_digest_auth(
|
||||
request_builder: RequestBuilder,
|
||||
username: &str,
|
||||
password: &str,
|
||||
) -> Result<Response> {
|
||||
let first_response = try_clone_request_builder(&request_builder)?.send()?;
|
||||
match first_response.status() {
|
||||
StatusCode::UNAUTHORIZED => {
|
||||
try_digest_auth(request_builder, first_response, username, password)
|
||||
}
|
||||
_ => Ok(first_response),
|
||||
}
|
||||
}
|
||||
|
||||
fn try_digest_auth(
|
||||
request_builder: RequestBuilder,
|
||||
first_response: Response,
|
||||
username: &str,
|
||||
password: &str,
|
||||
) -> Result<Response> {
|
||||
if let Some(answer) = get_answer(
|
||||
&request_builder,
|
||||
first_response.headers(),
|
||||
username,
|
||||
password,
|
||||
)? {
|
||||
return Ok(request_builder
|
||||
.header(AUTHORIZATION, answer.to_header_string())
|
||||
.send()?);
|
||||
};
|
||||
|
||||
Ok(first_response)
|
||||
}
|
||||
|
||||
fn try_clone_request_builder(request_builder: &RequestBuilder) -> Result<RequestBuilder> {
|
||||
request_builder
|
||||
.try_clone()
|
||||
.ok_or_else(|| anyhow!("Request body must not be a stream"))
|
||||
}
|
||||
|
||||
fn get_answer(
|
||||
request_builder: &RequestBuilder,
|
||||
first_response: &HeaderMap,
|
||||
username: &str,
|
||||
password: &str,
|
||||
) -> Result<Option<AuthorizationHeader>> {
|
||||
let answer = calculate_answer(request_builder, first_response, username, password);
|
||||
match answer {
|
||||
Ok(answer) => Ok(Some(answer)),
|
||||
Err(error) => Err(error),
|
||||
}
|
||||
}
|
||||
|
||||
fn calculate_answer(
|
||||
request_builder: &RequestBuilder,
|
||||
headers: &HeaderMap,
|
||||
username: &str,
|
||||
password: &str,
|
||||
) -> Result<AuthorizationHeader> {
|
||||
let request = try_clone_request_builder(request_builder)?.build()?;
|
||||
let path = &request.url()[Position::AfterPort..];
|
||||
let method = HttpMethod::from(request.method().as_str());
|
||||
let body = request.body().and_then(|b| b.as_bytes());
|
||||
|
||||
parse_digest_auth_header(headers, path, method, body, username, password)
|
||||
}
|
||||
|
||||
fn parse_digest_auth_header(
|
||||
header: &HeaderMap,
|
||||
path: &str,
|
||||
method: HttpMethod,
|
||||
body: Option<&[u8]>,
|
||||
username: &str,
|
||||
password: &str,
|
||||
) -> Result<AuthorizationHeader> {
|
||||
let www_auth = header
|
||||
.get("www-authenticate")
|
||||
.ok_or_else(|| anyhow!("The header 'www-authenticate' is missing."))?
|
||||
.to_str()?;
|
||||
let context = AuthContext::new_with_method(username, password, path, body, method);
|
||||
let mut prompt = digest_auth::parse(www_auth)?;
|
||||
|
||||
Ok(prompt.respond(&context)?)
|
||||
}
|
||||
@@ -4,7 +4,7 @@ mod utils;
|
||||
use fixtures::{server, Error, TestServer, BIN_FILE};
|
||||
use rstest::rstest;
|
||||
use serde_json::Value;
|
||||
use utils::retrieve_edit_file;
|
||||
use utils::retrive_edit_file;
|
||||
|
||||
#[rstest]
|
||||
fn get_dir(server: TestServer) -> Result<(), Error> {
|
||||
@@ -189,21 +189,6 @@ fn head_file(server: TestServer) -> Result<(), Error> {
|
||||
Ok(())
|
||||
}
|
||||
|
||||
#[rstest]
|
||||
fn hash_file(server: TestServer) -> Result<(), Error> {
|
||||
let resp = reqwest::blocking::get(format!("{}index.html?hash", server.url()))?;
|
||||
assert_eq!(
|
||||
resp.headers().get("content-type").unwrap(),
|
||||
"text/html; charset=utf-8"
|
||||
);
|
||||
assert_eq!(resp.status(), 200);
|
||||
assert_eq!(
|
||||
resp.text()?,
|
||||
"c8dd395e3202674b9512f7b7f956e0d96a8ba8f572e785b0d5413ab83766dbc4"
|
||||
);
|
||||
Ok(())
|
||||
}
|
||||
|
||||
#[rstest]
|
||||
fn get_file_404(server: TestServer) -> Result<(), Error> {
|
||||
let resp = reqwest::blocking::get(format!("{}404", server.url()))?;
|
||||
@@ -238,7 +223,7 @@ fn get_file_newline_path(server: TestServer) -> Result<(), Error> {
|
||||
fn get_file_edit(server: TestServer) -> Result<(), Error> {
|
||||
let resp = fetch!(b"GET", format!("{}index.html?edit", server.url())).send()?;
|
||||
assert_eq!(resp.status(), 200);
|
||||
let editable = retrieve_edit_file(&resp.text().unwrap()).unwrap();
|
||||
let editable = retrive_edit_file(&resp.text().unwrap()).unwrap();
|
||||
assert!(editable);
|
||||
Ok(())
|
||||
}
|
||||
@@ -247,7 +232,7 @@ fn get_file_edit(server: TestServer) -> Result<(), Error> {
|
||||
fn get_file_edit_bin(server: TestServer) -> Result<(), Error> {
|
||||
let resp = fetch!(b"GET", format!("{}{BIN_FILE}?edit", server.url())).send()?;
|
||||
assert_eq!(resp.status(), 200);
|
||||
let editable = retrieve_edit_file(&resp.text().unwrap()).unwrap();
|
||||
let editable = retrive_edit_file(&resp.text().unwrap()).unwrap();
|
||||
assert!(!editable);
|
||||
Ok(())
|
||||
}
|
||||
|
||||
@@ -1,8 +1,7 @@
|
||||
mod digest_auth_util;
|
||||
mod fixtures;
|
||||
mod utils;
|
||||
|
||||
use digest_auth_util::send_with_digest_auth;
|
||||
use diqwest::blocking::WithDigestAuth;
|
||||
use fixtures::{port, tmpdir, wait_for_port, Error};
|
||||
|
||||
use assert_cmd::prelude::*;
|
||||
@@ -32,12 +31,12 @@ fn log_remote_user(
|
||||
|
||||
let stdout = child.stdout.as_mut().expect("Failed to get stdout");
|
||||
|
||||
let req_builder = fetch!(b"GET", &format!("http://localhost:{port}"));
|
||||
let req = fetch!(b"GET", &format!("http://localhost:{port}"));
|
||||
|
||||
let resp = if is_basic {
|
||||
req_builder.basic_auth("user", Some("pass")).send()?
|
||||
req.basic_auth("user", Some("pass")).send()?
|
||||
} else {
|
||||
send_with_digest_auth(req_builder, "user", "pass")?
|
||||
req.send_with_digest_auth("user", "pass")?
|
||||
};
|
||||
|
||||
assert_eq!(resp.status(), 200);
|
||||
|
||||
@@ -26,7 +26,7 @@ macro_rules! fetch {
|
||||
|
||||
#[allow(dead_code)]
|
||||
pub fn retrieve_index_paths(content: &str) -> IndexSet<String> {
|
||||
let value = retrieve_json(content).unwrap();
|
||||
let value = retrive_json(content).unwrap();
|
||||
let paths = value
|
||||
.get("paths")
|
||||
.unwrap()
|
||||
@@ -47,8 +47,8 @@ pub fn retrieve_index_paths(content: &str) -> IndexSet<String> {
|
||||
}
|
||||
|
||||
#[allow(dead_code)]
|
||||
pub fn retrieve_edit_file(content: &str) -> Option<bool> {
|
||||
let value = retrieve_json(content)?;
|
||||
pub fn retrive_edit_file(content: &str) -> Option<bool> {
|
||||
let value = retrive_json(content)?;
|
||||
let value = value.get("editable").unwrap();
|
||||
Some(value.as_bool().unwrap())
|
||||
}
|
||||
@@ -60,7 +60,7 @@ pub fn encode_uri(v: &str) -> String {
|
||||
}
|
||||
|
||||
#[allow(dead_code)]
|
||||
pub fn retrieve_json(content: &str) -> Option<Value> {
|
||||
pub fn retrive_json(content: &str) -> Option<Value> {
|
||||
let lines: Vec<&str> = content.lines().collect();
|
||||
let line = lines.iter().find(|v| v.contains("DATA ="))?;
|
||||
let line_col = line.find("DATA =").unwrap() + 6;
|
||||
|
||||
Reference in New Issue
Block a user