mirror of
https://github.com/sigoden/dufs.git
synced 2026-04-09 09:09:03 +03:00
Compare commits
51 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
6cc8a18a3d | ||
|
|
a387d727b2 | ||
|
|
19d65a5aa4 | ||
|
|
d37762d2b9 | ||
|
|
9c9fca75d3 | ||
|
|
089d30c5a5 | ||
|
|
459a4d4f4a | ||
|
|
f8b69f4df8 | ||
|
|
53f064c73b | ||
|
|
8a92a0cf1a | ||
|
|
59685da06e | ||
|
|
09200860b4 | ||
|
|
4fbdec2878 | ||
|
|
d0453b7591 | ||
|
|
eda9769b2a | ||
|
|
d255f1376a | ||
|
|
669c4f8811 | ||
|
|
e576ddcbea | ||
|
|
af95ea1cd7 | ||
|
|
cbc620481d | ||
|
|
f1c9776962 | ||
|
|
ac15ae4e8e | ||
|
|
ab4ef06cb8 | ||
|
|
bc6c573acb | ||
|
|
f27f9e997f | ||
|
|
835438fc2a | ||
|
|
d445b78f96 | ||
|
|
881a67e1a4 | ||
|
|
bb5a5564b4 | ||
|
|
2cf6d39032 | ||
|
|
c500ce7acc | ||
|
|
f87c52fda2 | ||
|
|
2c5cc60965 | ||
|
|
972cf2377f | ||
|
|
5b338c40da | ||
|
|
964bf61c37 | ||
|
|
4bf92cc47a | ||
|
|
7d17d9c415 | ||
|
|
1db263efae | ||
|
|
5d26103ea2 | ||
|
|
3727dec115 | ||
|
|
0311c9fb90 | ||
|
|
e9ce4b2dc3 | ||
|
|
7aba3fe0b6 | ||
|
|
ca5c3d7c54 | ||
|
|
ec2b064a9a | ||
|
|
cadea9a3bf | ||
|
|
3e0e6b2e8a | ||
|
|
632f7a41bf | ||
|
|
f1e90686dc | ||
|
|
dc7a7cbb3f |
16
.github/ISSUE_TEMPLATE/bug_report.md
vendored
16
.github/ISSUE_TEMPLATE/bug_report.md
vendored
@@ -5,20 +5,24 @@ about: Create a report to help us improve
|
||||
|
||||
**Problem**
|
||||
|
||||
<!-- A clear and concise description of what the bug is. -->
|
||||
<!-- Provide a clear and concise description of the bug you're experiencing. What did you expect to happen, and what actually happened? -->
|
||||
|
||||
**Configuration**
|
||||
|
||||
<!-- The dufs command-line arguments or configuration -->
|
||||
<!-- Please specify the Dufs command-line arguments or configuration used. -->
|
||||
|
||||
<!-- If the problems are related to auth/perm, please conceal only the user:pass, but do not hide the entire `auth` configuration. -->
|
||||
<!-- If the issue is related to authentication/permissions, include auth configurations while concealing sensitive information (e.g., passwords). -->
|
||||
|
||||
**Log**
|
||||
|
||||
The dufs log is crucial for locating the problem, so please do not omit it.
|
||||
<!-- Attach relevant log outputs that can help diagnose the issue. -->
|
||||
|
||||
**Environment:**
|
||||
**Screenshots/Media**
|
||||
|
||||
<!-- If applicable, add screenshots or videos that help illustrate the issue, especially for WebUI problems. -->
|
||||
|
||||
**Environment Information**
|
||||
- Dufs version:
|
||||
- Browser/Webdav info:
|
||||
- OS info:
|
||||
- Proxy server: e.g. nginx, cloudflare
|
||||
- Proxy server (if any): <!-- e.g. nginx, cloudflare -->
|
||||
70
CHANGELOG.md
70
CHANGELOG.md
@@ -2,6 +2,76 @@
|
||||
|
||||
All notable changes to this project will be documented in this file.
|
||||
|
||||
## [0.44.0] - 2025-08-02
|
||||
|
||||
### Bug Fixes
|
||||
|
||||
- No authentication check if no auth users ([#497](https://github.com/sigoden/dufs/issues/497))
|
||||
- Webui can't handle hash property of URL well ([#515](https://github.com/sigoden/dufs/issues/515))
|
||||
- Incorrect dir size due to hidden files ([#529](https://github.com/sigoden/dufs/issues/529))
|
||||
- Webui formatDirSize ([#568](https://github.com/sigoden/dufs/issues/568))
|
||||
- Follow symlinks when searching/archiving ([#572](https://github.com/sigoden/dufs/issues/572))
|
||||
- Incorrect separator for zip archives under windows ([#577](https://github.com/sigoden/dufs/issues/577))
|
||||
- Unexpected public auth asking for login info ([#583](https://github.com/sigoden/dufs/issues/583))
|
||||
|
||||
### Features
|
||||
|
||||
- Higher perm auth path shadows lower one ([#521](https://github.com/sigoden/dufs/issues/521))
|
||||
- Add cache-control:no-cache while sending file and index ([#528](https://github.com/sigoden/dufs/issues/528))
|
||||
- Support multipart ranges ([#535](https://github.com/sigoden/dufs/issues/535))
|
||||
- Limit sub directory item counting ([#556](https://github.com/sigoden/dufs/issues/556))
|
||||
- Tolerate the absence of mtime ([#559](https://github.com/sigoden/dufs/issues/559))
|
||||
- Support noscript fallback ([#602](https://github.com/sigoden/dufs/issues/602))
|
||||
- Support downloading via token auth ([#603](https://github.com/sigoden/dufs/issues/603))
|
||||
|
||||
### Refactor
|
||||
|
||||
- Change description for `--allow-archive` ([#511](https://github.com/sigoden/dufs/issues/511))
|
||||
- Removes clippy warnings ([#601](https://github.com/sigoden/dufs/issues/601))
|
||||
- Update deps ([#604](https://github.com/sigoden/dufs/issues/604))
|
||||
- Fix typos ([#605](https://github.com/sigoden/dufs/issues/605))
|
||||
|
||||
## [0.43.0] - 2024-11-04
|
||||
|
||||
### Bug Fixes
|
||||
|
||||
- Auth failed if password contains `:` ([#449](https://github.com/sigoden/dufs/issues/449))
|
||||
- Resolve speed bottleneck in 10G network ([#451](https://github.com/sigoden/dufs/issues/451))
|
||||
|
||||
### Features
|
||||
|
||||
- Webui displays subdirectory items ([#457](https://github.com/sigoden/dufs/issues/457))
|
||||
- Support binding abstract unix socket ([#468](https://github.com/sigoden/dufs/issues/468))
|
||||
- Provide healthcheck API ([#474](https://github.com/sigoden/dufs/issues/474))
|
||||
|
||||
### Refactor
|
||||
|
||||
- Do not show size for Dir ([#447](https://github.com/sigoden/dufs/issues/447))
|
||||
|
||||
## [0.42.0] - 2024-09-01
|
||||
|
||||
### Bug Fixes
|
||||
|
||||
- Garbled characters caused by atob ([#422](https://github.com/sigoden/dufs/issues/422))
|
||||
- Webui unexpected save-btn when file is non-editable ([#429](https://github.com/sigoden/dufs/issues/429))
|
||||
- Login succeeded but popup `Forbidden` ([#437](https://github.com/sigoden/dufs/issues/437))
|
||||
|
||||
### Features
|
||||
|
||||
- Implements remaining http cache conditionalss ([#407](https://github.com/sigoden/dufs/issues/407))
|
||||
- Base64 index-data to avoid misencoding ([#421](https://github.com/sigoden/dufs/issues/421))
|
||||
- Webui support logout ([#439](https://github.com/sigoden/dufs/issues/439))
|
||||
|
||||
### Refactor
|
||||
|
||||
- No inline scripts in HTML ([#391](https://github.com/sigoden/dufs/issues/391))
|
||||
- Return 400 for propfind request when depth is neither 0 nor 1 ([#403](https://github.com/sigoden/dufs/issues/403))
|
||||
- Remove sabredav-partialupdate from DAV res header ([#415](https://github.com/sigoden/dufs/issues/415))
|
||||
- Date formatting in cache tests ([#428](https://github.com/sigoden/dufs/issues/428))
|
||||
- Some query params work as flag and must not accept a value ([#431](https://github.com/sigoden/dufs/issues/431))
|
||||
- Improve logout at asserts/index.js ([#440](https://github.com/sigoden/dufs/issues/440))
|
||||
- Make logout works on safari ([#442](https://github.com/sigoden/dufs/issues/442))
|
||||
|
||||
## [0.41.0] - 2024-05-22
|
||||
|
||||
### Bug Fixes
|
||||
|
||||
1796
Cargo.lock
generated
1796
Cargo.lock
generated
File diff suppressed because it is too large
Load Diff
18
Cargo.toml
18
Cargo.toml
@@ -1,6 +1,6 @@
|
||||
[package]
|
||||
name = "dufs"
|
||||
version = "0.41.0"
|
||||
version = "0.44.0"
|
||||
edition = "2021"
|
||||
authors = ["sigoden <sigoden@gmail.com>"]
|
||||
description = "Dufs is a distinctive utility file server"
|
||||
@@ -14,26 +14,26 @@ keywords = ["static", "file", "server", "webdav", "cli"]
|
||||
clap = { version = "4.5", features = ["wrap_help", "env"] }
|
||||
clap_complete = "4.5"
|
||||
chrono = { version = "0.4", default-features = false, features = ["clock"] }
|
||||
tokio = { version = "1", features = ["rt-multi-thread", "macros", "fs", "io-util", "signal"]}
|
||||
tokio = { version = "1", features = ["rt-multi-thread", "macros", "fs", "io-util", "signal", "net"]}
|
||||
tokio-util = { version = "0.7", features = ["io-util", "compat"] }
|
||||
hyper = { version = "1", features = ["http1", "server"] }
|
||||
percent-encoding = "2.3"
|
||||
serde = { version = "1", features = ["derive"] }
|
||||
serde_json = "1"
|
||||
futures-util = { version = "0.3", default-features = false, features = ["alloc"] }
|
||||
async_zip = { version = "0.0.17", default-features = false, features = ["deflate", "bzip2", "xz", "chrono", "tokio"] }
|
||||
async_zip = { version = "0.0.18", default-features = false, features = ["deflate", "bzip2", "xz", "chrono", "tokio"] }
|
||||
headers = "0.4"
|
||||
mime_guess = "2.0"
|
||||
if-addrs = "0.12"
|
||||
if-addrs = "0.14"
|
||||
rustls-pemfile = { version = "2.0", optional = true }
|
||||
tokio-rustls = { version = "0.26", optional = true, default-features = false, features = ["ring", "tls12"]}
|
||||
md5 = "0.7"
|
||||
md5 = "0.8"
|
||||
lazy_static = "1.4"
|
||||
uuid = { version = "1.7", features = ["v4", "fast-rng"] }
|
||||
urlencoding = "2.1"
|
||||
xml-rs = "0.8"
|
||||
log = { version = "0.4", features = ["std"] }
|
||||
socket2 = "0.5"
|
||||
socket2 = "0.6"
|
||||
async-stream = "0.3"
|
||||
walkdir = "2.3"
|
||||
form_urlencoded = "1.2"
|
||||
@@ -53,6 +53,8 @@ http-body-util = "0.1"
|
||||
bytes = "1.5"
|
||||
pin-project-lite = "0.2"
|
||||
sha2 = "0.10.8"
|
||||
ed25519-dalek = "2.2.0"
|
||||
hex = "0.4.3"
|
||||
|
||||
[features]
|
||||
default = ["tls"]
|
||||
@@ -62,8 +64,8 @@ tls = ["rustls-pemfile", "tokio-rustls"]
|
||||
assert_cmd = "2"
|
||||
reqwest = { version = "0.12", features = ["blocking", "multipart", "rustls-tls"], default-features = false }
|
||||
assert_fs = "1"
|
||||
port_check = "0.2"
|
||||
rstest = "0.19"
|
||||
port_check = "0.3"
|
||||
rstest = "0.26.1"
|
||||
regex = "1"
|
||||
url = "2"
|
||||
predicates = "3"
|
||||
|
||||
27
README.md
27
README.md
@@ -66,7 +66,7 @@ Options:
|
||||
--allow-delete Allow delete files/folders
|
||||
--allow-search Allow search files/folders
|
||||
--allow-symlink Allow symlink to files/folders outside root directory
|
||||
--allow-archive Allow zip archive generation
|
||||
--allow-archive Allow download folders as archive file
|
||||
--enable-cors Enable CORS, sets `Access-Control-Allow-Origin: *`
|
||||
--render-index Serve index.html when requesting a directory, returns 404 if not found index.html
|
||||
--render-try-index Serve index.html when requesting a directory, returns directory listing if not found index.html
|
||||
@@ -216,8 +216,14 @@ dd skip=$upload_offset if=file status=none ibs=1 | \
|
||||
curl -X PATCH -H "X-Update-Range: append" --data-binary @- http://127.0.0.1:5000/file
|
||||
```
|
||||
|
||||
Health checks
|
||||
|
||||
```sh
|
||||
curl http://127.0.0.1:5000/__dufs__/health
|
||||
```
|
||||
|
||||
<details>
|
||||
<summary><h2>Advanced topics</h2></summary>
|
||||
<summary><h2>Advanced Topics</h2></summary>
|
||||
|
||||
### Access Control
|
||||
|
||||
@@ -238,24 +244,25 @@ dufs -a user:pass@/:rw,/dir1 -a @/
|
||||
- `-a user:pass@/:rw,/dir1`: `user` has read-write permissions for `/*`, has read-only permissions for `/dir1/*`.
|
||||
- `-a @/`: All paths is publicly accessible, everyone can view/download it.
|
||||
|
||||
> There are no restrictions on using ':' and '@' characters in a password. For example, `user:pa:ss@1@/:rw` is valid, the password is `pa:ss@1`.
|
||||
**Auth permissions are restricted by dufs global permissions.** If dufs does not enable upload permissions via `--allow-upload`, then the account will not have upload permissions even if it is granted `read-write`(`:rw`) permissions.
|
||||
|
||||
#### Hashed Password
|
||||
|
||||
DUFS supports the use of sha-512 hashed password.
|
||||
|
||||
Create hashed password
|
||||
Create hashed password:
|
||||
|
||||
```
|
||||
$ mkpasswd -m sha-512 -s
|
||||
Password: 123456
|
||||
```sh
|
||||
$ openssl passwd -6 123456 # or `mkpasswd -m sha-512 123456`
|
||||
$6$tWMB51u6Kb2ui3wd$5gVHP92V9kZcMwQeKTjyTRgySsYJu471Jb1I6iHQ8iZ6s07GgCIO69KcPBRuwPE5tDq05xMAzye0NxVKuJdYs/
|
||||
```
|
||||
|
||||
Use hashed password
|
||||
```
|
||||
Use hashed password:
|
||||
|
||||
```sh
|
||||
dufs -a 'admin:$6$tWMB51u6Kb2ui3wd$5gVHP92V9kZcMwQeKTjyTRgySsYJu471Jb1I6iHQ8iZ6s07GgCIO69KcPBRuwPE5tDq05xMAzye0NxVKuJdYs/@/:rw'
|
||||
```
|
||||
> The hashed password contains `$6`, which can expand to a variable in some shells, so you have to use **single quotes** to wrap it.
|
||||
|
||||
Two important things for hashed passwords:
|
||||
|
||||
@@ -396,6 +403,8 @@ Dufs allows users to customize the UI with your own assets.
|
||||
dufs --assets my-assets-dir/
|
||||
```
|
||||
|
||||
> If you only need to make slight adjustments to the current UI, you copy dufs's [assets](https://github.com/sigoden/dufs/tree/main/assets) directory and modify it accordingly. The current UI doesn't use any frameworks, just plain HTML/JS/CSS. As long as you have some basic knowledge of web development, it shouldn't be difficult to modify.
|
||||
|
||||
Your assets folder must contains a `index.html` file.
|
||||
|
||||
`index.html` can use the following placeholder variables to retrieve internal data.
|
||||
|
||||
@@ -222,15 +222,20 @@ body {
|
||||
margin-right: 2em;
|
||||
}
|
||||
|
||||
.login-btn {
|
||||
cursor: pointer;
|
||||
}
|
||||
|
||||
.save-btn {
|
||||
cursor: pointer;
|
||||
-webkit-user-select: none;
|
||||
user-select: none;
|
||||
}
|
||||
|
||||
.user-btn {
|
||||
display: flex;
|
||||
align-items: center;
|
||||
.logout-btn {
|
||||
cursor: pointer;
|
||||
display: flex;
|
||||
align-items: center;
|
||||
}
|
||||
|
||||
.user-name {
|
||||
|
||||
@@ -4,12 +4,11 @@
|
||||
<head>
|
||||
<meta charset="utf-8" />
|
||||
<meta name="viewport" content="width=device-width" />
|
||||
<noscript>
|
||||
<meta http-equiv="refresh" content="0; url=?noscript">
|
||||
</noscript>
|
||||
<link rel="icon" type="image/x-icon" href="__ASSETS_PREFIX__favicon.ico">
|
||||
<link rel="stylesheet" href="__ASSETS_PREFIX__index.css">
|
||||
<script>
|
||||
DATA = __INDEX_DATA__
|
||||
</script>
|
||||
<script src="__ASSETS_PREFIX__index.js"></script>
|
||||
</head>
|
||||
|
||||
<body>
|
||||
@@ -74,11 +73,12 @@
|
||||
d="M11.742 10.344a6.5 6.5 0 1 0-1.397 1.398h-.001c.03.04.062.078.098.115l3.85 3.85a1 1 0 0 0 1.415-1.414l-3.85-3.85a1.007 1.007 0 0 0-.115-.1zM12 6.5a5.5 5.5 0 1 1-11 0 5.5 5.5 0 0 1 11 0z" />
|
||||
</svg>
|
||||
</div>
|
||||
<input id="search" title="Searching for folders or files" name="q" type="text" maxlength="128" autocomplete="off" tabindex="1">
|
||||
<input id="search" title="Searching for folders or files" name="q" type="text" maxlength="128" autocomplete="off"
|
||||
tabindex="1">
|
||||
<input type="submit" hidden />
|
||||
</form>
|
||||
<div class="toolbox-right">
|
||||
<div class="login-btn hidden" title="Login for upload/move/delete/edit permissions">
|
||||
<div class="login-btn hidden" title="Login">
|
||||
<svg width="16" height="16" viewBox="0 0 16 16">
|
||||
<path fill-rule="evenodd"
|
||||
d="M6 3.5a.5.5 0 0 1 .5-.5h8a.5.5 0 0 1 .5.5v9a.5.5 0 0 1-.5.5h-8a.5.5 0 0 1-.5-.5v-2a.5.5 0 0 0-1 0v2A1.5 1.5 0 0 0 6.5 14h8a1.5 1.5 0 0 0 1.5-1.5v-9A1.5 1.5 0 0 0 14.5 2h-8A1.5 1.5 0 0 0 5 3.5v2a.5.5 0 0 0 1 0v-2z" />
|
||||
@@ -86,10 +86,10 @@
|
||||
d="M11.854 8.354a.5.5 0 0 0 0-.708l-3-3a.5.5 0 1 0-.708.708L10.293 7.5H1.5a.5.5 0 0 0 0 1h8.793l-2.147 2.146a.5.5 0 0 0 .708.708l3-3z" />
|
||||
</svg>
|
||||
</div>
|
||||
<div class="user-btn hidden">
|
||||
<div class="logout-btn hidden" title="Logout">
|
||||
<svg width="16" height="16" viewBox="0 0 16 16">
|
||||
<path
|
||||
d="M8 8a3 3 0 1 0 0-6 3 3 0 0 0 0 6Zm2-3a2 2 0 1 1-4 0 2 2 0 0 1 4 0Zm4 8c0 1-1 1-1 1H3s-1 0-1-1 1-4 6-4 6 3 6 4Zm-1-.004c-.001-.246-.154-.986-.832-1.664C11.516 10.68 10.289 10 8 10c-2.29 0-3.516.68-4.168 1.332-.678.678-.83 1.418-.832 1.664h10Z" />
|
||||
<path fill-rule="evenodd" d="M10 3.5a.5.5 0 0 0-.5-.5h-8a.5.5 0 0 0-.5.5v9a.5.5 0 0 0 .5.5h8a.5.5 0 0 0 .5-.5v-2a.5.5 0 0 1 1 0v2A1.5 1.5 0 0 1 9.5 14h-8A1.5 1.5 0 0 1 0 12.5v-9A1.5 1.5 0 0 1 1.5 2h8A1.5 1.5 0 0 1 11 3.5v2a.5.5 0 0 1-1 0z"/>
|
||||
<path fill-rule="evenodd" d="M4.146 8.354a.5.5 0 0 1 0-.708l3-3a.5.5 0 1 1 .708.708L5.707 7.5H14.5a.5.5 0 0 1 0 1H5.707l2.147 2.146a.5.5 0 0 1-.708.708z"/>
|
||||
</svg>
|
||||
<span class="user-name"></span>
|
||||
</div>
|
||||
@@ -125,9 +125,8 @@
|
||||
<textarea id="editor" class="editor hidden" aria-label="Editor" cols="10"></textarea>
|
||||
</div>
|
||||
</div>
|
||||
<script>
|
||||
window.addEventListener("DOMContentLoaded", ready);
|
||||
</script>
|
||||
<template id="index-data">__INDEX_DATA__</template>
|
||||
<script src="__ASSETS_PREFIX__index.js"></script>
|
||||
</body>
|
||||
|
||||
</html>
|
||||
223
assets/index.js
223
assets/index.js
@@ -29,6 +29,11 @@ var DUFS_MAX_UPLOADINGS = 1;
|
||||
*/
|
||||
var DATA;
|
||||
|
||||
/**
|
||||
* @type {string}
|
||||
*/
|
||||
var DIR_EMPTY_NOTE;
|
||||
|
||||
/**
|
||||
* @type {PARAMS}
|
||||
* @typedef {object} PARAMS
|
||||
@@ -45,7 +50,7 @@ const IFRAME_FORMATS = [
|
||||
".mp3", ".ogg", ".wav", ".m4a",
|
||||
];
|
||||
|
||||
const dirEmptyNote = PARAMS.q ? 'No results' : DATA.dir_exists ? 'Empty folder' : 'Folder will be created when a file is uploaded';
|
||||
const MAX_SUBPATHS_COUNT = 1000;
|
||||
|
||||
const ICONS = {
|
||||
dir: `<svg height="16" viewBox="0 0 14 16" width="14"><path fill-rule="evenodd" d="M13 4H7V3c0-.66-.31-1-1-1H1c-.55 0-1 .45-1 1v10c0 .55.45 1 1 1h12c.55 0 1-.45 1-1V5c0-.55-.45-1-1-1zM6 4H1V3h5v1z"></path></svg>`,
|
||||
@@ -91,43 +96,61 @@ let $editor;
|
||||
/**
|
||||
* @type Element
|
||||
*/
|
||||
let $userBtn;
|
||||
let $loginBtn;
|
||||
/**
|
||||
* @type Element
|
||||
*/
|
||||
let $logoutBtn;
|
||||
/**
|
||||
* @type Element
|
||||
*/
|
||||
let $userName;
|
||||
|
||||
function ready() {
|
||||
$pathsTable = document.querySelector(".paths-table")
|
||||
// Produce table when window loads
|
||||
window.addEventListener("DOMContentLoaded", async () => {
|
||||
const $indexData = document.getElementById('index-data');
|
||||
if (!$indexData) {
|
||||
alert("No data");
|
||||
return;
|
||||
}
|
||||
|
||||
DATA = JSON.parse(decodeBase64($indexData.innerHTML));
|
||||
DIR_EMPTY_NOTE = PARAMS.q ? 'No results' : DATA.dir_exists ? 'Empty folder' : 'Folder will be created when a file is uploaded';
|
||||
|
||||
await ready();
|
||||
});
|
||||
|
||||
async function ready() {
|
||||
$pathsTable = document.querySelector(".paths-table");
|
||||
$pathsTableHead = document.querySelector(".paths-table thead");
|
||||
$pathsTableBody = document.querySelector(".paths-table tbody");
|
||||
$uploadersTable = document.querySelector(".uploaders-table");
|
||||
$emptyFolder = document.querySelector(".empty-folder");
|
||||
$editor = document.querySelector(".editor");
|
||||
$userBtn = document.querySelector(".user-btn");
|
||||
$loginBtn = document.querySelector(".login-btn");
|
||||
$logoutBtn = document.querySelector(".logout-btn");
|
||||
$userName = document.querySelector(".user-name");
|
||||
|
||||
addBreadcrumb(DATA.href, DATA.uri_prefix);
|
||||
|
||||
if (DATA.kind == "Index") {
|
||||
if (DATA.kind === "Index") {
|
||||
document.title = `Index of ${DATA.href} - Dufs`;
|
||||
document.querySelector(".index-page").classList.remove("hidden");
|
||||
|
||||
setupIndexPage();
|
||||
} else if (DATA.kind == "Edit") {
|
||||
await setupIndexPage();
|
||||
} else if (DATA.kind === "Edit") {
|
||||
document.title = `Edit ${DATA.href} - Dufs`;
|
||||
document.querySelector(".editor-page").classList.remove("hidden");;
|
||||
document.querySelector(".editor-page").classList.remove("hidden");
|
||||
|
||||
setupEditorPage();
|
||||
} else if (DATA.kind == "View") {
|
||||
await setupEditorPage();
|
||||
} else if (DATA.kind === "View") {
|
||||
document.title = `View ${DATA.href} - Dufs`;
|
||||
document.querySelector(".editor-page").classList.remove("hidden");;
|
||||
document.querySelector(".editor-page").classList.remove("hidden");
|
||||
|
||||
setupEditorPage();
|
||||
await setupEditorPage();
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
class Uploader {
|
||||
/**
|
||||
*
|
||||
@@ -221,16 +244,16 @@ class Uploader {
|
||||
uploadOffset = parseInt(value) || 0;
|
||||
}
|
||||
this.uploadOffset = uploadOffset;
|
||||
this.ajax()
|
||||
this.ajax();
|
||||
}
|
||||
|
||||
progress(event) {
|
||||
const now = Date.now();
|
||||
const speed = (event.loaded - this.uploaded) / (now - this.lastUptime) * 1000;
|
||||
const [speedValue, speedUnit] = formatSize(speed);
|
||||
const [speedValue, speedUnit] = formatFileSize(speed);
|
||||
const speedText = `${speedValue} ${speedUnit}/s`;
|
||||
const progress = formatPercent(((event.loaded + this.uploadOffset) / this.file.size) * 100);
|
||||
const duration = formatDuration((event.total - event.loaded) / speed)
|
||||
const duration = formatDuration((event.total - event.loaded) / speed);
|
||||
this.$uploadStatus.innerHTML = `<span style="width: 80px;">${speedText}</span><span>${progress} ${duration}</span>`;
|
||||
this.uploaded = event.loaded;
|
||||
this.lastUptime = now;
|
||||
@@ -274,7 +297,7 @@ Uploader.runQueue = async () => {
|
||||
if (!Uploader.auth) {
|
||||
Uploader.auth = true;
|
||||
try {
|
||||
await checkAuth()
|
||||
await checkAuth();
|
||||
} catch {
|
||||
Uploader.auth = false;
|
||||
}
|
||||
@@ -319,11 +342,12 @@ function addBreadcrumb(href, uri_prefix) {
|
||||
}
|
||||
}
|
||||
|
||||
function setupIndexPage() {
|
||||
async function setupIndexPage() {
|
||||
if (DATA.allow_archive) {
|
||||
const $download = document.querySelector(".download");
|
||||
$download.href = baseUrl() + "?zip";
|
||||
$download.title = "Download folder as a .zip file";
|
||||
$download.classList.add("dlwt");
|
||||
$download.classList.remove("hidden");
|
||||
}
|
||||
|
||||
@@ -335,15 +359,19 @@ function setupIndexPage() {
|
||||
}
|
||||
|
||||
if (DATA.auth) {
|
||||
setupAuth();
|
||||
await setupAuth();
|
||||
}
|
||||
|
||||
if (DATA.allow_search) {
|
||||
setupSearch()
|
||||
setupSearch();
|
||||
}
|
||||
|
||||
renderPathsTableHead();
|
||||
renderPathsTableBody();
|
||||
|
||||
if (DATA.user) {
|
||||
setupDownloadWithToken();
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
@@ -402,7 +430,7 @@ function renderPathsTableBody() {
|
||||
addPath(DATA.paths[i], i);
|
||||
}
|
||||
} else {
|
||||
$emptyFolder.textContent = dirEmptyNote;
|
||||
$emptyFolder.textContent = DIR_EMPTY_NOTE;
|
||||
$emptyFolder.classList.remove("hidden");
|
||||
}
|
||||
}
|
||||
@@ -414,7 +442,7 @@ function renderPathsTableBody() {
|
||||
*/
|
||||
function addPath(file, index) {
|
||||
const encodedName = encodedStr(file.name);
|
||||
let url = newUrl(file.name)
|
||||
let url = newUrl(file.name);
|
||||
let actionDelete = "";
|
||||
let actionDownload = "";
|
||||
let actionMove = "";
|
||||
@@ -426,13 +454,13 @@ function addPath(file, index) {
|
||||
if (DATA.allow_archive) {
|
||||
actionDownload = `
|
||||
<div class="action-btn">
|
||||
<a href="${url}?zip" title="Download folder as a .zip file">${ICONS.download}</a>
|
||||
<a class="dlwt" href="${url}?zip" title="Download folder as a .zip file" download>${ICONS.download}</a>
|
||||
</div>`;
|
||||
}
|
||||
} else {
|
||||
actionDownload = `
|
||||
<div class="action-btn" >
|
||||
<a href="${url}" title="Download file" download>${ICONS.download}</a>
|
||||
<a class="dlwt" href="${url}" title="Download file" download>${ICONS.download}</a>
|
||||
</div>`;
|
||||
}
|
||||
if (DATA.allow_delete) {
|
||||
@@ -455,7 +483,9 @@ function addPath(file, index) {
|
||||
${actionMove}
|
||||
${actionDelete}
|
||||
${actionEdit}
|
||||
</td>`
|
||||
</td>`;
|
||||
|
||||
let sizeDisplay = isDir ? formatDirSize(file.size) : formatFileSize(file.size).join(" ");
|
||||
|
||||
$pathsTableBody.insertAdjacentHTML("beforeend", `
|
||||
<tr id="addPath${index}">
|
||||
@@ -466,9 +496,9 @@ function addPath(file, index) {
|
||||
<a href="${url}" ${isDir ? "" : `target="_blank"`}>${encodedName}</a>
|
||||
</td>
|
||||
<td class="cell-mtime">${formatMtime(file.mtime)}</td>
|
||||
<td class="cell-size">${formatSize(file.size).join(" ")}</td>
|
||||
<td class="cell-size">${sizeDisplay}</td>
|
||||
${actionCell}
|
||||
</tr>`)
|
||||
</tr>`);
|
||||
}
|
||||
|
||||
function setupDropzone() {
|
||||
@@ -480,7 +510,7 @@ function setupDropzone() {
|
||||
});
|
||||
document.addEventListener("drop", async e => {
|
||||
if (!e.dataTransfer.items[0].webkitGetAsEntry) {
|
||||
const files = e.dataTransfer.files.filter(v => v.size > 0);
|
||||
const files = Array.from(e.dataTransfer.files).filter(v => v.size > 0);
|
||||
for (const file of files) {
|
||||
new Uploader(file, []).upload();
|
||||
}
|
||||
@@ -490,29 +520,54 @@ function setupDropzone() {
|
||||
for (let i = 0; i < len; i++) {
|
||||
entries.push(e.dataTransfer.items[i].webkitGetAsEntry());
|
||||
}
|
||||
addFileEntries(entries, [])
|
||||
addFileEntries(entries, []);
|
||||
}
|
||||
});
|
||||
}
|
||||
|
||||
function setupAuth() {
|
||||
async function setupAuth() {
|
||||
if (DATA.user) {
|
||||
$userBtn.classList.remove("hidden");
|
||||
$logoutBtn.classList.remove("hidden");
|
||||
$logoutBtn.addEventListener("click", logout);
|
||||
$userName.textContent = DATA.user;
|
||||
} else {
|
||||
const $loginBtn = document.querySelector(".login-btn");
|
||||
$loginBtn.classList.remove("hidden");
|
||||
$loginBtn.addEventListener("click", async () => {
|
||||
try {
|
||||
await checkAuth()
|
||||
location.reload();
|
||||
} catch (err) {
|
||||
alert(err.message);
|
||||
}
|
||||
await checkAuth();
|
||||
} catch { }
|
||||
location.reload();
|
||||
});
|
||||
}
|
||||
}
|
||||
|
||||
function setupDownloadWithToken() {
|
||||
document.querySelectorAll("a.dlwt").forEach(link => {
|
||||
link.addEventListener("click", async e => {
|
||||
e.preventDefault();
|
||||
try {
|
||||
const link = e.currentTarget || e.target;
|
||||
const originalHref = link.getAttribute("href");
|
||||
const tokengenUrl = new URL(originalHref);
|
||||
tokengenUrl.searchParams.set("tokengen", "");
|
||||
const res = await fetch(tokengenUrl);
|
||||
if (!res.ok) throw new Error("Failed to fetch token");
|
||||
const token = await res.text();
|
||||
const downloadUrl = new URL(originalHref);
|
||||
downloadUrl.searchParams.set("token", token);
|
||||
const tempA = document.createElement("a");
|
||||
tempA.href = downloadUrl.toString();
|
||||
tempA.download = "";
|
||||
document.body.appendChild(tempA);
|
||||
tempA.click();
|
||||
document.body.removeChild(tempA);
|
||||
} catch (err) {
|
||||
alert(`Failed to download, ${err.message}`);
|
||||
}
|
||||
});
|
||||
});
|
||||
}
|
||||
|
||||
function setupSearch() {
|
||||
const $searchbar = document.querySelector(".searchbar");
|
||||
$searchbar.classList.remove("hidden");
|
||||
@@ -585,11 +640,13 @@ async function setupEditorPage() {
|
||||
await doDeletePath(name, url, () => {
|
||||
location.href = location.href.split("/").slice(0, -1).join("/");
|
||||
});
|
||||
})
|
||||
});
|
||||
|
||||
const $saveBtn = document.querySelector(".save-btn");
|
||||
$saveBtn.classList.remove("hidden");
|
||||
$saveBtn.addEventListener("click", saveChange);
|
||||
if (DATA.editable) {
|
||||
const $saveBtn = document.querySelector(".save-btn");
|
||||
$saveBtn.classList.remove("hidden");
|
||||
$saveBtn.addEventListener("click", saveChange);
|
||||
}
|
||||
} else if (DATA.kind == "View") {
|
||||
$editor.readonly = true;
|
||||
}
|
||||
@@ -599,7 +656,7 @@ async function setupEditorPage() {
|
||||
const url = baseUrl();
|
||||
const ext = extName(baseName(url));
|
||||
if (IFRAME_FORMATS.find(v => v === ext)) {
|
||||
$notEditable.insertAdjacentHTML("afterend", `<iframe src="${url}" sandbox width="100%" height="${window.innerHeight - 100}px"></iframe>`)
|
||||
$notEditable.insertAdjacentHTML("afterend", `<iframe src="${url}" sandbox width="100%" height="${window.innerHeight - 100}px"></iframe>`);
|
||||
} else {
|
||||
$notEditable.classList.remove("hidden");
|
||||
$notEditable.textContent = "Cannot edit because file is too large or binary.";
|
||||
@@ -616,12 +673,12 @@ async function setupEditorPage() {
|
||||
$editor.value = await res.text();
|
||||
} else {
|
||||
const bytes = await res.arrayBuffer();
|
||||
const dataView = new DataView(bytes)
|
||||
const decoder = new TextDecoder(encoding)
|
||||
const dataView = new DataView(bytes);
|
||||
const decoder = new TextDecoder(encoding);
|
||||
$editor.value = decoder.decode(dataView);
|
||||
}
|
||||
} catch (err) {
|
||||
alert(`Failed get file, ${err.message}`);
|
||||
alert(`Failed to get file, ${err.message}`);
|
||||
}
|
||||
}
|
||||
|
||||
@@ -638,10 +695,10 @@ async function deletePath(index) {
|
||||
DATA.paths[index] = null;
|
||||
if (!DATA.paths.find(v => !!v)) {
|
||||
$pathsTable.classList.add("hidden");
|
||||
$emptyFolder.textContent = dirEmptyNote;
|
||||
$emptyFolder.textContent = DIR_EMPTY_NOTE;
|
||||
$emptyFolder.classList.remove("hidden");
|
||||
}
|
||||
})
|
||||
});
|
||||
}
|
||||
|
||||
async function doDeletePath(name, url, cb) {
|
||||
@@ -674,13 +731,13 @@ async function movePath(index) {
|
||||
}
|
||||
|
||||
async function doMovePath(fileUrl) {
|
||||
const fileUrlObj = new URL(fileUrl)
|
||||
const fileUrlObj = new URL(fileUrl);
|
||||
|
||||
const prefix = DATA.uri_prefix.slice(0, -1);
|
||||
|
||||
const filePath = decodeURIComponent(fileUrlObj.pathname.slice(prefix.length));
|
||||
|
||||
let newPath = prompt("Enter new path", filePath)
|
||||
let newPath = prompt("Enter new path", filePath);
|
||||
if (!newPath) return;
|
||||
if (!newPath.startsWith("/")) newPath = "/" + newPath;
|
||||
if (filePath === newPath) return;
|
||||
@@ -728,12 +785,23 @@ async function saveChange() {
|
||||
async function checkAuth() {
|
||||
if (!DATA.auth) return;
|
||||
const res = await fetch(baseUrl(), {
|
||||
method: "WRITEABLE",
|
||||
method: "CHECKAUTH",
|
||||
});
|
||||
await assertResOK(res);
|
||||
document.querySelector(".login-btn").classList.add("hidden");
|
||||
$userBtn.classList.remove("hidden");
|
||||
$userName.textContent = "";
|
||||
$loginBtn.classList.add("hidden");
|
||||
$logoutBtn.classList.remove("hidden");
|
||||
$userName.textContent = await res.text();
|
||||
}
|
||||
|
||||
function logout() {
|
||||
if (!DATA.auth) return;
|
||||
const url = baseUrl();
|
||||
const xhr = new XMLHttpRequest();
|
||||
xhr.open("LOGOUT", url, true, DATA.user);
|
||||
xhr.onload = () => {
|
||||
location.href = url;
|
||||
}
|
||||
xhr.send();
|
||||
}
|
||||
|
||||
/**
|
||||
@@ -799,11 +867,11 @@ function newUrl(name) {
|
||||
}
|
||||
|
||||
function baseUrl() {
|
||||
return location.href.split('?')[0];
|
||||
return location.href.split(/[?#]/)[0];
|
||||
}
|
||||
|
||||
function baseName(url) {
|
||||
return decodeURIComponent(url.split("/").filter(v => v.length > 0).slice(-1)[0])
|
||||
return decodeURIComponent(url.split("/").filter(v => v.length > 0).slice(-1)[0]);
|
||||
}
|
||||
|
||||
function extName(filename) {
|
||||
@@ -830,7 +898,7 @@ function getPathSvg(path_type) {
|
||||
}
|
||||
|
||||
function formatMtime(mtime) {
|
||||
if (!mtime) return ""
|
||||
if (!mtime) return "";
|
||||
const date = new Date(mtime);
|
||||
const year = date.getFullYear();
|
||||
const month = padZero(date.getMonth() + 1, 2);
|
||||
@@ -841,17 +909,23 @@ function formatMtime(mtime) {
|
||||
}
|
||||
|
||||
function padZero(value, size) {
|
||||
return ("0".repeat(size) + value).slice(-1 * size)
|
||||
return ("0".repeat(size) + value).slice(-1 * size);
|
||||
}
|
||||
|
||||
function formatSize(size) {
|
||||
if (size == null) return [0, "B"]
|
||||
function formatDirSize(size) {
|
||||
const unit = size === 1 ? "item" : "items";
|
||||
const num = size >= MAX_SUBPATHS_COUNT ? `>${MAX_SUBPATHS_COUNT - 1}` : `${size}`;
|
||||
return ` ${num} ${unit}`;
|
||||
}
|
||||
|
||||
function formatFileSize(size) {
|
||||
if (size == null) return [0, "B"];
|
||||
const sizes = ['B', 'KB', 'MB', 'GB', 'TB'];
|
||||
if (size == 0) return [0, "B"];
|
||||
const i = parseInt(Math.floor(Math.log(size) / Math.log(1024)));
|
||||
ratio = 1
|
||||
let ratio = 1;
|
||||
if (i >= 3) {
|
||||
ratio = 100
|
||||
ratio = 100;
|
||||
}
|
||||
return [Math.round(size * ratio / Math.pow(1024, i), 2) / ratio, sizes[i]];
|
||||
}
|
||||
@@ -860,7 +934,7 @@ function formatDuration(seconds) {
|
||||
seconds = Math.ceil(seconds);
|
||||
const h = Math.floor(seconds / 3600);
|
||||
const m = Math.floor((seconds - h * 3600) / 60);
|
||||
const s = seconds - h * 3600 - m * 60
|
||||
const s = seconds - h * 3600 - m * 60;
|
||||
return `${padZero(h, 2)}:${padZero(m, 2)}:${padZero(s, 2)}`;
|
||||
}
|
||||
|
||||
@@ -889,8 +963,27 @@ function getEncoding(contentType) {
|
||||
if (/charset/i.test(charset)) {
|
||||
let encoding = charset.split("=")[1];
|
||||
if (encoding) {
|
||||
return encoding.toLowerCase()
|
||||
return encoding.toLowerCase();
|
||||
}
|
||||
}
|
||||
return 'utf-8'
|
||||
return 'utf-8';
|
||||
}
|
||||
|
||||
// Parsing base64 strings with Unicode characters
|
||||
function decodeBase64(base64String) {
|
||||
const binString = atob(base64String);
|
||||
const len = binString.length;
|
||||
const bytes = new Uint8Array(len);
|
||||
const arr = new Uint32Array(bytes.buffer, 0, Math.floor(len / 4));
|
||||
let i = 0;
|
||||
for (; i < arr.length; i++) {
|
||||
arr[i] = binString.charCodeAt(i * 4) |
|
||||
(binString.charCodeAt(i * 4 + 1) << 8) |
|
||||
(binString.charCodeAt(i * 4 + 2) << 16) |
|
||||
(binString.charCodeAt(i * 4 + 3) << 24);
|
||||
}
|
||||
for (i = i * 4; i < len; i++) {
|
||||
bytes[i] = binString.charCodeAt(i);
|
||||
}
|
||||
return new TextDecoder().decode(bytes);
|
||||
}
|
||||
|
||||
26
src/args.rs
26
src/args.rs
@@ -146,7 +146,7 @@ pub fn build_cli() -> Command {
|
||||
.hide_env(true)
|
||||
.long("allow-archive")
|
||||
.action(ArgAction::SetTrue)
|
||||
.help("Allow zip archive generation"),
|
||||
.help("Allow download folders as archive file"),
|
||||
)
|
||||
.arg(
|
||||
Arg::new("enable-cors")
|
||||
@@ -461,28 +461,30 @@ impl Args {
|
||||
|
||||
#[derive(Debug, Clone, PartialEq, Eq, PartialOrd, Ord)]
|
||||
pub enum BindAddr {
|
||||
Address(IpAddr),
|
||||
Path(PathBuf),
|
||||
IpAddr(IpAddr),
|
||||
#[cfg(unix)]
|
||||
SocketPath(String),
|
||||
}
|
||||
|
||||
impl BindAddr {
|
||||
fn parse_addrs(addrs: &[&str]) -> Result<Vec<Self>> {
|
||||
let mut bind_addrs = vec![];
|
||||
#[cfg(not(unix))]
|
||||
let mut invalid_addrs = vec![];
|
||||
for addr in addrs {
|
||||
match addr.parse::<IpAddr>() {
|
||||
Ok(v) => {
|
||||
bind_addrs.push(BindAddr::Address(v));
|
||||
bind_addrs.push(BindAddr::IpAddr(v));
|
||||
}
|
||||
Err(_) => {
|
||||
if cfg!(unix) {
|
||||
bind_addrs.push(BindAddr::Path(PathBuf::from(addr)));
|
||||
} else {
|
||||
invalid_addrs.push(*addr);
|
||||
}
|
||||
#[cfg(unix)]
|
||||
bind_addrs.push(BindAddr::SocketPath(addr.to_string()));
|
||||
#[cfg(not(unix))]
|
||||
invalid_addrs.push(*addr);
|
||||
}
|
||||
}
|
||||
}
|
||||
#[cfg(not(unix))]
|
||||
if !invalid_addrs.is_empty() {
|
||||
bail!("Invalid bind address `{}`", invalid_addrs.join(","));
|
||||
}
|
||||
@@ -710,7 +712,7 @@ hidden: tmp,*.log,*.lock
|
||||
assert_eq!(args.serve_path, Args::sanitize_path(&tmpdir).unwrap());
|
||||
assert_eq!(
|
||||
args.addrs,
|
||||
vec![BindAddr::Address("0.0.0.0".parse().unwrap())]
|
||||
vec![BindAddr::IpAddr("0.0.0.0".parse().unwrap())]
|
||||
);
|
||||
assert_eq!(args.hidden, ["tmp", "*.log", "*.lock"]);
|
||||
assert_eq!(args.port, 3000);
|
||||
@@ -740,8 +742,8 @@ hidden:
|
||||
assert_eq!(
|
||||
args.addrs,
|
||||
vec![
|
||||
BindAddr::Address("127.0.0.1".parse().unwrap()),
|
||||
BindAddr::Address("192.168.8.10".parse().unwrap())
|
||||
BindAddr::IpAddr("127.0.0.1".parse().unwrap()),
|
||||
BindAddr::IpAddr("192.168.8.10".parse().unwrap())
|
||||
]
|
||||
);
|
||||
assert_eq!(args.hidden, ["tmp", "*.log", "*.lock"]);
|
||||
|
||||
196
src/auth.rs
196
src/auth.rs
@@ -1,12 +1,14 @@
|
||||
use crate::{args::Args, server::Response, utils::unix_now};
|
||||
|
||||
use anyhow::{anyhow, bail, Result};
|
||||
use base64::{engine::general_purpose, Engine as _};
|
||||
use base64::{engine::general_purpose::STANDARD, Engine as _};
|
||||
use ed25519_dalek::{ed25519::signature::SignerMut, Signature, SigningKey};
|
||||
use headers::HeaderValue;
|
||||
use hyper::{header::WWW_AUTHENTICATE, Method};
|
||||
use indexmap::IndexMap;
|
||||
use lazy_static::lazy_static;
|
||||
use md5::Context;
|
||||
use sha2::{Digest, Sha256};
|
||||
use std::{
|
||||
collections::HashMap,
|
||||
path::{Path, PathBuf},
|
||||
@@ -14,7 +16,8 @@ use std::{
|
||||
use uuid::Uuid;
|
||||
|
||||
const REALM: &str = "DUFS";
|
||||
const DIGEST_AUTH_TIMEOUT: u32 = 604800; // 7 days
|
||||
const DIGEST_AUTH_TIMEOUT: u32 = 60 * 60 * 24 * 7; // 7 days
|
||||
const TOKEN_EXPIRATION: u64 = 1000 * 60 * 60 * 24 * 3; // 3 days
|
||||
|
||||
lazy_static! {
|
||||
static ref NONCESTARTHASH: Context = {
|
||||
@@ -69,15 +72,20 @@ impl AccessControl {
|
||||
let mut anonymous = None;
|
||||
if let Some(paths) = annoy_paths {
|
||||
let mut access_paths = AccessPaths::default();
|
||||
access_paths.merge(paths);
|
||||
access_paths
|
||||
.merge(paths)
|
||||
.ok_or_else(|| anyhow!("Invalid auth value `@{paths}"))?;
|
||||
anonymous = Some(access_paths);
|
||||
}
|
||||
let mut users = IndexMap::new();
|
||||
for (user, pass, paths) in account_paths_pairs.into_iter() {
|
||||
let mut access_paths = anonymous.clone().unwrap_or_default();
|
||||
let mut access_paths = AccessPaths::default();
|
||||
access_paths
|
||||
.merge(paths)
|
||||
.ok_or_else(|| anyhow!("Invalid auth `{user}:{pass}@{paths}"))?;
|
||||
.ok_or_else(|| anyhow!("Invalid auth value `{user}:{pass}@{paths}"))?;
|
||||
if let Some(paths) = annoy_paths {
|
||||
access_paths.merge(paths);
|
||||
}
|
||||
if pass.starts_with("$6$") {
|
||||
use_hashed_password = true;
|
||||
}
|
||||
@@ -100,16 +108,29 @@ impl AccessControl {
|
||||
path: &str,
|
||||
method: &Method,
|
||||
authorization: Option<&HeaderValue>,
|
||||
token: Option<&String>,
|
||||
guard_options: bool,
|
||||
) -> (Option<String>, Option<AccessPaths>) {
|
||||
if self.users.is_empty() {
|
||||
return (None, Some(AccessPaths::new(AccessPerm::ReadWrite)));
|
||||
}
|
||||
|
||||
if method == Method::GET {
|
||||
if let Some(token) = token {
|
||||
if let Ok((user, ap)) = self.verify_token(token, path) {
|
||||
return (Some(user), ap.guard(path, method));
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
if let Some(authorization) = authorization {
|
||||
if let Some(user) = get_auth_user(authorization) {
|
||||
if let Some((pass, paths)) = self.users.get(&user) {
|
||||
if let Some((pass, ap)) = self.users.get(&user) {
|
||||
if method == Method::OPTIONS {
|
||||
return (Some(user), Some(AccessPaths::new(AccessPerm::ReadOnly)));
|
||||
}
|
||||
if check_auth(authorization, method.as_str(), &user, pass).is_some() {
|
||||
return (Some(user), paths.find(path, !is_readonly_method(method)));
|
||||
return (Some(user), ap.guard(path, method));
|
||||
}
|
||||
}
|
||||
}
|
||||
@@ -121,12 +142,55 @@ impl AccessControl {
|
||||
return (None, Some(AccessPaths::new(AccessPerm::ReadOnly)));
|
||||
}
|
||||
|
||||
if let Some(paths) = self.anonymous.as_ref() {
|
||||
return (None, paths.find(path, !is_readonly_method(method)));
|
||||
if let Some(ap) = self.anonymous.as_ref() {
|
||||
return (None, ap.guard(path, method));
|
||||
}
|
||||
|
||||
(None, None)
|
||||
}
|
||||
|
||||
pub fn generate_token(&self, path: &str, user: &str) -> Result<String> {
|
||||
let (pass, _) = self
|
||||
.users
|
||||
.get(user)
|
||||
.ok_or_else(|| anyhow!("Not found user '{user}'"))?;
|
||||
let exp = unix_now().as_millis() as u64 + TOKEN_EXPIRATION;
|
||||
let message = format!("{path}:{exp}");
|
||||
let mut signing_key = derive_secret_key(user, pass);
|
||||
let sig = signing_key.sign(message.as_bytes()).to_bytes();
|
||||
|
||||
let mut raw = Vec::with_capacity(64 + 8 + user.len());
|
||||
raw.extend_from_slice(&sig);
|
||||
raw.extend_from_slice(&exp.to_be_bytes());
|
||||
raw.extend_from_slice(user.as_bytes());
|
||||
|
||||
Ok(hex::encode(raw))
|
||||
}
|
||||
|
||||
fn verify_token<'a>(&'a self, token: &str, path: &str) -> Result<(String, &'a AccessPaths)> {
|
||||
let raw = hex::decode(token)?;
|
||||
|
||||
let sig_bytes = &raw[..64];
|
||||
let exp_bytes = &raw[64..72];
|
||||
let user_bytes = &raw[72..];
|
||||
|
||||
let exp = u64::from_be_bytes(exp_bytes.try_into()?);
|
||||
if unix_now().as_millis() as u64 > exp {
|
||||
bail!("Token expired");
|
||||
}
|
||||
|
||||
let user = std::str::from_utf8(user_bytes)?;
|
||||
let (pass, ap) = self
|
||||
.users
|
||||
.get(user)
|
||||
.ok_or_else(|| anyhow!("Not found user '{user}'"))?;
|
||||
|
||||
let sig = Signature::from_bytes(&<[u8; 64]>::try_from(sig_bytes)?);
|
||||
|
||||
let message = format!("{path}:{exp}");
|
||||
derive_secret_key(user, pass).verify(message.as_bytes(), &sig)?;
|
||||
Ok((user.to_string(), ap))
|
||||
}
|
||||
}
|
||||
|
||||
#[derive(Debug, Default, Clone, PartialEq, Eq)]
|
||||
@@ -148,8 +212,9 @@ impl AccessPaths {
|
||||
}
|
||||
|
||||
pub fn set_perm(&mut self, perm: AccessPerm) {
|
||||
if !perm.indexonly() {
|
||||
if self.perm < perm {
|
||||
self.perm = perm;
|
||||
self.recursively_purge_children(perm);
|
||||
}
|
||||
}
|
||||
|
||||
@@ -166,6 +231,25 @@ impl AccessPaths {
|
||||
Some(())
|
||||
}
|
||||
|
||||
pub fn guard(&self, path: &str, method: &Method) -> Option<Self> {
|
||||
let target = self.find(path)?;
|
||||
if !is_readonly_method(method) && !target.perm().readwrite() {
|
||||
return None;
|
||||
}
|
||||
Some(target)
|
||||
}
|
||||
|
||||
fn recursively_purge_children(&mut self, perm: AccessPerm) {
|
||||
self.children.retain(|_, child| {
|
||||
if child.perm <= perm {
|
||||
false
|
||||
} else {
|
||||
child.recursively_purge_children(perm);
|
||||
true
|
||||
}
|
||||
});
|
||||
}
|
||||
|
||||
fn add(&mut self, path: &str, perm: AccessPerm) {
|
||||
let path = path.trim_matches('/');
|
||||
if path.is_empty() {
|
||||
@@ -182,21 +266,20 @@ impl AccessPaths {
|
||||
self.set_perm(perm);
|
||||
return;
|
||||
}
|
||||
if self.perm >= perm {
|
||||
return;
|
||||
}
|
||||
let child = self.children.entry(parts[0].to_string()).or_default();
|
||||
child.add_impl(&parts[1..], perm)
|
||||
}
|
||||
|
||||
pub fn find(&self, path: &str, writable: bool) -> Option<AccessPaths> {
|
||||
pub fn find(&self, path: &str) -> Option<AccessPaths> {
|
||||
let parts: Vec<&str> = path
|
||||
.trim_matches('/')
|
||||
.split('/')
|
||||
.filter(|v| !v.is_empty())
|
||||
.collect();
|
||||
let target = self.find_impl(&parts, self.perm)?;
|
||||
if writable && !target.perm().readwrite() {
|
||||
return None;
|
||||
}
|
||||
Some(target)
|
||||
self.find_impl(&parts, self.perm)
|
||||
}
|
||||
|
||||
fn find_impl(&self, parts: &[&str], perm: AccessPerm) -> Option<AccessPaths> {
|
||||
@@ -229,20 +312,20 @@ impl AccessPaths {
|
||||
self.children.keys().collect()
|
||||
}
|
||||
|
||||
pub fn child_paths(&self, base: &Path) -> Vec<PathBuf> {
|
||||
pub fn entry_paths(&self, base: &Path) -> Vec<PathBuf> {
|
||||
if !self.perm().indexonly() {
|
||||
return vec![base.to_path_buf()];
|
||||
}
|
||||
let mut output = vec![];
|
||||
self.child_paths_impl(&mut output, base);
|
||||
self.entry_paths_impl(&mut output, base);
|
||||
output
|
||||
}
|
||||
|
||||
fn child_paths_impl(&self, output: &mut Vec<PathBuf>, base: &Path) {
|
||||
fn entry_paths_impl(&self, output: &mut Vec<PathBuf>, base: &Path) {
|
||||
for (name, child) in self.children.iter() {
|
||||
let base = base.join(name);
|
||||
if child.perm().indexonly() {
|
||||
child.child_paths_impl(output, &base);
|
||||
child.entry_paths_impl(output, &base);
|
||||
} else {
|
||||
output.push(base)
|
||||
}
|
||||
@@ -270,15 +353,14 @@ impl AccessPerm {
|
||||
|
||||
pub fn www_authenticate(res: &mut Response, args: &Args) -> Result<()> {
|
||||
if args.auth.use_hashed_password {
|
||||
let basic = HeaderValue::from_str(&format!("Basic realm=\"{}\"", REALM))?;
|
||||
let basic = HeaderValue::from_str(&format!("Basic realm=\"{REALM}\""))?;
|
||||
res.headers_mut().insert(WWW_AUTHENTICATE, basic);
|
||||
} else {
|
||||
let nonce = create_nonce()?;
|
||||
let digest = HeaderValue::from_str(&format!(
|
||||
"Digest realm=\"{}\", nonce=\"{}\", qop=\"auth\"",
|
||||
REALM, nonce
|
||||
"Digest realm=\"{REALM}\", nonce=\"{nonce}\", qop=\"auth\""
|
||||
))?;
|
||||
let basic = HeaderValue::from_str(&format!("Basic realm=\"{}\"", REALM))?;
|
||||
let basic = HeaderValue::from_str(&format!("Basic realm=\"{REALM}\""))?;
|
||||
res.headers_mut().append(WWW_AUTHENTICATE, digest);
|
||||
res.headers_mut().append(WWW_AUTHENTICATE, basic);
|
||||
}
|
||||
@@ -287,7 +369,7 @@ pub fn www_authenticate(res: &mut Response, args: &Args) -> Result<()> {
|
||||
|
||||
pub fn get_auth_user(authorization: &HeaderValue) -> Option<String> {
|
||||
if let Some(value) = strip_prefix(authorization.as_bytes(), b"Basic ") {
|
||||
let value: Vec<u8> = general_purpose::STANDARD.decode(value).ok()?;
|
||||
let value: Vec<u8> = STANDARD.decode(value).ok()?;
|
||||
let parts: Vec<&str> = std::str::from_utf8(&value).ok()?.split(':').collect();
|
||||
Some(parts[0].to_string())
|
||||
} else if let Some(value) = strip_prefix(authorization.as_bytes(), b"Digest ") {
|
||||
@@ -306,18 +388,18 @@ pub fn check_auth(
|
||||
auth_pass: &str,
|
||||
) -> Option<()> {
|
||||
if let Some(value) = strip_prefix(authorization.as_bytes(), b"Basic ") {
|
||||
let value: Vec<u8> = general_purpose::STANDARD.decode(value).ok()?;
|
||||
let parts: Vec<&str> = std::str::from_utf8(&value).ok()?.split(':').collect();
|
||||
let value: Vec<u8> = STANDARD.decode(value).ok()?;
|
||||
let (user, pass) = std::str::from_utf8(&value).ok()?.split_once(':')?;
|
||||
|
||||
if parts[0] != auth_user {
|
||||
if user != auth_user {
|
||||
return None;
|
||||
}
|
||||
|
||||
if auth_pass.starts_with("$6$") {
|
||||
if let Ok(()) = sha_crypt::sha512_check(parts[1], auth_pass) {
|
||||
if let Ok(()) = sha_crypt::sha512_check(pass, auth_pass) {
|
||||
return Some(());
|
||||
}
|
||||
} else if parts[1] == auth_pass {
|
||||
} else if pass == auth_pass {
|
||||
return Some(());
|
||||
}
|
||||
|
||||
@@ -340,8 +422,8 @@ pub fn check_auth(
|
||||
}
|
||||
|
||||
let mut h = Context::new();
|
||||
h.consume(format!("{}:{}:{}", auth_user, REALM, auth_pass).as_bytes());
|
||||
let auth_pass = format!("{:x}", h.compute());
|
||||
h.consume(format!("{auth_user}:{REALM}:{auth_pass}").as_bytes());
|
||||
let auth_pass = format!("{:x}", h.finalize());
|
||||
|
||||
let mut ha = Context::new();
|
||||
ha.consume(method);
|
||||
@@ -349,7 +431,7 @@ pub fn check_auth(
|
||||
if let Some(uri) = digest_map.get(b"uri".as_ref()) {
|
||||
ha.consume(uri);
|
||||
}
|
||||
let ha = format!("{:x}", ha.compute());
|
||||
let ha = format!("{:x}", ha.finalize());
|
||||
let mut correct_response = None;
|
||||
if let Some(qop) = digest_map.get(b"qop".as_ref()) {
|
||||
if qop == &b"auth".as_ref() || qop == &b"auth-int".as_ref() {
|
||||
@@ -370,7 +452,7 @@ pub fn check_auth(
|
||||
c.consume(qop);
|
||||
c.consume(b":");
|
||||
c.consume(&*ha);
|
||||
format!("{:x}", c.compute())
|
||||
format!("{:x}", c.finalize())
|
||||
});
|
||||
}
|
||||
}
|
||||
@@ -383,7 +465,7 @@ pub fn check_auth(
|
||||
c.consume(nonce);
|
||||
c.consume(b":");
|
||||
c.consume(&*ha);
|
||||
format!("{:x}", c.compute())
|
||||
format!("{:x}", c.finalize())
|
||||
}
|
||||
};
|
||||
if correct_response.as_bytes() == *user_response {
|
||||
@@ -396,6 +478,13 @@ pub fn check_auth(
|
||||
}
|
||||
}
|
||||
|
||||
fn derive_secret_key(user: &str, pass: &str) -> SigningKey {
|
||||
let mut hasher = Sha256::new();
|
||||
hasher.update(format!("{user}:{pass}").as_bytes());
|
||||
let hash = hasher.finalize();
|
||||
SigningKey::from_bytes(&hash.into())
|
||||
}
|
||||
|
||||
/// Check if a nonce is still valid.
|
||||
/// Return an error if it was never valid
|
||||
fn validate_nonce(nonce: &[u8]) -> Result<bool> {
|
||||
@@ -407,14 +496,14 @@ fn validate_nonce(nonce: &[u8]) -> Result<bool> {
|
||||
//get time
|
||||
if let Ok(secs_nonce) = u32::from_str_radix(&n[..8], 16) {
|
||||
//check time
|
||||
let now = unix_now()?;
|
||||
let now = unix_now();
|
||||
let secs_now = now.as_secs() as u32;
|
||||
|
||||
if let Some(dur) = secs_now.checked_sub(secs_nonce) {
|
||||
//check hash
|
||||
let mut h = NONCESTARTHASH.clone();
|
||||
h.consume(secs_nonce.to_be_bytes());
|
||||
let h = format!("{:x}", h.compute());
|
||||
let h = format!("{:x}", h.finalize());
|
||||
if h[..26] == n[8..34] {
|
||||
return Ok(dur < DIGEST_AUTH_TIMEOUT);
|
||||
}
|
||||
@@ -429,6 +518,8 @@ fn is_readonly_method(method: &Method) -> bool {
|
||||
|| method == Method::OPTIONS
|
||||
|| method == Method::HEAD
|
||||
|| method.as_str() == "PROPFIND"
|
||||
|| method.as_str() == "CHECKAUTH"
|
||||
|| method.as_str() == "LOGOUT"
|
||||
}
|
||||
|
||||
fn strip_prefix<'a>(search: &'a [u8], prefix: &[u8]) -> Option<&'a [u8]> {
|
||||
@@ -485,12 +576,12 @@ fn to_headermap(header: &[u8]) -> Result<HashMap<&[u8], &[u8]>, ()> {
|
||||
}
|
||||
|
||||
fn create_nonce() -> Result<String> {
|
||||
let now = unix_now()?;
|
||||
let now = unix_now();
|
||||
let secs = now.as_secs() as u32;
|
||||
let mut h = NONCESTARTHASH.clone();
|
||||
h.consume(secs.to_be_bytes());
|
||||
|
||||
let n = format!("{:08x}{:032x}", secs, h.compute());
|
||||
let n = format!("{:08x}{:032x}", secs, h.finalize());
|
||||
Ok(n[..34].to_string())
|
||||
}
|
||||
|
||||
@@ -572,7 +663,7 @@ mod tests {
|
||||
paths.add("/dir2/dir22/dir221", AccessPerm::ReadWrite);
|
||||
paths.add("/dir2/dir23/dir231", AccessPerm::ReadWrite);
|
||||
assert_eq!(
|
||||
paths.child_paths(Path::new("/tmp")),
|
||||
paths.entry_paths(Path::new("/tmp")),
|
||||
[
|
||||
"/tmp/dir1",
|
||||
"/tmp/dir2/dir21",
|
||||
@@ -585,8 +676,8 @@ mod tests {
|
||||
);
|
||||
assert_eq!(
|
||||
paths
|
||||
.find("dir2", false)
|
||||
.map(|v| v.child_paths(Path::new("/tmp/dir2"))),
|
||||
.find("dir2")
|
||||
.map(|v| v.entry_paths(Path::new("/tmp/dir2"))),
|
||||
Some(
|
||||
[
|
||||
"/tmp/dir2/dir21",
|
||||
@@ -598,19 +689,30 @@ mod tests {
|
||||
.collect::<Vec<_>>()
|
||||
)
|
||||
);
|
||||
assert_eq!(paths.find("dir2", true), None);
|
||||
assert_eq!(
|
||||
paths.find("dir1/file", true),
|
||||
paths.find("dir1/file"),
|
||||
Some(AccessPaths::new(AccessPerm::ReadWrite))
|
||||
);
|
||||
assert_eq!(
|
||||
paths.find("dir2/dir21/file", true),
|
||||
paths.find("dir2/dir21/file"),
|
||||
Some(AccessPaths::new(AccessPerm::ReadWrite))
|
||||
);
|
||||
assert_eq!(
|
||||
paths.find("dir2/dir21/dir211/file", false),
|
||||
paths.find("dir2/dir21/dir211/file"),
|
||||
Some(AccessPaths::new(AccessPerm::ReadWrite))
|
||||
);
|
||||
assert_eq!(
|
||||
paths.find("dir2/dir22/file"),
|
||||
Some(AccessPaths::new(AccessPerm::ReadOnly))
|
||||
);
|
||||
assert_eq!(paths.find("dir2/dir21/dir211/file", true), None);
|
||||
assert_eq!(
|
||||
paths.find("dir2/dir22/dir221/file"),
|
||||
Some(AccessPaths::new(AccessPerm::ReadWrite))
|
||||
);
|
||||
assert_eq!(paths.find("dir2/dir23/file"), None);
|
||||
assert_eq!(
|
||||
paths.find("dir2/dir23//dir231/file"),
|
||||
Some(AccessPaths::new(AccessPerm::ReadWrite))
|
||||
);
|
||||
}
|
||||
}
|
||||
|
||||
@@ -64,8 +64,8 @@ impl HttpLogger {
|
||||
}
|
||||
}
|
||||
match err {
|
||||
Some(err) => error!("{} {}", output, err),
|
||||
None => info!("{}", output),
|
||||
Some(err) => error!("{output} {err}"),
|
||||
None => info!("{output}"),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
59
src/main.rs
59
src/main.rs
@@ -3,6 +3,7 @@ mod auth;
|
||||
mod http_logger;
|
||||
mod http_utils;
|
||||
mod logger;
|
||||
mod noscript;
|
||||
mod server;
|
||||
mod utils;
|
||||
|
||||
@@ -57,7 +58,7 @@ async fn main() -> Result<()> {
|
||||
ret = join_all(handles) => {
|
||||
for r in ret {
|
||||
if let Err(e) = r {
|
||||
error!("{}", e);
|
||||
error!("{e}");
|
||||
}
|
||||
}
|
||||
Ok(())
|
||||
@@ -78,7 +79,7 @@ fn serve(args: Args, running: Arc<AtomicBool>) -> Result<Vec<JoinHandle<()>>> {
|
||||
for bind_addr in addrs.iter() {
|
||||
let server_handle = server_handle.clone();
|
||||
match bind_addr {
|
||||
BindAddr::Address(ip) => {
|
||||
BindAddr::IpAddr(ip) => {
|
||||
let listener = create_listener(SocketAddr::new(*ip, port))
|
||||
.with_context(|| format!("Failed to bind `{ip}:{port}`"))?;
|
||||
|
||||
@@ -140,26 +141,32 @@ fn serve(args: Args, running: Arc<AtomicBool>) -> Result<Vec<JoinHandle<()>>> {
|
||||
}
|
||||
};
|
||||
}
|
||||
BindAddr::Path(path) => {
|
||||
if path.exists() {
|
||||
std::fs::remove_file(path)?;
|
||||
}
|
||||
#[cfg(unix)]
|
||||
#[cfg(unix)]
|
||||
BindAddr::SocketPath(path) => {
|
||||
let socket_path = if path.starts_with("@")
|
||||
&& cfg!(any(target_os = "linux", target_os = "android"))
|
||||
{
|
||||
let listener = tokio::net::UnixListener::bind(path)
|
||||
.with_context(|| format!("Failed to bind `{}`", path.display()))?;
|
||||
let handle = tokio::spawn(async move {
|
||||
loop {
|
||||
let Ok((stream, _addr)) = listener.accept().await else {
|
||||
continue;
|
||||
};
|
||||
let stream = TokioIo::new(stream);
|
||||
tokio::spawn(handle_stream(server_handle.clone(), stream, None));
|
||||
}
|
||||
});
|
||||
let mut path_buf = path.as_bytes().to_vec();
|
||||
path_buf[0] = b'\0';
|
||||
unsafe { std::ffi::OsStr::from_encoded_bytes_unchecked(&path_buf) }
|
||||
.to_os_string()
|
||||
} else {
|
||||
let _ = std::fs::remove_file(path);
|
||||
path.into()
|
||||
};
|
||||
let listener = tokio::net::UnixListener::bind(socket_path)
|
||||
.with_context(|| format!("Failed to bind `{path}`"))?;
|
||||
let handle = tokio::spawn(async move {
|
||||
loop {
|
||||
let Ok((stream, _addr)) = listener.accept().await else {
|
||||
continue;
|
||||
};
|
||||
let stream = TokioIo::new(stream);
|
||||
tokio::spawn(handle_stream(server_handle.clone(), stream, None));
|
||||
}
|
||||
});
|
||||
|
||||
handles.push(handle);
|
||||
}
|
||||
handles.push(handle);
|
||||
}
|
||||
}
|
||||
}
|
||||
@@ -207,7 +214,7 @@ fn check_addrs(args: &Args) -> Result<(Vec<BindAddr>, Vec<BindAddr>)> {
|
||||
let (ipv4_addrs, ipv6_addrs) = interface_addrs()?;
|
||||
for bind_addr in args.addrs.iter() {
|
||||
match bind_addr {
|
||||
BindAddr::Address(ip) => match &ip {
|
||||
BindAddr::IpAddr(ip) => match &ip {
|
||||
IpAddr::V4(_) => {
|
||||
if !ipv4_addrs.is_empty() {
|
||||
new_addrs.push(bind_addr.clone());
|
||||
@@ -229,6 +236,7 @@ fn check_addrs(args: &Args) -> Result<(Vec<BindAddr>, Vec<BindAddr>)> {
|
||||
}
|
||||
}
|
||||
},
|
||||
#[cfg(unix)]
|
||||
_ => {
|
||||
new_addrs.push(bind_addr.clone());
|
||||
print_addrs.push(bind_addr.clone())
|
||||
@@ -246,10 +254,10 @@ fn interface_addrs() -> Result<(Vec<BindAddr>, Vec<BindAddr>)> {
|
||||
for iface in ifaces.into_iter() {
|
||||
let ip = iface.ip();
|
||||
if ip.is_ipv4() {
|
||||
ipv4_addrs.push(BindAddr::Address(ip))
|
||||
ipv4_addrs.push(BindAddr::IpAddr(ip))
|
||||
}
|
||||
if ip.is_ipv6() {
|
||||
ipv6_addrs.push(BindAddr::Address(ip))
|
||||
ipv6_addrs.push(BindAddr::IpAddr(ip))
|
||||
}
|
||||
}
|
||||
Ok((ipv4_addrs, ipv6_addrs))
|
||||
@@ -260,7 +268,7 @@ fn print_listening(args: &Args, print_addrs: &[BindAddr]) -> Result<String> {
|
||||
let urls = print_addrs
|
||||
.iter()
|
||||
.map(|bind_addr| match bind_addr {
|
||||
BindAddr::Address(addr) => {
|
||||
BindAddr::IpAddr(addr) => {
|
||||
let addr = match addr {
|
||||
IpAddr::V4(_) => format!("{}:{}", addr, args.port),
|
||||
IpAddr::V6(_) => format!("[{}]:{}", addr, args.port),
|
||||
@@ -272,7 +280,8 @@ fn print_listening(args: &Args, print_addrs: &[BindAddr]) -> Result<String> {
|
||||
};
|
||||
format!("{}://{}{}", protocol, addr, args.uri_prefix)
|
||||
}
|
||||
BindAddr::Path(path) => path.display().to_string(),
|
||||
#[cfg(unix)]
|
||||
BindAddr::SocketPath(path) => path.to_string(),
|
||||
})
|
||||
.collect::<Vec<_>>();
|
||||
|
||||
|
||||
100
src/noscript.rs
Normal file
100
src/noscript.rs
Normal file
@@ -0,0 +1,100 @@
|
||||
use crate::{
|
||||
server::{IndexData, PathItem, PathType, MAX_SUBPATHS_COUNT},
|
||||
utils::encode_uri,
|
||||
};
|
||||
|
||||
use anyhow::Result;
|
||||
use chrono::{DateTime, Utc};
|
||||
use xml::escape::escape_str_pcdata;
|
||||
|
||||
pub fn detect_noscript(user_agent: &str) -> bool {
|
||||
[
|
||||
"lynx/", "w3m/", "links ", "elinks/", "curl/", "wget/", "httpie/", "aria2/",
|
||||
]
|
||||
.iter()
|
||||
.any(|v| user_agent.starts_with(v))
|
||||
}
|
||||
|
||||
pub fn generate_noscript_html(data: &IndexData) -> Result<String> {
|
||||
let mut html = String::new();
|
||||
|
||||
let title = format!("Index of {}", escape_str_pcdata(&data.href));
|
||||
|
||||
html.push_str("<html>\n");
|
||||
html.push_str("<head>\n");
|
||||
html.push_str(&format!("<title>{title}</title>\n"));
|
||||
html.push_str(
|
||||
r#"<style>
|
||||
td {
|
||||
padding: 0.2rem;
|
||||
text-align: left;
|
||||
}
|
||||
td:nth-child(3) {
|
||||
text-align: right;
|
||||
}
|
||||
</style>
|
||||
"#,
|
||||
);
|
||||
html.push_str("</head>\n");
|
||||
html.push_str("<body>\n");
|
||||
html.push_str(&format!("<h1>{title}</h1>\n"));
|
||||
html.push_str("<table>\n");
|
||||
html.push_str(" <tbody>\n");
|
||||
html.push_str(&format!(" {}\n", render_parent()));
|
||||
|
||||
for path in &data.paths {
|
||||
html.push_str(&format!(" {}\n", render_path_item(path)));
|
||||
}
|
||||
|
||||
html.push_str(" </tbody>\n");
|
||||
html.push_str("</table>\n");
|
||||
html.push_str("</body>\n");
|
||||
|
||||
Ok(html)
|
||||
}
|
||||
|
||||
fn render_parent() -> String {
|
||||
let value = "../";
|
||||
format!("<tr><td><a href=\"{value}\">{value}</a></td><td></td><td></td></tr>")
|
||||
}
|
||||
|
||||
fn render_path_item(path: &PathItem) -> String {
|
||||
let href = encode_uri(&path.name);
|
||||
let suffix = if path.path_type.is_dir() { "/" } else { "" };
|
||||
let name = escape_str_pcdata(&path.name);
|
||||
let mtime = format_mtime(path.mtime).unwrap_or_default();
|
||||
let size = format_size(path.size, path.path_type);
|
||||
|
||||
format!("<tr><td><a href=\"{href}{suffix}\">{name}{suffix}</a></td><td>{mtime}</td><td>{size}</td></tr>")
|
||||
}
|
||||
|
||||
fn format_mtime(mtime: u64) -> Option<String> {
|
||||
let datetime = DateTime::<Utc>::from_timestamp_millis(mtime as _)?;
|
||||
Some(datetime.format("%Y-%m-%dT%H:%M:%S.%3fZ").to_string())
|
||||
}
|
||||
|
||||
fn format_size(size: u64, path_type: PathType) -> String {
|
||||
if path_type.is_dir() {
|
||||
let unit = if size == 1 { "item" } else { "items" };
|
||||
let num = match size >= MAX_SUBPATHS_COUNT {
|
||||
true => format!(">{}", MAX_SUBPATHS_COUNT - 1),
|
||||
false => size.to_string(),
|
||||
};
|
||||
format!("{num} {unit}")
|
||||
} else {
|
||||
if size == 0 {
|
||||
return "0 B".to_string();
|
||||
}
|
||||
const UNITS: [&str; 5] = ["B", "KB", "MB", "GB", "TB"];
|
||||
let i = (size as f64).log2() / 10.0;
|
||||
let i = i.floor() as usize;
|
||||
|
||||
if i >= UNITS.len() {
|
||||
// Handle extremely large numbers beyond Terabytes
|
||||
return format!("{:.2} PB", size as f64 / 1024.0f64.powi(5));
|
||||
}
|
||||
|
||||
let size = size as f64 / 1024.0f64.powi(i as i32);
|
||||
format!("{:.2} {}", size, UNITS[i])
|
||||
}
|
||||
}
|
||||
597
src/server.rs
597
src/server.rs
@@ -2,6 +2,7 @@
|
||||
|
||||
use crate::auth::{www_authenticate, AccessPaths, AccessPerm};
|
||||
use crate::http_utils::{body_full, IncomingStream, LengthLimitedStream};
|
||||
use crate::noscript::{detect_noscript, generate_noscript_html};
|
||||
use crate::utils::{
|
||||
decode_uri, encode_uri, get_file_mtime_and_mode, get_file_name, glob, parse_range,
|
||||
try_get_file_name,
|
||||
@@ -10,13 +11,14 @@ use crate::Args;
|
||||
|
||||
use anyhow::{anyhow, Result};
|
||||
use async_zip::{tokio::write::ZipFileWriter, Compression, ZipDateTime, ZipEntryBuilder};
|
||||
use base64::{engine::general_purpose::STANDARD, Engine as _};
|
||||
use bytes::Bytes;
|
||||
use chrono::{LocalResult, TimeZone, Utc};
|
||||
use futures_util::{pin_mut, TryStreamExt};
|
||||
use headers::{
|
||||
AcceptRanges, AccessControlAllowCredentials, AccessControlAllowOrigin, CacheControl,
|
||||
ContentLength, ContentType, ETag, HeaderMap, HeaderMapExt, IfModifiedSince, IfNoneMatch,
|
||||
IfRange, LastModified, Range,
|
||||
ContentLength, ContentType, ETag, HeaderMap, HeaderMapExt, IfMatch, IfModifiedSince,
|
||||
IfNoneMatch, IfRange, IfUnmodifiedSince, LastModified, Range,
|
||||
};
|
||||
use http_body_util::{combinators::BoxBody, BodyExt, StreamBody};
|
||||
use hyper::body::Frame;
|
||||
@@ -36,7 +38,7 @@ use std::collections::HashMap;
|
||||
use std::fs::Metadata;
|
||||
use std::io::SeekFrom;
|
||||
use std::net::SocketAddr;
|
||||
use std::path::{Component, Path, PathBuf};
|
||||
use std::path::{Component, Path, PathBuf, MAIN_SEPARATOR};
|
||||
use std::sync::atomic::{self, AtomicBool};
|
||||
use std::sync::Arc;
|
||||
use std::time::SystemTime;
|
||||
@@ -47,7 +49,7 @@ use tokio::{fs, io};
|
||||
use tokio_util::compat::FuturesAsyncWriteCompatExt;
|
||||
use tokio_util::io::{ReaderStream, StreamReader};
|
||||
use uuid::Uuid;
|
||||
use walkdir::WalkDir;
|
||||
use walkdir::{DirEntry, WalkDir};
|
||||
use xml::escape::escape_str_pcdata;
|
||||
|
||||
pub type Request = hyper::Request<Incoming>;
|
||||
@@ -61,6 +63,8 @@ const INDEX_NAME: &str = "index.html";
|
||||
const BUF_SIZE: usize = 65536;
|
||||
const EDITABLE_TEXT_MAX_SIZE: u64 = 4194304; // 4M
|
||||
const RESUMABLE_UPLOAD_MIN_SIZE: u64 = 20971520; // 20M
|
||||
const HEALTH_CHECK_PATH: &str = "__dufs__/health";
|
||||
pub const MAX_SUBPATHS_COUNT: u64 = 1000;
|
||||
|
||||
pub struct Server {
|
||||
args: Args,
|
||||
@@ -107,18 +111,12 @@ impl Server {
|
||||
let uri = req.uri().clone();
|
||||
let assets_prefix = &self.assets_prefix;
|
||||
let enable_cors = self.args.enable_cors;
|
||||
let is_microsoft_webdav = req
|
||||
.headers()
|
||||
.get("user-agent")
|
||||
.and_then(|v| v.to_str().ok())
|
||||
.map(|v| v.starts_with("Microsoft-WebDAV-MiniRedir/"))
|
||||
.unwrap_or_default();
|
||||
let mut http_log_data = self.args.http_logger.data(&req);
|
||||
if let Some(addr) = addr {
|
||||
http_log_data.insert("remote_addr".to_string(), addr.ip().to_string());
|
||||
}
|
||||
|
||||
let mut res = match self.clone().handle(req, is_microsoft_webdav).await {
|
||||
let mut res = match self.clone().handle(req).await {
|
||||
Ok(res) => {
|
||||
http_log_data.insert("status".to_string(), res.status().as_u16().to_string());
|
||||
if !uri.path().starts_with(assets_prefix) {
|
||||
@@ -138,22 +136,13 @@ impl Server {
|
||||
}
|
||||
};
|
||||
|
||||
if is_microsoft_webdav {
|
||||
// microsoft webdav requires this.
|
||||
res.headers_mut()
|
||||
.insert(CONNECTION, HeaderValue::from_static("close"));
|
||||
}
|
||||
if enable_cors {
|
||||
add_cors(&mut res);
|
||||
}
|
||||
Ok(res)
|
||||
}
|
||||
|
||||
pub async fn handle(
|
||||
self: Arc<Self>,
|
||||
req: Request,
|
||||
is_microsoft_webdav: bool,
|
||||
) -> Result<Response> {
|
||||
pub async fn handle(self: Arc<Self>, req: Request) -> Result<Response> {
|
||||
let mut res = Response::default();
|
||||
|
||||
let req_path = req.uri().path();
|
||||
@@ -170,17 +159,40 @@ impl Server {
|
||||
|
||||
if method == Method::GET
|
||||
&& self
|
||||
.handle_assets(&relative_path, headers, &mut res)
|
||||
.handle_internal(&relative_path, headers, &mut res)
|
||||
.await?
|
||||
{
|
||||
return Ok(res);
|
||||
}
|
||||
|
||||
let user_agent = headers
|
||||
.get("user-agent")
|
||||
.and_then(|v| v.to_str().ok())
|
||||
.map(|v| v.to_lowercase())
|
||||
.unwrap_or_default();
|
||||
|
||||
let is_microsoft_webdav = user_agent.starts_with("microsoft-webdav-miniredir/");
|
||||
|
||||
if is_microsoft_webdav {
|
||||
// microsoft webdav requires this.
|
||||
res.headers_mut()
|
||||
.insert(CONNECTION, HeaderValue::from_static("close"));
|
||||
}
|
||||
|
||||
let authorization = headers.get(AUTHORIZATION);
|
||||
let guard =
|
||||
self.args
|
||||
.auth
|
||||
.guard(&relative_path, &method, authorization, is_microsoft_webdav);
|
||||
|
||||
let query = req.uri().query().unwrap_or_default();
|
||||
let mut query_params: HashMap<String, String> = form_urlencoded::parse(query.as_bytes())
|
||||
.map(|(k, v)| (k.to_string(), v.to_string()))
|
||||
.collect();
|
||||
|
||||
let guard = self.args.auth.guard(
|
||||
&relative_path,
|
||||
&method,
|
||||
authorization,
|
||||
query_params.get("token"),
|
||||
is_microsoft_webdav,
|
||||
);
|
||||
|
||||
let (user, access_paths) = match guard {
|
||||
(None, None) => {
|
||||
@@ -194,12 +206,20 @@ impl Server {
|
||||
(x, Some(y)) => (x, y),
|
||||
};
|
||||
|
||||
let query = req.uri().query().unwrap_or_default();
|
||||
let query_params: HashMap<String, String> = form_urlencoded::parse(query.as_bytes())
|
||||
.map(|(k, v)| (k.to_string(), v.to_string()))
|
||||
.collect();
|
||||
if detect_noscript(&user_agent) {
|
||||
query_params.insert("noscript".to_string(), String::new());
|
||||
}
|
||||
|
||||
if method.as_str() == "WRITEABLE" {
|
||||
if method.as_str() == "CHECKAUTH" {
|
||||
*res.body_mut() = body_full(user.clone().unwrap_or_default());
|
||||
return Ok(res);
|
||||
} else if method.as_str() == "LOGOUT" {
|
||||
self.auth_reject(&mut res)?;
|
||||
return Ok(res);
|
||||
}
|
||||
|
||||
if has_query_flag(&query_params, "tokengen") {
|
||||
self.handle_tokengen(&relative_path, user, &mut res).await?;
|
||||
return Ok(res);
|
||||
}
|
||||
|
||||
@@ -250,7 +270,7 @@ impl Server {
|
||||
Method::GET | Method::HEAD => {
|
||||
if is_dir {
|
||||
if render_try_index {
|
||||
if allow_archive && query_params.contains_key("zip") {
|
||||
if allow_archive && has_query_flag(&query_params, "zip") {
|
||||
if !allow_archive {
|
||||
status_not_found(&mut res);
|
||||
return Ok(res);
|
||||
@@ -290,7 +310,7 @@ impl Server {
|
||||
&mut res,
|
||||
)
|
||||
.await?;
|
||||
} else if query_params.contains_key("zip") {
|
||||
} else if has_query_flag(&query_params, "zip") {
|
||||
if !allow_archive {
|
||||
status_not_found(&mut res);
|
||||
return Ok(res);
|
||||
@@ -320,13 +340,13 @@ impl Server {
|
||||
.await?;
|
||||
}
|
||||
} else if is_file {
|
||||
if query_params.contains_key("edit") {
|
||||
if has_query_flag(&query_params, "edit") {
|
||||
self.handle_edit_file(path, DataKind::Edit, head_only, user, &mut res)
|
||||
.await?;
|
||||
} else if query_params.contains_key("view") {
|
||||
} else if has_query_flag(&query_params, "view") {
|
||||
self.handle_edit_file(path, DataKind::View, head_only, user, &mut res)
|
||||
.await?;
|
||||
} else if query_params.contains_key("hash") {
|
||||
} else if has_query_flag(&query_params, "hash") {
|
||||
self.handle_hash_file(path, head_only, &mut res).await?;
|
||||
} else {
|
||||
self.handle_send_file(path, headers, head_only, &mut res)
|
||||
@@ -495,7 +515,7 @@ impl Server {
|
||||
};
|
||||
let stream = IncomingStream::new(req.into_body());
|
||||
|
||||
let body_with_io_error = stream.map_err(|err| io::Error::new(io::ErrorKind::Other, err));
|
||||
let body_with_io_error = stream.map_err(io::Error::other);
|
||||
let body_reader = StreamReader::new(body_with_io_error);
|
||||
|
||||
pin_mut!(body_reader);
|
||||
@@ -580,47 +600,20 @@ impl Server {
|
||||
} else {
|
||||
let path_buf = path.to_path_buf();
|
||||
let hidden = Arc::new(self.args.hidden.to_vec());
|
||||
let hidden = hidden.clone();
|
||||
let running = self.running.clone();
|
||||
let search = search.clone();
|
||||
|
||||
let access_paths = access_paths.clone();
|
||||
let search_paths = tokio::task::spawn_blocking(move || {
|
||||
let mut paths: Vec<PathBuf> = vec![];
|
||||
for dir in access_paths.child_paths(&path_buf) {
|
||||
let mut it = WalkDir::new(&dir).into_iter();
|
||||
it.next();
|
||||
while let Some(Ok(entry)) = it.next() {
|
||||
if !running.load(atomic::Ordering::SeqCst) {
|
||||
break;
|
||||
}
|
||||
let entry_path = entry.path();
|
||||
let base_name = get_file_name(entry_path);
|
||||
let file_type = entry.file_type();
|
||||
let mut is_dir_type: bool = file_type.is_dir();
|
||||
if file_type.is_symlink() {
|
||||
match std::fs::symlink_metadata(entry_path) {
|
||||
Ok(meta) => {
|
||||
is_dir_type = meta.is_dir();
|
||||
}
|
||||
Err(_) => {
|
||||
continue;
|
||||
}
|
||||
}
|
||||
}
|
||||
if is_hidden(&hidden, base_name, is_dir_type) {
|
||||
if file_type.is_dir() {
|
||||
it.skip_current_dir();
|
||||
}
|
||||
continue;
|
||||
}
|
||||
if !base_name.to_lowercase().contains(&search) {
|
||||
continue;
|
||||
}
|
||||
paths.push(entry_path.to_path_buf());
|
||||
}
|
||||
}
|
||||
paths
|
||||
})
|
||||
let search_paths = tokio::spawn(collect_dir_entries(
|
||||
access_paths,
|
||||
self.running.clone(),
|
||||
path_buf,
|
||||
hidden,
|
||||
self.args.allow_symlink,
|
||||
self.args.serve_path.clone(),
|
||||
move |x| get_file_name(x.path()).to_lowercase().contains(&search),
|
||||
))
|
||||
.await?;
|
||||
|
||||
for search_path in search_paths.into_iter() {
|
||||
if let Ok(Some(item)) = self.to_pathitem(search_path, path.to_path_buf()).await {
|
||||
paths.push(item);
|
||||
@@ -648,7 +641,7 @@ impl Server {
|
||||
) -> Result<()> {
|
||||
let (mut writer, reader) = tokio::io::duplex(BUF_SIZE);
|
||||
let filename = try_get_file_name(path)?;
|
||||
set_content_disposition(res, false, &format!("{}.zip", filename))?;
|
||||
set_content_disposition(res, false, &format!("{filename}.zip"))?;
|
||||
res.headers_mut()
|
||||
.insert("content-type", HeaderValue::from_static("application/zip"));
|
||||
if head_only {
|
||||
@@ -658,6 +651,8 @@ impl Server {
|
||||
let hidden = self.args.hidden.clone();
|
||||
let running = self.running.clone();
|
||||
let compression = self.args.compress.to_compression();
|
||||
let follow_symlinks = self.args.allow_symlink;
|
||||
let serve_path = self.args.serve_path.clone();
|
||||
tokio::spawn(async move {
|
||||
if let Err(e) = zip_dir(
|
||||
&mut writer,
|
||||
@@ -665,14 +660,16 @@ impl Server {
|
||||
access_paths,
|
||||
&hidden,
|
||||
compression,
|
||||
follow_symlinks,
|
||||
serve_path,
|
||||
running,
|
||||
)
|
||||
.await
|
||||
{
|
||||
error!("Failed to zip {}, {}", path.display(), e);
|
||||
error!("Failed to zip {}, {e}", path.display());
|
||||
}
|
||||
});
|
||||
let reader_stream = ReaderStream::new(reader);
|
||||
let reader_stream = ReaderStream::with_capacity(reader, BUF_SIZE);
|
||||
let stream_body = StreamBody::new(
|
||||
reader_stream
|
||||
.map_ok(Frame::data)
|
||||
@@ -728,7 +725,7 @@ impl Server {
|
||||
Ok(())
|
||||
}
|
||||
|
||||
async fn handle_assets(
|
||||
async fn handle_internal(
|
||||
&self,
|
||||
req_path: &str,
|
||||
headers: &HeaderMap<HeaderValue>,
|
||||
@@ -779,6 +776,12 @@ impl Server {
|
||||
HeaderValue::from_static("nosniff"),
|
||||
);
|
||||
Ok(true)
|
||||
} else if req_path == HEALTH_CHECK_PATH {
|
||||
res.headers_mut()
|
||||
.typed_insert(ContentType::from(mime_guess::mime::APPLICATION_JSON));
|
||||
|
||||
*res.body_mut() = body_full(r#"{"status":"OK"}"#);
|
||||
Ok(true)
|
||||
} else {
|
||||
Ok(false)
|
||||
}
|
||||
@@ -796,20 +799,33 @@ impl Server {
|
||||
let size = meta.len();
|
||||
let mut use_range = true;
|
||||
if let Some((etag, last_modified)) = extract_cache_headers(&meta) {
|
||||
let cached = {
|
||||
if let Some(if_none_match) = headers.typed_get::<IfNoneMatch>() {
|
||||
!if_none_match.precondition_passes(&etag)
|
||||
} else if let Some(if_modified_since) = headers.typed_get::<IfModifiedSince>() {
|
||||
!if_modified_since.is_modified(last_modified.into())
|
||||
} else {
|
||||
false
|
||||
if let Some(if_unmodified_since) = headers.typed_get::<IfUnmodifiedSince>() {
|
||||
if !if_unmodified_since.precondition_passes(last_modified.into()) {
|
||||
*res.status_mut() = StatusCode::PRECONDITION_FAILED;
|
||||
return Ok(());
|
||||
}
|
||||
}
|
||||
if let Some(if_match) = headers.typed_get::<IfMatch>() {
|
||||
if !if_match.precondition_passes(&etag) {
|
||||
*res.status_mut() = StatusCode::PRECONDITION_FAILED;
|
||||
return Ok(());
|
||||
}
|
||||
}
|
||||
if let Some(if_modified_since) = headers.typed_get::<IfModifiedSince>() {
|
||||
if !if_modified_since.is_modified(last_modified.into()) {
|
||||
*res.status_mut() = StatusCode::NOT_MODIFIED;
|
||||
return Ok(());
|
||||
}
|
||||
}
|
||||
if let Some(if_none_match) = headers.typed_get::<IfNoneMatch>() {
|
||||
if !if_none_match.precondition_passes(&etag) {
|
||||
*res.status_mut() = StatusCode::NOT_MODIFIED;
|
||||
return Ok(());
|
||||
}
|
||||
};
|
||||
if cached {
|
||||
*res.status_mut() = StatusCode::NOT_MODIFIED;
|
||||
return Ok(());
|
||||
}
|
||||
|
||||
res.headers_mut()
|
||||
.typed_insert(CacheControl::new().with_no_cache());
|
||||
res.headers_mut().typed_insert(last_modified);
|
||||
res.headers_mut().typed_insert(etag.clone());
|
||||
|
||||
@@ -824,7 +840,7 @@ impl Server {
|
||||
}
|
||||
}
|
||||
|
||||
let range = if use_range {
|
||||
let ranges = if use_range {
|
||||
headers.get(RANGE).map(|range| {
|
||||
range
|
||||
.to_str()
|
||||
@@ -845,27 +861,59 @@ impl Server {
|
||||
|
||||
res.headers_mut().typed_insert(AcceptRanges::bytes());
|
||||
|
||||
if let Some(range) = range {
|
||||
if let Some((start, end)) = range {
|
||||
file.seek(SeekFrom::Start(start)).await?;
|
||||
let range_size = end - start + 1;
|
||||
*res.status_mut() = StatusCode::PARTIAL_CONTENT;
|
||||
let content_range = format!("bytes {}-{}/{}", start, end, size);
|
||||
res.headers_mut()
|
||||
.insert(CONTENT_RANGE, content_range.parse()?);
|
||||
res.headers_mut()
|
||||
.insert(CONTENT_LENGTH, format!("{range_size}").parse()?);
|
||||
if head_only {
|
||||
return Ok(());
|
||||
}
|
||||
if let Some(ranges) = ranges {
|
||||
if let Some(ranges) = ranges {
|
||||
if ranges.len() == 1 {
|
||||
let (start, end) = ranges[0];
|
||||
file.seek(SeekFrom::Start(start)).await?;
|
||||
let range_size = end - start + 1;
|
||||
*res.status_mut() = StatusCode::PARTIAL_CONTENT;
|
||||
let content_range = format!("bytes {start}-{end}/{size}");
|
||||
res.headers_mut()
|
||||
.insert(CONTENT_RANGE, content_range.parse()?);
|
||||
res.headers_mut()
|
||||
.insert(CONTENT_LENGTH, format!("{range_size}").parse()?);
|
||||
if head_only {
|
||||
return Ok(());
|
||||
}
|
||||
|
||||
let stream_body = StreamBody::new(
|
||||
LengthLimitedStream::new(file, range_size as usize)
|
||||
.map_ok(Frame::data)
|
||||
.map_err(|err| anyhow!("{err}")),
|
||||
);
|
||||
let boxed_body = stream_body.boxed();
|
||||
*res.body_mut() = boxed_body;
|
||||
let stream_body = StreamBody::new(
|
||||
LengthLimitedStream::new(file, range_size as usize)
|
||||
.map_ok(Frame::data)
|
||||
.map_err(|err| anyhow!("{err}")),
|
||||
);
|
||||
let boxed_body = stream_body.boxed();
|
||||
*res.body_mut() = boxed_body;
|
||||
} else {
|
||||
*res.status_mut() = StatusCode::PARTIAL_CONTENT;
|
||||
let boundary = Uuid::new_v4();
|
||||
let mut body = Vec::new();
|
||||
let content_type = get_content_type(path).await?;
|
||||
for (start, end) in ranges {
|
||||
file.seek(SeekFrom::Start(start)).await?;
|
||||
let range_size = end - start + 1;
|
||||
let content_range = format!("bytes {start}-{end}/{size}");
|
||||
let part_header = format!(
|
||||
"--{boundary}\r\nContent-Type: {content_type}\r\nContent-Range: {content_range}\r\n\r\n",
|
||||
);
|
||||
body.extend_from_slice(part_header.as_bytes());
|
||||
let mut buffer = vec![0; range_size as usize];
|
||||
file.read_exact(&mut buffer).await?;
|
||||
body.extend_from_slice(&buffer);
|
||||
body.extend_from_slice(b"\r\n");
|
||||
}
|
||||
body.extend_from_slice(format!("--{boundary}--\r\n").as_bytes());
|
||||
res.headers_mut().insert(
|
||||
CONTENT_TYPE,
|
||||
format!("multipart/byteranges; boundary={boundary}").parse()?,
|
||||
);
|
||||
res.headers_mut()
|
||||
.insert(CONTENT_LENGTH, format!("{}", body.len()).parse()?);
|
||||
if head_only {
|
||||
return Ok(());
|
||||
}
|
||||
*res.body_mut() = body_full(body);
|
||||
}
|
||||
} else {
|
||||
*res.status_mut() = StatusCode::RANGE_NOT_SATISFIABLE;
|
||||
res.headers_mut()
|
||||
@@ -878,7 +926,7 @@ impl Server {
|
||||
return Ok(());
|
||||
}
|
||||
|
||||
let reader_stream = ReaderStream::new(file);
|
||||
let reader_stream = ReaderStream::with_capacity(file, BUF_SIZE);
|
||||
let stream_body = StreamBody::new(
|
||||
reader_stream
|
||||
.map_ok(Frame::data)
|
||||
@@ -920,15 +968,18 @@ impl Server {
|
||||
};
|
||||
res.headers_mut()
|
||||
.typed_insert(ContentType::from(mime_guess::mime::TEXT_HTML_UTF_8));
|
||||
let index_data = STANDARD.encode(serde_json::to_string(&data)?);
|
||||
let output = self
|
||||
.html
|
||||
.replace(
|
||||
"__ASSETS_PREFIX__",
|
||||
&format!("{}{}", self.args.uri_prefix, self.assets_prefix),
|
||||
)
|
||||
.replace("__INDEX_DATA__", &serde_json::to_string(&data)?);
|
||||
.replace("__INDEX_DATA__", &index_data);
|
||||
res.headers_mut()
|
||||
.typed_insert(ContentLength(output.as_bytes().len() as u64));
|
||||
.typed_insert(ContentLength(output.len() as u64));
|
||||
res.headers_mut()
|
||||
.typed_insert(CacheControl::new().with_no_cache());
|
||||
if head_only {
|
||||
return Ok(());
|
||||
}
|
||||
@@ -946,7 +997,7 @@ impl Server {
|
||||
res.headers_mut()
|
||||
.typed_insert(ContentType::from(mime_guess::mime::TEXT_HTML_UTF_8));
|
||||
res.headers_mut()
|
||||
.typed_insert(ContentLength(output.as_bytes().len() as u64));
|
||||
.typed_insert(ContentLength(output.len() as u64));
|
||||
if head_only {
|
||||
return Ok(());
|
||||
}
|
||||
@@ -954,6 +1005,24 @@ impl Server {
|
||||
Ok(())
|
||||
}
|
||||
|
||||
async fn handle_tokengen(
|
||||
&self,
|
||||
relative_path: &str,
|
||||
user: Option<String>,
|
||||
res: &mut Response,
|
||||
) -> Result<()> {
|
||||
let output = self
|
||||
.args
|
||||
.auth
|
||||
.generate_token(relative_path, &user.unwrap_or_default())?;
|
||||
res.headers_mut()
|
||||
.typed_insert(ContentType::from(mime_guess::mime::TEXT_PLAIN_UTF_8));
|
||||
res.headers_mut()
|
||||
.typed_insert(ContentLength(output.len() as u64));
|
||||
*res.body_mut() = body_full(output);
|
||||
Ok(())
|
||||
}
|
||||
|
||||
async fn handle_propfind_dir(
|
||||
&self,
|
||||
path: &Path,
|
||||
@@ -963,9 +1032,10 @@ impl Server {
|
||||
) -> Result<()> {
|
||||
let depth: u32 = match headers.get("depth") {
|
||||
Some(v) => match v.to_str().ok().and_then(|v| v.parse().ok()) {
|
||||
Some(v) => v,
|
||||
None => {
|
||||
status_bad_request(res, "");
|
||||
Some(0) => 0,
|
||||
Some(1) => 1,
|
||||
_ => {
|
||||
status_bad_request(res, "Invalid depth: only 0 and 1 are allowed.");
|
||||
return Ok(());
|
||||
}
|
||||
},
|
||||
@@ -975,7 +1045,7 @@ impl Server {
|
||||
Some(v) => vec![v],
|
||||
None => vec![],
|
||||
};
|
||||
if depth != 0 {
|
||||
if depth == 1 {
|
||||
match self
|
||||
.list_dir(path, &self.args.serve_path, access_paths)
|
||||
.await
|
||||
@@ -1120,7 +1190,7 @@ impl Server {
|
||||
} else {
|
||||
paths.sort_by(|v1, v2| v1.sort_by_name(v2))
|
||||
}
|
||||
if query_params.contains_key("simple") {
|
||||
if has_query_flag(query_params, "simple") {
|
||||
let output = paths
|
||||
.into_iter()
|
||||
.map(|v| {
|
||||
@@ -1135,7 +1205,7 @@ impl Server {
|
||||
res.headers_mut()
|
||||
.typed_insert(ContentType::from(mime_guess::mime::TEXT_HTML_UTF_8));
|
||||
res.headers_mut()
|
||||
.typed_insert(ContentLength(output.as_bytes().len() as u64));
|
||||
.typed_insert(ContentLength(output.len() as u64));
|
||||
*res.body_mut() = body_full(output);
|
||||
if head_only {
|
||||
return Ok(());
|
||||
@@ -1160,22 +1230,28 @@ impl Server {
|
||||
user,
|
||||
paths,
|
||||
};
|
||||
let output = if query_params.contains_key("json") {
|
||||
let output = if has_query_flag(query_params, "json") {
|
||||
res.headers_mut()
|
||||
.typed_insert(ContentType::from(mime_guess::mime::APPLICATION_JSON));
|
||||
serde_json::to_string_pretty(&data)?
|
||||
} else if has_query_flag(query_params, "noscript") {
|
||||
res.headers_mut()
|
||||
.typed_insert(ContentType::from(mime_guess::mime::TEXT_HTML_UTF_8));
|
||||
generate_noscript_html(&data)?
|
||||
} else {
|
||||
res.headers_mut()
|
||||
.typed_insert(ContentType::from(mime_guess::mime::TEXT_HTML_UTF_8));
|
||||
|
||||
let index_data = STANDARD.encode(serde_json::to_string(&data)?);
|
||||
self.html
|
||||
.replace(
|
||||
"__ASSETS_PREFIX__",
|
||||
&format!("{}{}", self.args.uri_prefix, self.assets_prefix),
|
||||
)
|
||||
.replace("__INDEX_DATA__", &serde_json::to_string(&data)?)
|
||||
.replace("__INDEX_DATA__", &index_data)
|
||||
};
|
||||
res.headers_mut()
|
||||
.typed_insert(ContentLength(output.as_bytes().len() as u64));
|
||||
.typed_insert(ContentLength(output.len() as u64));
|
||||
res.headers_mut()
|
||||
.typed_insert(CacheControl::new().with_no_cache());
|
||||
res.headers_mut().insert(
|
||||
@@ -1222,7 +1298,7 @@ impl Server {
|
||||
let guard = self
|
||||
.args
|
||||
.auth
|
||||
.guard(&dest_path, req.method(), authorization, false);
|
||||
.guard(&dest_path, req.method(), authorization, None, false);
|
||||
|
||||
match guard {
|
||||
(_, Some(_)) => {}
|
||||
@@ -1336,10 +1412,33 @@ impl Server {
|
||||
(true, false) => PathType::SymlinkFile,
|
||||
(false, false) => PathType::File,
|
||||
};
|
||||
let mtime = to_timestamp(&meta.modified()?);
|
||||
let mtime = match meta.modified().ok().or_else(|| meta.created().ok()) {
|
||||
Some(v) => to_timestamp(&v),
|
||||
None => 0,
|
||||
};
|
||||
let size = match path_type {
|
||||
PathType::Dir | PathType::SymlinkDir => None,
|
||||
PathType::File | PathType::SymlinkFile => Some(meta.len()),
|
||||
PathType::Dir | PathType::SymlinkDir => {
|
||||
let mut count = 0;
|
||||
let mut entries = tokio::fs::read_dir(&path).await?;
|
||||
while let Some(entry) = entries.next_entry().await? {
|
||||
let entry_path = entry.path();
|
||||
let base_name = get_file_name(&entry_path);
|
||||
let is_dir = entry
|
||||
.file_type()
|
||||
.await
|
||||
.map(|v| v.is_dir())
|
||||
.unwrap_or_default();
|
||||
if is_hidden(&self.args.hidden, base_name, is_dir) {
|
||||
continue;
|
||||
}
|
||||
count += 1;
|
||||
if count >= MAX_SUBPATHS_COUNT {
|
||||
break;
|
||||
}
|
||||
}
|
||||
count
|
||||
}
|
||||
PathType::File | PathType::SymlinkFile => meta.len(),
|
||||
};
|
||||
let rel_path = path.strip_prefix(base_path)?;
|
||||
let name = normalize_path(rel_path);
|
||||
@@ -1353,45 +1452,33 @@ impl Server {
|
||||
}
|
||||
|
||||
#[derive(Debug, Serialize, PartialEq)]
|
||||
enum DataKind {
|
||||
pub enum DataKind {
|
||||
Index,
|
||||
Edit,
|
||||
View,
|
||||
}
|
||||
|
||||
#[derive(Debug, Serialize)]
|
||||
struct IndexData {
|
||||
href: String,
|
||||
kind: DataKind,
|
||||
uri_prefix: String,
|
||||
allow_upload: bool,
|
||||
allow_delete: bool,
|
||||
allow_search: bool,
|
||||
allow_archive: bool,
|
||||
dir_exists: bool,
|
||||
auth: bool,
|
||||
user: Option<String>,
|
||||
paths: Vec<PathItem>,
|
||||
}
|
||||
|
||||
#[derive(Debug, Serialize)]
|
||||
struct EditData {
|
||||
href: String,
|
||||
kind: DataKind,
|
||||
uri_prefix: String,
|
||||
allow_upload: bool,
|
||||
allow_delete: bool,
|
||||
auth: bool,
|
||||
user: Option<String>,
|
||||
editable: bool,
|
||||
pub struct IndexData {
|
||||
pub href: String,
|
||||
pub kind: DataKind,
|
||||
pub uri_prefix: String,
|
||||
pub allow_upload: bool,
|
||||
pub allow_delete: bool,
|
||||
pub allow_search: bool,
|
||||
pub allow_archive: bool,
|
||||
pub dir_exists: bool,
|
||||
pub auth: bool,
|
||||
pub user: Option<String>,
|
||||
pub paths: Vec<PathItem>,
|
||||
}
|
||||
|
||||
#[derive(Debug, Serialize, Eq, PartialEq, Ord, PartialOrd)]
|
||||
struct PathItem {
|
||||
path_type: PathType,
|
||||
name: String,
|
||||
mtime: u64,
|
||||
size: Option<u64>,
|
||||
pub struct PathItem {
|
||||
pub path_type: PathType,
|
||||
pub name: String,
|
||||
pub mtime: u64,
|
||||
pub size: u64,
|
||||
}
|
||||
|
||||
impl PathItem {
|
||||
@@ -1425,27 +1512,24 @@ impl PathItem {
|
||||
),
|
||||
PathType::File | PathType::SymlinkFile => format!(
|
||||
r#"<D:response>
|
||||
<D:href>{}</D:href>
|
||||
<D:href>{href}</D:href>
|
||||
<D:propstat>
|
||||
<D:prop>
|
||||
<D:displayname>{}</D:displayname>
|
||||
<D:displayname>{displayname}</D:displayname>
|
||||
<D:getcontentlength>{}</D:getcontentlength>
|
||||
<D:getlastmodified>{}</D:getlastmodified>
|
||||
<D:getlastmodified>{mtime}</D:getlastmodified>
|
||||
<D:resourcetype></D:resourcetype>
|
||||
</D:prop>
|
||||
<D:status>HTTP/1.1 200 OK</D:status>
|
||||
</D:propstat>
|
||||
</D:response>"#,
|
||||
href,
|
||||
displayname,
|
||||
self.size.unwrap_or_default(),
|
||||
mtime
|
||||
self.size
|
||||
),
|
||||
}
|
||||
}
|
||||
|
||||
pub fn base_name(&self) -> &str {
|
||||
self.name.split('/').last().unwrap_or_default()
|
||||
self.name.split('/').next_back().unwrap_or_default()
|
||||
}
|
||||
|
||||
pub fn sort_by_name(&self, other: &Self) -> Ordering {
|
||||
@@ -1466,29 +1550,26 @@ impl PathItem {
|
||||
|
||||
pub fn sort_by_size(&self, other: &Self) -> Ordering {
|
||||
match self.path_type.cmp(&other.path_type) {
|
||||
Ordering::Equal => {
|
||||
if self.is_dir() {
|
||||
alphanumeric_sort::compare_str(
|
||||
self.name.to_lowercase(),
|
||||
other.name.to_lowercase(),
|
||||
)
|
||||
} else {
|
||||
self.size.unwrap_or(0).cmp(&other.size.unwrap_or(0))
|
||||
}
|
||||
}
|
||||
Ordering::Equal => self.size.cmp(&other.size),
|
||||
v => v,
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
#[derive(Debug, Serialize, Eq, PartialEq)]
|
||||
enum PathType {
|
||||
#[derive(Debug, Serialize, Clone, Copy, Eq, PartialEq)]
|
||||
pub enum PathType {
|
||||
Dir,
|
||||
SymlinkDir,
|
||||
File,
|
||||
SymlinkFile,
|
||||
}
|
||||
|
||||
impl PathType {
|
||||
pub fn is_dir(&self) -> bool {
|
||||
matches!(self, Self::Dir | Self::SymlinkDir)
|
||||
}
|
||||
}
|
||||
|
||||
impl Ord for PathType {
|
||||
fn cmp(&self, other: &Self) -> Ordering {
|
||||
let to_value = |t: &Self| -> u8 {
|
||||
@@ -1507,6 +1588,18 @@ impl PartialOrd for PathType {
|
||||
}
|
||||
}
|
||||
|
||||
#[derive(Debug, Serialize)]
|
||||
struct EditData {
|
||||
href: String,
|
||||
kind: DataKind,
|
||||
uri_prefix: String,
|
||||
allow_upload: bool,
|
||||
allow_delete: bool,
|
||||
auth: bool,
|
||||
user: Option<String>,
|
||||
editable: bool,
|
||||
}
|
||||
|
||||
fn to_timestamp(time: &SystemTime) -> u64 {
|
||||
time.duration_since(SystemTime::UNIX_EPOCH)
|
||||
.unwrap_or_default()
|
||||
@@ -1570,54 +1663,29 @@ async fn zip_dir<W: AsyncWrite + Unpin>(
|
||||
access_paths: AccessPaths,
|
||||
hidden: &[String],
|
||||
compression: Compression,
|
||||
follow_symlinks: bool,
|
||||
serve_path: PathBuf,
|
||||
running: Arc<AtomicBool>,
|
||||
) -> Result<()> {
|
||||
let mut writer = ZipFileWriter::with_tokio(writer);
|
||||
let hidden = Arc::new(hidden.to_vec());
|
||||
let dir_clone = dir.to_path_buf();
|
||||
let zip_paths = tokio::task::spawn_blocking(move || {
|
||||
let mut paths: Vec<PathBuf> = vec![];
|
||||
for dir in access_paths.child_paths(&dir_clone) {
|
||||
let mut it = WalkDir::new(&dir).into_iter();
|
||||
it.next();
|
||||
while let Some(Ok(entry)) = it.next() {
|
||||
if !running.load(atomic::Ordering::SeqCst) {
|
||||
break;
|
||||
}
|
||||
let entry_path = entry.path();
|
||||
let base_name = get_file_name(entry_path);
|
||||
let file_type = entry.file_type();
|
||||
let mut is_dir_type: bool = file_type.is_dir();
|
||||
if file_type.is_symlink() {
|
||||
match std::fs::symlink_metadata(entry_path) {
|
||||
Ok(meta) => {
|
||||
is_dir_type = meta.is_dir();
|
||||
}
|
||||
Err(_) => {
|
||||
continue;
|
||||
}
|
||||
}
|
||||
}
|
||||
if is_hidden(&hidden, base_name, is_dir_type) {
|
||||
if file_type.is_dir() {
|
||||
it.skip_current_dir();
|
||||
}
|
||||
continue;
|
||||
}
|
||||
if entry.path().symlink_metadata().is_err() {
|
||||
continue;
|
||||
}
|
||||
if !file_type.is_file() {
|
||||
continue;
|
||||
}
|
||||
paths.push(entry_path.to_path_buf());
|
||||
}
|
||||
}
|
||||
paths
|
||||
})
|
||||
let zip_paths = tokio::task::spawn(collect_dir_entries(
|
||||
access_paths,
|
||||
running,
|
||||
dir.to_path_buf(),
|
||||
hidden,
|
||||
follow_symlinks,
|
||||
serve_path,
|
||||
move |x| x.path().symlink_metadata().is_ok() && x.file_type().is_file(),
|
||||
))
|
||||
.await?;
|
||||
for zip_path in zip_paths.into_iter() {
|
||||
let filename = match zip_path.strip_prefix(dir).ok().and_then(|v| v.to_str()) {
|
||||
let filename = match zip_path
|
||||
.strip_prefix(dir)
|
||||
.ok()
|
||||
.and_then(|v| v.to_str())
|
||||
.map(|v| v.replace(MAIN_SEPARATOR, "/"))
|
||||
{
|
||||
Some(v) => v,
|
||||
None => continue,
|
||||
};
|
||||
@@ -1635,7 +1703,7 @@ async fn zip_dir<W: AsyncWrite + Unpin>(
|
||||
}
|
||||
|
||||
fn extract_cache_headers(meta: &Metadata) -> Option<(ETag, LastModified)> {
|
||||
let mtime = meta.modified().ok()?;
|
||||
let mtime = meta.modified().ok().or_else(|| meta.created().ok())?;
|
||||
let timestamp = to_timestamp(&mtime);
|
||||
let size = meta.len();
|
||||
let etag = format!(r#""{timestamp}-{size}""#).parse::<ETag>().ok()?;
|
||||
@@ -1677,7 +1745,7 @@ fn set_content_disposition(res: &mut Response, inline: bool, filename: &str) ->
|
||||
})
|
||||
.collect();
|
||||
let value = if filename.is_ascii() {
|
||||
HeaderValue::from_str(&format!("{kind}; filename=\"{}\"", filename,))?
|
||||
HeaderValue::from_str(&format!("{kind}; filename=\"{filename}\"",))?
|
||||
} else {
|
||||
HeaderValue::from_str(&format!(
|
||||
"{kind}; filename=\"{}\"; filename*=UTF-8''{}",
|
||||
@@ -1689,9 +1757,9 @@ fn set_content_disposition(res: &mut Response, inline: bool, filename: &str) ->
|
||||
Ok(())
|
||||
}
|
||||
|
||||
fn is_hidden(hidden: &[String], file_name: &str, is_dir_type: bool) -> bool {
|
||||
fn is_hidden(hidden: &[String], file_name: &str, is_dir: bool) -> bool {
|
||||
hidden.iter().any(|v| {
|
||||
if is_dir_type {
|
||||
if is_dir {
|
||||
if let Some(x) = v.strip_suffix('/') {
|
||||
return glob(x, file_name);
|
||||
}
|
||||
@@ -1703,12 +1771,12 @@ fn is_hidden(hidden: &[String], file_name: &str, is_dir_type: bool) -> bool {
|
||||
fn set_webdav_headers(res: &mut Response) {
|
||||
res.headers_mut().insert(
|
||||
"Allow",
|
||||
HeaderValue::from_static("GET,HEAD,PUT,OPTIONS,DELETE,PATCH,PROPFIND,COPY,MOVE"),
|
||||
);
|
||||
res.headers_mut().insert(
|
||||
"DAV",
|
||||
HeaderValue::from_static("1, 2, 3, sabredav-partialupdate"),
|
||||
HeaderValue::from_static(
|
||||
"GET,HEAD,PUT,OPTIONS,DELETE,PATCH,PROPFIND,COPY,MOVE,CHECKAUTH,LOGOUT",
|
||||
),
|
||||
);
|
||||
res.headers_mut()
|
||||
.insert("DAV", HeaderValue::from_static("1, 2, 3"));
|
||||
}
|
||||
|
||||
async fn get_content_type(path: &Path) -> Result<String> {
|
||||
@@ -1752,8 +1820,10 @@ fn parse_upload_offset(headers: &HeaderMap<HeaderValue>, size: u64) -> Result<Op
|
||||
if value == "append" {
|
||||
return Ok(Some(size));
|
||||
}
|
||||
let (start, _) = parse_range(value, size).ok_or_else(err)?;
|
||||
Ok(Some(start))
|
||||
// use the first range
|
||||
let ranges = parse_range(value, size).ok_or_else(err)?;
|
||||
let (start, _) = ranges.first().ok_or_else(err)?;
|
||||
Ok(Some(*start))
|
||||
}
|
||||
|
||||
async fn sha256_file(path: &Path) -> Result<String> {
|
||||
@@ -1770,5 +1840,66 @@ async fn sha256_file(path: &Path) -> Result<String> {
|
||||
}
|
||||
|
||||
let result = hasher.finalize();
|
||||
Ok(format!("{:x}", result))
|
||||
Ok(format!("{result:x}"))
|
||||
}
|
||||
|
||||
fn has_query_flag(query_params: &HashMap<String, String>, name: &str) -> bool {
|
||||
query_params
|
||||
.get(name)
|
||||
.map(|v| v.is_empty())
|
||||
.unwrap_or_default()
|
||||
}
|
||||
|
||||
async fn collect_dir_entries<F>(
|
||||
access_paths: AccessPaths,
|
||||
running: Arc<AtomicBool>,
|
||||
path: PathBuf,
|
||||
hidden: Arc<Vec<String>>,
|
||||
follow_symlinks: bool,
|
||||
serve_path: PathBuf,
|
||||
include_entry: F,
|
||||
) -> Vec<PathBuf>
|
||||
where
|
||||
F: Fn(&DirEntry) -> bool,
|
||||
{
|
||||
let mut paths: Vec<PathBuf> = vec![];
|
||||
for dir in access_paths.entry_paths(&path) {
|
||||
let mut it = WalkDir::new(&dir).follow_links(true).into_iter();
|
||||
it.next();
|
||||
while let Some(Ok(entry)) = it.next() {
|
||||
if !running.load(atomic::Ordering::SeqCst) {
|
||||
break;
|
||||
}
|
||||
let entry_path = entry.path();
|
||||
let base_name = get_file_name(entry_path);
|
||||
let is_dir = entry.file_type().is_dir();
|
||||
if is_hidden(&hidden, base_name, is_dir) {
|
||||
if is_dir {
|
||||
it.skip_current_dir();
|
||||
}
|
||||
continue;
|
||||
}
|
||||
|
||||
if !follow_symlinks
|
||||
&& !fs::canonicalize(entry_path)
|
||||
.await
|
||||
.ok()
|
||||
.map(|v| v.starts_with(&serve_path))
|
||||
.unwrap_or_default()
|
||||
{
|
||||
// We walked outside the server's root. This could only have
|
||||
// happened if we followed a symlink, and hence we only allow it
|
||||
// if allow_symlink is enabled, otherwise we skip this entry.
|
||||
if is_dir {
|
||||
it.skip_current_dir();
|
||||
}
|
||||
continue;
|
||||
}
|
||||
if !include_entry(&entry) {
|
||||
continue;
|
||||
}
|
||||
paths.push(entry_path.to_path_buf());
|
||||
}
|
||||
}
|
||||
paths
|
||||
}
|
||||
|
||||
74
src/utils.rs
74
src/utils.rs
@@ -8,10 +8,10 @@ use std::{
|
||||
time::{Duration, SystemTime, UNIX_EPOCH},
|
||||
};
|
||||
|
||||
pub fn unix_now() -> Result<Duration> {
|
||||
pub fn unix_now() -> Duration {
|
||||
SystemTime::now()
|
||||
.duration_since(UNIX_EPOCH)
|
||||
.with_context(|| "Invalid system time")
|
||||
.expect("Unable to get unix epoch time")
|
||||
}
|
||||
|
||||
pub fn encode_uri(v: &str) -> String {
|
||||
@@ -19,7 +19,7 @@ pub fn encode_uri(v: &str) -> String {
|
||||
parts.join("/")
|
||||
}
|
||||
|
||||
pub fn decode_uri(v: &str) -> Option<Cow<str>> {
|
||||
pub fn decode_uri(v: &str) -> Option<Cow<'_, str>> {
|
||||
percent_encoding::percent_decode(v.as_bytes())
|
||||
.decode_utf8()
|
||||
.ok()
|
||||
@@ -100,36 +100,42 @@ pub fn load_private_key<T: AsRef<Path>>(filename: T) -> Result<PrivateKeyDer<'st
|
||||
anyhow::bail!("No supported private key in file");
|
||||
}
|
||||
|
||||
pub fn parse_range(range: &str, size: u64) -> Option<(u64, u64)> {
|
||||
let (unit, range) = range.split_once('=')?;
|
||||
if unit != "bytes" || range.contains(',') {
|
||||
pub fn parse_range(range: &str, size: u64) -> Option<Vec<(u64, u64)>> {
|
||||
let (unit, ranges) = range.split_once('=')?;
|
||||
if unit != "bytes" {
|
||||
return None;
|
||||
}
|
||||
let (start, end) = range.split_once('-')?;
|
||||
if start.is_empty() {
|
||||
let offset = end.parse::<u64>().ok()?;
|
||||
if offset <= size {
|
||||
Some((size - offset, size - 1))
|
||||
} else {
|
||||
None
|
||||
}
|
||||
} else {
|
||||
let start = start.parse::<u64>().ok()?;
|
||||
if start < size {
|
||||
if end.is_empty() {
|
||||
Some((start, size - 1))
|
||||
|
||||
let mut result = Vec::new();
|
||||
for range in ranges.split(',') {
|
||||
let (start, end) = range.trim().split_once('-')?;
|
||||
if start.is_empty() {
|
||||
let offset = end.parse::<u64>().ok()?;
|
||||
if offset <= size {
|
||||
result.push((size - offset, size - 1));
|
||||
} else {
|
||||
let end = end.parse::<u64>().ok()?;
|
||||
if end < size {
|
||||
Some((start, end))
|
||||
} else {
|
||||
None
|
||||
}
|
||||
return None;
|
||||
}
|
||||
} else {
|
||||
None
|
||||
let start = start.parse::<u64>().ok()?;
|
||||
if start < size {
|
||||
if end.is_empty() {
|
||||
result.push((start, size - 1));
|
||||
} else {
|
||||
let end = end.parse::<u64>().ok()?;
|
||||
if end < size {
|
||||
result.push((start, end));
|
||||
} else {
|
||||
return None;
|
||||
}
|
||||
}
|
||||
} else {
|
||||
return None;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
Some(result)
|
||||
}
|
||||
|
||||
#[cfg(test)]
|
||||
@@ -162,13 +168,19 @@ mod tests {
|
||||
|
||||
#[test]
|
||||
fn test_parse_range() {
|
||||
assert_eq!(parse_range("bytes=0-499", 500), Some((0, 499)));
|
||||
assert_eq!(parse_range("bytes=0-", 500), Some((0, 499)));
|
||||
assert_eq!(parse_range("bytes=299-", 500), Some((299, 499)));
|
||||
assert_eq!(parse_range("bytes=-500", 500), Some((0, 499)));
|
||||
assert_eq!(parse_range("bytes=-300", 500), Some((200, 499)));
|
||||
assert_eq!(parse_range("bytes=0-499", 500), Some(vec![(0, 499)]));
|
||||
assert_eq!(parse_range("bytes=0-", 500), Some(vec![(0, 499)]));
|
||||
assert_eq!(parse_range("bytes=299-", 500), Some(vec![(299, 499)]));
|
||||
assert_eq!(parse_range("bytes=-500", 500), Some(vec![(0, 499)]));
|
||||
assert_eq!(parse_range("bytes=-300", 500), Some(vec![(200, 499)]));
|
||||
assert_eq!(
|
||||
parse_range("bytes=0-199, 100-399, 400-, -200", 500),
|
||||
Some(vec![(0, 199), (100, 399), (400, 499), (300, 499)])
|
||||
);
|
||||
assert_eq!(parse_range("bytes=500-", 500), None);
|
||||
assert_eq!(parse_range("bytes=-501", 500), None);
|
||||
assert_eq!(parse_range("bytes=0-500", 500), None);
|
||||
assert_eq!(parse_range("bytes=0-199,", 500), None);
|
||||
assert_eq!(parse_range("bytes=0-199, 500-", 500), None);
|
||||
}
|
||||
}
|
||||
|
||||
@@ -115,7 +115,7 @@ fn assets_override(tmpdir: TempDir, port: u16) -> Result<(), Error> {
|
||||
let url = format!("http://localhost:{port}");
|
||||
let resp = reqwest::blocking::get(&url)?;
|
||||
assert!(resp.text()?.starts_with(&format!(
|
||||
"/__dufs_v{}__/index.js;DATA",
|
||||
"/__dufs_v{}__/index.js;<template id=\"index-data\">",
|
||||
env!("CARGO_PKG_VERSION")
|
||||
)));
|
||||
let resp = reqwest::blocking::get(&url)?;
|
||||
|
||||
@@ -57,17 +57,18 @@ fn invalid_auth(
|
||||
Ok(())
|
||||
}
|
||||
|
||||
const HASHED_PASSWORD_AUTH: &str = "user:$6$gQxZwKyWn/ZmWEA2$4uV7KKMnSUnET2BtWTj/9T5.Jq3h/MdkOlnIl5hdlTxDZ4MZKmJ.kl6C.NL9xnNPqC4lVHC1vuI0E5cLpTJX81@/:rw"; // user:pass
|
||||
|
||||
#[rstest]
|
||||
#[case(server(&["--auth", "user:$6$gQxZwKyWn/ZmWEA2$4uV7KKMnSUnET2BtWTj/9T5.Jq3h/MdkOlnIl5hdlTxDZ4MZKmJ.kl6C.NL9xnNPqC4lVHC1vuI0E5cLpTJX81@/:rw", "-A"]), "user", "pass")]
|
||||
#[case(server(&["--auth", "user:$6$YV1J6OHZAAgbzCbS$V55ZEgvJ6JFdz1nLO4AD696PRHAJYhfQf.Gy2HafrCz5itnbgNTtTgfUSqZrt4BJ7FcpRfSt/QZzAan68pido0@/:rw", "-A"]), "user", "pa:ss@1")]
|
||||
fn auth_hashed_password(
|
||||
#[with(&["--auth", HASHED_PASSWORD_AUTH, "-A"])] server: TestServer,
|
||||
#[case] server: TestServer,
|
||||
#[case] user: &str,
|
||||
#[case] pass: &str,
|
||||
) -> Result<(), Error> {
|
||||
let url = format!("{}file1", server.url());
|
||||
let resp = fetch!(b"PUT", &url).body(b"abc".to_vec()).send()?;
|
||||
assert_eq!(resp.status(), 401);
|
||||
if let Err(err) =
|
||||
send_with_digest_auth(fetch!(b"PUT", &url).body(b"abc".to_vec()), "user", "pass")
|
||||
if let Err(err) = send_with_digest_auth(fetch!(b"PUT", &url).body(b"abc".to_vec()), user, pass)
|
||||
{
|
||||
assert_eq!(
|
||||
err.to_string(),
|
||||
@@ -76,7 +77,7 @@ fn auth_hashed_password(
|
||||
}
|
||||
let resp = fetch!(b"PUT", &url)
|
||||
.body(b"abc".to_vec())
|
||||
.basic_auth("user", Some("pass"))
|
||||
.basic_auth(user, Some(pass))
|
||||
.send()?;
|
||||
assert_eq!(resp.status(), 201);
|
||||
Ok(())
|
||||
@@ -115,30 +116,52 @@ fn auth_skip_on_options_method(
|
||||
}
|
||||
|
||||
#[rstest]
|
||||
fn auth_check(
|
||||
#[with(&["--auth", "user:pass@/:rw", "--auth", "user2:pass2@/", "-A"])] server: TestServer,
|
||||
) -> Result<(), Error> {
|
||||
fn auth_skip_if_no_auth_user(server: TestServer) -> Result<(), Error> {
|
||||
let url = format!("{}index.html", server.url());
|
||||
let resp = fetch!(b"WRITEABLE", &url).send()?;
|
||||
assert_eq!(resp.status(), 401);
|
||||
let resp = send_with_digest_auth(fetch!(b"WRITEABLE", &url), "user2", "pass2")?;
|
||||
assert_eq!(resp.status(), 403);
|
||||
let resp = send_with_digest_auth(fetch!(b"WRITEABLE", &url), "user", "pass")?;
|
||||
let resp = fetch!(b"GET", &url)
|
||||
.basic_auth("user", Some("pass"))
|
||||
.send()?;
|
||||
assert_eq!(resp.status(), 200);
|
||||
Ok(())
|
||||
}
|
||||
|
||||
#[rstest]
|
||||
fn auth_compact_rules(
|
||||
fn auth_check(
|
||||
#[with(&["--auth", "user:pass@/:rw", "--auth", "user2:pass2@/", "-A"])] server: TestServer,
|
||||
) -> Result<(), Error> {
|
||||
let url = format!("{}index.html", server.url());
|
||||
let resp = fetch!(b"CHECKAUTH", &url).send()?;
|
||||
assert_eq!(resp.status(), 401);
|
||||
let resp = send_with_digest_auth(fetch!(b"CHECKAUTH", &url), "user", "pass")?;
|
||||
assert_eq!(resp.status(), 200);
|
||||
let resp = send_with_digest_auth(fetch!(b"CHECKAUTH", &url), "user2", "pass2")?;
|
||||
assert_eq!(resp.status(), 200);
|
||||
Ok(())
|
||||
}
|
||||
|
||||
#[rstest]
|
||||
fn auth_check2(
|
||||
#[with(&["--auth", "user:pass@/:rw|user2:pass2@/", "-A"])] server: TestServer,
|
||||
) -> Result<(), Error> {
|
||||
let url = format!("{}index.html", server.url());
|
||||
let resp = fetch!(b"WRITEABLE", &url).send()?;
|
||||
let resp = fetch!(b"CHECKAUTH", &url).send()?;
|
||||
assert_eq!(resp.status(), 401);
|
||||
let resp = send_with_digest_auth(fetch!(b"WRITEABLE", &url), "user2", "pass2")?;
|
||||
assert_eq!(resp.status(), 403);
|
||||
let resp = send_with_digest_auth(fetch!(b"WRITEABLE", &url), "user", "pass")?;
|
||||
let resp = send_with_digest_auth(fetch!(b"CHECKAUTH", &url), "user", "pass")?;
|
||||
assert_eq!(resp.status(), 200);
|
||||
let resp = send_with_digest_auth(fetch!(b"CHECKAUTH", &url), "user2", "pass2")?;
|
||||
assert_eq!(resp.status(), 200);
|
||||
Ok(())
|
||||
}
|
||||
|
||||
#[rstest]
|
||||
fn auth_logout(
|
||||
#[with(&["--auth", "user:pass@/:rw", "-A"])] server: TestServer,
|
||||
) -> Result<(), Error> {
|
||||
let url = format!("{}index.html", server.url());
|
||||
let resp = fetch!(b"LOGOUT", &url).send()?;
|
||||
assert_eq!(resp.status(), 401);
|
||||
let resp = send_with_digest_auth(fetch!(b"LOGOUT", &url), "user", "pass")?;
|
||||
assert_eq!(resp.status(), 401);
|
||||
Ok(())
|
||||
}
|
||||
|
||||
@@ -313,16 +336,31 @@ fn auth_data(
|
||||
}
|
||||
|
||||
#[rstest]
|
||||
fn auth_precedence(
|
||||
#[with(&["--auth", "user:pass@/dir1:rw,/dir1/test.txt", "-A"])] server: TestServer,
|
||||
fn auth_shadow(
|
||||
#[with(&["--auth", "user:pass@/:rw", "-a", "@/dir1", "-A"])] server: TestServer,
|
||||
) -> Result<(), Error> {
|
||||
let url = format!("{}dir1/test.txt", server.url());
|
||||
let resp = send_with_digest_auth(fetch!(b"PUT", &url).body(b"abc".to_vec()), "user", "pass")?;
|
||||
assert_eq!(resp.status(), 403);
|
||||
let resp = fetch!(b"PUT", &url).body(b"abc".to_vec()).send()?;
|
||||
assert_eq!(resp.status(), 401);
|
||||
|
||||
let url = format!("{}dir1/file1", server.url());
|
||||
let resp = send_with_digest_auth(fetch!(b"PUT", &url).body(b"abc".to_vec()), "user", "pass")?;
|
||||
assert_eq!(resp.status(), 201);
|
||||
|
||||
Ok(())
|
||||
}
|
||||
|
||||
#[rstest]
|
||||
fn token_auth(#[with(&["-a", "user:pass@/"])] server: TestServer) -> Result<(), Error> {
|
||||
let url = format!("{}index.html", server.url());
|
||||
let resp = fetch!(b"GET", &url).send()?;
|
||||
assert_eq!(resp.status(), 401);
|
||||
let url = format!("{}index.html?tokengen", server.url());
|
||||
let resp = fetch!(b"GET", &url)
|
||||
.basic_auth("user", Some("pass"))
|
||||
.send()?;
|
||||
let token = resp.text()?;
|
||||
let url = format!("{}index.html?token={token}", server.url());
|
||||
let resp = fetch!(b"GET", &url).send()?;
|
||||
assert_eq!(resp.status(), 200);
|
||||
Ok(())
|
||||
}
|
||||
|
||||
80
tests/cache.rs
Normal file
80
tests/cache.rs
Normal file
@@ -0,0 +1,80 @@
|
||||
mod fixtures;
|
||||
mod utils;
|
||||
|
||||
use chrono::{DateTime, Duration};
|
||||
use fixtures::{server, Error, TestServer};
|
||||
use reqwest::header::{
|
||||
HeaderName, ETAG, IF_MATCH, IF_MODIFIED_SINCE, IF_NONE_MATCH, IF_UNMODIFIED_SINCE,
|
||||
LAST_MODIFIED,
|
||||
};
|
||||
use reqwest::StatusCode;
|
||||
use rstest::rstest;
|
||||
|
||||
#[rstest]
|
||||
#[case(IF_UNMODIFIED_SINCE, Duration::days(1), StatusCode::OK)]
|
||||
#[case(IF_UNMODIFIED_SINCE, Duration::days(0), StatusCode::OK)]
|
||||
#[case(IF_UNMODIFIED_SINCE, Duration::days(-1), StatusCode::PRECONDITION_FAILED)]
|
||||
#[case(IF_MODIFIED_SINCE, Duration::days(1), StatusCode::NOT_MODIFIED)]
|
||||
#[case(IF_MODIFIED_SINCE, Duration::days(0), StatusCode::NOT_MODIFIED)]
|
||||
#[case(IF_MODIFIED_SINCE, Duration::days(-1), StatusCode::OK)]
|
||||
fn get_file_with_if_modified_since_condition(
|
||||
#[case] header_condition: HeaderName,
|
||||
#[case] duration_after_file_modified: Duration,
|
||||
#[case] expected_code: StatusCode,
|
||||
server: TestServer,
|
||||
) -> Result<(), Error> {
|
||||
let resp = fetch!(b"HEAD", format!("{}index.html", server.url())).send()?;
|
||||
|
||||
let last_modified = resp
|
||||
.headers()
|
||||
.get(LAST_MODIFIED)
|
||||
.and_then(|h| h.to_str().ok())
|
||||
.and_then(|s| DateTime::parse_from_rfc2822(s).ok())
|
||||
.expect("Received no valid last modified header");
|
||||
|
||||
let req_modified_time = (last_modified + duration_after_file_modified)
|
||||
.format("%a, %d %b %Y %T GMT")
|
||||
.to_string();
|
||||
|
||||
let resp = fetch!(b"GET", format!("{}index.html", server.url()))
|
||||
.header(header_condition, req_modified_time)
|
||||
.send()?;
|
||||
|
||||
assert_eq!(resp.status(), expected_code);
|
||||
Ok(())
|
||||
}
|
||||
|
||||
fn same_etag(etag: &str) -> String {
|
||||
etag.to_owned()
|
||||
}
|
||||
|
||||
fn different_etag(etag: &str) -> String {
|
||||
format!("{etag}1234")
|
||||
}
|
||||
|
||||
#[rstest]
|
||||
#[case(IF_MATCH, same_etag, StatusCode::OK)]
|
||||
#[case(IF_MATCH, different_etag, StatusCode::PRECONDITION_FAILED)]
|
||||
#[case(IF_NONE_MATCH, same_etag, StatusCode::NOT_MODIFIED)]
|
||||
#[case(IF_NONE_MATCH, different_etag, StatusCode::OK)]
|
||||
fn get_file_with_etag_match(
|
||||
#[case] header_condition: HeaderName,
|
||||
#[case] etag_modifier: fn(&str) -> String,
|
||||
#[case] expected_code: StatusCode,
|
||||
server: TestServer,
|
||||
) -> Result<(), Error> {
|
||||
let resp = fetch!(b"HEAD", format!("{}index.html", server.url())).send()?;
|
||||
|
||||
let etag = resp
|
||||
.headers()
|
||||
.get(ETAG)
|
||||
.and_then(|h| h.to_str().ok())
|
||||
.expect("Received no valid etag header");
|
||||
|
||||
let resp = fetch!(b"GET", format!("{}index.html", server.url()))
|
||||
.header(header_condition, etag_modifier(etag))
|
||||
.send()?;
|
||||
|
||||
assert_eq!(resp.status(), expected_code);
|
||||
Ok(())
|
||||
}
|
||||
@@ -65,7 +65,7 @@ pub fn tmpdir() -> TempDir {
|
||||
if *directory == DIR_ASSETS {
|
||||
tmpdir
|
||||
.child(format!("{}{}", directory, "index.html"))
|
||||
.write_str("__ASSETS_PREFIX__index.js;DATA = __INDEX_DATA__")
|
||||
.write_str("__ASSETS_PREFIX__index.js;<template id=\"index-data\">__INDEX_DATA__</template>")
|
||||
.unwrap();
|
||||
} else {
|
||||
for file in FILES {
|
||||
@@ -146,14 +146,14 @@ where
|
||||
TestServer::new(port, tmpdir, child, is_tls)
|
||||
}
|
||||
|
||||
/// Wait a max of 1s for the port to become available.
|
||||
/// Wait a max of 2s for the port to become available.
|
||||
pub fn wait_for_port(port: u16) {
|
||||
let start_wait = Instant::now();
|
||||
|
||||
while !port_check::is_port_reachable(format!("localhost:{port}")) {
|
||||
sleep(Duration::from_millis(100));
|
||||
sleep(Duration::from_millis(250));
|
||||
|
||||
if start_wait.elapsed().as_secs() > 1 {
|
||||
if start_wait.elapsed().as_secs() > 2 {
|
||||
panic!("timeout waiting for port {port}");
|
||||
}
|
||||
}
|
||||
|
||||
31
tests/health.rs
Normal file
31
tests/health.rs
Normal file
@@ -0,0 +1,31 @@
|
||||
mod fixtures;
|
||||
mod utils;
|
||||
|
||||
use fixtures::{server, Error, TestServer};
|
||||
use rstest::rstest;
|
||||
|
||||
const HEALTH_CHECK_PATH: &str = "__dufs__/health";
|
||||
const HEALTH_CHECK_RESPONSE: &str = r#"{"status":"OK"}"#;
|
||||
|
||||
#[rstest]
|
||||
fn normal_health(server: TestServer) -> Result<(), Error> {
|
||||
let resp = reqwest::blocking::get(format!("{}{HEALTH_CHECK_PATH}", server.url()))?;
|
||||
assert_eq!(resp.text()?, HEALTH_CHECK_RESPONSE);
|
||||
Ok(())
|
||||
}
|
||||
|
||||
#[rstest]
|
||||
fn auth_health(
|
||||
#[with(&["--auth", "user:pass@/:rw", "-A"])] server: TestServer,
|
||||
) -> Result<(), Error> {
|
||||
let resp = reqwest::blocking::get(format!("{}{HEALTH_CHECK_PATH}", server.url()))?;
|
||||
assert_eq!(resp.text()?, HEALTH_CHECK_RESPONSE);
|
||||
Ok(())
|
||||
}
|
||||
|
||||
#[rstest]
|
||||
fn path_prefix_health(#[with(&["--path-prefix", "xyz"])] server: TestServer) -> Result<(), Error> {
|
||||
let resp = reqwest::blocking::get(format!("{}xyz/{HEALTH_CHECK_PATH}", server.url()))?;
|
||||
assert_eq!(resp.text()?, HEALTH_CHECK_RESPONSE);
|
||||
Ok(())
|
||||
}
|
||||
@@ -82,6 +82,19 @@ fn get_dir_simple(#[with(&["-A"])] server: TestServer) -> Result<(), Error> {
|
||||
Ok(())
|
||||
}
|
||||
|
||||
#[rstest]
|
||||
fn get_dir_noscript(#[with(&["-A"])] server: TestServer) -> Result<(), Error> {
|
||||
let resp = reqwest::blocking::get(format!("{}?noscript", server.url()))?;
|
||||
assert_eq!(resp.status(), 200);
|
||||
assert_eq!(
|
||||
resp.headers().get("content-type").unwrap(),
|
||||
"text/html; charset=utf-8"
|
||||
);
|
||||
let text = resp.text().unwrap();
|
||||
assert!(text.contains(r#"<td><a href="index.html">index.html</a></td>"#));
|
||||
Ok(())
|
||||
}
|
||||
|
||||
#[rstest]
|
||||
fn head_dir_zip(#[with(&["-A"])] server: TestServer) -> Result<(), Error> {
|
||||
let resp = fetch!(b"HEAD", format!("{}?zip", server.url())).send()?;
|
||||
@@ -265,12 +278,9 @@ fn options_dir(server: TestServer) -> Result<(), Error> {
|
||||
assert_eq!(resp.status(), 200);
|
||||
assert_eq!(
|
||||
resp.headers().get("allow").unwrap(),
|
||||
"GET,HEAD,PUT,OPTIONS,DELETE,PATCH,PROPFIND,COPY,MOVE"
|
||||
);
|
||||
assert_eq!(
|
||||
resp.headers().get("dav").unwrap(),
|
||||
"1, 2, 3, sabredav-partialupdate"
|
||||
"GET,HEAD,PUT,OPTIONS,DELETE,PATCH,PROPFIND,COPY,MOVE,CHECKAUTH,LOGOUT"
|
||||
);
|
||||
assert_eq!(resp.headers().get("dav").unwrap(), "1, 2, 3");
|
||||
Ok(())
|
||||
}
|
||||
|
||||
|
||||
@@ -2,7 +2,7 @@ mod fixtures;
|
||||
mod utils;
|
||||
|
||||
use fixtures::{server, Error, TestServer};
|
||||
use reqwest::header::HeaderValue;
|
||||
use reqwest::header::{HeaderMap, HeaderName, HeaderValue};
|
||||
use rstest::rstest;
|
||||
|
||||
#[rstest]
|
||||
@@ -39,3 +39,68 @@ fn get_file_range_invalid(server: TestServer) -> Result<(), Error> {
|
||||
assert_eq!(resp.headers().get("content-range").unwrap(), "bytes */18");
|
||||
Ok(())
|
||||
}
|
||||
|
||||
fn parse_multipart_body<'a>(body: &'a str, boundary: &str) -> Vec<(HeaderMap, &'a str)> {
|
||||
body.split(&format!("--{boundary}"))
|
||||
.filter(|part| !part.is_empty() && *part != "--\r\n")
|
||||
.map(|part| {
|
||||
let (head, body) = part.trim_ascii().split_once("\r\n\r\n").unwrap();
|
||||
let headers = head
|
||||
.split("\r\n")
|
||||
.fold(HeaderMap::new(), |mut headers, header| {
|
||||
let (key, value) = header.split_once(":").unwrap();
|
||||
let key = HeaderName::from_bytes(key.as_bytes()).unwrap();
|
||||
let value = HeaderValue::from_str(value.trim_ascii_start()).unwrap();
|
||||
headers.insert(key, value);
|
||||
headers
|
||||
});
|
||||
(headers, body)
|
||||
})
|
||||
.collect()
|
||||
}
|
||||
|
||||
#[rstest]
|
||||
fn get_file_multipart_range(server: TestServer) -> Result<(), Error> {
|
||||
let resp = fetch!(b"GET", format!("{}index.html", server.url()))
|
||||
.header("range", HeaderValue::from_static("bytes=0-11, 6-17"))
|
||||
.send()?;
|
||||
assert_eq!(resp.status(), 206);
|
||||
assert_eq!(resp.headers().get("accept-ranges").unwrap(), "bytes");
|
||||
|
||||
let content_type = resp
|
||||
.headers()
|
||||
.get("content-type")
|
||||
.unwrap()
|
||||
.to_str()?
|
||||
.to_string();
|
||||
assert!(content_type.starts_with("multipart/byteranges; boundary="));
|
||||
|
||||
let boundary = content_type.split_once('=').unwrap().1.trim_ascii_start();
|
||||
assert!(!boundary.is_empty());
|
||||
|
||||
let body = resp.text()?;
|
||||
let parts = parse_multipart_body(&body, boundary);
|
||||
assert_eq!(parts.len(), 2);
|
||||
|
||||
let (headers, body) = &parts[0];
|
||||
assert_eq!(headers.get("content-range").unwrap(), "bytes 0-11/18");
|
||||
assert_eq!(*body, "This is inde");
|
||||
|
||||
let (headers, body) = &parts[1];
|
||||
assert_eq!(headers.get("content-range").unwrap(), "bytes 6-17/18");
|
||||
assert_eq!(*body, "s index.html");
|
||||
|
||||
Ok(())
|
||||
}
|
||||
|
||||
#[rstest]
|
||||
fn get_file_multipart_range_invalid(server: TestServer) -> Result<(), Error> {
|
||||
let resp = fetch!(b"GET", format!("{}index.html", server.url()))
|
||||
.header("range", HeaderValue::from_static("bytes=0-6, 20-30"))
|
||||
.send()?;
|
||||
assert_eq!(resp.status(), 416);
|
||||
assert_eq!(resp.headers().get("content-range").unwrap(), "bytes */18");
|
||||
assert_eq!(resp.headers().get("accept-ranges").unwrap(), "bytes");
|
||||
assert_eq!(resp.headers().get("content-length").unwrap(), "0");
|
||||
Ok(())
|
||||
}
|
||||
|
||||
@@ -1,3 +1,4 @@
|
||||
use base64::{engine::general_purpose::STANDARD, Engine as _};
|
||||
use indexmap::IndexSet;
|
||||
use serde_json::Value;
|
||||
|
||||
@@ -48,7 +49,7 @@ pub fn retrieve_index_paths(content: &str) -> IndexSet<String> {
|
||||
|
||||
#[allow(dead_code)]
|
||||
pub fn retrieve_edit_file(content: &str) -> Option<bool> {
|
||||
let value = retrieve_json(content)?;
|
||||
let value = retrieve_json(content).unwrap();
|
||||
let value = value.get("editable").unwrap();
|
||||
Some(value.as_bool().unwrap())
|
||||
}
|
||||
@@ -62,8 +63,20 @@ pub fn encode_uri(v: &str) -> String {
|
||||
#[allow(dead_code)]
|
||||
pub fn retrieve_json(content: &str) -> Option<Value> {
|
||||
let lines: Vec<&str> = content.lines().collect();
|
||||
let line = lines.iter().find(|v| v.contains("DATA ="))?;
|
||||
let line_col = line.find("DATA =").unwrap() + 6;
|
||||
let value: Value = line[line_col..].parse().unwrap();
|
||||
let start_tag = "<template id=\"index-data\">";
|
||||
let end_tag = "</template>";
|
||||
|
||||
let line = lines.iter().find(|v| v.contains(start_tag))?;
|
||||
|
||||
let start_index = line.find(start_tag)?;
|
||||
let start_content_index = start_index + start_tag.len();
|
||||
|
||||
let end_index = line[start_content_index..].find(end_tag)?;
|
||||
let end_content_index = start_content_index + end_index;
|
||||
|
||||
let value = &line[start_content_index..end_content_index];
|
||||
let value = STANDARD.decode(value).ok()?;
|
||||
let value = serde_json::from_slice(&value).ok()?;
|
||||
|
||||
Some(value)
|
||||
}
|
||||
|
||||
@@ -40,6 +40,17 @@ fn propfind_dir_depth0(server: TestServer) -> Result<(), Error> {
|
||||
Ok(())
|
||||
}
|
||||
|
||||
#[rstest]
|
||||
fn propfind_dir_depth2(server: TestServer) -> Result<(), Error> {
|
||||
let resp = fetch!(b"PROPFIND", format!("{}dir1", server.url()))
|
||||
.header("depth", "2")
|
||||
.send()?;
|
||||
assert_eq!(resp.status(), 400);
|
||||
let body = resp.text()?;
|
||||
assert_eq!(body, "Invalid depth: only 0 and 1 are allowed.");
|
||||
Ok(())
|
||||
}
|
||||
|
||||
#[rstest]
|
||||
fn propfind_404(server: TestServer) -> Result<(), Error> {
|
||||
let resp = fetch!(b"PROPFIND", format!("{}404", server.url())).send()?;
|
||||
|
||||
Reference in New Issue
Block a user