mirror of
https://github.com/bootandy/dust.git
synced 2025-12-13 08:00:46 -08:00
Compare commits
1 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
082f15a0e2 |
5
.github/workflows/CICD.yml
vendored
5
.github/workflows/CICD.yml
vendored
@@ -45,11 +45,6 @@ jobs:
|
|||||||
override: true
|
override: true
|
||||||
profile: minimal # minimal component installation (ie, no documentation)
|
profile: minimal # minimal component installation (ie, no documentation)
|
||||||
components: rustfmt, clippy
|
components: rustfmt, clippy
|
||||||
- name: Install wget for Windows
|
|
||||||
if: matrix.job.os == 'windows-latest'
|
|
||||||
run: choco install wget --no-progress
|
|
||||||
- name: typos-action
|
|
||||||
uses: crate-ci/typos@v1.28.4
|
|
||||||
- name: "`fmt` testing"
|
- name: "`fmt` testing"
|
||||||
if: steps.vars.outputs.JOB_DO_FORMAT_TESTING
|
if: steps.vars.outputs.JOB_DO_FORMAT_TESTING
|
||||||
uses: actions-rs/cargo@v1
|
uses: actions-rs/cargo@v1
|
||||||
|
|||||||
@@ -1,11 +0,0 @@
|
|||||||
repos:
|
|
||||||
- repo: https://github.com/doublify/pre-commit-rust
|
|
||||||
rev: v1.0
|
|
||||||
hooks:
|
|
||||||
- id: cargo-check
|
|
||||||
stages: [commit]
|
|
||||||
- id: fmt
|
|
||||||
stages: [commit]
|
|
||||||
- id: clippy
|
|
||||||
args: [--all-targets, --all-features]
|
|
||||||
stages: [commit]
|
|
||||||
578
Cargo.lock
generated
578
Cargo.lock
generated
File diff suppressed because it is too large
Load Diff
19
Cargo.toml
19
Cargo.toml
@@ -1,9 +1,9 @@
|
|||||||
[package]
|
[package]
|
||||||
name = "du-dust"
|
name = "du-dust"
|
||||||
description = "A more intuitive version of du"
|
description = "A more intuitive version of du"
|
||||||
version = "1.2.0"
|
version = "1.0.0"
|
||||||
authors = ["bootandy <bootandy@gmail.com>", "nebkor <code@ardent.nebcorp.com>"]
|
authors = ["bootandy <bootandy@gmail.com>", "nebkor <code@ardent.nebcorp.com>"]
|
||||||
edition = "2024"
|
edition = "2021"
|
||||||
readme = "README.md"
|
readme = "README.md"
|
||||||
|
|
||||||
documentation = "https://github.com/bootandy/dust"
|
documentation = "https://github.com/bootandy/dust"
|
||||||
@@ -38,14 +38,9 @@ stfu8 = "0.2"
|
|||||||
regex = "1"
|
regex = "1"
|
||||||
config-file = "0.2"
|
config-file = "0.2"
|
||||||
serde = { version = "1.0", features = ["derive"] }
|
serde = { version = "1.0", features = ["derive"] }
|
||||||
serde_json = "1.0"
|
|
||||||
directories = "4"
|
directories = "4"
|
||||||
sysinfo = "0.27"
|
sysinfo = "0.27"
|
||||||
ctrlc = "3.4"
|
ctrlc = "3.4"
|
||||||
chrono = "0.4"
|
|
||||||
|
|
||||||
[target.'cfg(not(target_has_atomic = "64"))'.dependencies]
|
|
||||||
portable-atomic = "1.4"
|
|
||||||
|
|
||||||
[target.'cfg(windows)'.dependencies]
|
[target.'cfg(windows)'.dependencies]
|
||||||
winapi-util = "0.1"
|
winapi-util = "0.1"
|
||||||
@@ -86,16 +81,6 @@ assets = [
|
|||||||
"usr/share/doc/du-dust/README",
|
"usr/share/doc/du-dust/README",
|
||||||
"644",
|
"644",
|
||||||
],
|
],
|
||||||
[
|
|
||||||
"man-page/dust.1",
|
|
||||||
"usr/share/man/man1/dust.1",
|
|
||||||
"644",
|
|
||||||
],
|
|
||||||
[
|
|
||||||
"completions/dust.bash",
|
|
||||||
"usr/share/bash-completion/completions/dust",
|
|
||||||
"644",
|
|
||||||
],
|
|
||||||
]
|
]
|
||||||
extended-description = """\
|
extended-description = """\
|
||||||
Dust is meant to give you an instant overview of which directories are using
|
Dust is meant to give you an instant overview of which directories are using
|
||||||
|
|||||||
15
README.md
15
README.md
@@ -27,17 +27,11 @@ Because I want an easy way to see where my disk is being used.
|
|||||||
|
|
||||||
- `brew install dust`
|
- `brew install dust`
|
||||||
|
|
||||||
#### [Snap](https://ubuntu.com/core/services/guide/snaps-intro) Ubuntu and [supported systems](https://snapcraft.io/docs/installing-snapd)
|
|
||||||
|
|
||||||
- `snap install dust`
|
|
||||||
|
|
||||||
Note: `dust` installed through `snap` can only access files stored in the `/home` directory. See daniejstriata/dust-snap#2 for more information.
|
|
||||||
|
|
||||||
#### [Pacstall](https://github.com/pacstall/pacstall) (Debian/Ubuntu)
|
#### [Pacstall](https://github.com/pacstall/pacstall) (Debian/Ubuntu)
|
||||||
|
|
||||||
- `pacstall -I dust-bin`
|
- `pacstall -I dust-bin`
|
||||||
|
|
||||||
#### Anaconda (conda-forge)
|
### Anaconda (conda-forge)
|
||||||
|
|
||||||
- `conda install -c conda-forge dust`
|
- `conda install -c conda-forge dust`
|
||||||
|
|
||||||
@@ -45,10 +39,6 @@ Note: `dust` installed through `snap` can only access files stored in the `/home
|
|||||||
|
|
||||||
- `deb-get install du-dust`
|
- `deb-get install du-dust`
|
||||||
|
|
||||||
#### [x-cmd](https://www.x-cmd.com/pkg/#VPContent)
|
|
||||||
|
|
||||||
- `x env use dust`
|
|
||||||
|
|
||||||
#### Windows:
|
#### Windows:
|
||||||
|
|
||||||
- `scoop install dust`
|
- `scoop install dust`
|
||||||
@@ -101,9 +91,6 @@ Usage: dust -R (For screen readers. Removes bars/symbols. Adds new column: depth
|
|||||||
Usage: dust -S (Custom Stack size - Use if you see: 'fatal runtime error: stack overflow' (default allocation: low memory=1048576, high memory=1073741824)"),
|
Usage: dust -S (Custom Stack size - Use if you see: 'fatal runtime error: stack overflow' (default allocation: low memory=1048576, high memory=1073741824)"),
|
||||||
Usage: dust --skip-total (No total row will be displayed)
|
Usage: dust --skip-total (No total row will be displayed)
|
||||||
Usage: dust -z 40000/30MB/20kib (Exclude output files/directories below size 40000 bytes / 30MB / 20KiB)
|
Usage: dust -z 40000/30MB/20kib (Exclude output files/directories below size 40000 bytes / 30MB / 20KiB)
|
||||||
Usage: dust -j (Prints JSON representation of directories, try: dust -j | jq)
|
|
||||||
Usage: dust --files0-from=FILE (Reads null-terminated file paths from FILE); If FILE is - then read from stdin
|
|
||||||
Usage: dust --collapse=node-modules will keep the node-modules folder collapsed in display instead of recursively opening it
|
|
||||||
```
|
```
|
||||||
|
|
||||||
## Config file
|
## Config file
|
||||||
|
|||||||
@@ -1,21 +1,14 @@
|
|||||||
# ----------- To do a release ---------
|
# ----------- To do a release ---------
|
||||||
|
|
||||||
# ----------- Pre release ---------
|
|
||||||
# Compare times of runs to check no drastic slow down:
|
# Compare times of runs to check no drastic slow down:
|
||||||
# hyperfine 'target/release/dust /home/andy'
|
# time target/release/dust ~/dev
|
||||||
# hyperfine 'dust /home/andy'
|
# time dust ~dev
|
||||||
|
|
||||||
# ----------- Release ---------
|
|
||||||
# inc version in cargo.toml
|
|
||||||
# cargo build --release
|
|
||||||
# commit changed files
|
|
||||||
# merge to master in github
|
|
||||||
|
|
||||||
|
# edit version in cargo.toml
|
||||||
# tag a commit and push (increment version in Cargo.toml first):
|
# tag a commit and push (increment version in Cargo.toml first):
|
||||||
# git tag v0.4.5
|
# git tag v0.4.5
|
||||||
# git push origin v0.4.5
|
# git push origin v0.4.5
|
||||||
|
|
||||||
# cargo publish to put it in crates.io
|
# cargo publish to put it in crates.io
|
||||||
|
|
||||||
# Optional: To install locally
|
# To install locally [Do before pushing it]
|
||||||
#cargo install --path .
|
#cargo install --path .
|
||||||
|
|||||||
@@ -14,40 +14,27 @@ _dust() {
|
|||||||
fi
|
fi
|
||||||
|
|
||||||
local context curcontext="$curcontext" state line
|
local context curcontext="$curcontext" state line
|
||||||
_arguments "${_arguments_options[@]}" : \
|
_arguments "${_arguments_options[@]}" \
|
||||||
'-d+[Depth to show]:DEPTH:_default' \
|
'-d+[Depth to show]: : ' \
|
||||||
'--depth=[Depth to show]:DEPTH:_default' \
|
'--depth=[Depth to show]: : ' \
|
||||||
'-T+[Number of threads to use]: :_default' \
|
'-n+[Number of lines of output to show. (Default is terminal_height - 10)]: : ' \
|
||||||
'--threads=[Number of threads to use]: :_default' \
|
'--number-of-lines=[Number of lines of output to show. (Default is terminal_height - 10)]: : ' \
|
||||||
'--config=[Specify a config file to use]:FILE:_files' \
|
'*-X+[Exclude any file or directory with this name]: : ' \
|
||||||
'-n+[Number of lines of output to show. (Default is terminal_height - 10)]:NUMBER:_default' \
|
'*--ignore-directory=[Exclude any file or directory with this name]: : ' \
|
||||||
'--number-of-lines=[Number of lines of output to show. (Default is terminal_height - 10)]:NUMBER:_default' \
|
'-I+[Exclude any file or directory with a regex matching that listed in this file, the file entries will be added to the ignore regexs provided by --invert_filter]: : ' \
|
||||||
'*-X+[Exclude any file or directory with this path]:PATH:_files' \
|
'--ignore-all-in-file=[Exclude any file or directory with a regex matching that listed in this file, the file entries will be added to the ignore regexs provided by --invert_filter]: : ' \
|
||||||
'*--ignore-directory=[Exclude any file or directory with this path]:PATH:_files' \
|
'-z+[Minimum size file to include in output]: : ' \
|
||||||
'-I+[Exclude any file or directory with a regex matching that listed in this file, the file entries will be added to the ignore regexs provided by --invert_filter]:FILE:_files' \
|
'--min-size=[Minimum size file to include in output]: : ' \
|
||||||
'--ignore-all-in-file=[Exclude any file or directory with a regex matching that listed in this file, the file entries will be added to the ignore regexs provided by --invert_filter]:FILE:_files' \
|
'(-e --filter -t --file_types)*-v+[Exclude filepaths matching this regex. To ignore png files type\: -v "\\.png\$" ]: : ' \
|
||||||
'-z+[Minimum size file to include in output]:MIN_SIZE:_default' \
|
'(-e --filter -t --file_types)*--invert-filter=[Exclude filepaths matching this regex. To ignore png files type\: -v "\\.png\$" ]: : ' \
|
||||||
'--min-size=[Minimum size file to include in output]:MIN_SIZE:_default' \
|
'(-t --file_types)*-e+[Only include filepaths matching this regex. For png files type\: -e "\\.png\$" ]: : ' \
|
||||||
'(-e --filter -t --file-types)*-v+[Exclude filepaths matching this regex. To ignore png files type\: -v "\\.png\$" ]:REGEX:_default' \
|
'(-t --file_types)*--filter=[Only include filepaths matching this regex. For png files type\: -e "\\.png\$" ]: : ' \
|
||||||
'(-e --filter -t --file-types)*--invert-filter=[Exclude filepaths matching this regex. To ignore png files type\: -v "\\.png\$" ]:REGEX:_default' \
|
'-w+[Specify width of output overriding the auto detection of terminal width]: : ' \
|
||||||
'(-t --file-types)*-e+[Only include filepaths matching this regex. For png files type\: -e "\\.png\$" ]:REGEX:_default' \
|
'--terminal_width=[Specify width of output overriding the auto detection of terminal width]: : ' \
|
||||||
'(-t --file-types)*--filter=[Only include filepaths matching this regex. For png files type\: -e "\\.png\$" ]:REGEX:_default' \
|
'-o+[Changes output display size. si will print sizes in powers of 1000. b/bytes kb kib mb mib gb gib will print the whole tree in that size]: : ' \
|
||||||
'-w+[Specify width of output overriding the auto detection of terminal width]:WIDTH:_default' \
|
'--output-format=[Changes output display size. si will print sizes in powers of 1000. b/bytes kb kib mb mib gb gib will print the whole tree in that size]: : ' \
|
||||||
'--terminal-width=[Specify width of output overriding the auto detection of terminal width]:WIDTH:_default' \
|
'-S+[Specify memory to use as stack size - use if you see\: '\''fatal runtime error\: stack overflow'\'' (default low memory=1048576, high memory=1073741824)]: : ' \
|
||||||
'-o+[Changes output display size. si will print sizes in powers of 1000. b k m g t kb mb gb tb will print the whole tree in that size.]:FORMAT:(si b k m g t kb mb gb tb)' \
|
'--stack-size=[Specify memory to use as stack size - use if you see\: '\''fatal runtime error\: stack overflow'\'' (default low memory=1048576, high memory=1073741824)]: : ' \
|
||||||
'--output-format=[Changes output display size. si will print sizes in powers of 1000. b k m g t kb mb gb tb will print the whole tree in that size.]:FORMAT:(si b k m g t kb mb gb tb)' \
|
|
||||||
'-S+[Specify memory to use as stack size - use if you see\: '\''fatal runtime error\: stack overflow'\'' (default low memory=1048576, high memory=1073741824)]:STACK_SIZE:_default' \
|
|
||||||
'--stack-size=[Specify memory to use as stack size - use if you see\: '\''fatal runtime error\: stack overflow'\'' (default low memory=1048576, high memory=1073741824)]:STACK_SIZE:_default' \
|
|
||||||
'-M+[+/-n matches files modified more/less than n days ago , and n matches files modified exactly n days ago, days are rounded down.That is +n => (−∞, curr−(n+1)), n => \[curr−(n+1), curr−n), and -n => (𝑐𝑢𝑟𝑟−𝑛, +∞)]: :_default' \
|
|
||||||
'--mtime=[+/-n matches files modified more/less than n days ago , and n matches files modified exactly n days ago, days are rounded down.That is +n => (−∞, curr−(n+1)), n => \[curr−(n+1), curr−n), and -n => (𝑐𝑢𝑟𝑟−𝑛, +∞)]: :_default' \
|
|
||||||
'-A+[just like -mtime, but based on file access time]: :_default' \
|
|
||||||
'--atime=[just like -mtime, but based on file access time]: :_default' \
|
|
||||||
'-y+[just like -mtime, but based on file change time]: :_default' \
|
|
||||||
'--ctime=[just like -mtime, but based on file change time]: :_default' \
|
|
||||||
'--files0-from=[run dust on NUL-terminated file names specified in file; if argument is -, then read names from standard input]: :_files' \
|
|
||||||
'*--collapse=[Keep these directories collapsed]: :_files' \
|
|
||||||
'-m+[Directory '\''size'\'' is max filetime of child files instead of disk size. while a/c/m for last accessed/changed/modified time]: :(a c m)' \
|
|
||||||
'--filetime=[Directory '\''size'\'' is max filetime of child files instead of disk size. while a/c/m for last accessed/changed/modified time]: :(a c m)' \
|
|
||||||
'-p[Subdirectories will not have their path shortened]' \
|
'-p[Subdirectories will not have their path shortened]' \
|
||||||
'--full-paths[Subdirectories will not have their path shortened]' \
|
'--full-paths[Subdirectories will not have their path shortened]' \
|
||||||
'-L[dereference sym links - Treat sym links as directories and go into them]' \
|
'-L[dereference sym links - Treat sym links as directories and go into them]' \
|
||||||
@@ -72,23 +59,20 @@ _dust() {
|
|||||||
'-f[Directory '\''size'\'' is number of child files instead of disk size]' \
|
'-f[Directory '\''size'\'' is number of child files instead of disk size]' \
|
||||||
'--filecount[Directory '\''size'\'' is number of child files instead of disk size]' \
|
'--filecount[Directory '\''size'\'' is number of child files instead of disk size]' \
|
||||||
'-i[Do not display hidden files]' \
|
'-i[Do not display hidden files]' \
|
||||||
'--ignore-hidden[Do not display hidden files]' \
|
'--ignore_hidden[Do not display hidden files]' \
|
||||||
'(-d --depth -D --only-dir)-t[show only these file types]' \
|
'(-d --depth -D --only-dir)-t[show only these file types]' \
|
||||||
'(-d --depth -D --only-dir)--file-types[show only these file types]' \
|
'(-d --depth -D --only-dir)--file_types[show only these file types]' \
|
||||||
'-P[Disable the progress indication.]' \
|
'-P[Disable the progress indication.]' \
|
||||||
'--no-progress[Disable the progress indication.]' \
|
'--no-progress[Disable the progress indication.]' \
|
||||||
'--print-errors[Print path with errors.]' \
|
'(-F --only-file -t --file_types)-D[Only directories will be displayed.]' \
|
||||||
'(-F --only-file -t --file-types)-D[Only directories will be displayed.]' \
|
'(-F --only-file -t --file_types)--only-dir[Only directories will be displayed.]' \
|
||||||
'(-F --only-file -t --file-types)--only-dir[Only directories will be displayed.]' \
|
|
||||||
'(-D --only-dir)-F[Only files will be displayed. (Finds your largest files)]' \
|
'(-D --only-dir)-F[Only files will be displayed. (Finds your largest files)]' \
|
||||||
'(-D --only-dir)--only-file[Only files will be displayed. (Finds your largest files)]' \
|
'(-D --only-dir)--only-file[Only files will be displayed. (Finds your largest files)]' \
|
||||||
'-j[Output the directory tree as json to the current directory]' \
|
|
||||||
'--output-json[Output the directory tree as json to the current directory]' \
|
|
||||||
'-h[Print help]' \
|
'-h[Print help]' \
|
||||||
'--help[Print help]' \
|
'--help[Print help]' \
|
||||||
'-V[Print version]' \
|
'-V[Print version]' \
|
||||||
'--version[Print version]' \
|
'--version[Print version]' \
|
||||||
'*::params:_files' \
|
'*::params:' \
|
||||||
&& ret=0
|
&& ret=0
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -21,79 +21,63 @@ Register-ArgumentCompleter -Native -CommandName 'dust' -ScriptBlock {
|
|||||||
|
|
||||||
$completions = @(switch ($command) {
|
$completions = @(switch ($command) {
|
||||||
'dust' {
|
'dust' {
|
||||||
[CompletionResult]::new('-d', '-d', [CompletionResultType]::ParameterName, 'Depth to show')
|
[CompletionResult]::new('-d', 'd', [CompletionResultType]::ParameterName, 'Depth to show')
|
||||||
[CompletionResult]::new('--depth', '--depth', [CompletionResultType]::ParameterName, 'Depth to show')
|
[CompletionResult]::new('--depth', 'depth', [CompletionResultType]::ParameterName, 'Depth to show')
|
||||||
[CompletionResult]::new('-T', '-T ', [CompletionResultType]::ParameterName, 'Number of threads to use')
|
[CompletionResult]::new('-n', 'n', [CompletionResultType]::ParameterName, 'Number of lines of output to show. (Default is terminal_height - 10)')
|
||||||
[CompletionResult]::new('--threads', '--threads', [CompletionResultType]::ParameterName, 'Number of threads to use')
|
[CompletionResult]::new('--number-of-lines', 'number-of-lines', [CompletionResultType]::ParameterName, 'Number of lines of output to show. (Default is terminal_height - 10)')
|
||||||
[CompletionResult]::new('--config', '--config', [CompletionResultType]::ParameterName, 'Specify a config file to use')
|
[CompletionResult]::new('-X', 'X ', [CompletionResultType]::ParameterName, 'Exclude any file or directory with this name')
|
||||||
[CompletionResult]::new('-n', '-n', [CompletionResultType]::ParameterName, 'Number of lines of output to show. (Default is terminal_height - 10)')
|
[CompletionResult]::new('--ignore-directory', 'ignore-directory', [CompletionResultType]::ParameterName, 'Exclude any file or directory with this name')
|
||||||
[CompletionResult]::new('--number-of-lines', '--number-of-lines', [CompletionResultType]::ParameterName, 'Number of lines of output to show. (Default is terminal_height - 10)')
|
[CompletionResult]::new('-I', 'I ', [CompletionResultType]::ParameterName, 'Exclude any file or directory with a regex matching that listed in this file, the file entries will be added to the ignore regexs provided by --invert_filter')
|
||||||
[CompletionResult]::new('-X', '-X ', [CompletionResultType]::ParameterName, 'Exclude any file or directory with this path')
|
[CompletionResult]::new('--ignore-all-in-file', 'ignore-all-in-file', [CompletionResultType]::ParameterName, 'Exclude any file or directory with a regex matching that listed in this file, the file entries will be added to the ignore regexs provided by --invert_filter')
|
||||||
[CompletionResult]::new('--ignore-directory', '--ignore-directory', [CompletionResultType]::ParameterName, 'Exclude any file or directory with this path')
|
[CompletionResult]::new('-z', 'z', [CompletionResultType]::ParameterName, 'Minimum size file to include in output')
|
||||||
[CompletionResult]::new('-I', '-I ', [CompletionResultType]::ParameterName, 'Exclude any file or directory with a regex matching that listed in this file, the file entries will be added to the ignore regexs provided by --invert_filter')
|
[CompletionResult]::new('--min-size', 'min-size', [CompletionResultType]::ParameterName, 'Minimum size file to include in output')
|
||||||
[CompletionResult]::new('--ignore-all-in-file', '--ignore-all-in-file', [CompletionResultType]::ParameterName, 'Exclude any file or directory with a regex matching that listed in this file, the file entries will be added to the ignore regexs provided by --invert_filter')
|
[CompletionResult]::new('-v', 'v', [CompletionResultType]::ParameterName, 'Exclude filepaths matching this regex. To ignore png files type: -v "\.png$" ')
|
||||||
[CompletionResult]::new('-z', '-z', [CompletionResultType]::ParameterName, 'Minimum size file to include in output')
|
[CompletionResult]::new('--invert-filter', 'invert-filter', [CompletionResultType]::ParameterName, 'Exclude filepaths matching this regex. To ignore png files type: -v "\.png$" ')
|
||||||
[CompletionResult]::new('--min-size', '--min-size', [CompletionResultType]::ParameterName, 'Minimum size file to include in output')
|
[CompletionResult]::new('-e', 'e', [CompletionResultType]::ParameterName, 'Only include filepaths matching this regex. For png files type: -e "\.png$" ')
|
||||||
[CompletionResult]::new('-v', '-v', [CompletionResultType]::ParameterName, 'Exclude filepaths matching this regex. To ignore png files type: -v "\.png$" ')
|
[CompletionResult]::new('--filter', 'filter', [CompletionResultType]::ParameterName, 'Only include filepaths matching this regex. For png files type: -e "\.png$" ')
|
||||||
[CompletionResult]::new('--invert-filter', '--invert-filter', [CompletionResultType]::ParameterName, 'Exclude filepaths matching this regex. To ignore png files type: -v "\.png$" ')
|
[CompletionResult]::new('-w', 'w', [CompletionResultType]::ParameterName, 'Specify width of output overriding the auto detection of terminal width')
|
||||||
[CompletionResult]::new('-e', '-e', [CompletionResultType]::ParameterName, 'Only include filepaths matching this regex. For png files type: -e "\.png$" ')
|
[CompletionResult]::new('--terminal_width', 'terminal_width', [CompletionResultType]::ParameterName, 'Specify width of output overriding the auto detection of terminal width')
|
||||||
[CompletionResult]::new('--filter', '--filter', [CompletionResultType]::ParameterName, 'Only include filepaths matching this regex. For png files type: -e "\.png$" ')
|
[CompletionResult]::new('-o', 'o', [CompletionResultType]::ParameterName, 'Changes output display size. si will print sizes in powers of 1000. b/bytes kb kib mb mib gb gib will print the whole tree in that size')
|
||||||
[CompletionResult]::new('-w', '-w', [CompletionResultType]::ParameterName, 'Specify width of output overriding the auto detection of terminal width')
|
[CompletionResult]::new('--output-format', 'output-format', [CompletionResultType]::ParameterName, 'Changes output display size. si will print sizes in powers of 1000. b/bytes kb kib mb mib gb gib will print the whole tree in that size')
|
||||||
[CompletionResult]::new('--terminal-width', '--terminal-width', [CompletionResultType]::ParameterName, 'Specify width of output overriding the auto detection of terminal width')
|
[CompletionResult]::new('-S', 'S ', [CompletionResultType]::ParameterName, 'Specify memory to use as stack size - use if you see: ''fatal runtime error: stack overflow'' (default low memory=1048576, high memory=1073741824)')
|
||||||
[CompletionResult]::new('-o', '-o', [CompletionResultType]::ParameterName, 'Changes output display size. si will print sizes in powers of 1000. b k m g t kb mb gb tb will print the whole tree in that size.')
|
[CompletionResult]::new('--stack-size', 'stack-size', [CompletionResultType]::ParameterName, 'Specify memory to use as stack size - use if you see: ''fatal runtime error: stack overflow'' (default low memory=1048576, high memory=1073741824)')
|
||||||
[CompletionResult]::new('--output-format', '--output-format', [CompletionResultType]::ParameterName, 'Changes output display size. si will print sizes in powers of 1000. b k m g t kb mb gb tb will print the whole tree in that size.')
|
[CompletionResult]::new('-p', 'p', [CompletionResultType]::ParameterName, 'Subdirectories will not have their path shortened')
|
||||||
[CompletionResult]::new('-S', '-S ', [CompletionResultType]::ParameterName, 'Specify memory to use as stack size - use if you see: ''fatal runtime error: stack overflow'' (default low memory=1048576, high memory=1073741824)')
|
[CompletionResult]::new('--full-paths', 'full-paths', [CompletionResultType]::ParameterName, 'Subdirectories will not have their path shortened')
|
||||||
[CompletionResult]::new('--stack-size', '--stack-size', [CompletionResultType]::ParameterName, 'Specify memory to use as stack size - use if you see: ''fatal runtime error: stack overflow'' (default low memory=1048576, high memory=1073741824)')
|
[CompletionResult]::new('-L', 'L ', [CompletionResultType]::ParameterName, 'dereference sym links - Treat sym links as directories and go into them')
|
||||||
[CompletionResult]::new('-M', '-M ', [CompletionResultType]::ParameterName, '+/-n matches files modified more/less than n days ago , and n matches files modified exactly n days ago, days are rounded down.That is +n => (−∞, curr−(n+1)), n => [curr−(n+1), curr−n), and -n => (𝑐𝑢𝑟𝑟−𝑛, +∞)')
|
[CompletionResult]::new('--dereference-links', 'dereference-links', [CompletionResultType]::ParameterName, 'dereference sym links - Treat sym links as directories and go into them')
|
||||||
[CompletionResult]::new('--mtime', '--mtime', [CompletionResultType]::ParameterName, '+/-n matches files modified more/less than n days ago , and n matches files modified exactly n days ago, days are rounded down.That is +n => (−∞, curr−(n+1)), n => [curr−(n+1), curr−n), and -n => (𝑐𝑢𝑟𝑟−𝑛, +∞)')
|
[CompletionResult]::new('-x', 'x', [CompletionResultType]::ParameterName, 'Only count the files and directories on the same filesystem as the supplied directory')
|
||||||
[CompletionResult]::new('-A', '-A ', [CompletionResultType]::ParameterName, 'just like -mtime, but based on file access time')
|
[CompletionResult]::new('--limit-filesystem', 'limit-filesystem', [CompletionResultType]::ParameterName, 'Only count the files and directories on the same filesystem as the supplied directory')
|
||||||
[CompletionResult]::new('--atime', '--atime', [CompletionResultType]::ParameterName, 'just like -mtime, but based on file access time')
|
[CompletionResult]::new('-s', 's', [CompletionResultType]::ParameterName, 'Use file length instead of blocks')
|
||||||
[CompletionResult]::new('-y', '-y', [CompletionResultType]::ParameterName, 'just like -mtime, but based on file change time')
|
[CompletionResult]::new('--apparent-size', 'apparent-size', [CompletionResultType]::ParameterName, 'Use file length instead of blocks')
|
||||||
[CompletionResult]::new('--ctime', '--ctime', [CompletionResultType]::ParameterName, 'just like -mtime, but based on file change time')
|
[CompletionResult]::new('-r', 'r', [CompletionResultType]::ParameterName, 'Print tree upside down (biggest highest)')
|
||||||
[CompletionResult]::new('--files0-from', '--files0-from', [CompletionResultType]::ParameterName, 'run dust on NUL-terminated file names specified in file; if argument is -, then read names from standard input')
|
[CompletionResult]::new('--reverse', 'reverse', [CompletionResultType]::ParameterName, 'Print tree upside down (biggest highest)')
|
||||||
[CompletionResult]::new('--collapse', '--collapse', [CompletionResultType]::ParameterName, 'Keep these directories collapsed')
|
[CompletionResult]::new('-c', 'c', [CompletionResultType]::ParameterName, 'No colors will be printed (Useful for commands like: watch)')
|
||||||
[CompletionResult]::new('-m', '-m', [CompletionResultType]::ParameterName, 'Directory ''size'' is max filetime of child files instead of disk size. while a/c/m for last accessed/changed/modified time')
|
[CompletionResult]::new('--no-colors', 'no-colors', [CompletionResultType]::ParameterName, 'No colors will be printed (Useful for commands like: watch)')
|
||||||
[CompletionResult]::new('--filetime', '--filetime', [CompletionResultType]::ParameterName, 'Directory ''size'' is max filetime of child files instead of disk size. while a/c/m for last accessed/changed/modified time')
|
[CompletionResult]::new('-C', 'C ', [CompletionResultType]::ParameterName, 'Force colors print')
|
||||||
[CompletionResult]::new('-p', '-p', [CompletionResultType]::ParameterName, 'Subdirectories will not have their path shortened')
|
[CompletionResult]::new('--force-colors', 'force-colors', [CompletionResultType]::ParameterName, 'Force colors print')
|
||||||
[CompletionResult]::new('--full-paths', '--full-paths', [CompletionResultType]::ParameterName, 'Subdirectories will not have their path shortened')
|
[CompletionResult]::new('-b', 'b', [CompletionResultType]::ParameterName, 'No percent bars or percentages will be displayed')
|
||||||
[CompletionResult]::new('-L', '-L ', [CompletionResultType]::ParameterName, 'dereference sym links - Treat sym links as directories and go into them')
|
[CompletionResult]::new('--no-percent-bars', 'no-percent-bars', [CompletionResultType]::ParameterName, 'No percent bars or percentages will be displayed')
|
||||||
[CompletionResult]::new('--dereference-links', '--dereference-links', [CompletionResultType]::ParameterName, 'dereference sym links - Treat sym links as directories and go into them')
|
[CompletionResult]::new('-B', 'B ', [CompletionResultType]::ParameterName, 'percent bars moved to right side of screen')
|
||||||
[CompletionResult]::new('-x', '-x', [CompletionResultType]::ParameterName, 'Only count the files and directories on the same filesystem as the supplied directory')
|
[CompletionResult]::new('--bars-on-right', 'bars-on-right', [CompletionResultType]::ParameterName, 'percent bars moved to right side of screen')
|
||||||
[CompletionResult]::new('--limit-filesystem', '--limit-filesystem', [CompletionResultType]::ParameterName, 'Only count the files and directories on the same filesystem as the supplied directory')
|
[CompletionResult]::new('-R', 'R ', [CompletionResultType]::ParameterName, 'For screen readers. Removes bars. Adds new column: depth level (May want to use -p too for full path)')
|
||||||
[CompletionResult]::new('-s', '-s', [CompletionResultType]::ParameterName, 'Use file length instead of blocks')
|
[CompletionResult]::new('--screen-reader', 'screen-reader', [CompletionResultType]::ParameterName, 'For screen readers. Removes bars. Adds new column: depth level (May want to use -p too for full path)')
|
||||||
[CompletionResult]::new('--apparent-size', '--apparent-size', [CompletionResultType]::ParameterName, 'Use file length instead of blocks')
|
[CompletionResult]::new('--skip-total', 'skip-total', [CompletionResultType]::ParameterName, 'No total row will be displayed')
|
||||||
[CompletionResult]::new('-r', '-r', [CompletionResultType]::ParameterName, 'Print tree upside down (biggest highest)')
|
[CompletionResult]::new('-f', 'f', [CompletionResultType]::ParameterName, 'Directory ''size'' is number of child files instead of disk size')
|
||||||
[CompletionResult]::new('--reverse', '--reverse', [CompletionResultType]::ParameterName, 'Print tree upside down (biggest highest)')
|
[CompletionResult]::new('--filecount', 'filecount', [CompletionResultType]::ParameterName, 'Directory ''size'' is number of child files instead of disk size')
|
||||||
[CompletionResult]::new('-c', '-c', [CompletionResultType]::ParameterName, 'No colors will be printed (Useful for commands like: watch)')
|
[CompletionResult]::new('-i', 'i', [CompletionResultType]::ParameterName, 'Do not display hidden files')
|
||||||
[CompletionResult]::new('--no-colors', '--no-colors', [CompletionResultType]::ParameterName, 'No colors will be printed (Useful for commands like: watch)')
|
[CompletionResult]::new('--ignore_hidden', 'ignore_hidden', [CompletionResultType]::ParameterName, 'Do not display hidden files')
|
||||||
[CompletionResult]::new('-C', '-C ', [CompletionResultType]::ParameterName, 'Force colors print')
|
[CompletionResult]::new('-t', 't', [CompletionResultType]::ParameterName, 'show only these file types')
|
||||||
[CompletionResult]::new('--force-colors', '--force-colors', [CompletionResultType]::ParameterName, 'Force colors print')
|
[CompletionResult]::new('--file_types', 'file_types', [CompletionResultType]::ParameterName, 'show only these file types')
|
||||||
[CompletionResult]::new('-b', '-b', [CompletionResultType]::ParameterName, 'No percent bars or percentages will be displayed')
|
[CompletionResult]::new('-P', 'P ', [CompletionResultType]::ParameterName, 'Disable the progress indication.')
|
||||||
[CompletionResult]::new('--no-percent-bars', '--no-percent-bars', [CompletionResultType]::ParameterName, 'No percent bars or percentages will be displayed')
|
[CompletionResult]::new('--no-progress', 'no-progress', [CompletionResultType]::ParameterName, 'Disable the progress indication.')
|
||||||
[CompletionResult]::new('-B', '-B ', [CompletionResultType]::ParameterName, 'percent bars moved to right side of screen')
|
[CompletionResult]::new('-D', 'D ', [CompletionResultType]::ParameterName, 'Only directories will be displayed.')
|
||||||
[CompletionResult]::new('--bars-on-right', '--bars-on-right', [CompletionResultType]::ParameterName, 'percent bars moved to right side of screen')
|
[CompletionResult]::new('--only-dir', 'only-dir', [CompletionResultType]::ParameterName, 'Only directories will be displayed.')
|
||||||
[CompletionResult]::new('-R', '-R ', [CompletionResultType]::ParameterName, 'For screen readers. Removes bars. Adds new column: depth level (May want to use -p too for full path)')
|
[CompletionResult]::new('-F', 'F ', [CompletionResultType]::ParameterName, 'Only files will be displayed. (Finds your largest files)')
|
||||||
[CompletionResult]::new('--screen-reader', '--screen-reader', [CompletionResultType]::ParameterName, 'For screen readers. Removes bars. Adds new column: depth level (May want to use -p too for full path)')
|
[CompletionResult]::new('--only-file', 'only-file', [CompletionResultType]::ParameterName, 'Only files will be displayed. (Finds your largest files)')
|
||||||
[CompletionResult]::new('--skip-total', '--skip-total', [CompletionResultType]::ParameterName, 'No total row will be displayed')
|
[CompletionResult]::new('-h', 'h', [CompletionResultType]::ParameterName, 'Print help')
|
||||||
[CompletionResult]::new('-f', '-f', [CompletionResultType]::ParameterName, 'Directory ''size'' is number of child files instead of disk size')
|
[CompletionResult]::new('--help', 'help', [CompletionResultType]::ParameterName, 'Print help')
|
||||||
[CompletionResult]::new('--filecount', '--filecount', [CompletionResultType]::ParameterName, 'Directory ''size'' is number of child files instead of disk size')
|
[CompletionResult]::new('-V', 'V ', [CompletionResultType]::ParameterName, 'Print version')
|
||||||
[CompletionResult]::new('-i', '-i', [CompletionResultType]::ParameterName, 'Do not display hidden files')
|
[CompletionResult]::new('--version', 'version', [CompletionResultType]::ParameterName, 'Print version')
|
||||||
[CompletionResult]::new('--ignore-hidden', '--ignore-hidden', [CompletionResultType]::ParameterName, 'Do not display hidden files')
|
|
||||||
[CompletionResult]::new('-t', '-t', [CompletionResultType]::ParameterName, 'show only these file types')
|
|
||||||
[CompletionResult]::new('--file-types', '--file-types', [CompletionResultType]::ParameterName, 'show only these file types')
|
|
||||||
[CompletionResult]::new('-P', '-P ', [CompletionResultType]::ParameterName, 'Disable the progress indication.')
|
|
||||||
[CompletionResult]::new('--no-progress', '--no-progress', [CompletionResultType]::ParameterName, 'Disable the progress indication.')
|
|
||||||
[CompletionResult]::new('--print-errors', '--print-errors', [CompletionResultType]::ParameterName, 'Print path with errors.')
|
|
||||||
[CompletionResult]::new('-D', '-D ', [CompletionResultType]::ParameterName, 'Only directories will be displayed.')
|
|
||||||
[CompletionResult]::new('--only-dir', '--only-dir', [CompletionResultType]::ParameterName, 'Only directories will be displayed.')
|
|
||||||
[CompletionResult]::new('-F', '-F ', [CompletionResultType]::ParameterName, 'Only files will be displayed. (Finds your largest files)')
|
|
||||||
[CompletionResult]::new('--only-file', '--only-file', [CompletionResultType]::ParameterName, 'Only files will be displayed. (Finds your largest files)')
|
|
||||||
[CompletionResult]::new('-j', '-j', [CompletionResultType]::ParameterName, 'Output the directory tree as json to the current directory')
|
|
||||||
[CompletionResult]::new('--output-json', '--output-json', [CompletionResultType]::ParameterName, 'Output the directory tree as json to the current directory')
|
|
||||||
[CompletionResult]::new('-h', '-h', [CompletionResultType]::ParameterName, 'Print help')
|
|
||||||
[CompletionResult]::new('--help', '--help', [CompletionResultType]::ParameterName, 'Print help')
|
|
||||||
[CompletionResult]::new('-V', '-V ', [CompletionResultType]::ParameterName, 'Print version')
|
|
||||||
[CompletionResult]::new('--version', '--version', [CompletionResultType]::ParameterName, 'Print version')
|
|
||||||
break
|
break
|
||||||
}
|
}
|
||||||
})
|
})
|
||||||
|
|||||||
@@ -19,7 +19,7 @@ _dust() {
|
|||||||
|
|
||||||
case "${cmd}" in
|
case "${cmd}" in
|
||||||
dust)
|
dust)
|
||||||
opts="-d -T -n -p -X -I -L -x -s -r -c -C -b -B -z -R -f -i -v -e -t -w -P -D -F -o -S -j -M -A -y -m -h -V --depth --threads --config --number-of-lines --full-paths --ignore-directory --ignore-all-in-file --dereference-links --limit-filesystem --apparent-size --reverse --no-colors --force-colors --no-percent-bars --bars-on-right --min-size --screen-reader --skip-total --filecount --ignore-hidden --invert-filter --filter --file-types --terminal-width --no-progress --print-errors --only-dir --only-file --output-format --stack-size --output-json --mtime --atime --ctime --files0-from --collapse --filetime --help --version [PATH]..."
|
opts="-d -n -p -X -I -L -x -s -r -c -C -b -B -z -R -f -i -v -e -t -w -P -D -F -o -S -h -V --depth --number-of-lines --full-paths --ignore-directory --ignore-all-in-file --dereference-links --limit-filesystem --apparent-size --reverse --no-colors --force-colors --no-percent-bars --bars-on-right --min-size --screen-reader --skip-total --filecount --ignore_hidden --invert-filter --filter --file_types --terminal_width --no-progress --only-dir --only-file --output-format --stack-size --help --version [params]..."
|
||||||
if [[ ${cur} == -* || ${COMP_CWORD} -eq 1 ]] ; then
|
if [[ ${cur} == -* || ${COMP_CWORD} -eq 1 ]] ; then
|
||||||
COMPREPLY=( $(compgen -W "${opts}" -- "${cur}") )
|
COMPREPLY=( $(compgen -W "${opts}" -- "${cur}") )
|
||||||
return 0
|
return 0
|
||||||
@@ -33,29 +33,6 @@ _dust() {
|
|||||||
COMPREPLY=($(compgen -f "${cur}"))
|
COMPREPLY=($(compgen -f "${cur}"))
|
||||||
return 0
|
return 0
|
||||||
;;
|
;;
|
||||||
--threads)
|
|
||||||
COMPREPLY=($(compgen -f "${cur}"))
|
|
||||||
return 0
|
|
||||||
;;
|
|
||||||
-T)
|
|
||||||
COMPREPLY=($(compgen -f "${cur}"))
|
|
||||||
return 0
|
|
||||||
;;
|
|
||||||
--config)
|
|
||||||
local oldifs
|
|
||||||
if [ -n "${IFS+x}" ]; then
|
|
||||||
oldifs="$IFS"
|
|
||||||
fi
|
|
||||||
IFS=$'\n'
|
|
||||||
COMPREPLY=($(compgen -f "${cur}"))
|
|
||||||
if [ -n "${oldifs+x}" ]; then
|
|
||||||
IFS="$oldifs"
|
|
||||||
fi
|
|
||||||
if [[ "${BASH_VERSINFO[0]}" -ge 4 ]]; then
|
|
||||||
compopt -o filenames
|
|
||||||
fi
|
|
||||||
return 0
|
|
||||||
;;
|
|
||||||
--number-of-lines)
|
--number-of-lines)
|
||||||
COMPREPLY=($(compgen -f "${cur}"))
|
COMPREPLY=($(compgen -f "${cur}"))
|
||||||
return 0
|
return 0
|
||||||
@@ -73,33 +50,11 @@ _dust() {
|
|||||||
return 0
|
return 0
|
||||||
;;
|
;;
|
||||||
--ignore-all-in-file)
|
--ignore-all-in-file)
|
||||||
local oldifs
|
|
||||||
if [ -n "${IFS+x}" ]; then
|
|
||||||
oldifs="$IFS"
|
|
||||||
fi
|
|
||||||
IFS=$'\n'
|
|
||||||
COMPREPLY=($(compgen -f "${cur}"))
|
COMPREPLY=($(compgen -f "${cur}"))
|
||||||
if [ -n "${oldifs+x}" ]; then
|
|
||||||
IFS="$oldifs"
|
|
||||||
fi
|
|
||||||
if [[ "${BASH_VERSINFO[0]}" -ge 4 ]]; then
|
|
||||||
compopt -o filenames
|
|
||||||
fi
|
|
||||||
return 0
|
return 0
|
||||||
;;
|
;;
|
||||||
-I)
|
-I)
|
||||||
local oldifs
|
|
||||||
if [ -n "${IFS+x}" ]; then
|
|
||||||
oldifs="$IFS"
|
|
||||||
fi
|
|
||||||
IFS=$'\n'
|
|
||||||
COMPREPLY=($(compgen -f "${cur}"))
|
COMPREPLY=($(compgen -f "${cur}"))
|
||||||
if [ -n "${oldifs+x}" ]; then
|
|
||||||
IFS="$oldifs"
|
|
||||||
fi
|
|
||||||
if [[ "${BASH_VERSINFO[0]}" -ge 4 ]]; then
|
|
||||||
compopt -o filenames
|
|
||||||
fi
|
|
||||||
return 0
|
return 0
|
||||||
;;
|
;;
|
||||||
--min-size)
|
--min-size)
|
||||||
@@ -126,7 +81,7 @@ _dust() {
|
|||||||
COMPREPLY=($(compgen -f "${cur}"))
|
COMPREPLY=($(compgen -f "${cur}"))
|
||||||
return 0
|
return 0
|
||||||
;;
|
;;
|
||||||
--terminal-width)
|
--terminal_width)
|
||||||
COMPREPLY=($(compgen -f "${cur}"))
|
COMPREPLY=($(compgen -f "${cur}"))
|
||||||
return 0
|
return 0
|
||||||
;;
|
;;
|
||||||
@@ -135,11 +90,11 @@ _dust() {
|
|||||||
return 0
|
return 0
|
||||||
;;
|
;;
|
||||||
--output-format)
|
--output-format)
|
||||||
COMPREPLY=($(compgen -W "si b k m g t kb mb gb tb" -- "${cur}"))
|
COMPREPLY=($(compgen -f "${cur}"))
|
||||||
return 0
|
return 0
|
||||||
;;
|
;;
|
||||||
-o)
|
-o)
|
||||||
COMPREPLY=($(compgen -W "si b k m g t kb mb gb tb" -- "${cur}"))
|
COMPREPLY=($(compgen -f "${cur}"))
|
||||||
return 0
|
return 0
|
||||||
;;
|
;;
|
||||||
--stack-size)
|
--stack-size)
|
||||||
@@ -150,46 +105,6 @@ _dust() {
|
|||||||
COMPREPLY=($(compgen -f "${cur}"))
|
COMPREPLY=($(compgen -f "${cur}"))
|
||||||
return 0
|
return 0
|
||||||
;;
|
;;
|
||||||
--mtime)
|
|
||||||
COMPREPLY=($(compgen -f "${cur}"))
|
|
||||||
return 0
|
|
||||||
;;
|
|
||||||
-M)
|
|
||||||
COMPREPLY=($(compgen -f "${cur}"))
|
|
||||||
return 0
|
|
||||||
;;
|
|
||||||
--atime)
|
|
||||||
COMPREPLY=($(compgen -f "${cur}"))
|
|
||||||
return 0
|
|
||||||
;;
|
|
||||||
-A)
|
|
||||||
COMPREPLY=($(compgen -f "${cur}"))
|
|
||||||
return 0
|
|
||||||
;;
|
|
||||||
--ctime)
|
|
||||||
COMPREPLY=($(compgen -f "${cur}"))
|
|
||||||
return 0
|
|
||||||
;;
|
|
||||||
-y)
|
|
||||||
COMPREPLY=($(compgen -f "${cur}"))
|
|
||||||
return 0
|
|
||||||
;;
|
|
||||||
--files0-from)
|
|
||||||
COMPREPLY=($(compgen -f "${cur}"))
|
|
||||||
return 0
|
|
||||||
;;
|
|
||||||
--collapse)
|
|
||||||
COMPREPLY=($(compgen -f "${cur}"))
|
|
||||||
return 0
|
|
||||||
;;
|
|
||||||
--filetime)
|
|
||||||
COMPREPLY=($(compgen -W "a c m" -- "${cur}"))
|
|
||||||
return 0
|
|
||||||
;;
|
|
||||||
-m)
|
|
||||||
COMPREPLY=($(compgen -W "a c m" -- "${cur}"))
|
|
||||||
return 0
|
|
||||||
;;
|
|
||||||
*)
|
*)
|
||||||
COMPREPLY=()
|
COMPREPLY=()
|
||||||
;;
|
;;
|
||||||
|
|||||||
@@ -20,13 +20,10 @@ set edit:completion:arg-completer[dust] = {|@words|
|
|||||||
&'dust'= {
|
&'dust'= {
|
||||||
cand -d 'Depth to show'
|
cand -d 'Depth to show'
|
||||||
cand --depth 'Depth to show'
|
cand --depth 'Depth to show'
|
||||||
cand -T 'Number of threads to use'
|
|
||||||
cand --threads 'Number of threads to use'
|
|
||||||
cand --config 'Specify a config file to use'
|
|
||||||
cand -n 'Number of lines of output to show. (Default is terminal_height - 10)'
|
cand -n 'Number of lines of output to show. (Default is terminal_height - 10)'
|
||||||
cand --number-of-lines 'Number of lines of output to show. (Default is terminal_height - 10)'
|
cand --number-of-lines 'Number of lines of output to show. (Default is terminal_height - 10)'
|
||||||
cand -X 'Exclude any file or directory with this path'
|
cand -X 'Exclude any file or directory with this name'
|
||||||
cand --ignore-directory 'Exclude any file or directory with this path'
|
cand --ignore-directory 'Exclude any file or directory with this name'
|
||||||
cand -I 'Exclude any file or directory with a regex matching that listed in this file, the file entries will be added to the ignore regexs provided by --invert_filter'
|
cand -I 'Exclude any file or directory with a regex matching that listed in this file, the file entries will be added to the ignore regexs provided by --invert_filter'
|
||||||
cand --ignore-all-in-file 'Exclude any file or directory with a regex matching that listed in this file, the file entries will be added to the ignore regexs provided by --invert_filter'
|
cand --ignore-all-in-file 'Exclude any file or directory with a regex matching that listed in this file, the file entries will be added to the ignore regexs provided by --invert_filter'
|
||||||
cand -z 'Minimum size file to include in output'
|
cand -z 'Minimum size file to include in output'
|
||||||
@@ -36,21 +33,11 @@ set edit:completion:arg-completer[dust] = {|@words|
|
|||||||
cand -e 'Only include filepaths matching this regex. For png files type: -e "\.png$" '
|
cand -e 'Only include filepaths matching this regex. For png files type: -e "\.png$" '
|
||||||
cand --filter 'Only include filepaths matching this regex. For png files type: -e "\.png$" '
|
cand --filter 'Only include filepaths matching this regex. For png files type: -e "\.png$" '
|
||||||
cand -w 'Specify width of output overriding the auto detection of terminal width'
|
cand -w 'Specify width of output overriding the auto detection of terminal width'
|
||||||
cand --terminal-width 'Specify width of output overriding the auto detection of terminal width'
|
cand --terminal_width 'Specify width of output overriding the auto detection of terminal width'
|
||||||
cand -o 'Changes output display size. si will print sizes in powers of 1000. b k m g t kb mb gb tb will print the whole tree in that size.'
|
cand -o 'Changes output display size. si will print sizes in powers of 1000. b/bytes kb kib mb mib gb gib will print the whole tree in that size'
|
||||||
cand --output-format 'Changes output display size. si will print sizes in powers of 1000. b k m g t kb mb gb tb will print the whole tree in that size.'
|
cand --output-format 'Changes output display size. si will print sizes in powers of 1000. b/bytes kb kib mb mib gb gib will print the whole tree in that size'
|
||||||
cand -S 'Specify memory to use as stack size - use if you see: ''fatal runtime error: stack overflow'' (default low memory=1048576, high memory=1073741824)'
|
cand -S 'Specify memory to use as stack size - use if you see: ''fatal runtime error: stack overflow'' (default low memory=1048576, high memory=1073741824)'
|
||||||
cand --stack-size 'Specify memory to use as stack size - use if you see: ''fatal runtime error: stack overflow'' (default low memory=1048576, high memory=1073741824)'
|
cand --stack-size 'Specify memory to use as stack size - use if you see: ''fatal runtime error: stack overflow'' (default low memory=1048576, high memory=1073741824)'
|
||||||
cand -M '+/-n matches files modified more/less than n days ago , and n matches files modified exactly n days ago, days are rounded down.That is +n => (−∞, curr−(n+1)), n => [curr−(n+1), curr−n), and -n => (𝑐𝑢𝑟𝑟−𝑛, +∞)'
|
|
||||||
cand --mtime '+/-n matches files modified more/less than n days ago , and n matches files modified exactly n days ago, days are rounded down.That is +n => (−∞, curr−(n+1)), n => [curr−(n+1), curr−n), and -n => (𝑐𝑢𝑟𝑟−𝑛, +∞)'
|
|
||||||
cand -A 'just like -mtime, but based on file access time'
|
|
||||||
cand --atime 'just like -mtime, but based on file access time'
|
|
||||||
cand -y 'just like -mtime, but based on file change time'
|
|
||||||
cand --ctime 'just like -mtime, but based on file change time'
|
|
||||||
cand --files0-from 'run dust on NUL-terminated file names specified in file; if argument is -, then read names from standard input'
|
|
||||||
cand --collapse 'Keep these directories collapsed'
|
|
||||||
cand -m 'Directory ''size'' is max filetime of child files instead of disk size. while a/c/m for last accessed/changed/modified time'
|
|
||||||
cand --filetime 'Directory ''size'' is max filetime of child files instead of disk size. while a/c/m for last accessed/changed/modified time'
|
|
||||||
cand -p 'Subdirectories will not have their path shortened'
|
cand -p 'Subdirectories will not have their path shortened'
|
||||||
cand --full-paths 'Subdirectories will not have their path shortened'
|
cand --full-paths 'Subdirectories will not have their path shortened'
|
||||||
cand -L 'dereference sym links - Treat sym links as directories and go into them'
|
cand -L 'dereference sym links - Treat sym links as directories and go into them'
|
||||||
@@ -75,18 +62,15 @@ set edit:completion:arg-completer[dust] = {|@words|
|
|||||||
cand -f 'Directory ''size'' is number of child files instead of disk size'
|
cand -f 'Directory ''size'' is number of child files instead of disk size'
|
||||||
cand --filecount 'Directory ''size'' is number of child files instead of disk size'
|
cand --filecount 'Directory ''size'' is number of child files instead of disk size'
|
||||||
cand -i 'Do not display hidden files'
|
cand -i 'Do not display hidden files'
|
||||||
cand --ignore-hidden 'Do not display hidden files'
|
cand --ignore_hidden 'Do not display hidden files'
|
||||||
cand -t 'show only these file types'
|
cand -t 'show only these file types'
|
||||||
cand --file-types 'show only these file types'
|
cand --file_types 'show only these file types'
|
||||||
cand -P 'Disable the progress indication.'
|
cand -P 'Disable the progress indication.'
|
||||||
cand --no-progress 'Disable the progress indication.'
|
cand --no-progress 'Disable the progress indication.'
|
||||||
cand --print-errors 'Print path with errors.'
|
|
||||||
cand -D 'Only directories will be displayed.'
|
cand -D 'Only directories will be displayed.'
|
||||||
cand --only-dir 'Only directories will be displayed.'
|
cand --only-dir 'Only directories will be displayed.'
|
||||||
cand -F 'Only files will be displayed. (Finds your largest files)'
|
cand -F 'Only files will be displayed. (Finds your largest files)'
|
||||||
cand --only-file 'Only files will be displayed. (Finds your largest files)'
|
cand --only-file 'Only files will be displayed. (Finds your largest files)'
|
||||||
cand -j 'Output the directory tree as json to the current directory'
|
|
||||||
cand --output-json 'Output the directory tree as json to the current directory'
|
|
||||||
cand -h 'Print help'
|
cand -h 'Print help'
|
||||||
cand --help 'Print help'
|
cand --help 'Print help'
|
||||||
cand -V 'Print version'
|
cand -V 'Print version'
|
||||||
|
|||||||
@@ -1,32 +1,13 @@
|
|||||||
complete -c dust -s d -l depth -d 'Depth to show' -r
|
complete -c dust -s d -l depth -d 'Depth to show' -r
|
||||||
complete -c dust -s T -l threads -d 'Number of threads to use' -r
|
|
||||||
complete -c dust -l config -d 'Specify a config file to use' -r -F
|
|
||||||
complete -c dust -s n -l number-of-lines -d 'Number of lines of output to show. (Default is terminal_height - 10)' -r
|
complete -c dust -s n -l number-of-lines -d 'Number of lines of output to show. (Default is terminal_height - 10)' -r
|
||||||
complete -c dust -s X -l ignore-directory -d 'Exclude any file or directory with this path' -r -F
|
complete -c dust -s X -l ignore-directory -d 'Exclude any file or directory with this name' -r
|
||||||
complete -c dust -s I -l ignore-all-in-file -d 'Exclude any file or directory with a regex matching that listed in this file, the file entries will be added to the ignore regexs provided by --invert_filter' -r -F
|
complete -c dust -s I -l ignore-all-in-file -d 'Exclude any file or directory with a regex matching that listed in this file, the file entries will be added to the ignore regexs provided by --invert_filter' -r
|
||||||
complete -c dust -s z -l min-size -d 'Minimum size file to include in output' -r
|
complete -c dust -s z -l min-size -d 'Minimum size file to include in output' -r
|
||||||
complete -c dust -s v -l invert-filter -d 'Exclude filepaths matching this regex. To ignore png files type: -v "\\.png$" ' -r
|
complete -c dust -s v -l invert-filter -d 'Exclude filepaths matching this regex. To ignore png files type: -v "\\.png$" ' -r
|
||||||
complete -c dust -s e -l filter -d 'Only include filepaths matching this regex. For png files type: -e "\\.png$" ' -r
|
complete -c dust -s e -l filter -d 'Only include filepaths matching this regex. For png files type: -e "\\.png$" ' -r
|
||||||
complete -c dust -s w -l terminal-width -d 'Specify width of output overriding the auto detection of terminal width' -r
|
complete -c dust -s w -l terminal_width -d 'Specify width of output overriding the auto detection of terminal width' -r
|
||||||
complete -c dust -s o -l output-format -d 'Changes output display size. si will print sizes in powers of 1000. b k m g t kb mb gb tb will print the whole tree in that size.' -r -f -a "si\t''
|
complete -c dust -s o -l output-format -d 'Changes output display size. si will print sizes in powers of 1000. b/bytes kb kib mb mib gb gib will print the whole tree in that size' -r
|
||||||
b\t''
|
|
||||||
k\t''
|
|
||||||
m\t''
|
|
||||||
g\t''
|
|
||||||
t\t''
|
|
||||||
kb\t''
|
|
||||||
mb\t''
|
|
||||||
gb\t''
|
|
||||||
tb\t''"
|
|
||||||
complete -c dust -s S -l stack-size -d 'Specify memory to use as stack size - use if you see: \'fatal runtime error: stack overflow\' (default low memory=1048576, high memory=1073741824)' -r
|
complete -c dust -s S -l stack-size -d 'Specify memory to use as stack size - use if you see: \'fatal runtime error: stack overflow\' (default low memory=1048576, high memory=1073741824)' -r
|
||||||
complete -c dust -s M -l mtime -d '+/-n matches files modified more/less than n days ago , and n matches files modified exactly n days ago, days are rounded down.That is +n => (−∞, curr−(n+1)), n => [curr−(n+1), curr−n), and -n => (𝑐𝑢𝑟𝑟−𝑛, +∞)' -r
|
|
||||||
complete -c dust -s A -l atime -d 'just like -mtime, but based on file access time' -r
|
|
||||||
complete -c dust -s y -l ctime -d 'just like -mtime, but based on file change time' -r
|
|
||||||
complete -c dust -l files0-from -d 'run dust on NUL-terminated file names specified in file; if argument is -, then read names from standard input' -r -F
|
|
||||||
complete -c dust -l collapse -d 'Keep these directories collapsed' -r -F
|
|
||||||
complete -c dust -s m -l filetime -d 'Directory \'size\' is max filetime of child files instead of disk size. while a/c/m for last accessed/changed/modified time' -r -f -a "a\t''
|
|
||||||
c\t''
|
|
||||||
m\t''"
|
|
||||||
complete -c dust -s p -l full-paths -d 'Subdirectories will not have their path shortened'
|
complete -c dust -s p -l full-paths -d 'Subdirectories will not have their path shortened'
|
||||||
complete -c dust -s L -l dereference-links -d 'dereference sym links - Treat sym links as directories and go into them'
|
complete -c dust -s L -l dereference-links -d 'dereference sym links - Treat sym links as directories and go into them'
|
||||||
complete -c dust -s x -l limit-filesystem -d 'Only count the files and directories on the same filesystem as the supplied directory'
|
complete -c dust -s x -l limit-filesystem -d 'Only count the files and directories on the same filesystem as the supplied directory'
|
||||||
@@ -39,12 +20,10 @@ complete -c dust -s B -l bars-on-right -d 'percent bars moved to right side of s
|
|||||||
complete -c dust -s R -l screen-reader -d 'For screen readers. Removes bars. Adds new column: depth level (May want to use -p too for full path)'
|
complete -c dust -s R -l screen-reader -d 'For screen readers. Removes bars. Adds new column: depth level (May want to use -p too for full path)'
|
||||||
complete -c dust -l skip-total -d 'No total row will be displayed'
|
complete -c dust -l skip-total -d 'No total row will be displayed'
|
||||||
complete -c dust -s f -l filecount -d 'Directory \'size\' is number of child files instead of disk size'
|
complete -c dust -s f -l filecount -d 'Directory \'size\' is number of child files instead of disk size'
|
||||||
complete -c dust -s i -l ignore-hidden -d 'Do not display hidden files'
|
complete -c dust -s i -l ignore_hidden -d 'Do not display hidden files'
|
||||||
complete -c dust -s t -l file-types -d 'show only these file types'
|
complete -c dust -s t -l file_types -d 'show only these file types'
|
||||||
complete -c dust -s P -l no-progress -d 'Disable the progress indication.'
|
complete -c dust -s P -l no-progress -d 'Disable the progress indication.'
|
||||||
complete -c dust -l print-errors -d 'Print path with errors.'
|
|
||||||
complete -c dust -s D -l only-dir -d 'Only directories will be displayed.'
|
complete -c dust -s D -l only-dir -d 'Only directories will be displayed.'
|
||||||
complete -c dust -s F -l only-file -d 'Only files will be displayed. (Finds your largest files)'
|
complete -c dust -s F -l only-file -d 'Only files will be displayed. (Finds your largest files)'
|
||||||
complete -c dust -s j -l output-json -d 'Output the directory tree as json to the current directory'
|
|
||||||
complete -c dust -s h -l help -d 'Print help'
|
complete -c dust -s h -l help -d 'Print help'
|
||||||
complete -c dust -s V -l version -d 'Print version'
|
complete -c dust -s V -l version -d 'Print version'
|
||||||
|
|||||||
@@ -25,4 +25,4 @@ skip-total=true
|
|||||||
ignore-hidden=true
|
ignore-hidden=true
|
||||||
|
|
||||||
# print sizes in powers of 1000 (e.g., 1.1G)
|
# print sizes in powers of 1000 (e.g., 1.1G)
|
||||||
output-format="si"
|
iso=true
|
||||||
|
|||||||
@@ -1,33 +1,27 @@
|
|||||||
.ie \n(.g .ds Aq \(aq
|
.ie \n(.g .ds Aq \(aq
|
||||||
.el .ds Aq '
|
.el .ds Aq '
|
||||||
.TH Dust 1 "Dust 1.2.0"
|
.TH Dust 1 "Dust 1.0.0"
|
||||||
.SH NAME
|
.SH NAME
|
||||||
Dust \- Like du but more intuitive
|
Dust \- Like du but more intuitive
|
||||||
.SH SYNOPSIS
|
.SH SYNOPSIS
|
||||||
\fBdust\fR [\fB\-d\fR|\fB\-\-depth\fR] [\fB\-T\fR|\fB\-\-threads\fR] [\fB\-\-config\fR] [\fB\-n\fR|\fB\-\-number\-of\-lines\fR] [\fB\-p\fR|\fB\-\-full\-paths\fR] [\fB\-X\fR|\fB\-\-ignore\-directory\fR] [\fB\-I\fR|\fB\-\-ignore\-all\-in\-file\fR] [\fB\-L\fR|\fB\-\-dereference\-links\fR] [\fB\-x\fR|\fB\-\-limit\-filesystem\fR] [\fB\-s\fR|\fB\-\-apparent\-size\fR] [\fB\-r\fR|\fB\-\-reverse\fR] [\fB\-c\fR|\fB\-\-no\-colors\fR] [\fB\-C\fR|\fB\-\-force\-colors\fR] [\fB\-b\fR|\fB\-\-no\-percent\-bars\fR] [\fB\-B\fR|\fB\-\-bars\-on\-right\fR] [\fB\-z\fR|\fB\-\-min\-size\fR] [\fB\-R\fR|\fB\-\-screen\-reader\fR] [\fB\-\-skip\-total\fR] [\fB\-f\fR|\fB\-\-filecount\fR] [\fB\-i\fR|\fB\-\-ignore\-hidden\fR] [\fB\-v\fR|\fB\-\-invert\-filter\fR] [\fB\-e\fR|\fB\-\-filter\fR] [\fB\-t\fR|\fB\-\-file\-types\fR] [\fB\-w\fR|\fB\-\-terminal\-width\fR] [\fB\-P\fR|\fB\-\-no\-progress\fR] [\fB\-\-print\-errors\fR] [\fB\-D\fR|\fB\-\-only\-dir\fR] [\fB\-F\fR|\fB\-\-only\-file\fR] [\fB\-o\fR|\fB\-\-output\-format\fR] [\fB\-S\fR|\fB\-\-stack\-size\fR] [\fB\-j\fR|\fB\-\-output\-json\fR] [\fB\-M\fR|\fB\-\-mtime\fR] [\fB\-A\fR|\fB\-\-atime\fR] [\fB\-y\fR|\fB\-\-ctime\fR] [\fB\-\-files0\-from\fR] [\fB\-\-collapse\fR] [\fB\-m\fR|\fB\-\-filetime\fR] [\fB\-h\fR|\fB\-\-help\fR] [\fB\-V\fR|\fB\-\-version\fR] [\fIPATH\fR]
|
\fBdust\fR [\fB\-d\fR|\fB\-\-depth\fR] [\fB\-n\fR|\fB\-\-number\-of\-lines\fR] [\fB\-p\fR|\fB\-\-full\-paths\fR] [\fB\-X\fR|\fB\-\-ignore\-directory\fR] [\fB\-I\fR|\fB\-\-ignore\-all\-in\-file\fR] [\fB\-L\fR|\fB\-\-dereference\-links\fR] [\fB\-x\fR|\fB\-\-limit\-filesystem\fR] [\fB\-s\fR|\fB\-\-apparent\-size\fR] [\fB\-r\fR|\fB\-\-reverse\fR] [\fB\-c\fR|\fB\-\-no\-colors\fR] [\fB\-C\fR|\fB\-\-force\-colors\fR] [\fB\-b\fR|\fB\-\-no\-percent\-bars\fR] [\fB\-B\fR|\fB\-\-bars\-on\-right\fR] [\fB\-z\fR|\fB\-\-min\-size\fR] [\fB\-R\fR|\fB\-\-screen\-reader\fR] [\fB\-\-skip\-total\fR] [\fB\-f\fR|\fB\-\-filecount\fR] [\fB\-i\fR|\fB\-\-ignore_hidden\fR] [\fB\-v\fR|\fB\-\-invert\-filter\fR] [\fB\-e\fR|\fB\-\-filter\fR] [\fB\-t\fR|\fB\-\-file_types\fR] [\fB\-w\fR|\fB\-\-terminal_width\fR] [\fB\-P\fR|\fB\-\-no\-progress\fR] [\fB\-D\fR|\fB\-\-only\-dir\fR] [\fB\-F\fR|\fB\-\-only\-file\fR] [\fB\-o\fR|\fB\-\-output\-format\fR] [\fB\-S\fR|\fB\-\-stack\-size\fR] [\fB\-h\fR|\fB\-\-help\fR] [\fB\-V\fR|\fB\-\-version\fR] [\fIparams\fR]
|
||||||
.SH DESCRIPTION
|
.SH DESCRIPTION
|
||||||
Like du but more intuitive
|
Like du but more intuitive
|
||||||
.SH OPTIONS
|
.SH OPTIONS
|
||||||
.TP
|
.TP
|
||||||
\fB\-d\fR, \fB\-\-depth\fR=\fIDEPTH\fR
|
\fB\-d\fR, \fB\-\-depth\fR
|
||||||
Depth to show
|
Depth to show
|
||||||
.TP
|
.TP
|
||||||
\fB\-T\fR, \fB\-\-threads\fR
|
\fB\-n\fR, \fB\-\-number\-of\-lines\fR
|
||||||
Number of threads to use
|
|
||||||
.TP
|
|
||||||
\fB\-\-config\fR=\fIFILE\fR
|
|
||||||
Specify a config file to use
|
|
||||||
.TP
|
|
||||||
\fB\-n\fR, \fB\-\-number\-of\-lines\fR=\fINUMBER\fR
|
|
||||||
Number of lines of output to show. (Default is terminal_height \- 10)
|
Number of lines of output to show. (Default is terminal_height \- 10)
|
||||||
.TP
|
.TP
|
||||||
\fB\-p\fR, \fB\-\-full\-paths\fR
|
\fB\-p\fR, \fB\-\-full\-paths\fR
|
||||||
Subdirectories will not have their path shortened
|
Subdirectories will not have their path shortened
|
||||||
.TP
|
.TP
|
||||||
\fB\-X\fR, \fB\-\-ignore\-directory\fR=\fIPATH\fR
|
\fB\-X\fR, \fB\-\-ignore\-directory\fR
|
||||||
Exclude any file or directory with this path
|
Exclude any file or directory with this name
|
||||||
.TP
|
.TP
|
||||||
\fB\-I\fR, \fB\-\-ignore\-all\-in\-file\fR=\fIFILE\fR
|
\fB\-I\fR, \fB\-\-ignore\-all\-in\-file\fR
|
||||||
Exclude any file or directory with a regex matching that listed in this file, the file entries will be added to the ignore regexs provided by \-\-invert_filter
|
Exclude any file or directory with a regex matching that listed in this file, the file entries will be added to the ignore regexs provided by \-\-invert_filter
|
||||||
.TP
|
.TP
|
||||||
\fB\-L\fR, \fB\-\-dereference\-links\fR
|
\fB\-L\fR, \fB\-\-dereference\-links\fR
|
||||||
@@ -54,7 +48,7 @@ No percent bars or percentages will be displayed
|
|||||||
\fB\-B\fR, \fB\-\-bars\-on\-right\fR
|
\fB\-B\fR, \fB\-\-bars\-on\-right\fR
|
||||||
percent bars moved to right side of screen
|
percent bars moved to right side of screen
|
||||||
.TP
|
.TP
|
||||||
\fB\-z\fR, \fB\-\-min\-size\fR=\fIMIN_SIZE\fR
|
\fB\-z\fR, \fB\-\-min\-size\fR
|
||||||
Minimum size file to include in output
|
Minimum size file to include in output
|
||||||
.TP
|
.TP
|
||||||
\fB\-R\fR, \fB\-\-screen\-reader\fR
|
\fB\-R\fR, \fB\-\-screen\-reader\fR
|
||||||
@@ -66,75 +60,43 @@ No total row will be displayed
|
|||||||
\fB\-f\fR, \fB\-\-filecount\fR
|
\fB\-f\fR, \fB\-\-filecount\fR
|
||||||
Directory \*(Aqsize\*(Aq is number of child files instead of disk size
|
Directory \*(Aqsize\*(Aq is number of child files instead of disk size
|
||||||
.TP
|
.TP
|
||||||
\fB\-i\fR, \fB\-\-ignore\-hidden\fR
|
\fB\-i\fR, \fB\-\-ignore_hidden\fR
|
||||||
Do not display hidden files
|
Do not display hidden files
|
||||||
.TP
|
.TP
|
||||||
\fB\-v\fR, \fB\-\-invert\-filter\fR=\fIREGEX\fR
|
\fB\-v\fR, \fB\-\-invert\-filter\fR
|
||||||
Exclude filepaths matching this regex. To ignore png files type: \-v "\\.png$"
|
Exclude filepaths matching this regex. To ignore png files type: \-v "\\.png$"
|
||||||
.TP
|
.TP
|
||||||
\fB\-e\fR, \fB\-\-filter\fR=\fIREGEX\fR
|
\fB\-e\fR, \fB\-\-filter\fR
|
||||||
Only include filepaths matching this regex. For png files type: \-e "\\.png$"
|
Only include filepaths matching this regex. For png files type: \-e "\\.png$"
|
||||||
.TP
|
.TP
|
||||||
\fB\-t\fR, \fB\-\-file\-types\fR
|
\fB\-t\fR, \fB\-\-file_types\fR
|
||||||
show only these file types
|
show only these file types
|
||||||
.TP
|
.TP
|
||||||
\fB\-w\fR, \fB\-\-terminal\-width\fR=\fIWIDTH\fR
|
\fB\-w\fR, \fB\-\-terminal_width\fR
|
||||||
Specify width of output overriding the auto detection of terminal width
|
Specify width of output overriding the auto detection of terminal width
|
||||||
.TP
|
.TP
|
||||||
\fB\-P\fR, \fB\-\-no\-progress\fR
|
\fB\-P\fR, \fB\-\-no\-progress\fR
|
||||||
Disable the progress indication.
|
Disable the progress indication.
|
||||||
.TP
|
.TP
|
||||||
\fB\-\-print\-errors\fR
|
|
||||||
Print path with errors.
|
|
||||||
.TP
|
|
||||||
\fB\-D\fR, \fB\-\-only\-dir\fR
|
\fB\-D\fR, \fB\-\-only\-dir\fR
|
||||||
Only directories will be displayed.
|
Only directories will be displayed.
|
||||||
.TP
|
.TP
|
||||||
\fB\-F\fR, \fB\-\-only\-file\fR
|
\fB\-F\fR, \fB\-\-only\-file\fR
|
||||||
Only files will be displayed. (Finds your largest files)
|
Only files will be displayed. (Finds your largest files)
|
||||||
.TP
|
.TP
|
||||||
\fB\-o\fR, \fB\-\-output\-format\fR=\fIFORMAT\fR
|
\fB\-o\fR, \fB\-\-output\-format\fR
|
||||||
Changes output display size. si will print sizes in powers of 1000. b k m g t kb mb gb tb will print the whole tree in that size.
|
Changes output display size. si will print sizes in powers of 1000. b/bytes kb kib mb mib gb gib will print the whole tree in that size
|
||||||
.br
|
|
||||||
|
|
||||||
.br
|
|
||||||
[\fIpossible values: \fRsi, b, k, m, g, t, kb, mb, gb, tb]
|
|
||||||
.TP
|
.TP
|
||||||
\fB\-S\fR, \fB\-\-stack\-size\fR=\fISTACK_SIZE\fR
|
\fB\-S\fR, \fB\-\-stack\-size\fR
|
||||||
Specify memory to use as stack size \- use if you see: \*(Aqfatal runtime error: stack overflow\*(Aq (default low memory=1048576, high memory=1073741824)
|
Specify memory to use as stack size \- use if you see: \*(Aqfatal runtime error: stack overflow\*(Aq (default low memory=1048576, high memory=1073741824)
|
||||||
.TP
|
.TP
|
||||||
\fB\-j\fR, \fB\-\-output\-json\fR
|
|
||||||
Output the directory tree as json to the current directory
|
|
||||||
.TP
|
|
||||||
\fB\-M\fR, \fB\-\-mtime\fR
|
|
||||||
+/\-n matches files modified more/less than n days ago , and n matches files modified exactly n days ago, days are rounded down.That is +n => (−∞, curr−(n+1)), n => [curr−(n+1), curr−n), and \-n => (𝑐𝑢𝑟𝑟−𝑛, +∞)
|
|
||||||
.TP
|
|
||||||
\fB\-A\fR, \fB\-\-atime\fR
|
|
||||||
just like \-mtime, but based on file access time
|
|
||||||
.TP
|
|
||||||
\fB\-y\fR, \fB\-\-ctime\fR
|
|
||||||
just like \-mtime, but based on file change time
|
|
||||||
.TP
|
|
||||||
\fB\-\-files0\-from\fR
|
|
||||||
run dust on NUL\-terminated file names specified in file; if argument is \-, then read names from standard input
|
|
||||||
.TP
|
|
||||||
\fB\-\-collapse\fR
|
|
||||||
Keep these directories collapsed
|
|
||||||
.TP
|
|
||||||
\fB\-m\fR, \fB\-\-filetime\fR
|
|
||||||
Directory \*(Aqsize\*(Aq is max filetime of child files instead of disk size. while a/c/m for last accessed/changed/modified time
|
|
||||||
.br
|
|
||||||
|
|
||||||
.br
|
|
||||||
[\fIpossible values: \fRa, c, m]
|
|
||||||
.TP
|
|
||||||
\fB\-h\fR, \fB\-\-help\fR
|
\fB\-h\fR, \fB\-\-help\fR
|
||||||
Print help
|
Print help
|
||||||
.TP
|
.TP
|
||||||
\fB\-V\fR, \fB\-\-version\fR
|
\fB\-V\fR, \fB\-\-version\fR
|
||||||
Print version
|
Print version
|
||||||
.TP
|
.TP
|
||||||
[\fIPATH\fR]
|
[\fIparams\fR]
|
||||||
|
|
||||||
.SH VERSION
|
.SH VERSION
|
||||||
v1.2.0
|
v1.0.0
|
||||||
|
|||||||
137
src/cli.rs
137
src/cli.rs
@@ -1,4 +1,4 @@
|
|||||||
use clap::{Arg, Command, builder::PossibleValue, value_parser};
|
use clap::{value_parser, Arg, Command};
|
||||||
|
|
||||||
// For single thread mode set this variable on your command line:
|
// For single thread mode set this variable on your command line:
|
||||||
// export RAYON_NUM_THREADS=1
|
// export RAYON_NUM_THREADS=1
|
||||||
@@ -7,37 +7,19 @@ pub fn build_cli() -> Command {
|
|||||||
Command::new("Dust")
|
Command::new("Dust")
|
||||||
.about("Like du but more intuitive")
|
.about("Like du but more intuitive")
|
||||||
.version(env!("CARGO_PKG_VERSION"))
|
.version(env!("CARGO_PKG_VERSION"))
|
||||||
|
.trailing_var_arg(true)
|
||||||
.arg(
|
.arg(
|
||||||
Arg::new("depth")
|
Arg::new("depth")
|
||||||
.short('d')
|
.short('d')
|
||||||
.long("depth")
|
.long("depth")
|
||||||
.value_name("DEPTH")
|
|
||||||
.value_parser(value_parser!(usize))
|
.value_parser(value_parser!(usize))
|
||||||
.help("Depth to show")
|
.help("Depth to show")
|
||||||
.num_args(1)
|
.num_args(1)
|
||||||
)
|
)
|
||||||
.arg(
|
|
||||||
Arg::new("threads")
|
|
||||||
.short('T')
|
|
||||||
.long("threads")
|
|
||||||
.value_parser(value_parser!(usize))
|
|
||||||
.help("Number of threads to use")
|
|
||||||
.num_args(1)
|
|
||||||
)
|
|
||||||
.arg(
|
|
||||||
Arg::new("config")
|
|
||||||
.long("config")
|
|
||||||
.help("Specify a config file to use")
|
|
||||||
.value_name("FILE")
|
|
||||||
.value_hint(clap::ValueHint::FilePath)
|
|
||||||
.value_parser(value_parser!(String))
|
|
||||||
.num_args(1)
|
|
||||||
)
|
|
||||||
.arg(
|
.arg(
|
||||||
Arg::new("number_of_lines")
|
Arg::new("number_of_lines")
|
||||||
.short('n')
|
.short('n')
|
||||||
.long("number-of-lines")
|
.long("number-of-lines")
|
||||||
.value_name("NUMBER")
|
|
||||||
.value_parser(value_parser!(usize))
|
.value_parser(value_parser!(usize))
|
||||||
.help("Number of lines of output to show. (Default is terminal_height - 10)")
|
.help("Number of lines of output to show. (Default is terminal_height - 10)")
|
||||||
.num_args(1)
|
.num_args(1)
|
||||||
@@ -53,17 +35,13 @@ pub fn build_cli() -> Command {
|
|||||||
Arg::new("ignore_directory")
|
Arg::new("ignore_directory")
|
||||||
.short('X')
|
.short('X')
|
||||||
.long("ignore-directory")
|
.long("ignore-directory")
|
||||||
.value_name("PATH")
|
|
||||||
.value_hint(clap::ValueHint::AnyPath)
|
|
||||||
.action(clap::ArgAction::Append)
|
.action(clap::ArgAction::Append)
|
||||||
.help("Exclude any file or directory with this path"),
|
.help("Exclude any file or directory with this name"),
|
||||||
)
|
)
|
||||||
.arg(
|
.arg(
|
||||||
Arg::new("ignore_all_in_file")
|
Arg::new("ignore_all_in_file")
|
||||||
.short('I')
|
.short('I')
|
||||||
.long("ignore-all-in-file")
|
.long("ignore-all-in-file")
|
||||||
.value_name("FILE")
|
|
||||||
.value_hint(clap::ValueHint::FilePath)
|
|
||||||
.value_parser(value_parser!(String))
|
.value_parser(value_parser!(String))
|
||||||
.help("Exclude any file or directory with a regex matching that listed in this file, the file entries will be added to the ignore regexs provided by --invert_filter"),
|
.help("Exclude any file or directory with a regex matching that listed in this file, the file entries will be added to the ignore regexs provided by --invert_filter"),
|
||||||
)
|
)
|
||||||
@@ -127,7 +105,6 @@ pub fn build_cli() -> Command {
|
|||||||
Arg::new("min_size")
|
Arg::new("min_size")
|
||||||
.short('z')
|
.short('z')
|
||||||
.long("min-size")
|
.long("min-size")
|
||||||
.value_name("MIN_SIZE")
|
|
||||||
.num_args(1)
|
.num_args(1)
|
||||||
.help("Minimum size file to include in output"),
|
.help("Minimum size file to include in output"),
|
||||||
)
|
)
|
||||||
@@ -154,7 +131,7 @@ pub fn build_cli() -> Command {
|
|||||||
.arg(
|
.arg(
|
||||||
Arg::new("ignore_hidden")
|
Arg::new("ignore_hidden")
|
||||||
.short('i') // Do not use 'h' this is used by 'help'
|
.short('i') // Do not use 'h' this is used by 'help'
|
||||||
.long("ignore-hidden")
|
.long("ignore_hidden")
|
||||||
.action(clap::ArgAction::SetTrue)
|
.action(clap::ArgAction::SetTrue)
|
||||||
.help("Do not display hidden files"),
|
.help("Do not display hidden files"),
|
||||||
)
|
)
|
||||||
@@ -162,7 +139,6 @@ pub fn build_cli() -> Command {
|
|||||||
Arg::new("invert_filter")
|
Arg::new("invert_filter")
|
||||||
.short('v')
|
.short('v')
|
||||||
.long("invert-filter")
|
.long("invert-filter")
|
||||||
.value_name("REGEX")
|
|
||||||
.action(clap::ArgAction::Append)
|
.action(clap::ArgAction::Append)
|
||||||
.conflicts_with("filter")
|
.conflicts_with("filter")
|
||||||
.conflicts_with("types")
|
.conflicts_with("types")
|
||||||
@@ -172,7 +148,6 @@ pub fn build_cli() -> Command {
|
|||||||
Arg::new("filter")
|
Arg::new("filter")
|
||||||
.short('e')
|
.short('e')
|
||||||
.long("filter")
|
.long("filter")
|
||||||
.value_name("REGEX")
|
|
||||||
.action(clap::ArgAction::Append)
|
.action(clap::ArgAction::Append)
|
||||||
.conflicts_with("types")
|
.conflicts_with("types")
|
||||||
.help("Only include filepaths matching this regex. For png files type: -e \"\\.png$\" "),
|
.help("Only include filepaths matching this regex. For png files type: -e \"\\.png$\" "),
|
||||||
@@ -180,7 +155,7 @@ pub fn build_cli() -> Command {
|
|||||||
.arg(
|
.arg(
|
||||||
Arg::new("types")
|
Arg::new("types")
|
||||||
.short('t')
|
.short('t')
|
||||||
.long("file-types")
|
.long("file_types")
|
||||||
.conflicts_with("depth")
|
.conflicts_with("depth")
|
||||||
.conflicts_with("only_dir")
|
.conflicts_with("only_dir")
|
||||||
.action(clap::ArgAction::SetTrue)
|
.action(clap::ArgAction::SetTrue)
|
||||||
@@ -189,10 +164,9 @@ pub fn build_cli() -> Command {
|
|||||||
.arg(
|
.arg(
|
||||||
Arg::new("width")
|
Arg::new("width")
|
||||||
.short('w')
|
.short('w')
|
||||||
.long("terminal-width")
|
.long("terminal_width")
|
||||||
.value_name("WIDTH")
|
|
||||||
.value_parser(value_parser!(usize))
|
|
||||||
.num_args(1)
|
.num_args(1)
|
||||||
|
.value_parser(value_parser!(usize))
|
||||||
.help("Specify width of output overriding the auto detection of terminal width"),
|
.help("Specify width of output overriding the auto detection of terminal width"),
|
||||||
)
|
)
|
||||||
.arg(
|
.arg(
|
||||||
@@ -202,12 +176,6 @@ pub fn build_cli() -> Command {
|
|||||||
.action(clap::ArgAction::SetTrue)
|
.action(clap::ArgAction::SetTrue)
|
||||||
.help("Disable the progress indication."),
|
.help("Disable the progress indication."),
|
||||||
)
|
)
|
||||||
.arg(
|
|
||||||
Arg::new("print_errors")
|
|
||||||
.long("print-errors")
|
|
||||||
.action(clap::ArgAction::SetTrue)
|
|
||||||
.help("Print path with errors."),
|
|
||||||
)
|
|
||||||
.arg(
|
.arg(
|
||||||
Arg::new("only_dir")
|
Arg::new("only_dir")
|
||||||
.short('D')
|
.short('D')
|
||||||
@@ -229,98 +197,17 @@ pub fn build_cli() -> Command {
|
|||||||
Arg::new("output_format")
|
Arg::new("output_format")
|
||||||
.short('o')
|
.short('o')
|
||||||
.long("output-format")
|
.long("output-format")
|
||||||
.value_name("FORMAT")
|
.value_parser(value_parser!(String))
|
||||||
.value_parser([
|
.help("Changes output display size. si will print sizes in powers of 1000. b/bytes kb kib mb mib gb gib will print the whole tree in that size")
|
||||||
PossibleValue::new("si"),
|
|
||||||
PossibleValue::new("b"),
|
|
||||||
PossibleValue::new("k").alias("kib"),
|
|
||||||
PossibleValue::new("m").alias("mib"),
|
|
||||||
PossibleValue::new("g").alias("gib"),
|
|
||||||
PossibleValue::new("t").alias("tib"),
|
|
||||||
PossibleValue::new("kb"),
|
|
||||||
PossibleValue::new("mb"),
|
|
||||||
PossibleValue::new("gb"),
|
|
||||||
PossibleValue::new("tb"),
|
|
||||||
])
|
|
||||||
.ignore_case(true)
|
|
||||||
.help("Changes output display size. si will print sizes in powers of 1000. b k m g t kb mb gb tb will print the whole tree in that size.")
|
|
||||||
)
|
)
|
||||||
.arg(
|
.arg(
|
||||||
Arg::new("stack_size")
|
Arg::new("stack_size")
|
||||||
.short('S')
|
.short('S')
|
||||||
.long("stack-size")
|
.long("stack-size")
|
||||||
.value_name("STACK_SIZE")
|
|
||||||
.value_parser(value_parser!(usize))
|
|
||||||
.num_args(1)
|
.num_args(1)
|
||||||
|
.value_parser(value_parser!(usize))
|
||||||
.help("Specify memory to use as stack size - use if you see: 'fatal runtime error: stack overflow' (default low memory=1048576, high memory=1073741824)"),
|
.help("Specify memory to use as stack size - use if you see: 'fatal runtime error: stack overflow' (default low memory=1048576, high memory=1073741824)"),
|
||||||
)
|
)
|
||||||
.arg(
|
.arg(Arg::new("params").num_args(1..)
|
||||||
Arg::new("params")
|
.value_parser(value_parser!(String)))
|
||||||
.value_name("PATH")
|
|
||||||
.value_hint(clap::ValueHint::AnyPath)
|
|
||||||
.value_parser(value_parser!(String))
|
|
||||||
.num_args(1..)
|
|
||||||
)
|
|
||||||
.arg(
|
|
||||||
Arg::new("output_json")
|
|
||||||
.short('j')
|
|
||||||
.long("output-json")
|
|
||||||
.action(clap::ArgAction::SetTrue)
|
|
||||||
.help("Output the directory tree as json to the current directory"),
|
|
||||||
)
|
|
||||||
.arg(
|
|
||||||
Arg::new("mtime")
|
|
||||||
.short('M')
|
|
||||||
.long("mtime")
|
|
||||||
.num_args(1)
|
|
||||||
.allow_hyphen_values(true)
|
|
||||||
.value_parser(value_parser!(String))
|
|
||||||
.help("+/-n matches files modified more/less than n days ago , and n matches files modified exactly n days ago, days are rounded down.That is +n => (−∞, curr−(n+1)), n => [curr−(n+1), curr−n), and -n => (𝑐𝑢𝑟𝑟−𝑛, +∞)")
|
|
||||||
)
|
|
||||||
.arg(
|
|
||||||
Arg::new("atime")
|
|
||||||
.short('A')
|
|
||||||
.long("atime")
|
|
||||||
.num_args(1)
|
|
||||||
.allow_hyphen_values(true)
|
|
||||||
.value_parser(value_parser!(String))
|
|
||||||
.help("just like -mtime, but based on file access time")
|
|
||||||
)
|
|
||||||
.arg(
|
|
||||||
Arg::new("ctime")
|
|
||||||
.short('y')
|
|
||||||
.long("ctime")
|
|
||||||
.num_args(1)
|
|
||||||
.allow_hyphen_values(true)
|
|
||||||
.value_parser(value_parser!(String))
|
|
||||||
.help("just like -mtime, but based on file change time")
|
|
||||||
)
|
|
||||||
.arg(
|
|
||||||
Arg::new("files0_from")
|
|
||||||
.long("files0-from")
|
|
||||||
.value_hint(clap::ValueHint::AnyPath)
|
|
||||||
.value_parser(value_parser!(String))
|
|
||||||
.num_args(1)
|
|
||||||
.help("run dust on NUL-terminated file names specified in file; if argument is -, then read names from standard input"),
|
|
||||||
)
|
|
||||||
.arg(
|
|
||||||
Arg::new("collapse")
|
|
||||||
.long("collapse")
|
|
||||||
.value_hint(clap::ValueHint::AnyPath)
|
|
||||||
.value_parser(value_parser!(String))
|
|
||||||
.action(clap::ArgAction::Append)
|
|
||||||
.help("Keep these directories collapsed"),
|
|
||||||
)
|
|
||||||
.arg(
|
|
||||||
Arg::new("filetime")
|
|
||||||
.short('m')
|
|
||||||
.long("filetime")
|
|
||||||
.num_args(1)
|
|
||||||
.value_parser([
|
|
||||||
PossibleValue::new("a").alias("accessed"),
|
|
||||||
PossibleValue::new("c").alias("changed"),
|
|
||||||
PossibleValue::new("m").alias("modified"),
|
|
||||||
])
|
|
||||||
.help("Directory 'size' is max filetime of child files instead of disk size. while a/c/m for last accessed/changed/modified time"),
|
|
||||||
)
|
|
||||||
}
|
}
|
||||||
|
|||||||
224
src/config.rs
224
src/config.rs
@@ -1,19 +1,16 @@
|
|||||||
use crate::node::FileTime;
|
|
||||||
use chrono::{Local, TimeZone};
|
|
||||||
use clap::ArgMatches;
|
use clap::ArgMatches;
|
||||||
use config_file::FromConfigFile;
|
use config_file::FromConfigFile;
|
||||||
use regex::Regex;
|
use regex::Regex;
|
||||||
use serde::Deserialize;
|
use serde::Deserialize;
|
||||||
|
use std::io::IsTerminal;
|
||||||
use std::path::Path;
|
use std::path::Path;
|
||||||
use std::path::PathBuf;
|
use std::path::PathBuf;
|
||||||
|
|
||||||
use crate::dir_walker::Operator;
|
|
||||||
use crate::display::get_number_format;
|
use crate::display::get_number_format;
|
||||||
|
|
||||||
pub static DAY_SECONDS: i64 = 24 * 60 * 60;
|
|
||||||
|
|
||||||
#[derive(Deserialize, Default)]
|
#[derive(Deserialize, Default)]
|
||||||
#[serde(rename_all = "kebab-case")]
|
#[serde(rename_all = "kebab-case")]
|
||||||
|
#[serde(deny_unknown_fields)]
|
||||||
pub struct Config {
|
pub struct Config {
|
||||||
pub display_full_paths: Option<bool>,
|
pub display_full_paths: Option<bool>,
|
||||||
pub display_apparent_size: Option<bool>,
|
pub display_apparent_size: Option<bool>,
|
||||||
@@ -32,20 +29,9 @@ pub struct Config {
|
|||||||
pub depth: Option<usize>,
|
pub depth: Option<usize>,
|
||||||
pub bars_on_right: Option<bool>,
|
pub bars_on_right: Option<bool>,
|
||||||
pub stack_size: Option<usize>,
|
pub stack_size: Option<usize>,
|
||||||
pub threads: Option<usize>,
|
|
||||||
pub output_json: Option<bool>,
|
|
||||||
pub print_errors: Option<bool>,
|
|
||||||
pub files0_from: Option<String>,
|
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Config {
|
impl Config {
|
||||||
pub fn get_files_from(&self, options: &ArgMatches) -> Option<String> {
|
|
||||||
let from_file = options.get_one::<String>("files0_from");
|
|
||||||
match from_file {
|
|
||||||
None => self.files0_from.as_ref().map(|x| x.to_string()),
|
|
||||||
Some(x) => Some(x.to_string()),
|
|
||||||
}
|
|
||||||
}
|
|
||||||
pub fn get_no_colors(&self, options: &ArgMatches) -> bool {
|
pub fn get_no_colors(&self, options: &ArgMatches) -> bool {
|
||||||
Some(true) == self.no_colors || options.get_flag("no_colors")
|
Some(true) == self.no_colors || options.get_flag("no_colors")
|
||||||
}
|
}
|
||||||
@@ -53,7 +39,9 @@ impl Config {
|
|||||||
Some(true) == self.force_colors || options.get_flag("force_colors")
|
Some(true) == self.force_colors || options.get_flag("force_colors")
|
||||||
}
|
}
|
||||||
pub fn get_disable_progress(&self, options: &ArgMatches) -> bool {
|
pub fn get_disable_progress(&self, options: &ArgMatches) -> bool {
|
||||||
Some(true) == self.disable_progress || options.get_flag("disable_progress")
|
Some(true) == self.disable_progress
|
||||||
|
|| options.get_flag("disable_progress")
|
||||||
|
|| !std::io::stdout().is_terminal()
|
||||||
}
|
}
|
||||||
pub fn get_apparent_size(&self, options: &ArgMatches) -> bool {
|
pub fn get_apparent_size(&self, options: &ArgMatches) -> bool {
|
||||||
Some(true) == self.display_apparent_size || options.get_flag("display_apparent_size")
|
Some(true) == self.display_apparent_size || options.get_flag("display_apparent_size")
|
||||||
@@ -62,7 +50,10 @@ impl Config {
|
|||||||
Some(true) == self.ignore_hidden || options.get_flag("ignore_hidden")
|
Some(true) == self.ignore_hidden || options.get_flag("ignore_hidden")
|
||||||
}
|
}
|
||||||
pub fn get_full_paths(&self, options: &ArgMatches) -> bool {
|
pub fn get_full_paths(&self, options: &ArgMatches) -> bool {
|
||||||
Some(true) == self.display_full_paths || options.get_flag("display_full_paths")
|
// If we are only showing files, always show full paths
|
||||||
|
Some(true) == self.display_full_paths
|
||||||
|
|| options.get_flag("display_full_paths")
|
||||||
|
|| self.get_only_file(options)
|
||||||
}
|
}
|
||||||
pub fn get_reverse(&self, options: &ArgMatches) -> bool {
|
pub fn get_reverse(&self, options: &ArgMatches) -> bool {
|
||||||
Some(true) == self.reverse || options.get_flag("reverse")
|
Some(true) == self.reverse || options.get_flag("reverse")
|
||||||
@@ -81,20 +72,6 @@ impl Config {
|
|||||||
})
|
})
|
||||||
.to_lowercase()
|
.to_lowercase()
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn get_filetime(&self, options: &ArgMatches) -> Option<FileTime> {
|
|
||||||
let out_fmt = options.get_one::<String>("filetime");
|
|
||||||
match out_fmt {
|
|
||||||
None => None,
|
|
||||||
Some(x) => match x.as_str() {
|
|
||||||
"m" | "modified" => Some(FileTime::Modified),
|
|
||||||
"a" | "accessed" => Some(FileTime::Accessed),
|
|
||||||
"c" | "changed" => Some(FileTime::Changed),
|
|
||||||
_ => unreachable!(),
|
|
||||||
},
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
pub fn get_skip_total(&self, options: &ArgMatches) -> bool {
|
pub fn get_skip_total(&self, options: &ArgMatches) -> bool {
|
||||||
Some(true) == self.skip_total || options.get_flag("skip_total")
|
Some(true) == self.skip_total || options.get_flag("skip_total")
|
||||||
}
|
}
|
||||||
@@ -126,10 +103,6 @@ impl Config {
|
|||||||
pub fn get_only_dir(&self, options: &ArgMatches) -> bool {
|
pub fn get_only_dir(&self, options: &ArgMatches) -> bool {
|
||||||
Some(true) == self.only_dir || options.get_flag("only_dir")
|
Some(true) == self.only_dir || options.get_flag("only_dir")
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn get_print_errors(&self, options: &ArgMatches) -> bool {
|
|
||||||
Some(true) == self.print_errors || options.get_flag("print_errors")
|
|
||||||
}
|
|
||||||
pub fn get_only_file(&self, options: &ArgMatches) -> bool {
|
pub fn get_only_file(&self, options: &ArgMatches) -> bool {
|
||||||
Some(true) == self.only_file || options.get_flag("only_file")
|
Some(true) == self.only_file || options.get_flag("only_file")
|
||||||
}
|
}
|
||||||
@@ -144,72 +117,6 @@ impl Config {
|
|||||||
from_cmd_line.copied()
|
from_cmd_line.copied()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
pub fn get_threads(&self, options: &ArgMatches) -> Option<usize> {
|
|
||||||
let from_cmd_line = options.get_one::<usize>("threads");
|
|
||||||
if from_cmd_line.is_none() {
|
|
||||||
self.threads
|
|
||||||
} else {
|
|
||||||
from_cmd_line.copied()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
pub fn get_output_json(&self, options: &ArgMatches) -> bool {
|
|
||||||
Some(true) == self.output_json || options.get_flag("output_json")
|
|
||||||
}
|
|
||||||
|
|
||||||
pub fn get_modified_time_operator(&self, options: &ArgMatches) -> Option<(Operator, i64)> {
|
|
||||||
get_filter_time_operator(
|
|
||||||
options.get_one::<String>("mtime"),
|
|
||||||
get_current_date_epoch_seconds(),
|
|
||||||
)
|
|
||||||
}
|
|
||||||
|
|
||||||
pub fn get_accessed_time_operator(&self, options: &ArgMatches) -> Option<(Operator, i64)> {
|
|
||||||
get_filter_time_operator(
|
|
||||||
options.get_one::<String>("atime"),
|
|
||||||
get_current_date_epoch_seconds(),
|
|
||||||
)
|
|
||||||
}
|
|
||||||
|
|
||||||
pub fn get_changed_time_operator(&self, options: &ArgMatches) -> Option<(Operator, i64)> {
|
|
||||||
get_filter_time_operator(
|
|
||||||
options.get_one::<String>("ctime"),
|
|
||||||
get_current_date_epoch_seconds(),
|
|
||||||
)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
fn get_current_date_epoch_seconds() -> i64 {
|
|
||||||
// calculate current date epoch seconds
|
|
||||||
let now = Local::now();
|
|
||||||
let current_date = now.date_naive();
|
|
||||||
|
|
||||||
let current_date_time = current_date.and_hms_opt(0, 0, 0).unwrap();
|
|
||||||
Local
|
|
||||||
.from_local_datetime(¤t_date_time)
|
|
||||||
.unwrap()
|
|
||||||
.timestamp()
|
|
||||||
}
|
|
||||||
|
|
||||||
fn get_filter_time_operator(
|
|
||||||
option_value: Option<&String>,
|
|
||||||
current_date_epoch_seconds: i64,
|
|
||||||
) -> Option<(Operator, i64)> {
|
|
||||||
match option_value {
|
|
||||||
Some(val) => {
|
|
||||||
let time = current_date_epoch_seconds
|
|
||||||
- val
|
|
||||||
.parse::<i64>()
|
|
||||||
.unwrap_or_else(|_| panic!("invalid data format"))
|
|
||||||
.abs()
|
|
||||||
* DAY_SECONDS;
|
|
||||||
match val.chars().next().expect("Value should not be empty") {
|
|
||||||
'+' => Some((Operator::LessThan, time - DAY_SECONDS)),
|
|
||||||
'-' => Some((Operator::GreaterThan, time)),
|
|
||||||
_ => Some((Operator::Equal, time - DAY_SECONDS)),
|
|
||||||
}
|
|
||||||
}
|
|
||||||
None => None,
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
fn convert_min_size(input: &str) -> Option<usize> {
|
fn convert_min_size(input: &str) -> Option<usize> {
|
||||||
@@ -227,7 +134,7 @@ fn convert_min_size(input: &str) -> Option<usize> {
|
|||||||
match number_format {
|
match number_format {
|
||||||
Some((multiple, _)) => Some(parsed_digits * (multiple as usize)),
|
Some((multiple, _)) => Some(parsed_digits * (multiple as usize)),
|
||||||
None => {
|
None => {
|
||||||
if letters.is_empty() {
|
if letters.eq("") {
|
||||||
Some(parsed_digits)
|
Some(parsed_digits)
|
||||||
} else {
|
} else {
|
||||||
eprintln!("Ignoring invalid min-size: {input}");
|
eprintln!("Ignoring invalid min-size: {input}");
|
||||||
@@ -250,29 +157,12 @@ fn get_config_locations(base: &Path) -> Vec<PathBuf> {
|
|||||||
]
|
]
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn get_config(conf_path: Option<String>) -> Config {
|
pub fn get_config() -> Config {
|
||||||
match conf_path {
|
if let Some(home) = directories::BaseDirs::new() {
|
||||||
Some(path_str) => {
|
for path in get_config_locations(home.home_dir()) {
|
||||||
let path = Path::new(&path_str);
|
|
||||||
if path.exists() {
|
if path.exists() {
|
||||||
match Config::from_config_file(path) {
|
if let Ok(config) = Config::from_config_file(path) {
|
||||||
Ok(config) => return config,
|
return config;
|
||||||
Err(e) => {
|
|
||||||
eprintln!("Ignoring invalid config file '{}': {}", &path.display(), e)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
eprintln!("Config file {:?} doesn't exists", &path.display());
|
|
||||||
}
|
|
||||||
}
|
|
||||||
None => {
|
|
||||||
if let Some(home) = directories::BaseDirs::new() {
|
|
||||||
for path in get_config_locations(home.home_dir()) {
|
|
||||||
if path.exists() {
|
|
||||||
if let Ok(config) = Config::from_config_file(&path) {
|
|
||||||
return config;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -286,46 +176,30 @@ pub fn get_config(conf_path: Option<String>) -> Config {
|
|||||||
mod tests {
|
mod tests {
|
||||||
#[allow(unused_imports)]
|
#[allow(unused_imports)]
|
||||||
use super::*;
|
use super::*;
|
||||||
use chrono::{Datelike, Timelike};
|
use clap::{value_parser, Arg, ArgMatches, Command};
|
||||||
use clap::builder::PossibleValue;
|
|
||||||
use clap::{Arg, ArgMatches, Command, value_parser};
|
|
||||||
|
|
||||||
#[test]
|
|
||||||
fn test_get_current_date_epoch_seconds() {
|
|
||||||
let epoch_seconds = get_current_date_epoch_seconds();
|
|
||||||
let dt = Local.timestamp_opt(epoch_seconds, 0).unwrap();
|
|
||||||
|
|
||||||
assert_eq!(dt.hour(), 0);
|
|
||||||
assert_eq!(dt.minute(), 0);
|
|
||||||
assert_eq!(dt.second(), 0);
|
|
||||||
assert_eq!(dt.date_naive().day(), Local::now().date_naive().day());
|
|
||||||
assert_eq!(dt.date_naive().month(), Local::now().date_naive().month());
|
|
||||||
assert_eq!(dt.date_naive().year(), Local::now().date_naive().year());
|
|
||||||
}
|
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn test_conversion() {
|
fn test_conversion() {
|
||||||
assert_eq!(convert_min_size("55"), Some(55));
|
assert_eq!(convert_min_size("55"), Some(55));
|
||||||
assert_eq!(convert_min_size("12344321"), Some(12344321));
|
assert_eq!(convert_min_size("12344321"), Some(12344321));
|
||||||
assert_eq!(convert_min_size("95RUBBISH"), None);
|
assert_eq!(convert_min_size("95RUBBISH"), None);
|
||||||
assert_eq!(convert_min_size("10Ki"), Some(10 * 1024));
|
assert_eq!(convert_min_size("10K"), Some(10 * 1024));
|
||||||
assert_eq!(convert_min_size("10MiB"), Some(10 * 1024usize.pow(2)));
|
|
||||||
assert_eq!(convert_min_size("10M"), Some(10 * 1024usize.pow(2)));
|
assert_eq!(convert_min_size("10M"), Some(10 * 1024usize.pow(2)));
|
||||||
assert_eq!(convert_min_size("10Mb"), Some(10 * 1000usize.pow(2)));
|
assert_eq!(convert_min_size("10MiB"), Some(10 * 1000usize.pow(2)));
|
||||||
assert_eq!(convert_min_size("2Gi"), Some(2 * 1024usize.pow(3)));
|
assert_eq!(convert_min_size("2G"), Some(2 * 1024usize.pow(3)));
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn test_min_size_from_config_applied_or_overridden() {
|
fn test_min_size_from_config_applied_or_overridden() {
|
||||||
let c = Config {
|
let c = Config {
|
||||||
min_size: Some("1KiB".to_owned()),
|
min_size: Some("1K".to_owned()),
|
||||||
..Default::default()
|
..Default::default()
|
||||||
};
|
};
|
||||||
assert_eq!(c._get_min_size(None), Some(1024));
|
assert_eq!(c._get_min_size(None), Some(1024));
|
||||||
assert_eq!(c._get_min_size(Some(&"2KiB".into())), Some(2048));
|
assert_eq!(c._get_min_size(Some(&"2K".into())), Some(2048));
|
||||||
|
|
||||||
assert_eq!(c._get_min_size(Some(&"1kb".into())), Some(1000));
|
assert_eq!(c._get_min_size(Some(&"1kib".into())), Some(1000));
|
||||||
assert_eq!(c._get_min_size(Some(&"2KB".into())), Some(2000));
|
assert_eq!(c._get_min_size(Some(&"2KiB".into())), Some(2000));
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
@@ -367,56 +241,4 @@ mod tests {
|
|||||||
)
|
)
|
||||||
.get_matches_from(args)
|
.get_matches_from(args)
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
|
||||||
fn test_get_filetime() {
|
|
||||||
// No config and no flag.
|
|
||||||
let c = Config::default();
|
|
||||||
let args = get_filetime_args(vec!["dust"]);
|
|
||||||
assert_eq!(c.get_filetime(&args), None);
|
|
||||||
|
|
||||||
// Config is not defined and flag is defined as access time
|
|
||||||
let c = Config::default();
|
|
||||||
let args = get_filetime_args(vec!["dust", "--filetime", "a"]);
|
|
||||||
assert_eq!(c.get_filetime(&args), Some(FileTime::Accessed));
|
|
||||||
|
|
||||||
let c = Config::default();
|
|
||||||
let args = get_filetime_args(vec!["dust", "--filetime", "accessed"]);
|
|
||||||
assert_eq!(c.get_filetime(&args), Some(FileTime::Accessed));
|
|
||||||
|
|
||||||
// Config is not defined and flag is defined as modified time
|
|
||||||
let c = Config::default();
|
|
||||||
let args = get_filetime_args(vec!["dust", "--filetime", "m"]);
|
|
||||||
assert_eq!(c.get_filetime(&args), Some(FileTime::Modified));
|
|
||||||
|
|
||||||
let c = Config::default();
|
|
||||||
let args = get_filetime_args(vec!["dust", "--filetime", "modified"]);
|
|
||||||
assert_eq!(c.get_filetime(&args), Some(FileTime::Modified));
|
|
||||||
|
|
||||||
// Config is not defined and flag is defined as changed time
|
|
||||||
let c = Config::default();
|
|
||||||
let args = get_filetime_args(vec!["dust", "--filetime", "c"]);
|
|
||||||
assert_eq!(c.get_filetime(&args), Some(FileTime::Changed));
|
|
||||||
|
|
||||||
let c = Config::default();
|
|
||||||
let args = get_filetime_args(vec!["dust", "--filetime", "changed"]);
|
|
||||||
assert_eq!(c.get_filetime(&args), Some(FileTime::Changed));
|
|
||||||
}
|
|
||||||
|
|
||||||
fn get_filetime_args(args: Vec<&str>) -> ArgMatches {
|
|
||||||
Command::new("Dust")
|
|
||||||
.arg(
|
|
||||||
Arg::new("filetime")
|
|
||||||
.short('m')
|
|
||||||
.long("filetime")
|
|
||||||
.num_args(1)
|
|
||||||
.value_parser([
|
|
||||||
PossibleValue::new("a").alias("accessed"),
|
|
||||||
PossibleValue::new("c").alias("changed"),
|
|
||||||
PossibleValue::new("m").alias("modified"),
|
|
||||||
])
|
|
||||||
.help("Directory 'size' is max filetime of child files instead of disk size. while a/c/m for accessed/changed/modified time"),
|
|
||||||
)
|
|
||||||
.get_matches_from(args)
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,21 +1,17 @@
|
|||||||
use std::cmp::Ordering;
|
|
||||||
use std::fs;
|
use std::fs;
|
||||||
use std::io::Error;
|
|
||||||
use std::sync::Arc;
|
use std::sync::Arc;
|
||||||
use std::sync::Mutex;
|
use std::sync::Mutex;
|
||||||
|
|
||||||
use crate::node::Node;
|
use crate::node::Node;
|
||||||
use crate::progress::ORDERING;
|
|
||||||
use crate::progress::Operation;
|
use crate::progress::Operation;
|
||||||
use crate::progress::PAtomicInfo;
|
use crate::progress::PAtomicInfo;
|
||||||
use crate::progress::RuntimeErrors;
|
use crate::progress::RuntimeErrors;
|
||||||
use crate::utils::is_filtered_out_due_to_file_time;
|
use crate::progress::ORDERING;
|
||||||
use crate::utils::is_filtered_out_due_to_invert_regex;
|
use crate::utils::is_filtered_out_due_to_invert_regex;
|
||||||
use crate::utils::is_filtered_out_due_to_regex;
|
use crate::utils::is_filtered_out_due_to_regex;
|
||||||
use rayon::iter::ParallelBridge;
|
use rayon::iter::ParallelBridge;
|
||||||
use rayon::prelude::ParallelIterator;
|
use rayon::prelude::ParallelIterator;
|
||||||
use regex::Regex;
|
use regex::Regex;
|
||||||
use std::path::Path;
|
|
||||||
use std::path::PathBuf;
|
use std::path::PathBuf;
|
||||||
|
|
||||||
use std::collections::HashSet;
|
use std::collections::HashSet;
|
||||||
@@ -23,27 +19,14 @@ use std::collections::HashSet;
|
|||||||
use crate::node::build_node;
|
use crate::node::build_node;
|
||||||
use std::fs::DirEntry;
|
use std::fs::DirEntry;
|
||||||
|
|
||||||
use crate::node::FileTime;
|
|
||||||
use crate::platform::get_metadata;
|
use crate::platform::get_metadata;
|
||||||
|
|
||||||
#[derive(Debug)]
|
|
||||||
pub enum Operator {
|
|
||||||
Equal = 0,
|
|
||||||
LessThan = 1,
|
|
||||||
GreaterThan = 2,
|
|
||||||
}
|
|
||||||
|
|
||||||
pub struct WalkData<'a> {
|
pub struct WalkData<'a> {
|
||||||
pub ignore_directories: HashSet<PathBuf>,
|
pub ignore_directories: HashSet<PathBuf>,
|
||||||
pub filter_regex: &'a [Regex],
|
pub filter_regex: &'a [Regex],
|
||||||
pub invert_filter_regex: &'a [Regex],
|
pub invert_filter_regex: &'a [Regex],
|
||||||
pub allowed_filesystems: HashSet<u64>,
|
pub allowed_filesystems: HashSet<u64>,
|
||||||
pub filter_modified_time: Option<(Operator, i64)>,
|
|
||||||
pub filter_accessed_time: Option<(Operator, i64)>,
|
|
||||||
pub filter_changed_time: Option<(Operator, i64)>,
|
|
||||||
pub use_apparent_size: bool,
|
pub use_apparent_size: bool,
|
||||||
pub by_filecount: bool,
|
pub by_filecount: bool,
|
||||||
pub by_filetime: &'a Option<FileTime>,
|
|
||||||
pub ignore_hidden: bool,
|
pub ignore_hidden: bool,
|
||||||
pub follow_links: bool,
|
pub follow_links: bool,
|
||||||
pub progress_data: Arc<PAtomicInfo>,
|
pub progress_data: Arc<PAtomicInfo>,
|
||||||
@@ -61,15 +44,19 @@ pub fn walk_it(dirs: HashSet<PathBuf>, walk_data: &WalkData) -> Vec<Node> {
|
|||||||
|
|
||||||
prog_data.state.store(Operation::PREPARING, ORDERING);
|
prog_data.state.store(Operation::PREPARING, ORDERING);
|
||||||
|
|
||||||
clean_inodes(node, &mut inodes, walk_data)
|
clean_inodes(node, &mut inodes, walk_data.use_apparent_size)
|
||||||
})
|
})
|
||||||
.collect();
|
.collect();
|
||||||
top_level_nodes
|
top_level_nodes
|
||||||
}
|
}
|
||||||
|
|
||||||
// Remove files which have the same inode, we don't want to double count them.
|
// Remove files which have the same inode, we don't want to double count them.
|
||||||
fn clean_inodes(x: Node, inodes: &mut HashSet<(u64, u64)>, walk_data: &WalkData) -> Option<Node> {
|
fn clean_inodes(
|
||||||
if !walk_data.use_apparent_size {
|
x: Node,
|
||||||
|
inodes: &mut HashSet<(u64, u64)>,
|
||||||
|
use_apparent_size: bool,
|
||||||
|
) -> Option<Node> {
|
||||||
|
if !use_apparent_size {
|
||||||
if let Some(id) = x.inode_device {
|
if let Some(id) = x.inode_device {
|
||||||
if !inodes.insert(id) {
|
if !inodes.insert(id) {
|
||||||
return None;
|
return None;
|
||||||
@@ -82,25 +69,12 @@ fn clean_inodes(x: Node, inodes: &mut HashSet<(u64, u64)>, walk_data: &WalkData)
|
|||||||
tmp.sort_by(sort_by_inode);
|
tmp.sort_by(sort_by_inode);
|
||||||
let new_children: Vec<_> = tmp
|
let new_children: Vec<_> = tmp
|
||||||
.into_iter()
|
.into_iter()
|
||||||
.filter_map(|c| clean_inodes(c, inodes, walk_data))
|
.filter_map(|c| clean_inodes(c, inodes, use_apparent_size))
|
||||||
.collect();
|
.collect();
|
||||||
|
|
||||||
let actual_size = if walk_data.by_filetime.is_some() {
|
|
||||||
// If by_filetime is Some, directory 'size' is the maximum filetime among child files instead of disk size
|
|
||||||
new_children
|
|
||||||
.iter()
|
|
||||||
.map(|c| c.size)
|
|
||||||
.chain(std::iter::once(x.size))
|
|
||||||
.max()
|
|
||||||
.unwrap_or(0)
|
|
||||||
} else {
|
|
||||||
// If by_filetime is None, directory 'size' is the sum of disk sizes or file counts of child files
|
|
||||||
x.size + new_children.iter().map(|c| c.size).sum::<u64>()
|
|
||||||
};
|
|
||||||
|
|
||||||
Some(Node {
|
Some(Node {
|
||||||
name: x.name,
|
name: x.name,
|
||||||
size: actual_size,
|
size: x.size + new_children.iter().map(|c| c.size).sum::<u64>(),
|
||||||
children: new_children,
|
children: new_children,
|
||||||
inode_device: x.inode_device,
|
inode_device: x.inode_device,
|
||||||
depth: x.depth,
|
depth: x.depth,
|
||||||
@@ -109,80 +83,31 @@ fn clean_inodes(x: Node, inodes: &mut HashSet<(u64, u64)>, walk_data: &WalkData)
|
|||||||
|
|
||||||
fn sort_by_inode(a: &Node, b: &Node) -> std::cmp::Ordering {
|
fn sort_by_inode(a: &Node, b: &Node) -> std::cmp::Ordering {
|
||||||
// Sorting by inode is quicker than by sorting by name/size
|
// Sorting by inode is quicker than by sorting by name/size
|
||||||
match (a.inode_device, b.inode_device) {
|
if let Some(x) = a.inode_device {
|
||||||
(Some(x), Some(y)) => {
|
if let Some(y) = b.inode_device {
|
||||||
if x.0 != y.0 {
|
if x.0 != y.0 {
|
||||||
x.0.cmp(&y.0)
|
return x.0.cmp(&y.0);
|
||||||
} else if x.1 != y.1 {
|
} else if x.1 != y.1 {
|
||||||
x.1.cmp(&y.1)
|
return x.1.cmp(&y.1);
|
||||||
} else {
|
|
||||||
a.name.cmp(&b.name)
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
(Some(_), None) => Ordering::Greater,
|
|
||||||
(None, Some(_)) => Ordering::Less,
|
|
||||||
(None, None) => a.name.cmp(&b.name),
|
|
||||||
}
|
}
|
||||||
}
|
a.name.cmp(&b.name)
|
||||||
|
|
||||||
// Check if `path` is inside ignored directory
|
|
||||||
fn is_ignored_path(path: &Path, walk_data: &WalkData) -> bool {
|
|
||||||
if walk_data.ignore_directories.contains(path) {
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
|
|
||||||
// Entry is inside an ignored absolute path
|
|
||||||
// Absolute paths should be canonicalized before being added to `WalkData.ignore_directories`
|
|
||||||
for ignored_path in walk_data.ignore_directories.iter() {
|
|
||||||
if !ignored_path.is_absolute() {
|
|
||||||
continue;
|
|
||||||
}
|
|
||||||
let absolute_entry_path = std::fs::canonicalize(path).unwrap_or_default();
|
|
||||||
if absolute_entry_path.starts_with(ignored_path) {
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
false
|
|
||||||
}
|
}
|
||||||
|
|
||||||
fn ignore_file(entry: &DirEntry, walk_data: &WalkData) -> bool {
|
fn ignore_file(entry: &DirEntry, walk_data: &WalkData) -> bool {
|
||||||
if is_ignored_path(&entry.path(), walk_data) {
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
|
|
||||||
let is_dot_file = entry.file_name().to_str().unwrap_or("").starts_with('.');
|
let is_dot_file = entry.file_name().to_str().unwrap_or("").starts_with('.');
|
||||||
let follow_links = walk_data.follow_links && entry.file_type().is_ok_and(|ft| ft.is_symlink());
|
let is_ignored_path = walk_data.ignore_directories.contains(&entry.path());
|
||||||
|
|
||||||
if !walk_data.allowed_filesystems.is_empty() {
|
if !walk_data.allowed_filesystems.is_empty() {
|
||||||
let size_inode_device = get_metadata(entry.path(), false, follow_links);
|
let size_inode_device = get_metadata(&entry.path(), false);
|
||||||
if let Some((_size, Some((_id, dev)), _gunk)) = size_inode_device {
|
|
||||||
|
if let Some((_size, Some((_id, dev)))) = size_inode_device {
|
||||||
if !walk_data.allowed_filesystems.contains(&dev) {
|
if !walk_data.allowed_filesystems.contains(&dev) {
|
||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if walk_data.filter_accessed_time.is_some()
|
|
||||||
|| walk_data.filter_modified_time.is_some()
|
|
||||||
|| walk_data.filter_changed_time.is_some()
|
|
||||||
{
|
|
||||||
let size_inode_device = get_metadata(entry.path(), false, follow_links);
|
|
||||||
if let Some((_, _, (modified_time, accessed_time, changed_time))) = size_inode_device {
|
|
||||||
if entry.path().is_file()
|
|
||||||
&& [
|
|
||||||
(&walk_data.filter_modified_time, modified_time),
|
|
||||||
(&walk_data.filter_accessed_time, accessed_time),
|
|
||||||
(&walk_data.filter_changed_time, changed_time),
|
|
||||||
]
|
|
||||||
.iter()
|
|
||||||
.any(|(filter_time, actual_time)| {
|
|
||||||
is_filtered_out_due_to_file_time(filter_time, *actual_time)
|
|
||||||
})
|
|
||||||
{
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Keeping `walk_data.filter_regex.is_empty()` is important for performance reasons, it stops unnecessary work
|
// Keeping `walk_data.filter_regex.is_empty()` is important for performance reasons, it stops unnecessary work
|
||||||
if !walk_data.filter_regex.is_empty()
|
if !walk_data.filter_regex.is_empty()
|
||||||
@@ -199,12 +124,15 @@ fn ignore_file(entry: &DirEntry, walk_data: &WalkData) -> bool {
|
|||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
|
|
||||||
is_dot_file && walk_data.ignore_hidden
|
(is_dot_file && walk_data.ignore_hidden) || is_ignored_path
|
||||||
}
|
}
|
||||||
|
|
||||||
fn walk(dir: PathBuf, walk_data: &WalkData, depth: usize) -> Option<Node> {
|
fn walk(dir: PathBuf, walk_data: &WalkData, depth: usize) -> Option<Node> {
|
||||||
let prog_data = &walk_data.progress_data;
|
let prog_data = &walk_data.progress_data;
|
||||||
let errors = &walk_data.errors;
|
let errors = &walk_data.errors;
|
||||||
|
if errors.lock().unwrap().abort {
|
||||||
|
return None;
|
||||||
|
}
|
||||||
|
|
||||||
let children = if dir.is_dir() {
|
let children = if dir.is_dir() {
|
||||||
let read_dir = fs::read_dir(&dir);
|
let read_dir = fs::read_dir(&dir);
|
||||||
@@ -214,58 +142,63 @@ fn walk(dir: PathBuf, walk_data: &WalkData, depth: usize) -> Option<Node> {
|
|||||||
.into_iter()
|
.into_iter()
|
||||||
.par_bridge()
|
.par_bridge()
|
||||||
.filter_map(|entry| {
|
.filter_map(|entry| {
|
||||||
match entry {
|
if let Ok(ref entry) = entry {
|
||||||
Ok(ref entry) => {
|
// uncommenting the below line gives simpler code but
|
||||||
// uncommenting the below line gives simpler code but
|
// rayon doesn't parallelize as well giving a 3X performance drop
|
||||||
// rayon doesn't parallelize as well giving a 3X performance drop
|
// hence we unravel the recursion a bit
|
||||||
// hence we unravel the recursion a bit
|
|
||||||
|
|
||||||
// return walk(entry.path(), walk_data, depth)
|
// return walk(entry.path(), walk_data, depth)
|
||||||
|
|
||||||
if !ignore_file(entry, walk_data) {
|
if !ignore_file(entry, walk_data) {
|
||||||
if let Ok(data) = entry.file_type() {
|
if let Ok(data) = entry.file_type() {
|
||||||
if data.is_dir()
|
if data.is_dir()
|
||||||
|| (walk_data.follow_links && data.is_symlink())
|
|| (walk_data.follow_links && data.is_symlink())
|
||||||
{
|
{
|
||||||
return walk(entry.path(), walk_data, depth + 1);
|
return walk(entry.path(), walk_data, depth + 1);
|
||||||
}
|
|
||||||
|
|
||||||
let node = build_node(
|
|
||||||
entry.path(),
|
|
||||||
vec![],
|
|
||||||
data.is_symlink(),
|
|
||||||
data.is_file(),
|
|
||||||
depth,
|
|
||||||
walk_data,
|
|
||||||
);
|
|
||||||
|
|
||||||
prog_data.num_files.fetch_add(1, ORDERING);
|
|
||||||
if let Some(ref file) = node {
|
|
||||||
prog_data
|
|
||||||
.total_file_size
|
|
||||||
.fetch_add(file.size, ORDERING);
|
|
||||||
}
|
|
||||||
|
|
||||||
return node;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
let node = build_node(
|
||||||
|
entry.path(),
|
||||||
|
vec![],
|
||||||
|
walk_data.filter_regex,
|
||||||
|
walk_data.invert_filter_regex,
|
||||||
|
walk_data.use_apparent_size,
|
||||||
|
data.is_symlink(),
|
||||||
|
data.is_file(),
|
||||||
|
walk_data.by_filecount,
|
||||||
|
depth,
|
||||||
|
);
|
||||||
|
|
||||||
|
prog_data.num_files.fetch_add(1, ORDERING);
|
||||||
|
if let Some(ref file) = node {
|
||||||
|
prog_data.total_file_size.fetch_add(file.size, ORDERING);
|
||||||
|
}
|
||||||
|
|
||||||
|
return node;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
Err(ref failed) => {
|
} else {
|
||||||
if handle_error_and_retry(failed, &dir, walk_data) {
|
let mut editable_error = errors.lock().unwrap();
|
||||||
return walk(dir.clone(), walk_data, depth);
|
editable_error.no_permissions = true
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
None
|
None
|
||||||
})
|
})
|
||||||
.collect()
|
.collect()
|
||||||
}
|
}
|
||||||
Err(failed) => {
|
Err(failed) => {
|
||||||
if handle_error_and_retry(&failed, &dir, walk_data) {
|
let mut editable_error = errors.lock().unwrap();
|
||||||
return walk(dir, walk_data, depth);
|
match failed.kind() {
|
||||||
} else {
|
std::io::ErrorKind::PermissionDenied => {
|
||||||
vec![]
|
editable_error.no_permissions = true;
|
||||||
|
}
|
||||||
|
std::io::ErrorKind::NotFound => {
|
||||||
|
editable_error.file_not_found.insert(failed.to_string());
|
||||||
|
}
|
||||||
|
_ => {
|
||||||
|
editable_error.unknown_error.insert(failed.to_string());
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
vec![]
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
@@ -276,51 +209,20 @@ fn walk(dir: PathBuf, walk_data: &WalkData, depth: usize) -> Option<Node> {
|
|||||||
}
|
}
|
||||||
vec![]
|
vec![]
|
||||||
};
|
};
|
||||||
let is_symlink = if walk_data.follow_links {
|
build_node(
|
||||||
match fs::symlink_metadata(&dir) {
|
dir,
|
||||||
Ok(metadata) => metadata.file_type().is_symlink(),
|
children,
|
||||||
Err(_) => false,
|
walk_data.filter_regex,
|
||||||
}
|
walk_data.invert_filter_regex,
|
||||||
} else {
|
walk_data.use_apparent_size,
|
||||||
false
|
false,
|
||||||
};
|
false,
|
||||||
build_node(dir, children, is_symlink, false, depth, walk_data)
|
walk_data.by_filecount,
|
||||||
}
|
depth,
|
||||||
|
)
|
||||||
fn handle_error_and_retry(failed: &Error, dir: &Path, walk_data: &WalkData) -> bool {
|
|
||||||
let mut editable_error = walk_data.errors.lock().unwrap();
|
|
||||||
match failed.kind() {
|
|
||||||
std::io::ErrorKind::PermissionDenied => {
|
|
||||||
editable_error
|
|
||||||
.no_permissions
|
|
||||||
.insert(dir.to_string_lossy().into());
|
|
||||||
}
|
|
||||||
std::io::ErrorKind::InvalidInput => {
|
|
||||||
editable_error
|
|
||||||
.no_permissions
|
|
||||||
.insert(dir.to_string_lossy().into());
|
|
||||||
}
|
|
||||||
std::io::ErrorKind::NotFound => {
|
|
||||||
editable_error.file_not_found.insert(failed.to_string());
|
|
||||||
}
|
|
||||||
std::io::ErrorKind::Interrupted => {
|
|
||||||
let mut editable_error = walk_data.errors.lock().unwrap();
|
|
||||||
editable_error.interrupted_error += 1;
|
|
||||||
if editable_error.interrupted_error > 3 {
|
|
||||||
panic!("Multiple Interrupted Errors occurred while scanning filesystem. Aborting");
|
|
||||||
} else {
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
_ => {
|
|
||||||
editable_error.unknown_error.insert(failed.to_string());
|
|
||||||
}
|
|
||||||
}
|
|
||||||
false
|
|
||||||
}
|
}
|
||||||
|
|
||||||
mod tests {
|
mod tests {
|
||||||
|
|
||||||
#[allow(unused_imports)]
|
#[allow(unused_imports)]
|
||||||
use super::*;
|
use super::*;
|
||||||
|
|
||||||
@@ -335,43 +237,17 @@ mod tests {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
#[cfg(test)]
|
|
||||||
fn create_walker<'a>(use_apparent_size: bool) -> WalkData<'a> {
|
|
||||||
use crate::PIndicator;
|
|
||||||
let indicator = PIndicator::build_me();
|
|
||||||
WalkData {
|
|
||||||
ignore_directories: HashSet::new(),
|
|
||||||
filter_regex: &[],
|
|
||||||
invert_filter_regex: &[],
|
|
||||||
allowed_filesystems: HashSet::new(),
|
|
||||||
filter_modified_time: Some((Operator::GreaterThan, 0)),
|
|
||||||
filter_accessed_time: Some((Operator::GreaterThan, 0)),
|
|
||||||
filter_changed_time: Some((Operator::GreaterThan, 0)),
|
|
||||||
use_apparent_size,
|
|
||||||
by_filecount: false,
|
|
||||||
by_filetime: &None,
|
|
||||||
ignore_hidden: false,
|
|
||||||
follow_links: false,
|
|
||||||
progress_data: indicator.data.clone(),
|
|
||||||
errors: Arc::new(Mutex::new(RuntimeErrors::default())),
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
#[allow(clippy::redundant_clone)]
|
#[allow(clippy::redundant_clone)]
|
||||||
fn test_should_ignore_file() {
|
fn test_should_ignore_file() {
|
||||||
let mut inodes = HashSet::new();
|
let mut inodes = HashSet::new();
|
||||||
let n = create_node();
|
let n = create_node();
|
||||||
let walkdata = create_walker(false);
|
|
||||||
|
|
||||||
// First time we insert the node
|
// First time we insert the node
|
||||||
assert_eq!(
|
assert_eq!(clean_inodes(n.clone(), &mut inodes, false), Some(n.clone()));
|
||||||
clean_inodes(n.clone(), &mut inodes, &walkdata),
|
|
||||||
Some(n.clone())
|
|
||||||
);
|
|
||||||
|
|
||||||
// Second time is a duplicate - we ignore it
|
// Second time is a duplicate - we ignore it
|
||||||
assert_eq!(clean_inodes(n.clone(), &mut inodes, &walkdata), None);
|
assert_eq!(clean_inodes(n.clone(), &mut inodes, false), None);
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
@@ -379,53 +255,9 @@ mod tests {
|
|||||||
fn test_should_not_ignore_files_if_using_apparent_size() {
|
fn test_should_not_ignore_files_if_using_apparent_size() {
|
||||||
let mut inodes = HashSet::new();
|
let mut inodes = HashSet::new();
|
||||||
let n = create_node();
|
let n = create_node();
|
||||||
let walkdata = create_walker(true);
|
|
||||||
|
|
||||||
// If using apparent size we include Nodes, even if duplicate inodes
|
// If using apparent size we include Nodes, even if duplicate inodes
|
||||||
assert_eq!(
|
assert_eq!(clean_inodes(n.clone(), &mut inodes, true), Some(n.clone()));
|
||||||
clean_inodes(n.clone(), &mut inodes, &walkdata),
|
assert_eq!(clean_inodes(n.clone(), &mut inodes, true), Some(n.clone()));
|
||||||
Some(n.clone())
|
|
||||||
);
|
|
||||||
assert_eq!(
|
|
||||||
clean_inodes(n.clone(), &mut inodes, &walkdata),
|
|
||||||
Some(n.clone())
|
|
||||||
);
|
|
||||||
}
|
|
||||||
|
|
||||||
#[test]
|
|
||||||
fn test_total_ordering_of_sort_by_inode() {
|
|
||||||
use std::str::FromStr;
|
|
||||||
|
|
||||||
let a = Node {
|
|
||||||
name: PathBuf::from_str("a").unwrap(),
|
|
||||||
size: 0,
|
|
||||||
children: vec![],
|
|
||||||
inode_device: Some((3, 66310)),
|
|
||||||
depth: 0,
|
|
||||||
};
|
|
||||||
|
|
||||||
let b = Node {
|
|
||||||
name: PathBuf::from_str("b").unwrap(),
|
|
||||||
size: 0,
|
|
||||||
children: vec![],
|
|
||||||
inode_device: None,
|
|
||||||
depth: 0,
|
|
||||||
};
|
|
||||||
|
|
||||||
let c = Node {
|
|
||||||
name: PathBuf::from_str("c").unwrap(),
|
|
||||||
size: 0,
|
|
||||||
children: vec![],
|
|
||||||
inode_device: Some((1, 66310)),
|
|
||||||
depth: 0,
|
|
||||||
};
|
|
||||||
|
|
||||||
assert_eq!(sort_by_inode(&a, &b), Ordering::Greater);
|
|
||||||
assert_eq!(sort_by_inode(&a, &c), Ordering::Greater);
|
|
||||||
assert_eq!(sort_by_inode(&c, &b), Ordering::Greater);
|
|
||||||
|
|
||||||
assert_eq!(sort_by_inode(&b, &a), Ordering::Less);
|
|
||||||
assert_eq!(sort_by_inode(&c, &a), Ordering::Less);
|
|
||||||
assert_eq!(sort_by_inode(&b, &c), Ordering::Less);
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
111
src/display.rs
111
src/display.rs
@@ -1,5 +1,4 @@
|
|||||||
use crate::display_node::DisplayNode;
|
use crate::display_node::DisplayNode;
|
||||||
use crate::node::FileTime;
|
|
||||||
|
|
||||||
use ansi_term::Colour::Red;
|
use ansi_term::Colour::Red;
|
||||||
use lscolors::{LsColors, Style};
|
use lscolors::{LsColors, Style};
|
||||||
@@ -8,7 +7,6 @@ use unicode_width::UnicodeWidthStr;
|
|||||||
|
|
||||||
use stfu8::encode_u8;
|
use stfu8::encode_u8;
|
||||||
|
|
||||||
use chrono::{DateTime, Local, TimeZone, Utc};
|
|
||||||
use std::cmp::max;
|
use std::cmp::max;
|
||||||
use std::cmp::min;
|
use std::cmp::min;
|
||||||
use std::fs;
|
use std::fs;
|
||||||
@@ -16,16 +14,14 @@ use std::iter::repeat;
|
|||||||
use std::path::Path;
|
use std::path::Path;
|
||||||
use thousands::Separable;
|
use thousands::Separable;
|
||||||
|
|
||||||
pub static UNITS: [char; 5] = ['P', 'T', 'G', 'M', 'K'];
|
pub static UNITS: [char; 4] = ['T', 'G', 'M', 'K'];
|
||||||
static BLOCKS: [char; 5] = ['█', '▓', '▒', '░', ' '];
|
static BLOCKS: [char; 5] = ['█', '▓', '▒', '░', ' '];
|
||||||
const FILETIME_SHOW_LENGTH: usize = 19;
|
|
||||||
|
|
||||||
pub struct InitialDisplayData {
|
pub struct InitialDisplayData {
|
||||||
pub short_paths: bool,
|
pub short_paths: bool,
|
||||||
pub is_reversed: bool,
|
pub is_reversed: bool,
|
||||||
pub colors_on: bool,
|
pub colors_on: bool,
|
||||||
pub by_filecount: bool,
|
pub by_filecount: bool,
|
||||||
pub by_filetime: Option<FileTime>,
|
|
||||||
pub is_screen_reader: bool,
|
pub is_screen_reader: bool,
|
||||||
pub output_format: String,
|
pub output_format: String,
|
||||||
pub bars_on_right: bool,
|
pub bars_on_right: bool,
|
||||||
@@ -71,7 +67,11 @@ impl DisplayData {
|
|||||||
|
|
||||||
fn percent_size(&self, node: &DisplayNode) -> f32 {
|
fn percent_size(&self, node: &DisplayNode) -> f32 {
|
||||||
let result = node.size as f32 / self.base_size as f32;
|
let result = node.size as f32 / self.base_size as f32;
|
||||||
if result.is_normal() { result } else { 0.0 }
|
if result.is_normal() {
|
||||||
|
result
|
||||||
|
} else {
|
||||||
|
0.0
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -130,11 +130,17 @@ pub fn draw_it(
|
|||||||
root_node: &DisplayNode,
|
root_node: &DisplayNode,
|
||||||
skip_total: bool,
|
skip_total: bool,
|
||||||
) {
|
) {
|
||||||
|
let biggest = match skip_total {
|
||||||
|
false => root_node,
|
||||||
|
true => root_node
|
||||||
|
.get_children_from_node(false)
|
||||||
|
.next()
|
||||||
|
.unwrap_or(root_node),
|
||||||
|
};
|
||||||
|
|
||||||
let num_chars_needed_on_left_most = if idd.by_filecount {
|
let num_chars_needed_on_left_most = if idd.by_filecount {
|
||||||
let max_size = root_node.size;
|
let max_size = biggest.size;
|
||||||
max_size.separate_with_commas().chars().count()
|
max_size.separate_with_commas().chars().count()
|
||||||
} else if idd.by_filetime.is_some() {
|
|
||||||
FILETIME_SHOW_LENGTH
|
|
||||||
} else {
|
} else {
|
||||||
find_biggest_size_str(root_node, &idd.output_format)
|
find_biggest_size_str(root_node, &idd.output_format)
|
||||||
};
|
};
|
||||||
@@ -160,7 +166,7 @@ pub fn draw_it(
|
|||||||
let display_data = DisplayData {
|
let display_data = DisplayData {
|
||||||
initial: idd,
|
initial: idd,
|
||||||
num_chars_needed_on_left_most,
|
num_chars_needed_on_left_most,
|
||||||
base_size: root_node.size,
|
base_size: biggest.size,
|
||||||
longest_string_length,
|
longest_string_length,
|
||||||
ls_colors: LsColors::from_env().unwrap_or_default(),
|
ls_colors: LsColors::from_env().unwrap_or_default(),
|
||||||
};
|
};
|
||||||
@@ -269,7 +275,7 @@ fn clean_indentation_string(s: &str) -> String {
|
|||||||
is
|
is
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn get_printable_name<P: AsRef<Path>>(dir_name: &P, short_paths: bool) -> String {
|
fn get_printable_name<P: AsRef<Path>>(dir_name: &P, short_paths: bool) -> String {
|
||||||
let dir_name = dir_name.as_ref();
|
let dir_name = dir_name.as_ref();
|
||||||
let printable_name = {
|
let printable_name = {
|
||||||
if short_paths {
|
if short_paths {
|
||||||
@@ -336,8 +342,6 @@ pub fn format_string(
|
|||||||
if display_data.initial.is_screen_reader {
|
if display_data.initial.is_screen_reader {
|
||||||
// if screen_reader then bars is 'depth'
|
// if screen_reader then bars is 'depth'
|
||||||
format!("{pretty_name} {bars} {pretty_size}{percent}")
|
format!("{pretty_name} {bars} {pretty_size}{percent}")
|
||||||
} else if display_data.initial.by_filetime.is_some() {
|
|
||||||
format!("{pretty_size} {indent}{pretty_name}")
|
|
||||||
} else {
|
} else {
|
||||||
format!("{pretty_size} {indent} {pretty_name}{percent}")
|
format!("{pretty_size} {indent} {pretty_name}{percent}")
|
||||||
}
|
}
|
||||||
@@ -372,8 +376,6 @@ fn get_name_percent(
|
|||||||
fn get_pretty_size(node: &DisplayNode, is_biggest: bool, display_data: &DisplayData) -> String {
|
fn get_pretty_size(node: &DisplayNode, is_biggest: bool, display_data: &DisplayData) -> String {
|
||||||
let output = if display_data.initial.by_filecount {
|
let output = if display_data.initial.by_filecount {
|
||||||
node.size.separate_with_commas()
|
node.size.separate_with_commas()
|
||||||
} else if display_data.initial.by_filetime.is_some() {
|
|
||||||
get_pretty_file_modified_time(node.size as i64)
|
|
||||||
} else {
|
} else {
|
||||||
human_readable_number(node.size, &display_data.initial.output_format)
|
human_readable_number(node.size, &display_data.initial.output_format)
|
||||||
};
|
};
|
||||||
@@ -387,14 +389,6 @@ fn get_pretty_size(node: &DisplayNode, is_biggest: bool, display_data: &DisplayD
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
fn get_pretty_file_modified_time(timestamp: i64) -> String {
|
|
||||||
let datetime: DateTime<Utc> = Utc.timestamp_opt(timestamp, 0).unwrap();
|
|
||||||
|
|
||||||
let local_datetime = datetime.with_timezone(&Local);
|
|
||||||
|
|
||||||
local_datetime.format("%Y-%m-%dT%H:%M:%S").to_string()
|
|
||||||
}
|
|
||||||
|
|
||||||
fn get_pretty_name(
|
fn get_pretty_name(
|
||||||
node: &DisplayNode,
|
node: &DisplayNode,
|
||||||
name_and_padding: String,
|
name_and_padding: String,
|
||||||
@@ -417,14 +411,11 @@ fn get_pretty_name(
|
|||||||
|
|
||||||
// If we are working with SI units or not
|
// If we are working with SI units or not
|
||||||
pub fn get_type_of_thousand(output_str: &str) -> u64 {
|
pub fn get_type_of_thousand(output_str: &str) -> u64 {
|
||||||
if output_str.is_empty() {
|
let is_si = output_str.contains('i'); // si, KiB, MiB, etc
|
||||||
1024
|
if is_si {
|
||||||
} else if output_str == "si" {
|
|
||||||
1000
|
1000
|
||||||
} else if output_str.contains('i') || output_str.len() == 1 {
|
|
||||||
1024
|
|
||||||
} else {
|
} else {
|
||||||
1000
|
1024
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -475,7 +466,6 @@ mod tests {
|
|||||||
is_reversed: false,
|
is_reversed: false,
|
||||||
colors_on: false,
|
colors_on: false,
|
||||||
by_filecount: false,
|
by_filecount: false,
|
||||||
by_filetime: None,
|
|
||||||
is_screen_reader: false,
|
is_screen_reader: false,
|
||||||
output_format: "".into(),
|
output_format: "".into(),
|
||||||
bars_on_right: false,
|
bars_on_right: false,
|
||||||
@@ -554,14 +544,6 @@ mod tests {
|
|||||||
assert_eq!(human_readable_number(1024 * 1024 * 1024 - 1, ""), "1023M");
|
assert_eq!(human_readable_number(1024 * 1024 * 1024 - 1, ""), "1023M");
|
||||||
assert_eq!(human_readable_number(1024 * 1024 * 1024 * 20, ""), "20G");
|
assert_eq!(human_readable_number(1024 * 1024 * 1024 * 20, ""), "20G");
|
||||||
assert_eq!(human_readable_number(1024 * 1024 * 1024 * 1024, ""), "1.0T");
|
assert_eq!(human_readable_number(1024 * 1024 * 1024 * 1024, ""), "1.0T");
|
||||||
assert_eq!(
|
|
||||||
human_readable_number(1024 * 1024 * 1024 * 1024 * 234, ""),
|
|
||||||
"234T"
|
|
||||||
);
|
|
||||||
assert_eq!(
|
|
||||||
human_readable_number(1024 * 1024 * 1024 * 1024 * 1024, ""),
|
|
||||||
"1.0P"
|
|
||||||
);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
@@ -570,25 +552,23 @@ mod tests {
|
|||||||
assert_eq!(human_readable_number(1024 * 100, "si"), "102K");
|
assert_eq!(human_readable_number(1024 * 100, "si"), "102K");
|
||||||
}
|
}
|
||||||
|
|
||||||
// Refer to https://en.wikipedia.org/wiki/Byte#Multiple-byte_units
|
|
||||||
#[test]
|
#[test]
|
||||||
fn test_human_readable_number_kb() {
|
fn test_human_readable_number_kb() {
|
||||||
let hrn = human_readable_number;
|
let hrn = human_readable_number;
|
||||||
assert_eq!(hrn(1023, "b"), "1023B");
|
assert_eq!(hrn(1023, "b"), "1023B");
|
||||||
assert_eq!(hrn(1000 * 1000, "bytes"), "1000000B");
|
assert_eq!(hrn(1000 * 1000, "bytes"), "1000000B");
|
||||||
assert_eq!(hrn(1023, "kb"), "1K");
|
assert_eq!(hrn(1023, "kb"), "0K");
|
||||||
assert_eq!(hrn(1023, "k"), "0K");
|
assert_eq!(hrn(1023, "kib"), "1K");
|
||||||
assert_eq!(hrn(1023, "kib"), "0K");
|
assert_eq!(hrn(1024, "kb"), "1K");
|
||||||
assert_eq!(hrn(1024, "kib"), "1K");
|
assert_eq!(hrn(1024 * 512, "kb"), "512K");
|
||||||
assert_eq!(hrn(1024 * 512, "kib"), "512K");
|
assert_eq!(hrn(1024 * 1024, "kb"), "1024K");
|
||||||
assert_eq!(hrn(1024 * 1024, "kib"), "1024K");
|
assert_eq!(hrn(1024 * 1000 * 1000 * 20, "kb"), "20000000K");
|
||||||
assert_eq!(hrn(1024 * 1000 * 1000 * 20, "kib"), "20000000K");
|
assert_eq!(hrn(1024 * 1024 * 1000 * 20, "mb"), "20000M");
|
||||||
assert_eq!(hrn(1024 * 1024 * 1000 * 20, "mib"), "20000M");
|
assert_eq!(hrn(1024 * 1024 * 1024 * 20, "gb"), "20G");
|
||||||
assert_eq!(hrn(1024 * 1024 * 1024 * 20, "gib"), "20G");
|
|
||||||
}
|
}
|
||||||
|
|
||||||
#[cfg(test)]
|
#[cfg(test)]
|
||||||
fn build_draw_data(disp: &DisplayData, size: u32) -> (DrawData<'_>, DisplayNode) {
|
fn build_draw_data<'a>(disp: &'a DisplayData, size: u32) -> (DrawData<'a>, DisplayNode) {
|
||||||
let n = DisplayNode {
|
let n = DisplayNode {
|
||||||
name: PathBuf::from("/short"),
|
name: PathBuf::from("/short"),
|
||||||
size: 2_u64.pow(size),
|
size: 2_u64.pow(size),
|
||||||
@@ -640,37 +620,4 @@ mod tests {
|
|||||||
let bar = dd.generate_bar(&n, 5);
|
let bar = dd.generate_bar(&n, 5);
|
||||||
assert_eq!(bar, "████▓▓▓▓▓▓▓▓▓");
|
assert_eq!(bar, "████▓▓▓▓▓▓▓▓▓");
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
|
||||||
fn test_get_pretty_file_modified_time() {
|
|
||||||
// Create a timestamp for 2023-07-12 00:00:00 in local time
|
|
||||||
let local_dt = Local.with_ymd_and_hms(2023, 7, 12, 0, 0, 0).unwrap();
|
|
||||||
let timestamp = local_dt.timestamp();
|
|
||||||
|
|
||||||
// Format expected output
|
|
||||||
let expected_output = local_dt.format("%Y-%m-%dT%H:%M:%S").to_string();
|
|
||||||
|
|
||||||
assert_eq!(get_pretty_file_modified_time(timestamp), expected_output);
|
|
||||||
|
|
||||||
// Test another timestamp
|
|
||||||
let local_dt = Local.with_ymd_and_hms(2020, 1, 1, 12, 0, 0).unwrap();
|
|
||||||
let timestamp = local_dt.timestamp();
|
|
||||||
let expected_output = local_dt.format("%Y-%m-%dT%H:%M:%S").to_string();
|
|
||||||
|
|
||||||
assert_eq!(get_pretty_file_modified_time(timestamp), expected_output);
|
|
||||||
|
|
||||||
// Test timestamp for epoch start (1970-01-01T00:00:00)
|
|
||||||
let local_dt = Local.with_ymd_and_hms(1970, 1, 1, 0, 0, 0).unwrap();
|
|
||||||
let timestamp = local_dt.timestamp();
|
|
||||||
let expected_output = local_dt.format("%Y-%m-%dT%H:%M:%S").to_string();
|
|
||||||
|
|
||||||
assert_eq!(get_pretty_file_modified_time(timestamp), expected_output);
|
|
||||||
|
|
||||||
// Test a future timestamp
|
|
||||||
let local_dt = Local.with_ymd_and_hms(2030, 12, 25, 6, 30, 0).unwrap();
|
|
||||||
let timestamp = local_dt.timestamp();
|
|
||||||
let expected_output = local_dt.format("%Y-%m-%dT%H:%M:%S").to_string();
|
|
||||||
|
|
||||||
assert_eq!(get_pretty_file_modified_time(timestamp), expected_output);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,11 +1,5 @@
|
|||||||
use std::cell::RefCell;
|
|
||||||
use std::path::PathBuf;
|
use std::path::PathBuf;
|
||||||
|
|
||||||
use serde::ser::SerializeStruct;
|
|
||||||
use serde::{Serialize, Serializer};
|
|
||||||
|
|
||||||
use crate::display::human_readable_number;
|
|
||||||
|
|
||||||
#[derive(Debug, PartialEq, Eq, PartialOrd, Ord, Clone)]
|
#[derive(Debug, PartialEq, Eq, PartialOrd, Ord, Clone)]
|
||||||
pub struct DisplayNode {
|
pub struct DisplayNode {
|
||||||
// Note: the order of fields in important here, for PartialEq and PartialOrd
|
// Note: the order of fields in important here, for PartialEq and PartialOrd
|
||||||
@@ -29,30 +23,3 @@ impl DisplayNode {
|
|||||||
out
|
out
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Only used for -j 'json' flag combined with -o 'output_type' flag
|
|
||||||
// Used to pass the output_type into the custom Serde serializer
|
|
||||||
thread_local! {
|
|
||||||
pub static OUTPUT_TYPE: RefCell<String> = const { RefCell::new(String::new()) };
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
|
||||||
We need the custom Serialize incase someone uses the -o flag to pass a custom output type in
|
|
||||||
(show size in Mb / Gb etc).
|
|
||||||
Sadly this also necessitates a global variable OUTPUT_TYPE as we can not pass the output_type flag
|
|
||||||
into the serialize method
|
|
||||||
*/
|
|
||||||
impl Serialize for DisplayNode {
|
|
||||||
fn serialize<S>(&self, serializer: S) -> Result<S::Ok, S::Error>
|
|
||||||
where
|
|
||||||
S: Serializer,
|
|
||||||
{
|
|
||||||
let readable_size = OUTPUT_TYPE
|
|
||||||
.with(|output_type| human_readable_number(self.size, output_type.borrow().as_str()));
|
|
||||||
let mut state = serializer.serialize_struct("DisplayNode", 2)?;
|
|
||||||
state.serialize_field("size", &(readable_size))?;
|
|
||||||
state.serialize_field("name", &self.name)?;
|
|
||||||
state.serialize_field("children", &self.children)?;
|
|
||||||
state.end()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|||||||
103
src/filter.rs
103
src/filter.rs
@@ -1,12 +1,7 @@
|
|||||||
use stfu8::encode_u8;
|
|
||||||
|
|
||||||
use crate::display::get_printable_name;
|
|
||||||
use crate::display_node::DisplayNode;
|
use crate::display_node::DisplayNode;
|
||||||
use crate::node::FileTime;
|
|
||||||
use crate::node::Node;
|
use crate::node::Node;
|
||||||
use std::collections::BinaryHeap;
|
use std::collections::BinaryHeap;
|
||||||
use std::collections::HashMap;
|
use std::collections::HashMap;
|
||||||
use std::collections::HashSet;
|
|
||||||
use std::path::Path;
|
use std::path::Path;
|
||||||
use std::path::PathBuf;
|
use std::path::PathBuf;
|
||||||
|
|
||||||
@@ -17,15 +12,9 @@ pub struct AggregateData {
|
|||||||
pub number_of_lines: usize,
|
pub number_of_lines: usize,
|
||||||
pub depth: usize,
|
pub depth: usize,
|
||||||
pub using_a_filter: bool,
|
pub using_a_filter: bool,
|
||||||
pub short_paths: bool,
|
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn get_biggest(
|
pub fn get_biggest(top_level_nodes: Vec<Node>, display_data: AggregateData) -> Option<DisplayNode> {
|
||||||
top_level_nodes: Vec<Node>,
|
|
||||||
display_data: AggregateData,
|
|
||||||
by_filetime: &Option<FileTime>,
|
|
||||||
keep_collapsed: HashSet<PathBuf>,
|
|
||||||
) -> Option<DisplayNode> {
|
|
||||||
if top_level_nodes.is_empty() {
|
if top_level_nodes.is_empty() {
|
||||||
// perhaps change this, bring back Error object?
|
// perhaps change this, bring back Error object?
|
||||||
return None;
|
return None;
|
||||||
@@ -35,26 +24,14 @@ pub fn get_biggest(
|
|||||||
let root;
|
let root;
|
||||||
|
|
||||||
if number_top_level_nodes > 1 {
|
if number_top_level_nodes > 1 {
|
||||||
let size = if by_filetime.is_some() {
|
let size = top_level_nodes.iter().map(|node| node.size).sum();
|
||||||
top_level_nodes
|
|
||||||
.iter()
|
|
||||||
.map(|node| node.size)
|
|
||||||
.max()
|
|
||||||
.unwrap_or(0)
|
|
||||||
} else {
|
|
||||||
top_level_nodes.iter().map(|node| node.size).sum()
|
|
||||||
};
|
|
||||||
|
|
||||||
let nodes = handle_duplicate_top_level_names(top_level_nodes, display_data.short_paths);
|
|
||||||
|
|
||||||
root = Node {
|
root = Node {
|
||||||
name: PathBuf::from("(total)"),
|
name: PathBuf::from("(total)"),
|
||||||
size,
|
size,
|
||||||
children: nodes,
|
children: top_level_nodes,
|
||||||
inode_device: None,
|
inode_device: None,
|
||||||
depth: 0,
|
depth: 0,
|
||||||
};
|
};
|
||||||
|
|
||||||
// Always include the base nodes if we add a 'parent' (total) node
|
// Always include the base nodes if we add a 'parent' (total) node
|
||||||
heap = always_add_children(&display_data, &root, heap);
|
heap = always_add_children(&display_data, &root, heap);
|
||||||
} else {
|
} else {
|
||||||
@@ -62,19 +39,13 @@ pub fn get_biggest(
|
|||||||
heap = add_children(&display_data, &root, heap);
|
heap = add_children(&display_data, &root, heap);
|
||||||
}
|
}
|
||||||
|
|
||||||
Some(fill_remaining_lines(
|
Some(fill_remaining_lines(heap, &root, display_data))
|
||||||
heap,
|
|
||||||
&root,
|
|
||||||
display_data,
|
|
||||||
keep_collapsed,
|
|
||||||
))
|
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn fill_remaining_lines<'a>(
|
pub fn fill_remaining_lines<'a>(
|
||||||
mut heap: BinaryHeap<&'a Node>,
|
mut heap: BinaryHeap<&'a Node>,
|
||||||
root: &'a Node,
|
root: &'a Node,
|
||||||
display_data: AggregateData,
|
display_data: AggregateData,
|
||||||
keep_collapsed: HashSet<PathBuf>,
|
|
||||||
) -> DisplayNode {
|
) -> DisplayNode {
|
||||||
let mut allowed_nodes = HashMap::new();
|
let mut allowed_nodes = HashMap::new();
|
||||||
|
|
||||||
@@ -82,14 +53,10 @@ pub fn fill_remaining_lines<'a>(
|
|||||||
let line = heap.pop();
|
let line = heap.pop();
|
||||||
match line {
|
match line {
|
||||||
Some(line) => {
|
Some(line) => {
|
||||||
// If we are not doing only_file OR if we are doing
|
|
||||||
// only_file and it has no children (ie is a file not a dir)
|
|
||||||
if !display_data.only_file || line.children.is_empty() {
|
if !display_data.only_file || line.children.is_empty() {
|
||||||
allowed_nodes.insert(line.name.as_path(), line);
|
allowed_nodes.insert(line.name.as_path(), line);
|
||||||
}
|
}
|
||||||
if !keep_collapsed.contains(&line.name) {
|
heap = add_children(&display_data, line, heap);
|
||||||
heap = add_children(&display_data, line, heap);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
None => break,
|
None => break,
|
||||||
}
|
}
|
||||||
@@ -147,7 +114,7 @@ fn recursive_rebuilder(allowed_nodes: &HashMap<&Path, &Node>, current: &Node) ->
|
|||||||
.map(|c| recursive_rebuilder(allowed_nodes, c))
|
.map(|c| recursive_rebuilder(allowed_nodes, c))
|
||||||
.collect();
|
.collect();
|
||||||
|
|
||||||
build_display_node(new_children, current)
|
build_node(new_children, current)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Applies all allowed nodes as children to current node
|
// Applies all allowed nodes as children to current node
|
||||||
@@ -160,10 +127,10 @@ fn flat_rebuilder(allowed_nodes: HashMap<&Path, &Node>, current: &Node) -> Displ
|
|||||||
children: vec![],
|
children: vec![],
|
||||||
})
|
})
|
||||||
.collect::<Vec<DisplayNode>>();
|
.collect::<Vec<DisplayNode>>();
|
||||||
build_display_node(new_children, current)
|
build_node(new_children, current)
|
||||||
}
|
}
|
||||||
|
|
||||||
fn build_display_node(mut new_children: Vec<DisplayNode>, current: &Node) -> DisplayNode {
|
fn build_node(mut new_children: Vec<DisplayNode>, current: &Node) -> DisplayNode {
|
||||||
new_children.sort_by(|lhs, rhs| lhs.cmp(rhs).reverse());
|
new_children.sort_by(|lhs, rhs| lhs.cmp(rhs).reverse());
|
||||||
DisplayNode {
|
DisplayNode {
|
||||||
name: current.name.clone(),
|
name: current.name.clone(),
|
||||||
@@ -171,57 +138,3 @@ fn build_display_node(mut new_children: Vec<DisplayNode>, current: &Node) -> Dis
|
|||||||
children: new_children,
|
children: new_children,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
fn names_have_dup(top_level_nodes: &Vec<Node>) -> bool {
|
|
||||||
let mut stored = HashSet::new();
|
|
||||||
for node in top_level_nodes {
|
|
||||||
let name = get_printable_name(&node.name, true);
|
|
||||||
if stored.contains(&name) {
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
stored.insert(name);
|
|
||||||
}
|
|
||||||
false
|
|
||||||
}
|
|
||||||
|
|
||||||
fn handle_duplicate_top_level_names(top_level_nodes: Vec<Node>, short_paths: bool) -> Vec<Node> {
|
|
||||||
// If we have top level names that are the same - we need to tweak them:
|
|
||||||
if short_paths && names_have_dup(&top_level_nodes) {
|
|
||||||
let mut new_top_nodes = top_level_nodes.clone();
|
|
||||||
let mut dir_walk_up_count = 0;
|
|
||||||
|
|
||||||
while names_have_dup(&new_top_nodes) && dir_walk_up_count < 10 {
|
|
||||||
dir_walk_up_count += 1;
|
|
||||||
let mut newer = vec![];
|
|
||||||
|
|
||||||
for node in new_top_nodes.iter() {
|
|
||||||
let mut folders = node.name.iter().rev();
|
|
||||||
// Get parent folder (if second time round get grandparent and so on)
|
|
||||||
for _ in 0..dir_walk_up_count {
|
|
||||||
folders.next();
|
|
||||||
}
|
|
||||||
match folders.next() {
|
|
||||||
// Add (parent_name) to path of Node
|
|
||||||
Some(data) => {
|
|
||||||
let parent = encode_u8(data.as_encoded_bytes());
|
|
||||||
let current_node = node.name.display();
|
|
||||||
let n = Node {
|
|
||||||
name: PathBuf::from(format!("{current_node}({parent})")),
|
|
||||||
size: node.size,
|
|
||||||
children: node.children.clone(),
|
|
||||||
inode_device: node.inode_device,
|
|
||||||
depth: node.depth,
|
|
||||||
};
|
|
||||||
newer.push(n)
|
|
||||||
}
|
|
||||||
// Node does not have a parent
|
|
||||||
None => newer.push(node.clone()),
|
|
||||||
}
|
|
||||||
}
|
|
||||||
new_top_nodes = newer;
|
|
||||||
}
|
|
||||||
new_top_nodes
|
|
||||||
} else {
|
|
||||||
top_level_nodes
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|||||||
@@ -1,5 +1,4 @@
|
|||||||
use crate::display_node::DisplayNode;
|
use crate::display_node::DisplayNode;
|
||||||
use crate::node::FileTime;
|
|
||||||
use crate::node::Node;
|
use crate::node::Node;
|
||||||
use std::collections::HashMap;
|
use std::collections::HashMap;
|
||||||
use std::ffi::OsStr;
|
use std::ffi::OsStr;
|
||||||
@@ -11,11 +10,7 @@ struct ExtensionNode<'a> {
|
|||||||
extension: Option<&'a OsStr>,
|
extension: Option<&'a OsStr>,
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn get_all_file_types(
|
pub fn get_all_file_types(top_level_nodes: &[Node], n: usize) -> Option<DisplayNode> {
|
||||||
top_level_nodes: &[Node],
|
|
||||||
n: usize,
|
|
||||||
by_filetime: &Option<FileTime>,
|
|
||||||
) -> Option<DisplayNode> {
|
|
||||||
let ext_nodes = {
|
let ext_nodes = {
|
||||||
let mut extension_cumulative_sizes = HashMap::new();
|
let mut extension_cumulative_sizes = HashMap::new();
|
||||||
build_by_all_file_types(top_level_nodes, &mut extension_cumulative_sizes);
|
build_by_all_file_types(top_level_nodes, &mut extension_cumulative_sizes);
|
||||||
@@ -49,27 +44,16 @@ pub fn get_all_file_types(
|
|||||||
|
|
||||||
// ...then, aggregate the remaining nodes (if any) into a single "(others)" node
|
// ...then, aggregate the remaining nodes (if any) into a single "(others)" node
|
||||||
if ext_nodes_iter.len() > 0 {
|
if ext_nodes_iter.len() > 0 {
|
||||||
let actual_size = if by_filetime.is_some() {
|
|
||||||
ext_nodes_iter.map(|node| node.size).max().unwrap_or(0)
|
|
||||||
} else {
|
|
||||||
ext_nodes_iter.map(|node| node.size).sum()
|
|
||||||
};
|
|
||||||
displayed.push(DisplayNode {
|
displayed.push(DisplayNode {
|
||||||
name: PathBuf::from("(others)"),
|
name: PathBuf::from("(others)"),
|
||||||
size: actual_size,
|
size: ext_nodes_iter.map(|node| node.size).sum(),
|
||||||
children: vec![],
|
children: vec![],
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
let actual_size: u64 = if by_filetime.is_some() {
|
|
||||||
displayed.iter().map(|node| node.size).max().unwrap_or(0)
|
|
||||||
} else {
|
|
||||||
displayed.iter().map(|node| node.size).sum()
|
|
||||||
};
|
|
||||||
|
|
||||||
let result = DisplayNode {
|
let result = DisplayNode {
|
||||||
name: PathBuf::from("(total)"),
|
name: PathBuf::from("(total)"),
|
||||||
size: actual_size,
|
size: displayed.iter().map(|node| node.size).sum(),
|
||||||
children: displayed,
|
children: displayed,
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|||||||
204
src/main.rs
204
src/main.rs
@@ -21,24 +21,21 @@ use regex::Error;
|
|||||||
use std::collections::HashSet;
|
use std::collections::HashSet;
|
||||||
use std::env;
|
use std::env;
|
||||||
use std::fs::read_to_string;
|
use std::fs::read_to_string;
|
||||||
use std::io;
|
|
||||||
use std::panic;
|
use std::panic;
|
||||||
use std::process;
|
use std::process;
|
||||||
use std::sync::Arc;
|
use std::sync::Arc;
|
||||||
use std::sync::Mutex;
|
use std::sync::Mutex;
|
||||||
use sysinfo::{System, SystemExt};
|
use sysinfo::{System, SystemExt};
|
||||||
use utils::canonicalize_absolute_path;
|
|
||||||
|
|
||||||
use self::display::draw_it;
|
use self::display::draw_it;
|
||||||
use config::get_config;
|
use config::get_config;
|
||||||
use dir_walker::walk_it;
|
use dir_walker::walk_it;
|
||||||
use display_node::OUTPUT_TYPE;
|
|
||||||
use filter::get_biggest;
|
use filter::get_biggest;
|
||||||
use filter_type::get_all_file_types;
|
use filter_type::get_all_file_types;
|
||||||
use regex::Regex;
|
use regex::Regex;
|
||||||
use std::cmp::max;
|
use std::cmp::max;
|
||||||
use std::path::PathBuf;
|
use std::path::PathBuf;
|
||||||
use terminal_size::{Height, Width, terminal_size};
|
use terminal_size::{terminal_size, Height, Width};
|
||||||
use utils::get_filesystem_devices;
|
use utils::get_filesystem_devices;
|
||||||
use utils::simplify_dir_names;
|
use utils::simplify_dir_names;
|
||||||
|
|
||||||
@@ -81,19 +78,23 @@ fn should_init_color(no_color: bool, force_color: bool) -> bool {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn get_height_of_terminal() -> usize {
|
fn get_height_of_terminal() -> usize {
|
||||||
|
// Simplify once https://github.com/eminence/terminal-size/pull/41 is
|
||||||
|
// merged
|
||||||
terminal_size()
|
terminal_size()
|
||||||
// Windows CI runners detect a terminal height of 0
|
// Windows CI runners detect a terminal height of 0
|
||||||
.map(|(_, Height(h))| max(h.into(), DEFAULT_NUMBER_OF_LINES))
|
.map(|(_, Height(h))| max(h as usize, DEFAULT_NUMBER_OF_LINES))
|
||||||
.unwrap_or(DEFAULT_NUMBER_OF_LINES)
|
.unwrap_or(DEFAULT_NUMBER_OF_LINES)
|
||||||
- 10
|
- 10
|
||||||
}
|
}
|
||||||
|
|
||||||
fn get_width_of_terminal() -> usize {
|
fn get_width_of_terminal() -> usize {
|
||||||
|
// Simplify once https://github.com/eminence/terminal-size/pull/41 is
|
||||||
|
// merged
|
||||||
terminal_size()
|
terminal_size()
|
||||||
.map(|(Width(w), _)| match cfg!(windows) {
|
.map(|(Width(w), _)| match cfg!(windows) {
|
||||||
// Windows CI runners detect a very low terminal width
|
// Windows CI runners detect a very low terminal width
|
||||||
true => max(w.into(), DEFAULT_TERMINAL_WIDTH),
|
true => max(w as usize, DEFAULT_TERMINAL_WIDTH),
|
||||||
false => w.into(),
|
false => w as usize,
|
||||||
})
|
})
|
||||||
.unwrap_or(DEFAULT_TERMINAL_WIDTH)
|
.unwrap_or(DEFAULT_TERMINAL_WIDTH)
|
||||||
}
|
}
|
||||||
@@ -112,46 +113,22 @@ fn get_regex_value(maybe_value: Option<ValuesRef<String>>) -> Vec<Regex> {
|
|||||||
|
|
||||||
fn main() {
|
fn main() {
|
||||||
let options = build_cli().get_matches();
|
let options = build_cli().get_matches();
|
||||||
let config = get_config(options.get_one::<String>("config").cloned());
|
let config = get_config();
|
||||||
|
|
||||||
let errors = RuntimeErrors::default();
|
let errors = RuntimeErrors::default();
|
||||||
let error_listen_for_ctrlc = Arc::new(Mutex::new(errors));
|
let error_listen_for_ctrlc = Arc::new(Mutex::new(errors));
|
||||||
let errors_for_rayon = error_listen_for_ctrlc.clone();
|
let errors_for_rayon = error_listen_for_ctrlc.clone();
|
||||||
|
let errors_final = error_listen_for_ctrlc.clone();
|
||||||
|
|
||||||
ctrlc::set_handler(move || {
|
ctrlc::set_handler(move || {
|
||||||
|
error_listen_for_ctrlc.lock().unwrap().abort = true;
|
||||||
println!("\nAborting");
|
println!("\nAborting");
|
||||||
process::exit(1);
|
|
||||||
})
|
})
|
||||||
.expect("Error setting Ctrl-C handler");
|
.expect("Error setting Ctrl-C handler");
|
||||||
|
|
||||||
let target_dirs = match config.get_files_from(&options) {
|
let target_dirs = match options.get_many::<String>("params") {
|
||||||
Some(path) => {
|
Some(values) => values.map(|v| v.as_str()).collect::<Vec<&str>>(),
|
||||||
if path == "-" {
|
None => vec!["."],
|
||||||
let mut targets_to_add = io::stdin()
|
|
||||||
.lines()
|
|
||||||
.map_while(Result::ok)
|
|
||||||
.collect::<Vec<String>>();
|
|
||||||
|
|
||||||
if targets_to_add.is_empty() {
|
|
||||||
eprintln!("No input provided, defaulting to current directory");
|
|
||||||
targets_to_add.push(".".to_owned());
|
|
||||||
}
|
|
||||||
targets_to_add
|
|
||||||
} else {
|
|
||||||
// read file
|
|
||||||
match read_to_string(path) {
|
|
||||||
Ok(file_content) => file_content.lines().map(|x| x.to_string()).collect(),
|
|
||||||
Err(e) => {
|
|
||||||
eprintln!("Error reading file: {e}");
|
|
||||||
vec![".".to_owned()]
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
None => match options.get_many::<String>("params") {
|
|
||||||
Some(values) => values.cloned().collect(),
|
|
||||||
None => vec![".".to_owned()],
|
|
||||||
},
|
|
||||||
};
|
};
|
||||||
|
|
||||||
let summarize_file_types = options.get_flag("types");
|
let summarize_file_types = options.get_flag("types");
|
||||||
@@ -189,7 +166,6 @@ fn main() {
|
|||||||
Some(values) => values
|
Some(values) => values
|
||||||
.map(|v| v.as_str())
|
.map(|v| v.as_str())
|
||||||
.map(PathBuf::from)
|
.map(PathBuf::from)
|
||||||
.map(canonicalize_absolute_path)
|
|
||||||
.collect::<Vec<PathBuf>>(),
|
.collect::<Vec<PathBuf>>(),
|
||||||
None => vec![],
|
None => vec![],
|
||||||
};
|
};
|
||||||
@@ -213,14 +189,13 @@ fn main() {
|
|||||||
.collect::<Vec<Regex>>();
|
.collect::<Vec<Regex>>();
|
||||||
|
|
||||||
let by_filecount = options.get_flag("by_filecount");
|
let by_filecount = options.get_flag("by_filecount");
|
||||||
let by_filetime = config.get_filetime(&options);
|
|
||||||
let limit_filesystem = options.get_flag("limit_filesystem");
|
let limit_filesystem = options.get_flag("limit_filesystem");
|
||||||
let follow_links = options.get_flag("dereference_links");
|
let follow_links = options.get_flag("dereference_links");
|
||||||
|
|
||||||
|
let simplified_dirs = simplify_dir_names(target_dirs);
|
||||||
let allowed_filesystems = limit_filesystem
|
let allowed_filesystems = limit_filesystem
|
||||||
.then(|| get_filesystem_devices(&target_dirs, follow_links))
|
.then(|| get_filesystem_devices(simplified_dirs.iter()))
|
||||||
.unwrap_or_default();
|
.unwrap_or_default();
|
||||||
let simplified_dirs = simplify_dir_names(&target_dirs);
|
|
||||||
|
|
||||||
let ignored_full_path: HashSet<PathBuf> = ignore_directories
|
let ignored_full_path: HashSet<PathBuf> = ignore_directories
|
||||||
.into_iter()
|
.into_iter()
|
||||||
@@ -236,47 +211,25 @@ fn main() {
|
|||||||
indicator.spawn(output_format.clone())
|
indicator.spawn(output_format.clone())
|
||||||
}
|
}
|
||||||
|
|
||||||
let keep_collapsed: HashSet<PathBuf> = match options.get_many::<String>("collapse") {
|
|
||||||
Some(collapse) => {
|
|
||||||
let mut combined_dirs = HashSet::new();
|
|
||||||
for collapse_dir in collapse {
|
|
||||||
for target_dir in target_dirs.iter() {
|
|
||||||
combined_dirs.insert(PathBuf::from(target_dir).join(collapse_dir));
|
|
||||||
}
|
|
||||||
}
|
|
||||||
combined_dirs
|
|
||||||
}
|
|
||||||
None => HashSet::new(),
|
|
||||||
};
|
|
||||||
|
|
||||||
let filter_modified_time = config.get_modified_time_operator(&options);
|
|
||||||
let filter_accessed_time = config.get_accessed_time_operator(&options);
|
|
||||||
let filter_changed_time = config.get_changed_time_operator(&options);
|
|
||||||
|
|
||||||
let walk_data = WalkData {
|
let walk_data = WalkData {
|
||||||
ignore_directories: ignored_full_path,
|
ignore_directories: ignored_full_path,
|
||||||
filter_regex: &filter_regexs,
|
filter_regex: &filter_regexs,
|
||||||
invert_filter_regex: &invert_filter_regexs,
|
invert_filter_regex: &invert_filter_regexs,
|
||||||
allowed_filesystems,
|
allowed_filesystems,
|
||||||
filter_modified_time,
|
|
||||||
filter_accessed_time,
|
|
||||||
filter_changed_time,
|
|
||||||
use_apparent_size: config.get_apparent_size(&options),
|
use_apparent_size: config.get_apparent_size(&options),
|
||||||
by_filecount,
|
by_filecount,
|
||||||
by_filetime: &by_filetime,
|
|
||||||
ignore_hidden,
|
ignore_hidden,
|
||||||
follow_links,
|
follow_links,
|
||||||
progress_data: indicator.data.clone(),
|
progress_data: indicator.data.clone(),
|
||||||
errors: errors_for_rayon,
|
errors: errors_for_rayon,
|
||||||
};
|
};
|
||||||
let threads_to_use = config.get_threads(&options);
|
|
||||||
let stack_size = config.get_custom_stack_size(&options);
|
let stack_size = config.get_custom_stack_size(&options);
|
||||||
init_rayon(&stack_size, &threads_to_use);
|
init_rayon(&stack_size);
|
||||||
|
|
||||||
let top_level_nodes = walk_it(simplified_dirs, &walk_data);
|
let top_level_nodes = walk_it(simplified_dirs, &walk_data);
|
||||||
|
|
||||||
let tree = match summarize_file_types {
|
let tree = match summarize_file_types {
|
||||||
true => get_all_file_types(&top_level_nodes, number_of_lines, &by_filetime),
|
true => get_all_file_types(&top_level_nodes, number_of_lines),
|
||||||
false => {
|
false => {
|
||||||
let agg_data = AggregateData {
|
let agg_data = AggregateData {
|
||||||
min_size: config.get_min_size(&options),
|
min_size: config.get_min_size(&options),
|
||||||
@@ -285,16 +238,20 @@ fn main() {
|
|||||||
number_of_lines,
|
number_of_lines,
|
||||||
depth,
|
depth,
|
||||||
using_a_filter: !filter_regexs.is_empty() || !invert_filter_regexs.is_empty(),
|
using_a_filter: !filter_regexs.is_empty() || !invert_filter_regexs.is_empty(),
|
||||||
short_paths: !config.get_full_paths(&options),
|
|
||||||
};
|
};
|
||||||
get_biggest(top_level_nodes, agg_data, &by_filetime, keep_collapsed)
|
get_biggest(top_level_nodes, agg_data)
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
|
||||||
// Must have stopped indicator before we print to stderr
|
// Must have stopped indicator before we print to stderr
|
||||||
indicator.stop();
|
indicator.stop();
|
||||||
|
|
||||||
|
if errors_final.lock().unwrap().abort {
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
let final_errors = walk_data.errors.lock().unwrap();
|
let final_errors = walk_data.errors.lock().unwrap();
|
||||||
|
let failed_permissions = final_errors.no_permissions;
|
||||||
if !final_errors.file_not_found.is_empty() {
|
if !final_errors.file_not_found.is_empty() {
|
||||||
let err = final_errors
|
let err = final_errors
|
||||||
.file_not_found
|
.file_not_found
|
||||||
@@ -304,20 +261,8 @@ fn main() {
|
|||||||
.join(", ");
|
.join(", ");
|
||||||
eprintln!("No such file or directory: {}", err);
|
eprintln!("No such file or directory: {}", err);
|
||||||
}
|
}
|
||||||
if !final_errors.no_permissions.is_empty() {
|
if failed_permissions {
|
||||||
if config.get_print_errors(&options) {
|
eprintln!("Did not have permissions for all directories");
|
||||||
let err = final_errors
|
|
||||||
.no_permissions
|
|
||||||
.iter()
|
|
||||||
.map(|a| a.as_ref())
|
|
||||||
.collect::<Vec<&str>>()
|
|
||||||
.join(", ");
|
|
||||||
eprintln!("Did not have permissions for directories: {}", err);
|
|
||||||
} else {
|
|
||||||
eprintln!(
|
|
||||||
"Did not have permissions for all directories (add --print-errors to see errors)"
|
|
||||||
);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
if !final_errors.unknown_error.is_empty() {
|
if !final_errors.unknown_error.is_empty() {
|
||||||
let err = final_errors
|
let err = final_errors
|
||||||
@@ -330,72 +275,53 @@ fn main() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
if let Some(root_node) = tree {
|
if let Some(root_node) = tree {
|
||||||
if config.get_output_json(&options) {
|
let idd = InitialDisplayData {
|
||||||
OUTPUT_TYPE.with(|wrapped| {
|
short_paths: !config.get_full_paths(&options),
|
||||||
wrapped.replace(output_format);
|
is_reversed: !config.get_reverse(&options),
|
||||||
});
|
colors_on: is_colors,
|
||||||
println!("{}", serde_json::to_string(&root_node).unwrap());
|
by_filecount,
|
||||||
} else {
|
is_screen_reader: config.get_screen_reader(&options),
|
||||||
let idd = InitialDisplayData {
|
output_format,
|
||||||
short_paths: !config.get_full_paths(&options),
|
bars_on_right: config.get_bars_on_right(&options),
|
||||||
is_reversed: !config.get_reverse(&options),
|
};
|
||||||
colors_on: is_colors,
|
draw_it(
|
||||||
by_filecount,
|
idd,
|
||||||
by_filetime,
|
config.get_no_bars(&options),
|
||||||
is_screen_reader: config.get_screen_reader(&options),
|
terminal_width,
|
||||||
output_format,
|
&root_node,
|
||||||
bars_on_right: config.get_bars_on_right(&options),
|
config.get_skip_total(&options),
|
||||||
};
|
)
|
||||||
|
|
||||||
draw_it(
|
|
||||||
idd,
|
|
||||||
config.get_no_bars(&options),
|
|
||||||
terminal_width,
|
|
||||||
&root_node,
|
|
||||||
config.get_skip_total(&options),
|
|
||||||
)
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
fn init_rayon(stack_size: &Option<usize>, threads: &Option<usize>) {
|
fn init_rayon(stack_size: &Option<usize>) {
|
||||||
// Rayon seems to raise this error on 32-bit builds
|
// Rayon seems to raise this error on 32-bit builds
|
||||||
// The global thread pool has not been initialized.: ThreadPoolBuildError { kind: GlobalPoolAlreadyInitialized }
|
// The global thread pool has not been initialized.: ThreadPoolBuildError { kind: GlobalPoolAlreadyInitialized }
|
||||||
if cfg!(target_pointer_width = "64") {
|
if cfg!(target_pointer_width = "64") {
|
||||||
let result = panic::catch_unwind(|| build_thread_pool(*stack_size, *threads));
|
let result = panic::catch_unwind(|| {
|
||||||
|
match stack_size {
|
||||||
|
Some(n) => rayon::ThreadPoolBuilder::new()
|
||||||
|
.stack_size(*n)
|
||||||
|
.build_global(),
|
||||||
|
None => {
|
||||||
|
let large_stack = usize::pow(1024, 3);
|
||||||
|
let mut s = System::new();
|
||||||
|
s.refresh_memory();
|
||||||
|
let available = s.available_memory();
|
||||||
|
|
||||||
|
if available > large_stack.try_into().unwrap() {
|
||||||
|
// Larger stack size to handle cases with lots of nested directories
|
||||||
|
rayon::ThreadPoolBuilder::new()
|
||||||
|
.stack_size(large_stack)
|
||||||
|
.build_global()
|
||||||
|
} else {
|
||||||
|
rayon::ThreadPoolBuilder::new().build_global()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
});
|
||||||
if result.is_err() {
|
if result.is_err() {
|
||||||
eprintln!("Problem initializing rayon, try: export RAYON_NUM_THREADS=1")
|
eprintln!("Problem initializing rayon, try: export RAYON_NUM_THREADS=1")
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
fn build_thread_pool(
|
|
||||||
stack: Option<usize>,
|
|
||||||
threads: Option<usize>,
|
|
||||||
) -> Result<(), rayon::ThreadPoolBuildError> {
|
|
||||||
let mut pool = rayon::ThreadPoolBuilder::new();
|
|
||||||
|
|
||||||
if let Some(thread_count) = threads {
|
|
||||||
pool = pool.num_threads(thread_count);
|
|
||||||
}
|
|
||||||
|
|
||||||
let stack_size = match stack {
|
|
||||||
Some(s) => Some(s),
|
|
||||||
None => {
|
|
||||||
let large_stack = usize::pow(1024, 3);
|
|
||||||
let mut s = System::new();
|
|
||||||
s.refresh_memory();
|
|
||||||
// Larger stack size if possible to handle cases with lots of nested directories
|
|
||||||
let available = s.available_memory();
|
|
||||||
if available > large_stack.try_into().unwrap() {
|
|
||||||
Some(large_stack)
|
|
||||||
} else {
|
|
||||||
None
|
|
||||||
}
|
|
||||||
}
|
|
||||||
};
|
|
||||||
if let Some(stack_size_param) = stack_size {
|
|
||||||
pool = pool.stack_size(stack_size_param);
|
|
||||||
}
|
|
||||||
pool.build_global()
|
|
||||||
}
|
|
||||||
|
|||||||
54
src/node.rs
54
src/node.rs
@@ -1,9 +1,8 @@
|
|||||||
use crate::dir_walker::WalkData;
|
|
||||||
use crate::platform::get_metadata;
|
use crate::platform::get_metadata;
|
||||||
use crate::utils::is_filtered_out_due_to_file_time;
|
|
||||||
use crate::utils::is_filtered_out_due_to_invert_regex;
|
use crate::utils::is_filtered_out_due_to_invert_regex;
|
||||||
use crate::utils::is_filtered_out_due_to_regex;
|
use crate::utils::is_filtered_out_due_to_regex;
|
||||||
|
|
||||||
|
use regex::Regex;
|
||||||
use std::cmp::Ordering;
|
use std::cmp::Ordering;
|
||||||
use std::path::PathBuf;
|
use std::path::PathBuf;
|
||||||
|
|
||||||
@@ -16,56 +15,33 @@ pub struct Node {
|
|||||||
pub depth: usize,
|
pub depth: usize,
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug, PartialEq)]
|
|
||||||
pub enum FileTime {
|
|
||||||
Modified,
|
|
||||||
Accessed,
|
|
||||||
Changed,
|
|
||||||
}
|
|
||||||
|
|
||||||
#[allow(clippy::too_many_arguments)]
|
#[allow(clippy::too_many_arguments)]
|
||||||
pub fn build_node(
|
pub fn build_node(
|
||||||
dir: PathBuf,
|
dir: PathBuf,
|
||||||
children: Vec<Node>,
|
children: Vec<Node>,
|
||||||
|
filter_regex: &[Regex],
|
||||||
|
invert_filter_regex: &[Regex],
|
||||||
|
use_apparent_size: bool,
|
||||||
is_symlink: bool,
|
is_symlink: bool,
|
||||||
is_file: bool,
|
is_file: bool,
|
||||||
|
by_filecount: bool,
|
||||||
depth: usize,
|
depth: usize,
|
||||||
walk_data: &WalkData,
|
|
||||||
) -> Option<Node> {
|
) -> Option<Node> {
|
||||||
let use_apparent_size = walk_data.use_apparent_size;
|
get_metadata(&dir, use_apparent_size).map(|data| {
|
||||||
let by_filecount = walk_data.by_filecount;
|
let inode_device = if is_symlink && !use_apparent_size {
|
||||||
let by_filetime = &walk_data.by_filetime;
|
None
|
||||||
|
} else {
|
||||||
|
data.1
|
||||||
|
};
|
||||||
|
|
||||||
get_metadata(
|
let size = if is_filtered_out_due_to_regex(filter_regex, &dir)
|
||||||
&dir,
|
|| is_filtered_out_due_to_invert_regex(invert_filter_regex, &dir)
|
||||||
use_apparent_size,
|
|| (is_symlink && !use_apparent_size)
|
||||||
walk_data.follow_links && is_symlink,
|
|
||||||
)
|
|
||||||
.map(|data| {
|
|
||||||
let inode_device = data.1;
|
|
||||||
|
|
||||||
let size = if is_filtered_out_due_to_regex(walk_data.filter_regex, &dir)
|
|
||||||
|| is_filtered_out_due_to_invert_regex(walk_data.invert_filter_regex, &dir)
|
|
||||||
|| by_filecount && !is_file
|
|| by_filecount && !is_file
|
||||||
|| [
|
{
|
||||||
(&walk_data.filter_modified_time, data.2.0),
|
|
||||||
(&walk_data.filter_accessed_time, data.2.1),
|
|
||||||
(&walk_data.filter_changed_time, data.2.2),
|
|
||||||
]
|
|
||||||
.iter()
|
|
||||||
.any(|(filter_time, actual_time)| {
|
|
||||||
is_filtered_out_due_to_file_time(filter_time, *actual_time)
|
|
||||||
}) {
|
|
||||||
0
|
0
|
||||||
} else if by_filecount {
|
} else if by_filecount {
|
||||||
1
|
1
|
||||||
} else if by_filetime.is_some() {
|
|
||||||
match by_filetime {
|
|
||||||
Some(FileTime::Modified) => data.2.0.unsigned_abs(),
|
|
||||||
Some(FileTime::Accessed) => data.2.1.unsigned_abs(),
|
|
||||||
Some(FileTime::Changed) => data.2.2.unsigned_abs(),
|
|
||||||
None => unreachable!(),
|
|
||||||
}
|
|
||||||
} else {
|
} else {
|
||||||
data.0
|
data.0
|
||||||
};
|
};
|
||||||
|
|||||||
@@ -10,35 +10,15 @@ fn get_block_size() -> u64 {
|
|||||||
512
|
512
|
||||||
}
|
}
|
||||||
|
|
||||||
type InodeAndDevice = (u64, u64);
|
|
||||||
type FileTime = (i64, i64, i64);
|
|
||||||
|
|
||||||
#[cfg(target_family = "unix")]
|
#[cfg(target_family = "unix")]
|
||||||
pub fn get_metadata<P: AsRef<Path>>(
|
pub fn get_metadata(d: &Path, use_apparent_size: bool) -> Option<(u64, Option<(u64, u64)>)> {
|
||||||
path: P,
|
|
||||||
use_apparent_size: bool,
|
|
||||||
follow_links: bool,
|
|
||||||
) -> Option<(u64, Option<InodeAndDevice>, FileTime)> {
|
|
||||||
use std::os::unix::fs::MetadataExt;
|
use std::os::unix::fs::MetadataExt;
|
||||||
let metadata = if follow_links {
|
match d.metadata() {
|
||||||
path.as_ref().metadata()
|
|
||||||
} else {
|
|
||||||
path.as_ref().symlink_metadata()
|
|
||||||
};
|
|
||||||
match metadata {
|
|
||||||
Ok(md) => {
|
Ok(md) => {
|
||||||
if use_apparent_size {
|
if use_apparent_size {
|
||||||
Some((
|
Some((md.len(), Some((md.ino(), md.dev()))))
|
||||||
md.len(),
|
|
||||||
Some((md.ino(), md.dev())),
|
|
||||||
(md.mtime(), md.atime(), md.ctime()),
|
|
||||||
))
|
|
||||||
} else {
|
} else {
|
||||||
Some((
|
Some((md.blocks() * get_block_size(), Some((md.ino(), md.dev()))))
|
||||||
md.blocks() * get_block_size(),
|
|
||||||
Some((md.ino(), md.dev())),
|
|
||||||
(md.mtime(), md.atime(), md.ctime()),
|
|
||||||
))
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
Err(_e) => None,
|
Err(_e) => None,
|
||||||
@@ -46,11 +26,7 @@ pub fn get_metadata<P: AsRef<Path>>(
|
|||||||
}
|
}
|
||||||
|
|
||||||
#[cfg(target_family = "windows")]
|
#[cfg(target_family = "windows")]
|
||||||
pub fn get_metadata<P: AsRef<Path>>(
|
pub fn get_metadata(d: &Path, use_apparent_size: bool) -> Option<(u64, Option<(u64, u64)>)> {
|
||||||
path: P,
|
|
||||||
use_apparent_size: bool,
|
|
||||||
follow_links: bool,
|
|
||||||
) -> Option<(u64, Option<InodeAndDevice>, FileTime)> {
|
|
||||||
// On windows opening the file to get size, file ID and volume can be very
|
// On windows opening the file to get size, file ID and volume can be very
|
||||||
// expensive because 1) it causes a few system calls, and more importantly 2) it can cause
|
// expensive because 1) it causes a few system calls, and more importantly 2) it can cause
|
||||||
// windows defender to scan the file.
|
// windows defender to scan the file.
|
||||||
@@ -89,7 +65,7 @@ pub fn get_metadata<P: AsRef<Path>>(
|
|||||||
|
|
||||||
use std::io;
|
use std::io;
|
||||||
use winapi_util::Handle;
|
use winapi_util::Handle;
|
||||||
fn handle_from_path_limited(path: &Path) -> io::Result<Handle> {
|
fn handle_from_path_limited<P: AsRef<Path>>(path: P) -> io::Result<Handle> {
|
||||||
use std::fs::OpenOptions;
|
use std::fs::OpenOptions;
|
||||||
use std::os::windows::fs::OpenOptionsExt;
|
use std::os::windows::fs::OpenOptionsExt;
|
||||||
const FILE_READ_ATTRIBUTES: u32 = 0x0080;
|
const FILE_READ_ATTRIBUTES: u32 = 0x0080;
|
||||||
@@ -115,46 +91,30 @@ pub fn get_metadata<P: AsRef<Path>>(
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn get_metadata_expensive(
|
fn get_metadata_expensive(
|
||||||
path: &Path,
|
d: &Path,
|
||||||
use_apparent_size: bool,
|
use_apparent_size: bool,
|
||||||
) -> Option<(u64, Option<InodeAndDevice>, FileTime)> {
|
) -> Option<(u64, Option<(u64, u64)>)> {
|
||||||
use winapi_util::file::information;
|
use winapi_util::file::information;
|
||||||
|
|
||||||
let h = handle_from_path_limited(path).ok()?;
|
let h = handle_from_path_limited(d).ok()?;
|
||||||
let info = information(&h).ok()?;
|
let info = information(&h).ok()?;
|
||||||
|
|
||||||
if use_apparent_size {
|
if use_apparent_size {
|
||||||
use filesize::PathExt;
|
use filesize::PathExt;
|
||||||
Some((
|
Some((
|
||||||
path.size_on_disk().ok()?,
|
d.size_on_disk().ok()?,
|
||||||
Some((info.file_index(), info.volume_serial_number())),
|
Some((info.file_index(), info.volume_serial_number())),
|
||||||
(
|
|
||||||
info.last_write_time().unwrap() as i64,
|
|
||||||
info.last_access_time().unwrap() as i64,
|
|
||||||
info.creation_time().unwrap() as i64,
|
|
||||||
),
|
|
||||||
))
|
))
|
||||||
} else {
|
} else {
|
||||||
Some((
|
Some((
|
||||||
info.file_size(),
|
info.file_size(),
|
||||||
Some((info.file_index(), info.volume_serial_number())),
|
Some((info.file_index(), info.volume_serial_number())),
|
||||||
(
|
|
||||||
info.last_write_time().unwrap() as i64,
|
|
||||||
info.last_access_time().unwrap() as i64,
|
|
||||||
info.creation_time().unwrap() as i64,
|
|
||||||
),
|
|
||||||
))
|
))
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
use std::os::windows::fs::MetadataExt;
|
use std::os::windows::fs::MetadataExt;
|
||||||
let path = path.as_ref();
|
match d.metadata() {
|
||||||
let metadata = if follow_links {
|
|
||||||
path.metadata()
|
|
||||||
} else {
|
|
||||||
path.symlink_metadata()
|
|
||||||
};
|
|
||||||
match metadata {
|
|
||||||
Ok(ref md) => {
|
Ok(ref md) => {
|
||||||
const FILE_ATTRIBUTE_ARCHIVE: u32 = 0x20;
|
const FILE_ATTRIBUTE_ARCHIVE: u32 = 0x20;
|
||||||
const FILE_ATTRIBUTE_READONLY: u32 = 0x01;
|
const FILE_ATTRIBUTE_READONLY: u32 = 0x01;
|
||||||
@@ -182,19 +142,11 @@ pub fn get_metadata<P: AsRef<Path>>(
|
|||||||
|| md.file_attributes() == FILE_ATTRIBUTE_NORMAL)
|
|| md.file_attributes() == FILE_ATTRIBUTE_NORMAL)
|
||||||
&& !((attr_filtered & IS_PROBABLY_ONEDRIVE != 0) && use_apparent_size)
|
&& !((attr_filtered & IS_PROBABLY_ONEDRIVE != 0) && use_apparent_size)
|
||||||
{
|
{
|
||||||
Some((
|
Some((md.len(), None))
|
||||||
md.len(),
|
|
||||||
None,
|
|
||||||
(
|
|
||||||
md.last_write_time() as i64,
|
|
||||||
md.last_access_time() as i64,
|
|
||||||
md.creation_time() as i64,
|
|
||||||
),
|
|
||||||
))
|
|
||||||
} else {
|
} else {
|
||||||
get_metadata_expensive(path, use_apparent_size)
|
get_metadata_expensive(d, use_apparent_size)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
_ => get_metadata_expensive(path, use_apparent_size),
|
_ => get_metadata_expensive(d, use_apparent_size),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -3,19 +3,14 @@ use std::{
|
|||||||
io::Write,
|
io::Write,
|
||||||
path::Path,
|
path::Path,
|
||||||
sync::{
|
sync::{
|
||||||
Arc, RwLock,
|
atomic::{AtomicU64, AtomicU8, AtomicUsize, Ordering},
|
||||||
atomic::{AtomicU8, AtomicUsize, Ordering},
|
|
||||||
mpsc::{self, RecvTimeoutError, Sender},
|
mpsc::{self, RecvTimeoutError, Sender},
|
||||||
|
Arc, RwLock,
|
||||||
},
|
},
|
||||||
thread::JoinHandle,
|
thread::JoinHandle,
|
||||||
time::Duration,
|
time::Duration,
|
||||||
};
|
};
|
||||||
|
|
||||||
#[cfg(not(target_has_atomic = "64"))]
|
|
||||||
use portable_atomic::AtomicU64;
|
|
||||||
#[cfg(target_has_atomic = "64")]
|
|
||||||
use std::sync::atomic::AtomicU64;
|
|
||||||
|
|
||||||
use crate::display::human_readable_number;
|
use crate::display::human_readable_number;
|
||||||
|
|
||||||
/* -------------------------------------------------------------------------- */
|
/* -------------------------------------------------------------------------- */
|
||||||
@@ -75,10 +70,10 @@ impl PAtomicInfo {
|
|||||||
|
|
||||||
#[derive(Default)]
|
#[derive(Default)]
|
||||||
pub struct RuntimeErrors {
|
pub struct RuntimeErrors {
|
||||||
pub no_permissions: HashSet<String>,
|
pub no_permissions: bool,
|
||||||
pub file_not_found: HashSet<String>,
|
pub file_not_found: HashSet<String>,
|
||||||
pub unknown_error: HashSet<String>,
|
pub unknown_error: HashSet<String>,
|
||||||
pub interrupted_error: i32,
|
pub abort: bool,
|
||||||
}
|
}
|
||||||
|
|
||||||
/* -------------------------------------------------------------------------- */
|
/* -------------------------------------------------------------------------- */
|
||||||
@@ -118,7 +113,7 @@ impl PIndicator {
|
|||||||
|
|
||||||
let time_info_thread = std::thread::spawn(move || {
|
let time_info_thread = std::thread::spawn(move || {
|
||||||
let mut progress_char_i: usize = 0;
|
let mut progress_char_i: usize = 0;
|
||||||
let mut stderr = std::io::stderr();
|
let mut stdout = std::io::stdout();
|
||||||
let mut msg = "".to_string();
|
let mut msg = "".to_string();
|
||||||
|
|
||||||
// While the timeout triggers we go round the loop
|
// While the timeout triggers we go round the loop
|
||||||
@@ -127,8 +122,7 @@ impl PIndicator {
|
|||||||
receiver.recv_timeout(Duration::from_millis(SPINNER_SLEEP_TIME))
|
receiver.recv_timeout(Duration::from_millis(SPINNER_SLEEP_TIME))
|
||||||
{
|
{
|
||||||
// Clear the text written by 'write!'& Return at the start of line
|
// Clear the text written by 'write!'& Return at the start of line
|
||||||
let clear = format!("\r{:width$}", " ", width = msg.len());
|
print!("\r{:width$}", " ", width = msg.len());
|
||||||
write!(stderr, "{clear}").unwrap();
|
|
||||||
let prog_char = PROGRESS_CHARS[progress_char_i];
|
let prog_char = PROGRESS_CHARS[progress_char_i];
|
||||||
|
|
||||||
msg = match data.state.load(ORDERING) {
|
msg = match data.state.load(ORDERING) {
|
||||||
@@ -137,17 +131,15 @@ impl PIndicator {
|
|||||||
_ => panic!("Unknown State"),
|
_ => panic!("Unknown State"),
|
||||||
};
|
};
|
||||||
|
|
||||||
write!(stderr, "\r{msg}").unwrap();
|
write!(stdout, "\r{msg}").unwrap();
|
||||||
stderr.flush().unwrap();
|
stdout.flush().unwrap();
|
||||||
|
|
||||||
progress_char_i += 1;
|
progress_char_i += 1;
|
||||||
progress_char_i %= PROGRESS_CHARS_LEN;
|
progress_char_i %= PROGRESS_CHARS_LEN;
|
||||||
}
|
}
|
||||||
|
print!("\r{:width$}", " ", width = msg.len());
|
||||||
let clear = format!("\r{:width$}", " ", width = msg.len());
|
print!("\r");
|
||||||
write!(stderr, "{clear}").unwrap();
|
stdout.flush().unwrap();
|
||||||
write!(stderr, "\r").unwrap();
|
|
||||||
stderr.flush().unwrap();
|
|
||||||
});
|
});
|
||||||
self.thread = Some((stop_handler, time_info_thread))
|
self.thread = Some((stop_handler, time_info_thread))
|
||||||
}
|
}
|
||||||
|
|||||||
70
src/utils.rs
70
src/utils.rs
@@ -2,16 +2,13 @@ use platform::get_metadata;
|
|||||||
use std::collections::HashSet;
|
use std::collections::HashSet;
|
||||||
use std::path::{Path, PathBuf};
|
use std::path::{Path, PathBuf};
|
||||||
|
|
||||||
use crate::config::DAY_SECONDS;
|
|
||||||
|
|
||||||
use crate::dir_walker::Operator;
|
|
||||||
use crate::platform;
|
use crate::platform;
|
||||||
use regex::Regex;
|
use regex::Regex;
|
||||||
|
|
||||||
pub fn simplify_dir_names<P: AsRef<Path>>(dirs: &[P]) -> HashSet<PathBuf> {
|
pub fn simplify_dir_names<P: AsRef<Path>>(filenames: Vec<P>) -> HashSet<PathBuf> {
|
||||||
let mut top_level_names: HashSet<PathBuf> = HashSet::with_capacity(dirs.len());
|
let mut top_level_names: HashSet<PathBuf> = HashSet::with_capacity(filenames.len());
|
||||||
|
|
||||||
for t in dirs {
|
for t in filenames {
|
||||||
let top_level_name = normalize_path(t);
|
let top_level_name = normalize_path(t);
|
||||||
let mut can_add = true;
|
let mut can_add = true;
|
||||||
let mut to_remove: Vec<PathBuf> = Vec::new();
|
let mut to_remove: Vec<PathBuf> = Vec::new();
|
||||||
@@ -34,25 +31,13 @@ pub fn simplify_dir_names<P: AsRef<Path>>(dirs: &[P]) -> HashSet<PathBuf> {
|
|||||||
top_level_names
|
top_level_names
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn get_filesystem_devices<P: AsRef<Path>>(paths: &[P], follow_links: bool) -> HashSet<u64> {
|
pub fn get_filesystem_devices<'a, P: IntoIterator<Item = &'a PathBuf>>(paths: P) -> HashSet<u64> {
|
||||||
use std::fs;
|
|
||||||
// Gets the device ids for the filesystems which are used by the argument paths
|
// Gets the device ids for the filesystems which are used by the argument paths
|
||||||
paths
|
paths
|
||||||
.iter()
|
.into_iter()
|
||||||
.filter_map(|p| {
|
.filter_map(|p| match get_metadata(p, false) {
|
||||||
let follow_links = if follow_links {
|
Some((_size, Some((_id, dev)))) => Some(dev),
|
||||||
// slow path: If dereference-links is set, then we check if the file is a symbolic link
|
_ => None,
|
||||||
match fs::symlink_metadata(p) {
|
|
||||||
Ok(metadata) => metadata.file_type().is_symlink(),
|
|
||||||
Err(_) => false,
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
false
|
|
||||||
};
|
|
||||||
match get_metadata(p, false, follow_links) {
|
|
||||||
Some((_size, Some((_id, dev)), _time)) => Some(dev),
|
|
||||||
_ => None,
|
|
||||||
}
|
|
||||||
})
|
})
|
||||||
.collect()
|
.collect()
|
||||||
}
|
}
|
||||||
@@ -67,17 +52,6 @@ pub fn normalize_path<P: AsRef<Path>>(path: P) -> PathBuf {
|
|||||||
path.as_ref().components().collect()
|
path.as_ref().components().collect()
|
||||||
}
|
}
|
||||||
|
|
||||||
// Canonicalize the path only if it is an absolute path
|
|
||||||
pub fn canonicalize_absolute_path(path: PathBuf) -> PathBuf {
|
|
||||||
if !path.is_absolute() {
|
|
||||||
return path;
|
|
||||||
}
|
|
||||||
match std::fs::canonicalize(&path) {
|
|
||||||
Ok(canonicalized_path) => canonicalized_path,
|
|
||||||
Err(_) => path,
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
pub fn is_filtered_out_due_to_regex(filter_regex: &[Regex], dir: &Path) -> bool {
|
pub fn is_filtered_out_due_to_regex(filter_regex: &[Regex], dir: &Path) -> bool {
|
||||||
if filter_regex.is_empty() {
|
if filter_regex.is_empty() {
|
||||||
false
|
false
|
||||||
@@ -88,20 +62,6 @@ pub fn is_filtered_out_due_to_regex(filter_regex: &[Regex], dir: &Path) -> bool
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn is_filtered_out_due_to_file_time(
|
|
||||||
filter_time: &Option<(Operator, i64)>,
|
|
||||||
actual_time: i64,
|
|
||||||
) -> bool {
|
|
||||||
match filter_time {
|
|
||||||
None => false,
|
|
||||||
Some((Operator::Equal, bound_time)) => {
|
|
||||||
!(actual_time >= *bound_time && actual_time < *bound_time + DAY_SECONDS)
|
|
||||||
}
|
|
||||||
Some((Operator::GreaterThan, bound_time)) => actual_time < *bound_time,
|
|
||||||
Some((Operator::LessThan, bound_time)) => actual_time > *bound_time,
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
pub fn is_filtered_out_due_to_invert_regex(filter_regex: &[Regex], dir: &Path) -> bool {
|
pub fn is_filtered_out_due_to_invert_regex(filter_regex: &[Regex], dir: &Path) -> bool {
|
||||||
filter_regex
|
filter_regex
|
||||||
.iter()
|
.iter()
|
||||||
@@ -122,15 +82,15 @@ mod tests {
|
|||||||
fn test_simplify_dir() {
|
fn test_simplify_dir() {
|
||||||
let mut correct = HashSet::new();
|
let mut correct = HashSet::new();
|
||||||
correct.insert(PathBuf::from("a"));
|
correct.insert(PathBuf::from("a"));
|
||||||
assert_eq!(simplify_dir_names(&["a"]), correct);
|
assert_eq!(simplify_dir_names(vec!["a"]), correct);
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn test_simplify_dir_rm_subdir() {
|
fn test_simplify_dir_rm_subdir() {
|
||||||
let mut correct = HashSet::new();
|
let mut correct = HashSet::new();
|
||||||
correct.insert(["a", "b"].iter().collect::<PathBuf>());
|
correct.insert(["a", "b"].iter().collect::<PathBuf>());
|
||||||
assert_eq!(simplify_dir_names(&["a/b/c", "a/b", "a/b/d/f"]), correct);
|
assert_eq!(simplify_dir_names(vec!["a/b/c", "a/b", "a/b/d/f"]), correct);
|
||||||
assert_eq!(simplify_dir_names(&["a/b", "a/b/c", "a/b/d/f"]), correct);
|
assert_eq!(simplify_dir_names(vec!["a/b", "a/b/c", "a/b/d/f"]), correct);
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
@@ -139,7 +99,7 @@ mod tests {
|
|||||||
correct.insert(["a", "b"].iter().collect::<PathBuf>());
|
correct.insert(["a", "b"].iter().collect::<PathBuf>());
|
||||||
correct.insert(PathBuf::from("c"));
|
correct.insert(PathBuf::from("c"));
|
||||||
assert_eq!(
|
assert_eq!(
|
||||||
simplify_dir_names(&[
|
simplify_dir_names(vec![
|
||||||
"a/b",
|
"a/b",
|
||||||
"a/b//",
|
"a/b//",
|
||||||
"a/././b///",
|
"a/././b///",
|
||||||
@@ -158,14 +118,14 @@ mod tests {
|
|||||||
correct.insert(PathBuf::from("b"));
|
correct.insert(PathBuf::from("b"));
|
||||||
correct.insert(["c", "a", "b"].iter().collect::<PathBuf>());
|
correct.insert(["c", "a", "b"].iter().collect::<PathBuf>());
|
||||||
correct.insert(["a", "b"].iter().collect::<PathBuf>());
|
correct.insert(["a", "b"].iter().collect::<PathBuf>());
|
||||||
assert_eq!(simplify_dir_names(&["a/b", "c/a/b/", "b"]), correct);
|
assert_eq!(simplify_dir_names(vec!["a/b", "c/a/b/", "b"]), correct);
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn test_simplify_dir_dots() {
|
fn test_simplify_dir_dots() {
|
||||||
let mut correct = HashSet::new();
|
let mut correct = HashSet::new();
|
||||||
correct.insert(PathBuf::from("src"));
|
correct.insert(PathBuf::from("src"));
|
||||||
assert_eq!(simplify_dir_names(&["src/."]), correct);
|
assert_eq!(simplify_dir_names(vec!["src/."]), correct);
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
@@ -173,7 +133,7 @@ mod tests {
|
|||||||
let mut correct = HashSet::new();
|
let mut correct = HashSet::new();
|
||||||
correct.insert(PathBuf::from("src"));
|
correct.insert(PathBuf::from("src"));
|
||||||
correct.insert(PathBuf::from("src_v2"));
|
correct.insert(PathBuf::from("src_v2"));
|
||||||
assert_eq!(simplify_dir_names(&["src/", "src_v2"]), correct);
|
assert_eq!(simplify_dir_names(vec!["src/", "src_v2"]), correct);
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
|
|||||||
@@ -1,11 +1,9 @@
|
|||||||
use assert_cmd::Command;
|
use assert_cmd::Command;
|
||||||
use std::ffi::OsStr;
|
use std::ffi::OsStr;
|
||||||
use std::process::Output;
|
use std::str;
|
||||||
use std::sync::Once;
|
use std::sync::Once;
|
||||||
use std::{io, str};
|
|
||||||
|
|
||||||
static INIT: Once = Once::new();
|
static INIT: Once = Once::new();
|
||||||
static UNREADABLE_DIR_PATH: &str = "/tmp/unreadable_dir";
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* This file contains tests that verify the exact output of the command.
|
* This file contains tests that verify the exact output of the command.
|
||||||
@@ -35,59 +33,34 @@ fn copy_test_data(dir: &str) {
|
|||||||
.map_err(|err| eprintln!("Error copying directory for test setup\n{:?}", err));
|
.map_err(|err| eprintln!("Error copying directory for test setup\n{:?}", err));
|
||||||
}
|
}
|
||||||
|
|
||||||
fn create_unreadable_directory() -> io::Result<()> {
|
|
||||||
#[cfg(unix)]
|
|
||||||
{
|
|
||||||
use std::fs;
|
|
||||||
use std::fs::Permissions;
|
|
||||||
use std::os::unix::fs::PermissionsExt;
|
|
||||||
fs::create_dir_all(UNREADABLE_DIR_PATH)?;
|
|
||||||
fs::set_permissions(UNREADABLE_DIR_PATH, Permissions::from_mode(0))?;
|
|
||||||
}
|
|
||||||
Ok(())
|
|
||||||
}
|
|
||||||
|
|
||||||
fn initialize() {
|
fn initialize() {
|
||||||
INIT.call_once(|| {
|
INIT.call_once(|| {
|
||||||
copy_test_data("tests/test_dir");
|
copy_test_data("tests/test_dir");
|
||||||
copy_test_data("tests/test_dir2");
|
copy_test_data("tests/test_dir2");
|
||||||
copy_test_data("tests/test_dir_unicode");
|
copy_test_data("tests/test_dir_unicode");
|
||||||
|
|
||||||
if let Err(e) = create_unreadable_directory() {
|
|
||||||
panic!("Failed to create unreadable directory: {}", e);
|
|
||||||
}
|
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
fn run_cmd<T: AsRef<OsStr>>(command_args: &[T]) -> Output {
|
fn exact_output_test<T: AsRef<OsStr>>(valid_outputs: Vec<String>, command_args: Vec<T>) {
|
||||||
initialize();
|
initialize();
|
||||||
let mut to_run = &mut Command::cargo_bin("dust").unwrap();
|
|
||||||
|
let mut a = &mut Command::cargo_bin("dust").unwrap();
|
||||||
|
|
||||||
for p in command_args {
|
for p in command_args {
|
||||||
to_run = to_run.arg(p);
|
a = a.arg(p);
|
||||||
}
|
}
|
||||||
to_run.unwrap()
|
|
||||||
}
|
|
||||||
|
|
||||||
fn exact_stdout_test<T: AsRef<OsStr>>(command_args: &[T], valid_stdout: Vec<String>) {
|
let output = str::from_utf8(&a.unwrap().stdout).unwrap().to_owned();
|
||||||
let to_run = run_cmd(command_args);
|
|
||||||
|
|
||||||
let stdout_output = str::from_utf8(&to_run.stdout).unwrap().to_owned();
|
let will_fail = valid_outputs.iter().any(|i| output.contains(i));
|
||||||
let will_fail = valid_stdout.iter().any(|i| stdout_output.contains(i));
|
|
||||||
if !will_fail {
|
if !will_fail {
|
||||||
eprintln!(
|
eprintln!(
|
||||||
"output(stdout):\n{}\ndoes not contain any of:\n{}",
|
"output:\n{}\ndoes not contain any of:\n{}",
|
||||||
stdout_output,
|
output,
|
||||||
valid_stdout.join("\n\n")
|
valid_outputs.join("\n\n")
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
assert!(will_fail);
|
assert!(will_fail)
|
||||||
}
|
|
||||||
|
|
||||||
fn exact_stderr_test<T: AsRef<OsStr>>(command_args: &[T], valid_stderr: String) {
|
|
||||||
let to_run = run_cmd(command_args);
|
|
||||||
|
|
||||||
let stderr_output = str::from_utf8(&to_run.stderr).unwrap().trim();
|
|
||||||
assert_eq!(stderr_output, valid_stderr);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// "windows" result data can vary by host (size seems to be variable by one byte); fix code vs test and re-enable
|
// "windows" result data can vary by host (size seems to be variable by one byte); fix code vs test and re-enable
|
||||||
@@ -95,20 +68,20 @@ fn exact_stderr_test<T: AsRef<OsStr>>(command_args: &[T], valid_stderr: String)
|
|||||||
#[test]
|
#[test]
|
||||||
pub fn test_main_basic() {
|
pub fn test_main_basic() {
|
||||||
// -c is no color mode - This makes testing much simpler
|
// -c is no color mode - This makes testing much simpler
|
||||||
exact_stdout_test(&["-c", "-B", "/tmp/test_dir/"], main_output());
|
exact_output_test(main_output(), vec!["-c", "-B", "/tmp/test_dir/"])
|
||||||
}
|
}
|
||||||
|
|
||||||
#[cfg_attr(target_os = "windows", ignore)]
|
#[cfg_attr(target_os = "windows", ignore)]
|
||||||
#[test]
|
#[test]
|
||||||
pub fn test_main_multi_arg() {
|
pub fn test_main_multi_arg() {
|
||||||
let command_args = [
|
let command_args = vec![
|
||||||
"-c",
|
"-c",
|
||||||
"-B",
|
"-B",
|
||||||
"/tmp/test_dir/many/",
|
"/tmp/test_dir/many/",
|
||||||
"/tmp/test_dir",
|
"/tmp/test_dir",
|
||||||
"/tmp/test_dir",
|
"/tmp/test_dir",
|
||||||
];
|
];
|
||||||
exact_stdout_test(&command_args, main_output());
|
exact_output_test(main_output(), command_args);
|
||||||
}
|
}
|
||||||
|
|
||||||
fn main_output() -> Vec<String> {
|
fn main_output() -> Vec<String> {
|
||||||
@@ -138,8 +111,8 @@ fn main_output() -> Vec<String> {
|
|||||||
#[cfg_attr(target_os = "windows", ignore)]
|
#[cfg_attr(target_os = "windows", ignore)]
|
||||||
#[test]
|
#[test]
|
||||||
pub fn test_main_long_paths() {
|
pub fn test_main_long_paths() {
|
||||||
let command_args = ["-c", "-p", "-B", "/tmp/test_dir/"];
|
let command_args = vec!["-c", "-p", "-B", "/tmp/test_dir/"];
|
||||||
exact_stdout_test(&command_args, main_output_long_paths());
|
exact_output_test(main_output_long_paths(), command_args);
|
||||||
}
|
}
|
||||||
|
|
||||||
fn main_output_long_paths() -> Vec<String> {
|
fn main_output_long_paths() -> Vec<String> {
|
||||||
@@ -166,8 +139,8 @@ fn main_output_long_paths() -> Vec<String> {
|
|||||||
#[cfg_attr(target_os = "windows", ignore)]
|
#[cfg_attr(target_os = "windows", ignore)]
|
||||||
#[test]
|
#[test]
|
||||||
pub fn test_substring_of_names_and_long_names() {
|
pub fn test_substring_of_names_and_long_names() {
|
||||||
let command_args = ["-c", "-B", "/tmp/test_dir2"];
|
let command_args = vec!["-c", "-B", "/tmp/test_dir2"];
|
||||||
exact_stdout_test(&command_args, no_substring_of_names_output());
|
exact_output_test(no_substring_of_names_output(), command_args);
|
||||||
}
|
}
|
||||||
|
|
||||||
fn no_substring_of_names_output() -> Vec<String> {
|
fn no_substring_of_names_output() -> Vec<String> {
|
||||||
@@ -200,8 +173,8 @@ fn no_substring_of_names_output() -> Vec<String> {
|
|||||||
#[cfg_attr(target_os = "windows", ignore)]
|
#[cfg_attr(target_os = "windows", ignore)]
|
||||||
#[test]
|
#[test]
|
||||||
pub fn test_unicode_directories() {
|
pub fn test_unicode_directories() {
|
||||||
let command_args = ["-c", "-B", "/tmp/test_dir_unicode"];
|
let command_args = vec!["-c", "-B", "/tmp/test_dir_unicode"];
|
||||||
exact_stdout_test(&command_args, unicode_dir());
|
exact_output_test(unicode_dir(), command_args);
|
||||||
}
|
}
|
||||||
|
|
||||||
fn unicode_dir() -> Vec<String> {
|
fn unicode_dir() -> Vec<String> {
|
||||||
@@ -227,8 +200,8 @@ fn unicode_dir() -> Vec<String> {
|
|||||||
#[cfg_attr(target_os = "windows", ignore)]
|
#[cfg_attr(target_os = "windows", ignore)]
|
||||||
#[test]
|
#[test]
|
||||||
pub fn test_apparent_size() {
|
pub fn test_apparent_size() {
|
||||||
let command_args = ["-c", "-s", "-b", "/tmp/test_dir"];
|
let command_args = vec!["-c", "-s", "-b", "/tmp/test_dir"];
|
||||||
exact_stdout_test(&command_args, apparent_size_output());
|
exact_output_test(apparent_size_output(), command_args);
|
||||||
}
|
}
|
||||||
|
|
||||||
fn apparent_size_output() -> Vec<String> {
|
fn apparent_size_output() -> Vec<String> {
|
||||||
@@ -249,26 +222,3 @@ fn apparent_size_output() -> Vec<String> {
|
|||||||
|
|
||||||
vec![one_space_before, two_space_before]
|
vec![one_space_before, two_space_before]
|
||||||
}
|
}
|
||||||
|
|
||||||
#[cfg_attr(target_os = "windows", ignore)]
|
|
||||||
#[test]
|
|
||||||
pub fn test_permission_normal() {
|
|
||||||
let command_args = [UNREADABLE_DIR_PATH];
|
|
||||||
let permission_msg =
|
|
||||||
r#"Did not have permissions for all directories (add --print-errors to see errors)"#
|
|
||||||
.trim()
|
|
||||||
.to_string();
|
|
||||||
exact_stderr_test(&command_args, permission_msg);
|
|
||||||
}
|
|
||||||
|
|
||||||
#[cfg_attr(target_os = "windows", ignore)]
|
|
||||||
#[test]
|
|
||||||
pub fn test_permission_flag() {
|
|
||||||
// add the flag to CLI
|
|
||||||
let command_args = ["--print-errors", UNREADABLE_DIR_PATH];
|
|
||||||
let permission_msg = format!(
|
|
||||||
"Did not have permissions for directories: {}",
|
|
||||||
UNREADABLE_DIR_PATH
|
|
||||||
);
|
|
||||||
exact_stderr_test(&command_args, permission_msg);
|
|
||||||
}
|
|
||||||
|
|||||||
@@ -10,9 +10,6 @@ use std::str;
|
|||||||
|
|
||||||
fn build_command<T: AsRef<OsStr>>(command_args: Vec<T>) -> String {
|
fn build_command<T: AsRef<OsStr>>(command_args: Vec<T>) -> String {
|
||||||
let mut cmd = &mut Command::cargo_bin("dust").unwrap();
|
let mut cmd = &mut Command::cargo_bin("dust").unwrap();
|
||||||
// Hide progress bar
|
|
||||||
cmd = cmd.arg("-P");
|
|
||||||
|
|
||||||
for p in command_args {
|
for p in command_args {
|
||||||
cmd = cmd.arg(p);
|
cmd = cmd.arg(p);
|
||||||
}
|
}
|
||||||
@@ -62,12 +59,6 @@ pub fn test_d_flag_works() {
|
|||||||
assert!(!output.contains("hello_file"));
|
assert!(!output.contains("hello_file"));
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
|
||||||
pub fn test_threads_flag_works() {
|
|
||||||
let output = build_command(vec!["-T", "1", "tests/test_dir/"]);
|
|
||||||
assert!(output.contains("hello_file"));
|
|
||||||
}
|
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
pub fn test_d_flag_works_and_still_recurses_down() {
|
pub fn test_d_flag_works_and_still_recurses_down() {
|
||||||
// We had a bug where running with '-d 1' would stop at the first directory and the code
|
// We had a bug where running with '-d 1' would stop at the first directory and the code
|
||||||
@@ -142,9 +133,9 @@ pub fn test_show_files_by_type() {
|
|||||||
#[cfg(target_family = "unix")]
|
#[cfg(target_family = "unix")]
|
||||||
pub fn test_show_files_only() {
|
pub fn test_show_files_only() {
|
||||||
let output = build_command(vec!["-c", "-F", "tests/test_dir"]);
|
let output = build_command(vec!["-c", "-F", "tests/test_dir"]);
|
||||||
assert!(output.contains("a_file"));
|
assert!(output.contains("tests/test_dir/many/a_file"));
|
||||||
assert!(output.contains("hello_file"));
|
assert!(output.contains("tests/test_dir/many/hello_file"));
|
||||||
assert!(!output.contains("many"));
|
assert!(!output.contains("tests/test_dir/many "));
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
@@ -257,26 +248,3 @@ pub fn test_force_color() {
|
|||||||
assert!(output.contains("\x1B[31m"));
|
assert!(output.contains("\x1B[31m"));
|
||||||
assert!(output.contains("\x1B[0m"));
|
assert!(output.contains("\x1B[0m"));
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
|
||||||
pub fn test_collapse() {
|
|
||||||
let output = build_command(vec!["--collapse", "many", "tests/test_dir/"]);
|
|
||||||
assert!(output.contains("many"));
|
|
||||||
assert!(!output.contains("hello_file"));
|
|
||||||
}
|
|
||||||
|
|
||||||
#[test]
|
|
||||||
pub fn test_handle_duplicate_names() {
|
|
||||||
// Check that even if we run on a multiple directories with the same name
|
|
||||||
// we still show the distinct parent dir in the output
|
|
||||||
let output = build_command(vec![
|
|
||||||
"tests/test_dir_matching/dave/dup_name",
|
|
||||||
"tests/test_dir_matching/andy/dup_name",
|
|
||||||
"ci",
|
|
||||||
]);
|
|
||||||
assert!(output.contains("andy"));
|
|
||||||
assert!(output.contains("dave"));
|
|
||||||
assert!(output.contains("ci"));
|
|
||||||
assert!(output.contains("dup_name"));
|
|
||||||
assert!(!output.contains("test_dir_matching"));
|
|
||||||
}
|
|
||||||
|
|||||||
Reference in New Issue
Block a user