mirror of
https://github.com/bootandy/dust.git
synced 2025-12-12 23:59:58 -08:00
Compare commits
1 Commits
upgrade_li
...
v1.1.0
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
3428c5c35c |
5
.github/workflows/CICD.yml
vendored
5
.github/workflows/CICD.yml
vendored
@@ -45,11 +45,6 @@ jobs:
|
|||||||
override: true
|
override: true
|
||||||
profile: minimal # minimal component installation (ie, no documentation)
|
profile: minimal # minimal component installation (ie, no documentation)
|
||||||
components: rustfmt, clippy
|
components: rustfmt, clippy
|
||||||
- name: Install wget for Windows
|
|
||||||
if: matrix.job.os == 'windows-latest'
|
|
||||||
run: choco install wget --no-progress
|
|
||||||
- name: typos-action
|
|
||||||
uses: crate-ci/typos@v1.28.4
|
|
||||||
- name: "`fmt` testing"
|
- name: "`fmt` testing"
|
||||||
if: steps.vars.outputs.JOB_DO_FORMAT_TESTING
|
if: steps.vars.outputs.JOB_DO_FORMAT_TESTING
|
||||||
uses: actions-rs/cargo@v1
|
uses: actions-rs/cargo@v1
|
||||||
|
|||||||
@@ -1,11 +0,0 @@
|
|||||||
repos:
|
|
||||||
- repo: https://github.com/doublify/pre-commit-rust
|
|
||||||
rev: v1.0
|
|
||||||
hooks:
|
|
||||||
- id: cargo-check
|
|
||||||
stages: [commit]
|
|
||||||
- id: fmt
|
|
||||||
stages: [commit]
|
|
||||||
- id: clippy
|
|
||||||
args: [--all-targets, --all-features]
|
|
||||||
stages: [commit]
|
|
||||||
462
Cargo.lock
generated
462
Cargo.lock
generated
File diff suppressed because it is too large
Load Diff
17
Cargo.toml
17
Cargo.toml
@@ -1,9 +1,9 @@
|
|||||||
[package]
|
[package]
|
||||||
name = "du-dust"
|
name = "du-dust"
|
||||||
description = "A more intuitive version of du"
|
description = "A more intuitive version of du"
|
||||||
version = "1.2.0"
|
version = "1.1.0"
|
||||||
authors = ["bootandy <bootandy@gmail.com>", "nebkor <code@ardent.nebcorp.com>"]
|
authors = ["bootandy <bootandy@gmail.com>", "nebkor <code@ardent.nebcorp.com>"]
|
||||||
edition = "2024"
|
edition = "2021"
|
||||||
readme = "README.md"
|
readme = "README.md"
|
||||||
|
|
||||||
documentation = "https://github.com/bootandy/dust"
|
documentation = "https://github.com/bootandy/dust"
|
||||||
@@ -44,9 +44,6 @@ sysinfo = "0.27"
|
|||||||
ctrlc = "3.4"
|
ctrlc = "3.4"
|
||||||
chrono = "0.4"
|
chrono = "0.4"
|
||||||
|
|
||||||
[target.'cfg(not(target_has_atomic = "64"))'.dependencies]
|
|
||||||
portable-atomic = "1.4"
|
|
||||||
|
|
||||||
[target.'cfg(windows)'.dependencies]
|
[target.'cfg(windows)'.dependencies]
|
||||||
winapi-util = "0.1"
|
winapi-util = "0.1"
|
||||||
filesize = "0.2.0"
|
filesize = "0.2.0"
|
||||||
@@ -86,16 +83,6 @@ assets = [
|
|||||||
"usr/share/doc/du-dust/README",
|
"usr/share/doc/du-dust/README",
|
||||||
"644",
|
"644",
|
||||||
],
|
],
|
||||||
[
|
|
||||||
"man-page/dust.1",
|
|
||||||
"usr/share/man/man1/dust.1",
|
|
||||||
"644",
|
|
||||||
],
|
|
||||||
[
|
|
||||||
"completions/dust.bash",
|
|
||||||
"usr/share/bash-completion/completions/dust",
|
|
||||||
"644",
|
|
||||||
],
|
|
||||||
]
|
]
|
||||||
extended-description = """\
|
extended-description = """\
|
||||||
Dust is meant to give you an instant overview of which directories are using
|
Dust is meant to give you an instant overview of which directories are using
|
||||||
|
|||||||
13
README.md
13
README.md
@@ -27,17 +27,11 @@ Because I want an easy way to see where my disk is being used.
|
|||||||
|
|
||||||
- `brew install dust`
|
- `brew install dust`
|
||||||
|
|
||||||
#### [Snap](https://ubuntu.com/core/services/guide/snaps-intro) Ubuntu and [supported systems](https://snapcraft.io/docs/installing-snapd)
|
|
||||||
|
|
||||||
- `snap install dust`
|
|
||||||
|
|
||||||
Note: `dust` installed through `snap` can only access files stored in the `/home` directory. See daniejstriata/dust-snap#2 for more information.
|
|
||||||
|
|
||||||
#### [Pacstall](https://github.com/pacstall/pacstall) (Debian/Ubuntu)
|
#### [Pacstall](https://github.com/pacstall/pacstall) (Debian/Ubuntu)
|
||||||
|
|
||||||
- `pacstall -I dust-bin`
|
- `pacstall -I dust-bin`
|
||||||
|
|
||||||
#### Anaconda (conda-forge)
|
### Anaconda (conda-forge)
|
||||||
|
|
||||||
- `conda install -c conda-forge dust`
|
- `conda install -c conda-forge dust`
|
||||||
|
|
||||||
@@ -45,10 +39,6 @@ Note: `dust` installed through `snap` can only access files stored in the `/home
|
|||||||
|
|
||||||
- `deb-get install du-dust`
|
- `deb-get install du-dust`
|
||||||
|
|
||||||
#### [x-cmd](https://www.x-cmd.com/pkg/#VPContent)
|
|
||||||
|
|
||||||
- `x env use dust`
|
|
||||||
|
|
||||||
#### Windows:
|
#### Windows:
|
||||||
|
|
||||||
- `scoop install dust`
|
- `scoop install dust`
|
||||||
@@ -103,7 +93,6 @@ Usage: dust --skip-total (No total row will be displayed)
|
|||||||
Usage: dust -z 40000/30MB/20kib (Exclude output files/directories below size 40000 bytes / 30MB / 20KiB)
|
Usage: dust -z 40000/30MB/20kib (Exclude output files/directories below size 40000 bytes / 30MB / 20KiB)
|
||||||
Usage: dust -j (Prints JSON representation of directories, try: dust -j | jq)
|
Usage: dust -j (Prints JSON representation of directories, try: dust -j | jq)
|
||||||
Usage: dust --files0-from=FILE (Reads null-terminated file paths from FILE); If FILE is - then read from stdin
|
Usage: dust --files0-from=FILE (Reads null-terminated file paths from FILE); If FILE is - then read from stdin
|
||||||
Usage: dust --collapse=node-modules will keep the node-modules folder collapsed in display instead of recursively opening it
|
|
||||||
```
|
```
|
||||||
|
|
||||||
## Config file
|
## Config file
|
||||||
|
|||||||
@@ -1,21 +1,14 @@
|
|||||||
# ----------- To do a release ---------
|
# ----------- To do a release ---------
|
||||||
|
|
||||||
# ----------- Pre release ---------
|
|
||||||
# Compare times of runs to check no drastic slow down:
|
# Compare times of runs to check no drastic slow down:
|
||||||
# hyperfine 'target/release/dust /home/andy'
|
# time target/release/dust ~/dev
|
||||||
# hyperfine 'dust /home/andy'
|
# time dust ~dev
|
||||||
|
|
||||||
# ----------- Release ---------
|
|
||||||
# inc version in cargo.toml
|
|
||||||
# cargo build --release
|
|
||||||
# commit changed files
|
|
||||||
# merge to master in github
|
|
||||||
|
|
||||||
|
# edit version in cargo.toml
|
||||||
# tag a commit and push (increment version in Cargo.toml first):
|
# tag a commit and push (increment version in Cargo.toml first):
|
||||||
# git tag v0.4.5
|
# git tag v0.4.5
|
||||||
# git push origin v0.4.5
|
# git push origin v0.4.5
|
||||||
|
|
||||||
# cargo publish to put it in crates.io
|
# cargo publish to put it in crates.io
|
||||||
|
|
||||||
# Optional: To install locally
|
# To install locally [Do before pushing it]
|
||||||
#cargo install --path .
|
#cargo install --path .
|
||||||
|
|||||||
@@ -14,40 +14,36 @@ _dust() {
|
|||||||
fi
|
fi
|
||||||
|
|
||||||
local context curcontext="$curcontext" state line
|
local context curcontext="$curcontext" state line
|
||||||
_arguments "${_arguments_options[@]}" : \
|
_arguments "${_arguments_options[@]}" \
|
||||||
'-d+[Depth to show]:DEPTH:_default' \
|
'-d+[Depth to show]:DEPTH: ' \
|
||||||
'--depth=[Depth to show]:DEPTH:_default' \
|
'--depth=[Depth to show]:DEPTH: ' \
|
||||||
'-T+[Number of threads to use]: :_default' \
|
'-T+[Number of threads to use]: : ' \
|
||||||
'--threads=[Number of threads to use]: :_default' \
|
'--threads=[Number of threads to use]: : ' \
|
||||||
'--config=[Specify a config file to use]:FILE:_files' \
|
'-n+[Number of lines of output to show. (Default is terminal_height - 10)]:NUMBER: ' \
|
||||||
'-n+[Number of lines of output to show. (Default is terminal_height - 10)]:NUMBER:_default' \
|
'--number-of-lines=[Number of lines of output to show. (Default is terminal_height - 10)]:NUMBER: ' \
|
||||||
'--number-of-lines=[Number of lines of output to show. (Default is terminal_height - 10)]:NUMBER:_default' \
|
'*-X+[Exclude any file or directory with this name]:PATH:_files' \
|
||||||
'*-X+[Exclude any file or directory with this path]:PATH:_files' \
|
'*--ignore-directory=[Exclude any file or directory with this name]:PATH:_files' \
|
||||||
'*--ignore-directory=[Exclude any file or directory with this path]:PATH:_files' \
|
|
||||||
'-I+[Exclude any file or directory with a regex matching that listed in this file, the file entries will be added to the ignore regexs provided by --invert_filter]:FILE:_files' \
|
'-I+[Exclude any file or directory with a regex matching that listed in this file, the file entries will be added to the ignore regexs provided by --invert_filter]:FILE:_files' \
|
||||||
'--ignore-all-in-file=[Exclude any file or directory with a regex matching that listed in this file, the file entries will be added to the ignore regexs provided by --invert_filter]:FILE:_files' \
|
'--ignore-all-in-file=[Exclude any file or directory with a regex matching that listed in this file, the file entries will be added to the ignore regexs provided by --invert_filter]:FILE:_files' \
|
||||||
'-z+[Minimum size file to include in output]:MIN_SIZE:_default' \
|
'-z+[Minimum size file to include in output]:MIN_SIZE: ' \
|
||||||
'--min-size=[Minimum size file to include in output]:MIN_SIZE:_default' \
|
'--min-size=[Minimum size file to include in output]:MIN_SIZE: ' \
|
||||||
'(-e --filter -t --file-types)*-v+[Exclude filepaths matching this regex. To ignore png files type\: -v "\\.png\$" ]:REGEX:_default' \
|
'(-e --filter -t --file_types)*-v+[Exclude filepaths matching this regex. To ignore png files type\: -v "\\.png\$" ]:REGEX: ' \
|
||||||
'(-e --filter -t --file-types)*--invert-filter=[Exclude filepaths matching this regex. To ignore png files type\: -v "\\.png\$" ]:REGEX:_default' \
|
'(-e --filter -t --file_types)*--invert-filter=[Exclude filepaths matching this regex. To ignore png files type\: -v "\\.png\$" ]:REGEX: ' \
|
||||||
'(-t --file-types)*-e+[Only include filepaths matching this regex. For png files type\: -e "\\.png\$" ]:REGEX:_default' \
|
'(-t --file_types)*-e+[Only include filepaths matching this regex. For png files type\: -e "\\.png\$" ]:REGEX: ' \
|
||||||
'(-t --file-types)*--filter=[Only include filepaths matching this regex. For png files type\: -e "\\.png\$" ]:REGEX:_default' \
|
'(-t --file_types)*--filter=[Only include filepaths matching this regex. For png files type\: -e "\\.png\$" ]:REGEX: ' \
|
||||||
'-w+[Specify width of output overriding the auto detection of terminal width]:WIDTH:_default' \
|
'-w+[Specify width of output overriding the auto detection of terminal width]:WIDTH: ' \
|
||||||
'--terminal-width=[Specify width of output overriding the auto detection of terminal width]:WIDTH:_default' \
|
'--terminal_width=[Specify width of output overriding the auto detection of terminal width]:WIDTH: ' \
|
||||||
'-o+[Changes output display size. si will print sizes in powers of 1000. b k m g t kb mb gb tb will print the whole tree in that size.]:FORMAT:(si b k m g t kb mb gb tb)' \
|
'-o+[Changes output display size. si will print sizes in powers of 1000. b k m g t kb mb gb tb will print the whole tree in that size.]:FORMAT:(si b k m g t kb mb gb tb)' \
|
||||||
'--output-format=[Changes output display size. si will print sizes in powers of 1000. b k m g t kb mb gb tb will print the whole tree in that size.]:FORMAT:(si b k m g t kb mb gb tb)' \
|
'--output-format=[Changes output display size. si will print sizes in powers of 1000. b k m g t kb mb gb tb will print the whole tree in that size.]:FORMAT:(si b k m g t kb mb gb tb)' \
|
||||||
'-S+[Specify memory to use as stack size - use if you see\: '\''fatal runtime error\: stack overflow'\'' (default low memory=1048576, high memory=1073741824)]:STACK_SIZE:_default' \
|
'-S+[Specify memory to use as stack size - use if you see\: '\''fatal runtime error\: stack overflow'\'' (default low memory=1048576, high memory=1073741824)]:STACK_SIZE: ' \
|
||||||
'--stack-size=[Specify memory to use as stack size - use if you see\: '\''fatal runtime error\: stack overflow'\'' (default low memory=1048576, high memory=1073741824)]:STACK_SIZE:_default' \
|
'--stack-size=[Specify memory to use as stack size - use if you see\: '\''fatal runtime error\: stack overflow'\'' (default low memory=1048576, high memory=1073741824)]:STACK_SIZE: ' \
|
||||||
'-M+[+/-n matches files modified more/less than n days ago , and n matches files modified exactly n days ago, days are rounded down.That is +n => (−∞, curr−(n+1)), n => \[curr−(n+1), curr−n), and -n => (𝑐𝑢𝑟𝑟−𝑛, +∞)]: :_default' \
|
'-M+[+/-n matches files modified more/less than n days ago , and n matches files modified exactly n days ago, days are rounded down.That is +n => (−∞, curr−(n+1)), n => \[curr−(n+1), curr−n), and -n => (𝑐𝑢𝑟𝑟−𝑛, +∞)]: : ' \
|
||||||
'--mtime=[+/-n matches files modified more/less than n days ago , and n matches files modified exactly n days ago, days are rounded down.That is +n => (−∞, curr−(n+1)), n => \[curr−(n+1), curr−n), and -n => (𝑐𝑢𝑟𝑟−𝑛, +∞)]: :_default' \
|
'--mtime=[+/-n matches files modified more/less than n days ago , and n matches files modified exactly n days ago, days are rounded down.That is +n => (−∞, curr−(n+1)), n => \[curr−(n+1), curr−n), and -n => (𝑐𝑢𝑟𝑟−𝑛, +∞)]: : ' \
|
||||||
'-A+[just like -mtime, but based on file access time]: :_default' \
|
'-A+[just like -mtime, but based on file access time]: : ' \
|
||||||
'--atime=[just like -mtime, but based on file access time]: :_default' \
|
'--atime=[just like -mtime, but based on file access time]: : ' \
|
||||||
'-y+[just like -mtime, but based on file change time]: :_default' \
|
'-y+[just like -mtime, but based on file change time]: : ' \
|
||||||
'--ctime=[just like -mtime, but based on file change time]: :_default' \
|
'--ctime=[just like -mtime, but based on file change time]: : ' \
|
||||||
'--files0-from=[run dust on NUL-terminated file names specified in file; if argument is -, then read names from standard input]: :_files' \
|
'--files0-from=[run dust on NUL-terminated file names specified in file; if argument is -, then read names from standard input]: :_files' \
|
||||||
'*--collapse=[Keep these directories collapsed]: :_files' \
|
|
||||||
'-m+[Directory '\''size'\'' is max filetime of child files instead of disk size. while a/c/m for last accessed/changed/modified time]: :(a c m)' \
|
|
||||||
'--filetime=[Directory '\''size'\'' is max filetime of child files instead of disk size. while a/c/m for last accessed/changed/modified time]: :(a c m)' \
|
|
||||||
'-p[Subdirectories will not have their path shortened]' \
|
'-p[Subdirectories will not have their path shortened]' \
|
||||||
'--full-paths[Subdirectories will not have their path shortened]' \
|
'--full-paths[Subdirectories will not have their path shortened]' \
|
||||||
'-L[dereference sym links - Treat sym links as directories and go into them]' \
|
'-L[dereference sym links - Treat sym links as directories and go into them]' \
|
||||||
@@ -72,14 +68,14 @@ _dust() {
|
|||||||
'-f[Directory '\''size'\'' is number of child files instead of disk size]' \
|
'-f[Directory '\''size'\'' is number of child files instead of disk size]' \
|
||||||
'--filecount[Directory '\''size'\'' is number of child files instead of disk size]' \
|
'--filecount[Directory '\''size'\'' is number of child files instead of disk size]' \
|
||||||
'-i[Do not display hidden files]' \
|
'-i[Do not display hidden files]' \
|
||||||
'--ignore-hidden[Do not display hidden files]' \
|
'--ignore_hidden[Do not display hidden files]' \
|
||||||
'(-d --depth -D --only-dir)-t[show only these file types]' \
|
'(-d --depth -D --only-dir)-t[show only these file types]' \
|
||||||
'(-d --depth -D --only-dir)--file-types[show only these file types]' \
|
'(-d --depth -D --only-dir)--file_types[show only these file types]' \
|
||||||
'-P[Disable the progress indication.]' \
|
'-P[Disable the progress indication.]' \
|
||||||
'--no-progress[Disable the progress indication.]' \
|
'--no-progress[Disable the progress indication.]' \
|
||||||
'--print-errors[Print path with errors.]' \
|
'--print-errors[Print path with errors.]' \
|
||||||
'(-F --only-file -t --file-types)-D[Only directories will be displayed.]' \
|
'(-F --only-file -t --file_types)-D[Only directories will be displayed.]' \
|
||||||
'(-F --only-file -t --file-types)--only-dir[Only directories will be displayed.]' \
|
'(-F --only-file -t --file_types)--only-dir[Only directories will be displayed.]' \
|
||||||
'(-D --only-dir)-F[Only files will be displayed. (Finds your largest files)]' \
|
'(-D --only-dir)-F[Only files will be displayed. (Finds your largest files)]' \
|
||||||
'(-D --only-dir)--only-file[Only files will be displayed. (Finds your largest files)]' \
|
'(-D --only-dir)--only-file[Only files will be displayed. (Finds your largest files)]' \
|
||||||
'-j[Output the directory tree as json to the current directory]' \
|
'-j[Output the directory tree as json to the current directory]' \
|
||||||
|
|||||||
@@ -21,79 +21,75 @@ Register-ArgumentCompleter -Native -CommandName 'dust' -ScriptBlock {
|
|||||||
|
|
||||||
$completions = @(switch ($command) {
|
$completions = @(switch ($command) {
|
||||||
'dust' {
|
'dust' {
|
||||||
[CompletionResult]::new('-d', '-d', [CompletionResultType]::ParameterName, 'Depth to show')
|
[CompletionResult]::new('-d', 'd', [CompletionResultType]::ParameterName, 'Depth to show')
|
||||||
[CompletionResult]::new('--depth', '--depth', [CompletionResultType]::ParameterName, 'Depth to show')
|
[CompletionResult]::new('--depth', 'depth', [CompletionResultType]::ParameterName, 'Depth to show')
|
||||||
[CompletionResult]::new('-T', '-T ', [CompletionResultType]::ParameterName, 'Number of threads to use')
|
[CompletionResult]::new('-T', 'T ', [CompletionResultType]::ParameterName, 'Number of threads to use')
|
||||||
[CompletionResult]::new('--threads', '--threads', [CompletionResultType]::ParameterName, 'Number of threads to use')
|
[CompletionResult]::new('--threads', 'threads', [CompletionResultType]::ParameterName, 'Number of threads to use')
|
||||||
[CompletionResult]::new('--config', '--config', [CompletionResultType]::ParameterName, 'Specify a config file to use')
|
[CompletionResult]::new('-n', 'n', [CompletionResultType]::ParameterName, 'Number of lines of output to show. (Default is terminal_height - 10)')
|
||||||
[CompletionResult]::new('-n', '-n', [CompletionResultType]::ParameterName, 'Number of lines of output to show. (Default is terminal_height - 10)')
|
[CompletionResult]::new('--number-of-lines', 'number-of-lines', [CompletionResultType]::ParameterName, 'Number of lines of output to show. (Default is terminal_height - 10)')
|
||||||
[CompletionResult]::new('--number-of-lines', '--number-of-lines', [CompletionResultType]::ParameterName, 'Number of lines of output to show. (Default is terminal_height - 10)')
|
[CompletionResult]::new('-X', 'X ', [CompletionResultType]::ParameterName, 'Exclude any file or directory with this name')
|
||||||
[CompletionResult]::new('-X', '-X ', [CompletionResultType]::ParameterName, 'Exclude any file or directory with this path')
|
[CompletionResult]::new('--ignore-directory', 'ignore-directory', [CompletionResultType]::ParameterName, 'Exclude any file or directory with this name')
|
||||||
[CompletionResult]::new('--ignore-directory', '--ignore-directory', [CompletionResultType]::ParameterName, 'Exclude any file or directory with this path')
|
[CompletionResult]::new('-I', 'I ', [CompletionResultType]::ParameterName, 'Exclude any file or directory with a regex matching that listed in this file, the file entries will be added to the ignore regexs provided by --invert_filter')
|
||||||
[CompletionResult]::new('-I', '-I ', [CompletionResultType]::ParameterName, 'Exclude any file or directory with a regex matching that listed in this file, the file entries will be added to the ignore regexs provided by --invert_filter')
|
[CompletionResult]::new('--ignore-all-in-file', 'ignore-all-in-file', [CompletionResultType]::ParameterName, 'Exclude any file or directory with a regex matching that listed in this file, the file entries will be added to the ignore regexs provided by --invert_filter')
|
||||||
[CompletionResult]::new('--ignore-all-in-file', '--ignore-all-in-file', [CompletionResultType]::ParameterName, 'Exclude any file or directory with a regex matching that listed in this file, the file entries will be added to the ignore regexs provided by --invert_filter')
|
[CompletionResult]::new('-z', 'z', [CompletionResultType]::ParameterName, 'Minimum size file to include in output')
|
||||||
[CompletionResult]::new('-z', '-z', [CompletionResultType]::ParameterName, 'Minimum size file to include in output')
|
[CompletionResult]::new('--min-size', 'min-size', [CompletionResultType]::ParameterName, 'Minimum size file to include in output')
|
||||||
[CompletionResult]::new('--min-size', '--min-size', [CompletionResultType]::ParameterName, 'Minimum size file to include in output')
|
[CompletionResult]::new('-v', 'v', [CompletionResultType]::ParameterName, 'Exclude filepaths matching this regex. To ignore png files type: -v "\.png$" ')
|
||||||
[CompletionResult]::new('-v', '-v', [CompletionResultType]::ParameterName, 'Exclude filepaths matching this regex. To ignore png files type: -v "\.png$" ')
|
[CompletionResult]::new('--invert-filter', 'invert-filter', [CompletionResultType]::ParameterName, 'Exclude filepaths matching this regex. To ignore png files type: -v "\.png$" ')
|
||||||
[CompletionResult]::new('--invert-filter', '--invert-filter', [CompletionResultType]::ParameterName, 'Exclude filepaths matching this regex. To ignore png files type: -v "\.png$" ')
|
[CompletionResult]::new('-e', 'e', [CompletionResultType]::ParameterName, 'Only include filepaths matching this regex. For png files type: -e "\.png$" ')
|
||||||
[CompletionResult]::new('-e', '-e', [CompletionResultType]::ParameterName, 'Only include filepaths matching this regex. For png files type: -e "\.png$" ')
|
[CompletionResult]::new('--filter', 'filter', [CompletionResultType]::ParameterName, 'Only include filepaths matching this regex. For png files type: -e "\.png$" ')
|
||||||
[CompletionResult]::new('--filter', '--filter', [CompletionResultType]::ParameterName, 'Only include filepaths matching this regex. For png files type: -e "\.png$" ')
|
[CompletionResult]::new('-w', 'w', [CompletionResultType]::ParameterName, 'Specify width of output overriding the auto detection of terminal width')
|
||||||
[CompletionResult]::new('-w', '-w', [CompletionResultType]::ParameterName, 'Specify width of output overriding the auto detection of terminal width')
|
[CompletionResult]::new('--terminal_width', 'terminal_width', [CompletionResultType]::ParameterName, 'Specify width of output overriding the auto detection of terminal width')
|
||||||
[CompletionResult]::new('--terminal-width', '--terminal-width', [CompletionResultType]::ParameterName, 'Specify width of output overriding the auto detection of terminal width')
|
[CompletionResult]::new('-o', 'o', [CompletionResultType]::ParameterName, 'Changes output display size. si will print sizes in powers of 1000. b k m g t kb mb gb tb will print the whole tree in that size.')
|
||||||
[CompletionResult]::new('-o', '-o', [CompletionResultType]::ParameterName, 'Changes output display size. si will print sizes in powers of 1000. b k m g t kb mb gb tb will print the whole tree in that size.')
|
[CompletionResult]::new('--output-format', 'output-format', [CompletionResultType]::ParameterName, 'Changes output display size. si will print sizes in powers of 1000. b k m g t kb mb gb tb will print the whole tree in that size.')
|
||||||
[CompletionResult]::new('--output-format', '--output-format', [CompletionResultType]::ParameterName, 'Changes output display size. si will print sizes in powers of 1000. b k m g t kb mb gb tb will print the whole tree in that size.')
|
[CompletionResult]::new('-S', 'S ', [CompletionResultType]::ParameterName, 'Specify memory to use as stack size - use if you see: ''fatal runtime error: stack overflow'' (default low memory=1048576, high memory=1073741824)')
|
||||||
[CompletionResult]::new('-S', '-S ', [CompletionResultType]::ParameterName, 'Specify memory to use as stack size - use if you see: ''fatal runtime error: stack overflow'' (default low memory=1048576, high memory=1073741824)')
|
[CompletionResult]::new('--stack-size', 'stack-size', [CompletionResultType]::ParameterName, 'Specify memory to use as stack size - use if you see: ''fatal runtime error: stack overflow'' (default low memory=1048576, high memory=1073741824)')
|
||||||
[CompletionResult]::new('--stack-size', '--stack-size', [CompletionResultType]::ParameterName, 'Specify memory to use as stack size - use if you see: ''fatal runtime error: stack overflow'' (default low memory=1048576, high memory=1073741824)')
|
[CompletionResult]::new('-M', 'M ', [CompletionResultType]::ParameterName, '+/-n matches files modified more/less than n days ago , and n matches files modified exactly n days ago, days are rounded down.That is +n => (−∞, curr−(n+1)), n => [curr−(n+1), curr−n), and -n => (𝑐𝑢𝑟𝑟−𝑛, +∞)')
|
||||||
[CompletionResult]::new('-M', '-M ', [CompletionResultType]::ParameterName, '+/-n matches files modified more/less than n days ago , and n matches files modified exactly n days ago, days are rounded down.That is +n => (−∞, curr−(n+1)), n => [curr−(n+1), curr−n), and -n => (𝑐𝑢𝑟𝑟−𝑛, +∞)')
|
[CompletionResult]::new('--mtime', 'mtime', [CompletionResultType]::ParameterName, '+/-n matches files modified more/less than n days ago , and n matches files modified exactly n days ago, days are rounded down.That is +n => (−∞, curr−(n+1)), n => [curr−(n+1), curr−n), and -n => (𝑐𝑢𝑟𝑟−𝑛, +∞)')
|
||||||
[CompletionResult]::new('--mtime', '--mtime', [CompletionResultType]::ParameterName, '+/-n matches files modified more/less than n days ago , and n matches files modified exactly n days ago, days are rounded down.That is +n => (−∞, curr−(n+1)), n => [curr−(n+1), curr−n), and -n => (𝑐𝑢𝑟𝑟−𝑛, +∞)')
|
[CompletionResult]::new('-A', 'A ', [CompletionResultType]::ParameterName, 'just like -mtime, but based on file access time')
|
||||||
[CompletionResult]::new('-A', '-A ', [CompletionResultType]::ParameterName, 'just like -mtime, but based on file access time')
|
[CompletionResult]::new('--atime', 'atime', [CompletionResultType]::ParameterName, 'just like -mtime, but based on file access time')
|
||||||
[CompletionResult]::new('--atime', '--atime', [CompletionResultType]::ParameterName, 'just like -mtime, but based on file access time')
|
[CompletionResult]::new('-y', 'y', [CompletionResultType]::ParameterName, 'just like -mtime, but based on file change time')
|
||||||
[CompletionResult]::new('-y', '-y', [CompletionResultType]::ParameterName, 'just like -mtime, but based on file change time')
|
[CompletionResult]::new('--ctime', 'ctime', [CompletionResultType]::ParameterName, 'just like -mtime, but based on file change time')
|
||||||
[CompletionResult]::new('--ctime', '--ctime', [CompletionResultType]::ParameterName, 'just like -mtime, but based on file change time')
|
[CompletionResult]::new('--files0-from', 'files0-from', [CompletionResultType]::ParameterName, 'run dust on NUL-terminated file names specified in file; if argument is -, then read names from standard input')
|
||||||
[CompletionResult]::new('--files0-from', '--files0-from', [CompletionResultType]::ParameterName, 'run dust on NUL-terminated file names specified in file; if argument is -, then read names from standard input')
|
[CompletionResult]::new('-p', 'p', [CompletionResultType]::ParameterName, 'Subdirectories will not have their path shortened')
|
||||||
[CompletionResult]::new('--collapse', '--collapse', [CompletionResultType]::ParameterName, 'Keep these directories collapsed')
|
[CompletionResult]::new('--full-paths', 'full-paths', [CompletionResultType]::ParameterName, 'Subdirectories will not have their path shortened')
|
||||||
[CompletionResult]::new('-m', '-m', [CompletionResultType]::ParameterName, 'Directory ''size'' is max filetime of child files instead of disk size. while a/c/m for last accessed/changed/modified time')
|
[CompletionResult]::new('-L', 'L ', [CompletionResultType]::ParameterName, 'dereference sym links - Treat sym links as directories and go into them')
|
||||||
[CompletionResult]::new('--filetime', '--filetime', [CompletionResultType]::ParameterName, 'Directory ''size'' is max filetime of child files instead of disk size. while a/c/m for last accessed/changed/modified time')
|
[CompletionResult]::new('--dereference-links', 'dereference-links', [CompletionResultType]::ParameterName, 'dereference sym links - Treat sym links as directories and go into them')
|
||||||
[CompletionResult]::new('-p', '-p', [CompletionResultType]::ParameterName, 'Subdirectories will not have their path shortened')
|
[CompletionResult]::new('-x', 'x', [CompletionResultType]::ParameterName, 'Only count the files and directories on the same filesystem as the supplied directory')
|
||||||
[CompletionResult]::new('--full-paths', '--full-paths', [CompletionResultType]::ParameterName, 'Subdirectories will not have their path shortened')
|
[CompletionResult]::new('--limit-filesystem', 'limit-filesystem', [CompletionResultType]::ParameterName, 'Only count the files and directories on the same filesystem as the supplied directory')
|
||||||
[CompletionResult]::new('-L', '-L ', [CompletionResultType]::ParameterName, 'dereference sym links - Treat sym links as directories and go into them')
|
[CompletionResult]::new('-s', 's', [CompletionResultType]::ParameterName, 'Use file length instead of blocks')
|
||||||
[CompletionResult]::new('--dereference-links', '--dereference-links', [CompletionResultType]::ParameterName, 'dereference sym links - Treat sym links as directories and go into them')
|
[CompletionResult]::new('--apparent-size', 'apparent-size', [CompletionResultType]::ParameterName, 'Use file length instead of blocks')
|
||||||
[CompletionResult]::new('-x', '-x', [CompletionResultType]::ParameterName, 'Only count the files and directories on the same filesystem as the supplied directory')
|
[CompletionResult]::new('-r', 'r', [CompletionResultType]::ParameterName, 'Print tree upside down (biggest highest)')
|
||||||
[CompletionResult]::new('--limit-filesystem', '--limit-filesystem', [CompletionResultType]::ParameterName, 'Only count the files and directories on the same filesystem as the supplied directory')
|
[CompletionResult]::new('--reverse', 'reverse', [CompletionResultType]::ParameterName, 'Print tree upside down (biggest highest)')
|
||||||
[CompletionResult]::new('-s', '-s', [CompletionResultType]::ParameterName, 'Use file length instead of blocks')
|
[CompletionResult]::new('-c', 'c', [CompletionResultType]::ParameterName, 'No colors will be printed (Useful for commands like: watch)')
|
||||||
[CompletionResult]::new('--apparent-size', '--apparent-size', [CompletionResultType]::ParameterName, 'Use file length instead of blocks')
|
[CompletionResult]::new('--no-colors', 'no-colors', [CompletionResultType]::ParameterName, 'No colors will be printed (Useful for commands like: watch)')
|
||||||
[CompletionResult]::new('-r', '-r', [CompletionResultType]::ParameterName, 'Print tree upside down (biggest highest)')
|
[CompletionResult]::new('-C', 'C ', [CompletionResultType]::ParameterName, 'Force colors print')
|
||||||
[CompletionResult]::new('--reverse', '--reverse', [CompletionResultType]::ParameterName, 'Print tree upside down (biggest highest)')
|
[CompletionResult]::new('--force-colors', 'force-colors', [CompletionResultType]::ParameterName, 'Force colors print')
|
||||||
[CompletionResult]::new('-c', '-c', [CompletionResultType]::ParameterName, 'No colors will be printed (Useful for commands like: watch)')
|
[CompletionResult]::new('-b', 'b', [CompletionResultType]::ParameterName, 'No percent bars or percentages will be displayed')
|
||||||
[CompletionResult]::new('--no-colors', '--no-colors', [CompletionResultType]::ParameterName, 'No colors will be printed (Useful for commands like: watch)')
|
[CompletionResult]::new('--no-percent-bars', 'no-percent-bars', [CompletionResultType]::ParameterName, 'No percent bars or percentages will be displayed')
|
||||||
[CompletionResult]::new('-C', '-C ', [CompletionResultType]::ParameterName, 'Force colors print')
|
[CompletionResult]::new('-B', 'B ', [CompletionResultType]::ParameterName, 'percent bars moved to right side of screen')
|
||||||
[CompletionResult]::new('--force-colors', '--force-colors', [CompletionResultType]::ParameterName, 'Force colors print')
|
[CompletionResult]::new('--bars-on-right', 'bars-on-right', [CompletionResultType]::ParameterName, 'percent bars moved to right side of screen')
|
||||||
[CompletionResult]::new('-b', '-b', [CompletionResultType]::ParameterName, 'No percent bars or percentages will be displayed')
|
[CompletionResult]::new('-R', 'R ', [CompletionResultType]::ParameterName, 'For screen readers. Removes bars. Adds new column: depth level (May want to use -p too for full path)')
|
||||||
[CompletionResult]::new('--no-percent-bars', '--no-percent-bars', [CompletionResultType]::ParameterName, 'No percent bars or percentages will be displayed')
|
[CompletionResult]::new('--screen-reader', 'screen-reader', [CompletionResultType]::ParameterName, 'For screen readers. Removes bars. Adds new column: depth level (May want to use -p too for full path)')
|
||||||
[CompletionResult]::new('-B', '-B ', [CompletionResultType]::ParameterName, 'percent bars moved to right side of screen')
|
[CompletionResult]::new('--skip-total', 'skip-total', [CompletionResultType]::ParameterName, 'No total row will be displayed')
|
||||||
[CompletionResult]::new('--bars-on-right', '--bars-on-right', [CompletionResultType]::ParameterName, 'percent bars moved to right side of screen')
|
[CompletionResult]::new('-f', 'f', [CompletionResultType]::ParameterName, 'Directory ''size'' is number of child files instead of disk size')
|
||||||
[CompletionResult]::new('-R', '-R ', [CompletionResultType]::ParameterName, 'For screen readers. Removes bars. Adds new column: depth level (May want to use -p too for full path)')
|
[CompletionResult]::new('--filecount', 'filecount', [CompletionResultType]::ParameterName, 'Directory ''size'' is number of child files instead of disk size')
|
||||||
[CompletionResult]::new('--screen-reader', '--screen-reader', [CompletionResultType]::ParameterName, 'For screen readers. Removes bars. Adds new column: depth level (May want to use -p too for full path)')
|
[CompletionResult]::new('-i', 'i', [CompletionResultType]::ParameterName, 'Do not display hidden files')
|
||||||
[CompletionResult]::new('--skip-total', '--skip-total', [CompletionResultType]::ParameterName, 'No total row will be displayed')
|
[CompletionResult]::new('--ignore_hidden', 'ignore_hidden', [CompletionResultType]::ParameterName, 'Do not display hidden files')
|
||||||
[CompletionResult]::new('-f', '-f', [CompletionResultType]::ParameterName, 'Directory ''size'' is number of child files instead of disk size')
|
[CompletionResult]::new('-t', 't', [CompletionResultType]::ParameterName, 'show only these file types')
|
||||||
[CompletionResult]::new('--filecount', '--filecount', [CompletionResultType]::ParameterName, 'Directory ''size'' is number of child files instead of disk size')
|
[CompletionResult]::new('--file_types', 'file_types', [CompletionResultType]::ParameterName, 'show only these file types')
|
||||||
[CompletionResult]::new('-i', '-i', [CompletionResultType]::ParameterName, 'Do not display hidden files')
|
[CompletionResult]::new('-P', 'P ', [CompletionResultType]::ParameterName, 'Disable the progress indication.')
|
||||||
[CompletionResult]::new('--ignore-hidden', '--ignore-hidden', [CompletionResultType]::ParameterName, 'Do not display hidden files')
|
[CompletionResult]::new('--no-progress', 'no-progress', [CompletionResultType]::ParameterName, 'Disable the progress indication.')
|
||||||
[CompletionResult]::new('-t', '-t', [CompletionResultType]::ParameterName, 'show only these file types')
|
[CompletionResult]::new('--print-errors', 'print-errors', [CompletionResultType]::ParameterName, 'Print path with errors.')
|
||||||
[CompletionResult]::new('--file-types', '--file-types', [CompletionResultType]::ParameterName, 'show only these file types')
|
[CompletionResult]::new('-D', 'D ', [CompletionResultType]::ParameterName, 'Only directories will be displayed.')
|
||||||
[CompletionResult]::new('-P', '-P ', [CompletionResultType]::ParameterName, 'Disable the progress indication.')
|
[CompletionResult]::new('--only-dir', 'only-dir', [CompletionResultType]::ParameterName, 'Only directories will be displayed.')
|
||||||
[CompletionResult]::new('--no-progress', '--no-progress', [CompletionResultType]::ParameterName, 'Disable the progress indication.')
|
[CompletionResult]::new('-F', 'F ', [CompletionResultType]::ParameterName, 'Only files will be displayed. (Finds your largest files)')
|
||||||
[CompletionResult]::new('--print-errors', '--print-errors', [CompletionResultType]::ParameterName, 'Print path with errors.')
|
[CompletionResult]::new('--only-file', 'only-file', [CompletionResultType]::ParameterName, 'Only files will be displayed. (Finds your largest files)')
|
||||||
[CompletionResult]::new('-D', '-D ', [CompletionResultType]::ParameterName, 'Only directories will be displayed.')
|
[CompletionResult]::new('-j', 'j', [CompletionResultType]::ParameterName, 'Output the directory tree as json to the current directory')
|
||||||
[CompletionResult]::new('--only-dir', '--only-dir', [CompletionResultType]::ParameterName, 'Only directories will be displayed.')
|
[CompletionResult]::new('--output-json', 'output-json', [CompletionResultType]::ParameterName, 'Output the directory tree as json to the current directory')
|
||||||
[CompletionResult]::new('-F', '-F ', [CompletionResultType]::ParameterName, 'Only files will be displayed. (Finds your largest files)')
|
[CompletionResult]::new('-h', 'h', [CompletionResultType]::ParameterName, 'Print help')
|
||||||
[CompletionResult]::new('--only-file', '--only-file', [CompletionResultType]::ParameterName, 'Only files will be displayed. (Finds your largest files)')
|
[CompletionResult]::new('--help', 'help', [CompletionResultType]::ParameterName, 'Print help')
|
||||||
[CompletionResult]::new('-j', '-j', [CompletionResultType]::ParameterName, 'Output the directory tree as json to the current directory')
|
[CompletionResult]::new('-V', 'V ', [CompletionResultType]::ParameterName, 'Print version')
|
||||||
[CompletionResult]::new('--output-json', '--output-json', [CompletionResultType]::ParameterName, 'Output the directory tree as json to the current directory')
|
[CompletionResult]::new('--version', 'version', [CompletionResultType]::ParameterName, 'Print version')
|
||||||
[CompletionResult]::new('-h', '-h', [CompletionResultType]::ParameterName, 'Print help')
|
|
||||||
[CompletionResult]::new('--help', '--help', [CompletionResultType]::ParameterName, 'Print help')
|
|
||||||
[CompletionResult]::new('-V', '-V ', [CompletionResultType]::ParameterName, 'Print version')
|
|
||||||
[CompletionResult]::new('--version', '--version', [CompletionResultType]::ParameterName, 'Print version')
|
|
||||||
break
|
break
|
||||||
}
|
}
|
||||||
})
|
})
|
||||||
|
|||||||
@@ -19,7 +19,7 @@ _dust() {
|
|||||||
|
|
||||||
case "${cmd}" in
|
case "${cmd}" in
|
||||||
dust)
|
dust)
|
||||||
opts="-d -T -n -p -X -I -L -x -s -r -c -C -b -B -z -R -f -i -v -e -t -w -P -D -F -o -S -j -M -A -y -m -h -V --depth --threads --config --number-of-lines --full-paths --ignore-directory --ignore-all-in-file --dereference-links --limit-filesystem --apparent-size --reverse --no-colors --force-colors --no-percent-bars --bars-on-right --min-size --screen-reader --skip-total --filecount --ignore-hidden --invert-filter --filter --file-types --terminal-width --no-progress --print-errors --only-dir --only-file --output-format --stack-size --output-json --mtime --atime --ctime --files0-from --collapse --filetime --help --version [PATH]..."
|
opts="-d -T -n -p -X -I -L -x -s -r -c -C -b -B -z -R -f -i -v -e -t -w -P -D -F -o -S -j -M -A -y -h -V --depth --threads --number-of-lines --full-paths --ignore-directory --ignore-all-in-file --dereference-links --limit-filesystem --apparent-size --reverse --no-colors --force-colors --no-percent-bars --bars-on-right --min-size --screen-reader --skip-total --filecount --ignore_hidden --invert-filter --filter --file_types --terminal_width --no-progress --print-errors --only-dir --only-file --output-format --stack-size --output-json --mtime --atime --ctime --files0-from --help --version [PATH]..."
|
||||||
if [[ ${cur} == -* || ${COMP_CWORD} -eq 1 ]] ; then
|
if [[ ${cur} == -* || ${COMP_CWORD} -eq 1 ]] ; then
|
||||||
COMPREPLY=( $(compgen -W "${opts}" -- "${cur}") )
|
COMPREPLY=( $(compgen -W "${opts}" -- "${cur}") )
|
||||||
return 0
|
return 0
|
||||||
@@ -41,21 +41,6 @@ _dust() {
|
|||||||
COMPREPLY=($(compgen -f "${cur}"))
|
COMPREPLY=($(compgen -f "${cur}"))
|
||||||
return 0
|
return 0
|
||||||
;;
|
;;
|
||||||
--config)
|
|
||||||
local oldifs
|
|
||||||
if [ -n "${IFS+x}" ]; then
|
|
||||||
oldifs="$IFS"
|
|
||||||
fi
|
|
||||||
IFS=$'\n'
|
|
||||||
COMPREPLY=($(compgen -f "${cur}"))
|
|
||||||
if [ -n "${oldifs+x}" ]; then
|
|
||||||
IFS="$oldifs"
|
|
||||||
fi
|
|
||||||
if [[ "${BASH_VERSINFO[0]}" -ge 4 ]]; then
|
|
||||||
compopt -o filenames
|
|
||||||
fi
|
|
||||||
return 0
|
|
||||||
;;
|
|
||||||
--number-of-lines)
|
--number-of-lines)
|
||||||
COMPREPLY=($(compgen -f "${cur}"))
|
COMPREPLY=($(compgen -f "${cur}"))
|
||||||
return 0
|
return 0
|
||||||
@@ -74,12 +59,12 @@ _dust() {
|
|||||||
;;
|
;;
|
||||||
--ignore-all-in-file)
|
--ignore-all-in-file)
|
||||||
local oldifs
|
local oldifs
|
||||||
if [ -n "${IFS+x}" ]; then
|
if [[ -v IFS ]]; then
|
||||||
oldifs="$IFS"
|
oldifs="$IFS"
|
||||||
fi
|
fi
|
||||||
IFS=$'\n'
|
IFS=$'\n'
|
||||||
COMPREPLY=($(compgen -f "${cur}"))
|
COMPREPLY=($(compgen -f "${cur}"))
|
||||||
if [ -n "${oldifs+x}" ]; then
|
if [[ -v oldifs ]]; then
|
||||||
IFS="$oldifs"
|
IFS="$oldifs"
|
||||||
fi
|
fi
|
||||||
if [[ "${BASH_VERSINFO[0]}" -ge 4 ]]; then
|
if [[ "${BASH_VERSINFO[0]}" -ge 4 ]]; then
|
||||||
@@ -89,12 +74,12 @@ _dust() {
|
|||||||
;;
|
;;
|
||||||
-I)
|
-I)
|
||||||
local oldifs
|
local oldifs
|
||||||
if [ -n "${IFS+x}" ]; then
|
if [[ -v IFS ]]; then
|
||||||
oldifs="$IFS"
|
oldifs="$IFS"
|
||||||
fi
|
fi
|
||||||
IFS=$'\n'
|
IFS=$'\n'
|
||||||
COMPREPLY=($(compgen -f "${cur}"))
|
COMPREPLY=($(compgen -f "${cur}"))
|
||||||
if [ -n "${oldifs+x}" ]; then
|
if [[ -v oldifs ]]; then
|
||||||
IFS="$oldifs"
|
IFS="$oldifs"
|
||||||
fi
|
fi
|
||||||
if [[ "${BASH_VERSINFO[0]}" -ge 4 ]]; then
|
if [[ "${BASH_VERSINFO[0]}" -ge 4 ]]; then
|
||||||
@@ -126,7 +111,7 @@ _dust() {
|
|||||||
COMPREPLY=($(compgen -f "${cur}"))
|
COMPREPLY=($(compgen -f "${cur}"))
|
||||||
return 0
|
return 0
|
||||||
;;
|
;;
|
||||||
--terminal-width)
|
--terminal_width)
|
||||||
COMPREPLY=($(compgen -f "${cur}"))
|
COMPREPLY=($(compgen -f "${cur}"))
|
||||||
return 0
|
return 0
|
||||||
;;
|
;;
|
||||||
@@ -178,18 +163,6 @@ _dust() {
|
|||||||
COMPREPLY=($(compgen -f "${cur}"))
|
COMPREPLY=($(compgen -f "${cur}"))
|
||||||
return 0
|
return 0
|
||||||
;;
|
;;
|
||||||
--collapse)
|
|
||||||
COMPREPLY=($(compgen -f "${cur}"))
|
|
||||||
return 0
|
|
||||||
;;
|
|
||||||
--filetime)
|
|
||||||
COMPREPLY=($(compgen -W "a c m" -- "${cur}"))
|
|
||||||
return 0
|
|
||||||
;;
|
|
||||||
-m)
|
|
||||||
COMPREPLY=($(compgen -W "a c m" -- "${cur}"))
|
|
||||||
return 0
|
|
||||||
;;
|
|
||||||
*)
|
*)
|
||||||
COMPREPLY=()
|
COMPREPLY=()
|
||||||
;;
|
;;
|
||||||
|
|||||||
@@ -22,11 +22,10 @@ set edit:completion:arg-completer[dust] = {|@words|
|
|||||||
cand --depth 'Depth to show'
|
cand --depth 'Depth to show'
|
||||||
cand -T 'Number of threads to use'
|
cand -T 'Number of threads to use'
|
||||||
cand --threads 'Number of threads to use'
|
cand --threads 'Number of threads to use'
|
||||||
cand --config 'Specify a config file to use'
|
|
||||||
cand -n 'Number of lines of output to show. (Default is terminal_height - 10)'
|
cand -n 'Number of lines of output to show. (Default is terminal_height - 10)'
|
||||||
cand --number-of-lines 'Number of lines of output to show. (Default is terminal_height - 10)'
|
cand --number-of-lines 'Number of lines of output to show. (Default is terminal_height - 10)'
|
||||||
cand -X 'Exclude any file or directory with this path'
|
cand -X 'Exclude any file or directory with this name'
|
||||||
cand --ignore-directory 'Exclude any file or directory with this path'
|
cand --ignore-directory 'Exclude any file or directory with this name'
|
||||||
cand -I 'Exclude any file or directory with a regex matching that listed in this file, the file entries will be added to the ignore regexs provided by --invert_filter'
|
cand -I 'Exclude any file or directory with a regex matching that listed in this file, the file entries will be added to the ignore regexs provided by --invert_filter'
|
||||||
cand --ignore-all-in-file 'Exclude any file or directory with a regex matching that listed in this file, the file entries will be added to the ignore regexs provided by --invert_filter'
|
cand --ignore-all-in-file 'Exclude any file or directory with a regex matching that listed in this file, the file entries will be added to the ignore regexs provided by --invert_filter'
|
||||||
cand -z 'Minimum size file to include in output'
|
cand -z 'Minimum size file to include in output'
|
||||||
@@ -36,7 +35,7 @@ set edit:completion:arg-completer[dust] = {|@words|
|
|||||||
cand -e 'Only include filepaths matching this regex. For png files type: -e "\.png$" '
|
cand -e 'Only include filepaths matching this regex. For png files type: -e "\.png$" '
|
||||||
cand --filter 'Only include filepaths matching this regex. For png files type: -e "\.png$" '
|
cand --filter 'Only include filepaths matching this regex. For png files type: -e "\.png$" '
|
||||||
cand -w 'Specify width of output overriding the auto detection of terminal width'
|
cand -w 'Specify width of output overriding the auto detection of terminal width'
|
||||||
cand --terminal-width 'Specify width of output overriding the auto detection of terminal width'
|
cand --terminal_width 'Specify width of output overriding the auto detection of terminal width'
|
||||||
cand -o 'Changes output display size. si will print sizes in powers of 1000. b k m g t kb mb gb tb will print the whole tree in that size.'
|
cand -o 'Changes output display size. si will print sizes in powers of 1000. b k m g t kb mb gb tb will print the whole tree in that size.'
|
||||||
cand --output-format 'Changes output display size. si will print sizes in powers of 1000. b k m g t kb mb gb tb will print the whole tree in that size.'
|
cand --output-format 'Changes output display size. si will print sizes in powers of 1000. b k m g t kb mb gb tb will print the whole tree in that size.'
|
||||||
cand -S 'Specify memory to use as stack size - use if you see: ''fatal runtime error: stack overflow'' (default low memory=1048576, high memory=1073741824)'
|
cand -S 'Specify memory to use as stack size - use if you see: ''fatal runtime error: stack overflow'' (default low memory=1048576, high memory=1073741824)'
|
||||||
@@ -48,9 +47,6 @@ set edit:completion:arg-completer[dust] = {|@words|
|
|||||||
cand -y 'just like -mtime, but based on file change time'
|
cand -y 'just like -mtime, but based on file change time'
|
||||||
cand --ctime 'just like -mtime, but based on file change time'
|
cand --ctime 'just like -mtime, but based on file change time'
|
||||||
cand --files0-from 'run dust on NUL-terminated file names specified in file; if argument is -, then read names from standard input'
|
cand --files0-from 'run dust on NUL-terminated file names specified in file; if argument is -, then read names from standard input'
|
||||||
cand --collapse 'Keep these directories collapsed'
|
|
||||||
cand -m 'Directory ''size'' is max filetime of child files instead of disk size. while a/c/m for last accessed/changed/modified time'
|
|
||||||
cand --filetime 'Directory ''size'' is max filetime of child files instead of disk size. while a/c/m for last accessed/changed/modified time'
|
|
||||||
cand -p 'Subdirectories will not have their path shortened'
|
cand -p 'Subdirectories will not have their path shortened'
|
||||||
cand --full-paths 'Subdirectories will not have their path shortened'
|
cand --full-paths 'Subdirectories will not have their path shortened'
|
||||||
cand -L 'dereference sym links - Treat sym links as directories and go into them'
|
cand -L 'dereference sym links - Treat sym links as directories and go into them'
|
||||||
@@ -75,9 +71,9 @@ set edit:completion:arg-completer[dust] = {|@words|
|
|||||||
cand -f 'Directory ''size'' is number of child files instead of disk size'
|
cand -f 'Directory ''size'' is number of child files instead of disk size'
|
||||||
cand --filecount 'Directory ''size'' is number of child files instead of disk size'
|
cand --filecount 'Directory ''size'' is number of child files instead of disk size'
|
||||||
cand -i 'Do not display hidden files'
|
cand -i 'Do not display hidden files'
|
||||||
cand --ignore-hidden 'Do not display hidden files'
|
cand --ignore_hidden 'Do not display hidden files'
|
||||||
cand -t 'show only these file types'
|
cand -t 'show only these file types'
|
||||||
cand --file-types 'show only these file types'
|
cand --file_types 'show only these file types'
|
||||||
cand -P 'Disable the progress indication.'
|
cand -P 'Disable the progress indication.'
|
||||||
cand --no-progress 'Disable the progress indication.'
|
cand --no-progress 'Disable the progress indication.'
|
||||||
cand --print-errors 'Print path with errors.'
|
cand --print-errors 'Print path with errors.'
|
||||||
|
|||||||
@@ -1,32 +1,18 @@
|
|||||||
complete -c dust -s d -l depth -d 'Depth to show' -r
|
complete -c dust -s d -l depth -d 'Depth to show' -r
|
||||||
complete -c dust -s T -l threads -d 'Number of threads to use' -r
|
complete -c dust -s T -l threads -d 'Number of threads to use' -r
|
||||||
complete -c dust -l config -d 'Specify a config file to use' -r -F
|
|
||||||
complete -c dust -s n -l number-of-lines -d 'Number of lines of output to show. (Default is terminal_height - 10)' -r
|
complete -c dust -s n -l number-of-lines -d 'Number of lines of output to show. (Default is terminal_height - 10)' -r
|
||||||
complete -c dust -s X -l ignore-directory -d 'Exclude any file or directory with this path' -r -F
|
complete -c dust -s X -l ignore-directory -d 'Exclude any file or directory with this name' -r -F
|
||||||
complete -c dust -s I -l ignore-all-in-file -d 'Exclude any file or directory with a regex matching that listed in this file, the file entries will be added to the ignore regexs provided by --invert_filter' -r -F
|
complete -c dust -s I -l ignore-all-in-file -d 'Exclude any file or directory with a regex matching that listed in this file, the file entries will be added to the ignore regexs provided by --invert_filter' -r -F
|
||||||
complete -c dust -s z -l min-size -d 'Minimum size file to include in output' -r
|
complete -c dust -s z -l min-size -d 'Minimum size file to include in output' -r
|
||||||
complete -c dust -s v -l invert-filter -d 'Exclude filepaths matching this regex. To ignore png files type: -v "\\.png$" ' -r
|
complete -c dust -s v -l invert-filter -d 'Exclude filepaths matching this regex. To ignore png files type: -v "\\.png$" ' -r
|
||||||
complete -c dust -s e -l filter -d 'Only include filepaths matching this regex. For png files type: -e "\\.png$" ' -r
|
complete -c dust -s e -l filter -d 'Only include filepaths matching this regex. For png files type: -e "\\.png$" ' -r
|
||||||
complete -c dust -s w -l terminal-width -d 'Specify width of output overriding the auto detection of terminal width' -r
|
complete -c dust -s w -l terminal_width -d 'Specify width of output overriding the auto detection of terminal width' -r
|
||||||
complete -c dust -s o -l output-format -d 'Changes output display size. si will print sizes in powers of 1000. b k m g t kb mb gb tb will print the whole tree in that size.' -r -f -a "si\t''
|
complete -c dust -s o -l output-format -d 'Changes output display size. si will print sizes in powers of 1000. b k m g t kb mb gb tb will print the whole tree in that size.' -r -f -a "{si '',b '',k '',m '',g '',t '',kb '',mb '',gb '',tb ''}"
|
||||||
b\t''
|
|
||||||
k\t''
|
|
||||||
m\t''
|
|
||||||
g\t''
|
|
||||||
t\t''
|
|
||||||
kb\t''
|
|
||||||
mb\t''
|
|
||||||
gb\t''
|
|
||||||
tb\t''"
|
|
||||||
complete -c dust -s S -l stack-size -d 'Specify memory to use as stack size - use if you see: \'fatal runtime error: stack overflow\' (default low memory=1048576, high memory=1073741824)' -r
|
complete -c dust -s S -l stack-size -d 'Specify memory to use as stack size - use if you see: \'fatal runtime error: stack overflow\' (default low memory=1048576, high memory=1073741824)' -r
|
||||||
complete -c dust -s M -l mtime -d '+/-n matches files modified more/less than n days ago , and n matches files modified exactly n days ago, days are rounded down.That is +n => (−∞, curr−(n+1)), n => [curr−(n+1), curr−n), and -n => (𝑐𝑢𝑟𝑟−𝑛, +∞)' -r
|
complete -c dust -s M -l mtime -d '+/-n matches files modified more/less than n days ago , and n matches files modified exactly n days ago, days are rounded down.That is +n => (−∞, curr−(n+1)), n => [curr−(n+1), curr−n), and -n => (𝑐𝑢𝑟𝑟−𝑛, +∞)' -r
|
||||||
complete -c dust -s A -l atime -d 'just like -mtime, but based on file access time' -r
|
complete -c dust -s A -l atime -d 'just like -mtime, but based on file access time' -r
|
||||||
complete -c dust -s y -l ctime -d 'just like -mtime, but based on file change time' -r
|
complete -c dust -s y -l ctime -d 'just like -mtime, but based on file change time' -r
|
||||||
complete -c dust -l files0-from -d 'run dust on NUL-terminated file names specified in file; if argument is -, then read names from standard input' -r -F
|
complete -c dust -l files0-from -d 'run dust on NUL-terminated file names specified in file; if argument is -, then read names from standard input' -r -F
|
||||||
complete -c dust -l collapse -d 'Keep these directories collapsed' -r -F
|
|
||||||
complete -c dust -s m -l filetime -d 'Directory \'size\' is max filetime of child files instead of disk size. while a/c/m for last accessed/changed/modified time' -r -f -a "a\t''
|
|
||||||
c\t''
|
|
||||||
m\t''"
|
|
||||||
complete -c dust -s p -l full-paths -d 'Subdirectories will not have their path shortened'
|
complete -c dust -s p -l full-paths -d 'Subdirectories will not have their path shortened'
|
||||||
complete -c dust -s L -l dereference-links -d 'dereference sym links - Treat sym links as directories and go into them'
|
complete -c dust -s L -l dereference-links -d 'dereference sym links - Treat sym links as directories and go into them'
|
||||||
complete -c dust -s x -l limit-filesystem -d 'Only count the files and directories on the same filesystem as the supplied directory'
|
complete -c dust -s x -l limit-filesystem -d 'Only count the files and directories on the same filesystem as the supplied directory'
|
||||||
@@ -39,8 +25,8 @@ complete -c dust -s B -l bars-on-right -d 'percent bars moved to right side of s
|
|||||||
complete -c dust -s R -l screen-reader -d 'For screen readers. Removes bars. Adds new column: depth level (May want to use -p too for full path)'
|
complete -c dust -s R -l screen-reader -d 'For screen readers. Removes bars. Adds new column: depth level (May want to use -p too for full path)'
|
||||||
complete -c dust -l skip-total -d 'No total row will be displayed'
|
complete -c dust -l skip-total -d 'No total row will be displayed'
|
||||||
complete -c dust -s f -l filecount -d 'Directory \'size\' is number of child files instead of disk size'
|
complete -c dust -s f -l filecount -d 'Directory \'size\' is number of child files instead of disk size'
|
||||||
complete -c dust -s i -l ignore-hidden -d 'Do not display hidden files'
|
complete -c dust -s i -l ignore_hidden -d 'Do not display hidden files'
|
||||||
complete -c dust -s t -l file-types -d 'show only these file types'
|
complete -c dust -s t -l file_types -d 'show only these file types'
|
||||||
complete -c dust -s P -l no-progress -d 'Disable the progress indication.'
|
complete -c dust -s P -l no-progress -d 'Disable the progress indication.'
|
||||||
complete -c dust -l print-errors -d 'Print path with errors.'
|
complete -c dust -l print-errors -d 'Print path with errors.'
|
||||||
complete -c dust -s D -l only-dir -d 'Only directories will be displayed.'
|
complete -c dust -s D -l only-dir -d 'Only directories will be displayed.'
|
||||||
|
|||||||
@@ -25,4 +25,4 @@ skip-total=true
|
|||||||
ignore-hidden=true
|
ignore-hidden=true
|
||||||
|
|
||||||
# print sizes in powers of 1000 (e.g., 1.1G)
|
# print sizes in powers of 1000 (e.g., 1.1G)
|
||||||
output-format="si"
|
iso=true
|
||||||
|
|||||||
@@ -1,10 +1,10 @@
|
|||||||
.ie \n(.g .ds Aq \(aq
|
.ie \n(.g .ds Aq \(aq
|
||||||
.el .ds Aq '
|
.el .ds Aq '
|
||||||
.TH Dust 1 "Dust 1.2.0"
|
.TH Dust 1 "Dust 1.0.0"
|
||||||
.SH NAME
|
.SH NAME
|
||||||
Dust \- Like du but more intuitive
|
Dust \- Like du but more intuitive
|
||||||
.SH SYNOPSIS
|
.SH SYNOPSIS
|
||||||
\fBdust\fR [\fB\-d\fR|\fB\-\-depth\fR] [\fB\-T\fR|\fB\-\-threads\fR] [\fB\-\-config\fR] [\fB\-n\fR|\fB\-\-number\-of\-lines\fR] [\fB\-p\fR|\fB\-\-full\-paths\fR] [\fB\-X\fR|\fB\-\-ignore\-directory\fR] [\fB\-I\fR|\fB\-\-ignore\-all\-in\-file\fR] [\fB\-L\fR|\fB\-\-dereference\-links\fR] [\fB\-x\fR|\fB\-\-limit\-filesystem\fR] [\fB\-s\fR|\fB\-\-apparent\-size\fR] [\fB\-r\fR|\fB\-\-reverse\fR] [\fB\-c\fR|\fB\-\-no\-colors\fR] [\fB\-C\fR|\fB\-\-force\-colors\fR] [\fB\-b\fR|\fB\-\-no\-percent\-bars\fR] [\fB\-B\fR|\fB\-\-bars\-on\-right\fR] [\fB\-z\fR|\fB\-\-min\-size\fR] [\fB\-R\fR|\fB\-\-screen\-reader\fR] [\fB\-\-skip\-total\fR] [\fB\-f\fR|\fB\-\-filecount\fR] [\fB\-i\fR|\fB\-\-ignore\-hidden\fR] [\fB\-v\fR|\fB\-\-invert\-filter\fR] [\fB\-e\fR|\fB\-\-filter\fR] [\fB\-t\fR|\fB\-\-file\-types\fR] [\fB\-w\fR|\fB\-\-terminal\-width\fR] [\fB\-P\fR|\fB\-\-no\-progress\fR] [\fB\-\-print\-errors\fR] [\fB\-D\fR|\fB\-\-only\-dir\fR] [\fB\-F\fR|\fB\-\-only\-file\fR] [\fB\-o\fR|\fB\-\-output\-format\fR] [\fB\-S\fR|\fB\-\-stack\-size\fR] [\fB\-j\fR|\fB\-\-output\-json\fR] [\fB\-M\fR|\fB\-\-mtime\fR] [\fB\-A\fR|\fB\-\-atime\fR] [\fB\-y\fR|\fB\-\-ctime\fR] [\fB\-\-files0\-from\fR] [\fB\-\-collapse\fR] [\fB\-m\fR|\fB\-\-filetime\fR] [\fB\-h\fR|\fB\-\-help\fR] [\fB\-V\fR|\fB\-\-version\fR] [\fIPATH\fR]
|
\fBdust\fR [\fB\-d\fR|\fB\-\-depth\fR] [\fB\-T\fR|\fB\-\-threads\fR] [\fB\-n\fR|\fB\-\-number\-of\-lines\fR] [\fB\-p\fR|\fB\-\-full\-paths\fR] [\fB\-X\fR|\fB\-\-ignore\-directory\fR] [\fB\-I\fR|\fB\-\-ignore\-all\-in\-file\fR] [\fB\-L\fR|\fB\-\-dereference\-links\fR] [\fB\-x\fR|\fB\-\-limit\-filesystem\fR] [\fB\-s\fR|\fB\-\-apparent\-size\fR] [\fB\-r\fR|\fB\-\-reverse\fR] [\fB\-c\fR|\fB\-\-no\-colors\fR] [\fB\-C\fR|\fB\-\-force\-colors\fR] [\fB\-b\fR|\fB\-\-no\-percent\-bars\fR] [\fB\-B\fR|\fB\-\-bars\-on\-right\fR] [\fB\-z\fR|\fB\-\-min\-size\fR] [\fB\-R\fR|\fB\-\-screen\-reader\fR] [\fB\-\-skip\-total\fR] [\fB\-f\fR|\fB\-\-filecount\fR] [\fB\-i\fR|\fB\-\-ignore_hidden\fR] [\fB\-v\fR|\fB\-\-invert\-filter\fR] [\fB\-e\fR|\fB\-\-filter\fR] [\fB\-t\fR|\fB\-\-file_types\fR] [\fB\-w\fR|\fB\-\-terminal_width\fR] [\fB\-P\fR|\fB\-\-no\-progress\fR] [\fB\-\-print\-errors\fR] [\fB\-D\fR|\fB\-\-only\-dir\fR] [\fB\-F\fR|\fB\-\-only\-file\fR] [\fB\-o\fR|\fB\-\-output\-format\fR] [\fB\-S\fR|\fB\-\-stack\-size\fR] [\fB\-j\fR|\fB\-\-output\-json\fR] [\fB\-M\fR|\fB\-\-mtime\fR] [\fB\-A\fR|\fB\-\-atime\fR] [\fB\-y\fR|\fB\-\-ctime\fR] [\fB\-\-files0\-from\fR] [\fB\-h\fR|\fB\-\-help\fR] [\fB\-V\fR|\fB\-\-version\fR] [\fIPATH\fR]
|
||||||
.SH DESCRIPTION
|
.SH DESCRIPTION
|
||||||
Like du but more intuitive
|
Like du but more intuitive
|
||||||
.SH OPTIONS
|
.SH OPTIONS
|
||||||
@@ -15,9 +15,6 @@ Depth to show
|
|||||||
\fB\-T\fR, \fB\-\-threads\fR
|
\fB\-T\fR, \fB\-\-threads\fR
|
||||||
Number of threads to use
|
Number of threads to use
|
||||||
.TP
|
.TP
|
||||||
\fB\-\-config\fR=\fIFILE\fR
|
|
||||||
Specify a config file to use
|
|
||||||
.TP
|
|
||||||
\fB\-n\fR, \fB\-\-number\-of\-lines\fR=\fINUMBER\fR
|
\fB\-n\fR, \fB\-\-number\-of\-lines\fR=\fINUMBER\fR
|
||||||
Number of lines of output to show. (Default is terminal_height \- 10)
|
Number of lines of output to show. (Default is terminal_height \- 10)
|
||||||
.TP
|
.TP
|
||||||
@@ -25,7 +22,7 @@ Number of lines of output to show. (Default is terminal_height \- 10)
|
|||||||
Subdirectories will not have their path shortened
|
Subdirectories will not have their path shortened
|
||||||
.TP
|
.TP
|
||||||
\fB\-X\fR, \fB\-\-ignore\-directory\fR=\fIPATH\fR
|
\fB\-X\fR, \fB\-\-ignore\-directory\fR=\fIPATH\fR
|
||||||
Exclude any file or directory with this path
|
Exclude any file or directory with this name
|
||||||
.TP
|
.TP
|
||||||
\fB\-I\fR, \fB\-\-ignore\-all\-in\-file\fR=\fIFILE\fR
|
\fB\-I\fR, \fB\-\-ignore\-all\-in\-file\fR=\fIFILE\fR
|
||||||
Exclude any file or directory with a regex matching that listed in this file, the file entries will be added to the ignore regexs provided by \-\-invert_filter
|
Exclude any file or directory with a regex matching that listed in this file, the file entries will be added to the ignore regexs provided by \-\-invert_filter
|
||||||
@@ -66,7 +63,7 @@ No total row will be displayed
|
|||||||
\fB\-f\fR, \fB\-\-filecount\fR
|
\fB\-f\fR, \fB\-\-filecount\fR
|
||||||
Directory \*(Aqsize\*(Aq is number of child files instead of disk size
|
Directory \*(Aqsize\*(Aq is number of child files instead of disk size
|
||||||
.TP
|
.TP
|
||||||
\fB\-i\fR, \fB\-\-ignore\-hidden\fR
|
\fB\-i\fR, \fB\-\-ignore_hidden\fR
|
||||||
Do not display hidden files
|
Do not display hidden files
|
||||||
.TP
|
.TP
|
||||||
\fB\-v\fR, \fB\-\-invert\-filter\fR=\fIREGEX\fR
|
\fB\-v\fR, \fB\-\-invert\-filter\fR=\fIREGEX\fR
|
||||||
@@ -75,10 +72,10 @@ Exclude filepaths matching this regex. To ignore png files type: \-v "\\.png$"
|
|||||||
\fB\-e\fR, \fB\-\-filter\fR=\fIREGEX\fR
|
\fB\-e\fR, \fB\-\-filter\fR=\fIREGEX\fR
|
||||||
Only include filepaths matching this regex. For png files type: \-e "\\.png$"
|
Only include filepaths matching this regex. For png files type: \-e "\\.png$"
|
||||||
.TP
|
.TP
|
||||||
\fB\-t\fR, \fB\-\-file\-types\fR
|
\fB\-t\fR, \fB\-\-file_types\fR
|
||||||
show only these file types
|
show only these file types
|
||||||
.TP
|
.TP
|
||||||
\fB\-w\fR, \fB\-\-terminal\-width\fR=\fIWIDTH\fR
|
\fB\-w\fR, \fB\-\-terminal_width\fR=\fIWIDTH\fR
|
||||||
Specify width of output overriding the auto detection of terminal width
|
Specify width of output overriding the auto detection of terminal width
|
||||||
.TP
|
.TP
|
||||||
\fB\-P\fR, \fB\-\-no\-progress\fR
|
\fB\-P\fR, \fB\-\-no\-progress\fR
|
||||||
@@ -118,16 +115,6 @@ just like \-mtime, but based on file change time
|
|||||||
\fB\-\-files0\-from\fR
|
\fB\-\-files0\-from\fR
|
||||||
run dust on NUL\-terminated file names specified in file; if argument is \-, then read names from standard input
|
run dust on NUL\-terminated file names specified in file; if argument is \-, then read names from standard input
|
||||||
.TP
|
.TP
|
||||||
\fB\-\-collapse\fR
|
|
||||||
Keep these directories collapsed
|
|
||||||
.TP
|
|
||||||
\fB\-m\fR, \fB\-\-filetime\fR
|
|
||||||
Directory \*(Aqsize\*(Aq is max filetime of child files instead of disk size. while a/c/m for last accessed/changed/modified time
|
|
||||||
.br
|
|
||||||
|
|
||||||
.br
|
|
||||||
[\fIpossible values: \fRa, c, m]
|
|
||||||
.TP
|
|
||||||
\fB\-h\fR, \fB\-\-help\fR
|
\fB\-h\fR, \fB\-\-help\fR
|
||||||
Print help
|
Print help
|
||||||
.TP
|
.TP
|
||||||
@@ -137,4 +124,4 @@ Print version
|
|||||||
[\fIPATH\fR]
|
[\fIPATH\fR]
|
||||||
|
|
||||||
.SH VERSION
|
.SH VERSION
|
||||||
v1.2.0
|
v1.0.0
|
||||||
|
|||||||
39
src/cli.rs
39
src/cli.rs
@@ -1,4 +1,4 @@
|
|||||||
use clap::{Arg, Command, builder::PossibleValue, value_parser};
|
use clap::{builder::PossibleValue, value_parser, Arg, Command};
|
||||||
|
|
||||||
// For single thread mode set this variable on your command line:
|
// For single thread mode set this variable on your command line:
|
||||||
// export RAYON_NUM_THREADS=1
|
// export RAYON_NUM_THREADS=1
|
||||||
@@ -24,15 +24,6 @@ pub fn build_cli() -> Command {
|
|||||||
.help("Number of threads to use")
|
.help("Number of threads to use")
|
||||||
.num_args(1)
|
.num_args(1)
|
||||||
)
|
)
|
||||||
.arg(
|
|
||||||
Arg::new("config")
|
|
||||||
.long("config")
|
|
||||||
.help("Specify a config file to use")
|
|
||||||
.value_name("FILE")
|
|
||||||
.value_hint(clap::ValueHint::FilePath)
|
|
||||||
.value_parser(value_parser!(String))
|
|
||||||
.num_args(1)
|
|
||||||
)
|
|
||||||
.arg(
|
.arg(
|
||||||
Arg::new("number_of_lines")
|
Arg::new("number_of_lines")
|
||||||
.short('n')
|
.short('n')
|
||||||
@@ -56,7 +47,7 @@ pub fn build_cli() -> Command {
|
|||||||
.value_name("PATH")
|
.value_name("PATH")
|
||||||
.value_hint(clap::ValueHint::AnyPath)
|
.value_hint(clap::ValueHint::AnyPath)
|
||||||
.action(clap::ArgAction::Append)
|
.action(clap::ArgAction::Append)
|
||||||
.help("Exclude any file or directory with this path"),
|
.help("Exclude any file or directory with this name"),
|
||||||
)
|
)
|
||||||
.arg(
|
.arg(
|
||||||
Arg::new("ignore_all_in_file")
|
Arg::new("ignore_all_in_file")
|
||||||
@@ -154,7 +145,7 @@ pub fn build_cli() -> Command {
|
|||||||
.arg(
|
.arg(
|
||||||
Arg::new("ignore_hidden")
|
Arg::new("ignore_hidden")
|
||||||
.short('i') // Do not use 'h' this is used by 'help'
|
.short('i') // Do not use 'h' this is used by 'help'
|
||||||
.long("ignore-hidden")
|
.long("ignore_hidden")
|
||||||
.action(clap::ArgAction::SetTrue)
|
.action(clap::ArgAction::SetTrue)
|
||||||
.help("Do not display hidden files"),
|
.help("Do not display hidden files"),
|
||||||
)
|
)
|
||||||
@@ -180,7 +171,7 @@ pub fn build_cli() -> Command {
|
|||||||
.arg(
|
.arg(
|
||||||
Arg::new("types")
|
Arg::new("types")
|
||||||
.short('t')
|
.short('t')
|
||||||
.long("file-types")
|
.long("file_types")
|
||||||
.conflicts_with("depth")
|
.conflicts_with("depth")
|
||||||
.conflicts_with("only_dir")
|
.conflicts_with("only_dir")
|
||||||
.action(clap::ArgAction::SetTrue)
|
.action(clap::ArgAction::SetTrue)
|
||||||
@@ -189,7 +180,7 @@ pub fn build_cli() -> Command {
|
|||||||
.arg(
|
.arg(
|
||||||
Arg::new("width")
|
Arg::new("width")
|
||||||
.short('w')
|
.short('w')
|
||||||
.long("terminal-width")
|
.long("terminal_width")
|
||||||
.value_name("WIDTH")
|
.value_name("WIDTH")
|
||||||
.value_parser(value_parser!(usize))
|
.value_parser(value_parser!(usize))
|
||||||
.num_args(1)
|
.num_args(1)
|
||||||
@@ -303,24 +294,4 @@ pub fn build_cli() -> Command {
|
|||||||
.num_args(1)
|
.num_args(1)
|
||||||
.help("run dust on NUL-terminated file names specified in file; if argument is -, then read names from standard input"),
|
.help("run dust on NUL-terminated file names specified in file; if argument is -, then read names from standard input"),
|
||||||
)
|
)
|
||||||
.arg(
|
|
||||||
Arg::new("collapse")
|
|
||||||
.long("collapse")
|
|
||||||
.value_hint(clap::ValueHint::AnyPath)
|
|
||||||
.value_parser(value_parser!(String))
|
|
||||||
.action(clap::ArgAction::Append)
|
|
||||||
.help("Keep these directories collapsed"),
|
|
||||||
)
|
|
||||||
.arg(
|
|
||||||
Arg::new("filetime")
|
|
||||||
.short('m')
|
|
||||||
.long("filetime")
|
|
||||||
.num_args(1)
|
|
||||||
.value_parser([
|
|
||||||
PossibleValue::new("a").alias("accessed"),
|
|
||||||
PossibleValue::new("c").alias("changed"),
|
|
||||||
PossibleValue::new("m").alias("modified"),
|
|
||||||
])
|
|
||||||
.help("Directory 'size' is max filetime of child files instead of disk size. while a/c/m for last accessed/changed/modified time"),
|
|
||||||
)
|
|
||||||
}
|
}
|
||||||
|
|||||||
125
src/config.rs
125
src/config.rs
@@ -1,19 +1,20 @@
|
|||||||
use crate::node::FileTime;
|
|
||||||
use chrono::{Local, TimeZone};
|
use chrono::{Local, TimeZone};
|
||||||
use clap::ArgMatches;
|
use clap::ArgMatches;
|
||||||
use config_file::FromConfigFile;
|
use config_file::FromConfigFile;
|
||||||
use regex::Regex;
|
use regex::Regex;
|
||||||
use serde::Deserialize;
|
use serde::Deserialize;
|
||||||
|
use std::io::IsTerminal;
|
||||||
use std::path::Path;
|
use std::path::Path;
|
||||||
use std::path::PathBuf;
|
use std::path::PathBuf;
|
||||||
|
|
||||||
use crate::dir_walker::Operator;
|
use crate::dir_walker::Operater;
|
||||||
use crate::display::get_number_format;
|
use crate::display::get_number_format;
|
||||||
|
|
||||||
pub static DAY_SECONDS: i64 = 24 * 60 * 60;
|
pub static DAY_SECONDS: i64 = 24 * 60 * 60;
|
||||||
|
|
||||||
#[derive(Deserialize, Default)]
|
#[derive(Deserialize, Default)]
|
||||||
#[serde(rename_all = "kebab-case")]
|
#[serde(rename_all = "kebab-case")]
|
||||||
|
#[serde(deny_unknown_fields)]
|
||||||
pub struct Config {
|
pub struct Config {
|
||||||
pub display_full_paths: Option<bool>,
|
pub display_full_paths: Option<bool>,
|
||||||
pub display_apparent_size: Option<bool>,
|
pub display_apparent_size: Option<bool>,
|
||||||
@@ -53,7 +54,9 @@ impl Config {
|
|||||||
Some(true) == self.force_colors || options.get_flag("force_colors")
|
Some(true) == self.force_colors || options.get_flag("force_colors")
|
||||||
}
|
}
|
||||||
pub fn get_disable_progress(&self, options: &ArgMatches) -> bool {
|
pub fn get_disable_progress(&self, options: &ArgMatches) -> bool {
|
||||||
Some(true) == self.disable_progress || options.get_flag("disable_progress")
|
Some(true) == self.disable_progress
|
||||||
|
|| options.get_flag("disable_progress")
|
||||||
|
|| !std::io::stdout().is_terminal()
|
||||||
}
|
}
|
||||||
pub fn get_apparent_size(&self, options: &ArgMatches) -> bool {
|
pub fn get_apparent_size(&self, options: &ArgMatches) -> bool {
|
||||||
Some(true) == self.display_apparent_size || options.get_flag("display_apparent_size")
|
Some(true) == self.display_apparent_size || options.get_flag("display_apparent_size")
|
||||||
@@ -81,20 +84,6 @@ impl Config {
|
|||||||
})
|
})
|
||||||
.to_lowercase()
|
.to_lowercase()
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn get_filetime(&self, options: &ArgMatches) -> Option<FileTime> {
|
|
||||||
let out_fmt = options.get_one::<String>("filetime");
|
|
||||||
match out_fmt {
|
|
||||||
None => None,
|
|
||||||
Some(x) => match x.as_str() {
|
|
||||||
"m" | "modified" => Some(FileTime::Modified),
|
|
||||||
"a" | "accessed" => Some(FileTime::Accessed),
|
|
||||||
"c" | "changed" => Some(FileTime::Changed),
|
|
||||||
_ => unreachable!(),
|
|
||||||
},
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
pub fn get_skip_total(&self, options: &ArgMatches) -> bool {
|
pub fn get_skip_total(&self, options: &ArgMatches) -> bool {
|
||||||
Some(true) == self.skip_total || options.get_flag("skip_total")
|
Some(true) == self.skip_total || options.get_flag("skip_total")
|
||||||
}
|
}
|
||||||
@@ -156,21 +145,21 @@ impl Config {
|
|||||||
Some(true) == self.output_json || options.get_flag("output_json")
|
Some(true) == self.output_json || options.get_flag("output_json")
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn get_modified_time_operator(&self, options: &ArgMatches) -> Option<(Operator, i64)> {
|
pub fn get_modified_time_operator(&self, options: &ArgMatches) -> (Operater, i64) {
|
||||||
get_filter_time_operator(
|
get_filter_time_operator(
|
||||||
options.get_one::<String>("mtime"),
|
options.get_one::<String>("mtime"),
|
||||||
get_current_date_epoch_seconds(),
|
get_current_date_epoch_seconds(),
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn get_accessed_time_operator(&self, options: &ArgMatches) -> Option<(Operator, i64)> {
|
pub fn get_accessed_time_operator(&self, options: &ArgMatches) -> (Operater, i64) {
|
||||||
get_filter_time_operator(
|
get_filter_time_operator(
|
||||||
options.get_one::<String>("atime"),
|
options.get_one::<String>("atime"),
|
||||||
get_current_date_epoch_seconds(),
|
get_current_date_epoch_seconds(),
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn get_changed_time_operator(&self, options: &ArgMatches) -> Option<(Operator, i64)> {
|
pub fn get_created_time_operator(&self, options: &ArgMatches) -> (Operater, i64) {
|
||||||
get_filter_time_operator(
|
get_filter_time_operator(
|
||||||
options.get_one::<String>("ctime"),
|
options.get_one::<String>("ctime"),
|
||||||
get_current_date_epoch_seconds(),
|
get_current_date_epoch_seconds(),
|
||||||
@@ -179,7 +168,7 @@ impl Config {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn get_current_date_epoch_seconds() -> i64 {
|
fn get_current_date_epoch_seconds() -> i64 {
|
||||||
// calculate current date epoch seconds
|
// calcurate current date epoch seconds
|
||||||
let now = Local::now();
|
let now = Local::now();
|
||||||
let current_date = now.date_naive();
|
let current_date = now.date_naive();
|
||||||
|
|
||||||
@@ -193,7 +182,7 @@ fn get_current_date_epoch_seconds() -> i64 {
|
|||||||
fn get_filter_time_operator(
|
fn get_filter_time_operator(
|
||||||
option_value: Option<&String>,
|
option_value: Option<&String>,
|
||||||
current_date_epoch_seconds: i64,
|
current_date_epoch_seconds: i64,
|
||||||
) -> Option<(Operator, i64)> {
|
) -> (Operater, i64) {
|
||||||
match option_value {
|
match option_value {
|
||||||
Some(val) => {
|
Some(val) => {
|
||||||
let time = current_date_epoch_seconds
|
let time = current_date_epoch_seconds
|
||||||
@@ -203,12 +192,12 @@ fn get_filter_time_operator(
|
|||||||
.abs()
|
.abs()
|
||||||
* DAY_SECONDS;
|
* DAY_SECONDS;
|
||||||
match val.chars().next().expect("Value should not be empty") {
|
match val.chars().next().expect("Value should not be empty") {
|
||||||
'+' => Some((Operator::LessThan, time - DAY_SECONDS)),
|
'+' => (Operater::LessThan, time - DAY_SECONDS),
|
||||||
'-' => Some((Operator::GreaterThan, time)),
|
'-' => (Operater::GreaterThan, time),
|
||||||
_ => Some((Operator::Equal, time - DAY_SECONDS)),
|
_ => (Operater::Equal, time - DAY_SECONDS),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
None => None,
|
None => (Operater::GreaterThan, 0),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -227,7 +216,7 @@ fn convert_min_size(input: &str) -> Option<usize> {
|
|||||||
match number_format {
|
match number_format {
|
||||||
Some((multiple, _)) => Some(parsed_digits * (multiple as usize)),
|
Some((multiple, _)) => Some(parsed_digits * (multiple as usize)),
|
||||||
None => {
|
None => {
|
||||||
if letters.is_empty() {
|
if letters.eq("") {
|
||||||
Some(parsed_digits)
|
Some(parsed_digits)
|
||||||
} else {
|
} else {
|
||||||
eprintln!("Ignoring invalid min-size: {input}");
|
eprintln!("Ignoring invalid min-size: {input}");
|
||||||
@@ -250,29 +239,12 @@ fn get_config_locations(base: &Path) -> Vec<PathBuf> {
|
|||||||
]
|
]
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn get_config(conf_path: Option<String>) -> Config {
|
pub fn get_config() -> Config {
|
||||||
match conf_path {
|
if let Some(home) = directories::BaseDirs::new() {
|
||||||
Some(path_str) => {
|
for path in get_config_locations(home.home_dir()) {
|
||||||
let path = Path::new(&path_str);
|
|
||||||
if path.exists() {
|
if path.exists() {
|
||||||
match Config::from_config_file(path) {
|
if let Ok(config) = Config::from_config_file(path) {
|
||||||
Ok(config) => return config,
|
return config;
|
||||||
Err(e) => {
|
|
||||||
eprintln!("Ignoring invalid config file '{}': {}", &path.display(), e)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
eprintln!("Config file {:?} doesn't exists", &path.display());
|
|
||||||
}
|
|
||||||
}
|
|
||||||
None => {
|
|
||||||
if let Some(home) = directories::BaseDirs::new() {
|
|
||||||
for path in get_config_locations(home.home_dir()) {
|
|
||||||
if path.exists() {
|
|
||||||
if let Ok(config) = Config::from_config_file(&path) {
|
|
||||||
return config;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -287,8 +259,7 @@ mod tests {
|
|||||||
#[allow(unused_imports)]
|
#[allow(unused_imports)]
|
||||||
use super::*;
|
use super::*;
|
||||||
use chrono::{Datelike, Timelike};
|
use chrono::{Datelike, Timelike};
|
||||||
use clap::builder::PossibleValue;
|
use clap::{value_parser, Arg, ArgMatches, Command};
|
||||||
use clap::{Arg, ArgMatches, Command, value_parser};
|
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn test_get_current_date_epoch_seconds() {
|
fn test_get_current_date_epoch_seconds() {
|
||||||
@@ -367,56 +338,4 @@ mod tests {
|
|||||||
)
|
)
|
||||||
.get_matches_from(args)
|
.get_matches_from(args)
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
|
||||||
fn test_get_filetime() {
|
|
||||||
// No config and no flag.
|
|
||||||
let c = Config::default();
|
|
||||||
let args = get_filetime_args(vec!["dust"]);
|
|
||||||
assert_eq!(c.get_filetime(&args), None);
|
|
||||||
|
|
||||||
// Config is not defined and flag is defined as access time
|
|
||||||
let c = Config::default();
|
|
||||||
let args = get_filetime_args(vec!["dust", "--filetime", "a"]);
|
|
||||||
assert_eq!(c.get_filetime(&args), Some(FileTime::Accessed));
|
|
||||||
|
|
||||||
let c = Config::default();
|
|
||||||
let args = get_filetime_args(vec!["dust", "--filetime", "accessed"]);
|
|
||||||
assert_eq!(c.get_filetime(&args), Some(FileTime::Accessed));
|
|
||||||
|
|
||||||
// Config is not defined and flag is defined as modified time
|
|
||||||
let c = Config::default();
|
|
||||||
let args = get_filetime_args(vec!["dust", "--filetime", "m"]);
|
|
||||||
assert_eq!(c.get_filetime(&args), Some(FileTime::Modified));
|
|
||||||
|
|
||||||
let c = Config::default();
|
|
||||||
let args = get_filetime_args(vec!["dust", "--filetime", "modified"]);
|
|
||||||
assert_eq!(c.get_filetime(&args), Some(FileTime::Modified));
|
|
||||||
|
|
||||||
// Config is not defined and flag is defined as changed time
|
|
||||||
let c = Config::default();
|
|
||||||
let args = get_filetime_args(vec!["dust", "--filetime", "c"]);
|
|
||||||
assert_eq!(c.get_filetime(&args), Some(FileTime::Changed));
|
|
||||||
|
|
||||||
let c = Config::default();
|
|
||||||
let args = get_filetime_args(vec!["dust", "--filetime", "changed"]);
|
|
||||||
assert_eq!(c.get_filetime(&args), Some(FileTime::Changed));
|
|
||||||
}
|
|
||||||
|
|
||||||
fn get_filetime_args(args: Vec<&str>) -> ArgMatches {
|
|
||||||
Command::new("Dust")
|
|
||||||
.arg(
|
|
||||||
Arg::new("filetime")
|
|
||||||
.short('m')
|
|
||||||
.long("filetime")
|
|
||||||
.num_args(1)
|
|
||||||
.value_parser([
|
|
||||||
PossibleValue::new("a").alias("accessed"),
|
|
||||||
PossibleValue::new("c").alias("changed"),
|
|
||||||
PossibleValue::new("m").alias("modified"),
|
|
||||||
])
|
|
||||||
.help("Directory 'size' is max filetime of child files instead of disk size. while a/c/m for accessed/changed/modified time"),
|
|
||||||
)
|
|
||||||
.get_matches_from(args)
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,21 +1,19 @@
|
|||||||
use std::cmp::Ordering;
|
use std::cmp::Ordering;
|
||||||
use std::fs;
|
use std::fs;
|
||||||
use std::io::Error;
|
|
||||||
use std::sync::Arc;
|
use std::sync::Arc;
|
||||||
use std::sync::Mutex;
|
use std::sync::Mutex;
|
||||||
|
|
||||||
use crate::node::Node;
|
use crate::node::Node;
|
||||||
use crate::progress::ORDERING;
|
|
||||||
use crate::progress::Operation;
|
use crate::progress::Operation;
|
||||||
use crate::progress::PAtomicInfo;
|
use crate::progress::PAtomicInfo;
|
||||||
use crate::progress::RuntimeErrors;
|
use crate::progress::RuntimeErrors;
|
||||||
|
use crate::progress::ORDERING;
|
||||||
use crate::utils::is_filtered_out_due_to_file_time;
|
use crate::utils::is_filtered_out_due_to_file_time;
|
||||||
use crate::utils::is_filtered_out_due_to_invert_regex;
|
use crate::utils::is_filtered_out_due_to_invert_regex;
|
||||||
use crate::utils::is_filtered_out_due_to_regex;
|
use crate::utils::is_filtered_out_due_to_regex;
|
||||||
use rayon::iter::ParallelBridge;
|
use rayon::iter::ParallelBridge;
|
||||||
use rayon::prelude::ParallelIterator;
|
use rayon::prelude::ParallelIterator;
|
||||||
use regex::Regex;
|
use regex::Regex;
|
||||||
use std::path::Path;
|
|
||||||
use std::path::PathBuf;
|
use std::path::PathBuf;
|
||||||
|
|
||||||
use std::collections::HashSet;
|
use std::collections::HashSet;
|
||||||
@@ -23,11 +21,10 @@ use std::collections::HashSet;
|
|||||||
use crate::node::build_node;
|
use crate::node::build_node;
|
||||||
use std::fs::DirEntry;
|
use std::fs::DirEntry;
|
||||||
|
|
||||||
use crate::node::FileTime;
|
|
||||||
use crate::platform::get_metadata;
|
use crate::platform::get_metadata;
|
||||||
|
|
||||||
#[derive(Debug)]
|
#[derive(Debug)]
|
||||||
pub enum Operator {
|
pub enum Operater {
|
||||||
Equal = 0,
|
Equal = 0,
|
||||||
LessThan = 1,
|
LessThan = 1,
|
||||||
GreaterThan = 2,
|
GreaterThan = 2,
|
||||||
@@ -38,12 +35,11 @@ pub struct WalkData<'a> {
|
|||||||
pub filter_regex: &'a [Regex],
|
pub filter_regex: &'a [Regex],
|
||||||
pub invert_filter_regex: &'a [Regex],
|
pub invert_filter_regex: &'a [Regex],
|
||||||
pub allowed_filesystems: HashSet<u64>,
|
pub allowed_filesystems: HashSet<u64>,
|
||||||
pub filter_modified_time: Option<(Operator, i64)>,
|
pub filter_modified_time: (Operater, i64),
|
||||||
pub filter_accessed_time: Option<(Operator, i64)>,
|
pub filter_accessed_time: (Operater, i64),
|
||||||
pub filter_changed_time: Option<(Operator, i64)>,
|
pub filter_changed_time: (Operater, i64),
|
||||||
pub use_apparent_size: bool,
|
pub use_apparent_size: bool,
|
||||||
pub by_filecount: bool,
|
pub by_filecount: bool,
|
||||||
pub by_filetime: &'a Option<FileTime>,
|
|
||||||
pub ignore_hidden: bool,
|
pub ignore_hidden: bool,
|
||||||
pub follow_links: bool,
|
pub follow_links: bool,
|
||||||
pub progress_data: Arc<PAtomicInfo>,
|
pub progress_data: Arc<PAtomicInfo>,
|
||||||
@@ -61,15 +57,19 @@ pub fn walk_it(dirs: HashSet<PathBuf>, walk_data: &WalkData) -> Vec<Node> {
|
|||||||
|
|
||||||
prog_data.state.store(Operation::PREPARING, ORDERING);
|
prog_data.state.store(Operation::PREPARING, ORDERING);
|
||||||
|
|
||||||
clean_inodes(node, &mut inodes, walk_data)
|
clean_inodes(node, &mut inodes, walk_data.use_apparent_size)
|
||||||
})
|
})
|
||||||
.collect();
|
.collect();
|
||||||
top_level_nodes
|
top_level_nodes
|
||||||
}
|
}
|
||||||
|
|
||||||
// Remove files which have the same inode, we don't want to double count them.
|
// Remove files which have the same inode, we don't want to double count them.
|
||||||
fn clean_inodes(x: Node, inodes: &mut HashSet<(u64, u64)>, walk_data: &WalkData) -> Option<Node> {
|
fn clean_inodes(
|
||||||
if !walk_data.use_apparent_size {
|
x: Node,
|
||||||
|
inodes: &mut HashSet<(u64, u64)>,
|
||||||
|
use_apparent_size: bool,
|
||||||
|
) -> Option<Node> {
|
||||||
|
if !use_apparent_size {
|
||||||
if let Some(id) = x.inode_device {
|
if let Some(id) = x.inode_device {
|
||||||
if !inodes.insert(id) {
|
if !inodes.insert(id) {
|
||||||
return None;
|
return None;
|
||||||
@@ -82,25 +82,12 @@ fn clean_inodes(x: Node, inodes: &mut HashSet<(u64, u64)>, walk_data: &WalkData)
|
|||||||
tmp.sort_by(sort_by_inode);
|
tmp.sort_by(sort_by_inode);
|
||||||
let new_children: Vec<_> = tmp
|
let new_children: Vec<_> = tmp
|
||||||
.into_iter()
|
.into_iter()
|
||||||
.filter_map(|c| clean_inodes(c, inodes, walk_data))
|
.filter_map(|c| clean_inodes(c, inodes, use_apparent_size))
|
||||||
.collect();
|
.collect();
|
||||||
|
|
||||||
let actual_size = if walk_data.by_filetime.is_some() {
|
|
||||||
// If by_filetime is Some, directory 'size' is the maximum filetime among child files instead of disk size
|
|
||||||
new_children
|
|
||||||
.iter()
|
|
||||||
.map(|c| c.size)
|
|
||||||
.chain(std::iter::once(x.size))
|
|
||||||
.max()
|
|
||||||
.unwrap_or(0)
|
|
||||||
} else {
|
|
||||||
// If by_filetime is None, directory 'size' is the sum of disk sizes or file counts of child files
|
|
||||||
x.size + new_children.iter().map(|c| c.size).sum::<u64>()
|
|
||||||
};
|
|
||||||
|
|
||||||
Some(Node {
|
Some(Node {
|
||||||
name: x.name,
|
name: x.name,
|
||||||
size: actual_size,
|
size: x.size + new_children.iter().map(|c| c.size).sum::<u64>(),
|
||||||
children: new_children,
|
children: new_children,
|
||||||
inode_device: x.inode_device,
|
inode_device: x.inode_device,
|
||||||
depth: x.depth,
|
depth: x.depth,
|
||||||
@@ -125,62 +112,31 @@ fn sort_by_inode(a: &Node, b: &Node) -> std::cmp::Ordering {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Check if `path` is inside ignored directory
|
fn ignore_file(entry: &DirEntry, walk_data: &WalkData) -> bool {
|
||||||
fn is_ignored_path(path: &Path, walk_data: &WalkData) -> bool {
|
let is_dot_file = entry.file_name().to_str().unwrap_or("").starts_with('.');
|
||||||
if walk_data.ignore_directories.contains(path) {
|
let is_ignored_path = walk_data.ignore_directories.contains(&entry.path());
|
||||||
return true;
|
|
||||||
}
|
|
||||||
|
|
||||||
// Entry is inside an ignored absolute path
|
let size_inode_device = get_metadata(entry.path(), false);
|
||||||
// Absolute paths should be canonicalized before being added to `WalkData.ignore_directories`
|
if let Some((_size, Some((_id, dev)), (modified_time, accessed_time, changed_time))) =
|
||||||
for ignored_path in walk_data.ignore_directories.iter() {
|
size_inode_device
|
||||||
if !ignored_path.is_absolute() {
|
{
|
||||||
continue;
|
if !walk_data.allowed_filesystems.is_empty()
|
||||||
}
|
&& !walk_data.allowed_filesystems.contains(&dev)
|
||||||
let absolute_entry_path = std::fs::canonicalize(path).unwrap_or_default();
|
{
|
||||||
if absolute_entry_path.starts_with(ignored_path) {
|
|
||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
}
|
if entry.path().is_file()
|
||||||
|
&& [
|
||||||
false
|
(&walk_data.filter_modified_time, modified_time),
|
||||||
}
|
(&walk_data.filter_accessed_time, accessed_time),
|
||||||
|
(&walk_data.filter_changed_time, changed_time),
|
||||||
fn ignore_file(entry: &DirEntry, walk_data: &WalkData) -> bool {
|
]
|
||||||
if is_ignored_path(&entry.path(), walk_data) {
|
.iter()
|
||||||
return true;
|
.any(|(filter_time, actual_time)| {
|
||||||
}
|
is_filtered_out_due_to_file_time(filter_time, *actual_time)
|
||||||
|
})
|
||||||
let is_dot_file = entry.file_name().to_str().unwrap_or("").starts_with('.');
|
{
|
||||||
let follow_links = walk_data.follow_links && entry.file_type().is_ok_and(|ft| ft.is_symlink());
|
return true;
|
||||||
|
|
||||||
if !walk_data.allowed_filesystems.is_empty() {
|
|
||||||
let size_inode_device = get_metadata(entry.path(), false, follow_links);
|
|
||||||
if let Some((_size, Some((_id, dev)), _gunk)) = size_inode_device {
|
|
||||||
if !walk_data.allowed_filesystems.contains(&dev) {
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
if walk_data.filter_accessed_time.is_some()
|
|
||||||
|| walk_data.filter_modified_time.is_some()
|
|
||||||
|| walk_data.filter_changed_time.is_some()
|
|
||||||
{
|
|
||||||
let size_inode_device = get_metadata(entry.path(), false, follow_links);
|
|
||||||
if let Some((_, _, (modified_time, accessed_time, changed_time))) = size_inode_device {
|
|
||||||
if entry.path().is_file()
|
|
||||||
&& [
|
|
||||||
(&walk_data.filter_modified_time, modified_time),
|
|
||||||
(&walk_data.filter_accessed_time, accessed_time),
|
|
||||||
(&walk_data.filter_changed_time, changed_time),
|
|
||||||
]
|
|
||||||
.iter()
|
|
||||||
.any(|(filter_time, actual_time)| {
|
|
||||||
is_filtered_out_due_to_file_time(filter_time, *actual_time)
|
|
||||||
})
|
|
||||||
{
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -199,13 +155,17 @@ fn ignore_file(entry: &DirEntry, walk_data: &WalkData) -> bool {
|
|||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
|
|
||||||
is_dot_file && walk_data.ignore_hidden
|
(is_dot_file && walk_data.ignore_hidden) || is_ignored_path
|
||||||
}
|
}
|
||||||
|
|
||||||
fn walk(dir: PathBuf, walk_data: &WalkData, depth: usize) -> Option<Node> {
|
fn walk(dir: PathBuf, walk_data: &WalkData, depth: usize) -> Option<Node> {
|
||||||
let prog_data = &walk_data.progress_data;
|
let prog_data = &walk_data.progress_data;
|
||||||
let errors = &walk_data.errors;
|
let errors = &walk_data.errors;
|
||||||
|
|
||||||
|
if errors.lock().unwrap().abort {
|
||||||
|
return None;
|
||||||
|
}
|
||||||
|
|
||||||
let children = if dir.is_dir() {
|
let children = if dir.is_dir() {
|
||||||
let read_dir = fs::read_dir(&dir);
|
let read_dir = fs::read_dir(&dir);
|
||||||
match read_dir {
|
match read_dir {
|
||||||
@@ -251,9 +211,8 @@ fn walk(dir: PathBuf, walk_data: &WalkData, depth: usize) -> Option<Node> {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
Err(ref failed) => {
|
Err(ref failed) => {
|
||||||
if handle_error_and_retry(failed, &dir, walk_data) {
|
let mut editable_error = errors.lock().unwrap();
|
||||||
return walk(dir.clone(), walk_data, depth);
|
editable_error.no_permissions.insert(failed.to_string());
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
None
|
None
|
||||||
@@ -261,11 +220,21 @@ fn walk(dir: PathBuf, walk_data: &WalkData, depth: usize) -> Option<Node> {
|
|||||||
.collect()
|
.collect()
|
||||||
}
|
}
|
||||||
Err(failed) => {
|
Err(failed) => {
|
||||||
if handle_error_and_retry(&failed, &dir, walk_data) {
|
let mut editable_error = errors.lock().unwrap();
|
||||||
return walk(dir, walk_data, depth);
|
match failed.kind() {
|
||||||
} else {
|
std::io::ErrorKind::PermissionDenied => {
|
||||||
vec![]
|
editable_error
|
||||||
|
.no_permissions
|
||||||
|
.insert(dir.to_string_lossy().into());
|
||||||
|
}
|
||||||
|
std::io::ErrorKind::NotFound => {
|
||||||
|
editable_error.file_not_found.insert(failed.to_string());
|
||||||
|
}
|
||||||
|
_ => {
|
||||||
|
editable_error.unknown_error.insert(failed.to_string());
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
vec![]
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
@@ -276,47 +245,7 @@ fn walk(dir: PathBuf, walk_data: &WalkData, depth: usize) -> Option<Node> {
|
|||||||
}
|
}
|
||||||
vec![]
|
vec![]
|
||||||
};
|
};
|
||||||
let is_symlink = if walk_data.follow_links {
|
build_node(dir, children, false, false, depth, walk_data)
|
||||||
match fs::symlink_metadata(&dir) {
|
|
||||||
Ok(metadata) => metadata.file_type().is_symlink(),
|
|
||||||
Err(_) => false,
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
false
|
|
||||||
};
|
|
||||||
build_node(dir, children, is_symlink, false, depth, walk_data)
|
|
||||||
}
|
|
||||||
|
|
||||||
fn handle_error_and_retry(failed: &Error, dir: &Path, walk_data: &WalkData) -> bool {
|
|
||||||
let mut editable_error = walk_data.errors.lock().unwrap();
|
|
||||||
match failed.kind() {
|
|
||||||
std::io::ErrorKind::PermissionDenied => {
|
|
||||||
editable_error
|
|
||||||
.no_permissions
|
|
||||||
.insert(dir.to_string_lossy().into());
|
|
||||||
}
|
|
||||||
std::io::ErrorKind::InvalidInput => {
|
|
||||||
editable_error
|
|
||||||
.no_permissions
|
|
||||||
.insert(dir.to_string_lossy().into());
|
|
||||||
}
|
|
||||||
std::io::ErrorKind::NotFound => {
|
|
||||||
editable_error.file_not_found.insert(failed.to_string());
|
|
||||||
}
|
|
||||||
std::io::ErrorKind::Interrupted => {
|
|
||||||
let mut editable_error = walk_data.errors.lock().unwrap();
|
|
||||||
editable_error.interrupted_error += 1;
|
|
||||||
if editable_error.interrupted_error > 3 {
|
|
||||||
panic!("Multiple Interrupted Errors occurred while scanning filesystem. Aborting");
|
|
||||||
} else {
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
_ => {
|
|
||||||
editable_error.unknown_error.insert(failed.to_string());
|
|
||||||
}
|
|
||||||
}
|
|
||||||
false
|
|
||||||
}
|
}
|
||||||
|
|
||||||
mod tests {
|
mod tests {
|
||||||
@@ -335,43 +264,17 @@ mod tests {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
#[cfg(test)]
|
|
||||||
fn create_walker<'a>(use_apparent_size: bool) -> WalkData<'a> {
|
|
||||||
use crate::PIndicator;
|
|
||||||
let indicator = PIndicator::build_me();
|
|
||||||
WalkData {
|
|
||||||
ignore_directories: HashSet::new(),
|
|
||||||
filter_regex: &[],
|
|
||||||
invert_filter_regex: &[],
|
|
||||||
allowed_filesystems: HashSet::new(),
|
|
||||||
filter_modified_time: Some((Operator::GreaterThan, 0)),
|
|
||||||
filter_accessed_time: Some((Operator::GreaterThan, 0)),
|
|
||||||
filter_changed_time: Some((Operator::GreaterThan, 0)),
|
|
||||||
use_apparent_size,
|
|
||||||
by_filecount: false,
|
|
||||||
by_filetime: &None,
|
|
||||||
ignore_hidden: false,
|
|
||||||
follow_links: false,
|
|
||||||
progress_data: indicator.data.clone(),
|
|
||||||
errors: Arc::new(Mutex::new(RuntimeErrors::default())),
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
#[allow(clippy::redundant_clone)]
|
#[allow(clippy::redundant_clone)]
|
||||||
fn test_should_ignore_file() {
|
fn test_should_ignore_file() {
|
||||||
let mut inodes = HashSet::new();
|
let mut inodes = HashSet::new();
|
||||||
let n = create_node();
|
let n = create_node();
|
||||||
let walkdata = create_walker(false);
|
|
||||||
|
|
||||||
// First time we insert the node
|
// First time we insert the node
|
||||||
assert_eq!(
|
assert_eq!(clean_inodes(n.clone(), &mut inodes, false), Some(n.clone()));
|
||||||
clean_inodes(n.clone(), &mut inodes, &walkdata),
|
|
||||||
Some(n.clone())
|
|
||||||
);
|
|
||||||
|
|
||||||
// Second time is a duplicate - we ignore it
|
// Second time is a duplicate - we ignore it
|
||||||
assert_eq!(clean_inodes(n.clone(), &mut inodes, &walkdata), None);
|
assert_eq!(clean_inodes(n.clone(), &mut inodes, false), None);
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
@@ -379,17 +282,10 @@ mod tests {
|
|||||||
fn test_should_not_ignore_files_if_using_apparent_size() {
|
fn test_should_not_ignore_files_if_using_apparent_size() {
|
||||||
let mut inodes = HashSet::new();
|
let mut inodes = HashSet::new();
|
||||||
let n = create_node();
|
let n = create_node();
|
||||||
let walkdata = create_walker(true);
|
|
||||||
|
|
||||||
// If using apparent size we include Nodes, even if duplicate inodes
|
// If using apparent size we include Nodes, even if duplicate inodes
|
||||||
assert_eq!(
|
assert_eq!(clean_inodes(n.clone(), &mut inodes, true), Some(n.clone()));
|
||||||
clean_inodes(n.clone(), &mut inodes, &walkdata),
|
assert_eq!(clean_inodes(n.clone(), &mut inodes, true), Some(n.clone()));
|
||||||
Some(n.clone())
|
|
||||||
);
|
|
||||||
assert_eq!(
|
|
||||||
clean_inodes(n.clone(), &mut inodes, &walkdata),
|
|
||||||
Some(n.clone())
|
|
||||||
);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
|
|||||||
@@ -1,5 +1,4 @@
|
|||||||
use crate::display_node::DisplayNode;
|
use crate::display_node::DisplayNode;
|
||||||
use crate::node::FileTime;
|
|
||||||
|
|
||||||
use ansi_term::Colour::Red;
|
use ansi_term::Colour::Red;
|
||||||
use lscolors::{LsColors, Style};
|
use lscolors::{LsColors, Style};
|
||||||
@@ -8,7 +7,6 @@ use unicode_width::UnicodeWidthStr;
|
|||||||
|
|
||||||
use stfu8::encode_u8;
|
use stfu8::encode_u8;
|
||||||
|
|
||||||
use chrono::{DateTime, Local, TimeZone, Utc};
|
|
||||||
use std::cmp::max;
|
use std::cmp::max;
|
||||||
use std::cmp::min;
|
use std::cmp::min;
|
||||||
use std::fs;
|
use std::fs;
|
||||||
@@ -16,16 +14,14 @@ use std::iter::repeat;
|
|||||||
use std::path::Path;
|
use std::path::Path;
|
||||||
use thousands::Separable;
|
use thousands::Separable;
|
||||||
|
|
||||||
pub static UNITS: [char; 5] = ['P', 'T', 'G', 'M', 'K'];
|
pub static UNITS: [char; 4] = ['T', 'G', 'M', 'K'];
|
||||||
static BLOCKS: [char; 5] = ['█', '▓', '▒', '░', ' '];
|
static BLOCKS: [char; 5] = ['█', '▓', '▒', '░', ' '];
|
||||||
const FILETIME_SHOW_LENGTH: usize = 19;
|
|
||||||
|
|
||||||
pub struct InitialDisplayData {
|
pub struct InitialDisplayData {
|
||||||
pub short_paths: bool,
|
pub short_paths: bool,
|
||||||
pub is_reversed: bool,
|
pub is_reversed: bool,
|
||||||
pub colors_on: bool,
|
pub colors_on: bool,
|
||||||
pub by_filecount: bool,
|
pub by_filecount: bool,
|
||||||
pub by_filetime: Option<FileTime>,
|
|
||||||
pub is_screen_reader: bool,
|
pub is_screen_reader: bool,
|
||||||
pub output_format: String,
|
pub output_format: String,
|
||||||
pub bars_on_right: bool,
|
pub bars_on_right: bool,
|
||||||
@@ -71,7 +67,11 @@ impl DisplayData {
|
|||||||
|
|
||||||
fn percent_size(&self, node: &DisplayNode) -> f32 {
|
fn percent_size(&self, node: &DisplayNode) -> f32 {
|
||||||
let result = node.size as f32 / self.base_size as f32;
|
let result = node.size as f32 / self.base_size as f32;
|
||||||
if result.is_normal() { result } else { 0.0 }
|
if result.is_normal() {
|
||||||
|
result
|
||||||
|
} else {
|
||||||
|
0.0
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -130,11 +130,17 @@ pub fn draw_it(
|
|||||||
root_node: &DisplayNode,
|
root_node: &DisplayNode,
|
||||||
skip_total: bool,
|
skip_total: bool,
|
||||||
) {
|
) {
|
||||||
|
let biggest = match skip_total {
|
||||||
|
false => root_node,
|
||||||
|
true => root_node
|
||||||
|
.get_children_from_node(false)
|
||||||
|
.next()
|
||||||
|
.unwrap_or(root_node),
|
||||||
|
};
|
||||||
|
|
||||||
let num_chars_needed_on_left_most = if idd.by_filecount {
|
let num_chars_needed_on_left_most = if idd.by_filecount {
|
||||||
let max_size = root_node.size;
|
let max_size = biggest.size;
|
||||||
max_size.separate_with_commas().chars().count()
|
max_size.separate_with_commas().chars().count()
|
||||||
} else if idd.by_filetime.is_some() {
|
|
||||||
FILETIME_SHOW_LENGTH
|
|
||||||
} else {
|
} else {
|
||||||
find_biggest_size_str(root_node, &idd.output_format)
|
find_biggest_size_str(root_node, &idd.output_format)
|
||||||
};
|
};
|
||||||
@@ -160,7 +166,7 @@ pub fn draw_it(
|
|||||||
let display_data = DisplayData {
|
let display_data = DisplayData {
|
||||||
initial: idd,
|
initial: idd,
|
||||||
num_chars_needed_on_left_most,
|
num_chars_needed_on_left_most,
|
||||||
base_size: root_node.size,
|
base_size: biggest.size,
|
||||||
longest_string_length,
|
longest_string_length,
|
||||||
ls_colors: LsColors::from_env().unwrap_or_default(),
|
ls_colors: LsColors::from_env().unwrap_or_default(),
|
||||||
};
|
};
|
||||||
@@ -269,7 +275,7 @@ fn clean_indentation_string(s: &str) -> String {
|
|||||||
is
|
is
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn get_printable_name<P: AsRef<Path>>(dir_name: &P, short_paths: bool) -> String {
|
fn get_printable_name<P: AsRef<Path>>(dir_name: &P, short_paths: bool) -> String {
|
||||||
let dir_name = dir_name.as_ref();
|
let dir_name = dir_name.as_ref();
|
||||||
let printable_name = {
|
let printable_name = {
|
||||||
if short_paths {
|
if short_paths {
|
||||||
@@ -336,8 +342,6 @@ pub fn format_string(
|
|||||||
if display_data.initial.is_screen_reader {
|
if display_data.initial.is_screen_reader {
|
||||||
// if screen_reader then bars is 'depth'
|
// if screen_reader then bars is 'depth'
|
||||||
format!("{pretty_name} {bars} {pretty_size}{percent}")
|
format!("{pretty_name} {bars} {pretty_size}{percent}")
|
||||||
} else if display_data.initial.by_filetime.is_some() {
|
|
||||||
format!("{pretty_size} {indent}{pretty_name}")
|
|
||||||
} else {
|
} else {
|
||||||
format!("{pretty_size} {indent} {pretty_name}{percent}")
|
format!("{pretty_size} {indent} {pretty_name}{percent}")
|
||||||
}
|
}
|
||||||
@@ -372,8 +376,6 @@ fn get_name_percent(
|
|||||||
fn get_pretty_size(node: &DisplayNode, is_biggest: bool, display_data: &DisplayData) -> String {
|
fn get_pretty_size(node: &DisplayNode, is_biggest: bool, display_data: &DisplayData) -> String {
|
||||||
let output = if display_data.initial.by_filecount {
|
let output = if display_data.initial.by_filecount {
|
||||||
node.size.separate_with_commas()
|
node.size.separate_with_commas()
|
||||||
} else if display_data.initial.by_filetime.is_some() {
|
|
||||||
get_pretty_file_modified_time(node.size as i64)
|
|
||||||
} else {
|
} else {
|
||||||
human_readable_number(node.size, &display_data.initial.output_format)
|
human_readable_number(node.size, &display_data.initial.output_format)
|
||||||
};
|
};
|
||||||
@@ -387,14 +389,6 @@ fn get_pretty_size(node: &DisplayNode, is_biggest: bool, display_data: &DisplayD
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
fn get_pretty_file_modified_time(timestamp: i64) -> String {
|
|
||||||
let datetime: DateTime<Utc> = Utc.timestamp_opt(timestamp, 0).unwrap();
|
|
||||||
|
|
||||||
let local_datetime = datetime.with_timezone(&Local);
|
|
||||||
|
|
||||||
local_datetime.format("%Y-%m-%dT%H:%M:%S").to_string()
|
|
||||||
}
|
|
||||||
|
|
||||||
fn get_pretty_name(
|
fn get_pretty_name(
|
||||||
node: &DisplayNode,
|
node: &DisplayNode,
|
||||||
name_and_padding: String,
|
name_and_padding: String,
|
||||||
@@ -475,7 +469,6 @@ mod tests {
|
|||||||
is_reversed: false,
|
is_reversed: false,
|
||||||
colors_on: false,
|
colors_on: false,
|
||||||
by_filecount: false,
|
by_filecount: false,
|
||||||
by_filetime: None,
|
|
||||||
is_screen_reader: false,
|
is_screen_reader: false,
|
||||||
output_format: "".into(),
|
output_format: "".into(),
|
||||||
bars_on_right: false,
|
bars_on_right: false,
|
||||||
@@ -554,14 +547,6 @@ mod tests {
|
|||||||
assert_eq!(human_readable_number(1024 * 1024 * 1024 - 1, ""), "1023M");
|
assert_eq!(human_readable_number(1024 * 1024 * 1024 - 1, ""), "1023M");
|
||||||
assert_eq!(human_readable_number(1024 * 1024 * 1024 * 20, ""), "20G");
|
assert_eq!(human_readable_number(1024 * 1024 * 1024 * 20, ""), "20G");
|
||||||
assert_eq!(human_readable_number(1024 * 1024 * 1024 * 1024, ""), "1.0T");
|
assert_eq!(human_readable_number(1024 * 1024 * 1024 * 1024, ""), "1.0T");
|
||||||
assert_eq!(
|
|
||||||
human_readable_number(1024 * 1024 * 1024 * 1024 * 234, ""),
|
|
||||||
"234T"
|
|
||||||
);
|
|
||||||
assert_eq!(
|
|
||||||
human_readable_number(1024 * 1024 * 1024 * 1024 * 1024, ""),
|
|
||||||
"1.0P"
|
|
||||||
);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
@@ -640,37 +625,4 @@ mod tests {
|
|||||||
let bar = dd.generate_bar(&n, 5);
|
let bar = dd.generate_bar(&n, 5);
|
||||||
assert_eq!(bar, "████▓▓▓▓▓▓▓▓▓");
|
assert_eq!(bar, "████▓▓▓▓▓▓▓▓▓");
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
|
||||||
fn test_get_pretty_file_modified_time() {
|
|
||||||
// Create a timestamp for 2023-07-12 00:00:00 in local time
|
|
||||||
let local_dt = Local.with_ymd_and_hms(2023, 7, 12, 0, 0, 0).unwrap();
|
|
||||||
let timestamp = local_dt.timestamp();
|
|
||||||
|
|
||||||
// Format expected output
|
|
||||||
let expected_output = local_dt.format("%Y-%m-%dT%H:%M:%S").to_string();
|
|
||||||
|
|
||||||
assert_eq!(get_pretty_file_modified_time(timestamp), expected_output);
|
|
||||||
|
|
||||||
// Test another timestamp
|
|
||||||
let local_dt = Local.with_ymd_and_hms(2020, 1, 1, 12, 0, 0).unwrap();
|
|
||||||
let timestamp = local_dt.timestamp();
|
|
||||||
let expected_output = local_dt.format("%Y-%m-%dT%H:%M:%S").to_string();
|
|
||||||
|
|
||||||
assert_eq!(get_pretty_file_modified_time(timestamp), expected_output);
|
|
||||||
|
|
||||||
// Test timestamp for epoch start (1970-01-01T00:00:00)
|
|
||||||
let local_dt = Local.with_ymd_and_hms(1970, 1, 1, 0, 0, 0).unwrap();
|
|
||||||
let timestamp = local_dt.timestamp();
|
|
||||||
let expected_output = local_dt.format("%Y-%m-%dT%H:%M:%S").to_string();
|
|
||||||
|
|
||||||
assert_eq!(get_pretty_file_modified_time(timestamp), expected_output);
|
|
||||||
|
|
||||||
// Test a future timestamp
|
|
||||||
let local_dt = Local.with_ymd_and_hms(2030, 12, 25, 6, 30, 0).unwrap();
|
|
||||||
let timestamp = local_dt.timestamp();
|
|
||||||
let expected_output = local_dt.format("%Y-%m-%dT%H:%M:%S").to_string();
|
|
||||||
|
|
||||||
assert_eq!(get_pretty_file_modified_time(timestamp), expected_output);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,12 +1,8 @@
|
|||||||
use std::cell::RefCell;
|
|
||||||
use std::path::PathBuf;
|
use std::path::PathBuf;
|
||||||
|
|
||||||
use serde::ser::SerializeStruct;
|
use serde::Serialize;
|
||||||
use serde::{Serialize, Serializer};
|
|
||||||
|
|
||||||
use crate::display::human_readable_number;
|
#[derive(Debug, PartialEq, Eq, PartialOrd, Ord, Clone, Serialize)]
|
||||||
|
|
||||||
#[derive(Debug, PartialEq, Eq, PartialOrd, Ord, Clone)]
|
|
||||||
pub struct DisplayNode {
|
pub struct DisplayNode {
|
||||||
// Note: the order of fields in important here, for PartialEq and PartialOrd
|
// Note: the order of fields in important here, for PartialEq and PartialOrd
|
||||||
pub size: u64,
|
pub size: u64,
|
||||||
@@ -29,30 +25,3 @@ impl DisplayNode {
|
|||||||
out
|
out
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Only used for -j 'json' flag combined with -o 'output_type' flag
|
|
||||||
// Used to pass the output_type into the custom Serde serializer
|
|
||||||
thread_local! {
|
|
||||||
pub static OUTPUT_TYPE: RefCell<String> = const { RefCell::new(String::new()) };
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
|
||||||
We need the custom Serialize incase someone uses the -o flag to pass a custom output type in
|
|
||||||
(show size in Mb / Gb etc).
|
|
||||||
Sadly this also necessitates a global variable OUTPUT_TYPE as we can not pass the output_type flag
|
|
||||||
into the serialize method
|
|
||||||
*/
|
|
||||||
impl Serialize for DisplayNode {
|
|
||||||
fn serialize<S>(&self, serializer: S) -> Result<S::Ok, S::Error>
|
|
||||||
where
|
|
||||||
S: Serializer,
|
|
||||||
{
|
|
||||||
let readable_size = OUTPUT_TYPE
|
|
||||||
.with(|output_type| human_readable_number(self.size, output_type.borrow().as_str()));
|
|
||||||
let mut state = serializer.serialize_struct("DisplayNode", 2)?;
|
|
||||||
state.serialize_field("size", &(readable_size))?;
|
|
||||||
state.serialize_field("name", &self.name)?;
|
|
||||||
state.serialize_field("children", &self.children)?;
|
|
||||||
state.end()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|||||||
103
src/filter.rs
103
src/filter.rs
@@ -1,12 +1,7 @@
|
|||||||
use stfu8::encode_u8;
|
|
||||||
|
|
||||||
use crate::display::get_printable_name;
|
|
||||||
use crate::display_node::DisplayNode;
|
use crate::display_node::DisplayNode;
|
||||||
use crate::node::FileTime;
|
|
||||||
use crate::node::Node;
|
use crate::node::Node;
|
||||||
use std::collections::BinaryHeap;
|
use std::collections::BinaryHeap;
|
||||||
use std::collections::HashMap;
|
use std::collections::HashMap;
|
||||||
use std::collections::HashSet;
|
|
||||||
use std::path::Path;
|
use std::path::Path;
|
||||||
use std::path::PathBuf;
|
use std::path::PathBuf;
|
||||||
|
|
||||||
@@ -17,15 +12,9 @@ pub struct AggregateData {
|
|||||||
pub number_of_lines: usize,
|
pub number_of_lines: usize,
|
||||||
pub depth: usize,
|
pub depth: usize,
|
||||||
pub using_a_filter: bool,
|
pub using_a_filter: bool,
|
||||||
pub short_paths: bool,
|
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn get_biggest(
|
pub fn get_biggest(top_level_nodes: Vec<Node>, display_data: AggregateData) -> Option<DisplayNode> {
|
||||||
top_level_nodes: Vec<Node>,
|
|
||||||
display_data: AggregateData,
|
|
||||||
by_filetime: &Option<FileTime>,
|
|
||||||
keep_collapsed: HashSet<PathBuf>,
|
|
||||||
) -> Option<DisplayNode> {
|
|
||||||
if top_level_nodes.is_empty() {
|
if top_level_nodes.is_empty() {
|
||||||
// perhaps change this, bring back Error object?
|
// perhaps change this, bring back Error object?
|
||||||
return None;
|
return None;
|
||||||
@@ -35,26 +24,14 @@ pub fn get_biggest(
|
|||||||
let root;
|
let root;
|
||||||
|
|
||||||
if number_top_level_nodes > 1 {
|
if number_top_level_nodes > 1 {
|
||||||
let size = if by_filetime.is_some() {
|
let size = top_level_nodes.iter().map(|node| node.size).sum();
|
||||||
top_level_nodes
|
|
||||||
.iter()
|
|
||||||
.map(|node| node.size)
|
|
||||||
.max()
|
|
||||||
.unwrap_or(0)
|
|
||||||
} else {
|
|
||||||
top_level_nodes.iter().map(|node| node.size).sum()
|
|
||||||
};
|
|
||||||
|
|
||||||
let nodes = handle_duplicate_top_level_names(top_level_nodes, display_data.short_paths);
|
|
||||||
|
|
||||||
root = Node {
|
root = Node {
|
||||||
name: PathBuf::from("(total)"),
|
name: PathBuf::from("(total)"),
|
||||||
size,
|
size,
|
||||||
children: nodes,
|
children: top_level_nodes,
|
||||||
inode_device: None,
|
inode_device: None,
|
||||||
depth: 0,
|
depth: 0,
|
||||||
};
|
};
|
||||||
|
|
||||||
// Always include the base nodes if we add a 'parent' (total) node
|
// Always include the base nodes if we add a 'parent' (total) node
|
||||||
heap = always_add_children(&display_data, &root, heap);
|
heap = always_add_children(&display_data, &root, heap);
|
||||||
} else {
|
} else {
|
||||||
@@ -62,19 +39,13 @@ pub fn get_biggest(
|
|||||||
heap = add_children(&display_data, &root, heap);
|
heap = add_children(&display_data, &root, heap);
|
||||||
}
|
}
|
||||||
|
|
||||||
Some(fill_remaining_lines(
|
Some(fill_remaining_lines(heap, &root, display_data))
|
||||||
heap,
|
|
||||||
&root,
|
|
||||||
display_data,
|
|
||||||
keep_collapsed,
|
|
||||||
))
|
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn fill_remaining_lines<'a>(
|
pub fn fill_remaining_lines<'a>(
|
||||||
mut heap: BinaryHeap<&'a Node>,
|
mut heap: BinaryHeap<&'a Node>,
|
||||||
root: &'a Node,
|
root: &'a Node,
|
||||||
display_data: AggregateData,
|
display_data: AggregateData,
|
||||||
keep_collapsed: HashSet<PathBuf>,
|
|
||||||
) -> DisplayNode {
|
) -> DisplayNode {
|
||||||
let mut allowed_nodes = HashMap::new();
|
let mut allowed_nodes = HashMap::new();
|
||||||
|
|
||||||
@@ -82,14 +53,10 @@ pub fn fill_remaining_lines<'a>(
|
|||||||
let line = heap.pop();
|
let line = heap.pop();
|
||||||
match line {
|
match line {
|
||||||
Some(line) => {
|
Some(line) => {
|
||||||
// If we are not doing only_file OR if we are doing
|
|
||||||
// only_file and it has no children (ie is a file not a dir)
|
|
||||||
if !display_data.only_file || line.children.is_empty() {
|
if !display_data.only_file || line.children.is_empty() {
|
||||||
allowed_nodes.insert(line.name.as_path(), line);
|
allowed_nodes.insert(line.name.as_path(), line);
|
||||||
}
|
}
|
||||||
if !keep_collapsed.contains(&line.name) {
|
heap = add_children(&display_data, line, heap);
|
||||||
heap = add_children(&display_data, line, heap);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
None => break,
|
None => break,
|
||||||
}
|
}
|
||||||
@@ -147,7 +114,7 @@ fn recursive_rebuilder(allowed_nodes: &HashMap<&Path, &Node>, current: &Node) ->
|
|||||||
.map(|c| recursive_rebuilder(allowed_nodes, c))
|
.map(|c| recursive_rebuilder(allowed_nodes, c))
|
||||||
.collect();
|
.collect();
|
||||||
|
|
||||||
build_display_node(new_children, current)
|
build_node(new_children, current)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Applies all allowed nodes as children to current node
|
// Applies all allowed nodes as children to current node
|
||||||
@@ -160,10 +127,10 @@ fn flat_rebuilder(allowed_nodes: HashMap<&Path, &Node>, current: &Node) -> Displ
|
|||||||
children: vec![],
|
children: vec![],
|
||||||
})
|
})
|
||||||
.collect::<Vec<DisplayNode>>();
|
.collect::<Vec<DisplayNode>>();
|
||||||
build_display_node(new_children, current)
|
build_node(new_children, current)
|
||||||
}
|
}
|
||||||
|
|
||||||
fn build_display_node(mut new_children: Vec<DisplayNode>, current: &Node) -> DisplayNode {
|
fn build_node(mut new_children: Vec<DisplayNode>, current: &Node) -> DisplayNode {
|
||||||
new_children.sort_by(|lhs, rhs| lhs.cmp(rhs).reverse());
|
new_children.sort_by(|lhs, rhs| lhs.cmp(rhs).reverse());
|
||||||
DisplayNode {
|
DisplayNode {
|
||||||
name: current.name.clone(),
|
name: current.name.clone(),
|
||||||
@@ -171,57 +138,3 @@ fn build_display_node(mut new_children: Vec<DisplayNode>, current: &Node) -> Dis
|
|||||||
children: new_children,
|
children: new_children,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
fn names_have_dup(top_level_nodes: &Vec<Node>) -> bool {
|
|
||||||
let mut stored = HashSet::new();
|
|
||||||
for node in top_level_nodes {
|
|
||||||
let name = get_printable_name(&node.name, true);
|
|
||||||
if stored.contains(&name) {
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
stored.insert(name);
|
|
||||||
}
|
|
||||||
false
|
|
||||||
}
|
|
||||||
|
|
||||||
fn handle_duplicate_top_level_names(top_level_nodes: Vec<Node>, short_paths: bool) -> Vec<Node> {
|
|
||||||
// If we have top level names that are the same - we need to tweak them:
|
|
||||||
if short_paths && names_have_dup(&top_level_nodes) {
|
|
||||||
let mut new_top_nodes = top_level_nodes.clone();
|
|
||||||
let mut dir_walk_up_count = 0;
|
|
||||||
|
|
||||||
while names_have_dup(&new_top_nodes) && dir_walk_up_count < 10 {
|
|
||||||
dir_walk_up_count += 1;
|
|
||||||
let mut newer = vec![];
|
|
||||||
|
|
||||||
for node in new_top_nodes.iter() {
|
|
||||||
let mut folders = node.name.iter().rev();
|
|
||||||
// Get parent folder (if second time round get grandparent and so on)
|
|
||||||
for _ in 0..dir_walk_up_count {
|
|
||||||
folders.next();
|
|
||||||
}
|
|
||||||
match folders.next() {
|
|
||||||
// Add (parent_name) to path of Node
|
|
||||||
Some(data) => {
|
|
||||||
let parent = encode_u8(data.as_encoded_bytes());
|
|
||||||
let current_node = node.name.display();
|
|
||||||
let n = Node {
|
|
||||||
name: PathBuf::from(format!("{current_node}({parent})")),
|
|
||||||
size: node.size,
|
|
||||||
children: node.children.clone(),
|
|
||||||
inode_device: node.inode_device,
|
|
||||||
depth: node.depth,
|
|
||||||
};
|
|
||||||
newer.push(n)
|
|
||||||
}
|
|
||||||
// Node does not have a parent
|
|
||||||
None => newer.push(node.clone()),
|
|
||||||
}
|
|
||||||
}
|
|
||||||
new_top_nodes = newer;
|
|
||||||
}
|
|
||||||
new_top_nodes
|
|
||||||
} else {
|
|
||||||
top_level_nodes
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|||||||
@@ -1,5 +1,4 @@
|
|||||||
use crate::display_node::DisplayNode;
|
use crate::display_node::DisplayNode;
|
||||||
use crate::node::FileTime;
|
|
||||||
use crate::node::Node;
|
use crate::node::Node;
|
||||||
use std::collections::HashMap;
|
use std::collections::HashMap;
|
||||||
use std::ffi::OsStr;
|
use std::ffi::OsStr;
|
||||||
@@ -11,11 +10,7 @@ struct ExtensionNode<'a> {
|
|||||||
extension: Option<&'a OsStr>,
|
extension: Option<&'a OsStr>,
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn get_all_file_types(
|
pub fn get_all_file_types(top_level_nodes: &[Node], n: usize) -> Option<DisplayNode> {
|
||||||
top_level_nodes: &[Node],
|
|
||||||
n: usize,
|
|
||||||
by_filetime: &Option<FileTime>,
|
|
||||||
) -> Option<DisplayNode> {
|
|
||||||
let ext_nodes = {
|
let ext_nodes = {
|
||||||
let mut extension_cumulative_sizes = HashMap::new();
|
let mut extension_cumulative_sizes = HashMap::new();
|
||||||
build_by_all_file_types(top_level_nodes, &mut extension_cumulative_sizes);
|
build_by_all_file_types(top_level_nodes, &mut extension_cumulative_sizes);
|
||||||
@@ -49,27 +44,16 @@ pub fn get_all_file_types(
|
|||||||
|
|
||||||
// ...then, aggregate the remaining nodes (if any) into a single "(others)" node
|
// ...then, aggregate the remaining nodes (if any) into a single "(others)" node
|
||||||
if ext_nodes_iter.len() > 0 {
|
if ext_nodes_iter.len() > 0 {
|
||||||
let actual_size = if by_filetime.is_some() {
|
|
||||||
ext_nodes_iter.map(|node| node.size).max().unwrap_or(0)
|
|
||||||
} else {
|
|
||||||
ext_nodes_iter.map(|node| node.size).sum()
|
|
||||||
};
|
|
||||||
displayed.push(DisplayNode {
|
displayed.push(DisplayNode {
|
||||||
name: PathBuf::from("(others)"),
|
name: PathBuf::from("(others)"),
|
||||||
size: actual_size,
|
size: ext_nodes_iter.map(|node| node.size).sum(),
|
||||||
children: vec![],
|
children: vec![],
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
let actual_size: u64 = if by_filetime.is_some() {
|
|
||||||
displayed.iter().map(|node| node.size).max().unwrap_or(0)
|
|
||||||
} else {
|
|
||||||
displayed.iter().map(|node| node.size).sum()
|
|
||||||
};
|
|
||||||
|
|
||||||
let result = DisplayNode {
|
let result = DisplayNode {
|
||||||
name: PathBuf::from("(total)"),
|
name: PathBuf::from("(total)"),
|
||||||
size: actual_size,
|
size: displayed.iter().map(|node| node.size).sum(),
|
||||||
children: displayed,
|
children: displayed,
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|||||||
81
src/main.rs
81
src/main.rs
@@ -24,21 +24,21 @@ use std::fs::read_to_string;
|
|||||||
use std::io;
|
use std::io;
|
||||||
use std::panic;
|
use std::panic;
|
||||||
use std::process;
|
use std::process;
|
||||||
|
use std::sync::atomic::AtomicBool;
|
||||||
|
use std::sync::atomic::Ordering;
|
||||||
use std::sync::Arc;
|
use std::sync::Arc;
|
||||||
use std::sync::Mutex;
|
use std::sync::Mutex;
|
||||||
use sysinfo::{System, SystemExt};
|
use sysinfo::{System, SystemExt};
|
||||||
use utils::canonicalize_absolute_path;
|
|
||||||
|
|
||||||
use self::display::draw_it;
|
use self::display::draw_it;
|
||||||
use config::get_config;
|
use config::get_config;
|
||||||
use dir_walker::walk_it;
|
use dir_walker::walk_it;
|
||||||
use display_node::OUTPUT_TYPE;
|
|
||||||
use filter::get_biggest;
|
use filter::get_biggest;
|
||||||
use filter_type::get_all_file_types;
|
use filter_type::get_all_file_types;
|
||||||
use regex::Regex;
|
use regex::Regex;
|
||||||
use std::cmp::max;
|
use std::cmp::max;
|
||||||
use std::path::PathBuf;
|
use std::path::PathBuf;
|
||||||
use terminal_size::{Height, Width, terminal_size};
|
use terminal_size::{terminal_size, Height, Width};
|
||||||
use utils::get_filesystem_devices;
|
use utils::get_filesystem_devices;
|
||||||
use utils::simplify_dir_names;
|
use utils::simplify_dir_names;
|
||||||
|
|
||||||
@@ -81,19 +81,23 @@ fn should_init_color(no_color: bool, force_color: bool) -> bool {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn get_height_of_terminal() -> usize {
|
fn get_height_of_terminal() -> usize {
|
||||||
|
// Simplify once https://github.com/eminence/terminal-size/pull/41 is
|
||||||
|
// merged
|
||||||
terminal_size()
|
terminal_size()
|
||||||
// Windows CI runners detect a terminal height of 0
|
// Windows CI runners detect a terminal height of 0
|
||||||
.map(|(_, Height(h))| max(h.into(), DEFAULT_NUMBER_OF_LINES))
|
.map(|(_, Height(h))| max(h as usize, DEFAULT_NUMBER_OF_LINES))
|
||||||
.unwrap_or(DEFAULT_NUMBER_OF_LINES)
|
.unwrap_or(DEFAULT_NUMBER_OF_LINES)
|
||||||
- 10
|
- 10
|
||||||
}
|
}
|
||||||
|
|
||||||
fn get_width_of_terminal() -> usize {
|
fn get_width_of_terminal() -> usize {
|
||||||
|
// Simplify once https://github.com/eminence/terminal-size/pull/41 is
|
||||||
|
// merged
|
||||||
terminal_size()
|
terminal_size()
|
||||||
.map(|(Width(w), _)| match cfg!(windows) {
|
.map(|(Width(w), _)| match cfg!(windows) {
|
||||||
// Windows CI runners detect a very low terminal width
|
// Windows CI runners detect a very low terminal width
|
||||||
true => max(w.into(), DEFAULT_TERMINAL_WIDTH),
|
true => max(w as usize, DEFAULT_TERMINAL_WIDTH),
|
||||||
false => w.into(),
|
false => w as usize,
|
||||||
})
|
})
|
||||||
.unwrap_or(DEFAULT_TERMINAL_WIDTH)
|
.unwrap_or(DEFAULT_TERMINAL_WIDTH)
|
||||||
}
|
}
|
||||||
@@ -112,18 +116,25 @@ fn get_regex_value(maybe_value: Option<ValuesRef<String>>) -> Vec<Regex> {
|
|||||||
|
|
||||||
fn main() {
|
fn main() {
|
||||||
let options = build_cli().get_matches();
|
let options = build_cli().get_matches();
|
||||||
let config = get_config(options.get_one::<String>("config").cloned());
|
let config = get_config();
|
||||||
|
|
||||||
let errors = RuntimeErrors::default();
|
let errors = RuntimeErrors::default();
|
||||||
let error_listen_for_ctrlc = Arc::new(Mutex::new(errors));
|
let error_listen_for_ctrlc = Arc::new(Mutex::new(errors));
|
||||||
let errors_for_rayon = error_listen_for_ctrlc.clone();
|
let errors_for_rayon = error_listen_for_ctrlc.clone();
|
||||||
|
let errors_final = error_listen_for_ctrlc.clone();
|
||||||
|
let is_in_listing = Arc::new(AtomicBool::new(false));
|
||||||
|
let cloned_is_in_listing = Arc::clone(&is_in_listing);
|
||||||
|
|
||||||
ctrlc::set_handler(move || {
|
ctrlc::set_handler(move || {
|
||||||
|
error_listen_for_ctrlc.lock().unwrap().abort = true;
|
||||||
println!("\nAborting");
|
println!("\nAborting");
|
||||||
process::exit(1);
|
if cloned_is_in_listing.load(Ordering::Relaxed) {
|
||||||
|
process::exit(1);
|
||||||
|
}
|
||||||
})
|
})
|
||||||
.expect("Error setting Ctrl-C handler");
|
.expect("Error setting Ctrl-C handler");
|
||||||
|
|
||||||
|
is_in_listing.store(true, Ordering::Relaxed);
|
||||||
let target_dirs = match config.get_files_from(&options) {
|
let target_dirs = match config.get_files_from(&options) {
|
||||||
Some(path) => {
|
Some(path) => {
|
||||||
if path == "-" {
|
if path == "-" {
|
||||||
@@ -153,6 +164,7 @@ fn main() {
|
|||||||
None => vec![".".to_owned()],
|
None => vec![".".to_owned()],
|
||||||
},
|
},
|
||||||
};
|
};
|
||||||
|
is_in_listing.store(false, Ordering::Relaxed);
|
||||||
|
|
||||||
let summarize_file_types = options.get_flag("types");
|
let summarize_file_types = options.get_flag("types");
|
||||||
|
|
||||||
@@ -189,7 +201,6 @@ fn main() {
|
|||||||
Some(values) => values
|
Some(values) => values
|
||||||
.map(|v| v.as_str())
|
.map(|v| v.as_str())
|
||||||
.map(PathBuf::from)
|
.map(PathBuf::from)
|
||||||
.map(canonicalize_absolute_path)
|
|
||||||
.collect::<Vec<PathBuf>>(),
|
.collect::<Vec<PathBuf>>(),
|
||||||
None => vec![],
|
None => vec![],
|
||||||
};
|
};
|
||||||
@@ -213,12 +224,11 @@ fn main() {
|
|||||||
.collect::<Vec<Regex>>();
|
.collect::<Vec<Regex>>();
|
||||||
|
|
||||||
let by_filecount = options.get_flag("by_filecount");
|
let by_filecount = options.get_flag("by_filecount");
|
||||||
let by_filetime = config.get_filetime(&options);
|
|
||||||
let limit_filesystem = options.get_flag("limit_filesystem");
|
let limit_filesystem = options.get_flag("limit_filesystem");
|
||||||
let follow_links = options.get_flag("dereference_links");
|
let follow_links = options.get_flag("dereference_links");
|
||||||
|
|
||||||
let allowed_filesystems = limit_filesystem
|
let allowed_filesystems = limit_filesystem
|
||||||
.then(|| get_filesystem_devices(&target_dirs, follow_links))
|
.then(|| get_filesystem_devices(&target_dirs))
|
||||||
.unwrap_or_default();
|
.unwrap_or_default();
|
||||||
let simplified_dirs = simplify_dir_names(&target_dirs);
|
let simplified_dirs = simplify_dir_names(&target_dirs);
|
||||||
|
|
||||||
@@ -236,22 +246,9 @@ fn main() {
|
|||||||
indicator.spawn(output_format.clone())
|
indicator.spawn(output_format.clone())
|
||||||
}
|
}
|
||||||
|
|
||||||
let keep_collapsed: HashSet<PathBuf> = match options.get_many::<String>("collapse") {
|
|
||||||
Some(collapse) => {
|
|
||||||
let mut combined_dirs = HashSet::new();
|
|
||||||
for collapse_dir in collapse {
|
|
||||||
for target_dir in target_dirs.iter() {
|
|
||||||
combined_dirs.insert(PathBuf::from(target_dir).join(collapse_dir));
|
|
||||||
}
|
|
||||||
}
|
|
||||||
combined_dirs
|
|
||||||
}
|
|
||||||
None => HashSet::new(),
|
|
||||||
};
|
|
||||||
|
|
||||||
let filter_modified_time = config.get_modified_time_operator(&options);
|
let filter_modified_time = config.get_modified_time_operator(&options);
|
||||||
let filter_accessed_time = config.get_accessed_time_operator(&options);
|
let filter_accessed_time = config.get_accessed_time_operator(&options);
|
||||||
let filter_changed_time = config.get_changed_time_operator(&options);
|
let filter_changed_time = config.get_created_time_operator(&options);
|
||||||
|
|
||||||
let walk_data = WalkData {
|
let walk_data = WalkData {
|
||||||
ignore_directories: ignored_full_path,
|
ignore_directories: ignored_full_path,
|
||||||
@@ -263,7 +260,6 @@ fn main() {
|
|||||||
filter_changed_time,
|
filter_changed_time,
|
||||||
use_apparent_size: config.get_apparent_size(&options),
|
use_apparent_size: config.get_apparent_size(&options),
|
||||||
by_filecount,
|
by_filecount,
|
||||||
by_filetime: &by_filetime,
|
|
||||||
ignore_hidden,
|
ignore_hidden,
|
||||||
follow_links,
|
follow_links,
|
||||||
progress_data: indicator.data.clone(),
|
progress_data: indicator.data.clone(),
|
||||||
@@ -276,7 +272,7 @@ fn main() {
|
|||||||
let top_level_nodes = walk_it(simplified_dirs, &walk_data);
|
let top_level_nodes = walk_it(simplified_dirs, &walk_data);
|
||||||
|
|
||||||
let tree = match summarize_file_types {
|
let tree = match summarize_file_types {
|
||||||
true => get_all_file_types(&top_level_nodes, number_of_lines, &by_filetime),
|
true => get_all_file_types(&top_level_nodes, number_of_lines),
|
||||||
false => {
|
false => {
|
||||||
let agg_data = AggregateData {
|
let agg_data = AggregateData {
|
||||||
min_size: config.get_min_size(&options),
|
min_size: config.get_min_size(&options),
|
||||||
@@ -285,15 +281,18 @@ fn main() {
|
|||||||
number_of_lines,
|
number_of_lines,
|
||||||
depth,
|
depth,
|
||||||
using_a_filter: !filter_regexs.is_empty() || !invert_filter_regexs.is_empty(),
|
using_a_filter: !filter_regexs.is_empty() || !invert_filter_regexs.is_empty(),
|
||||||
short_paths: !config.get_full_paths(&options),
|
|
||||||
};
|
};
|
||||||
get_biggest(top_level_nodes, agg_data, &by_filetime, keep_collapsed)
|
get_biggest(top_level_nodes, agg_data)
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
|
||||||
// Must have stopped indicator before we print to stderr
|
// Must have stopped indicator before we print to stderr
|
||||||
indicator.stop();
|
indicator.stop();
|
||||||
|
|
||||||
|
if errors_final.lock().unwrap().abort {
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
let final_errors = walk_data.errors.lock().unwrap();
|
let final_errors = walk_data.errors.lock().unwrap();
|
||||||
if !final_errors.file_not_found.is_empty() {
|
if !final_errors.file_not_found.is_empty() {
|
||||||
let err = final_errors
|
let err = final_errors
|
||||||
@@ -330,23 +329,19 @@ fn main() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
if let Some(root_node) = tree {
|
if let Some(root_node) = tree {
|
||||||
|
let idd = InitialDisplayData {
|
||||||
|
short_paths: !config.get_full_paths(&options),
|
||||||
|
is_reversed: !config.get_reverse(&options),
|
||||||
|
colors_on: is_colors,
|
||||||
|
by_filecount,
|
||||||
|
is_screen_reader: config.get_screen_reader(&options),
|
||||||
|
output_format,
|
||||||
|
bars_on_right: config.get_bars_on_right(&options),
|
||||||
|
};
|
||||||
|
|
||||||
if config.get_output_json(&options) {
|
if config.get_output_json(&options) {
|
||||||
OUTPUT_TYPE.with(|wrapped| {
|
|
||||||
wrapped.replace(output_format);
|
|
||||||
});
|
|
||||||
println!("{}", serde_json::to_string(&root_node).unwrap());
|
println!("{}", serde_json::to_string(&root_node).unwrap());
|
||||||
} else {
|
} else {
|
||||||
let idd = InitialDisplayData {
|
|
||||||
short_paths: !config.get_full_paths(&options),
|
|
||||||
is_reversed: !config.get_reverse(&options),
|
|
||||||
colors_on: is_colors,
|
|
||||||
by_filecount,
|
|
||||||
by_filetime,
|
|
||||||
is_screen_reader: config.get_screen_reader(&options),
|
|
||||||
output_format,
|
|
||||||
bars_on_right: config.get_bars_on_right(&options),
|
|
||||||
};
|
|
||||||
|
|
||||||
draw_it(
|
draw_it(
|
||||||
idd,
|
idd,
|
||||||
config.get_no_bars(&options),
|
config.get_no_bars(&options),
|
||||||
|
|||||||
35
src/node.rs
35
src/node.rs
@@ -16,13 +16,6 @@ pub struct Node {
|
|||||||
pub depth: usize,
|
pub depth: usize,
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug, PartialEq)]
|
|
||||||
pub enum FileTime {
|
|
||||||
Modified,
|
|
||||||
Accessed,
|
|
||||||
Changed,
|
|
||||||
}
|
|
||||||
|
|
||||||
#[allow(clippy::too_many_arguments)]
|
#[allow(clippy::too_many_arguments)]
|
||||||
pub fn build_node(
|
pub fn build_node(
|
||||||
dir: PathBuf,
|
dir: PathBuf,
|
||||||
@@ -34,23 +27,22 @@ pub fn build_node(
|
|||||||
) -> Option<Node> {
|
) -> Option<Node> {
|
||||||
let use_apparent_size = walk_data.use_apparent_size;
|
let use_apparent_size = walk_data.use_apparent_size;
|
||||||
let by_filecount = walk_data.by_filecount;
|
let by_filecount = walk_data.by_filecount;
|
||||||
let by_filetime = &walk_data.by_filetime;
|
|
||||||
|
|
||||||
get_metadata(
|
get_metadata(&dir, use_apparent_size).map(|data| {
|
||||||
&dir,
|
let inode_device = if is_symlink && !use_apparent_size {
|
||||||
use_apparent_size,
|
None
|
||||||
walk_data.follow_links && is_symlink,
|
} else {
|
||||||
)
|
data.1
|
||||||
.map(|data| {
|
};
|
||||||
let inode_device = data.1;
|
|
||||||
|
|
||||||
let size = if is_filtered_out_due_to_regex(walk_data.filter_regex, &dir)
|
let size = if is_filtered_out_due_to_regex(walk_data.filter_regex, &dir)
|
||||||
|| is_filtered_out_due_to_invert_regex(walk_data.invert_filter_regex, &dir)
|
|| is_filtered_out_due_to_invert_regex(walk_data.invert_filter_regex, &dir)
|
||||||
|
|| (is_symlink && !use_apparent_size)
|
||||||
|| by_filecount && !is_file
|
|| by_filecount && !is_file
|
||||||
|| [
|
|| [
|
||||||
(&walk_data.filter_modified_time, data.2.0),
|
(&walk_data.filter_modified_time, data.2 .0),
|
||||||
(&walk_data.filter_accessed_time, data.2.1),
|
(&walk_data.filter_accessed_time, data.2 .1),
|
||||||
(&walk_data.filter_changed_time, data.2.2),
|
(&walk_data.filter_changed_time, data.2 .2),
|
||||||
]
|
]
|
||||||
.iter()
|
.iter()
|
||||||
.any(|(filter_time, actual_time)| {
|
.any(|(filter_time, actual_time)| {
|
||||||
@@ -59,13 +51,6 @@ pub fn build_node(
|
|||||||
0
|
0
|
||||||
} else if by_filecount {
|
} else if by_filecount {
|
||||||
1
|
1
|
||||||
} else if by_filetime.is_some() {
|
|
||||||
match by_filetime {
|
|
||||||
Some(FileTime::Modified) => data.2.0.unsigned_abs(),
|
|
||||||
Some(FileTime::Accessed) => data.2.1.unsigned_abs(),
|
|
||||||
Some(FileTime::Changed) => data.2.2.unsigned_abs(),
|
|
||||||
None => unreachable!(),
|
|
||||||
}
|
|
||||||
} else {
|
} else {
|
||||||
data.0
|
data.0
|
||||||
};
|
};
|
||||||
|
|||||||
@@ -17,15 +17,9 @@ type FileTime = (i64, i64, i64);
|
|||||||
pub fn get_metadata<P: AsRef<Path>>(
|
pub fn get_metadata<P: AsRef<Path>>(
|
||||||
path: P,
|
path: P,
|
||||||
use_apparent_size: bool,
|
use_apparent_size: bool,
|
||||||
follow_links: bool,
|
|
||||||
) -> Option<(u64, Option<InodeAndDevice>, FileTime)> {
|
) -> Option<(u64, Option<InodeAndDevice>, FileTime)> {
|
||||||
use std::os::unix::fs::MetadataExt;
|
use std::os::unix::fs::MetadataExt;
|
||||||
let metadata = if follow_links {
|
match path.as_ref().metadata() {
|
||||||
path.as_ref().metadata()
|
|
||||||
} else {
|
|
||||||
path.as_ref().symlink_metadata()
|
|
||||||
};
|
|
||||||
match metadata {
|
|
||||||
Ok(md) => {
|
Ok(md) => {
|
||||||
if use_apparent_size {
|
if use_apparent_size {
|
||||||
Some((
|
Some((
|
||||||
@@ -49,7 +43,6 @@ pub fn get_metadata<P: AsRef<Path>>(
|
|||||||
pub fn get_metadata<P: AsRef<Path>>(
|
pub fn get_metadata<P: AsRef<Path>>(
|
||||||
path: P,
|
path: P,
|
||||||
use_apparent_size: bool,
|
use_apparent_size: bool,
|
||||||
follow_links: bool,
|
|
||||||
) -> Option<(u64, Option<InodeAndDevice>, FileTime)> {
|
) -> Option<(u64, Option<InodeAndDevice>, FileTime)> {
|
||||||
// On windows opening the file to get size, file ID and volume can be very
|
// On windows opening the file to get size, file ID and volume can be very
|
||||||
// expensive because 1) it causes a few system calls, and more importantly 2) it can cause
|
// expensive because 1) it causes a few system calls, and more importantly 2) it can cause
|
||||||
@@ -149,12 +142,7 @@ pub fn get_metadata<P: AsRef<Path>>(
|
|||||||
|
|
||||||
use std::os::windows::fs::MetadataExt;
|
use std::os::windows::fs::MetadataExt;
|
||||||
let path = path.as_ref();
|
let path = path.as_ref();
|
||||||
let metadata = if follow_links {
|
match path.metadata() {
|
||||||
path.metadata()
|
|
||||||
} else {
|
|
||||||
path.symlink_metadata()
|
|
||||||
};
|
|
||||||
match metadata {
|
|
||||||
Ok(ref md) => {
|
Ok(ref md) => {
|
||||||
const FILE_ATTRIBUTE_ARCHIVE: u32 = 0x20;
|
const FILE_ATTRIBUTE_ARCHIVE: u32 = 0x20;
|
||||||
const FILE_ATTRIBUTE_READONLY: u32 = 0x01;
|
const FILE_ATTRIBUTE_READONLY: u32 = 0x01;
|
||||||
|
|||||||
@@ -3,19 +3,14 @@ use std::{
|
|||||||
io::Write,
|
io::Write,
|
||||||
path::Path,
|
path::Path,
|
||||||
sync::{
|
sync::{
|
||||||
Arc, RwLock,
|
atomic::{AtomicU64, AtomicU8, AtomicUsize, Ordering},
|
||||||
atomic::{AtomicU8, AtomicUsize, Ordering},
|
|
||||||
mpsc::{self, RecvTimeoutError, Sender},
|
mpsc::{self, RecvTimeoutError, Sender},
|
||||||
|
Arc, RwLock,
|
||||||
},
|
},
|
||||||
thread::JoinHandle,
|
thread::JoinHandle,
|
||||||
time::Duration,
|
time::Duration,
|
||||||
};
|
};
|
||||||
|
|
||||||
#[cfg(not(target_has_atomic = "64"))]
|
|
||||||
use portable_atomic::AtomicU64;
|
|
||||||
#[cfg(target_has_atomic = "64")]
|
|
||||||
use std::sync::atomic::AtomicU64;
|
|
||||||
|
|
||||||
use crate::display::human_readable_number;
|
use crate::display::human_readable_number;
|
||||||
|
|
||||||
/* -------------------------------------------------------------------------- */
|
/* -------------------------------------------------------------------------- */
|
||||||
@@ -78,7 +73,7 @@ pub struct RuntimeErrors {
|
|||||||
pub no_permissions: HashSet<String>,
|
pub no_permissions: HashSet<String>,
|
||||||
pub file_not_found: HashSet<String>,
|
pub file_not_found: HashSet<String>,
|
||||||
pub unknown_error: HashSet<String>,
|
pub unknown_error: HashSet<String>,
|
||||||
pub interrupted_error: i32,
|
pub abort: bool,
|
||||||
}
|
}
|
||||||
|
|
||||||
/* -------------------------------------------------------------------------- */
|
/* -------------------------------------------------------------------------- */
|
||||||
@@ -118,7 +113,7 @@ impl PIndicator {
|
|||||||
|
|
||||||
let time_info_thread = std::thread::spawn(move || {
|
let time_info_thread = std::thread::spawn(move || {
|
||||||
let mut progress_char_i: usize = 0;
|
let mut progress_char_i: usize = 0;
|
||||||
let mut stderr = std::io::stderr();
|
let mut stdout = std::io::stdout();
|
||||||
let mut msg = "".to_string();
|
let mut msg = "".to_string();
|
||||||
|
|
||||||
// While the timeout triggers we go round the loop
|
// While the timeout triggers we go round the loop
|
||||||
@@ -127,8 +122,7 @@ impl PIndicator {
|
|||||||
receiver.recv_timeout(Duration::from_millis(SPINNER_SLEEP_TIME))
|
receiver.recv_timeout(Duration::from_millis(SPINNER_SLEEP_TIME))
|
||||||
{
|
{
|
||||||
// Clear the text written by 'write!'& Return at the start of line
|
// Clear the text written by 'write!'& Return at the start of line
|
||||||
let clear = format!("\r{:width$}", " ", width = msg.len());
|
print!("\r{:width$}", " ", width = msg.len());
|
||||||
write!(stderr, "{clear}").unwrap();
|
|
||||||
let prog_char = PROGRESS_CHARS[progress_char_i];
|
let prog_char = PROGRESS_CHARS[progress_char_i];
|
||||||
|
|
||||||
msg = match data.state.load(ORDERING) {
|
msg = match data.state.load(ORDERING) {
|
||||||
@@ -137,17 +131,15 @@ impl PIndicator {
|
|||||||
_ => panic!("Unknown State"),
|
_ => panic!("Unknown State"),
|
||||||
};
|
};
|
||||||
|
|
||||||
write!(stderr, "\r{msg}").unwrap();
|
write!(stdout, "\r{msg}").unwrap();
|
||||||
stderr.flush().unwrap();
|
stdout.flush().unwrap();
|
||||||
|
|
||||||
progress_char_i += 1;
|
progress_char_i += 1;
|
||||||
progress_char_i %= PROGRESS_CHARS_LEN;
|
progress_char_i %= PROGRESS_CHARS_LEN;
|
||||||
}
|
}
|
||||||
|
print!("\r{:width$}", " ", width = msg.len());
|
||||||
let clear = format!("\r{:width$}", " ", width = msg.len());
|
print!("\r");
|
||||||
write!(stderr, "{clear}").unwrap();
|
stdout.flush().unwrap();
|
||||||
write!(stderr, "\r").unwrap();
|
|
||||||
stderr.flush().unwrap();
|
|
||||||
});
|
});
|
||||||
self.thread = Some((stop_handler, time_info_thread))
|
self.thread = Some((stop_handler, time_info_thread))
|
||||||
}
|
}
|
||||||
|
|||||||
45
src/utils.rs
45
src/utils.rs
@@ -4,7 +4,7 @@ use std::path::{Path, PathBuf};
|
|||||||
|
|
||||||
use crate::config::DAY_SECONDS;
|
use crate::config::DAY_SECONDS;
|
||||||
|
|
||||||
use crate::dir_walker::Operator;
|
use crate::dir_walker::Operater;
|
||||||
use crate::platform;
|
use crate::platform;
|
||||||
use regex::Regex;
|
use regex::Regex;
|
||||||
|
|
||||||
@@ -34,25 +34,13 @@ pub fn simplify_dir_names<P: AsRef<Path>>(dirs: &[P]) -> HashSet<PathBuf> {
|
|||||||
top_level_names
|
top_level_names
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn get_filesystem_devices<P: AsRef<Path>>(paths: &[P], follow_links: bool) -> HashSet<u64> {
|
pub fn get_filesystem_devices<P: AsRef<Path>>(paths: &[P]) -> HashSet<u64> {
|
||||||
use std::fs;
|
|
||||||
// Gets the device ids for the filesystems which are used by the argument paths
|
// Gets the device ids for the filesystems which are used by the argument paths
|
||||||
paths
|
paths
|
||||||
.iter()
|
.iter()
|
||||||
.filter_map(|p| {
|
.filter_map(|p| match get_metadata(p, false) {
|
||||||
let follow_links = if follow_links {
|
Some((_size, Some((_id, dev)), _time)) => Some(dev),
|
||||||
// slow path: If dereference-links is set, then we check if the file is a symbolic link
|
_ => None,
|
||||||
match fs::symlink_metadata(p) {
|
|
||||||
Ok(metadata) => metadata.file_type().is_symlink(),
|
|
||||||
Err(_) => false,
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
false
|
|
||||||
};
|
|
||||||
match get_metadata(p, false, follow_links) {
|
|
||||||
Some((_size, Some((_id, dev)), _time)) => Some(dev),
|
|
||||||
_ => None,
|
|
||||||
}
|
|
||||||
})
|
})
|
||||||
.collect()
|
.collect()
|
||||||
}
|
}
|
||||||
@@ -67,17 +55,6 @@ pub fn normalize_path<P: AsRef<Path>>(path: P) -> PathBuf {
|
|||||||
path.as_ref().components().collect()
|
path.as_ref().components().collect()
|
||||||
}
|
}
|
||||||
|
|
||||||
// Canonicalize the path only if it is an absolute path
|
|
||||||
pub fn canonicalize_absolute_path(path: PathBuf) -> PathBuf {
|
|
||||||
if !path.is_absolute() {
|
|
||||||
return path;
|
|
||||||
}
|
|
||||||
match std::fs::canonicalize(&path) {
|
|
||||||
Ok(canonicalized_path) => canonicalized_path,
|
|
||||||
Err(_) => path,
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
pub fn is_filtered_out_due_to_regex(filter_regex: &[Regex], dir: &Path) -> bool {
|
pub fn is_filtered_out_due_to_regex(filter_regex: &[Regex], dir: &Path) -> bool {
|
||||||
if filter_regex.is_empty() {
|
if filter_regex.is_empty() {
|
||||||
false
|
false
|
||||||
@@ -88,17 +65,13 @@ pub fn is_filtered_out_due_to_regex(filter_regex: &[Regex], dir: &Path) -> bool
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn is_filtered_out_due_to_file_time(
|
pub fn is_filtered_out_due_to_file_time(filter_time: &(Operater, i64), actual_time: i64) -> bool {
|
||||||
filter_time: &Option<(Operator, i64)>,
|
|
||||||
actual_time: i64,
|
|
||||||
) -> bool {
|
|
||||||
match filter_time {
|
match filter_time {
|
||||||
None => false,
|
(Operater::Equal, bound_time) => {
|
||||||
Some((Operator::Equal, bound_time)) => {
|
|
||||||
!(actual_time >= *bound_time && actual_time < *bound_time + DAY_SECONDS)
|
!(actual_time >= *bound_time && actual_time < *bound_time + DAY_SECONDS)
|
||||||
}
|
}
|
||||||
Some((Operator::GreaterThan, bound_time)) => actual_time < *bound_time,
|
(Operater::GreaterThan, bound_time) => actual_time < *bound_time,
|
||||||
Some((Operator::LessThan, bound_time)) => actual_time > *bound_time,
|
(Operater::LessThan, bound_time) => actual_time > *bound_time,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -2,7 +2,7 @@ use assert_cmd::Command;
|
|||||||
use std::ffi::OsStr;
|
use std::ffi::OsStr;
|
||||||
use std::process::Output;
|
use std::process::Output;
|
||||||
use std::sync::Once;
|
use std::sync::Once;
|
||||||
use std::{io, str};
|
use std::{fs, io, str};
|
||||||
|
|
||||||
static INIT: Once = Once::new();
|
static INIT: Once = Once::new();
|
||||||
static UNREADABLE_DIR_PATH: &str = "/tmp/unreadable_dir";
|
static UNREADABLE_DIR_PATH: &str = "/tmp/unreadable_dir";
|
||||||
@@ -38,7 +38,6 @@ fn copy_test_data(dir: &str) {
|
|||||||
fn create_unreadable_directory() -> io::Result<()> {
|
fn create_unreadable_directory() -> io::Result<()> {
|
||||||
#[cfg(unix)]
|
#[cfg(unix)]
|
||||||
{
|
{
|
||||||
use std::fs;
|
|
||||||
use std::fs::Permissions;
|
use std::fs::Permissions;
|
||||||
use std::os::unix::fs::PermissionsExt;
|
use std::os::unix::fs::PermissionsExt;
|
||||||
fs::create_dir_all(UNREADABLE_DIR_PATH)?;
|
fs::create_dir_all(UNREADABLE_DIR_PATH)?;
|
||||||
|
|||||||
@@ -10,9 +10,6 @@ use std::str;
|
|||||||
|
|
||||||
fn build_command<T: AsRef<OsStr>>(command_args: Vec<T>) -> String {
|
fn build_command<T: AsRef<OsStr>>(command_args: Vec<T>) -> String {
|
||||||
let mut cmd = &mut Command::cargo_bin("dust").unwrap();
|
let mut cmd = &mut Command::cargo_bin("dust").unwrap();
|
||||||
// Hide progress bar
|
|
||||||
cmd = cmd.arg("-P");
|
|
||||||
|
|
||||||
for p in command_args {
|
for p in command_args {
|
||||||
cmd = cmd.arg(p);
|
cmd = cmd.arg(p);
|
||||||
}
|
}
|
||||||
@@ -257,26 +254,3 @@ pub fn test_force_color() {
|
|||||||
assert!(output.contains("\x1B[31m"));
|
assert!(output.contains("\x1B[31m"));
|
||||||
assert!(output.contains("\x1B[0m"));
|
assert!(output.contains("\x1B[0m"));
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
|
||||||
pub fn test_collapse() {
|
|
||||||
let output = build_command(vec!["--collapse", "many", "tests/test_dir/"]);
|
|
||||||
assert!(output.contains("many"));
|
|
||||||
assert!(!output.contains("hello_file"));
|
|
||||||
}
|
|
||||||
|
|
||||||
#[test]
|
|
||||||
pub fn test_handle_duplicate_names() {
|
|
||||||
// Check that even if we run on a multiple directories with the same name
|
|
||||||
// we still show the distinct parent dir in the output
|
|
||||||
let output = build_command(vec![
|
|
||||||
"tests/test_dir_matching/dave/dup_name",
|
|
||||||
"tests/test_dir_matching/andy/dup_name",
|
|
||||||
"ci",
|
|
||||||
]);
|
|
||||||
assert!(output.contains("andy"));
|
|
||||||
assert!(output.contains("dave"));
|
|
||||||
assert!(output.contains("ci"));
|
|
||||||
assert!(output.contains("dup_name"));
|
|
||||||
assert!(!output.contains("test_dir_matching"));
|
|
||||||
}
|
|
||||||
|
|||||||
Reference in New Issue
Block a user