Comment by pxc

Comment by pxc 2 days ago

4 replies

I do it this way but I indent the rest of the pipeline (like one typically would in a functional language with a pipeline operator or thread macro, or in an OOP language with method chaining via `.`):

  foo --long-something \
    | bar --long-other
and if the lines in subsequent commands in the pipeline start to get long, I also indent their arguments:

  foo --long-flag \
    | bar \
        --long-other-flag \
        --long-option a \
        --long-option b \
    | baz \
        --long-another-flag \
        --long-flag-again \
    > /path/to/some/file
I really like to use && on its own line like that. One of my favorite things about Fish is how it turns && and || from special syntax into commands, which it calls combiners, so you could write:

  foo \
    | bar \
    | baz
  and echo done
I use this often for conditions whose bodies would only be one line, avoiding the nesting and indentation:

  test -n "$SOMETHING"
  or set -x SOMETHING some-default

  command $SOMETHING
In Bash, I usually use parameter substitution for this, but in other situations (other than setting default values for vars) I throw a backslash at the end of a line, indent and use && or ||, imitating the Fish style.

One of my favorite patterns for readability is to use indented, long-form pipelines like this to set variables. They work fine inside subshells, but for uniformity and clarity I prefer to do

  shopt -s lastpipe

  foo \
    | bar \
    | baz \
    | read SOMEVAR
I really enjoy 'maximizing' pipelines like this because it makes it possible to use long pipelines everywhere without making your program terse and mysterious, or creating unwieldy long lines.

If you do this, you end up with a script is mostly 'flat' (having very little nested control flow, largely avoiding loops), has very few variable assignments, and predictably locates the variable assignments it does have at the ends of pipelines. Each pipeline is a singular train of thought requiring you to consider context and state only at the very beginning and very end, and you can typically likewise think of all the intermediate steps/commands in functional terms.

I tend to write all of my shell scripts this way, including the ones I write interactively at my prompt. One really cool thing about shell languages is that unlike in 'real' programming languages, loops are actually composable! So you can freely mix ad-hoc/imperative and pipeline-centric styles like this (example is Fish):

  find -name whatever -exec basename '{}' \;
      | while read -l data
          set -l temp (some-series $data)
          set -l another (some-command $temp)
          blahdiblah --something $temp --other $another
      end \
      | bar \
      | baz \
      > some-list.txt
(I typically use 2 spaces to indent when writing Bash scripts, but Fish defaults to 4 in the prompt, which it also automatically indents for you. I'm not sure if that's configurable but I haven't attempted to change it.)

I tend to follow my guideline suggested earlier and do this only close to the very beginning or very end of a pipeline if that loop actually modifies the filesystem or non-local variables, but it's really nice to have that flexibility imo. (It's especially handy if you want to embed some testing or conditional logic into the pipeline to filter results in a complex way.)

stouset 2 days ago

Shell script authors like yourself make me very happy. The pipe-to-read is a fun idea, I’ll use it.

One stanza I have at the beginning of every script:

    [[ -n “${TRACE:-}” ]] && set -o xtrace
This lets you trace any script just by setting the environment variable. And it’s nounset-safe.

This was typed from memory on mobile so if the above is bugged, my bad :)

  • pxc a day ago

    > Shell script authors like yourself make me very happy.

    Shell script is really good for some things, so we're gonna end up writing it. And if we write it, it might as well be legible, right? Shell scripts deserve the same care that other programs do.

    > This lets you trace any script just by setting the environment variable. And it’s nounset-safe.

    Nice! I think I'll start adding the same.

    > This was typed from memory on mobile so if the above is bugged, my bad :)

    I think it's fine aside from the smart quote characters your phone inserted instead of plain vertical double quotes around the parameter expansion you use to make the line nounset-friendly!

  • ramses0 2 days ago

    Same w/ the "pipe to read" (although it doesn't seem to work right with OSX's bash-3.2). I found this gizmo somewhere and it's worth sharing as well...

        # MAGIC DEBUGGING LINE
        #trap '(read -p "[$BASH_SOURCE:$LINENO] $BASH_COMMAND? ")' DEBUG
    
    ...basically interactive prompting while running.

    I need to write up some thoughts on bash being effectively a lisp if you stare at it the right way.

    • pxc a day ago

      > Same w/ the "pipe to read" (although it doesn't seem to work right with OSX's bash-3.2). I found this gizmo somewhere and it's worth sharing as well...

      Yes, you need a recent Bash for this (and lots of other nice things like 'associative arrays' (maps)). To free myself up to write more concise and legible scripts, I lean into bashisms and make use of external programs without restriction, regardless of what's shipped by default on any system. To recover portability where I need it, I use Nix to fix my Bash interpreter versions as well as the versions of external tools.

      When my scripts are 'packaged' by Nix, their bodies are written as multiline strings in Nix. In that case, Nix takes care of setting the shebang to a fixed version of Bash, and I interpolate external commands in like this:

        "${pkgs.coreutils}/bin/realpath" .      # these do
        "${pkgs.coreutils}/bin/pwd" --physical  # the same thing
      
      and in that way my scripts use only their fixed dependencies at fixed versions without modifying the environment at all. This also works nicely across platforms, so when these scripts run on macOS they get the same versions of the GNU coreutils as they do when they run on Linux, just compiled for the appropriate platform and architecture. Same goes for different architectures. So this way your script runs 'natively' (no virtualization) but you still pin all its dependencies.

      In other contexts (e.g., cloud-init), I use bash a bit more restrictively depending on what versions I'm targeting. But I still use Nix to provide dependencies so that my scripts use the same versions of external tools regardless of distro or platform:

        nix shell nixpkgs#toybox --command which which # these do
        nix run nixpkgs#which -- which                 # the same thing
      
      `nix run` and `nix shell` both behave as desired in pipelines and subshells and all that. (To get the same level of determinism as with the method of use outlined earlier, you'd want to either pin the `nixpkgs` ref in your local flake registry or replace it with a reference that pinned the ref down to a commit hash.)

      There are is a really cool tool[1] by the Nixer abathur for automagically converting naively-written scripts to ones that pin their deps via Nix, as in the first example. I'm not using it yet but I likely will if the scripts I use for our development environments at work get much bigger-- that way I can store them as normal scripts without any special escaping/templating and linters will know how to read them and all that.

      Anyhow, it's totally safe to install a newer Bash on macOS, and I recommend doing it for personal interactive use and private scripts. Pkgsrc and MacPorts both have Bash 5.x, if you don't have or want Nix.

      > I need to write up some thoughts on bash being effectively a lisp if you stare at it the right way.

      You should! It's totally true, since external commands and shell builtins are in a prefix syntax just like Lisp function calls. Some shells accentuate this a bit, like Fish, where subshells are just parens with no dollar signs so nested subcommands look very Lisp-y, and Elvish, whose author (xiaq) has tried to lean into that syntactic convergence designing a new shell (drawing inspiration from Scheme in various places).

      > I found this gizmo somewhere and it's worth sharing as well...

          # MAGIC DEBUGGING LINE
          #trap '(read -p "[$BASH_SOURCE:$LINENO] $BASH_COMMAND? ")' DEBUG
      
      Okay that looks really nifty. I will definitely find a use for that literally tomorrow, if not today.

      --

      1: https://github.com/abathur/resholve