mirror of
https://github.com/LukeSmithxyz/st.git
synced 2024-11-24 16:05:47 +01:00
copy/open url use same regex. slimming.
This commit is contained in:
parent
7a7c5f8bfd
commit
dcaad2ceba
2 changed files with 8 additions and 27 deletions
6
config.h
6
config.h
|
@ -235,13 +235,7 @@ MouseKey mkeys[] = {
|
||||||
};
|
};
|
||||||
|
|
||||||
static char *openurlcmd[] = { "/bin/sh", "-c", "st-urlhandler -o", "externalpipe", NULL };
|
static char *openurlcmd[] = { "/bin/sh", "-c", "st-urlhandler -o", "externalpipe", NULL };
|
||||||
|
|
||||||
static char *copyurlcmd[] = { "/bin/sh", "-c", "st-urlhandler -c", "externalpipe", NULL };
|
static char *copyurlcmd[] = { "/bin/sh", "-c", "st-urlhandler -c", "externalpipe", NULL };
|
||||||
|
|
||||||
/* static char *copyurlcmd[] = { "/bin/sh", "-c", */
|
|
||||||
/* "tmp=$(sed 's/.*│//g' | tr -d '\n' | grep -aEo '(((http|https|gopher|gemini|ftp|ftps|git)://|www\\.)[a-zA-Z0-9.]*[:]?[a-zA-Z0-9./@$&%?$#=_-~]*)|((magnet:\\?xt=urn:btih:)[a-zA-Z0-9]*)' | uniq | sed 's/^www./http:\\/\\/www\\./g' ); IFS=; [ ! -z $tmp ] && echo $tmp | dmenu -i -p 'Copy which url?' -l 10 | tr -d '\n' | xclip -selection clipboard", */
|
|
||||||
/* "externalpipe", NULL }; */
|
|
||||||
|
|
||||||
static char *copyoutput[] = { "/bin/sh", "-c", "st-copyout", "externalpipe", NULL };
|
static char *copyoutput[] = { "/bin/sh", "-c", "st-copyout", "externalpipe", NULL };
|
||||||
|
|
||||||
static Shortcut shortcuts[] = {
|
static Shortcut shortcuts[] = {
|
||||||
|
|
|
@ -1,31 +1,18 @@
|
||||||
#!/bin/sh
|
#!/bin/sh
|
||||||
|
|
||||||
geturl() {
|
urlregex="(((http|https|gopher|gemini|ftp|ftps|git)://|www\\.)[a-zA-Z0-9.]*[:]?[a-zA-Z0-9./@$&%?$\#=_~-]*)|((magnet:\\?xt=urn:btih:)[a-zA-Z0-9]*)"
|
||||||
# First remove linebreaks and mutt sidebars:
|
|
||||||
urls="$(sed 's/.*│//g' | tr -d '\n' |
|
urls="$(sed 's/.*│//g' | tr -d '\n' | # First remove linebreaks and mutt sidebars:
|
||||||
grep -aEo "$urlregex" | # grep only urls as defined above.
|
grep -aEo "$urlregex" | # grep only urls as defined above.
|
||||||
uniq | # Ignore neighboring duplicates.
|
uniq | # Ignore neighboring duplicates.
|
||||||
sed 's/^www./http:\/\/www\./g')"
|
sed 's/^www./http:\/\/www\./g')" # xdg-open will not detect url without http://
|
||||||
|
|
||||||
[ -z "$urls" ] && exit
|
[ -z "$urls" ] && exit 1
|
||||||
}
|
|
||||||
|
|
||||||
openurl() {
|
|
||||||
urlregex="(((http|https)://|www\\.)[a-zA-Z0-9.]*[:]?[a-zA-Z0-9./@$&%?$\#=_~-]*)|((magnet:\\?xt=urn:btih:)[a-zA-Z0-9]*)"
|
|
||||||
geturl
|
|
||||||
chosen="$(echo "$urls" | dmenu -i -p 'Follow which url?' -l 10)"
|
|
||||||
setsid xdg-open "$chosen" >/dev/null 2>&1 &
|
|
||||||
}
|
|
||||||
|
|
||||||
copyurl() {
|
|
||||||
urlregex="(((http|https|gopher|gemini|ftp|ftps|git)://|www\\.)[a-zA-Z0-9.]*[:]?[a-zA-Z0-9./@$&%?$\#=_~-]*)|((magnet:\\?xt=urn:btih:)[a-zA-Z0-9]*)"
|
|
||||||
geturl
|
|
||||||
echo $urls | dmenu -i -p 'Copy which url?' -l 10 | tr -d '\n' | xclip -selection clipboard
|
|
||||||
}
|
|
||||||
|
|
||||||
while getopts "hoc" o; do case "${o}" in
|
while getopts "hoc" o; do case "${o}" in
|
||||||
h) printf "Optional arguments for custom use:\\n -c: copy\\n -o: xdg-open\\n -h: Show this message\\n" && exit 1 ;;
|
h) printf "Optional arguments for custom use:\\n -c: copy\\n -o: xdg-open\\n -h: Show this message\\n" && exit 1 ;;
|
||||||
o) openurl ;;
|
o) chosen="$(echo "$urls" | dmenu -i -p 'Follow which url?' -l 10)"
|
||||||
c) copyurl ;;
|
setsid xdg-open "$chosen" >/dev/null 2>&1 & ;;
|
||||||
|
c) echo "$urls" | dmenu -i -p 'Copy which url?' -l 10 | tr -d '\n' | xclip -selection clipboard ;;
|
||||||
*) printf "Invalid option: -%s\\n" "$OPTARG" && exit 1 ;;
|
*) printf "Invalid option: -%s\\n" "$OPTARG" && exit 1 ;;
|
||||||
esac done
|
esac done
|
||||||
|
|
Loading…
Reference in a new issue