sites

public wiki contents of suckless.org
git clone git://git.suckless.org/sites
Log | Files | Refs

commit 90ac09c942c36b7bf8c508b37be649afea2e9090
parent a36f1c294c2677674520216a17236beaceb89de7
Author: Santtu Lakkala <inz@inz.fi>
Date:   Mon, 21 Feb 2022 16:19:15 +0200

[st][patch][copyurl] Slight update and multiline version

The previous version of my copyurl patch worked incorrectly due to
relying on the snapping selection; use double selextend() instead.

Also a new version that uses a (reverse) DFA to detect links without
copying to buffer, enabling the detection of multiple line urls.

Diffstat:
Mst.suckless.org/patches/copyurl/index.md | 8+++++++-
Dst.suckless.org/patches/copyurl/st-copyurl-20220217-0.8.5.diff | 136-------------------------------------------------------------------------------
Ast.suckless.org/patches/copyurl/st-copyurl-20220221-0.8.5.diff | 137+++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
Ast.suckless.org/patches/copyurl/st-copyurl-multiline-20220221-0.8.5.diff | 156+++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
4 files changed, 300 insertions(+), 137 deletions(-)

diff --git a/st.suckless.org/patches/copyurl/index.md b/st.suckless.org/patches/copyurl/index.md @@ -22,7 +22,12 @@ Following patches also highlight the selected urls: * [st-copyurl-20190202-3be4cf1.diff](st-copyurl-20190202-3be4cf1.diff) * [st-copyurl-0.8.4.diff](st-copyurl-0.8.4.diff) -* [st-copyurl-20220217-0.8.5.diff](st-copyurl-20220217-0.8.5.diff) +* [st-copyurl-20220221-0.8.5.diff](st-copyurl-20220221-0.8.5.diff) + +Following patches also allow urls that span multiple lines: + +* [st-copyurl-multiline-20220221-0.8.5.diff](st-copyurl-multiline-20220221-0.8.5.diff) + Authors ------- @@ -31,3 +36,4 @@ Authors * David Phillips - <david@sighup.nz> (git port) * Matthew Parnell - <matt@parnmatt.co.uk> (0.7, git ports) * Michael Buch - <michaelbuch12@gmail.com> (0.8+ git port, url highlighting) +* Santtu Lakkala - <inz@inz.fi> (multiline url support) diff --git a/st.suckless.org/patches/copyurl/st-copyurl-20220217-0.8.5.diff b/st.suckless.org/patches/copyurl/st-copyurl-20220217-0.8.5.diff @@ -1,136 +0,0 @@ -From eacbcf7de84e81a27759d086ad1a5ea9e41f9894 Mon Sep 17 00:00:00 2001 -From: Santtu Lakkala <inz@inz.fi> -Date: Wed, 16 Feb 2022 20:34:20 +0200 -Subject: [PATCH] Loop through urls on screen and copy to clipboard - -Based on the previous highlighting patches, slightly simplified and -fixes graphical issues with mixed copyurl and selection. ---- - config.def.h | 1 + - st.c | 81 ++++++++++++++++++++++++++++++++++++++++++++++++++++ - st.h | 1 + - 3 files changed, 83 insertions(+) - -diff --git a/config.def.h b/config.def.h -index 91ab8ca..3f365c7 100644 ---- a/config.def.h -+++ b/config.def.h -@@ -201,6 +201,7 @@ static Shortcut shortcuts[] = { - { TERMMOD, XK_Y, selpaste, {.i = 0} }, - { ShiftMask, XK_Insert, selpaste, {.i = 0} }, - { TERMMOD, XK_Num_Lock, numlock, {.i = 0} }, -+ { MODKEY, XK_l, copyurl, {.i = 0} }, - }; - - /* -diff --git a/st.c b/st.c -index 51049ba..931b29f 100644 ---- a/st.c -+++ b/st.c -@@ -200,6 +200,7 @@ static void tdefutf8(char); - static int32_t tdefcolor(const int *, int *, int); - static void tdeftran(char); - static void tstrsequence(uchar); -+static const char *findlastany(const char *, const char**, size_t); - - static void drawregion(int, int, int, int); - -@@ -2688,3 +2689,83 @@ redraw(void) - tfulldirt(); - draw(); - } -+ -+const char * -+findlastany(const char *str, const char**find, size_t len) -+{ -+ const char *found = NULL; -+ int i = 0; -+ -+ for (found = str + strlen(str) - 1; found >= str; --found) { -+ for(i = 0; i < len; i++) { -+ if (strncmp(found, find[i], strlen(find[i])) == 0) { -+ return found; -+ } -+ } -+ } -+ -+ return NULL; -+} -+ -+/* -+** Select and copy the previous url on screen (do nothing if there's no url). -+** -+** FIXME: doesn't handle urls that span multiple lines; will need to add support -+** for multiline "getsel()" first -+*/ -+void -+copyurl(const Arg *arg) { -+ /* () and [] can appear in urls, but excluding them here will reduce false -+ * positives when figuring out where a given url ends. -+ */ -+ static const char URLCHARS[] = "ABCDEFGHIJKLMNOPQRSTUVWXYZ" -+ "abcdefghijklmnopqrstuvwxyz" -+ "0123456789-._~:/?#@!$&'*+,;=%"; -+ -+ static const char* URLSTRINGS[] = {"http://", "https://"}; -+ -+ int row = 0, /* row of current URL */ -+ col = 0, /* column of current URL start */ -+ colend = 0, /* column of last occurrence */ -+ passes = 0; /* how many rows have been scanned */ -+ -+ char linestr[term.col + 1]; -+ const char *c = NULL, -+ *match = NULL; -+ -+ row = (sel.ob.x >= 0 && sel.nb.y > 0) ? sel.nb.y : term.bot; -+ LIMIT(row, term.top, term.bot); -+ -+ colend = (sel.ob.x >= 0 && sel.nb.y > 0) ? sel.nb.x : term.col; -+ LIMIT(colend, 0, term.col); -+ -+ /* -+ ** Scan from (term.row - 1,term.col - 1) to (0,0) and find -+ ** next occurrance of a URL -+ */ -+ for (passes = 0; passes < term.row; passes++) { -+ /* Read in each column of every row until -+ ** we hit previous occurrence of URL -+ */ -+ for (col = 0; col < colend; ++col) -+ linestr[col] = term.line[row][col].u < 128 ? term.line[row][col].u : ' '; -+ linestr[col] = '\0'; -+ -+ if ((match = findlastany(linestr, URLSTRINGS, -+ sizeof(URLSTRINGS)/sizeof(URLSTRINGS[0])))) -+ break; -+ -+ if (--row < 0) -+ row = term.row - 1; -+ -+ colend = term.col; -+ }; -+ -+ if (match) { -+ size_t l = strspn(match, URLCHARS); -+ selstart(match - linestr, row, 1); -+ selextend(match - linestr + l - 1, row, 1, SEL_REGULAR); -+ xsetsel(getsel()); -+ xclipcopy(); -+ } -+} -diff --git a/st.h b/st.h -index 519b9bd..0458005 100644 ---- a/st.h -+++ b/st.h -@@ -85,6 +85,7 @@ void printscreen(const Arg *); - void printsel(const Arg *); - void sendbreak(const Arg *); - void toggleprinter(const Arg *); -+void copyurl(const Arg *); - - int tattrset(int); - void tnew(int, int); --- -2.32.0 - diff --git a/st.suckless.org/patches/copyurl/st-copyurl-20220221-0.8.5.diff b/st.suckless.org/patches/copyurl/st-copyurl-20220221-0.8.5.diff @@ -0,0 +1,137 @@ +From 897c3958d01d8df80ebf1666b972b8a658b419ba Mon Sep 17 00:00:00 2001 +From: Santtu Lakkala <inz@inz.fi> +Date: Wed, 16 Feb 2022 20:34:20 +0200 +Subject: [PATCH] Loop through urls on screen and copy to clipboard + +Based on the previous highlighting patches, slightly simplified and +fixes graphical issues with mixed copyurl and selection. +--- + config.def.h | 1 + + st.c | 82 ++++++++++++++++++++++++++++++++++++++++++++++++++++ + st.h | 1 + + 3 files changed, 84 insertions(+) + +diff --git a/config.def.h b/config.def.h +index 91ab8ca..3f365c7 100644 +--- a/config.def.h ++++ b/config.def.h +@@ -201,6 +201,7 @@ static Shortcut shortcuts[] = { + { TERMMOD, XK_Y, selpaste, {.i = 0} }, + { ShiftMask, XK_Insert, selpaste, {.i = 0} }, + { TERMMOD, XK_Num_Lock, numlock, {.i = 0} }, ++ { MODKEY, XK_l, copyurl, {.i = 0} }, + }; + + /* +diff --git a/st.c b/st.c +index 51049ba..bf3d81a 100644 +--- a/st.c ++++ b/st.c +@@ -200,6 +200,7 @@ static void tdefutf8(char); + static int32_t tdefcolor(const int *, int *, int); + static void tdeftran(char); + static void tstrsequence(uchar); ++static const char *findlastany(const char *, const char**, size_t); + + static void drawregion(int, int, int, int); + +@@ -2688,3 +2689,84 @@ redraw(void) + tfulldirt(); + draw(); + } ++ ++const char * ++findlastany(const char *str, const char**find, size_t len) ++{ ++ const char *found = NULL; ++ int i = 0; ++ ++ for (found = str + strlen(str) - 1; found >= str; --found) { ++ for(i = 0; i < len; i++) { ++ if (strncmp(found, find[i], strlen(find[i])) == 0) { ++ return found; ++ } ++ } ++ } ++ ++ return NULL; ++} ++ ++/* ++** Select and copy the previous url on screen (do nothing if there's no url). ++** ++** FIXME: doesn't handle urls that span multiple lines; will need to add support ++** for multiline "getsel()" first ++*/ ++void ++copyurl(const Arg *arg) { ++ /* () and [] can appear in urls, but excluding them here will reduce false ++ * positives when figuring out where a given url ends. ++ */ ++ static const char URLCHARS[] = "ABCDEFGHIJKLMNOPQRSTUVWXYZ" ++ "abcdefghijklmnopqrstuvwxyz" ++ "0123456789-._~:/?#@!$&'*+,;=%"; ++ ++ static const char* URLSTRINGS[] = {"http://", "https://"}; ++ ++ int row = 0, /* row of current URL */ ++ col = 0, /* column of current URL start */ ++ colend = 0, /* column of last occurrence */ ++ passes = 0; /* how many rows have been scanned */ ++ ++ char linestr[term.col + 1]; ++ const char *c = NULL, ++ *match = NULL; ++ ++ row = (sel.ob.x >= 0 && sel.nb.y > 0) ? sel.nb.y : term.bot; ++ LIMIT(row, term.top, term.bot); ++ ++ colend = (sel.ob.x >= 0 && sel.nb.y > 0) ? sel.nb.x : term.col; ++ LIMIT(colend, 0, term.col); ++ ++ /* ++ ** Scan from (term.row - 1,term.col - 1) to (0,0) and find ++ ** next occurrance of a URL ++ */ ++ for (passes = 0; passes < term.row; passes++) { ++ /* Read in each column of every row until ++ ** we hit previous occurrence of URL ++ */ ++ for (col = 0; col < colend; ++col) ++ linestr[col] = term.line[row][col].u < 128 ? term.line[row][col].u : ' '; ++ linestr[col] = '\0'; ++ ++ if ((match = findlastany(linestr, URLSTRINGS, ++ sizeof(URLSTRINGS)/sizeof(URLSTRINGS[0])))) ++ break; ++ ++ if (--row < 0) ++ row = term.row - 1; ++ ++ colend = term.col; ++ }; ++ ++ if (match) { ++ size_t l = strspn(match, URLCHARS); ++ selstart(match - linestr, row, 0); ++ selextend(match - linestr + l - 1, row, SEL_REGULAR, 0); ++ selextend(match - linestr + l - 1, row, SEL_REGULAR, 1); ++ xsetsel(getsel()); ++ xclipcopy(); ++ } ++} +diff --git a/st.h b/st.h +index 519b9bd..0458005 100644 +--- a/st.h ++++ b/st.h +@@ -85,6 +85,7 @@ void printscreen(const Arg *); + void printsel(const Arg *); + void sendbreak(const Arg *); + void toggleprinter(const Arg *); ++void copyurl(const Arg *); + + int tattrset(int); + void tnew(int, int); +-- +2.32.0 + diff --git a/st.suckless.org/patches/copyurl/st-copyurl-multiline-20220221-0.8.5.diff b/st.suckless.org/patches/copyurl/st-copyurl-multiline-20220221-0.8.5.diff @@ -0,0 +1,156 @@ +From 30a04d9ecb3998953bdbe42e5617d00d6002869b Mon Sep 17 00:00:00 2001 +From: Santtu Lakkala <inz@inz.fi> +Date: Wed, 16 Feb 2022 20:34:20 +0200 +Subject: [PATCH] Loop through urls on screen and copy to clipboard + +Replace url detection heuristics with a DFA, enabling urls that span +multiple lines. Also fix the selection not to use snapping so that urls +are selected exactly. +--- + config.def.h | 1 + + st.c | 93 ++++++++++++++++++++++++++++++++++++++++++++++++++++ + st.h | 1 + + 3 files changed, 95 insertions(+) + +diff --git a/config.def.h b/config.def.h +index 91ab8ca..3f365c7 100644 +--- a/config.def.h ++++ b/config.def.h +@@ -201,6 +201,7 @@ static Shortcut shortcuts[] = { + { TERMMOD, XK_Y, selpaste, {.i = 0} }, + { ShiftMask, XK_Insert, selpaste, {.i = 0} }, + { TERMMOD, XK_Num_Lock, numlock, {.i = 0} }, ++ { MODKEY, XK_l, copyurl, {.i = 0} }, + }; + + /* +diff --git a/st.c b/st.c +index 51049ba..5b6d919 100644 +--- a/st.c ++++ b/st.c +@@ -152,6 +152,11 @@ typedef struct { + int narg; /* nb of args */ + } STREscape; + ++typedef struct { ++ int state; ++ size_t length; ++} URLdfa; ++ + static void execsh(char *, char **); + static void stty(char **); + static void sigchld(int); +@@ -200,6 +205,7 @@ static void tdefutf8(char); + static int32_t tdefcolor(const int *, int *, int); + static void tdeftran(char); + static void tstrsequence(uchar); ++static int daddch(URLdfa *, char); + + static void drawregion(int, int, int, int); + +@@ -2688,3 +2694,90 @@ redraw(void) + tfulldirt(); + draw(); + } ++ ++int ++daddch(URLdfa *dfa, char c) ++{ ++ /* () and [] can appear in urls, but excluding them here will reduce false ++ * positives when figuring out where a given url ends. ++ */ ++ static const char URLCHARS[] = "ABCDEFGHIJKLMNOPQRSTUVWXYZ" ++ "abcdefghijklmnopqrstuvwxyz" ++ "0123456789-._~:/?#@!$&'*+,;=%"; ++ static const char RPFX[] = "//:sptth"; ++ ++ if (!strchr(URLCHARS, c)) { ++ dfa->length = 0; ++ dfa->state = 0; ++ ++ return 0; ++ } ++ ++ dfa->length++; ++ ++ if (dfa->state == 2 && c == '/') { ++ dfa->state = 0; ++ } else if (dfa->state == 3 && c == 'p') { ++ dfa->state++; ++ } else if (c != RPFX[dfa->state]) { ++ dfa->state = 0; ++ return 0; ++ } ++ ++ if (dfa->state++ == 7) { ++ dfa->state = 0; ++ return 1; ++ } ++ ++ return 0; ++} ++ ++/* ++** Select and copy the previous url on screen (do nothing if there's no url). ++*/ ++void ++copyurl(const Arg *arg) { ++ int row = 0, /* row of current URL */ ++ col = 0, /* column of current URL start */ ++ colend = 0, /* column of last occurrence */ ++ passes = 0; /* how many rows have been scanned */ ++ ++ const char *c = NULL, ++ *match = NULL; ++ URLdfa dfa = { 0 }; ++ ++ row = (sel.ob.x >= 0 && sel.nb.y > 0) ? sel.nb.y : term.bot; ++ LIMIT(row, term.top, term.bot); ++ ++ colend = (sel.ob.x >= 0 && sel.nb.y > 0) ? sel.nb.x : term.col; ++ LIMIT(colend, 0, term.col); ++ ++ /* ++ ** Scan from (term.row - 1,term.col - 1) to (0,0) and find ++ ** next occurrance of a URL ++ */ ++ for (passes = 0; passes < term.row; passes++) { ++ /* Read in each column of every row until ++ ** we hit previous occurrence of URL ++ */ ++ for (col = colend; col--;) ++ if (daddch(&dfa, term.line[row][col].u < 128 ? term.line[row][col].u : ' ')) ++ break; ++ ++ if (col >= 0) ++ break; ++ ++ if (--row < 0) ++ row = term.row - 1; ++ ++ colend = term.col; ++ } ++ ++ if (passes < term.row) { ++ selstart(col, row, 0); ++ selextend((col + dfa.length - 1) % term.col, row + (col + dfa.length - 1) / term.col, SEL_REGULAR, 0); ++ selextend((col + dfa.length - 1) % term.col, row + (col + dfa.length - 1) / term.col, SEL_REGULAR, 1); ++ xsetsel(getsel()); ++ xclipcopy(); ++ } ++} +diff --git a/st.h b/st.h +index 519b9bd..0458005 100644 +--- a/st.h ++++ b/st.h +@@ -85,6 +85,7 @@ void printscreen(const Arg *); + void printsel(const Arg *); + void sendbreak(const Arg *); + void toggleprinter(const Arg *); ++void copyurl(const Arg *); + + int tattrset(int); + void tnew(int, int); +-- +2.32.0 +