sites

public wiki contents of suckless.org
git clone git://git.suckless.org/sites
Log | Files | Refs

commit 1ac951c28e68c02b808204d193e8ce64a019e571
parent 928f5bf043dcd956e44877a04ec6c153b8464190
Author: Michael Buch <michaelbuch12@gmail.com>
Date:   Sat,  2 Feb 2019 14:28:14 +0000

fix broken git refs for copyurl patches

Diffstat:
Mst.suckless.org/patches/copyurl/index.md | 8+++-----
Dst.suckless.org/patches/copyurl/st-copyurl-20180418-7e810b5.diff | 109-------------------------------------------------------------------------------
Dst.suckless.org/patches/copyurl/st-copyurl-20180514-a7bd977.diff | 143-------------------------------------------------------------------------------
Dst.suckless.org/patches/copyurl/st-copyurl-20180520-73f375a.diff | 185-------------------------------------------------------------------------------
Ast.suckless.org/patches/copyurl/st-copyurl-20190202-0.8.1.diff | 109+++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
Ast.suckless.org/patches/copyurl/st-copyurl-20190202-3be4cf1.diff | 177+++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
6 files changed, 289 insertions(+), 442 deletions(-)

diff --git a/st.suckless.org/patches/copyurl/index.md b/st.suckless.org/patches/copyurl/index.md @@ -8,8 +8,7 @@ Multiple invocations cycle through the available URLs. Notes ----- -URLs spanning multiple lines are not handled and only the first -URL on each line is selected. +* URLs spanning multiple lines are not handled Download -------- @@ -17,12 +16,11 @@ Download * [st-copyurl-0.7.diff](st-copyurl-0.7.diff) * [st-copyurl-20161105-8c99915.diff](st-copyurl-20161105-8c99915.diff) * [st-copyurl-20170802-e2ee5ee.diff](st-copyurl-20170802-e2ee5ee.diff) -* [st-copyurl-20180418-7e810b5.diff](st-copyurl-20180418-7e810b5.diff) +* [st-copyurl-20190202-3be4cf1.diff](st-copyurl-20190202-3be4cf1.diff) Following patches also highlight the selected urls: -* [st-copyurl-20180514-a7bd977.diff](st-copyurl-20180514-a7bd977.diff) -* [st-copyurl-20180520-73f375a.diff](st-copyurl-20180520-73f375a.diff) +* [st-copyurl-20190202-0.8.1.diff](st-copyurl-20190202-0.8.1.diff) Authors ------- diff --git a/st.suckless.org/patches/copyurl/st-copyurl-20180418-7e810b5.diff b/st.suckless.org/patches/copyurl/st-copyurl-20180418-7e810b5.diff @@ -1,109 +0,0 @@ -From 7e810b5d11e16aa8b79e6057cc4b9fa55df1fc57 Mon Sep 17 00:00:00 2001 -From: Michael Buch <michaelbuch12@gmail.com> -Date: Wed, 18 Apr 2018 01:12:09 +0100 -Subject: [PATCH] Port the copyurl patch to the 0.8.1 st release. Mainly fix - usage of depracted selcopy - ---- - config.def.h | 1 + - st.c | 62 ++++++++++++++++++++++++++++++++++++++++++++++++++++ - st.h | 1 + - 3 files changed, 64 insertions(+) - -diff --git a/config.def.h b/config.def.h -index 82b1b09..cbe923e 100644 ---- a/config.def.h -+++ b/config.def.h -@@ -178,6 +178,7 @@ static Shortcut shortcuts[] = { - { TERMMOD, XK_Y, selpaste, {.i = 0} }, - { TERMMOD, XK_Num_Lock, numlock, {.i = 0} }, - { TERMMOD, XK_I, iso14755, {.i = 0} }, -+ { MODKEY, XK_l, copyurl, {.i = 0} }, - }; - - /* -diff --git a/st.c b/st.c -index 0628707..3df99d9 100644 ---- a/st.c -+++ b/st.c -@@ -2617,3 +2617,65 @@ redraw(void) - tfulldirt(); - draw(); - } -+ -+/* select and copy the previous url on screen (do nothing if there's no url). -+ * known bug: doesn't handle urls that span multiple lines (wontfix) -+ * known bug: only finds first url on line (mightfix) -+ */ -+void -+copyurl(const Arg *arg) { -+ /* () and [] can appear in urls, but excluding them here will reduce false -+ * positives when figuring out where a given url ends. -+ */ -+ static char URLCHARS[] = "ABCDEFGHIJKLMNOPQRSTUVWXYZ" -+ "abcdefghijklmnopqrstuvwxyz" -+ "0123456789-._~:/?#@!$&'*+,;=%"; -+ -+ int i, row, startrow; -+ char *linestr = calloc(sizeof(char), term.col+1); /* assume ascii */ -+ char *c, *match = NULL; -+ -+ row = (sel.ob.x >= 0 && sel.nb.y > 0) ? sel.nb.y-1 : term.bot; -+ LIMIT(row, term.top, term.bot); -+ startrow = row; -+ -+ /* find the start of the last url before selection */ -+ do { -+ for (i = 0; i < term.col; ++i) { -+ if (term.line[row][i].u > 127) /* assume ascii */ -+ continue; -+ linestr[i] = term.line[row][i].u; -+ } -+ linestr[term.col] = '\0'; -+ if ((match = strstr(linestr, "http://")) -+ || (match = strstr(linestr, "https://"))) -+ break; -+ if (--row < term.top) -+ row = term.bot; -+ } while (row != startrow); -+ -+ if (match) { -+ /* must happen before trim */ -+ selclear(); -+ sel.ob.x = strlen(linestr) - strlen(match); -+ -+ /* trim the rest of the line from the url match */ -+ for (c = match; *c != '\0'; ++c) -+ if (!strchr(URLCHARS, *c)) { -+ *c = '\0'; -+ break; -+ } -+ -+ /* select and copy */ -+ sel.mode = 1; -+ sel.type = SEL_REGULAR; -+ sel.oe.x = sel.ob.x + strlen(match)-1; -+ sel.ob.y = sel.oe.y = row; -+ selnormalize(); -+ tsetdirt(sel.nb.y, sel.ne.y); -+ xsetsel(getsel()); -+ xclipcopy(); -+ } -+ -+ free(linestr); -+} -diff --git a/st.h b/st.h -index dac64d8..5a58f8f 100644 ---- a/st.h -+++ b/st.h -@@ -85,6 +85,7 @@ void printscreen(const Arg *); - void printsel(const Arg *); - void sendbreak(const Arg *); - void toggleprinter(const Arg *); -+void copyurl(const Arg *); - - int tattrset(int); - void tnew(int, int); --- -2.17.0 - diff --git a/st.suckless.org/patches/copyurl/st-copyurl-20180514-a7bd977.diff b/st.suckless.org/patches/copyurl/st-copyurl-20180514-a7bd977.diff @@ -1,143 +0,0 @@ -From a7bd977e7f53d53223dc0a844406ad706d5fb803 Mon Sep 17 00:00:00 2001 -From: Michael Buch <michaelbuch12@gmail.com> -Date: Mon, 14 May 2018 23:27:09 +0100 -Subject: [PATCH] [st][patch][copyurl] Prevent url highlighting from occasional - crash after resizing terminal. Also fix occasional partial highlighting. Will - replace the previous patch on the wiki site - ---- - config.def.h | 1 + - st.c | 81 ++++++++++++++++++++++++++++++++++++++++++++++++++++ - st.h | 1 + - 3 files changed, 83 insertions(+) - -diff --git a/config.def.h b/config.def.h -index 82b1b09..cbe923e 100644 ---- a/config.def.h -+++ b/config.def.h -@@ -178,6 +178,7 @@ static Shortcut shortcuts[] = { - { TERMMOD, XK_Y, selpaste, {.i = 0} }, - { TERMMOD, XK_Num_Lock, numlock, {.i = 0} }, - { TERMMOD, XK_I, iso14755, {.i = 0} }, -+ { MODKEY, XK_l, copyurl, {.i = 0} }, - }; - - /* -diff --git a/st.c b/st.c -index 0628707..afe3d9a 100644 ---- a/st.c -+++ b/st.c -@@ -204,6 +204,7 @@ static void tdefutf8(char); - static int32_t tdefcolor(int *, int *, int); - static void tdeftran(char); - static void tstrsequence(uchar); -+static void tsetcolor(int, int, int, uint32_t, uint32_t); - - static void drawregion(int, int, int, int); - -@@ -1600,6 +1601,17 @@ tsetmode(int priv, int set, int *args, int narg) - } - } - -+void -+tsetcolor( int row, int start, int end, uint32_t fg, uint32_t bg ) -+{ -+ int i = start; -+ for( ; i < end; ++i ) -+ { -+ term.line[row][i].fg = fg; -+ term.line[row][i].bg = bg; -+ } -+} -+ - void - csihandle(void) - { -@@ -2617,3 +2629,72 @@ redraw(void) - tfulldirt(); - draw(); - } -+ -+/* select and copy the previous url on screen (do nothing if there's no url). -+ * known bug: doesn't handle urls that span multiple lines (wontfix) -+ * known bug: only finds first url on line (mightfix) -+ */ -+void -+copyurl(const Arg *arg) { -+ /* () and [] can appear in urls, but excluding them here will reduce false -+ * positives when figuring out where a given url ends. -+ */ -+ static char URLCHARS[] = "ABCDEFGHIJKLMNOPQRSTUVWXYZ" -+ "abcdefghijklmnopqrstuvwxyz" -+ "0123456789-._~:/?#@!$&'*+,;=%"; -+ -+ // remove highlighting from previous selection if any -+ if(sel.ob.x >= 0 && sel.oe.x >= 0) -+ tsetcolor(sel.nb.y, sel.ob.x, sel.oe.x + 1, defaultfg, defaultbg); -+ -+ int i, row, startrow; -+ char *linestr = calloc(sizeof(char), term.col+1); /* assume ascii */ -+ char *c, *match = NULL; -+ -+ row = (sel.ob.x >= 0 && sel.nb.y > 0) ? sel.nb.y-1 : term.bot; -+ LIMIT(row, term.top, term.bot); -+ startrow = row; -+ -+ /* find the start of the last url before selection */ -+ do { -+ for (i = 0; i < term.col; ++i) { -+ if (term.line[row][i].u > 127) /* assume ascii */ -+ continue; -+ linestr[i] = term.line[row][i].u; -+ } -+ linestr[term.col] = '\0'; -+ if ((match = strstr(linestr, "http://")) -+ || (match = strstr(linestr, "https://"))) -+ break; -+ if (--row < term.top) -+ row = term.bot; -+ } while (row != startrow); -+ -+ if (match) { -+ /* must happen before trim */ -+ selclear(); -+ sel.ob.x = strlen(linestr) - strlen(match); -+ -+ /* trim the rest of the line from the url match */ -+ for (c = match; *c != '\0'; ++c) -+ if (!strchr(URLCHARS, *c)) { -+ *c = '\0'; -+ break; -+ } -+ -+ // highlight selection by inverting terminal colors -+ tsetcolor(row, sel.ob.x, sel.ob.x + strlen( match ), defaultbg, defaultfg); -+ -+ /* select and copy */ -+ sel.mode = 1; -+ sel.type = SEL_REGULAR; -+ sel.oe.x = sel.ob.x + strlen(match)-1; -+ sel.ob.y = sel.oe.y = row; -+ selnormalize(); -+ tsetdirt(sel.nb.y, sel.ne.y); -+ xsetsel(getsel()); -+ xclipcopy(); -+ } -+ -+ free(linestr); -+} -diff --git a/st.h b/st.h -index dac64d8..5a58f8f 100644 ---- a/st.h -+++ b/st.h -@@ -85,6 +85,7 @@ void printscreen(const Arg *); - void printsel(const Arg *); - void sendbreak(const Arg *); - void toggleprinter(const Arg *); -+void copyurl(const Arg *); - - int tattrset(int); - void tnew(int, int); --- -2.17.0 - diff --git a/st.suckless.org/patches/copyurl/st-copyurl-20180520-73f375a.diff b/st.suckless.org/patches/copyurl/st-copyurl-20180520-73f375a.diff @@ -1,185 +0,0 @@ -From 73f375a7c3256d32c62bdb7d616207071e89f982 Mon Sep 17 00:00:00 2001 -From: Michael Buch <michaelbuch12@gmail.com> -Date: Sun, 20 May 2018 16:14:14 +0100 -Subject: [PATCH] [copuryl patch] Cleanup URL determination logic - -Separate out into separate function and provide -ability to add more URLs to the list. ---- - config.def.h | 1 + - st.c | 123 +++++++++++++++++++++++++++++++++++++++++++++++++++ - st.h | 1 + - 3 files changed, 125 insertions(+) - -diff --git a/config.def.h b/config.def.h -index 82b1b09..cbe923e 100644 ---- a/config.def.h -+++ b/config.def.h -@@ -178,6 +178,7 @@ static Shortcut shortcuts[] = { - { TERMMOD, XK_Y, selpaste, {.i = 0} }, - { TERMMOD, XK_Num_Lock, numlock, {.i = 0} }, - { TERMMOD, XK_I, iso14755, {.i = 0} }, -+ { MODKEY, XK_l, copyurl, {.i = 0} }, - }; - - /* -diff --git a/st.c b/st.c -index 0628707..b08f454 100644 ---- a/st.c -+++ b/st.c -@@ -204,6 +204,7 @@ static void tdefutf8(char); - static int32_t tdefcolor(int *, int *, int); - static void tdeftran(char); - static void tstrsequence(uchar); -+static void tsetcolor(int, int, int, uint32_t, uint32_t); - - static void drawregion(int, int, int, int); - -@@ -1600,6 +1601,17 @@ tsetmode(int priv, int set, int *args, int narg) - } - } - -+void -+tsetcolor( int row, int start, int end, uint32_t fg, uint32_t bg ) -+{ -+ int i = start; -+ for( ; i < end; ++i ) -+ { -+ term.line[row][i].fg = fg; -+ term.line[row][i].bg = bg; -+ } -+} -+ - void - csihandle(void) - { -@@ -2617,3 +2629,114 @@ redraw(void) - tfulldirt(); - draw(); - } -+ -+char * -+findlastany(char *str, const char** find, size_t len) -+{ -+ char* found = NULL; -+ int i = 0; -+ for(found = str + strlen(str) - 1; found >= str; --found) { -+ for(i = 0; i < len; i++) { -+ if(strncmp(found, find[i], strlen(find[i])) == 0) { -+ return found; -+ } -+ } -+ } -+ -+ return NULL; -+} -+ -+/* -+** Select and copy the previous url on screen (do nothing if there's no url). -+** -+** FIXME: doesn't handle urls that span multiple lines; will need to add support -+** for multiline "getsel()" first -+*/ -+void -+copyurl(const Arg *arg) { -+ /* () and [] can appear in urls, but excluding them here will reduce false -+ * positives when figuring out where a given url ends. -+ */ -+ static char URLCHARS[] = "ABCDEFGHIJKLMNOPQRSTUVWXYZ" -+ "abcdefghijklmnopqrstuvwxyz" -+ "0123456789-._~:/?#@!$&'*+,;=%"; -+ -+ static const char* URLSTRINGS[] = {"http://", "https://"}; -+ -+ /* remove highlighting from previous selection if any */ -+ if(sel.ob.x >= 0 && sel.oe.x >= 0) -+ tsetcolor(sel.nb.y, sel.ob.x, sel.oe.x + 1, defaultfg, defaultbg); -+ -+ int i = 0, -+ row = 0, /* row of current URL */ -+ col = 0, /* column of current URL start */ -+ startrow = 0, /* row of last occurrence */ -+ colend = 0, /* column of last occurrence */ -+ passes = 0; /* how many rows have been scanned */ -+ -+ char *linestr = calloc(sizeof(char), term.col+1); /* assume ascii */ -+ char *c = NULL, -+ *match = NULL; -+ -+ row = (sel.ob.x >= 0 && sel.nb.y > 0) ? sel.nb.y : term.bot; -+ LIMIT(row, term.top, term.bot); -+ startrow = row; -+ -+ colend = (sel.ob.x >= 0 && sel.nb.y > 0) ? sel.nb.x : term.col; -+ LIMIT(colend, 0, term.col); -+ -+ /* -+ ** Scan from (term.bot,term.col) to (0,0) and find -+ ** next occurrance of a URL -+ */ -+ while(passes !=term.bot + 2) { -+ /* Read in each column of every row until -+ ** we hit previous occurrence of URL -+ */ -+ for (col = 0, i = 0; col < colend; ++col,++i) { -+ /* assume ascii */ -+ if (term.line[row][col].u > 127) -+ continue; -+ linestr[i] = term.line[row][col].u; -+ } -+ linestr[term.col] = '\0'; -+ -+ if ((match = findlastany(linestr, URLSTRINGS, -+ sizeof(URLSTRINGS)/sizeof(URLSTRINGS[0])))) -+ break; -+ -+ if (--row < term.top) -+ row = term.bot; -+ -+ colend = term.col; -+ passes++; -+ }; -+ -+ if (match) { -+ /* must happen before trim */ -+ selclear(); -+ sel.ob.x = strlen(linestr) - strlen(match); -+ -+ /* trim the rest of the line from the url match */ -+ for (c = match; *c != '\0'; ++c) -+ if (!strchr(URLCHARS, *c)) { -+ *c = '\0'; -+ break; -+ } -+ -+ /* highlight selection by inverting terminal colors */ -+ tsetcolor(row, sel.ob.x, sel.ob.x + strlen( match ), defaultbg, defaultfg); -+ -+ /* select and copy */ -+ sel.mode = 1; -+ sel.type = SEL_REGULAR; -+ sel.oe.x = sel.ob.x + strlen(match)-1; -+ sel.ob.y = sel.oe.y = row; -+ selnormalize(); -+ tsetdirt(sel.nb.y, sel.ne.y); -+ xsetsel(getsel()); -+ xclipcopy(); -+ } -+ -+ free(linestr); -+} -diff --git a/st.h b/st.h -index dac64d8..5a58f8f 100644 ---- a/st.h -+++ b/st.h -@@ -85,6 +85,7 @@ void printscreen(const Arg *); - void printsel(const Arg *); - void sendbreak(const Arg *); - void toggleprinter(const Arg *); -+void copyurl(const Arg *); - - int tattrset(int); - void tnew(int, int); --- -2.17.0 - diff --git a/st.suckless.org/patches/copyurl/st-copyurl-20190202-0.8.1.diff b/st.suckless.org/patches/copyurl/st-copyurl-20190202-0.8.1.diff @@ -0,0 +1,109 @@ +From be408247f1c1ff8ccf7ab128b126f54d19bd4392 Mon Sep 17 00:00:00 2001 +From: Michael Buch <michaelbuch12@gmail.com> +Date: Sat, 2 Feb 2019 14:20:52 +0000 +Subject: [PATCH] Port the copyurl patch to the 0.8.1 st release. Mainly fix + usage of depracted selcopy + +--- + config.def.h | 1 + + st.c | 62 ++++++++++++++++++++++++++++++++++++++++++++++++++++ + st.h | 1 + + 3 files changed, 64 insertions(+) + +diff --git a/config.def.h b/config.def.h +index 82b1b09..cbe923e 100644 +--- a/config.def.h ++++ b/config.def.h +@@ -178,6 +178,7 @@ static Shortcut shortcuts[] = { + { TERMMOD, XK_Y, selpaste, {.i = 0} }, + { TERMMOD, XK_Num_Lock, numlock, {.i = 0} }, + { TERMMOD, XK_I, iso14755, {.i = 0} }, ++ { MODKEY, XK_l, copyurl, {.i = 0} }, + }; + + /* +diff --git a/st.c b/st.c +index 46c954b..476eb31 100644 +--- a/st.c ++++ b/st.c +@@ -2616,3 +2616,65 @@ redraw(void) + tfulldirt(); + draw(); + } ++ ++/* select and copy the previous url on screen (do nothing if there's no url). ++ * known bug: doesn't handle urls that span multiple lines (wontfix) ++ * known bug: only finds first url on line (mightfix) ++ */ ++void ++copyurl(const Arg *arg) { ++ /* () and [] can appear in urls, but excluding them here will reduce false ++ * positives when figuring out where a given url ends. ++ */ ++ static char URLCHARS[] = "ABCDEFGHIJKLMNOPQRSTUVWXYZ" ++ "abcdefghijklmnopqrstuvwxyz" ++ "0123456789-._~:/?#@!$&'*+,;=%"; ++ ++ int i, row, startrow; ++ char *linestr = calloc(sizeof(char), term.col+1); /* assume ascii */ ++ char *c, *match = NULL; ++ ++ row = (sel.ob.x >= 0 && sel.nb.y > 0) ? sel.nb.y-1 : term.bot; ++ LIMIT(row, term.top, term.bot); ++ startrow = row; ++ ++ /* find the start of the last url before selection */ ++ do { ++ for (i = 0; i < term.col; ++i) { ++ if (term.line[row][i].u > 127) /* assume ascii */ ++ continue; ++ linestr[i] = term.line[row][i].u; ++ } ++ linestr[term.col] = '\0'; ++ if ((match = strstr(linestr, "http://")) ++ || (match = strstr(linestr, "https://"))) ++ break; ++ if (--row < term.top) ++ row = term.bot; ++ } while (row != startrow); ++ ++ if (match) { ++ /* must happen before trim */ ++ selclear(); ++ sel.ob.x = strlen(linestr) - strlen(match); ++ ++ /* trim the rest of the line from the url match */ ++ for (c = match; *c != '\0'; ++c) ++ if (!strchr(URLCHARS, *c)) { ++ *c = '\0'; ++ break; ++ } ++ ++ /* select and copy */ ++ sel.mode = 1; ++ sel.type = SEL_REGULAR; ++ sel.oe.x = sel.ob.x + strlen(match)-1; ++ sel.ob.y = sel.oe.y = row; ++ selnormalize(); ++ tsetdirt(sel.nb.y, sel.ne.y); ++ xsetsel(getsel()); ++ xclipcopy(); ++ } ++ ++ free(linestr); ++} +diff --git a/st.h b/st.h +index dac64d8..5a58f8f 100644 +--- a/st.h ++++ b/st.h +@@ -85,6 +85,7 @@ void printscreen(const Arg *); + void printsel(const Arg *); + void sendbreak(const Arg *); + void toggleprinter(const Arg *); ++void copyurl(const Arg *); + + int tattrset(int); + void tnew(int, int); +-- +2.20.1 + diff --git a/st.suckless.org/patches/copyurl/st-copyurl-20190202-3be4cf1.diff b/st.suckless.org/patches/copyurl/st-copyurl-20190202-3be4cf1.diff @@ -0,0 +1,177 @@ +From f0d27279e47dac9bc413830cc116662dde91837c Mon Sep 17 00:00:00 2001 +From: Michael Buch <michaelbuch12@gmail.com> +Date: Sat, 2 Feb 2019 14:10:31 +0000 +Subject: [PATCH] Highlight selected urls + +--- + config.def.h | 1 + + st.c | 124 +++++++++++++++++++++++++++++++++++++++++++++++++++ + st.h | 1 + + 3 files changed, 126 insertions(+) + +diff --git a/config.def.h b/config.def.h +index 0e01717..3fb13ec 100644 +--- a/config.def.h ++++ b/config.def.h +@@ -178,6 +178,7 @@ static Shortcut shortcuts[] = { + { TERMMOD, XK_Y, selpaste, {.i = 0} }, + { ShiftMask, XK_Insert, selpaste, {.i = 0} }, + { TERMMOD, XK_Num_Lock, numlock, {.i = 0} }, ++ { MODKEY, XK_l, copyurl, {.i = 0} }, + }; + + /* +diff --git a/st.c b/st.c +index b8e6077..6970e51 100644 +--- a/st.c ++++ b/st.c +@@ -200,6 +200,8 @@ static void tdefutf8(char); + static int32_t tdefcolor(int *, int *, int); + static void tdeftran(char); + static void tstrsequence(uchar); ++static void tsetcolor(int, int, int, uint32_t, uint32_t); ++static char * findlastany(char *, const char**, size_t); + + static void drawregion(int, int, int, int); + +@@ -2602,3 +2604,125 @@ redraw(void) + tfulldirt(); + draw(); + } ++ ++void ++tsetcolor( int row, int start, int end, uint32_t fg, uint32_t bg ) ++{ ++ int i = start; ++ for( ; i < end; ++i ) ++ { ++ term.line[row][i].fg = fg; ++ term.line[row][i].bg = bg; ++ } ++} ++ ++char * ++findlastany(char *str, const char** find, size_t len) ++{ ++ char* found = NULL; ++ int i = 0; ++ for(found = str + strlen(str) - 1; found >= str; --found) { ++ for(i = 0; i < len; i++) { ++ if(strncmp(found, find[i], strlen(find[i])) == 0) { ++ return found; ++ } ++ } ++ } ++ ++ return NULL; ++} ++ ++/* ++** Select and copy the previous url on screen (do nothing if there's no url). ++** ++** FIXME: doesn't handle urls that span multiple lines; will need to add support ++** for multiline "getsel()" first ++*/ ++void ++copyurl(const Arg *arg) { ++ /* () and [] can appear in urls, but excluding them here will reduce false ++ * positives when figuring out where a given url ends. ++ */ ++ static char URLCHARS[] = "ABCDEFGHIJKLMNOPQRSTUVWXYZ" ++ "abcdefghijklmnopqrstuvwxyz" ++ "0123456789-._~:/?#@!$&'*+,;=%"; ++ ++ static const char* URLSTRINGS[] = {"http://", "https://"}; ++ ++ /* remove highlighting from previous selection if any */ ++ if(sel.ob.x >= 0 && sel.oe.x >= 0) ++ tsetcolor(sel.nb.y, sel.ob.x, sel.oe.x + 1, defaultfg, defaultbg); ++ ++ int i = 0, ++ row = 0, /* row of current URL */ ++ col = 0, /* column of current URL start */ ++ startrow = 0, /* row of last occurrence */ ++ colend = 0, /* column of last occurrence */ ++ passes = 0; /* how many rows have been scanned */ ++ ++ char *linestr = calloc(sizeof(char), term.col+1); /* assume ascii */ ++ char *c = NULL, ++ *match = NULL; ++ ++ row = (sel.ob.x >= 0 && sel.nb.y > 0) ? sel.nb.y : term.bot; ++ LIMIT(row, term.top, term.bot); ++ startrow = row; ++ ++ colend = (sel.ob.x >= 0 && sel.nb.y > 0) ? sel.nb.x : term.col; ++ LIMIT(colend, 0, term.col); ++ ++ /* ++ ** Scan from (term.bot,term.col) to (0,0) and find ++ ** next occurrance of a URL ++ */ ++ while(passes !=term.bot + 2) { ++ /* Read in each column of every row until ++ ** we hit previous occurrence of URL ++ */ ++ for (col = 0, i = 0; col < colend; ++col,++i) { ++ /* assume ascii */ ++ if (term.line[row][col].u > 127) ++ continue; ++ linestr[i] = term.line[row][col].u; ++ } ++ linestr[term.col] = '\0'; ++ ++ if ((match = findlastany(linestr, URLSTRINGS, ++ sizeof(URLSTRINGS)/sizeof(URLSTRINGS[0])))) ++ break; ++ ++ if (--row < term.top) ++ row = term.bot; ++ ++ colend = term.col; ++ passes++; ++ }; ++ ++ if (match) { ++ /* must happen before trim */ ++ selclear(); ++ sel.ob.x = strlen(linestr) - strlen(match); ++ ++ /* trim the rest of the line from the url match */ ++ for (c = match; *c != '\0'; ++c) ++ if (!strchr(URLCHARS, *c)) { ++ *c = '\0'; ++ break; ++ } ++ ++ /* highlight selection by inverting terminal colors */ ++ tsetcolor(row, sel.ob.x, sel.ob.x + strlen( match ), defaultbg, defaultfg); ++ ++ /* select and copy */ ++ sel.mode = 1; ++ sel.type = SEL_REGULAR; ++ sel.oe.x = sel.ob.x + strlen(match)-1; ++ sel.ob.y = sel.oe.y = row; ++ selnormalize(); ++ tsetdirt(sel.nb.y, sel.ne.y); ++ xsetsel(getsel()); ++ xclipcopy(); ++ } ++ ++ free(linestr); ++} +diff --git a/st.h b/st.h +index 38c61c4..67e7419 100644 +--- a/st.h ++++ b/st.h +@@ -84,6 +84,7 @@ void printscreen(const Arg *); + void printsel(const Arg *); + void sendbreak(const Arg *); + void toggleprinter(const Arg *); ++void copyurl(const Arg *); + + int tattrset(int); + void tnew(int, int); +-- +2.20.1 +