buku/buku

1337 lines
40 KiB
Plaintext
Raw Normal View History

#!/usr/bin/env python3
#
# Bookmark management utility
#
# Copyright (C) 2015 Arun Prakash Jana <engineerarun@gmail.com>
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with buku. If not, see <http://www.gnu.org/licenses/>.
import sys
import os
import sqlite3
2016-04-24 14:19:32 -05:00
import argparse
import readline
import webbrowser
import html.parser as HTMLParser
2016-04-04 04:50:23 -05:00
from http.client import HTTPConnection, HTTPSConnection
from urllib.parse import urljoin, quote, unquote
import gzip
import io
import signal
# Import libraries needed for encryption
try:
import getpass
import hashlib
from Crypto.Cipher import AES
from Crypto import Random
import struct
no_crypto = False
BLOCKSIZE = 65536
SALT_SIZE = 32
CHUNKSIZE = 0x80000 # Read/write 512 KB chunks
except ImportError:
no_crypto = True
# Globals
2016-04-05 23:55:25 -05:00
update = False # Update a bookmark in DB
2016-05-18 12:23:08 -05:00
tagsearch = False # Search bookmarks by tag
2016-04-05 23:55:25 -05:00
titleData = None # Title fetched from a page
titleManual = None # Manually add a title offline
2016-05-17 15:11:31 -05:00
description = None # Description of the bookmark
2016-04-17 11:07:26 -05:00
jsonOutput = False # Output json formatted result
2016-04-25 11:23:03 -05:00
showOpt = 0 # Modify show. 1: show only URL, 2: show URL and tag
2016-04-24 16:18:56 -05:00
debug = False # Enable debug logs
2016-04-05 23:55:25 -05:00
pipeargs = [] # Holds arguments piped to the program
2016-05-01 11:20:56 -05:00
_VERSION_ = 2.0 # Program version
2016-04-05 23:55:25 -05:00
2016-04-05 06:25:40 -05:00
class BMHTMLParser(HTMLParser.HTMLParser):
2016-04-05 23:39:56 -05:00
"""Class to parse and fetch the title from a HTML page, if available"""
2016-04-05 06:25:40 -05:00
def __init__(self):
HTMLParser.HTMLParser.__init__(self)
self.inTitle = False
self.data = ""
self.lasttag = None
def handle_starttag(self, tag, attrs):
self.inTitle = False
if tag == "title":
self.inTitle = True
self.lasttag = tag
def handle_endtag(self, tag):
global titleData
if tag == "title":
self.inTitle = False
if self.data != "":
titleData = self.data
self.reset() # We have received title data, exit parsing
def handle_data(self, data):
if self.lasttag == "title" and self.inTitle == True:
self.data += data
def error(self, message):
pass
2016-04-05 06:25:40 -05:00
def getDataPath():
2016-04-10 07:28:49 -05:00
"""Determine the DB file path:
if $XDG_DATA_HOME is defined, use it
else if $HOME exists, use it
else use the current directory
"""
data_home = os.environ.get('XDG_DATA_HOME')
if data_home is None:
if os.environ.get('HOME') is None:
data_home = '.'
else:
data_home = os.path.join(os.environ.get('HOME'), '.local', 'share')
return os.path.join(data_home, 'buku')
def moveOldDatabase():
2016-04-10 07:28:49 -05:00
"""Move database file from earlier path used in versions <= 1.8
to new path. Errors out if both the old and new DB files exist.
"""
olddbpath = os.path.join(os.environ.get('HOME'), '.cache', 'buku')
olddbfile = os.path.join(olddbpath, 'bookmarks.db')
if not os.path.exists(olddbfile):
return
newdbpath = getDataPath()
newdbfile = os.path.join(newdbpath, 'bookmarks.db')
if os.path.exists(newdbfile):
print("Both old (%s) and new (%s) databases exist, need manual action" % (olddbfile, newdbfile))
sys.exit(1)
if not os.path.exists(newdbpath):
os.makedirs(newdbpath)
2016-05-14 10:29:03 -05:00
os.rename(olddbfile, newdbfile)
2016-04-08 09:15:47 -05:00
print("Database was moved from old (%s) to new (%s) location.\n" % (olddbfile, newdbfile))
os.rmdir(olddbpath)
def initdb():
2016-04-05 06:25:40 -05:00
"""Initialize the database connection. Create DB file and/or bookmarks table
if they don't exist. Alert on encryption options on first execution.
Returns: connection, cursor
"""
dbpath = getDataPath()
if not os.path.exists(dbpath):
os.makedirs(dbpath)
dbfile = os.path.join(dbpath, 'bookmarks.db')
encpath = os.path.join(dbpath, 'bookmarks.db.enc')
# Notify if DB file needs to be decrypted first
if os.path.exists(encpath) and not os.path.exists(dbfile):
print("Unlock database first")
sys.exit(1)
# Show info on first creation
if no_crypto == False and not os.path.exists(dbfile):
print("DB file is being created. You may want to encrypt it later.")
try:
# Create a connection
conn = sqlite3.connect(dbfile)
cur = conn.cursor()
# Create table if it doesn't exist
cur.execute('''CREATE TABLE if not exists bookmarks \
2016-05-17 15:11:31 -05:00
(id integer PRIMARY KEY, URL text NOT NULL UNIQUE, metadata text, tags text, desc text)''')
conn.commit()
except Exception as e:
print("\x1b[1mEXCEPTION\x1b[21m [initdb]: (%s) %s" % (type(e).__name__, e))
sys.exit(1)
2016-05-17 15:11:31 -05:00
# Add description column in existing DB (from version 2.1)
try:
cur.execute("""ALTER TABLE bookmarks ADD COLUMN desc text default \'\'""")
conn.commit()
except:
pass
return (conn, cur)
2016-04-09 09:10:30 -05:00
def getPageResp(url, fullurl=False):
2016-04-05 06:25:40 -05:00
"""Connect to a server and fetch the requested page data.
Supports gzip compression.
2016-04-10 07:41:00 -05:00
2016-04-05 06:25:40 -05:00
Params: URL to fetch, redirection status
Returns: connection, HTTP(S) GET response
"""
if url.find("%20") != -1:
2016-04-19 12:02:15 -05:00
url = unquote(url).replace(" ", "%20")
else:
url = unquote(url)
if debug:
print("unquoted: %s" % url)
if url.find("https://") >= 0: # Secure connection
server = url[8:]
marker = server.find("/")
if marker > 0:
2016-04-09 09:10:30 -05:00
if fullurl == False:
url = server[marker:]
server = server[:marker]
urlconn = HTTPSConnection(server, timeout=30)
elif url.find("http://") >= 0: # Insecure connection
server = url[7:]
marker = server.find("/")
if marker > 0:
2016-04-09 09:10:30 -05:00
if fullurl == False:
url = server[marker:]
server = server[:marker]
urlconn = HTTPConnection(server, timeout=30)
else:
printmsg("Not a valid HTTP(S) url", "WARNING")
2016-05-01 12:49:02 -05:00
if url.find(":") == -1:
printmsg("Doesn't appear to be a valid url either", "WARNING")
return (None, None)
if debug:
2016-04-14 09:04:02 -05:00
print("server [%s] url [%s]" % (server, url))
# Handle URLs passed with %xx escape
try:
url.encode('ascii')
except:
url = quote(url)
urlconn.request("GET", url, None, {
"Accept-encoding": "gzip",
})
return (urlconn, urlconn.getresponse())
2016-04-05 06:25:40 -05:00
def getTitleData(resp):
"""Invoke HTML parser and extract title from HTTP response
2016-04-10 07:41:00 -05:00
2016-04-05 06:25:40 -05:00
Params: GET response
"""
data = None
2016-04-05 06:25:40 -05:00
charset = resp.headers.get_content_charset()
if resp.headers.get('Content-Encoding') == 'gzip':
if debug:
print("gzip response")
data = gzip.GzipFile(fileobj=io.BytesIO(resp.read())).read()
else:
data = resp.read()
2016-04-05 06:25:40 -05:00
if charset == None:
charset = 'utf-8'
2016-04-14 09:04:02 -05:00
if debug:
printmsg("Charset missing in response", "WARNING")
2016-04-05 06:25:40 -05:00
if debug:
print("charset: %s" % charset)
parser = BMHTMLParser()
try:
if charset == 'utf-8':
parser.feed(data.decode(charset, "replace"))
2016-04-05 06:25:40 -05:00
else:
parser.feed(data.decode(charset))
2016-04-05 06:25:40 -05:00
except Exception as e:
if debug and str(e) != "we should not get here!":
# Suppress Exception due to intentional self.reset() in HTMLParser
2016-04-09 12:39:00 -05:00
print("\x1b[1mEXCEPTION\x1b[21m [getTitleData]: (%s) %s" % (type(e).__name__, e))
2016-04-05 06:25:40 -05:00
def fetchTitle(url):
2016-04-05 06:25:40 -05:00
"""Handle server connection and redirections
2016-04-10 07:41:00 -05:00
2016-04-05 06:25:40 -05:00
Params: URL to fetch
Returns: page title or empty string, if not found
"""
global titleData
titleData = None
urlconn = None
retry = False
try:
urlconn, resp = getPageResp(url, False)
while 1:
if resp is None:
break
elif resp.status == 200:
getTitleData(resp)
break
elif resp.status in [301, 302]:
redirurl = urljoin(url, resp.getheader('location', ''))
printmsg(redirurl, "REDIRECTION")
2016-04-10 08:01:11 -05:00
if redirurl.find("sorry/IndexRedirect?") >= 0: # gracefully handle Google blocks
printmsg("Connection blocked due to unusual activity", "ERROR")
break
marker = redirurl.find("redirectUrl=")
if marker != -1:
redirurl = redirurl[marker + 12:]
# break same URL redirection loop
if url == redirurl:
printmsg("Detected repeated redirection to same URL", "ERROR")
break
url = redirurl
urlconn.close()
2016-04-09 09:10:30 -05:00
# Try with complete URL on redirection
urlconn, resp = getPageResp(url, True)
elif resp.status == 500 and retry == False:
"""Retry on status 500 (Internal Server Error) with truncated
URL. Some servers support truncated request URL on redirection.
"""
urlconn.close()
if debug:
print("Received status 500: retrying.")
urlconn, resp = getPageResp(url, False)
retry = True
else:
printmsg(("[" + str(resp.status) + "] " + resp.reason), "ERROR")
break
except Exception as e:
2016-04-09 12:39:00 -05:00
print("\x1b[1mEXCEPTION\x1b[21m [fetchTitle]: (%s) %s" % (type(e).__name__, e))
finally:
if urlconn is not None:
urlconn.close()
if titleData is None:
return ''
return titleData.strip().replace("\n","")
def isBookmarkAdded(cur, url):
"""Check if URL already exists in DB
Params: cursor, URL to search
Returns: DB index if URL found, else -1
"""
cur.execute("SELECT id FROM bookmarks WHERE URL = ?", (url,))
resultset = cur.fetchall()
if len(resultset) == 0:
return -1
return resultset[0][0]
def AddUpdateEntry(conn, cur, keywords, updateindex, insertindex=0):
"""Add a new bookmark or update an existing record at
updateindex or insert a new record at insertindex (if empty)
2016-04-10 07:41:00 -05:00
Params: connection, cursor, keywords, index to update, index to insert at
2016-04-05 06:25:40 -05:00
"""
global titleManual
2016-05-17 15:11:31 -05:00
global description
tags = ','
meta = ''
url = keywords[0]
"""In case of an add or insert operation ensure
that the URL does not exist in DB already
"""
2016-04-25 11:23:03 -05:00
if updateindex == 0:
id = isBookmarkAdded(cur, url)
if id != -1:
print("URL already exists at index %d" % id)
return
# Cleanse and get the tags
if len(keywords) > 1:
for tag in keywords[1:]:
if tag[-1] == ',':
2016-04-19 12:02:15 -05:00
tag = tag.strip(',') + ',' # if delimiter is present, maintain it
else:
tag = tag.strip(',') # a token in a multi-word tag
if tag == ',':
continue
if tags[-1] == ',':
tags += tag
else:
tags += ' ' + tag
if tags[-1] != ',':
tags += ','
if titleManual is not None:
meta = titleManual
2016-04-21 09:24:06 -05:00
else:
meta = fetchTitle(url)
if meta == '':
print("\x1B[91mTitle: []\x1B[0m")
else:
print("Title: [%s]" % meta)
2016-04-25 11:23:03 -05:00
if updateindex == 0: # Add or insert a new entry
2016-05-17 15:11:31 -05:00
if description is None:
description = ''
try:
if insertindex == 0: # insertindex is index number to insert record at
2016-05-17 15:11:31 -05:00
cur.execute('INSERT INTO bookmarks(URL, metadata, tags, desc) VALUES (?, ?, ?, ?)', (url, meta, tags, description))
else:
2016-05-17 15:11:31 -05:00
cur.execute('INSERT INTO bookmarks(id, URL, metadata, tags, desc) VALUES (?, ?, ?, ?, ?)', (insertindex, url, meta, tags, description))
conn.commit()
print("Added at index %d\n" % cur.lastrowid)
2016-04-24 17:05:46 -05:00
printdb(cur, cur.lastrowid)
except sqlite3.IntegrityError:
for row in cur.execute("SELECT id from bookmarks where URL LIKE ?", (url,)):
print("URL already exists at index %s" % row[0])
return
print("Index %d exists" % insertindex)
else: # Update an existing entry
try:
2016-05-17 15:11:31 -05:00
if description is None:
cur.execute("UPDATE bookmarks SET URL = ?, metadata = ?, tags = ? WHERE id = ?", (url, meta, tags, updateindex,))
else:
cur.execute("UPDATE bookmarks SET URL = ?, metadata = ?, tags = ?, desc = ? WHERE id = ?", (url, meta, tags, description, updateindex,))
conn.commit()
if cur.rowcount == 1:
2016-04-25 11:23:03 -05:00
print("Updated index %d\n" % updateindex)
printdb(cur, updateindex)
else:
print("No matching index")
except sqlite3.IntegrityError:
print("URL already exists")
2016-04-21 22:12:17 -05:00
def dbRefresh(conn, cur, index):
2016-04-05 06:25:40 -05:00
"""Refresh ALL records in the database. Fetch title for each
bookmark from the web and update the records. Doesn't udpate
the record if title is empty.
This API doesn't change DB index, URL or tags of a bookmark.
2016-04-10 07:41:00 -05:00
2016-04-05 06:25:40 -05:00
Params: connection, cursor
"""
2016-04-21 23:53:38 -05:00
global titleManual
2016-04-21 22:12:17 -05:00
if index == 0:
cur.execute("SELECT id, url FROM bookmarks ORDER BY id ASC")
else:
cur.execute("SELECT id, url FROM bookmarks WHERE id = ?", (index,))
2016-04-21 23:53:38 -05:00
resultset = cur.fetchall()
2016-04-22 09:40:58 -05:00
if titleManual is None:
for row in resultset:
2016-04-21 23:53:38 -05:00
title = fetchTitle(row[1])
if title == '':
print("\x1B[91mTitle: []")
print("\x1b[1mNOT updating index %d\x1b[21m\x1B[0m\n" % row[0])
continue
else:
print("Title: [%s]" % title)
2016-04-22 09:40:58 -05:00
cur.execute("UPDATE bookmarks SET metadata = ? WHERE id = ?", (title, row[0],))
conn.commit()
print("Updated index %d\n" % row[0])
else:
title = titleManual
2016-04-22 09:40:58 -05:00
for row in resultset:
cur.execute("UPDATE bookmarks SET metadata = ? WHERE id = ?", (title, row[0],))
conn.commit()
print("Updated index %d\n" % row[0])
2016-04-24 15:33:59 -05:00
def searchdb(cur, keywords, all_keywords=False):
2016-04-05 06:25:40 -05:00
"""Search the database for an entries with tags or URL
or title info matching keywords and list those.
2016-04-10 07:41:00 -05:00
Params: cursor, keywords to search, search any or all keywords
2016-04-05 06:25:40 -05:00
"""
global jsonOutput
arguments = []
placeholder = "'%' || ? || '%'"
2016-05-17 15:11:31 -05:00
query = "SELECT id, url, metadata, tags, desc FROM bookmarks WHERE"
2016-04-24 15:33:59 -05:00
if all_keywords == True: # Match all keywords in URL or Title
for token in keywords:
2016-05-17 15:11:31 -05:00
query += " (tags LIKE (%s) OR URL LIKE (%s) OR metadata LIKE (%s) OR desc LIKE (%s)) AND" % (placeholder, placeholder, placeholder, placeholder)
arguments.append(token)
arguments.append(token)
arguments.append(token)
arguments.append(token)
query = query[:-4]
else: # Match any keyword in URL or Title
for token in keywords:
2016-05-17 15:11:31 -05:00
query += " tags LIKE (%s) OR URL LIKE (%s) OR metadata LIKE (%s) OR desc LIKE (%s) OR" % (placeholder, placeholder, placeholder, placeholder)
arguments.append(token)
arguments.append(token)
arguments.append(token)
arguments.append(token)
query = query[:-3]
if debug:
print("\"%s\", (%s)" % (query, arguments))
2016-05-16 12:30:47 -05:00
cur.execute(query, arguments)
2016-05-17 10:57:12 -05:00
results = cur.fetchall()
if len(results) == 0:
return
if jsonOutput == False:
2016-05-18 10:46:08 -05:00
showPrompt(results)
else:
print(formatJson(results))
2016-05-17 10:57:12 -05:00
2016-05-18 10:46:08 -05:00
def searchTag(cur, tag):
"""Search and list bookmarks with a tag
2016-05-17 10:57:12 -05:00
2016-05-18 10:46:08 -05:00
Params: cursor, tag to search
"""
2016-05-17 10:57:12 -05:00
2016-05-18 10:46:08 -05:00
global jsonOutput
cur.execute("SELECT id, url, metadata, tags, desc FROM bookmarks WHERE tags LIKE '%' || ? || '%'", (tag,))
results = cur.fetchall()
if len(results) == 0:
return
if jsonOutput == False:
showPrompt(results)
2016-05-17 10:57:12 -05:00
else:
print(formatJson(results))
2016-04-17 09:45:29 -05:00
def compactDB(conn, cur, index):
2016-04-05 06:25:40 -05:00
"""When an entry at index is deleted, move the last
entry in DB to index, if index is lesser.
2016-04-10 07:41:00 -05:00
2016-04-05 06:25:40 -05:00
Params: connection, cursor, index of deleted entry
"""
cur.execute('SELECT MAX(id) from bookmarks')
if cur.rowcount < 1:
return
results = cur.fetchall()
for row in results:
if row[0] > index:
2016-05-17 15:11:31 -05:00
cur.execute('SELECT id, URL, metadata, tags, desc FROM bookmarks WHERE id = ?', (row[0],))
results = cur.fetchall()
for row in results:
cur.execute('DELETE FROM bookmarks WHERE id = ?', (row[0],))
conn.commit()
2016-05-17 15:11:31 -05:00
cur.execute('INSERT INTO bookmarks(id, URL, metadata, tags, desc) VALUES (?, ?, ?, ?, ?)', (index, row[1], row[2], row[3], row[4],))
conn.commit()
print("Index %d moved to %d" % (row[0], index))
def cleardb(conn, cur, index):
2016-04-05 06:25:40 -05:00
"""Delete a single record or remove the table if index is None
2016-04-10 07:41:00 -05:00
2016-04-05 06:25:40 -05:00
Params: connection, cursor, index to delete
"""
2016-04-24 15:33:59 -05:00
if index == 0: # Remove the table
resp = input("ALL bookmarks will be removed. Enter \x1b[1my\x1b[21m to confirm: ")
if resp != 'y':
print("No bookmarks deleted")
return
cur.execute('DROP TABLE if exists bookmarks')
conn.commit()
print("All bookmarks deleted")
else: # Remove a single entry
try:
2016-04-24 15:33:59 -05:00
cur.execute('DELETE FROM bookmarks WHERE id = ?', (index,))
conn.commit()
if cur.rowcount == 1:
2016-04-24 15:33:59 -05:00
print("Removed index %d" % index)
compactDB(conn, cur, index)
else:
print("No matching index")
except IndexError:
print("Index out of bound")
2016-05-18 10:46:08 -05:00
def showPrompt(results):
"""Show each matching result from a search"""
count = 0
for row in results:
count += 1
printRecord(row, count)
while True:
try:
nav = input("Result number to open: ")
except EOFError:
return
if is_int(nav):
index = int(nav) - 1
if index < 0 or index >= count:
print("Index out of bound")
continue
try:
browser_open(unquote(results[index][1]))
except Exception as e:
print("\x1b[1mEXCEPTION\x1b[21m [searchdb]: (%s) %s" % (type(e).__name__, e))
else:
break
2016-05-17 15:11:31 -05:00
def printRecord(row, count=0):
"""Print a single DB record
Handles differently for search and print (count = 0)
"""
# Print index and URL
if count != 0:
print("\x1B[1m\x1B[93m%d. \x1B[0m\x1B[92m%s\x1B[0m\t[%d]" % (count, row[1], row[0]))
else:
print("\x1B[1m\x1B[93m%d. \x1B[0m\x1B[92m%s\x1B[0m" % (row[0], row[1]))
# Print title
if row[2] != '':
2016-05-17 17:35:01 -05:00
print(" \x1B[91m>\x1B[0m %s" % row[2])
2016-05-17 15:11:31 -05:00
# Print description
if row[4] != '':
2016-05-17 17:35:01 -05:00
print(" \x1B[91m+\x1B[0m %s" % row[4])
2016-05-17 15:11:31 -05:00
# Print tags
if row[3] != ',':
2016-05-17 17:35:01 -05:00
print(" \x1B[91m#\x1B[0m %s" % row[3][1:-1])
2016-05-17 15:11:31 -05:00
print("")
def printdb(cur, index, empty=False):
2016-04-05 06:25:40 -05:00
"""Print bookmark details at index or all bookmarks if index is None
2016-04-24 17:05:46 -05:00
Print only bookmarks with blank title or tag if empty is True
2016-04-24 16:18:56 -05:00
Note: URL is printed on top because title may be blank
2016-04-10 07:41:00 -05:00
2016-04-05 06:25:40 -05:00
Params: cursor, index to print, flag to show only bookmarks with no title or tags
"""
global showOpt
2016-03-22 18:31:17 -05:00
global jsonOutput
resultset = None
2016-04-24 17:05:46 -05:00
if index == 0: # Show all entries
if empty == False:
cur.execute('SELECT * FROM bookmarks')
resultset = cur.fetchall()
else:
cur.execute("SELECT * FROM bookmarks WHERE metadata = '' OR tags = ','")
resultset = cur.fetchall()
print("\x1b[1m%d records found\x1b[21m\n" % len(resultset))
if jsonOutput == False:
2016-04-19 12:02:15 -05:00
if showOpt == 0:
for row in resultset:
2016-05-17 15:11:31 -05:00
printRecord(row)
2016-04-19 12:02:15 -05:00
elif showOpt == 1:
for row in resultset:
print("%s %s" % (row[0], row[1]))
2016-04-19 12:02:15 -05:00
elif showOpt == 2:
for row in resultset:
print("%s %s %s" % (row[0], row[1], row[3][1:-1]))
2016-04-19 12:02:15 -05:00
else:
print(formatJson(resultset))
else: # Show record at index
try:
2016-04-24 17:05:46 -05:00
resultset = cur.execute("SELECT * FROM bookmarks WHERE id = ?", (index,))
except IndexError:
print("Index out of bound")
return
if jsonOutput == False:
for row in resultset:
2016-05-17 15:11:31 -05:00
printRecord(row)
return
print("No matching index")
else:
print(formatJson(resultset, True))
2016-03-23 07:10:08 -05:00
2016-03-22 18:29:45 -05:00
def formatJson(resultset, single=False):
2016-05-16 09:39:01 -05:00
"""Return results in Json format"""
2016-03-22 18:29:45 -05:00
global showOpt
if single == False:
marks = []
for row in resultset:
if showOpt == 1:
record = { 'url': row[1] }
elif showOpt == 2:
record = { 'url': row[1], 'tags': row[3][1:-1] }
else:
2016-05-18 14:15:57 -05:00
record = { 'url': row[1], 'title': row[2], 'comment': row[4], 'tags': row[3][1:-1]}
2016-03-22 18:29:45 -05:00
marks.append(record)
else:
marks = {}
for row in resultset:
if showOpt == 1:
marks['url'] = row[1]
elif showOpt == 2:
marks['title'] = row[2]
marks['tags'] = row[3][1:-1]
else:
marks['url'] = row[1]
marks['title'] = row[2]
2016-05-18 14:15:57 -05:00
marks['comment'] = row[4]
2016-03-22 18:29:45 -05:00
marks['tags'] = row[3][1:-1]
2016-03-22 18:29:45 -05:00
return json.dumps(marks, sort_keys=True, indent=4)
2016-05-16 09:39:01 -05:00
def showUniqueTags(cur):
2016-04-05 06:25:40 -05:00
"""Print all unique tags ordered alphabetically
2016-04-10 07:41:00 -05:00
2016-04-05 06:25:40 -05:00
Params: cursor
"""
count = 1
Tags = []
uniqueTags = []
for row in cur.execute('SELECT DISTINCT tags FROM bookmarks'):
if row[0] == ',':
continue
Tags.extend(row[0].strip(',').split(','))
for tag in Tags:
if tag not in uniqueTags:
uniqueTags.append(tag)
Tags = sorted(uniqueTags, key=str.lower)
for tag in Tags:
print("%6d. %s" % (count, tag))
count += 1
2016-04-29 12:29:06 -05:00
def replaceTags(conn, cur, orig, new=None):
2016-04-05 06:25:40 -05:00
"""Replace orig tags with new tags in DB for all records.
Remove orig tag is new tag is empty.
2016-04-10 07:41:00 -05:00
2016-04-05 06:25:40 -05:00
Params: connection, cursor, original and new tags
"""
update = False
delete = False
orig = ',' + orig + ','
2016-04-29 12:29:06 -05:00
if new is None:
newtags = ','
delete = True
else:
2016-04-29 12:29:06 -05:00
newtags = ','
for tag in new:
if tag[-1] == ',':
tag = tag.strip(',') + ',' # if delimiter is present, maintain it
else:
tag = tag.strip(',') # a token in a multi-word tag
2016-04-29 12:29:06 -05:00
if tag == ',':
continue
if newtags[-1] == ',':
newtags += tag
else:
newtags += ' ' + tag
if newtags[-1] != ',':
newtags += ','
if newtags == ',':
delete = True
2016-04-29 12:29:06 -05:00
if orig == newtags:
print("Tags are same.")
return
cur.execute("SELECT id, tags FROM bookmarks WHERE tags LIKE ?", ('%' + orig + '%',))
results = cur.fetchall()
for row in results:
if delete == False:
2016-04-29 12:29:06 -05:00
# Check if tag newtags is already added
if row[1].find(newtags) >= 0:
newtags = ','
2016-04-29 12:29:06 -05:00
newtags = row[1].replace(orig, newtags)
cur.execute("UPDATE bookmarks SET tags = ? WHERE id = ?", (newtags, row[0],))
print("Updated index %d" % row[0])
update = True
if update:
conn.commit()
def fetchopen(index):
2016-04-05 06:25:40 -05:00
"""Fetch URL at index and open in browser
2016-04-10 07:41:00 -05:00
2016-04-05 06:25:40 -05:00
Params: index
"""
try:
2016-04-24 16:18:56 -05:00
for row in cur.execute("SELECT URL FROM bookmarks WHERE id = ?", (index,)):
url = unquote(row[0])
browser_open(url)
return
print("No matching index")
except IndexError:
print("Index out of bound")
def is_int(string):
2016-04-05 06:25:40 -05:00
"""Check if a string is a digit
2016-04-10 07:41:00 -05:00
2016-04-05 06:25:40 -05:00
Params: string
"""
try:
int(string)
return True
except:
return False
def browser_open(url):
"""Duplicate stdin, stdout (to suppress showing errors
on the terminal) and open URL in default browser
Params: url to open
"""
url = url.replace("%22", "\"")
_stderr = os.dup(2)
os.close(2)
_stdout = os.dup(1)
os.close(1)
fd = os.open(os.devnull, os.O_RDWR)
os.dup2(fd, 2)
os.dup2(fd, 1)
try:
webbrowser.open(url)
except Exception as e:
2016-04-09 12:39:00 -05:00
print("\x1b[1mEXCEPTION\x1b[21m [browser_open]: (%s) %s" % (type(e).__name__, e))
finally:
os.close(fd)
os.dup2(_stderr, 2)
os.dup2(_stdout, 1)
def get_filehash(filepath):
"""Get the SHA256 hash of a file
Params: path to the file
"""
with open(filepath, 'rb') as f:
hasher = hashlib.sha256()
buf = f.read(BLOCKSIZE)
while len(buf) > 0:
hasher.update(buf)
buf = f.read(BLOCKSIZE)
return hasher.digest()
def encrypt_file(iterations):
"""Encrypt the bookmarks database file"""
dbpath = os.path.join(getDataPath(), 'bookmarks.db')
encpath = dbpath + '.enc'
if not os.path.exists(dbpath):
print("%s missing. Already encrypted?" % dbpath)
sys.exit(1)
# If both encrypted file and flat file exist, error out
if os.path.exists(dbpath) and os.path.exists(encpath):
printmsg("Both encrypted and flat DB files exist!", "ERROR")
sys.exit(1)
password = ''
password = getpass.getpass()
passconfirm = getpass.getpass()
if password == '':
print("Empty password");
sys.exit(1)
if password != passconfirm:
print("Passwords don't match");
sys.exit(1)
# Get SHA256 hash of DB file
dbhash = get_filehash(dbpath)
2016-04-10 08:01:11 -05:00
# Generate random 256-bit salt and key
2016-04-10 07:41:00 -05:00
salt = Random.get_random_bytes(SALT_SIZE)
key = (password + salt.decode('utf-8', "replace")).encode('utf-8')
for i in range(iterations):
key = hashlib.sha256(key).digest()
iv = Random.get_random_bytes(16)
cipher = AES.new(key, AES.MODE_CBC, iv)
filesize = os.path.getsize(dbpath)
with open(dbpath, 'rb') as infile:
with open(encpath, 'wb') as outfile:
outfile.write(struct.pack('<Q', filesize))
outfile.write(salt)
outfile.write(iv)
# Embed DB file hash in encrypted file
outfile.write(dbhash)
while True:
chunk = infile.read(CHUNKSIZE)
if len(chunk) == 0:
break
elif len(chunk) % 16 != 0:
chunk += ' ' * (16 - len(chunk) % 16)
outfile.write(cipher.encrypt(chunk))
os.remove(dbpath)
print("File encrypted")
sys.exit(0)
def decrypt_file(iterations):
"""Decrypt the bookmarks database file"""
dbpath = os.path.join(getDataPath(), 'bookmarks.db')
encpath = dbpath + '.enc'
if not os.path.exists(encpath):
printmsg((encpath + " missing"), "ERROR")
sys.exit(1)
# If both encrypted file and flat file exist, error out
if os.path.exists(dbpath) and os.path.exists(encpath):
printmsg("Both encrypted and flat DB files exist!", "ERROR")
sys.exit(1)
password = ''
password = getpass.getpass()
if password == '':
2016-04-24 17:05:46 -05:00
printmsg("Decryption failed", "ERROR");
sys.exit(1)
with open(encpath, 'rb') as infile:
origsize = struct.unpack('<Q', infile.read(struct.calcsize('Q')))[0]
# Read 256-bit salt and generate key
salt = infile.read(32)
key = (password + salt.decode('utf-8', "replace")).encode('utf-8')
for i in range(iterations):
key = hashlib.sha256(key).digest()
iv = infile.read(16)
cipher = AES.new(key, AES.MODE_CBC, iv)
# Get original DB file's SHA256 hash from encrypted file
enchash = infile.read(32)
with open(dbpath, 'wb') as outfile:
while True:
chunk = infile.read(CHUNKSIZE)
if len(chunk) == 0:
break;
outfile.write(cipher.decrypt(chunk))
outfile.truncate(origsize)
# Match hash of generated file with that of original DB file
dbhash = get_filehash(dbpath)
if dbhash != enchash:
os.remove(dbpath)
2016-04-24 17:05:46 -05:00
printmsg("Decryption failed", "ERROR");
2016-04-24 16:18:56 -05:00
sys.exit(1)
else:
os.remove(encpath)
print("File decrypted")
def sigint_handler(signum, frame):
"""Custom SIGINT handler"""
print('\nInterrupted.', file=sys.stderr)
sys.exit(1)
signal.signal(signal.SIGINT, sigint_handler)
2016-05-01 12:29:46 -05:00
def closequit(conn=None, exitval=0):
"""Close a DB connection and exit"""
if conn is not None:
conn.close()
sys.exit(exitval)
def printmsg(msg, level=None):
"""Print a message in 2 parts, with the level in bold
Params: msg, level
"""
if level is not None:
print("\x1b[1m%s:\x1b[21m %s" % (level, msg))
else:
print("%s" % msg)
class customUpdateAction(argparse.Action):
2016-05-18 12:23:08 -05:00
"""Class to capture if optional param 'update'
is actually used, even if sans arguments
"""
2016-04-25 11:23:03 -05:00
def __call__(self, parser, args, values, option_string=None):
global update
update = True
# NOTE: the following converts a None argument to an empty array []
setattr(args, self.dest, values)
2016-04-26 12:23:48 -05:00
2016-05-18 12:23:08 -05:00
class customTagSearchAction(argparse.Action):
"""Class to capture if optional param 'stag'
is actually used, even if sans arguments
"""
def __call__(self, parser, args, values, option_string=None):
global tagsearch
tagsearch = True
setattr(args, self.dest, values)
class customTitleAction(argparse.Action):
2016-05-18 12:23:08 -05:00
"""Class to capture if optional param 'title'
is actually used, even if sans arguments
"""
def __call__(self, parser, args, values, option_string=None):
global titleManual
titleManual = ''
setattr(args, self.dest, values)
2016-05-17 15:11:31 -05:00
class customDescAction(argparse.Action):
2016-05-18 12:23:08 -05:00
"""Class to capture if optional param 'comment'
2016-05-17 15:11:31 -05:00
is actually used, even if sans arguments
"""
def __call__(self, parser, args, values, option_string=None):
global description
description = ''
setattr(args, self.dest, values)
2016-04-24 14:19:32 -05:00
class ExtendedArgumentParser(argparse.ArgumentParser):
"""Extend classic argument parser"""
2016-04-24 14:19:32 -05:00
2016-04-25 15:45:10 -05:00
# Print additional help and info
@staticmethod
def print_extended_help(file=None):
2016-05-14 10:33:21 -05:00
file.write('''
prompt keys:
1-N open the Nth search result in web browser
2016-05-17 10:57:12 -05:00
Enter exit buku
2016-05-14 10:33:21 -05:00
Version %.1f
Copyright (C) 2015-2016 Arun Prakash Jana <engineerarun@gmail.com>
License: GPLv3
Webpage: https://github.com/jarun/buku
''' % _VERSION_)
2016-04-25 15:45:10 -05:00
# Help
2016-04-24 14:19:32 -05:00
def print_help(self, file=None):
super(ExtendedArgumentParser, self).print_help(file)
2016-04-25 15:45:10 -05:00
self.print_extended_help(file)
2016-04-05 23:55:25 -05:00
"""main starts here"""
2016-04-26 12:23:48 -05:00
# Handle piped input
def main(argv = sys.argv):
if not sys.stdin.isatty():
pipeargs.extend(sys.argv)
for s in sys.stdin.readlines():
pipeargs.extend(s.split())
if __name__ == "__main__":
try:
main(sys.argv)
except KeyboardInterrupt:
pass
2016-04-26 12:23:48 -05:00
# If piped input, set argument vector
if len(pipeargs) > 0:
sys.argv = pipeargs
# Setup custom argument parser
2016-04-24 14:19:32 -05:00
argparser = ExtendedArgumentParser(
2016-05-12 03:58:31 -05:00
description='A private command-line bookmark manager. Your mini web!',
2016-04-26 14:22:59 -05:00
formatter_class=argparse.RawTextHelpFormatter,
usage='''buku [-a URL [tags ...]] [-u [N [URL tags ...]]]
2016-05-17 15:11:31 -05:00
[-t [...]] [-c [...]] [-d [N]] [-h]
2016-05-18 12:23:08 -05:00
[-s keyword [...]] [-S keyword [...]] [--st [...]]
2016-04-26 14:22:59 -05:00
[-k [N]] [-l [N]] [-p [N]] [-f N]
2016-04-29 12:29:06 -05:00
[-r oldtag [newtag ...]] [-j] [-o N] [-z]''',
2016-04-26 14:22:59 -05:00
add_help=False
2016-04-24 14:19:32 -05:00
)
2016-04-26 16:13:27 -05:00
# General options
general_group = argparser.add_argument_group(title="general options",
description='''-a, --add URL [tags ...]
bookmark URL with comma separated tags
-u, --update [N [URL tags ...]]
update fields of bookmark at DB index N
refresh all titles, if no arguments
if URL omitted and -t is unused, update
title of bookmark at index N from web
-t, --title [...] manually set title, works with -a, -u
2016-04-29 12:29:06 -05:00
do not set title, if no arguments
2016-05-17 15:11:31 -05:00
-c, --comment [...] description of the bookmark, works with
-a, -u; clears comment, if no arguments
2016-04-26 16:13:27 -05:00
-d, --delete [N] delete bookmark at DB index N
delete all bookmarks, if no arguments
-h, --help show this information''')
general_group.add_argument('-a', '--add', nargs='+', dest='addurl', metavar=('URL', 'tags'), help=argparse.SUPPRESS)
general_group.add_argument('-u', '--update', nargs='*', dest='update', action=customUpdateAction, metavar=('N', 'URL tags'), help=argparse.SUPPRESS)
general_group.add_argument('-t', '--title', nargs='*', dest='title', action=customTitleAction, metavar='title', help=argparse.SUPPRESS)
2016-05-17 15:11:31 -05:00
general_group.add_argument('-c', '--comment', nargs='*', dest='desc', type=str, action=customDescAction, metavar='desc', help=argparse.SUPPRESS)
2016-04-26 16:13:27 -05:00
general_group.add_argument('-d', '--delete', nargs='?', dest='delete', type=int, const=0, metavar='N', help=argparse.SUPPRESS)
general_group.add_argument('-h', '--help', dest='help', action='store_true', help=argparse.SUPPRESS)
# Search options
search_group=argparser.add_argument_group(title="search options",
description='''-s, --sany keyword [...]
search bookmarks for ANY matching keyword
-S, --sall keyword [...]
search bookmarks with ALL keywords
2016-05-18 12:23:08 -05:00
special keyword -
2016-05-18 10:46:08 -05:00
"blank": list entries with empty title/tag
2016-05-18 12:23:08 -05:00
--st, --stag [...] search bookmarks by tag
list all tags alphabetically, if no arguments''')
2016-04-26 16:13:27 -05:00
search_group.add_argument('-s', '--sany', nargs='+', metavar='keyword', help=argparse.SUPPRESS)
search_group.add_argument('-S', '--sall', nargs='+', metavar='keyword', help=argparse.SUPPRESS)
2016-05-18 12:23:08 -05:00
search_group.add_argument('--st', '--stag', nargs='*', dest='stag', action=customTagSearchAction, metavar='keyword', help=argparse.SUPPRESS)
2016-04-26 16:13:27 -05:00
# Encryption options
crypto_group=argparser.add_argument_group(title="encryption options",
description='''-l, --lock [N] encrypt DB file with N (> 0, default 8)
hash iterations to generate key
-k, --unlock [N] decrypt DB file with N (> 0, default 8)
hash iterations to generate key''')
crypto_group.add_argument('-k', '--unlock', nargs='?', dest='decrypt', type=int, const=8, metavar='N', help=argparse.SUPPRESS)
crypto_group.add_argument('-l', '--lock', nargs='?', dest='encrypt', type=int, const=8, metavar='N', help=argparse.SUPPRESS)
# Power toys
power_group=argparser.add_argument_group(title="power toys",
description='''-p, --print [N] show details of bookmark at DB index N
show all bookmarks, if no arguments
-f, --format N modify -p output
N=1: show only URL, N=2: show URL and tag
2016-04-29 12:29:06 -05:00
-r, --replace oldtag [newtag ...]
replace oldtag with newtag in all bookmarks
2016-04-29 12:29:06 -05:00
delete oldtag, if no newtag
2016-05-18 14:15:57 -05:00
-j, --json Json formatted output, for -p, -s, -S, --st
2016-05-15 01:03:02 -05:00
-o, --open N open bookmark at DB index N in web browser
2016-04-26 16:13:27 -05:00
-z, --debug show debug information and additional logs''')
power_group.add_argument('-p', '--print', nargs='?', dest='printindex', type=int, const=0, metavar='N', help=argparse.SUPPRESS)
power_group.add_argument('-f', '--format', dest='showOpt', type=int, choices=[1, 2], metavar='N', help=argparse.SUPPRESS)
2016-04-29 12:29:06 -05:00
power_group.add_argument('-r', '--replace', nargs='+', dest='replace', metavar=('oldtag', 'newtag'), help=argparse.SUPPRESS)
2016-04-26 16:13:27 -05:00
power_group.add_argument('-j', '--json', dest='jsonOutput', action='store_true', help=argparse.SUPPRESS)
power_group.add_argument('-o', '--open', dest='openurl', type=int, metavar='N', help=argparse.SUPPRESS)
power_group.add_argument('-z', '--debug', dest='debug', action='store_true', help=argparse.SUPPRESS)
2016-05-02 12:42:29 -05:00
"""
# NOTE: Insert is functional but commented because DB compaction serves the purpose.
addarg = argparser.add_argument
addarg('-i', '--insert', nargs='+', dest='insert', metavar=('N', 'URL tags'),
help=" insert new bookmark with URL and tags at free DB index N; frees index if URL and tags are omitted")
"""
2016-04-26 14:22:59 -05:00
2016-05-02 12:42:29 -05:00
# Show help and exit if no arguments
if len(sys.argv) < 2:
2016-04-24 14:19:32 -05:00
argparser.print_help(sys.stderr)
sys.exit(1)
2016-04-26 12:23:48 -05:00
# Parse the arguments
2016-04-24 14:19:32 -05:00
args = argparser.parse_args()
2016-04-26 14:22:59 -05:00
# Show help and exit if help requested
if args.help == True:
argparser.print_help(sys.stderr)
sys.exit(0)
2016-04-24 17:05:46 -05:00
# Assign the values to globals
if args.showOpt is not None:
showOpt = args.showOpt
if titleManual is not None and len(args.title) > 0:
titleManual = " ".join(args.title)
2016-05-17 15:11:31 -05:00
if description is not None and len(args.desc) > 0:
description = " ".join(args.desc)
2016-05-14 10:17:38 -05:00
if args.jsonOutput:
import json
jsonOutput = args.jsonOutput
2016-04-24 14:19:32 -05:00
debug = args.debug
2016-04-24 14:19:32 -05:00
# Show version in debug logs
if debug:
print("Version %.1f" % _VERSION_)
# Move pre-1.9 database to new location
moveOldDatabase()
# Handle encrypt/decrypt options at top priority
if args.encrypt is not None:
2016-04-24 15:33:59 -05:00
if no_crypto:
printmsg("PyCrypto missing", "ERROR")
sys.exit(1)
if args.encrypt < 1:
2016-04-24 17:05:46 -05:00
printmsg("Iterations must be >= 1", "ERROR")
sys.exit(1)
encrypt_file(args.encrypt)
if args.decrypt is not None:
2016-04-24 15:33:59 -05:00
if no_crypto:
printmsg("PyCrypto missing", "ERROR")
sys.exit(1)
if args.decrypt < 1:
2016-04-24 17:05:46 -05:00
printmsg("Decryption failed", "ERROR");
sys.exit(1)
decrypt_file(args.decrypt)
2016-04-24 16:18:56 -05:00
# Initialize the database and get handles
conn, cur = initdb()
2016-04-24 15:33:59 -05:00
# Add a record
if args.addurl is not None:
2016-04-25 11:23:03 -05:00
AddUpdateEntry(conn, cur, args.addurl, 0)
2016-05-02 12:42:29 -05:00
# Delete record(s)
2016-04-24 15:33:59 -05:00
if args.delete is not None:
2016-04-24 17:05:46 -05:00
if args.delete < 0:
printmsg("Index must be >= 0", "ERROR")
closequit(conn, 1)
2016-04-24 15:33:59 -05:00
cleardb(conn, cur, args.delete)
2016-04-25 13:43:28 -05:00
# Search URLs, titles, tags for any keyword
2016-05-17 10:57:12 -05:00
if args.sany is not None:
searchdb(cur, args.sany)
2016-04-25 13:43:28 -05:00
2016-05-17 10:57:12 -05:00
# Search URLs, titles, tags with all keywords
2016-04-24 15:33:59 -05:00
if args.sall is not None:
2016-05-18 12:23:08 -05:00
if args.sall[0] == 'blank' and len(args.sall) == 1:
2016-04-25 13:43:28 -05:00
printdb(cur, 0, True)
else:
2016-04-25 13:43:28 -05:00
searchdb(cur, args.sall, True)
2016-05-18 10:46:08 -05:00
# Search bookmarks by tag
2016-05-18 12:23:08 -05:00
if tagsearch == True:
if len(args.stag) > 0:
tag = ',' + " ".join(args.stag) + ','
searchTag(cur, tag)
else:
showUniqueTags(cur)
2016-05-18 10:46:08 -05:00
2016-04-21 12:48:42 -05:00
# Update record
if update == True:
2016-04-25 11:23:03 -05:00
if len(args.update) == 0:
dbRefresh(conn, cur, 0)
elif not args.update[0].isdigit():
printmsg("Index must be a number >= 0", "ERROR")
closequit(conn, 1)
2016-04-25 11:23:03 -05:00
elif int(args.update[0]) == 0:
dbRefresh(conn, cur, 0)
elif len(args.update) == 1:
printmsg("At least URL should be provided for non-zero index", "ERROR")
closequit(conn, 1)
2016-04-21 12:48:42 -05:00
else:
2016-04-25 11:23:03 -05:00
AddUpdateEntry(conn, cur, args.update[1:], int(args.update[0]))
# Print all records
2016-04-24 17:05:46 -05:00
if args.printindex is not None:
if args.printindex < 0:
printmsg("Index must be >= 0", "ERROR")
closequit(conn, 1)
2016-04-24 17:05:46 -05:00
printdb(cur, args.printindex)
2016-04-24 17:52:15 -05:00
# Replace a tag in DB
if args.replace is not None:
2016-04-29 12:29:06 -05:00
if len(args.replace) == 1:
replaceTags(conn, cur, args.replace[0])
2016-04-24 17:52:15 -05:00
else:
2016-04-29 12:29:06 -05:00
replaceTags(conn, cur, args.replace[0], args.replace[1:])
# Open URL in browser
2016-04-25 13:43:28 -05:00
if args.openurl is not None:
if args.openurl < 1:
printmsg("Index must be >= 1", "ERROR")
closequit(conn, 1)
2016-04-25 13:43:28 -05:00
fetchopen(args.openurl)
2016-05-02 12:42:29 -05:00
"""
# NOTE: Insert is functional but commented because DB compaction serves the purpose.
# Insert a record at an index
if args.insert is not None:
if not args.insert[0].isdigit():
printmsg("Index must be a number >= 1", "ERROR")
closequit(conn, 1)
insertindex = int(args.insert[0])
if insertindex < 1:
printmsg("Index must be a number >= 1", "ERROR")
closequit(conn, 1)
if len(args.insert) == 1:
2016-04-25 13:43:28 -05:00
pass # No operation
else:
2016-04-25 11:23:03 -05:00
AddUpdateEntry(conn, cur, args.insert[1:], 0, insertindex)
2016-04-25 12:33:07 -05:00
"""
# Close the connection before exiting
conn.close()