Mercurial > hg
view hgext/infinitepush/sqlindexapi.py @ 39772:ae531f5e583c
testing: add interface unit tests for file storage
Our strategy for supporting alternate storage backends is to define
interfaces for everything then "code to the interface."
We already have interfaces for various primitives, including file
and manifest storage.
What we don't have is generic unit tests for those interfaces. Up
to this point we've been relying on high-level integration tests
(mainly in the form of existing .t tests) to test alternate storage
backends. And my experience with developing the "simple store" test
extension is that such testing is very tedious: it takes several
minutes to run all tests and when you find a failure, it is often
non-trivial to debug.
This commit starts to change that.
This commit introduces the mercurial.testing.storage module. It
contains testing code for storage. Currently, it defines some
unittest.TestCase classes for testing the file storage interfaces.
It also defines some factory functions that allow a caller to easily
spawn a custom TestCase "bound" to a specific file storage backend
implementation.
A new .py test has been added. It simply defines a callable to produce
filelog and transaction instances on demand and then "registers" the
various test classes so the filelog class can be tested with the
storage interface unit tests.
As part of writing the tests, I identified a couple of apparent
bugs in revlog.py and filelog.py! These are tracked with inline
TODO comments.
Writing the tests makes it more obvious where the storage interface
is lacking. For example, we raise either IndexError or
error.LookupError for missing revisions depending on whether we
use an integer revision or a node. Also, we raise error.RevlogError
in various places when we should be raising a storage-agnostic
error type.
The storage interfaces are currently far from perfect and there is much
work to be done to improve them. But at least with this commit we
finally have the start of unit tests that can be used to "qualify"
the behavior of a storage backend. And when implementing and debugging
new storage backends, we now have an obvious place to define new
tests and have obvious places to insert breakpoints to facilitate
debugging. This should be invaluable when implementing new storage
backends.
I added the mercurial.testing package because these interface
conformance tests are generic and need to be usable by all storage
backends. Having the code live in tests/ would make it difficult for
storage backends implemented in extensions to test their interface
conformance. First, it would require obtaining a copy of Mercurial's
storage test code in order to test. Second, it would make testing
against multiple Mercurial versions difficult, as you would need to
import N copies of the storage testing code in order to achieve test
coverage. By making the test code part of the Mercurial distribution
itself, extensions can `import mercurial.testing.*` to access and run
the test code. The test will run against whatever Mercurial version
is active.
FWIW I've always wanted to move parts of run-tests.py into the
mercurial.* package to make the testing story simpler (e.g. imagine an
`hg debugruntests` command that could invoke the test harness). While I
have no plans to do that in the near future, establishing the
mercurial.testing package does provide a natural home for that code
should someone do this in the future.
Differential Revision: https://phab.mercurial-scm.org/D4650
author | Gregory Szorc <gregory.szorc@gmail.com> |
---|---|
date | Tue, 18 Sep 2018 16:52:11 -0700 |
parents | c1fac3878196 |
children | 2372284d9457 |
line wrap: on
line source
# Infinite push # # Copyright 2016 Facebook, Inc. # # This software may be used and distributed according to the terms of the # GNU General Public License version 2 or any later version. from __future__ import absolute_import import logging import os import time import warnings import mysql.connector from . import indexapi def _convertbookmarkpattern(pattern): pattern = pattern.replace('_', '\\_') pattern = pattern.replace('%', '\\%') if pattern.endswith('*'): pattern = pattern[:-1] + '%' return pattern class sqlindexapi(indexapi.indexapi): ''' Sql backend for infinitepush index. See schema.sql ''' def __init__(self, reponame, host, port, database, user, password, logfile, loglevel, waittimeout=300, locktimeout=120): super(sqlindexapi, self).__init__() self.reponame = reponame self.sqlargs = { 'host': host, 'port': port, 'database': database, 'user': user, 'password': password, } self.sqlconn = None self.sqlcursor = None if not logfile: logfile = os.devnull logging.basicConfig(filename=logfile) self.log = logging.getLogger() self.log.setLevel(loglevel) self._connected = False self._waittimeout = waittimeout self._locktimeout = locktimeout def sqlconnect(self): if self.sqlconn: raise indexapi.indexexception("SQL connection already open") if self.sqlcursor: raise indexapi.indexexception("SQL cursor already open without" " connection") retry = 3 while True: try: self.sqlconn = mysql.connector.connect(**self.sqlargs) # Code is copy-pasted from hgsql. Bug fixes need to be # back-ported! # The default behavior is to return byte arrays, when we # need strings. This custom convert returns strings. self.sqlconn.set_converter_class(CustomConverter) self.sqlconn.autocommit = False break except mysql.connector.errors.Error: # mysql can be flakey occasionally, so do some minimal # retrying. retry -= 1 if retry == 0: raise time.sleep(0.2) waittimeout = self.sqlconn.converter.escape('%s' % self._waittimeout) self.sqlcursor = self.sqlconn.cursor() self.sqlcursor.execute("SET wait_timeout=%s" % waittimeout) self.sqlcursor.execute("SET innodb_lock_wait_timeout=%s" % self._locktimeout) self._connected = True def close(self): """Cleans up the metadata store connection.""" with warnings.catch_warnings(): warnings.simplefilter("ignore") self.sqlcursor.close() self.sqlconn.close() self.sqlcursor = None self.sqlconn = None def __enter__(self): if not self._connected: self.sqlconnect() return self def __exit__(self, exc_type, exc_val, exc_tb): if exc_type is None: self.sqlconn.commit() else: self.sqlconn.rollback() def addbundle(self, bundleid, nodesctx): if not self._connected: self.sqlconnect() self.log.info("ADD BUNDLE %r %r" % (self.reponame, bundleid)) self.sqlcursor.execute( "INSERT INTO bundles(bundle, reponame) VALUES " "(%s, %s)", params=(bundleid, self.reponame)) for ctx in nodesctx: self.sqlcursor.execute( "INSERT INTO nodestobundle(node, bundle, reponame) " "VALUES (%s, %s, %s) ON DUPLICATE KEY UPDATE " "bundle=VALUES(bundle)", params=(ctx.hex(), bundleid, self.reponame)) extra = ctx.extra() author_name = ctx.user() committer_name = extra.get('committer', ctx.user()) author_date = int(ctx.date()[0]) committer_date = int(extra.get('committer_date', author_date)) self.sqlcursor.execute( "INSERT IGNORE INTO nodesmetadata(node, message, p1, p2, " "author, committer, author_date, committer_date, " "reponame) VALUES " "(%s, %s, %s, %s, %s, %s, %s, %s, %s)", params=(ctx.hex(), ctx.description(), ctx.p1().hex(), ctx.p2().hex(), author_name, committer_name, author_date, committer_date, self.reponame) ) def addbookmark(self, bookmark, node): """Takes a bookmark name and hash, and records mapping in the metadata store.""" if not self._connected: self.sqlconnect() self.log.info( "ADD BOOKMARKS %r bookmark: %r node: %r" % (self.reponame, bookmark, node)) self.sqlcursor.execute( "INSERT INTO bookmarkstonode(bookmark, node, reponame) " "VALUES (%s, %s, %s) ON DUPLICATE KEY UPDATE node=VALUES(node)", params=(bookmark, node, self.reponame)) def addmanybookmarks(self, bookmarks): if not self._connected: self.sqlconnect() args = [] values = [] for bookmark, node in bookmarks.iteritems(): args.append('(%s, %s, %s)') values.extend((bookmark, node, self.reponame)) args = ','.join(args) self.sqlcursor.execute( "INSERT INTO bookmarkstonode(bookmark, node, reponame) " "VALUES %s ON DUPLICATE KEY UPDATE node=VALUES(node)" % args, params=values) def deletebookmarks(self, patterns): """Accepts list of bookmark patterns and deletes them. If `commit` is set then bookmark will actually be deleted. Otherwise deletion will be delayed until the end of transaction. """ if not self._connected: self.sqlconnect() self.log.info("DELETE BOOKMARKS: %s" % patterns) for pattern in patterns: pattern = _convertbookmarkpattern(pattern) self.sqlcursor.execute( "DELETE from bookmarkstonode WHERE bookmark LIKE (%s) " "and reponame = %s", params=(pattern, self.reponame)) def getbundle(self, node): """Returns the bundleid for the bundle that contains the given node.""" if not self._connected: self.sqlconnect() self.log.info("GET BUNDLE %r %r" % (self.reponame, node)) self.sqlcursor.execute( "SELECT bundle from nodestobundle " "WHERE node = %s AND reponame = %s", params=(node, self.reponame)) result = self.sqlcursor.fetchall() if len(result) != 1 or len(result[0]) != 1: self.log.info("No matching node") return None bundle = result[0][0] self.log.info("Found bundle %r" % bundle) return bundle def getnode(self, bookmark): """Returns the node for the given bookmark. None if it doesn't exist.""" if not self._connected: self.sqlconnect() self.log.info( "GET NODE reponame: %r bookmark: %r" % (self.reponame, bookmark)) self.sqlcursor.execute( "SELECT node from bookmarkstonode WHERE " "bookmark = %s AND reponame = %s", params=(bookmark, self.reponame)) result = self.sqlcursor.fetchall() if len(result) != 1 or len(result[0]) != 1: self.log.info("No matching bookmark") return None node = result[0][0] self.log.info("Found node %r" % node) return node def getbookmarks(self, query): if not self._connected: self.sqlconnect() self.log.info( "QUERY BOOKMARKS reponame: %r query: %r" % (self.reponame, query)) query = _convertbookmarkpattern(query) self.sqlcursor.execute( "SELECT bookmark, node from bookmarkstonode WHERE " "reponame = %s AND bookmark LIKE %s", params=(self.reponame, query)) result = self.sqlcursor.fetchall() bookmarks = {} for row in result: if len(row) != 2: self.log.info("Bad row returned: %s" % row) continue bookmarks[row[0]] = row[1] return bookmarks def saveoptionaljsonmetadata(self, node, jsonmetadata): if not self._connected: self.sqlconnect() self.log.info( ("INSERT METADATA, QUERY BOOKMARKS reponame: %r " + "node: %r, jsonmetadata: %s") % (self.reponame, node, jsonmetadata)) self.sqlcursor.execute( "UPDATE nodesmetadata SET optional_json_metadata=%s WHERE " "reponame=%s AND node=%s", params=(jsonmetadata, self.reponame, node)) class CustomConverter(mysql.connector.conversion.MySQLConverter): """Ensure that all values being returned are returned as python string (versus the default byte arrays).""" def _STRING_to_python(self, value, dsc=None): return str(value) def _VAR_STRING_to_python(self, value, dsc=None): return str(value) def _BLOB_to_python(self, value, dsc=None): return str(value)