# # Copyright (C) 2006-2010 Nexedi SA # # This program is free software; you can redistribute it and/or # modify it under the terms of the GNU General Public License # as published by the Free Software Foundation; either version 2 # of the License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program; if not, write to the Free Software # Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. from cPickle import dumps, loads from zlib import compress as real_compress, decompress from neo.lib.locking import Empty from random import shuffle import time import os from ZODB.POSException import UndoError, StorageTransactionError, ConflictError from ZODB.POSException import ReadConflictError from ZODB.ConflictResolution import ResolvedSerial from persistent.TimeStamp import TimeStamp import neo.lib from neo.lib.protocol import NodeTypes, Packets, \ INVALID_PARTITION, ZERO_HASH, ZERO_TID from neo.lib.event import EventManager from neo.lib.util import makeChecksum as real_makeChecksum, dump from neo.lib.locking import Lock from neo.lib.connection import MTClientConnection, OnTimeout, ConnectionClosed from neo.lib.node import NodeManager from neo.lib.connector import getConnectorHandler from .exception import NEOStorageError, NEOStorageCreationUndoneError from .exception import NEOStorageNotFoundError from neo.lib.exception import NeoException from .handlers import storage, master from neo.lib.dispatcher import Dispatcher, ForgottenPacket from .poll import ThreadedPoll, psThreadedPoll from .iterator import Iterator from .cache import ClientCache from .pool import ConnectionPool from neo.lib.util import u64, parseMasterList from neo.lib.profiling import profiler_decorator, PROFILING_ENABLED from neo.lib.debug import register as registerLiveDebugger from .container import ThreadContainer, TransactionContainer if PROFILING_ENABLED: # Those functions require a "real" python function wrapper before they can # be decorated. @profiler_decorator def compress(data): return real_compress(data) @profiler_decorator def makeChecksum(data): return real_makeChecksum(data) else: # If profiling is disabled, directly use original functions. compress = real_compress makeChecksum = real_makeChecksum CHECKED_SERIAL = object() class Application(object): """The client node application.""" def __init__(self, master_nodes, name, compress=True, **kw): # Start polling thread self.em = EventManager() self.poll_thread = ThreadedPoll(self.em, name=name) psThreadedPoll() # Internal Attributes common to all thread self._db = None self.name = name master_addresses, connector_name = parseMasterList(master_nodes) self.connector_handler = getConnectorHandler(connector_name) self.dispatcher = Dispatcher(self.poll_thread) self.nm = NodeManager() self.cp = ConnectionPool(self) self.pt = None self.master_conn = None self.primary_master_node = None self.trying_master_node = None # load master node list for address in master_addresses: self.nm.createMaster(address=address) # no self-assigned UUID, primary master will supply us one self.uuid = None self._cache = ClientCache() self.new_oid_list = [] self.last_oid = '\0' * 8 self.storage_event_handler = storage.StorageEventHandler(self) self.storage_bootstrap_handler = storage.StorageBootstrapHandler(self) self.storage_handler = storage.StorageAnswersHandler(self) self.primary_handler = master.PrimaryAnswersHandler(self) self.primary_bootstrap_handler = master.PrimaryBootstrapHandler(self) self.notifications_handler = master.PrimaryNotificationsHandler( self) # Internal attribute distinct between thread self._thread_container = ThreadContainer() self._txn_container = TransactionContainer() # Lock definition : # _load_lock is used to make loading and storing atomic lock = Lock() self._load_lock_acquire = lock.acquire self._load_lock_release = lock.release # _oid_lock is used in order to not call multiple oid # generation at the same time lock = Lock() self._oid_lock_acquire = lock.acquire self._oid_lock_release = lock.release lock = Lock() # _cache_lock is used for the client cache self._cache_lock_acquire = lock.acquire self._cache_lock_release = lock.release lock = Lock() # _connecting_to_master_node is used to prevent simultaneous master # node connection attemps self._connecting_to_master_node_acquire = lock.acquire self._connecting_to_master_node_release = lock.release # _nm ensure exclusive access to the node manager lock = Lock() self._nm_acquire = lock.acquire self._nm_release = lock.release self.compress = compress registerLiveDebugger(on_log=self.log) def getHandlerData(self): return self._thread_container.get()['answer'] def setHandlerData(self, data): self._thread_container.get()['answer'] = data def _getThreadQueue(self): return self._thread_container.get()['queue'] def log(self): self.em.log() self.nm.log() if self.pt is not None: self.pt.log() @profiler_decorator def _handlePacket(self, conn, packet, handler=None): """ conn The connection which received the packet (forwarded to handler). packet The packet to handle. handler The handler to use to handle packet. If not given, it will be guessed from connection's not type. """ if handler is None: # Guess the handler to use based on the type of node on the # connection node = self.nm.getByAddress(conn.getAddress()) if node is None: raise ValueError, 'Expecting an answer from a node ' \ 'which type is not known... Is this right ?' if node.isStorage(): handler = self.storage_handler elif node.isMaster(): handler = self.primary_handler else: raise ValueError, 'Unknown node type: %r' % (node.__class__, ) conn.lock() try: handler.dispatch(conn, packet) finally: conn.unlock() @profiler_decorator def _waitAnyMessage(self, queue, block=True): """ Handle all pending packets. block If True (default), will block until at least one packet was received. """ pending = self.dispatcher.pending get = queue.get _handlePacket = self._handlePacket while pending(queue): try: conn, packet = get(block) except Empty: break if packet is None or isinstance(packet, ForgottenPacket): # connection was closed or some packet was forgotten continue block = False try: _handlePacket(conn, packet) except ConnectionClosed: pass def _waitAnyTransactionMessage(self, txn_context, block=True): """ Just like _waitAnyMessage, but for per-transaction exchanges, rather than per-thread. """ queue = txn_context['queue'] self.setHandlerData(txn_context) try: self._waitAnyMessage(queue, block=block) finally: # Don't leave access to thread context, even if a raise happens. self.setHandlerData(None) @profiler_decorator def _ask(self, conn, packet, handler=None): self.setHandlerData(None) queue = self._getThreadQueue() msg_id = conn.ask(packet, queue=queue) get = queue.get _handlePacket = self._handlePacket while True: qconn, qpacket = get(True) is_forgotten = isinstance(qpacket, ForgottenPacket) if conn is qconn: # check fake packet if qpacket is None: raise ConnectionClosed if msg_id == qpacket.getId(): if is_forgotten: raise ValueError, 'ForgottenPacket for an ' \ 'explicitely expected packet.' _handlePacket(qconn, qpacket, handler=handler) break if not is_forgotten and qpacket is not None: _handlePacket(qconn, qpacket) return self.getHandlerData() @profiler_decorator def _askStorage(self, conn, packet): """ Send a request to a storage node and process its answer """ return self._ask(conn, packet, handler=self.storage_handler) @profiler_decorator def _askPrimary(self, packet): """ Send a request to the primary master and process its answer """ return self._ask(self._getMasterConnection(), packet, handler=self.primary_handler) @profiler_decorator def _getMasterConnection(self): """ Connect to the primary master node on demand """ # acquire the lock to allow only one thread to connect to the primary result = self.master_conn if result is None: self._connecting_to_master_node_acquire() try: self.new_oid_list = [] result = self._connectToPrimaryNode() self.master_conn = result finally: self._connecting_to_master_node_release() return result def getPartitionTable(self): """ Return the partition table manager, reconnect the PMN if needed """ # this ensure the master connection is established and the partition # table is up to date. self._getMasterConnection() return self.pt @profiler_decorator def _connectToPrimaryNode(self): """ Lookup for the current primary master node """ neo.lib.logging.debug('connecting to primary master...') ready = False nm = self.nm packet = Packets.AskPrimary() while not ready: # Get network connection to primary master index = 0 connected = False while not connected: if self.primary_master_node is not None: # If I know a primary master node, pinpoint it. self.trying_master_node = self.primary_master_node self.primary_master_node = None else: # Otherwise, check one by one. master_list = nm.getMasterList() try: self.trying_master_node = master_list[index] except IndexError: time.sleep(1) index = 0 self.trying_master_node = master_list[0] index += 1 # Connect to master conn = MTClientConnection(self.em, self.notifications_handler, addr=self.trying_master_node.getAddress(), connector=self.connector_handler(), dispatcher=self.dispatcher) # Query for primary master node if conn.getConnector() is None: # This happens if a connection could not be established. neo.lib.logging.error( 'Connection to master node %s failed', self.trying_master_node) continue try: self._ask(conn, packet, handler=self.primary_bootstrap_handler) except ConnectionClosed: continue # If we reached the primary master node, mark as connected connected = self.primary_master_node is not None and \ self.primary_master_node is self.trying_master_node neo.lib.logging.info( 'Connected to %s' % (self.primary_master_node, )) try: ready = self.identifyToPrimaryNode(conn) except ConnectionClosed: neo.lib.logging.error('Connection to %s lost', self.trying_master_node) self.primary_master_node = None neo.lib.logging.info("Connected and ready") return conn def identifyToPrimaryNode(self, conn): """ Request identification and required informations to be operational. Might raise ConnectionClosed so that the new primary can be looked-up again. """ neo.lib.logging.info('Initializing from master') ask = self._ask handler = self.primary_bootstrap_handler # Identify to primary master and request initial data p = Packets.RequestIdentification(NodeTypes.CLIENT, self.uuid, None, self.name) while conn.getUUID() is None: ask(conn, p, handler=handler) if conn.getUUID() is None: # Node identification was refused by master, it is considered # as the primary as long as we are connected to it. time.sleep(1) ask(conn, Packets.AskNodeInformation(), handler=handler) ask(conn, Packets.AskPartitionTable(), handler=handler) return self.pt.operational() def registerDB(self, db, limit): self._db = db def getDB(self): return self._db @profiler_decorator def new_oid(self): """Get a new OID.""" self._oid_lock_acquire() try: if len(self.new_oid_list) == 0: # Get new oid list from master node # we manage a list of oid here to prevent # from asking too many time new oid one by one # from master node self._askPrimary(Packets.AskNewOIDs(100)) if len(self.new_oid_list) <= 0: raise NEOStorageError('new_oid failed') self.last_oid = self.new_oid_list.pop(0) return self.last_oid finally: self._oid_lock_release() def getStorageSize(self): # return the last OID used, this is innacurate return int(u64(self.last_oid)) @profiler_decorator def load(self, oid, tid=None, before_tid=None): """ Internal method which manage load, loadSerial and loadBefore. OID and TID (serial) parameters are expected packed. oid OID of object to get. tid If given, the exact serial at which OID is desired. before_tid should be None. before_tid If given, the excluded upper bound serial at which OID is desired. serial should be None. Return value: (3-tuple) - Object data (None if object creation was undone). - Serial of given data. - Next serial at which object exists, or None. Only set when tid parameter is not None. Exceptions: NEOStorageError technical problem NEOStorageNotFoundError object exists but no data satisfies given parameters NEOStorageDoesNotExistError object doesn't exist NEOStorageCreationUndoneError object existed, but its creation was undone Note that loadSerial is used during conflict resolution to load object's current version, which is not visible to us normaly (it was committed after our snapshot was taken). """ # TODO: # - rename parameters (here? and in handlers & packet definitions) self._load_lock_acquire() try: result = self._loadFromCache(oid, tid, before_tid) if not result: result = self._loadFromStorage(oid, tid, before_tid) self._cache_lock_acquire() try: self._cache.store(oid, *result) finally: self._cache_lock_release() return result finally: self._load_lock_release() @profiler_decorator def _loadFromStorage(self, oid, at_tid, before_tid): packet = Packets.AskObject(oid, at_tid, before_tid) for node, conn in self.cp.iterateForObject(oid, readable=True): try: noid, tid, next_tid, compression, checksum, data \ = self._askStorage(conn, packet) except ConnectionClosed: continue if data or checksum != ZERO_HASH: if checksum != makeChecksum(data): neo.lib.logging.error('wrong checksum from %s for oid %s', conn, dump(oid)) continue if compression: data = decompress(data) return data, tid, next_tid raise NEOStorageCreationUndoneError(dump(oid)) # We didn't got any object from all storage node because of # connection error raise NEOStorageError('connection failure') @profiler_decorator def _loadFromCache(self, oid, at_tid=None, before_tid=None): """ Load from local cache, return None if not found. """ self._cache_lock_acquire() try: if at_tid: result = self._cache.load(oid, at_tid + '*') assert not result or result[1] == at_tid return result return self._cache.load(oid, before_tid) finally: self._cache_lock_release() @profiler_decorator def tpc_begin(self, transaction, tid=None, status=' '): """Begin a new transaction.""" txn_container = self._txn_container # First get a transaction, only one is allowed at a time if txn_container.get(transaction) is not None: # We already begin the same transaction raise StorageTransactionError('Duplicate tpc_begin calls') txn_context = txn_container.new(transaction) # use the given TID or request a new one to the master answer_ttid = self._askPrimary(Packets.AskBeginTransaction(tid)) if answer_ttid is None: raise NEOStorageError('tpc_begin failed') assert tid in (None, answer_ttid), (tid, answer_ttid) txn_context['txn'] = transaction txn_context['ttid'] = answer_ttid @profiler_decorator def store(self, oid, serial, data, version, transaction): """Store object.""" txn_context = self._txn_container.get(transaction) if txn_context is None: raise StorageTransactionError(self, transaction) neo.lib.logging.debug( 'storing oid %s serial %s', dump(oid), dump(serial)) self._store(txn_context, oid, serial, data) return None def _store(self, txn_context, oid, serial, data, data_serial=None, unlock=False): ttid = txn_context['ttid'] if data is None: # This is some undo: either a no-data object (undoing object # creation) or a back-pointer to an earlier revision (going back to # an older object revision). compressed_data = '' compression = 0 checksum = ZERO_HASH else: assert data_serial is None compression = self.compress compressed_data = data if self.compress: compressed_data = compress(data) if len(compressed_data) > len(data): compressed_data = data compression = 0 else: compression = 1 checksum = makeChecksum(compressed_data) on_timeout = OnTimeout(self.onStoreTimeout, txn_context, oid) # Store object in tmp cache data_dict = txn_context['data_dict'] if oid not in data_dict: txn_context['data_list'].append(oid) data_dict[oid] = data # Store data on each node txn_context['object_stored_counter_dict'][oid] = {} object_base_serial_dict = txn_context['object_base_serial_dict'] if oid not in object_base_serial_dict: object_base_serial_dict[oid] = serial txn_context['object_serial_dict'][oid] = serial queue = txn_context['queue'] involved_nodes = txn_context['involved_nodes'] add_involved_nodes = involved_nodes.add packet = Packets.AskStoreObject(oid, serial, compression, checksum, compressed_data, data_serial, ttid, unlock) for node, conn in self.cp.iterateForObject(oid, writable=True): try: conn.ask(packet, on_timeout=on_timeout, queue=queue) add_involved_nodes(node) except ConnectionClosed: continue if not involved_nodes: raise NEOStorageError("Store failed") self._waitAnyTransactionMessage(txn_context, False) def onStoreTimeout(self, conn, msg_id, txn_context, oid): # NOTE: this method is called from poll thread, don't use # thread-specific value ! txn_context.setdefault('timeout_dict', {})[oid] = msg_id # Ask the storage if someone locks the object. # By sending a message with a smaller timeout, # the connection will be kept open. conn.ask(Packets.AskHasLock(txn_context['ttid'], oid), timeout=5, queue=txn_context['queue']) @profiler_decorator def _handleConflicts(self, txn_context, tryToResolveConflict): result = [] append = result.append # Check for conflicts data_dict = txn_context['data_dict'] object_base_serial_dict = txn_context['object_base_serial_dict'] object_serial_dict = txn_context['object_serial_dict'] conflict_serial_dict = txn_context['conflict_serial_dict'].copy() txn_context['conflict_serial_dict'].clear() resolved_conflict_serial_dict = txn_context[ 'resolved_conflict_serial_dict'] for oid, conflict_serial_set in conflict_serial_dict.iteritems(): conflict_serial = max(conflict_serial_set) serial = object_serial_dict[oid] data = data_dict[oid] if ZERO_TID in conflict_serial_set: if 1: # XXX: disable deadlock avoidance code until it is fixed neo.lib.logging.info('Deadlock avoidance on %r:%r', dump(oid), dump(serial)) else: # Storage refused us from taking object lock, to avoid a # possible deadlock. TID is actually used for some kind of # "locking priority": when a higher value has the lock, # this means we stored objects "too late", and we would # otherwise cause a deadlock. # To recover, we must ask storages to release locks we # hold (to let possibly-competing transactions acquire # them), and requeue our already-sent store requests. # XXX: currently, brute-force is implemented: we send # object data again. neo.lib.logging.info('Deadlock avoidance triggered on %r:%r', dump(oid), dump(serial)) for store_oid, store_data in data_dict.iteritems(): store_serial = object_serial_dict[store_oid] if store_data is CHECKED_SERIAL: self._checkCurrentSerialInTransaction(txn_context, store_oid, store_serial) else: if store_data is None: # Some undo neo.lib.logging.warning('Deadlock avoidance cannot' ' reliably work with undo, this must be ' 'implemented.') conflict_serial = ZERO_TID break self._store(txn_context, store_oid, store_serial, store_data, unlock=True) else: continue elif data is not CHECKED_SERIAL: resolved_serial_set = resolved_conflict_serial_dict.setdefault( oid, set()) if resolved_serial_set and conflict_serial <= max( resolved_serial_set): # A later serial has already been resolved, skip. resolved_serial_set.update(conflict_serial_set) continue new_data = tryToResolveConflict(oid, conflict_serial, serial, data) if new_data is not None: neo.lib.logging.info('Conflict resolution succeed for ' \ '%r:%r with %r', dump(oid), dump(serial), dump(conflict_serial)) # Mark this conflict as resolved resolved_serial_set.update(conflict_serial_set) # Base serial changes too, as we resolved a conflict object_base_serial_dict[oid] = conflict_serial # Try to store again self._store(txn_context, oid, conflict_serial, new_data) append(oid) continue else: neo.lib.logging.info('Conflict resolution failed for ' \ '%r:%r with %r', dump(oid), dump(serial), dump(conflict_serial)) # XXX: Is it really required to remove from data_dict ? del data_dict[oid] txn_context['data_list'].remove(oid) if data is CHECKED_SERIAL: raise ReadConflictError(oid=oid, serials=(conflict_serial, serial)) raise ConflictError(oid=oid, serials=(txn_context['ttid'], serial), data=data) return result @profiler_decorator def waitResponses(self, queue, handler_data): """Wait for all requests to be answered (or their connection to be detected as closed)""" pending = self.dispatcher.pending _waitAnyMessage = self._waitAnyMessage self.setHandlerData(handler_data) while pending(queue): _waitAnyMessage(queue) @profiler_decorator def waitStoreResponses(self, txn_context, tryToResolveConflict): result = [] append = result.append resolved_oid_set = set() update = resolved_oid_set.update ttid = txn_context['ttid'] _handleConflicts = self._handleConflicts queue = txn_context['queue'] conflict_serial_dict = txn_context['conflict_serial_dict'] pending = self.dispatcher.pending _waitAnyTransactionMessage = self._waitAnyTransactionMessage while pending(queue) or conflict_serial_dict: # Note: handler data can be overwritten by _handleConflicts # so we must set it for each iteration. _waitAnyTransactionMessage(txn_context) if conflict_serial_dict: conflicts = _handleConflicts(txn_context, tryToResolveConflict) if conflicts: update(conflicts) # Check for never-stored objects, and update result for all others for oid, store_dict in \ txn_context['object_stored_counter_dict'].iteritems(): if not store_dict: neo.lib.logging.error('tpc_store failed') raise NEOStorageError('tpc_store failed') elif oid in resolved_oid_set: append((oid, ResolvedSerial)) return result @profiler_decorator def tpc_vote(self, transaction, tryToResolveConflict): """Store current transaction.""" txn_context = self._txn_container.get(transaction) if txn_context is None or transaction is not txn_context['txn']: raise StorageTransactionError(self, transaction) result = self.waitStoreResponses(txn_context, tryToResolveConflict) ttid = txn_context['ttid'] # Store data on each node txn_stored_counter = 0 packet = Packets.AskStoreTransaction(ttid, str(transaction.user), str(transaction.description), dumps(transaction._extension), txn_context['data_list']) add_involved_nodes = txn_context['involved_nodes'].add for node, conn in self.cp.iterateForObject(ttid, writable=True): neo.lib.logging.debug("voting object %s on %s", dump(ttid), dump(conn.getUUID())) try: self._askStorage(conn, packet) except ConnectionClosed: continue add_involved_nodes(node) txn_stored_counter += 1 # check at least one storage node accepted if txn_stored_counter == 0: neo.lib.logging.error('tpc_vote failed') raise NEOStorageError('tpc_vote failed') # Check if master connection is still alive. # This is just here to lower the probability of detecting a problem # in tpc_finish, as we should do our best to detect problem before # tpc_finish. self._getMasterConnection() txn_context['txn_voted'] = True return result @profiler_decorator def tpc_abort(self, transaction): """Abort current transaction.""" txn_container = self._txn_container txn_context = txn_container.get(transaction) if txn_context is None: return ttid = txn_context['ttid'] p = Packets.AbortTransaction(ttid) getConnForNode = self.cp.getConnForNode # cancel transaction one all those nodes for node in txn_context['involved_nodes']: conn = getConnForNode(node) if conn is None: continue try: conn.notify(p) except: neo.lib.logging.error( 'Exception in tpc_abort while notifying' \ 'storage node %r of abortion, ignoring.', conn, exc_info=1) self._getMasterConnection().notify(p) queue = txn_context['queue'] # We don't need to flush queue, as it won't be reused by future # transactions (deleted on next line & indexed by transaction object # instance). self.dispatcher.forget_queue(queue, flush_queue=False) txn_container.delete(transaction) @profiler_decorator def tpc_finish(self, transaction, tryToResolveConflict, f=None): """Finish current transaction.""" txn_container = self._txn_container txn_context = txn_container.get(transaction) if txn_context is None: raise StorageTransactionError('tpc_finish called for wrong ' 'transaction') if not txn_context['txn_voted']: self.tpc_vote(transaction, tryToResolveConflict) self._load_lock_acquire() try: # Call finish on master oid_list = txn_context['data_list'] p = Packets.AskFinishTransaction(txn_context['ttid'], oid_list) tid = self._askPrimary(p) # Call function given by ZODB if f is not None: f(tid) # Update cache self._cache_lock_acquire() try: cache = self._cache for oid, data in txn_context['data_dict'].iteritems(): if data is CHECKED_SERIAL: # this is just a remain of # checkCurrentSerialInTransaction call, ignore (no data # was modified). continue # Update ex-latest value in cache cache.invalidate(oid, tid) if data is not None: # Store in cache with no next_tid cache.store(oid, data, tid, None) finally: self._cache_lock_release() txn_container.delete(transaction) return tid finally: self._load_lock_release() def undo(self, snapshot_tid, undone_tid, txn, tryToResolveConflict): txn_context = self._txn_container.get(txn) if txn_context is None: raise StorageTransactionError(self, undone_tid) txn_info, txn_ext = self._getTransactionInformation(undone_tid) txn_oid_list = txn_info['oids'] # Regroup objects per partition, to ask a minimum set of storage. partition_oid_dict = {} pt = self.getPartitionTable() for oid in txn_oid_list: partition = pt.getPartition(oid) try: oid_list = partition_oid_dict[partition] except KeyError: oid_list = partition_oid_dict[partition] = [] oid_list.append(oid) # Ask storage the undo serial (serial at which object's previous data # is) getCellList = pt.getCellList getCellSortKey = self.cp.getCellSortKey getConnForCell = self.cp.getConnForCell queue = self._getThreadQueue() ttid = txn_context['ttid'] for partition, oid_list in partition_oid_dict.iteritems(): cell_list = getCellList(partition, readable=True) # We do want to shuffle before getting one with the smallest # key, so that all cells with the same (smallest) key has # identical chance to be chosen. shuffle(cell_list) storage_conn = getConnForCell(min(cell_list, key=getCellSortKey)) storage_conn.ask(Packets.AskObjectUndoSerial(ttid, snapshot_tid, undone_tid, oid_list), queue=queue) # Wait for all AnswerObjectUndoSerial. We might get OidNotFoundError, # meaning that objects in transaction's oid_list do not exist any # longer. This is the symptom of a pack, so forbid undoing transaction # when it happens. undo_object_tid_dict = {} try: self.waitResponses(queue, undo_object_tid_dict) except NEOStorageNotFoundError: self.dispatcher.forget_queue(queue) raise UndoError('non-undoable transaction') # Send undo data to all storage nodes. for oid in txn_oid_list: current_serial, undo_serial, is_current = undo_object_tid_dict[oid] if is_current: data = None else: # Serial being undone is not the latest version for this # object. This is an undo conflict, try to resolve it. try: # Load the latest version we are supposed to see data = self.load(oid, current_serial)[0] # Load the version we were undoing to undo_data = self.load(oid, undo_serial)[0] except NEOStorageNotFoundError: raise UndoError('Object not found while resolving undo ' 'conflict') # Resolve conflict try: data = tryToResolveConflict(oid, current_serial, undone_tid, undo_data, data) except ConflictError: data = None if data is None: raise UndoError('Some data were modified by a later ' \ 'transaction', oid) undo_serial = None self._store(txn_context, oid, current_serial, data, undo_serial) return None, txn_oid_list def _insertMetadata(self, txn_info, extension): for k, v in loads(extension).items(): txn_info[k] = v def _getTransactionInformation(self, tid): packet = Packets.AskTransactionInformation(tid) for node, conn in self.cp.iterateForObject(tid, readable=True): try: txn_info, txn_ext = self._askStorage(conn, packet) except ConnectionClosed: continue except NEOStorageNotFoundError: # TID not found continue break else: raise NEOStorageError('Transaction %r not found' % (tid, )) return (txn_info, txn_ext) # XXX: The following 2 methods fail when they reconnect to a storage after # they already sent a request to a previous storage. # See also testStorageReconnectDuringXxx def undoLog(self, first, last, filter=None, block=0): # XXX: undoLog is broken if last < 0: # See FileStorage.py for explanation last = first - last # First get a list of transactions from all storage nodes. # Each storage node will return TIDs only for UP_TO_DATE state and # FEEDING state cells pt = self.getPartitionTable() storage_node_list = pt.getNodeList() queue = self._getThreadQueue() packet = Packets.AskTIDs(first, last, INVALID_PARTITION) for storage_node in storage_node_list: conn = self.cp.getConnForNode(storage_node) if conn is None: continue conn.ask(packet, queue=queue) # Wait for answers from all storages. tid_set = set() self.waitResponses(queue, tid_set) # Reorder tids ordered_tids = sorted(tid_set, reverse=True) neo.lib.logging.debug("UndoLog tids %s", map(dump, ordered_tids)) # For each transaction, get info undo_info = [] append = undo_info.append for tid in ordered_tids: (txn_info, txn_ext) = self._getTransactionInformation(tid) if filter is None or filter(txn_info): txn_info.pop('packed') txn_info.pop("oids") self._insertMetadata(txn_info, txn_ext) append(txn_info) if len(undo_info) >= last - first: break # Check we return at least one element, otherwise call # again but extend offset if len(undo_info) == 0 and not block: undo_info = self.undoLog(first=first, last=last*5, filter=filter, block=1) return undo_info def transactionLog(self, start, stop, limit): node_map = self.pt.getNodeMap() node_list = node_map.keys() node_list.sort(key=self.cp.getCellSortKey) partition_set = set(range(self.pt.getPartitions())) queue = self._getThreadQueue() # request a tid list for each partition for node in node_list: conn = self.cp.getConnForNode(node) request_set = set(node_map[node]) & partition_set if conn is None or not request_set: continue partition_set -= set(request_set) packet = Packets.AskTIDsFrom(start, stop, limit, request_set) conn.ask(packet, queue=queue) if not partition_set: break assert not partition_set tid_set = set() self.waitResponses(queue, tid_set) # request transactions informations txn_list = [] append = txn_list.append tid = None for tid in sorted(tid_set): (txn_info, txn_ext) = self._getTransactionInformation(tid) txn_info['ext'] = loads(txn_ext) append(txn_info) return (tid, txn_list) def history(self, oid, size=1, filter=None): # Get history informations for object first packet = Packets.AskObjectHistory(oid, 0, size) for node, conn in self.cp.iterateForObject(oid, readable=True): try: history_list = self._askStorage(conn, packet) except ConnectionClosed: continue # Now that we have object informations, get txn informations result = [] # history_list is already sorted descending (by the storage) for serial, size in history_list: txn_info, txn_ext = self._getTransactionInformation(serial) # create history dict txn_info.pop('id') txn_info.pop('oids') txn_info.pop('packed') txn_info['tid'] = serial txn_info['version'] = '' txn_info['size'] = size if filter is None or filter(txn_info): result.append(txn_info) self._insertMetadata(txn_info, txn_ext) return result @profiler_decorator def importFrom(self, source, start, stop, tryToResolveConflict): serials = {} transaction_iter = source.iterator(start, stop) for transaction in transaction_iter: tid = transaction.tid self.tpc_begin(transaction, tid, transaction.status) for r in transaction: oid = r.oid pre = serials.get(oid, None) # TODO: bypass conflict resolution, locks... self.store(oid, pre, r.data, r.version, transaction) serials[oid] = tid conflicted = self.tpc_vote(transaction, tryToResolveConflict) assert not conflicted, conflicted real_tid = self.tpc_finish(transaction, tryToResolveConflict) assert real_tid == tid, (real_tid, tid) transaction_iter.close() def iterator(self, start, stop): if start is None: start = ZERO_TID return Iterator(self, start, stop) def lastTransaction(self): return self._askPrimary(Packets.AskLastTransaction()) def abortVersion(self, src, transaction): if self._txn_container.get(transaction) is None: raise StorageTransactionError(self, transaction) return '', [] def commitVersion(self, src, dest, transaction): if self._txn_container.get(transaction) is None: raise StorageTransactionError(self, transaction) return '', [] def __del__(self): """Clear all connection.""" # Due to bug in ZODB, close is not always called when shutting # down zope, so use __del__ to close connections for conn in self.em.getConnectionList(): conn.close() self.cp.flush() self.master_conn = None # Stop polling thread neo.lib.logging.debug('Stopping %s', self.poll_thread) self.poll_thread.stop() psThreadedPoll() close = __del__ def pack(self, t): tid = repr(TimeStamp(*time.gmtime(t)[:5] + (t % 60, ))) if tid == ZERO_TID: raise NEOStorageError('Invalid pack time') self._askPrimary(Packets.AskPack(tid)) # XXX: this is only needed to make ZODB unit tests pass. # It should not be otherwise required (clients should be free to load # old data as long as it is available in cache, event if it was pruned # by a pack), so don't bother invalidating on other clients. self._cache_lock_acquire() try: self._cache.clear() finally: self._cache_lock_release() def getLastTID(self, oid): return self.load(oid)[1] def checkCurrentSerialInTransaction(self, oid, serial, transaction): txn_context = self._txn_container.get(transaction) if txn_context is None: raise StorageTransactionError(self, transaction) self._checkCurrentSerialInTransaction(txn_context, oid, serial) def _checkCurrentSerialInTransaction(self, txn_context, oid, serial): ttid = txn_context['ttid'] txn_context['object_serial_dict'][oid] = serial # Placeholders queue = txn_context['queue'] txn_context['object_stored_counter_dict'][oid] = {} data_dict = txn_context['data_dict'] if oid not in data_dict: # Marker value so we don't try to resolve conflicts. data_dict[oid] = CHECKED_SERIAL txn_context['data_list'].append(oid) packet = Packets.AskCheckCurrentSerial(ttid, serial, oid) for node, conn in self.cp.iterateForObject(oid, writable=True): try: conn.ask(packet, queue=queue) except ConnectionClosed: continue self._waitAnyTransactionMessage(txn_context, False)