1. 13 Dec, 2023 2 commits
    • Kirill Smelkov's avatar
      app: Remember SSL credentials so that it is possible to retrieve them · 6e19ba3b
      Kirill Smelkov authored
      Unfortunately after creating SSL context it is not possible, or at least
      I could not find how, to retrieve original credentials with which the
      context was created. However wendelin.core needs to be able to take a
      client storage, reconstruct zurl to refer to that particular storage,
      and pass that zurl to wcfs, so that wcfs, in turn, could access the same
      ZODB database.
      
      Given a NEO client instance, it is already possible to retrieve
      master_nodes, cluster name, and detect whether SSL is being in use.
      However without being able to retrieve original SSL credentials,
      reconstructed zurl will not be full and wcfs won't be able to use
      exactly the same secrets as python part does.
      
      -> Help wendelin.core by remembering which ca/cert/key were used to
      build SSL context.
      
      This information is used by zstor_2zurl in wendelin.core here:
      
      https://lab.nexedi.com/nexedi/wendelin.core/blob/885b3556/lib/zodb.py#L390-418
      6e19ba3b
    • Kirill Smelkov's avatar
      client: Allow to force TLS via neos:// scheme · 57233d76
      Kirill Smelkov authored
      Similarly to how it is done with e.g. http:// and https:// - if neos://
      is given TLS usage is forced and ca/cert/key must be there either in the
      URI itself, or in $NEO_CA, $NEO_CERT and $NEO_KEY environment variables
      mimicking the way how e.g. for https:// TLS credentials are taken from
      host environment, not from the uri.
      
      The latter might be usability convenience, but is also useful for WCFS
      which needs to be able to remove secrets from uri on zurl normalization.
      
      Please see discussion at nexedi/neoppod!18 (comment 184439)
      for details.
      57233d76
  2. 11 Dec, 2023 1 commit
  3. 08 Nov, 2023 1 commit
    • Julien Muchembled's avatar
      master: fix crash when aborting early e.g. when failing to open listening socket · 9a3898e4
      Julien Muchembled authored
      Pre-mortem data:
      Traceback (most recent call last):
      File "neo/master/app.py", line 172, in run
      self._run()
      File "neo/master/app.py", line 180, in _run
      self.listening_conn = ListeningConnection(self, None, self.server)
      File "neo/lib/connection.py", line 298, in __init__
      connector.makeListeningConnection()
      File "neo/lib/connector.py", line 133, in makeListeningConnection
      self._error('listen', e)
      File "neo/lib/connector.py", line 93, in _error
      raise ConnectorException
      ConnectorException
      Traceback (most recent call last):
        File "neomaster", line 50, in <module>
          sys.exit(neo.scripts.neomaster.main())
        File "neo/scripts/neomaster.py", line 31, in main
          app.run()
        File "neo/master/app.py", line 175, in run
          self.log()
        File "neo/master/app.py", line 167, in log
          if self.pt is not None:
      AttributeError: 'Application' object has no attribute 'pt'
      9a3898e4
  4. 16 Oct, 2023 5 commits
    • Julien Muchembled's avatar
      b6f821a2
    • Julien Muchembled's avatar
      d112bfbd
    • Julien Muchembled's avatar
    • Julien Muchembled's avatar
      Bump protocol version · 0fc95175
      Julien Muchembled authored
      0fc95175
    • Julien Muchembled's avatar
      Reimplement pack in a scalable way, partial pack & approval/reject of pack orders · 4c3b6c4d
      Julien Muchembled authored
      This is still pack without garbage collection, and without deleting
      any transaction metadata ('trans' table).
      
      Partial pack means that the client can take a list of oids: only these
      oids will be packed. No API is defined yet at IStorage level.
      
      Storage nodes pack in background, independently from other storage
      nodes, partition by partition, and calling IStorage.pack() returns
      immediately (though internally, NEO does have a mechanism to wait
      until it's done, which can be required for some ZODB unit tests).
      
      This new implementation also introduces the concept of signing pack
      orders. The idea is that calling IStorage.pack() only records a pack
      order in the database, that can be reviewed/approved/rejected using
      an UI that is left to be done. For the moment, pack orders are
      automatically approved (by the master).
      
      Internally, pack orders are stored as extra metadata of a transaction.
      IOW, IStorage.pack() implies the commit of an (empty) transaction.
      
      IStorage.pack() can be called without waiting for the previous one
      to be completed. Pack orders processed in the same order as they are
      requested:
      - an unsigned pack order blocks the processing of any newer pack order;
      - rejected pack order are ignored.
      
      Approving a pack order also triggers pack on backup clusters.
      That's the simplest way to have everything consistent.
      Maybe later we could identify scenarios where it would be ok
      to unsign pack orders during asynchronous replication.
      
      The feature to check replicas is marked as experimental because it is
      not aware of differences that can happen during pack operations.
      _______________________________________________________________________
      
      About concurrency within the storage node, a first implementation
      extended what was done to delete partitions in background (see
      previous commit). But here, the job can't be easily split in splices
      that are never too big:
      - it's simpler to never split the processing of an oid but this can
        freeze the application for a long time when packing an oid that was
        modified many times (e.g. 30 min for an oid with 20 millions
        historical records);
      - then an attempt so that an oid can be processed in several times was
        inefficient, maybe due to a limit in RocksDB (packing the oid in the
        above example would take days during which NEO is significantly
        slower).
      
      So background database jobs were moved to a separate thread, using a
      separate connection to the underlying database. This is obviously
      only useful for the MySQL backend. In order to share as much code as
      possible between backends, SQLite also does the work in a separate
      thread but sharing the main connection instead of opening a separate
      one (so such backend would not be suited in the above example).
      
      But deleting raw data with a secondary connection is not possible
      without fsyncing too often (or transaction isolation issues...): these
      deletions are deferred by recording them in a new table, which is
      processed later with the main connection. This is not so bad because
      the actual deletion of raw data is usually more efficient this way
      (more sequential IO).
      
      Here are a few numbers:
      - without load: 10h45 (12h for the first reimplementation)
      - with a load that normally takes 6h58:
        - load: 7h33 (so 8.4% slower)
        - pack: 15h36 (+4h51)
      
      As explained above, the pack of a partition is split in 2 steps:
      - the longest one (here 78% without load) should have negligible
        peformance impact on the application because the work is done in a
        separate thread with a secondary connection, and also with something
        to minimize GIL impact by prioritizing the main thread;
      - the shortest one (22%) to process the deferred deletions,
        with even lower priority than replication: it tries to split
        the work in tasks that take ~10ms.
      4c3b6c4d
  5. 11 Oct, 2023 1 commit
  6. 04 Apr, 2023 9 commits
  7. 09 Mar, 2023 2 commits
  8. 19 Feb, 2023 1 commit
  9. 16 Feb, 2023 2 commits
  10. 14 Feb, 2023 3 commits
  11. 10 Feb, 2023 1 commit
  12. 02 Feb, 2022 1 commit
    • Kirill Smelkov's avatar
      Fix breakage with zodbpickle >= 2 · d5afef8e
      Kirill Smelkov authored
      Starting from zodbpickle 2 its binary class does not allow users to set
      arbitrary attributes and so
      
      	binary._pack = bytes.__str__
      
      fails with
      
      	TypeError: can't set attributes of built-in/extension type 'zodbpickle.binary'
      
      -> Fix it by explicitly checking for binary type on encoding instead of
      setting binary._pack
      
      See nexedi/slapos@27f574bc for pre-history.
      
      /cc @jerome
      d5afef8e
  13. 04 Jun, 2021 1 commit
    • Julien Muchembled's avatar
      admin: fix crash if not operational and a downstream cluster is RUNNING · 7f81ac2d
      Julien Muchembled authored
      Traceback (most recent call last):
        ...
        File ".../neo/lib/handler.py", line 75, in dispatch
          method(conn, *args, **kw)
        File ".../neo/admin/handler.py", line 174, in wrapper
          return func(self, name, *args, **kw)
        File ".../neo/admin/handler.py", line 190, in notifyMonitorInformation
          self.app.updateMonitorInformation(name, **info)
        File ".../neo/admin/app.py", line 290, in updateMonitorInformation
          self._notify(self.operational)
        File ".../neo/admin/app.py", line 315, in _notify
          body += '', name, '    ' + backup.formatSummary(upstream)[1]
        File ".../neo/admin/app.py", line 83, in formatSummary
          tid = self.ltid
      AttributeError: 'Backup' object has no attribute 'ltid'
      7f81ac2d
  14. 11 May, 2021 1 commit
  15. 02 Apr, 2021 5 commits
  16. 22 Mar, 2021 1 commit
  17. 04 Mar, 2021 2 commits
  18. 15 Jan, 2021 1 commit
    • Julien Muchembled's avatar
      ssl: don't care whether EOF is ragged or not · d98205d0
      Julien Muchembled authored
      The purpose of suppress_ragged_eofs=False was to micro-optimize the
      normal case: when there's no EOF.
      
      But commit 061cd5d8 showed that this
      option only turns ragged EOF into an exception. It may be easier for
      alternate NEO implementations to close the SSL connection properly. Or
      the performance benefit was not worth the risk to freeze a NEO process.
      d98205d0