Ejemplo n.º 1
0
def test_logrotateworker(logs_model):
  worker = LogRotateWorker()
  days = 90
  start_timestamp = datetime(2019, 1, 1)

  # Make sure there are no existing logs
  found = _lookup_logs(logs_model, start_timestamp - timedelta(days=1000), start_timestamp + timedelta(days=1000))
  assert not found

  # Create some logs
  for day in range(0, days):
    logs_model.log_action('push_repo', namespace_name='devtable', repository_name='simple',
                          ip='1.2.3.4', timestamp=start_timestamp-timedelta(days=day))

  # Ensure there are logs.
  logs = _lookup_logs(logs_model,
                      start_timestamp - timedelta(days=1000),
                      start_timestamp + timedelta(days=1000))

  assert len(logs) == days

  # Archive all the logs.
  assert worker._perform_archiving(start_timestamp + timedelta(days=1))

  # Ensure all the logs were archived.
  found = _lookup_logs(logs_model, start_timestamp - timedelta(days=1000), start_timestamp + timedelta(days=1000))
  assert not found
Ejemplo n.º 2
0
def test_logrotateworker_with_cutoff(logs_model):
    days = 60
    start_timestamp = datetime(2019, 1, 1)

    # Make sure there are no existing logs
    found = _lookup_logs(logs_model, start_timestamp - timedelta(days=365),
                         start_timestamp + timedelta(days=365))
    assert not found

    # Create a new set of logs/indices.
    for day in range(0, days):
        logs_model.log_action(
            "push_repo",
            namespace_name="devtable",
            repository_name="simple",
            ip="1.2.3.4",
            timestamp=start_timestamp + timedelta(days=day),
        )

    # Get all logs
    logs = _lookup_logs(
        logs_model,
        start_timestamp - timedelta(days=days - 1),
        start_timestamp + timedelta(days=days + 1),
    )

    assert len(logs) == days

    # Set the cutoff datetime to be the midpoint of the logs
    midpoint = logs[0:len(logs) // 2]
    assert midpoint
    assert len(midpoint) < len(logs)

    worker = LogRotateWorker()
    cutoff_date = midpoint[-1].datetime

    # Archive the indices at or older than the cutoff date
    archived_files = worker._perform_archiving(cutoff_date)

    # Ensure the eariler logs were archived
    found = _lookup_logs(logs_model, start_timestamp,
                         cutoff_date - timedelta(seconds=1))
    assert not found

    # Check that the files were written to storage
    for archived_file in archived_files:
        assert storage.exists([SAVE_LOCATION],
                              os.path.join(SAVE_PATH, archived_file))

    # If current model uses ES, check that the indices were also deleted
    if isinstance(logs_model, DocumentLogsModel):
        assert len(logs_model.list_indices()) == days - (len(logs) // 2)
        for index in logs_model.list_indices():
            dt = datetime.strptime(index[len(INDEX_NAME_PREFIX):],
                                   INDEX_DATE_FORMAT)
            assert dt >= cutoff_date