def test_bad_header(self):
        """
        Ensure that bad header is skipped when it exists.
        """

        # This case tests against a header that does not match
        # 0000 0000 0000 0100 0000 0000 0000 0151
        file_path = os.path.join(RESOURCE_PATH, 'E0000001-BAD-HEADER1.DAT')
        stream_handle = open(file_path, 'rb')

        with self.assertRaises(SampleException):
            GlobalWfpEFileParser(self.config, None, stream_handle,
                                 self.state_callback, self.pub_callback,
                                 self.exception_callback)

        stream_handle.close()

        # This case tests against a header that does not match global E-type data, but matches coastal
        # E-type data: 0001 0000 0000 0000 0001 0001 0000 0000
        file_path = os.path.join(RESOURCE_PATH, 'E0000001-BAD-HEADER2.DAT')
        stream_handle = open(file_path, 'rb')

        with self.assertRaises(SampleException):
            GlobalWfpEFileParser(self.config, None, stream_handle,
                                 self.state_callback, self.pub_callback,
                                 self.exception_callback)

        stream_handle.close()
    def test_bad_header(self):
        """
        Ensure that bad header is skipped when it exists.
        """

        # This case tests against a header that does not match
        # 0000 0000 0000 0100 0000 0000 0000 0151
        file_path = os.path.join(RESOURCE_PATH, 'E0000001-BAD-HEADER1.DAT')
        self.stream_handle = open(file_path, 'rb')

        with self.assertRaises(SampleException):
            self.parser = GlobalWfpEFileParser(self.config, self.start_state, self.stream_handle,
                                               self.state_callback, self.pub_callback, self.exception_callback)

        self.stream_handle.close()

        # This case tests against a header that does not match global E-type data, but matches coastal
        # E-type data: 0001 0000 0000 0000 0001 0001 0000 0000
        file_path = os.path.join(RESOURCE_PATH, 'E0000001-BAD-HEADER2.DAT')
        self.stream_handle = open(file_path, 'rb')

        with self.assertRaises(SampleException):
            self.parser = GlobalWfpEFileParser(self.config, self.start_state, self.stream_handle,
                                               self.state_callback, self.pub_callback, self.exception_callback)

        self.stream_handle.close()
    def test_mid_state_start(self):
        """
        Test starting the parser in a state in the middle of processing
        """
        file_path = os.path.join(RESOURCE_PATH, 'E0000001.DAT')
        self.stream_handle = open(file_path, 'rb')

        # Moving the file position past the header and two records
        new_state = {
            StateKey.POSITION:
            HEADER_BYTES + (WFP_E_GLOBAL_RECOVERED_ENG_DATA_SAMPLE_BYTES * 2)
        }

        self.parser = GlobalWfpEFileParser(self.config, new_state,
                                           self.stream_handle,
                                           self.state_callback,
                                           self.pub_callback,
                                           self.exception_callback)

        particles = self.parser.get_records(4)

        # Should end up with 4 particles
        self.assertTrue(len(particles) == 4)

        # Compare with 6th data record
        self.assert_result(self.test_particle1, particles[3])

        self.stream_handle.close()
    def test_get_many(self):
        """
        Read test data and pull out multiple data particles at one time.
        Assert that the results are those we expected.
        """
        file_path = os.path.join(RESOURCE_PATH, 'E0000001.DAT')
        self.stream_handle = open(file_path, 'rb')

        self.parser = GlobalWfpEFileParser(self.config, self.start_state,
                                           self.stream_handle,
                                           self.state_callback,
                                           self.pub_callback,
                                           self.exception_callback)

        particles = self.parser.get_records(20)

        # Should end up with 20 particles
        self.assertTrue(len(particles) == 20)

        # Compare test_particle3 with the 20th particle from sample data
        self.assert_result(self.test_particle3, particles[19])

        particles = self.parser.get_records(30)

        # Should end up with 30 particles
        self.assertTrue(len(particles) == 30)

        # Compare test_particle4 with the 30th particle from sample data
        self.assert_result(self.test_particle4, particles[29])

        self.stream_handle.close()
    def test_simple(self):
        """
        Read test data and pull out data particles six at a time.
        Assert that the results of sixth particle are those we expected.
        """

        file_path = os.path.join(RESOURCE_PATH, 'small.DAT')
        self.stream_handle = open(file_path, 'rb')

        self.parser = GlobalWfpEFileParser(self.config, self.start_state,
                                           self.stream_handle,
                                           self.state_callback,
                                           self.pub_callback,
                                           self.exception_callback)

        particles = self.parser.get_records(6)

        log.debug("particles: %s", particles)
        for particle in particles:
            log.info("*** test particle: %s", particle.generate_dict())

        # Make sure the sixth particle has the correct values
        test_data = self.get_dict_from_yml('good.yml')
        self.assert_result(test_data['data'][0], particles[5])

        self.stream_handle.close()
    def test_bad_data(self):
        """
        Ensure that bad data is skipped when it exists.
        """
        file_path = os.path.join(RESOURCE_PATH, 'E0000001-BAD-DATA.DAT')
        stream_handle = open(file_path, 'rb')

        parser = GlobalWfpEFileParser(self.config, None, stream_handle,
                                      self.state_callback, self.pub_callback, self.exception_callback)

        with self.assertRaises(SampleException):
            parser.get_records(1)

        stream_handle.close()
Esempio n. 7
0
    def test_bad_data(self):
        """
        Ensure that bad data is skipped when it exists.
        """
        file_path = os.path.join(RESOURCE_PATH, 'E0000001-BAD-DATA.DAT')
        stream_handle = open(file_path, 'rb')

        parser = GlobalWfpEFileParser(self.config, None, stream_handle,
                                      self.state_callback, self.pub_callback, self.exception_callback)

        with self.assertRaises(SampleException):
            parser.get_records(1)

        stream_handle.close()
    def test_set_state(self):
        """
        Test changing to a new state after initializing the parser and
        reading data, as if new data has been found and the state has
        changed
        """
        filepath = os.path.join(RESOURCE_PATH, 'E0000001.DAT')
        self.stream_handle = open(filepath, 'rb')

        # Moving the file position past the header and two records
        new_state = {
            StateKey.POSITION:
            HEADER_BYTES + (WFP_E_GLOBAL_RECOVERED_ENG_DATA_SAMPLE_BYTES * 2)
        }

        log.debug("new_state: %s", new_state)
        self.parser = GlobalWfpEFileParser(self.config, new_state,
                                           self.stream_handle,
                                           self.state_callback,
                                           self.pub_callback,
                                           self.exception_callback)

        particles = self.parser.get_records(4)

        # Should end up with 4 particles
        self.assertTrue(len(particles) == 4)

        # particles[3] is the 7th data record in the test file
        self.assert_result(self.test_particle1, particles[3])

        # Simulate State has changed:
        #   Move the file position past the header and three records
        #   Set the state of the existing parser
        new_state = {
            StateKey.POSITION:
            HEADER_BYTES + (WFP_E_GLOBAL_RECOVERED_ENG_DATA_SAMPLE_BYTES * 3)
        }
        self.parser.set_state(new_state)

        particles = self.parser.get_records(10)

        # Should end up with 10 particles
        self.assertTrue(len(particles) == 10)

        # particles[9] is the 13th data record in test file
        self.assert_result(self.test_particle2, particles[9])

        self.stream_handle.close()
Esempio n. 9
0
def parse(basePythonCodePath, sourceFilePath, particleDataHdlrObj):

    config.add_configuration(
        os.path.join(basePythonCodePath, 'res', 'config', 'mi-logging.yml'))

    log = get_logger()

    parser_config = {
        DataSetDriverConfigKeys.PARTICLE_MODULE:
        'mi.dataset.parser.flord_l_wfp',
        DataSetDriverConfigKeys.PARTICLE_CLASS:
        'FlordLWfpInstrumentParserDataParticle'
    }

    def exception_callback(exception):
        log.debug("ERROR: %r", exception)
        particleDataHdlrObj.setParticleDataCaptureFailure()

    with open(sourceFilePath, 'r') as stream_handle:
        parser = GlobalWfpEFileParser(
            parser_config, None, stream_handle, lambda state, ingested: None,
            lambda data: log.trace("Found data: %s", data), exception_callback)

        driver = DataSetDriver(parser, particleDataHdlrObj)
        driver.processFileStream()

    return particleDataHdlrObj
    def test_get_many(self):
        """
        Read test data and pull out multiple data particles at one time.
        Assert that the results are those we expected.
        """
        file_path = os.path.join(RESOURCE_PATH, 'E0000001.DAT')
        self.stream_handle = open(file_path, 'rb')

        self.parser = GlobalWfpEFileParser(self.config, self.start_state, self.stream_handle,
                                           self.state_callback, self.pub_callback, self.exception_callback)

        particles = self.parser.get_records(20)

        # Should end up with 20 particles
        self.assertTrue(len(particles) == 20)

        # Compare test_particle3 with the 20th particle from sample data
        self.assert_result(self.test_particle3, particles[19])

        particles = self.parser.get_records(30)

        # Should end up with 30 particles
        self.assertTrue(len(particles) == 30)

        # Compare test_particle4 with the 30th particle from sample data
        self.assert_result(self.test_particle4, particles[29])

        self.stream_handle.close()
Esempio n. 11
0
    def test_get_many(self):
        """
        Read test data and pull out multiple data particles at one time.
        Assert that the results are those we expected.
        """
        file_path = os.path.join(RESOURCE_PATH, 'E0000001.DAT')
        stream_handle = open(file_path, 'rb')

        parser = GlobalWfpEFileParser(self.config, None, stream_handle,
                                      self.state_callback, self.pub_callback, self.exception_callback)

        particles = parser.get_records(50)

        # Should end up with 20 particles
        self.assertTrue(len(particles) == 50)

        self.assert_particles(particles, 'E0000001_50.yml', RESOURCE_PATH)
        stream_handle.close()
Esempio n. 12
0
    def test_long_stream(self):
        """
        Test a long stream
        """
        file_path = os.path.join(RESOURCE_PATH, 'E0000001.DAT')
        stream_handle = open(file_path, 'rb')

        parser = GlobalWfpEFileParser(self.config, None, stream_handle,
                                      self.state_callback, self.pub_callback, self.exception_callback)

        particles = parser.get_records(1000)
        # File is 20,530 bytes
        #   minus 24 header
        #   minus 16 footer
        #   each particle is 30 bytes
        # Should end up with 683 particles
        self.assertTrue(len(particles) == 683)
        stream_handle.close()
Esempio n. 13
0
    def test_get_many(self):
        """
        Read test data and pull out multiple data particles at one time.
        Assert that the results are those we expected.
        """
        file_path = os.path.join(RESOURCE_PATH, 'E0000001.DAT')
        stream_handle = open(file_path, 'rb')

        parser = GlobalWfpEFileParser(self.config, None, stream_handle,
                                      self.state_callback, self.pub_callback, self.exception_callback)

        particles = parser.get_records(50)

        # Should end up with 20 particles
        self.assertTrue(len(particles) == 50)

        self.assert_particles(particles, 'E0000001_50.yml', RESOURCE_PATH)
        stream_handle.close()
Esempio n. 14
0
    def test_long_stream(self):
        """
        Test a long stream
        """
        file_path = os.path.join(RESOURCE_PATH, 'E0000001.DAT')
        stream_handle = open(file_path, 'rb')

        parser = GlobalWfpEFileParser(self.config, None, stream_handle,
                                      self.state_callback, self.pub_callback, self.exception_callback)

        particles = parser.get_records(1000)
        # File is 20,530 bytes
        #   minus 24 header
        #   minus 16 footer
        #   each particle is 30 bytes
        # Should end up with 683 particles
        self.assertTrue(len(particles) == 683)
        stream_handle.close()
Esempio n. 15
0
    def test_simple(self):
        """
        Read test data and pull out data particles six at a time.
        Assert that the results of sixth particle are those we expected.
        """

        file_path = os.path.join(RESOURCE_PATH, 'small.DAT')
        stream_handle = open(file_path, 'rb')

        parser = GlobalWfpEFileParser(self.config, None, stream_handle,
                                      self.state_callback, self.pub_callback, self.exception_callback)

        particles = parser.get_records(6)

        log.debug("particles: %s", particles)
        for particle in particles:
            log.info("*** test particle: %s", particle.generate_dict())

        self.assert_particles(particles, 'good.yml', RESOURCE_PATH)

        stream_handle.close()
Esempio n. 16
0
    def test_simple(self):
        """
        Read test data and pull out data particles six at a time.
        Assert that the results of sixth particle are those we expected.
        """

        file_path = os.path.join(RESOURCE_PATH, 'small.DAT')
        stream_handle = open(file_path, 'rb')

        parser = GlobalWfpEFileParser(self.config, None, stream_handle,
                                      self.state_callback, self.pub_callback, self.exception_callback)

        particles = parser.get_records(6)

        log.debug("particles: %s", particles)
        for particle in particles:
            log.info("*** test particle: %s", particle.generate_dict())

        self.assert_particles(particles, 'good.yml', RESOURCE_PATH)

        stream_handle.close()
Esempio n. 17
0
    def _build_recovered_parser(self, parser_state, stream_in):
        """
        Build and return the parser
        """
        config = self._parser_config.get(DataSourceKey.FLORD_L_WFP)
        config.update({
            DataSetDriverConfigKeys.PARTICLE_MODULE:
            'mi.dataset.parser.flord_l_wfp',
            DataSetDriverConfigKeys.PARTICLE_CLASS:
            'FlordLWfpInstrumentParserDataParticle'
        })

        parser = GlobalWfpEFileParser(
            config, parser_state, stream_in,
            lambda state, ingested: self._save_parser_state(
                state, DataSourceKey.FLORD_L_WFP, ingested),
            self._data_callback, self._sample_exception_callback)
        return parser
Esempio n. 18
0
def parse(unused, source_file_path, particle_data_handler):
    parser_config = {
        DataSetDriverConfigKeys.PARTICLE_MODULE:
        'mi.dataset.parser.flord_l_wfp',
        DataSetDriverConfigKeys.PARTICLE_CLASS:
        'FlordLWfpInstrumentParserDataParticle'
    }

    def exception_callback(exception):
        log.debug("ERROR: %r", exception)
        particle_data_handler.setParticleDataCaptureFailure()

    with open(source_file_path, 'r') as stream_handle:
        parser = GlobalWfpEFileParser(
            parser_config, None, stream_handle, lambda state, ingested: None,
            lambda data: log.trace("Found data: %s", data), exception_callback)

        driver = DataSetDriver(parser, particle_data_handler)
        driver.processFileStream()

    return particle_data_handler
    def test_set_state(self):
        """
        Test changing to a new state after initializing the parser and
        reading data, as if new data has been found and the state has
        changed
        """
        filepath = os.path.join(RESOURCE_PATH, 'E0000001.DAT')
        self.stream_handle = open(filepath, 'rb')

        # Moving the file position past the header and two records
        new_state = {StateKey.POSITION: HEADER_BYTES+(WFP_E_GLOBAL_RECOVERED_ENG_DATA_SAMPLE_BYTES*2)}

        log.debug("new_state: %s", new_state)
        self.parser = GlobalWfpEFileParser(self.config, new_state, self.stream_handle,
                                           self.state_callback, self.pub_callback, self.exception_callback)

        particles = self.parser.get_records(4)

        # Should end up with 4 particles
        self.assertTrue(len(particles) == 4)

        # particles[3] is the 7th data record in the test file
        self.assert_result(self.test_particle1, particles[3])

        # Simulate State has changed:
        #   Move the file position past the header and three records
        #   Set the state of the existing parser
        new_state = {StateKey.POSITION: HEADER_BYTES+(WFP_E_GLOBAL_RECOVERED_ENG_DATA_SAMPLE_BYTES*3)}
        self.parser.set_state(new_state)

        particles = self.parser.get_records(10)

        # Should end up with 10 particles
        self.assertTrue(len(particles) == 10)

        # particles[9] is the 13th data record in test file
        self.assert_result(self.test_particle2, particles[9])

        self.stream_handle.close()
    def test_mid_state_start(self):
        """
        Test starting the parser in a state in the middle of processing
        """
        file_path = os.path.join(RESOURCE_PATH, 'E0000001.DAT')
        self.stream_handle = open(file_path, 'rb')

        # Moving the file position past the header and two records
        new_state = {StateKey.POSITION: HEADER_BYTES+(WFP_E_GLOBAL_RECOVERED_ENG_DATA_SAMPLE_BYTES*2)}

        self.parser = GlobalWfpEFileParser(self.config, new_state, self.stream_handle,
                                           self.state_callback, self.pub_callback, self.exception_callback)

        particles = self.parser.get_records(4)

        # Should end up with 4 particles
        self.assertTrue(len(particles) == 4)

        # Compare with 6th data record
        self.assert_result(self.test_particle1, particles[3])

        self.stream_handle.close()
    def test_simple(self):
        """
        Read test data and pull out data particles six at a time.
        Assert that the results of sixth particle are those we expected.
        """

        file_path = os.path.join(RESOURCE_PATH, 'small.DAT')
        self.stream_handle = open(file_path, 'rb')

        self.parser = GlobalWfpEFileParser(self.config, self.start_state, self.stream_handle,
                                           self.state_callback, self.pub_callback, self.exception_callback)

        particles = self.parser.get_records(6)

        log.debug("particles: %s", particles)
        for particle in particles:
            log.info("*** test particle: %s", particle.generate_dict())

        # Make sure the sixth particle has the correct values
        test_data = self.get_dict_from_yml('good.yml')
        self.assert_result(test_data['data'][0], particles[5])

        self.stream_handle.close()
class FlordLWfpParserUnitTestCase(ParserUnitTestCase):
    """
    Parser unit test suite
    """
    def state_callback(self, state, file_ingested):
        """ Call back method to watch what comes in via the position callback """
        self.state_callback_value = state
        self.file_ingested_value = file_ingested

    def pub_callback(self, pub):
        """ Call back method to watch what comes in via the publish callback """
        self.publish_callback_value = pub

    def exception_callback(self, exception):
        """ Call back method to match what comes in via the exception callback """
        self.exception_callback_value = exception

    def setUp(self):
        ParserUnitTestCase.setUp(self)
        self.config = {
            DataSetDriverConfigKeys.PARTICLE_MODULE:
            'mi.dataset.parser.flord_l_wfp',
            DataSetDriverConfigKeys.PARTICLE_CLASS:
            'FlordLWfpInstrumentParserDataParticle'
        }

        # Define test data particles and their associated timestamps which will be
        # compared with returned results

        self.start_state = {StateKey.POSITION: 0}

        self.file_ingested_value = None
        self.state_callback_value = None
        self.publish_callback_value = None

        self.test_particle1 = {}
        self.test_particle1['internal_timestamp'] = 3583638177.0
        self.test_particle1[StateKey.POSITION] = 204
        self.test_particle1[
            FlordLWfpInstrumentParserDataParticleKey.RAW_SIGNAL_CHL] = 54
        self.test_particle1[
            FlordLWfpInstrumentParserDataParticleKey.RAW_SIGNAL_BETA] = 112
        self.test_particle1[
            FlordLWfpInstrumentParserDataParticleKey.RAW_INTERNAL_TEMP] = 571
        self.test_particle1[FlordLWfpInstrumentParserDataParticleKey.
                            WFP_TIMESTAMP] = 1374649377

        self.test_particle2 = {}
        self.test_particle2['internal_timestamp'] = 3583638247
        self.test_particle2[StateKey.POSITION] = 414
        self.test_particle2[
            FlordLWfpInstrumentParserDataParticleKey.RAW_SIGNAL_CHL] = 55
        self.test_particle2[
            FlordLWfpInstrumentParserDataParticleKey.RAW_SIGNAL_BETA] = 112
        self.test_particle2[
            FlordLWfpInstrumentParserDataParticleKey.RAW_INTERNAL_TEMP] = 570
        self.test_particle2[FlordLWfpInstrumentParserDataParticleKey.
                            WFP_TIMESTAMP] = 1374649447

        self.test_particle3 = {}
        self.test_particle3['internal_timestamp'] = 3583638317
        self.test_particle3[StateKey.POSITION] = 624
        self.test_particle3[
            FlordLWfpInstrumentParserDataParticleKey.RAW_SIGNAL_CHL] = 56
        self.test_particle3[
            FlordLWfpInstrumentParserDataParticleKey.RAW_SIGNAL_BETA] = 114
        self.test_particle3[
            FlordLWfpInstrumentParserDataParticleKey.RAW_INTERNAL_TEMP] = 570
        self.test_particle3[FlordLWfpInstrumentParserDataParticleKey.
                            WFP_TIMESTAMP] = 1374649517

        self.test_particle4 = {}
        self.test_particle4['internal_timestamp'] = 3583638617
        self.test_particle4[StateKey.POSITION] = 1524
        self.test_particle4[
            FlordLWfpInstrumentParserDataParticleKey.RAW_SIGNAL_CHL] = 54
        self.test_particle4[
            FlordLWfpInstrumentParserDataParticleKey.RAW_SIGNAL_BETA] = 110
        self.test_particle4[
            FlordLWfpInstrumentParserDataParticleKey.RAW_INTERNAL_TEMP] = 569
        self.test_particle4[FlordLWfpInstrumentParserDataParticleKey.
                            WFP_TIMESTAMP] = 1374649817

    def test_simple(self):
        """
        Read test data and pull out data particles six at a time.
        Assert that the results of sixth particle are those we expected.
        """

        file_path = os.path.join(RESOURCE_PATH, 'small.DAT')
        self.stream_handle = open(file_path, 'rb')

        self.parser = GlobalWfpEFileParser(self.config, self.start_state,
                                           self.stream_handle,
                                           self.state_callback,
                                           self.pub_callback,
                                           self.exception_callback)

        particles = self.parser.get_records(6)

        log.debug("particles: %s", particles)
        for particle in particles:
            log.info("*** test particle: %s", particle.generate_dict())

        # Make sure the sixth particle has the correct values
        test_data = self.get_dict_from_yml('good.yml')
        self.assert_result(test_data['data'][0], particles[5])

        self.stream_handle.close()

    def test_get_many(self):
        """
        Read test data and pull out multiple data particles at one time.
        Assert that the results are those we expected.
        """
        file_path = os.path.join(RESOURCE_PATH, 'E0000001.DAT')
        self.stream_handle = open(file_path, 'rb')

        self.parser = GlobalWfpEFileParser(self.config, self.start_state,
                                           self.stream_handle,
                                           self.state_callback,
                                           self.pub_callback,
                                           self.exception_callback)

        particles = self.parser.get_records(20)

        # Should end up with 20 particles
        self.assertTrue(len(particles) == 20)

        # Compare test_particle3 with the 20th particle from sample data
        self.assert_result(self.test_particle3, particles[19])

        particles = self.parser.get_records(30)

        # Should end up with 30 particles
        self.assertTrue(len(particles) == 30)

        # Compare test_particle4 with the 30th particle from sample data
        self.assert_result(self.test_particle4, particles[29])

        self.stream_handle.close()

    def test_long_stream(self):
        """
        Test a long stream
        """
        file_path = os.path.join(RESOURCE_PATH, 'E0000001.DAT')
        self.stream_handle = open(file_path, 'rb')

        self.parser = GlobalWfpEFileParser(self.config, self.start_state,
                                           self.stream_handle,
                                           self.state_callback,
                                           self.pub_callback,
                                           self.exception_callback)

        particles = self.parser.get_records(1000)
        # File is 20,530 bytes
        #   minus 24 header
        #   minus 16 footer
        #   each particle is 30 bytes
        # Should end up with 683 particles
        self.assertTrue(len(particles) == 683)
        self.stream_handle.close()

    def test_mid_state_start(self):
        """
        Test starting the parser in a state in the middle of processing
        """
        file_path = os.path.join(RESOURCE_PATH, 'E0000001.DAT')
        self.stream_handle = open(file_path, 'rb')

        # Moving the file position past the header and two records
        new_state = {
            StateKey.POSITION:
            HEADER_BYTES + (WFP_E_GLOBAL_RECOVERED_ENG_DATA_SAMPLE_BYTES * 2)
        }

        self.parser = GlobalWfpEFileParser(self.config, new_state,
                                           self.stream_handle,
                                           self.state_callback,
                                           self.pub_callback,
                                           self.exception_callback)

        particles = self.parser.get_records(4)

        # Should end up with 4 particles
        self.assertTrue(len(particles) == 4)

        # Compare with 6th data record
        self.assert_result(self.test_particle1, particles[3])

        self.stream_handle.close()

    def test_set_state(self):
        """
        Test changing to a new state after initializing the parser and
        reading data, as if new data has been found and the state has
        changed
        """
        filepath = os.path.join(RESOURCE_PATH, 'E0000001.DAT')
        self.stream_handle = open(filepath, 'rb')

        # Moving the file position past the header and two records
        new_state = {
            StateKey.POSITION:
            HEADER_BYTES + (WFP_E_GLOBAL_RECOVERED_ENG_DATA_SAMPLE_BYTES * 2)
        }

        log.debug("new_state: %s", new_state)
        self.parser = GlobalWfpEFileParser(self.config, new_state,
                                           self.stream_handle,
                                           self.state_callback,
                                           self.pub_callback,
                                           self.exception_callback)

        particles = self.parser.get_records(4)

        # Should end up with 4 particles
        self.assertTrue(len(particles) == 4)

        # particles[3] is the 7th data record in the test file
        self.assert_result(self.test_particle1, particles[3])

        # Simulate State has changed:
        #   Move the file position past the header and three records
        #   Set the state of the existing parser
        new_state = {
            StateKey.POSITION:
            HEADER_BYTES + (WFP_E_GLOBAL_RECOVERED_ENG_DATA_SAMPLE_BYTES * 3)
        }
        self.parser.set_state(new_state)

        particles = self.parser.get_records(10)

        # Should end up with 10 particles
        self.assertTrue(len(particles) == 10)

        # particles[9] is the 13th data record in test file
        self.assert_result(self.test_particle2, particles[9])

        self.stream_handle.close()

    def test_bad_data(self):
        """
        Ensure that bad data is skipped when it exists.
        """
        file_path = os.path.join(RESOURCE_PATH, 'E0000001-BAD-DATA.DAT')
        self.stream_handle = open(file_path, 'rb')

        self.parser = GlobalWfpEFileParser(self.config, self.start_state,
                                           self.stream_handle,
                                           self.state_callback,
                                           self.pub_callback,
                                           self.exception_callback)

        with self.assertRaises(SampleException):
            self.parser.get_records(1)

        self.stream_handle.close()

    def test_bad_header(self):
        """
        Ensure that bad header is skipped when it exists.
        """

        # This case tests against a header that does not match
        # 0000 0000 0000 0100 0000 0000 0000 0151
        file_path = os.path.join(RESOURCE_PATH, 'E0000001-BAD-HEADER1.DAT')
        self.stream_handle = open(file_path, 'rb')

        with self.assertRaises(SampleException):
            self.parser = GlobalWfpEFileParser(self.config, self.start_state,
                                               self.stream_handle,
                                               self.state_callback,
                                               self.pub_callback,
                                               self.exception_callback)

        self.stream_handle.close()

        # This case tests against a header that does not match global E-type data, but matches coastal
        # E-type data: 0001 0000 0000 0000 0001 0001 0000 0000
        file_path = os.path.join(RESOURCE_PATH, 'E0000001-BAD-HEADER2.DAT')
        self.stream_handle = open(file_path, 'rb')

        with self.assertRaises(SampleException):
            self.parser = GlobalWfpEFileParser(self.config, self.start_state,
                                               self.stream_handle,
                                               self.state_callback,
                                               self.pub_callback,
                                               self.exception_callback)

        self.stream_handle.close()

    def assert_result(self, test, particle):
        """
        Suite of tests to run against each returned particle and expected
        results of the same.  The test parameter should be a dictionary
        that contains the keys to be tested in the particle
        the 'internal_timestamp' and 'position' keys are
        treated differently than others but can be verified if supplied
        """
        log.debug("test arg: %s", test)
        particle_dict = particle.generate_dict()
        log.debug("particle_dict: %s", particle_dict)
        #for efficiency turn the particle values list of dictionaries into a dictionary
        particle_values = {}
        for param in particle_dict.get('values'):
            particle_values[param['value_id']] = param['value']

        # compare each key in the test to the data in the particle
        for key in test:
            test_data = test[key]

            #get the correct data to compare to the test
            if key == 'internal_timestamp':
                particle_data = particle.get_value('internal_timestamp')
                #the timestamp is in the header part of the particle

                log.info("internal_timestamp %d", particle_data)

            elif key == 'position':
                particle_data = self.state_callback_value['position']
                #position corresponds to the position in the file

                log.info("position %d", particle_data)

            else:
                particle_data = particle_values.get(key)
                #others are all part of the parsed values part of the particle

            if particle_data is None:
                #generally OK to ignore index keys in the test data, verify others

                log.warning(
                    "\nWarning: assert_result ignoring test key %s, does not exist in particle",
                    key)
            else:
                if isinstance(test_data, float):
                    # slightly different test for these values as they are floats.
                    compare = numpy.abs(test_data - particle_data) <= 1e-5
                    self.assertTrue(compare)
                else:
                    # otherwise they are all ints and should be exactly equal
                    log.debug("test_data %s, particle_data %s", test_data,
                              particle_data)
                    self.assertEqual(test_data, particle_data)

    @staticmethod
    def get_dict_from_yml(filename):
        """
        This utility routine loads the contents of a yml file
        into a dictionary
        """

        fid = open(os.path.join(RESOURCE_PATH, filename), 'r')
        result = yaml.load(fid)
        fid.close()

        return result
class FlordLWfpParserUnitTestCase(ParserUnitTestCase):
    """
    Parser unit test suite
    """
    def state_callback(self, state, file_ingested):
        """ Call back method to watch what comes in via the position callback """
        self.state_callback_value = state
        self.file_ingested_value = file_ingested

    def pub_callback(self, pub):
        """ Call back method to watch what comes in via the publish callback """
        self.publish_callback_value = pub

    def exception_callback(self, exception):
        """ Call back method to match what comes in via the exception callback """
        self.exception_callback_value = exception

    def setUp(self):
        ParserUnitTestCase.setUp(self)
        self.config = {
            DataSetDriverConfigKeys.PARTICLE_MODULE: 'mi.dataset.parser.flord_l_wfp',
            DataSetDriverConfigKeys.PARTICLE_CLASS: 'FlordLWfpInstrumentParserDataParticle'
        }

        # Define test data particles and their associated timestamps which will be
        # compared with returned results

        self.start_state = {StateKey.POSITION: 0}

        self.file_ingested_value = None
        self.state_callback_value = None
        self.publish_callback_value = None

        self.test_particle1 = {}
        self.test_particle1['internal_timestamp'] = 3583638177.0
        self.test_particle1[StateKey.POSITION] = 204
        self.test_particle1[FlordLWfpInstrumentParserDataParticleKey.RAW_SIGNAL_CHL] = 54
        self.test_particle1[FlordLWfpInstrumentParserDataParticleKey.RAW_SIGNAL_BETA] = 112
        self.test_particle1[FlordLWfpInstrumentParserDataParticleKey.RAW_INTERNAL_TEMP] = 571
        self.test_particle1[FlordLWfpInstrumentParserDataParticleKey.WFP_TIMESTAMP] = 1374649377

        self.test_particle2 = {}
        self.test_particle2['internal_timestamp'] = 3583638247
        self.test_particle2[StateKey.POSITION] = 414
        self.test_particle2[FlordLWfpInstrumentParserDataParticleKey.RAW_SIGNAL_CHL] = 55
        self.test_particle2[FlordLWfpInstrumentParserDataParticleKey.RAW_SIGNAL_BETA] = 112
        self.test_particle2[FlordLWfpInstrumentParserDataParticleKey.RAW_INTERNAL_TEMP] = 570
        self.test_particle2[FlordLWfpInstrumentParserDataParticleKey.WFP_TIMESTAMP] = 1374649447

        self.test_particle3 = {}
        self.test_particle3['internal_timestamp'] = 3583638317
        self.test_particle3[StateKey.POSITION] = 624
        self.test_particle3[FlordLWfpInstrumentParserDataParticleKey.RAW_SIGNAL_CHL] = 56
        self.test_particle3[FlordLWfpInstrumentParserDataParticleKey.RAW_SIGNAL_BETA] = 114
        self.test_particle3[FlordLWfpInstrumentParserDataParticleKey.RAW_INTERNAL_TEMP] = 570
        self.test_particle3[FlordLWfpInstrumentParserDataParticleKey.WFP_TIMESTAMP] = 1374649517

        self.test_particle4 = {}
        self.test_particle4['internal_timestamp'] = 3583638617
        self.test_particle4[StateKey.POSITION] = 1524
        self.test_particle4[FlordLWfpInstrumentParserDataParticleKey.RAW_SIGNAL_CHL] = 54
        self.test_particle4[FlordLWfpInstrumentParserDataParticleKey.RAW_SIGNAL_BETA] = 110
        self.test_particle4[FlordLWfpInstrumentParserDataParticleKey.RAW_INTERNAL_TEMP] = 569
        self.test_particle4[FlordLWfpInstrumentParserDataParticleKey.WFP_TIMESTAMP] = 1374649817

    def test_simple(self):
        """
        Read test data and pull out data particles six at a time.
        Assert that the results of sixth particle are those we expected.
        """

        file_path = os.path.join(RESOURCE_PATH, 'small.DAT')
        self.stream_handle = open(file_path, 'rb')

        self.parser = GlobalWfpEFileParser(self.config, self.start_state, self.stream_handle,
                                           self.state_callback, self.pub_callback, self.exception_callback)

        particles = self.parser.get_records(6)

        log.debug("particles: %s", particles)
        for particle in particles:
            log.info("*** test particle: %s", particle.generate_dict())

        # Make sure the sixth particle has the correct values
        test_data = self.get_dict_from_yml('good.yml')
        self.assert_result(test_data['data'][0], particles[5])

        self.stream_handle.close()

    def test_get_many(self):
        """
        Read test data and pull out multiple data particles at one time.
        Assert that the results are those we expected.
        """
        file_path = os.path.join(RESOURCE_PATH, 'E0000001.DAT')
        self.stream_handle = open(file_path, 'rb')

        self.parser = GlobalWfpEFileParser(self.config, self.start_state, self.stream_handle,
                                           self.state_callback, self.pub_callback, self.exception_callback)

        particles = self.parser.get_records(20)

        # Should end up with 20 particles
        self.assertTrue(len(particles) == 20)

        # Compare test_particle3 with the 20th particle from sample data
        self.assert_result(self.test_particle3, particles[19])

        particles = self.parser.get_records(30)

        # Should end up with 30 particles
        self.assertTrue(len(particles) == 30)

        # Compare test_particle4 with the 30th particle from sample data
        self.assert_result(self.test_particle4, particles[29])

        self.stream_handle.close()

    def test_long_stream(self):
        """
        Test a long stream
        """
        file_path = os.path.join(RESOURCE_PATH, 'E0000001.DAT')
        self.stream_handle = open(file_path, 'rb')

        self.parser = GlobalWfpEFileParser(self.config, self.start_state, self.stream_handle,
                                           self.state_callback, self.pub_callback, self.exception_callback)

        particles = self.parser.get_records(1000)
        # File is 20,530 bytes
        #   minus 24 header
        #   minus 16 footer
        #   each particle is 30 bytes
        # Should end up with 683 particles
        self.assertTrue(len(particles) == 683)
        self.stream_handle.close()

    def test_mid_state_start(self):
        """
        Test starting the parser in a state in the middle of processing
        """
        file_path = os.path.join(RESOURCE_PATH, 'E0000001.DAT')
        self.stream_handle = open(file_path, 'rb')

        # Moving the file position past the header and two records
        new_state = {StateKey.POSITION: HEADER_BYTES+(WFP_E_GLOBAL_RECOVERED_ENG_DATA_SAMPLE_BYTES*2)}

        self.parser = GlobalWfpEFileParser(self.config, new_state, self.stream_handle,
                                           self.state_callback, self.pub_callback, self.exception_callback)

        particles = self.parser.get_records(4)

        # Should end up with 4 particles
        self.assertTrue(len(particles) == 4)

        # Compare with 6th data record
        self.assert_result(self.test_particle1, particles[3])

        self.stream_handle.close()

    def test_set_state(self):
        """
        Test changing to a new state after initializing the parser and
        reading data, as if new data has been found and the state has
        changed
        """
        filepath = os.path.join(RESOURCE_PATH, 'E0000001.DAT')
        self.stream_handle = open(filepath, 'rb')

        # Moving the file position past the header and two records
        new_state = {StateKey.POSITION: HEADER_BYTES+(WFP_E_GLOBAL_RECOVERED_ENG_DATA_SAMPLE_BYTES*2)}

        log.debug("new_state: %s", new_state)
        self.parser = GlobalWfpEFileParser(self.config, new_state, self.stream_handle,
                                           self.state_callback, self.pub_callback, self.exception_callback)

        particles = self.parser.get_records(4)

        # Should end up with 4 particles
        self.assertTrue(len(particles) == 4)

        # particles[3] is the 7th data record in the test file
        self.assert_result(self.test_particle1, particles[3])

        # Simulate State has changed:
        #   Move the file position past the header and three records
        #   Set the state of the existing parser
        new_state = {StateKey.POSITION: HEADER_BYTES+(WFP_E_GLOBAL_RECOVERED_ENG_DATA_SAMPLE_BYTES*3)}
        self.parser.set_state(new_state)

        particles = self.parser.get_records(10)

        # Should end up with 10 particles
        self.assertTrue(len(particles) == 10)

        # particles[9] is the 13th data record in test file
        self.assert_result(self.test_particle2, particles[9])

        self.stream_handle.close()

    def test_bad_data(self):
        """
        Ensure that bad data is skipped when it exists.
        """
        file_path = os.path.join(RESOURCE_PATH, 'E0000001-BAD-DATA.DAT')
        self.stream_handle = open(file_path, 'rb')

        self.parser = GlobalWfpEFileParser(self.config, self.start_state, self.stream_handle,
                                           self.state_callback, self.pub_callback, self.exception_callback)

        with self.assertRaises(SampleException):
            self.parser.get_records(1)

        self.stream_handle.close()

    def test_bad_header(self):
        """
        Ensure that bad header is skipped when it exists.
        """

        # This case tests against a header that does not match
        # 0000 0000 0000 0100 0000 0000 0000 0151
        file_path = os.path.join(RESOURCE_PATH, 'E0000001-BAD-HEADER1.DAT')
        self.stream_handle = open(file_path, 'rb')

        with self.assertRaises(SampleException):
            self.parser = GlobalWfpEFileParser(self.config, self.start_state, self.stream_handle,
                                               self.state_callback, self.pub_callback, self.exception_callback)

        self.stream_handle.close()

        # This case tests against a header that does not match global E-type data, but matches coastal
        # E-type data: 0001 0000 0000 0000 0001 0001 0000 0000
        file_path = os.path.join(RESOURCE_PATH, 'E0000001-BAD-HEADER2.DAT')
        self.stream_handle = open(file_path, 'rb')

        with self.assertRaises(SampleException):
            self.parser = GlobalWfpEFileParser(self.config, self.start_state, self.stream_handle,
                                               self.state_callback, self.pub_callback, self.exception_callback)

        self.stream_handle.close()

    def assert_result(self, test, particle):
        """
        Suite of tests to run against each returned particle and expected
        results of the same.  The test parameter should be a dictionary
        that contains the keys to be tested in the particle
        the 'internal_timestamp' and 'position' keys are
        treated differently than others but can be verified if supplied
        """
        log.debug("test arg: %s", test)
        particle_dict = particle.generate_dict()
        log.debug("particle_dict: %s", particle_dict)
        #for efficiency turn the particle values list of dictionaries into a dictionary
        particle_values = {}
        for param in particle_dict.get('values'):
            particle_values[param['value_id']] = param['value']

        # compare each key in the test to the data in the particle
        for key in test:
            test_data = test[key]

            #get the correct data to compare to the test
            if key == 'internal_timestamp':
                particle_data = particle.get_value('internal_timestamp')
                #the timestamp is in the header part of the particle

                log.info("internal_timestamp %d", particle_data)

            elif key == 'position':
                particle_data = self.state_callback_value['position']
                #position corresponds to the position in the file

                log.info("position %d", particle_data)

            else:
                particle_data = particle_values.get(key)
                #others are all part of the parsed values part of the particle

            if particle_data is None:
                #generally OK to ignore index keys in the test data, verify others

                log.warning("\nWarning: assert_result ignoring test key %s, does not exist in particle", key)
            else:
                if isinstance(test_data, float):
                    # slightly different test for these values as they are floats.
                    compare = numpy.abs(test_data - particle_data) <= 1e-5
                    self.assertTrue(compare)
                else:
                    # otherwise they are all ints and should be exactly equal
                    log.debug("test_data %s, particle_data %s", test_data, particle_data)
                    self.assertEqual(test_data, particle_data)

    @staticmethod
    def get_dict_from_yml(filename):
        """
        This utility routine loads the contents of a yml file
        into a dictionary
        """

        fid = open(os.path.join(RESOURCE_PATH, filename), 'r')
        result = yaml.load(fid)
        fid.close()

        return result