Class SampleDataForSplitPoints

    • Constructor Detail

      • SampleDataForSplitPoints

        public SampleDataForSplitPoints()
    • Method Detail

      • validate

        public uk.gov.gchq.koryphe.ValidationResult validate()
        Description copied from interface: Operation
        Validates an operation. This should be used to validate that fields have been be configured correctly. By default no validation is applied. Override this method to implement validation.
        Specified by:
        validate in interface Operation
        Returns:
        validation result.
      • isValidate

        public boolean isValidate()
      • setValidate

        public void setValidate​(boolean validate)
      • getNumSplits

        public Integer getNumSplits()
      • setNumSplits

        public void setNumSplits​(Integer numSplits)
      • getProportionToSample

        public float getProportionToSample()
      • setProportionToSample

        public void setProportionToSample​(float proportionToSample)
      • getJobInitialiser

        public JobInitialiser getJobInitialiser()
        Description copied from interface: MapReduce
        A job initialiser allows additional job initialisation to be carried out in addition to that done by the store. Most stores will probably require the Job Input to be configured in this initialiser as this is specific to the type of data store in Hdfs. For Avro data see AvroJobInitialiser. For Text data see TextJobInitialiser.
        Specified by:
        getJobInitialiser in interface MapReduce
        Returns:
        the job initialiser
      • setUseProvidedSplits

        public void setUseProvidedSplits​(boolean useProvidedSplits)
        Specified by:
        setUseProvidedSplits in interface MapReduce
      • getPartitioner

        public Class<? extends org.apache.hadoop.mapreduce.Partitioner> getPartitioner()
        Specified by:
        getPartitioner in interface MapReduce
      • setPartitioner

        public void setPartitioner​(Class<? extends org.apache.hadoop.mapreduce.Partitioner> partitioner)
        Specified by:
        setPartitioner in interface MapReduce
      • getCompressionCodec

        public Class<? extends org.apache.hadoop.io.compress.CompressionCodec> getCompressionCodec()
      • setCompressionCodec

        public void setCompressionCodec​(Class<? extends org.apache.hadoop.io.compress.CompressionCodec> compressionCodec)
      • getOptions

        public Map<String,​String> getOptions()
        Specified by:
        getOptions in interface Operation
        Returns:
        the operation options. This may contain store specific options such as authorisation strings or and other properties required for the operation to be executed. Note these options will probably not be interpreted in the same way by every store implementation.
      • setOptions

        public void setOptions​(Map<String,​String> options)
        Specified by:
        setOptions in interface Operation
        Parameters:
        options - the operation options. This may contain store specific options such as authorisation strings or and other properties required for the operation to be executed. Note these options will probably not be interpreted in the same way by every store implementation.
      • shallowClone

        public SampleDataForSplitPoints shallowClone()
        Description copied from interface: Operation
        Operation implementations should ensure a ShallowClone method is implemented. Performs a shallow clone. Creates a new instance and copies the fields across. It does not clone the fields. If the operation contains nested operations, these must also be cloned.
        Specified by:
        shallowClone in interface Operation
        Returns:
        shallow clone