You can not select more than 25 topics
			Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
		
		
		
		
		
			
		
			
				
					
					
						
							198 lines
						
					
					
						
							5.1 KiB
						
					
					
				
			
		
		
	
	
							198 lines
						
					
					
						
							5.1 KiB
						
					
					
				class BulkUpload::Sales::Validator | 
						|
  include ActiveModel::Validations | 
						|
  include Rails.application.routes.url_helpers | 
						|
 | 
						|
  attr_reader :bulk_upload, :path | 
						|
 | 
						|
  validate :validate_file_not_empty | 
						|
  validate :validate_field_numbers_count | 
						|
  validate :validate_max_columns | 
						|
  validate :validate_missing_required_headers | 
						|
  validate :validate_correct_template | 
						|
 | 
						|
  def initialize(bulk_upload:, path:) | 
						|
    @bulk_upload = bulk_upload | 
						|
    @path = path | 
						|
  end | 
						|
 | 
						|
  def call | 
						|
    row_parsers.each(&:valid?) | 
						|
 | 
						|
    validate_duplicate_rows if FeatureToggle.bulk_upload_duplicate_log_check_enabled? | 
						|
 | 
						|
    row_parsers.each_with_index do |row_parser, index| | 
						|
      row = index + row_offset + 1 | 
						|
 | 
						|
      row_parser.errors.each do |error| | 
						|
        col = csv_parser.column_for_field(error.attribute.to_s) | 
						|
 | 
						|
        bulk_upload.bulk_upload_errors.create!( | 
						|
          field: error.attribute, | 
						|
          error: error.message, | 
						|
          purchaser_code: row_parser.purchaser_code, | 
						|
          row:, | 
						|
          cell: "#{col}#{row}", | 
						|
          col:, | 
						|
          category: error.options[:category], | 
						|
        ) | 
						|
      end | 
						|
    end | 
						|
  end | 
						|
 | 
						|
  def block_log_creation_reason | 
						|
    return "setup_errors" if any_setup_errors? | 
						|
 | 
						|
    if row_parsers.any?(&:block_log_creation?) | 
						|
      Sentry.capture_message("Bulk upload log creation blocked: #{bulk_upload.id}.") | 
						|
      return "row_parser_block_log_creation" | 
						|
    end | 
						|
 | 
						|
    if any_logs_already_exist? && FeatureToggle.bulk_upload_duplicate_log_check_enabled? | 
						|
      Sentry.capture_message("Bulk upload log creation blocked due to duplicate logs: #{bulk_upload.id}.") | 
						|
      return "duplicate_logs" | 
						|
    end | 
						|
 | 
						|
    row_parsers.each do |row_parser| | 
						|
      row_parser.log.blank_invalid_non_setup_fields! | 
						|
    end | 
						|
 | 
						|
    if any_logs_invalid? | 
						|
      Sentry.capture_message("Bulk upload log creation blocked due to invalid logs after blanking non setup fields: #{bulk_upload.id}.") | 
						|
      "logs_invalid" | 
						|
    end | 
						|
  end | 
						|
 | 
						|
  def any_setup_errors? | 
						|
    bulk_upload | 
						|
      .bulk_upload_errors | 
						|
      .where(category: "setup") | 
						|
      .count | 
						|
      .positive? | 
						|
  end | 
						|
 | 
						|
  def any_logs_already_exist? | 
						|
    row_parsers.any?(&:log_already_exists?) | 
						|
  end | 
						|
 | 
						|
  def soft_validation_errors_only? | 
						|
    errors = bulk_upload.bulk_upload_errors | 
						|
    errors.count == errors.where(category: "soft_validation").count && errors.count.positive? | 
						|
  end | 
						|
 | 
						|
  def total_logs_count | 
						|
    csv_parser.body_rows.count | 
						|
  end | 
						|
 | 
						|
private | 
						|
 | 
						|
  # n^2 algo | 
						|
  def validate_duplicate_rows | 
						|
    row_parsers.each do |rp| | 
						|
      dupe = row_parsers.reject { |r| r.object_id.equal?(rp.object_id) }.any? do |rp_counter| | 
						|
        rp.spreadsheet_duplicate_hash == rp_counter.spreadsheet_duplicate_hash | 
						|
      end | 
						|
 | 
						|
      if dupe | 
						|
        rp.add_duplicate_found_in_spreadsheet_errors | 
						|
      end | 
						|
    end | 
						|
  end | 
						|
 | 
						|
  def any_logs_invalid? | 
						|
    row_parsers.any? { |row_parser| row_parser.log.invalid? } | 
						|
  end | 
						|
 | 
						|
  def csv_parser | 
						|
    @csv_parser ||= case bulk_upload.year | 
						|
                    when 2023 | 
						|
                      BulkUpload::Sales::Year2023::CsvParser.new(path:) | 
						|
                    when 2024 | 
						|
                      BulkUpload::Sales::Year2024::CsvParser.new(path:) | 
						|
                    else | 
						|
                      raise "csv parser not found" | 
						|
                    end | 
						|
  end | 
						|
 | 
						|
  def row_offset | 
						|
    csv_parser.row_offset | 
						|
  end | 
						|
 | 
						|
  def col_offset | 
						|
    csv_parser.col_offset | 
						|
  end | 
						|
 | 
						|
  def field_number_for_attribute(attribute) | 
						|
    attribute.to_s.split("_").last.to_i | 
						|
  end | 
						|
 | 
						|
  def cols | 
						|
    csv_parser.cols | 
						|
  end | 
						|
 | 
						|
  def row_parsers | 
						|
    return @row_parsers if @row_parsers | 
						|
 | 
						|
    @row_parsers = csv_parser.row_parsers | 
						|
 | 
						|
    @row_parsers.each do |row_parser| | 
						|
      row_parser.bulk_upload = bulk_upload | 
						|
    end | 
						|
 | 
						|
    @row_parsers | 
						|
  end | 
						|
 | 
						|
  def rows | 
						|
    csv_parser.rows | 
						|
  end | 
						|
 | 
						|
  def body_rows | 
						|
    csv_parser.body_rows | 
						|
  end | 
						|
 | 
						|
  def validate_file_not_empty | 
						|
    if File.size(path).zero? || csv_parser.body_rows.flatten.compact.empty? | 
						|
      errors.add(:base, I18n.t("validations.sales.#{@bulk_upload.year}.bulk_upload.blank_file")) | 
						|
 | 
						|
      halt_validations! | 
						|
    end | 
						|
  end | 
						|
 | 
						|
  def validate_max_columns | 
						|
    return if halt_validations? | 
						|
 | 
						|
    column_count = rows.map(&:size).max | 
						|
 | 
						|
    errors.add(:base, I18n.t("validations.sales.#{@bulk_upload.year}.bulk_upload.wrong_template.over_max_column_count")) if column_count > csv_parser.class::MAX_COLUMNS | 
						|
  end | 
						|
 | 
						|
  def validate_correct_template | 
						|
    return if halt_validations? | 
						|
 | 
						|
    errors.add(:base, I18n.t("validations.sales.#{@bulk_upload.year}.bulk_upload.wrong_template.wrong_template")) if csv_parser.wrong_template_for_year? | 
						|
  end | 
						|
 | 
						|
  def validate_missing_required_headers | 
						|
    return if halt_validations? | 
						|
 | 
						|
    if csv_parser.missing_required_headers? | 
						|
      errors.add :base, I18n.t("validations.sales.#{@bulk_upload.year}.bulk_upload.wrong_template.no_headers", guidance_link: bulk_upload_sales_log_url(id: "guidance", form: { year: bulk_upload.year }, host: ENV["APP_HOST"], anchor: "using-the-bulk-upload-template")) | 
						|
    end | 
						|
  end | 
						|
 | 
						|
  def validate_field_numbers_count | 
						|
    return if halt_validations? | 
						|
 | 
						|
    unless csv_parser.correct_field_count? | 
						|
      errors.add(:base, I18n.t("validations.sales.#{@bulk_upload.year}.bulk_upload.wrong_template.wrong_field_numbers_count")) | 
						|
      halt_validations! | 
						|
    end | 
						|
  end | 
						|
 | 
						|
  def halt_validations! | 
						|
    @halt_validations = true | 
						|
  end | 
						|
 | 
						|
  def halt_validations? | 
						|
    @halt_validations ||= false | 
						|
  end | 
						|
end
 | 
						|
 |