mirror of
https://github.com/Freika/dawarich.git
synced 2026-01-09 08:47:11 -05:00
* fix: move foreman to global gems to fix startup crash (#1971) * Update exporting code to stream points data to file in batches to red… (#1980) * Update exporting code to stream points data to file in batches to reduce memory usage * Update changelog * Update changelog * Feature/maplibre frontend (#1953) * Add a plan to use MapLibre GL JS for the frontend map rendering, replacing Leaflet * Implement phase 1 * Phases 1-3 + part of 4 * Fix e2e tests * Phase 6 * Implement fog of war * Phase 7 * Next step: fix specs, phase 7 done * Use our own map tiles * Extract v2 map logic to separate manager classes * Update settings panel on v2 map * Update v2 e2e tests structure * Reimplement location search in maps v2 * Update speed routes * Implement visits and places creation in v2 * Fix last failing test * Implement visits merging * Fix a routes e2e test and simplify the routes layer styling. * Extract js to modules from maps_v2_controller.js * Implement area creation * Fix spec problem * Fix some e2e tests * Implement live mode in v2 map * Update icons and panel * Extract some styles * Remove unused file * Start adding dark theme to popups on MapLibre maps * Make popups respect dark theme * Move v2 maps to maplibre namespace * Update v2 references to maplibre * Put place, area and visit info into side panel * Update API to use safe settings config method * Fix specs * Fix method name to config in SafeSettings and update usages accordingly * Add missing public files * Add handling for real time points * Fix remembering enabled/disabled layers of the v2 map * Fix lots of e2e tests * Add settings to select map version * Use maps/v2 as main path for MapLibre maps * Update routing * Update live mode * Update maplibre controller * Update changelog * Remove some console.log statements * Pull only necessary data for map v2 points * Feature/raw data archive (#2009) * 0.36.2 (#2007) * fix: move foreman to global gems to fix startup crash (#1971) * Update exporting code to stream points data to file in batches to red… (#1980) * Update exporting code to stream points data to file in batches to reduce memory usage * Update changelog * Update changelog * Feature/maplibre frontend (#1953) * Add a plan to use MapLibre GL JS for the frontend map rendering, replacing Leaflet * Implement phase 1 * Phases 1-3 + part of 4 * Fix e2e tests * Phase 6 * Implement fog of war * Phase 7 * Next step: fix specs, phase 7 done * Use our own map tiles * Extract v2 map logic to separate manager classes * Update settings panel on v2 map * Update v2 e2e tests structure * Reimplement location search in maps v2 * Update speed routes * Implement visits and places creation in v2 * Fix last failing test * Implement visits merging * Fix a routes e2e test and simplify the routes layer styling. * Extract js to modules from maps_v2_controller.js * Implement area creation * Fix spec problem * Fix some e2e tests * Implement live mode in v2 map * Update icons and panel * Extract some styles * Remove unused file * Start adding dark theme to popups on MapLibre maps * Make popups respect dark theme * Move v2 maps to maplibre namespace * Update v2 references to maplibre * Put place, area and visit info into side panel * Update API to use safe settings config method * Fix specs * Fix method name to config in SafeSettings and update usages accordingly * Add missing public files * Add handling for real time points * Fix remembering enabled/disabled layers of the v2 map * Fix lots of e2e tests * Add settings to select map version * Use maps/v2 as main path for MapLibre maps * Update routing * Update live mode * Update maplibre controller * Update changelog * Remove some console.log statements --------- Co-authored-by: Robin Tuszik <mail@robin.gg> * Remove esbuild scripts from package.json * Remove sideEffects field from package.json * Raw data archivation * Add tests * Fix tests * Fix tests * Update ExceptionReporter * Add schedule to run raw data archival job monthly * Change file structure for raw data archival feature * Update changelog and version for raw data archival feature --------- Co-authored-by: Robin Tuszik <mail@robin.gg> * Set raw_data to an empty hash instead of nil when archiving * Fix storage configuration and file extraction * Consider MIN_MINUTES_SPENT_IN_CITY during stats calculation (#2018) * Consider MIN_MINUTES_SPENT_IN_CITY during stats calculation * Remove raw data from visited cities api endpoint * Use user timezone to show dates on maps (#2020) * Fix/pre epoch time (#2019) * Use user timezone to show dates on maps * Limit timestamps to valid range to prevent database errors when users enter pre-epoch dates. * Limit timestamps to valid range to prevent database errors when users enter pre-epoch dates. * Fix tests failing due to new index on stats table * Fix failing specs * Update redis client configuration to support unix socket connection * Update changelog * Fix kml kmz import issues (#2023) * Fix kml kmz import issues * Refactor KML importer to improve readability and maintainability * Implement moving points in map v2 and fix route rendering logic to ma… (#2027) * Implement moving points in map v2 and fix route rendering logic to match map v1. * Fix route spec * fix(maplibre): update date format to ISO 8601 (#2029) * Add verification step to raw data archival process (#2028) * Add verification step to raw data archival process * Add actual verification of raw data archives after creation, and only clear raw_data for verified archives. * Fix failing specs * Eliminate zip-bomb risk * Fix potential memory leak in js * Return .keep files * Use Toast instead of alert for notifications * Add help section to navbar dropdown * Update changelog * Remove raw_data_archival_job * Ensure file is being closed properly after reading in Archivable concern * Add composite index to stats table if not exists * Update changelog * Update entrypoint to always sync static assets (not only new ones) * Add family layer to MapLibre maps (#2055) * Add family layer to MapLibre maps * Update migration * Don't show family toggle if feature is disabled * Update changelog * Return changelog * Update changelog * Update tailwind file * Bump sentry-rails from 6.0.0 to 6.1.0 (#1945) Bumps [sentry-rails](https://github.com/getsentry/sentry-ruby) from 6.0.0 to 6.1.0. - [Release notes](https://github.com/getsentry/sentry-ruby/releases) - [Changelog](https://github.com/getsentry/sentry-ruby/blob/master/CHANGELOG.md) - [Commits](https://github.com/getsentry/sentry-ruby/compare/6.0.0...6.1.0) --- updated-dependencies: - dependency-name: sentry-rails dependency-version: 6.1.0 dependency-type: direct:production update-type: version-update:semver-minor ... Signed-off-by: dependabot[bot] <support@github.com> Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> * Bump turbo-rails from 2.0.17 to 2.0.20 (#1944) Bumps [turbo-rails](https://github.com/hotwired/turbo-rails) from 2.0.17 to 2.0.20. - [Release notes](https://github.com/hotwired/turbo-rails/releases) - [Commits](https://github.com/hotwired/turbo-rails/compare/v2.0.17...v2.0.20) --- updated-dependencies: - dependency-name: turbo-rails dependency-version: 2.0.20 dependency-type: direct:production update-type: version-update:semver-patch ... Signed-off-by: dependabot[bot] <support@github.com> Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> Co-authored-by: Evgenii Burmakin <Freika@users.noreply.github.com> * Bump webmock from 3.25.1 to 3.26.1 (#1943) Bumps [webmock](https://github.com/bblimke/webmock) from 3.25.1 to 3.26.1. - [Release notes](https://github.com/bblimke/webmock/releases) - [Changelog](https://github.com/bblimke/webmock/blob/master/CHANGELOG.md) - [Commits](https://github.com/bblimke/webmock/compare/v3.25.1...v3.26.1) --- updated-dependencies: - dependency-name: webmock dependency-version: 3.26.1 dependency-type: direct:development update-type: version-update:semver-minor ... Signed-off-by: dependabot[bot] <support@github.com> Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> Co-authored-by: Evgenii Burmakin <Freika@users.noreply.github.com> * Bump brakeman from 7.1.0 to 7.1.1 (#1942) Bumps [brakeman](https://github.com/presidentbeef/brakeman) from 7.1.0 to 7.1.1. - [Release notes](https://github.com/presidentbeef/brakeman/releases) - [Changelog](https://github.com/presidentbeef/brakeman/blob/main/CHANGES.md) - [Commits](https://github.com/presidentbeef/brakeman/compare/v7.1.0...v7.1.1) --- updated-dependencies: - dependency-name: brakeman dependency-version: 7.1.1 dependency-type: direct:development update-type: version-update:semver-patch ... Signed-off-by: dependabot[bot] <support@github.com> Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> * Bump redis from 5.4.0 to 5.4.1 (#1941) Bumps [redis](https://github.com/redis/redis-rb) from 5.4.0 to 5.4.1. - [Changelog](https://github.com/redis/redis-rb/blob/master/CHANGELOG.md) - [Commits](https://github.com/redis/redis-rb/compare/v5.4.0...v5.4.1) --- updated-dependencies: - dependency-name: redis dependency-version: 5.4.1 dependency-type: direct:production update-type: version-update:semver-patch ... Signed-off-by: dependabot[bot] <support@github.com> Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> * Put import deletion into background job (#2045) * Put import deletion into background job * Update changelog * fix null type error and update heatmap styling (#2037) * fix: use constant weight for maplibre heatmap layer * fix null type, update heatmap styling * improve heatmap styling * fix typo * Fix stats calculation to recursively reduce H3 resolution when too ma… (#2065) * Fix stats calculation to recursively reduce H3 resolution when too many hexagons are generated * Update CHANGELOG.md * Validate trip start and end dates (#2066) * Validate trip start and end dates * Update changelog * Update migration to clean up duplicate stats before adding unique index * Fix fog of war radius setting being ignored and applying settings causing errors (#2068) * Update changelog * Add Rack::Deflater middleware to config/application.rb to enable gzip compression for responses. * Add composite index to points on user_id and timestamp * Deduplicte points based on timestamp brought to unix time * Fix/stats cache invalidation (#2072) * Fix family layer toggle in Map v2 settings for non-selfhosted env * Invalidate cache * Remove comments * Remove comment * Add new indicies to improve performance and remove unused ones to opt… (#2078) * Add new indicies to improve performance and remove unused ones to optimize database. * Remove comments * Update map search suggestions panel styling * Add yearly digest (#2073) * Add yearly digest * Rename YearlyDigests to Users::Digests * Minor changes * Update yearly digest layout and styles * Add flags and chart to email * Update colors * Fix layout of stats in yearly digest view * Remove cron job for yearly digest scheduling * Update CHANGELOG.md * Update digest email setting handling * Allow sharing digest for 1 week or 1 month * Change Digests Distance to Bigint * Fix settings page * Update changelog * Add RailsPulse (#2079) * Add RailsPulse * Add RailsPulse monitoring tool with basic HTTP authentication * Bring points_count to integer * Update migration and version * Update rubocop issues --------- Signed-off-by: dependabot[bot] <support@github.com> Co-authored-by: Robin Tuszik <mail@robin.gg> Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
278 lines
8.1 KiB
Ruby
278 lines
8.1 KiB
Ruby
# frozen_string_literal: true
|
|
|
|
require 'time'
|
|
|
|
class Users::ImportData::Points
|
|
BATCH_SIZE = 5000
|
|
|
|
def initialize(user, points_data = nil, batch_size: BATCH_SIZE, logger: Rails.logger)
|
|
@user = user
|
|
@points_data = points_data
|
|
@batch_size = batch_size
|
|
@logger = logger
|
|
|
|
@buffer = []
|
|
@total_created = 0
|
|
@processed_count = 0
|
|
@skipped_count = 0
|
|
@preloaded = false
|
|
|
|
@imports_lookup = {}
|
|
@countries_lookup = {}
|
|
@visits_lookup = {}
|
|
end
|
|
|
|
def call
|
|
return 0 unless points_data.respond_to?(:each)
|
|
|
|
logger.info "Importing #{collection_description(points_data)} points for user: #{user.email}"
|
|
|
|
enumerate(points_data) do |point_data|
|
|
add(point_data)
|
|
end
|
|
|
|
finalize
|
|
end
|
|
|
|
# Allows streamed usage by pushing a single point at a time.
|
|
def add(point_data)
|
|
preload_reference_data unless @preloaded
|
|
|
|
if valid_point_data?(point_data)
|
|
prepared_attributes = prepare_point_attributes(point_data)
|
|
|
|
if prepared_attributes
|
|
@buffer << prepared_attributes
|
|
@processed_count += 1
|
|
|
|
flush_batch if @buffer.size >= batch_size
|
|
else
|
|
@skipped_count += 1
|
|
end
|
|
else
|
|
@skipped_count += 1
|
|
logger.debug "Skipped point: invalid data - #{point_data.inspect}"
|
|
end
|
|
end
|
|
|
|
def finalize
|
|
preload_reference_data unless @preloaded
|
|
flush_batch
|
|
|
|
logger.info "Points import completed. Created: #{@total_created}. Processed #{@processed_count} valid points, skipped #{@skipped_count}."
|
|
@total_created
|
|
end
|
|
|
|
private
|
|
|
|
attr_reader :user, :points_data, :batch_size, :logger, :imports_lookup, :countries_lookup, :visits_lookup
|
|
|
|
def enumerate(collection, &block)
|
|
collection.each(&block)
|
|
end
|
|
|
|
def collection_description(collection)
|
|
return collection.size if collection.respond_to?(:size)
|
|
|
|
'streamed'
|
|
end
|
|
|
|
def flush_batch
|
|
return if @buffer.empty?
|
|
|
|
logger.debug "Processing batch of #{@buffer.size} points"
|
|
logger.debug "First point in batch: #{@buffer.first.inspect}"
|
|
|
|
normalized_batch = normalize_point_keys(@buffer)
|
|
|
|
begin
|
|
result = Point.upsert_all(
|
|
normalized_batch,
|
|
unique_by: %i[lonlat timestamp user_id],
|
|
returning: %w[id],
|
|
on_duplicate: :skip
|
|
)
|
|
|
|
batch_created = result&.count.to_i
|
|
@total_created += batch_created
|
|
|
|
logger.debug "Processed batch of #{@buffer.size} points, created #{batch_created}, total created: #{@total_created}"
|
|
rescue StandardError => e
|
|
logger.error "Failed to process point batch: #{e.message}"
|
|
logger.error "Batch size: #{@buffer.size}"
|
|
logger.error "First point in failed batch: #{@buffer.first.inspect}"
|
|
logger.error "Backtrace: #{e.backtrace.first(5).join('\n')}"
|
|
ensure
|
|
@buffer.clear
|
|
end
|
|
end
|
|
|
|
def preload_reference_data
|
|
return if @preloaded
|
|
|
|
logger.debug 'Preloading reference data for points import'
|
|
|
|
@imports_lookup = {}
|
|
user.imports.reload.each do |import|
|
|
string_key = [import.name, import.source, import.created_at.utc.iso8601]
|
|
integer_key = [import.name, Import.sources[import.source], import.created_at.utc.iso8601]
|
|
|
|
@imports_lookup[string_key] = import
|
|
@imports_lookup[integer_key] = import
|
|
end
|
|
logger.debug "Loaded #{user.imports.size} imports with #{@imports_lookup.size} lookup keys"
|
|
|
|
@countries_lookup = {}
|
|
Country.all.each do |country|
|
|
@countries_lookup[[country.name, country.iso_a2, country.iso_a3]] = country
|
|
@countries_lookup[country.name] = country
|
|
end
|
|
logger.debug "Loaded #{Country.count} countries for lookup"
|
|
|
|
@visits_lookup = user.visits.reload.index_by do |visit|
|
|
[visit.name, visit.started_at.utc.iso8601, visit.ended_at.utc.iso8601]
|
|
end
|
|
logger.debug "Loaded #{@visits_lookup.size} visits for lookup"
|
|
|
|
@preloaded = true
|
|
end
|
|
|
|
def normalize_point_keys(points)
|
|
all_keys = points.flat_map(&:keys).uniq
|
|
|
|
points.map do |point|
|
|
all_keys.each_with_object({}) do |key, normalized|
|
|
normalized[key] = point[key]
|
|
end
|
|
end
|
|
end
|
|
|
|
def valid_point_data?(point_data)
|
|
return false unless point_data.is_a?(Hash)
|
|
return false unless point_data['timestamp'].present?
|
|
|
|
has_lonlat = point_data['lonlat'].present? && point_data['lonlat'].is_a?(String) && point_data['lonlat'].start_with?('POINT(')
|
|
has_coordinates = point_data['longitude'].present? && point_data['latitude'].present?
|
|
|
|
has_lonlat || has_coordinates
|
|
rescue StandardError => e
|
|
logger.debug "Point validation failed: #{e.message} for data: #{point_data.inspect}"
|
|
false
|
|
end
|
|
|
|
def prepare_point_attributes(point_data)
|
|
attributes = point_data.except(
|
|
'created_at',
|
|
'updated_at',
|
|
'import_reference',
|
|
'country_info',
|
|
'visit_reference',
|
|
'country'
|
|
)
|
|
|
|
ensure_lonlat_field(attributes, point_data)
|
|
|
|
attributes.delete('longitude')
|
|
attributes.delete('latitude')
|
|
|
|
attributes['user_id'] = user.id
|
|
attributes['created_at'] = Time.current
|
|
attributes['updated_at'] = Time.current
|
|
|
|
resolve_import_reference(attributes, point_data['import_reference'])
|
|
resolve_country_reference(attributes, point_data['country_info'])
|
|
resolve_visit_reference(attributes, point_data['visit_reference'])
|
|
|
|
result = attributes.symbolize_keys
|
|
|
|
logger.debug "Prepared point attributes: #{result.slice(:lonlat, :timestamp, :import_id, :country_id, :visit_id)}"
|
|
result
|
|
rescue StandardError => e
|
|
ExceptionReporter.call(e, 'Failed to prepare point attributes')
|
|
nil
|
|
end
|
|
|
|
def resolve_import_reference(attributes, import_reference)
|
|
return unless import_reference.is_a?(Hash)
|
|
|
|
created_at = normalize_timestamp_for_lookup(import_reference['created_at'])
|
|
|
|
import_key = [
|
|
import_reference['name'],
|
|
import_reference['source'],
|
|
created_at
|
|
]
|
|
|
|
import = imports_lookup[import_key]
|
|
if import
|
|
attributes['import_id'] = import.id
|
|
logger.debug "Resolved import reference: #{import_reference['name']} -> #{import.id}"
|
|
else
|
|
logger.debug "Import not found for reference: #{import_reference.inspect}"
|
|
logger.debug "Available imports: #{imports_lookup.keys.inspect}"
|
|
end
|
|
end
|
|
|
|
def resolve_country_reference(attributes, country_info)
|
|
return unless country_info.is_a?(Hash)
|
|
|
|
country_key = [country_info['name'], country_info['iso_a2'], country_info['iso_a3']]
|
|
country = countries_lookup[country_key]
|
|
|
|
country = countries_lookup[country_info['name']] if country.nil? && country_info['name'].present?
|
|
|
|
if country
|
|
attributes['country_id'] = country.id
|
|
logger.debug "Resolved country reference: #{country_info['name']} -> #{country.id}"
|
|
else
|
|
logger.debug "Country not found for: #{country_info.inspect}"
|
|
end
|
|
end
|
|
|
|
def resolve_visit_reference(attributes, visit_reference)
|
|
return unless visit_reference.is_a?(Hash)
|
|
|
|
started_at = normalize_timestamp_for_lookup(visit_reference['started_at'])
|
|
ended_at = normalize_timestamp_for_lookup(visit_reference['ended_at'])
|
|
|
|
visit_key = [
|
|
visit_reference['name'],
|
|
started_at,
|
|
ended_at
|
|
]
|
|
|
|
visit = visits_lookup[visit_key]
|
|
if visit
|
|
attributes['visit_id'] = visit.id
|
|
logger.debug "Resolved visit reference: #{visit_reference['name']} -> #{visit.id}"
|
|
else
|
|
logger.debug "Visit not found for reference: #{visit_reference.inspect}"
|
|
logger.debug "Available visits: #{visits_lookup.keys.inspect}"
|
|
end
|
|
end
|
|
|
|
def ensure_lonlat_field(attributes, point_data)
|
|
return unless attributes['lonlat'].blank? && point_data['longitude'].present? && point_data['latitude'].present?
|
|
|
|
longitude = point_data['longitude'].to_f
|
|
latitude = point_data['latitude'].to_f
|
|
attributes['lonlat'] = "POINT(#{longitude} #{latitude})"
|
|
logger.debug "Reconstructed lonlat: #{attributes['lonlat']}"
|
|
end
|
|
|
|
def normalize_timestamp_for_lookup(timestamp)
|
|
return nil if timestamp.blank?
|
|
|
|
case timestamp
|
|
when String
|
|
Time.parse(timestamp).utc.iso8601
|
|
when Time, DateTime
|
|
timestamp.utc.iso8601
|
|
else
|
|
timestamp.to_s
|
|
end
|
|
rescue StandardError => e
|
|
logger.debug "Failed to normalize timestamp #{timestamp}: #{e.message}"
|
|
timestamp.to_s
|
|
end
|
|
end
|