dawarich/app/services/users/export_data.rb
2025-06-25 21:44:36 +02:00

434 lines
13 KiB
Ruby

# frozen_string_literal: true
require 'zip'
# Users::ExportData - Exports complete user data with preserved relationships
#
# Output JSON Structure Example:
# {
# "settings": {
# "distance_unit": "km",
# "timezone": "UTC",
# "immich_url": "https://immich.example.com",
# // ... other user settings
# },
# "areas": [
# {
# "name": "Home",
# "latitude": "40.7128",
# "longitude": "-74.0060",
# "radius": 100,
# "created_at": "2024-01-01T00:00:00Z"
# }
# ],
# "imports": [
# {
# "name": "2023_MARCH.json",
# "source": "google_semantic_history",
# "created_at": "2024-01-01T00:00:00Z",
# "processed": true,
# "points_count": 1500,
# "file_name": "import_1_2023_MARCH.json",
# "original_filename": "2023_MARCH.json",
# "file_size": 2048576,
# "content_type": "application/json"
# }
# ],
# "exports": [
# {
# "name": "export_2024-01-01_to_2024-01-31.json",
# "status": "completed",
# "file_format": "json",
# "file_type": "points",
# "created_at": "2024-02-01T00:00:00Z",
# "file_name": "export_1_export_2024-01-01_to_2024-01-31.json",
# "original_filename": "export_2024-01-01_to_2024-01-31.json",
# "file_size": 1048576,
# "content_type": "application/json"
# }
# ],
# "trips": [
# {
# "name": "Business Trip to NYC",
# "started_at": "2024-01-15T08:00:00Z",
# "ended_at": "2024-01-18T20:00:00Z",
# "distance": 1245.67,
# "created_at": "2024-01-19T00:00:00Z"
# }
# ],
# "stats": [
# {
# "year": 2024,
# "month": 1,
# "distance": 456.78,
# "toponyms": [
# {"country": "United States", "cities": [{"city": "New York"}]}
# ],
# "created_at": "2024-02-01T00:00:00Z"
# }
# ],
# "notifications": [
# {
# "kind": "info",
# "title": "Import completed",
# "content": "Your data import has been processed successfully",
# "read": true,
# "created_at": "2024-01-01T12:00:00Z"
# }
# ],
# "points": [
# {
# "battery_status": "charging",
# "battery": 85,
# "timestamp": 1704067200,
# "altitude": 15.5,
# "velocity": 25.5,
# "accuracy": 5.0,
# "created_at": "2024-01-01T00:00:00Z",
# "import_reference": {
# "name": "2023_MARCH.json",
# "source": "google_semantic_history",
# "created_at": "2024-01-01T00:00:00Z"
# },
# "country_info": {
# "name": "United States",
# "iso_a2": "US",
# "iso_a3": "USA"
# },
# "visit_reference": {
# "name": "Work Visit",
# "started_at": "2024-01-01T08:00:00Z",
# "ended_at": "2024-01-01T17:00:00Z"
# }
# },
# {
# // Example of point without relationships (edge cases)
# "timestamp": 1704070800,
# "altitude": 10.0,
# "import_reference": null, // Orphaned point
# "country_info": null, // No country data
# "visit_reference": null // Not part of a visit
# }
# ],
# "visits": [
# {
# "name": "Work Visit",
# "started_at": "2024-01-01T08:00:00Z",
# "ended_at": "2024-01-01T17:00:00Z",
# "duration": 32400,
# "status": "suggested",
# "place_reference": {
# "name": "Office Building",
# "latitude": "40.7589",
# "longitude": "-73.9851",
# "source": "manual"
# }
# },
# {
# // Example of visit without place
# "name": "Unknown Location",
# "started_at": "2024-01-02T10:00:00Z",
# "ended_at": "2024-01-02T12:00:00Z",
# "place_reference": null // No associated place
# }
# ],
# "places": [
# {
# "name": "Office Building",
# "lonlat": "POINT(-73.9851 40.7589)",
# "source": "manual",
# "geodata": {"properties": {"name": "Office Building"}},
# "created_at": "2024-01-01T00:00:00Z"
# }
# ]
# }
#
# Import Strategy Notes:
# 1. Countries: Look up by name/ISO codes, create if missing
# 2. Imports: Match by name + source + created_at, create new import records
# 3. Places: Match by name + coordinates, create if missing
# 4. Visits: Match by name + timestamps + place_reference, create if missing
# 5. Points: Import with reconstructed foreign keys from references
# 6. Files: Import files are available in the files/ directory with names from file_name fields
class Users::ExportData
def initialize(user)
@user = user
@export_directory = export_directory
@files_directory = files_directory
end
def export
# TODO: Implement
# 1. Export user settings
# 2. Export user points
# 4. Export user visits
# 8. Export user places
# 11. Zip all the files
FileUtils.mkdir_p(files_directory)
begin
data = {}
data[:settings] = user.safe_settings.settings
data[:areas] = serialized_areas
data[:imports] = serialized_imports
data[:exports] = serialized_exports
data[:trips] = serialized_trips
data[:stats] = serialized_stats
data[:notifications] = serialized_notifications
data[:points] = serialized_points
data[:visits] = serialized_visits
data[:places] = serialized_places
json_file_path = export_directory.join('data.json')
File.write(json_file_path, data.to_json)
zip_file_path = export_directory.join('export.zip')
create_zip_archive(zip_file_path)
# Move the zip file to a final location (e.g., tmp root) before cleanup
final_zip_path = Rails.root.join('tmp', "#{user.email}_export_#{Time.current.strftime('%Y%m%d_%H%M%S')}.zip")
FileUtils.mv(zip_file_path, final_zip_path)
final_zip_path
ensure
cleanup_temporary_files
end
end
private
attr_reader :user
def export_directory
@export_directory ||= Rails.root.join('tmp', "#{user.email}_#{Time.current.strftime('%Y%m%d_%H%M%S')}")
end
def files_directory
@files_directory ||= export_directory.join('files')
end
def serialized_exports
exports_data = user.exports.includes(:file_attachment).map do |export|
process_export(export)
end
exports_data
end
def process_export(export)
Rails.logger.info "Processing export #{export.name}"
# Only include essential attributes, exclude any potentially large fields
export_hash = export.as_json(except: %w[user_id])
if export.file.attached?
add_file_data_to_export(export, export_hash)
else
add_empty_file_data_to_export(export_hash)
end
Rails.logger.info "Export #{export.name} processed"
export_hash
end
def add_file_data_to_export(export, export_hash)
sanitized_filename = generate_sanitized_export_filename(export)
file_path = files_directory.join(sanitized_filename)
begin
download_and_save_export_file(export, file_path)
add_file_metadata_to_export(export, export_hash, sanitized_filename)
rescue StandardError => e
Rails.logger.error "Failed to download export file #{export.id}: #{e.message}"
export_hash['file_error'] = "Failed to download: #{e.message}"
end
end
def add_empty_file_data_to_export(export_hash)
export_hash['file_name'] = nil
export_hash['original_filename'] = nil
end
def generate_sanitized_export_filename(export)
"export_#{export.id}_#{export.file.blob.filename}".gsub(/[^0-9A-Za-z._-]/, '_')
end
def download_and_save_export_file(export, file_path)
file_content = Imports::SecureFileDownloader.new(export.file).download_with_verification
File.write(file_path, file_content, mode: 'wb')
end
def add_file_metadata_to_export(export, export_hash, sanitized_filename)
export_hash['file_name'] = sanitized_filename
export_hash['original_filename'] = export.file.blob.filename.to_s
export_hash['file_size'] = export.file.blob.byte_size
export_hash['content_type'] = export.file.blob.content_type
end
def serialized_imports
imports_data = user.imports.includes(:file_attachment).map do |import|
process_import(import)
end
imports_data
end
def process_import(import)
Rails.logger.info "Processing import #{import.name}"
# Only include essential attributes, exclude large fields like raw_data
import_hash = import.as_json(except: %w[user_id raw_data])
if import.file.attached?
add_file_data_to_import(import, import_hash)
else
add_empty_file_data_to_import(import_hash)
end
Rails.logger.info "Import #{import.name} processed"
import_hash
end
def add_file_data_to_import(import, import_hash)
sanitized_filename = generate_sanitized_filename(import)
file_path = files_directory.join(sanitized_filename)
begin
download_and_save_import_file(import, file_path)
add_file_metadata_to_import(import, import_hash, sanitized_filename)
rescue StandardError => e
Rails.logger.error "Failed to download import file #{import.id}: #{e.message}"
import_hash['file_error'] = "Failed to download: #{e.message}"
end
end
def add_empty_file_data_to_import(import_hash)
import_hash['file_name'] = nil
import_hash['original_filename'] = nil
end
def generate_sanitized_filename(import)
"import_#{import.id}_#{import.file.blob.filename}".gsub(/[^0-9A-Za-z._-]/, '_')
end
def download_and_save_import_file(import, file_path)
file_content = Imports::SecureFileDownloader.new(import.file).download_with_verification
File.write(file_path, file_content, mode: 'wb')
end
def add_file_metadata_to_import(import, import_hash, sanitized_filename)
import_hash['file_name'] = sanitized_filename
import_hash['original_filename'] = import.file.blob.filename.to_s
import_hash['file_size'] = import.file.blob.byte_size
import_hash['content_type'] = import.file.blob.content_type
end
def create_zip_archive(zip_file_path)
Zip::File.open(zip_file_path, Zip::File::CREATE) do |zipfile|
Dir.glob(export_directory.join('**', '*')).each do |file|
next if File.directory?(file) || file == zip_file_path.to_s
relative_path = file.sub(export_directory.to_s + '/', '')
zipfile.add(relative_path, file)
end
end
end
def cleanup_temporary_files
return unless File.directory?(export_directory)
Rails.logger.info "Cleaning up temporary export directory: #{export_directory}"
FileUtils.rm_rf(export_directory)
rescue StandardError => e
Rails.logger.error "Failed to cleanup temporary files: #{e.message}"
# Don't re-raise the error as cleanup failure shouldn't break the export
end
def serialized_trips
user.trips.as_json(except: %w[user_id id])
end
def serialized_areas
user.areas.as_json(except: %w[user_id id])
end
def serialized_stats
user.stats.as_json(except: %w[user_id id])
end
def serialized_notifications
user.notifications.as_json(except: %w[user_id id])
end
def serialized_points
# Include relationship with country to avoid N+1 queries
user.tracked_points.includes(:country, :import, :visit).find_each(batch_size: 1000).map do |point|
point_hash = point.as_json(except: %w[user_id import_id country_id visit_id id])
# Replace import_id with import natural key
if point.import
point_hash['import_reference'] = {
'name' => point.import.name,
'source' => point.import.source,
'created_at' => point.import.created_at.iso8601
}
else
point_hash['import_reference'] = nil
end
# Replace country_id with country information
if point.country
point_hash['country_info'] = {
'name' => point.country.name,
'iso_a2' => point.country.iso_a2,
'iso_a3' => point.country.iso_a3
}
else
point_hash['country_info'] = nil
end
# Replace visit_id with visit natural key
if point.visit
point_hash['visit_reference'] = {
'name' => point.visit.name,
'started_at' => point.visit.started_at&.iso8601,
'ended_at' => point.visit.ended_at&.iso8601
}
else
point_hash['visit_reference'] = nil
end
point_hash
end
end
def serialized_visits
user.visits.includes(:place).map do |visit|
visit_hash = visit.as_json(except: %w[user_id place_id id])
# Replace place_id with place natural key
if visit.place
visit_hash['place_reference'] = {
'name' => visit.place.name,
'latitude' => visit.place.lat.to_s,
'longitude' => visit.place.lon.to_s,
'source' => visit.place.source
}
else
visit_hash['place_reference'] = nil
end
visit_hash
end
end
def serialized_places
user.places.as_json(except: %w[user_id id])
end
end