diff --git a/.env.sample b/.env.sample index fdfbfe4..97bd1f8 100644 --- a/.env.sample +++ b/.env.sample @@ -1,3 +1,23 @@ +# ================================================================================================ +# BrickTracker Configuration File +# ================================================================================================ +# +# FILE LOCATION (v1.3+): +# ---------------------- +# This file can be placed in two locations: +# 1. data/.env (RECOMMENDED) - Included in data volume backup, settings persist via admin panel +# 2. .env (root) - Backward compatible +# +# Priority: data/.env > .env (root) +# +# The application automatically detects and uses the correct location at runtime. +# +# For Docker: +# - Recommended: Place this file as data/.env (included in data volume) +# - Backward compatible: Keep as .env in root (add "env_file: .env" to compose.yaml) +# +# ================================================================================================ +# # Note on *_DEFAULT_ORDER # If set, it will append a direct ORDER BY to the SQL query # while listing objects. You can look at the structure of the SQLite database to @@ -32,15 +52,20 @@ # Default: https://www.bricklink.com/v2/catalog/catalogitem.page?P={part}&C={color} # BK_BRICKLINK_LINK_PART_PATTERN= +# Optional: Pattern of the link to Bricklink for a set. Will be passed to Python .format() +# Supports {set_num} parameter. Set numbers in format like '10255-1' are used. +# Default: https://www.bricklink.com/v2/catalog/catalogitem.page?S={set_num} +# BK_BRICKLINK_LINK_SET_PATTERN= + # Optional: Display Bricklink links wherever applicable # Default: false # BK_BRICKLINK_LINKS=true -# Optional: Path to the database. +# Optional: Path to the database, relative to '/app/' folder # Useful if you need it mounted in a Docker volume. Keep in mind that it will not # do any check on the existence of the path, or if it is dangerous. -# Default: ./app.db -# BK_DATABASE_PATH=/var/lib/bricktracker/app.db +# Default: data/app.db +# BK_DATABASE_PATH=data/app.db # Optional: Format of the timestamp added to the database file when downloading it # Check https://docs.python.org/3/library/time.html#time.strftime for format details @@ -81,9 +106,9 @@ # Default: .pdf # BK_INSTRUCTIONS_ALLOWED_EXTENSIONS=.pdf, .docx, .png -# Optional: Folder where to store the instructions, relative to the '/app/static/' folder -# Default: instructions -# BK_INSTRUCTIONS_FOLDER=/var/lib/bricktracker/instructions/ +# Optional: Folder where to store the instructions, relative to '/app/' folder +# Default: data/instructions +# BK_INSTRUCTIONS_FOLDER=data/instructions # Optional: Hide the 'Add' entry from the menu. Does not disable the route. # Default: false @@ -97,6 +122,14 @@ # Default: false # BK_HIDE_ADMIN=true +# Optional: Admin sections to expand by default (comma-separated list) +# Valid sections: authentication, instructions, image, theme, retired, metadata, owner, purchase_location, status, storage, tag, database +# Default: database (maintains original behavior with database section expanded) +# Examples: +# BK_ADMIN_DEFAULT_EXPANDED_SECTIONS=database,theme +# BK_ADMIN_DEFAULT_EXPANDED_SECTIONS=instructions,metadata +# BK_ADMIN_DEFAULT_EXPANDED_SECTIONS= (all sections collapsed) + # Optional: Hide the 'Instructions' entry from the menu. Does not disable the route. # Default: false # BK_HIDE_ALL_INSTRUCTIONS=true @@ -122,6 +155,10 @@ # Default: false # BK_HIDE_ALL_STORAGES=true +# Optional: Hide the 'Statistics' entry from the menu. Does not disable the route. +# Default: false +# BK_HIDE_STATISTICS=true + # Optional: Hide the 'Instructions' entry in a Set card # Default: false # BK_HIDE_SET_INSTRUCTIONS=true @@ -134,21 +171,33 @@ # Default: false # BK_HIDE_TABLE_MISSING_PARTS=true +# Optional: Hide the 'Checked' column from the parts table. +# Default: false +# BK_HIDE_TABLE_CHECKED_PARTS=true + # Optional: Hide the 'Wishlist' entry from the menu. Does not disable the route. # Default: false # BK_HIDE_WISHES=true # Optional: Change the default order of minifigures. By default ordered by insertion order. # Useful column names for this option are: -# - "rebrickable_minifigures"."figure": minifigure ID (fig-xxxxx) -# - "rebrickable_minifigures"."number": minifigure ID as an integer (xxxxx) +# - "rebrickable_minifigures"."figure": minifigure ID (e.g., "fig-001234") +# - "rebrickable_minifigures"."number": minifigure ID as an integer (e.g., 1234) # - "rebrickable_minifigures"."name": minifigure name +# - "rebrickable_minifigures"."number_of_parts": number of parts in the minifigure +# - "bricktracker_minifigures"."quantity": quantity owned +# - "total_missing": number of missing parts (composite field) +# - "total_damaged": number of damaged parts (composite field) +# - "total_quantity": total quantity across all sets (composite field) +# - "total_sets": number of sets containing this minifigure (composite field) # Default: "rebrickable_minifigures"."name" ASC -# BK_MINIFIGURES_DEFAULT_ORDER="rebrickable_minifigures"."name" ASC +# Examples: +# BK_MINIFIGURES_DEFAULT_ORDER="rebrickable_minifigures"."number" DESC +# BK_MINIFIGURES_DEFAULT_ORDER="total_missing" DESC, "rebrickable_minifigures"."name" ASC -# Optional: Folder where to store the minifigures images, relative to the '/app/static/' folder -# Default: minifigs -# BK_MINIFIGURES_FOLDER=minifigures +# Optional: Folder where to store the minifigures images, relative to '/app/' folder +# Default: data/minifigures +# BK_MINIFIGURES_FOLDER=data/minifigures # Optional: Disable threading on the task executed by the socket. # You should not need to change this parameter unless you are debugging something with the @@ -158,17 +207,67 @@ # Optional: Change the default order of parts. By default ordered by insertion order. # Useful column names for this option are: -# - "bricktracker_parts"."part": part number -# - "bricktracker_parts"."spare": part is a spare part +# - "bricktracker_parts"."part": part number (e.g., "3001") +# - "bricktracker_parts"."spare": part is a spare part (0 or 1) +# - "bricktracker_parts"."quantity": quantity of this part +# - "bricktracker_parts"."missing": number of missing parts +# - "bricktracker_parts"."damaged": number of damaged parts # - "rebrickable_parts"."name": part name # - "rebrickable_parts"."color_name": part color name -# - "total_missing": number of missing parts +# - "total_missing": total missing across all sets (composite field) +# - "total_damaged": total damaged across all sets (composite field) +# - "total_quantity": total quantity across all sets (composite field) +# - "total_sets": number of sets containing this part (composite field) +# - "total_minifigures": number of minifigures with this part (composite field) # Default: "rebrickable_parts"."name" ASC, "rebrickable_parts"."color_name" ASC, "bricktracker_parts"."spare" ASC -# BK_PARTS_DEFAULT_ORDER="total_missing" DESC, "rebrickable_parts"."name"."name" ASC +# Examples: +# BK_PARTS_DEFAULT_ORDER="total_missing" DESC, "rebrickable_parts"."name" ASC +# BK_PARTS_DEFAULT_ORDER="rebrickable_parts"."color_name" ASC, "rebrickable_parts"."name" ASC -# Optional: Folder where to store the parts images, relative to the '/app/static/' folder -# Default: parts -# BK_PARTS_FOLDER=parts +# Optional: Folder where to store the parts images, relative to '/app/' folder +# Default: data/parts +# BK_PARTS_FOLDER=data/parts + +# Optional: Enable server-side pagination for individual pages (recommended for large collections) +# When enabled, pages use server-side pagination with configurable page sizes +# When disabled, pages load all data at once with instant client-side search +# Default: false for all +# BK_SETS_SERVER_SIDE_PAGINATION=true +# BK_PARTS_SERVER_SIDE_PAGINATION=true +# BK_MINIFIGURES_SERVER_SIDE_PAGINATION=true +# BK_PROBLEMS_SERVER_SIDE_PAGINATION=true + +# Optional: Number of parts to show per page on desktop devices (when server-side pagination is enabled) +# Default: 10 +# BK_PARTS_PAGINATION_SIZE_DESKTOP=10 + +# Optional: Number of parts to show per page on mobile devices (when server-side pagination is enabled) +# Default: 5 +# BK_PARTS_PAGINATION_SIZE_MOBILE=5 + +# Optional: Number of sets to show per page on desktop devices (when server-side pagination is enabled) +# Should be divisible by 4 for grid layout. Default: 12 +# BK_SETS_PAGINATION_SIZE_DESKTOP=12 + +# Optional: Number of sets to show per page on mobile devices (when server-side pagination is enabled) +# Default: 4 +# BK_SETS_PAGINATION_SIZE_MOBILE=4 + +# Optional: Number of minifigures to show per page on desktop devices (when server-side pagination is enabled) +# Default: 10 +# BK_MINIFIGURES_PAGINATION_SIZE_DESKTOP=10 + +# Optional: Number of minifigures to show per page on mobile devices (when server-side pagination is enabled) +# Default: 5 +# BK_MINIFIGURES_PAGINATION_SIZE_MOBILE=5 + +# Optional: Number of problems to show per page on desktop devices (when server-side pagination is enabled) +# Default: 10 +# BK_PROBLEMS_PAGINATION_SIZE_DESKTOP=10 + +# Optional: Number of problems to show per page on mobile devices (when server-side pagination is enabled) +# Default: 5 +# BK_PROBLEMS_PAGINATION_SIZE_MOBILE=5 # Optional: Port the server will listen on. # Default: 3333 @@ -185,9 +284,12 @@ # Optional: Change the default order of purchase locations. By default ordered by insertion order. # Useful column names for this option are: -# - "bricktracker_metadata_purchase_locations"."name" ASC: storage name +# - "bricktracker_metadata_purchase_locations"."name": purchase location name +# - "bricktracker_metadata_purchase_locations"."rowid": insertion order (special column) # Default: "bricktracker_metadata_purchase_locations"."name" ASC -# BK_PURCHASE_LOCATION_DEFAULT_ORDER="bricktracker_metadata_purchase_locations"."name" ASC +# Examples: +# BK_PURCHASE_LOCATION_DEFAULT_ORDER="bricktracker_metadata_purchase_locations"."name" DESC +# BK_PURCHASE_LOCATION_DEFAULT_ORDER="bricktracker_metadata_purchase_locations"."rowid" DESC # Optional: Shuffle the lists on the front page. # Default: false @@ -203,27 +305,54 @@ # Optional: URL of the image representing a missing image in Rebrickable # Default: https://rebrickable.com/static/img/nil.png -# BK_REBRICKABLE_IMAGE_NIL= +# BK_REBRICKABLE_IMAGE_NIL=https://rebrickable.com/static/img/nil.png # Optional: URL of the image representing a missing minifigure image in Rebrickable # Default: https://rebrickable.com/static/img/nil_mf.jpg -# BK_REBRICKABLE_IMAGE_NIL_MINIFIGURE= +# BK_REBRICKABLE_IMAGE_NIL_MINIFIGURE=https://rebrickable.com/static/img/nil_mf.jpg # Optional: Pattern of the link to Rebrickable for a minifigure. Will be passed to Python .format() # Default: https://rebrickable.com/minifigs/{figure} -# BK_REBRICKABLE_LINK_MINIFIGURE_PATTERN= +# BK_REBRICKABLE_LINK_MINIFIGURE_PATTERN=https://rebrickable.com/minifigs/{figure} # Optional: Pattern of the link to Rebrickable for a part. Will be passed to Python .format() # Default: https://rebrickable.com/parts/{part}/_/{color} -# BK_REBRICKABLE_LINK_PART_PATTERN= +# BK_REBRICKABLE_LINK_PART_PATTERN=https://rebrickable.com/parts/{part}/_/{color} # Optional: Pattern of the link to Rebrickable for instructions. Will be passed to Python .format() # Default: https://rebrickable.com/instructions/{path} -# BK_REBRICKABLE_LINK_INSTRUCTIONS_PATTERN= +# BK_REBRICKABLE_LINK_INSTRUCTIONS_PATTERN=https://rebrickable.com/instructions/{path} -# Optional: User-Agent to use when querying Rebrickable outside of the Rebrick python library -# Default: 'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/91.0.4472.124 Safari/537.36' -# BK_REBRICKABLE_USER_AGENT= +# Optional: User-Agent to use when querying Rebrickable and Peeron outside of the Rebrick python library +# Default: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/91.0.4472.124 Safari/537.36 +# BK_USER_AGENT=Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/91.0.4472.124 Safari/537.36 + +# Legacy: User-Agent for Rebrickable (use BK_USER_AGENT instead) +# BK_REBRICKABLE_USER_AGENT=Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/91.0.4472.124 Safari/537.36 + +# Optional: Delay in milliseconds between Peeron page downloads to avoid being potentially blocked +# Default: 1000 +# BK_PEERON_DOWNLOAD_DELAY=1000 + +# Optional: Minimum image size (width/height) for valid Peeron instruction pages +# Images smaller than this are considered error placeholders and will be rejected +# Default: 100 +# BK_PEERON_MIN_IMAGE_SIZE=100 + +# Optional: Pattern for Peeron instruction page URLs. Will be passed to Python .format() +# Supports {set_number} and {version_number} parameters +# Default: http://peeron.com/scans/{set_number}-{version_number} +# BK_PEERON_INSTRUCTION_PATTERN= + +# Optional: Pattern for Peeron thumbnail URLs. Will be passed to Python .format() +# Supports {set_number} and {version_number} parameters +# Default: http://belay.peeron.com/thumbs/{set_number}-{version_number}/ +# BK_PEERON_THUMBNAIL_PATTERN= + +# Optional: Pattern for Peeron scan URLs. Will be passed to Python .format() +# Supports {set_number} and {version_number} parameters +# Default: http://belay.peeron.com/scans/{set_number}-{version_number}/ +# BK_PEERON_SCAN_PATTERN= # Optional: Display Rebrickable links wherever applicable # Default: false @@ -238,27 +367,39 @@ # Default: https://docs.google.com/spreadsheets/d/1rlYfEXtNKxUOZt2Mfv0H17DvK7bj6Pe0CuYwq6ay8WA/gviz/tq?tqx=out:csv&sheet=Sorted%20by%20Retirement%20Date # BK_RETIRED_SETS_FILE_URL= -# Optional: Path to the unofficial retired sets lists +# Optional: Path to the unofficial retired sets lists, relative to '/app/' folder # You can name it whatever you want, but content has to be a CSV -# Default: ./retired_sets.csv -# BK_RETIRED_SETS_PATH=/var/lib/bricktracker/retired_sets.csv +# Default: data/retired_sets.csv +# BK_RETIRED_SETS_PATH=data/retired_sets.csv # Optional: Change the default order of sets. By default ordered by insertion order. # Useful column names for this option are: -# - "rebrickable_sets"."set": set number as a string -# - "rebrickable_sets"."number": the number part of set as an integer -# - "rebrickable_sets"."version": the version part of set as an integer +# - "rebrickable_sets"."set": set number as a string (e.g., "10255-1") +# - "rebrickable_sets"."number": the number part of set as text (e.g., "10255") +# - "rebrickable_sets"."version": the version part of set as an integer (e.g., 1) # - "rebrickable_sets"."name": set name # - "rebrickable_sets"."year": set release year # - "rebrickable_sets"."number_of_parts": set number of parts -# - "total_missing": number of missing parts -# - "total_minifigures": number of minifigures +# - "bricktracker_sets"."purchase_date": purchase date (as REAL/Julian day) +# - "bricktracker_sets"."purchase_price": purchase price +# - "total_missing": number of missing parts (composite field) +# - "total_damaged": number of damaged parts (composite field) +# - "total_minifigures": number of minifigures (composite field) # Default: "rebrickable_sets"."number" DESC, "rebrickable_sets"."version" ASC -# BK_SETS_DEFAULT_ORDER="rebrickable_sets"."year" ASC +# Examples: +# BK_SETS_DEFAULT_ORDER="rebrickable_sets"."year" DESC, "rebrickable_sets"."name" ASC +# BK_SETS_DEFAULT_ORDER="rebrickable_sets"."number_of_parts" DESC +# BK_SETS_DEFAULT_ORDER="total_missing" DESC, "rebrickable_sets"."year" ASC -# Optional: Folder where to store the sets images, relative to the '/app/static/' folder -# Default: sets -# BK_SETS_FOLDER=sets +# Optional: Folder where to store the sets images, relative to '/app/' folder +# Default: data/sets +# BK_SETS_FOLDER=data/sets + +# Optional: Enable set consolidation/grouping on the main sets page +# When enabled, multiple copies of the same set are grouped together showing instance count +# When disabled, each set copy is displayed individually (original behavior) +# Default: false +# BK_SETS_CONSOLIDATION=true # Optional: Make the grid filters displayed by default, rather than collapsed # Default: false @@ -268,10 +409,18 @@ # Default: false # BK_SHOW_GRID_SORT=true -# Optional: Skip saving or displaying spare parts +# Optional: Show duplicate filter button on sets page +# Default: true +# BK_SHOW_SETS_DUPLICATE_FILTER=true + +# Optional: Skip importing spare parts when downloading sets from Rebrickable # Default: false # BK_SKIP_SPARE_PARTS=true +# Optional: Hide spare parts from parts lists (spare parts must still be in database) +# Default: false +# BK_HIDE_SPARE_PARTS=true + # Optional: Namespace of the Socket.IO socket # Default: bricksocket # BK_SOCKET_NAMESPACE=customsocket @@ -282,18 +431,21 @@ # Optional: Change the default order of storages. By default ordered by insertion order. # Useful column names for this option are: -# - "bricktracker_metadata_storages"."name" ASC: storage name +# - "bricktracker_metadata_storages"."name": storage name +# - "bricktracker_metadata_storages"."rowid": insertion order (special column) # Default: "bricktracker_metadata_storages"."name" ASC -# BK_STORAGE_DEFAULT_ORDER="bricktracker_metadata_storages"."name" ASC +# Examples: +# BK_STORAGE_DEFAULT_ORDER="bricktracker_metadata_storages"."name" DESC +# BK_STORAGE_DEFAULT_ORDER="bricktracker_metadata_storages"."rowid" DESC # Optional: URL to the themes.csv.gz on Rebrickable # Default: https://cdn.rebrickable.com/media/downloads/themes.csv.gz # BK_THEMES_FILE_URL= -# Optional: Path to the themes file +# Optional: Path to the themes file, relative to '/app/' folder # You can name it whatever you want, but content has to be a CSV -# Default: ./themes.csv -# BK_THEMES_PATH=/var/lib/bricktracker/themes.csv +# Default: data/themes.csv +# BK_THEMES_PATH=data/themes.csv # Optional: Timezone to use to display datetimes # Check your system for available timezone/TZ values @@ -305,11 +457,30 @@ # Default: false # BK_USE_REMOTE_IMAGES=true -# Optional: Change the default order of sets. By default ordered by insertion order. +# Optional: Change the default order of wishlist sets. By default ordered by insertion order. # Useful column names for this option are: -# - "bricktracker_wishes"."set": set number as a string +# - "bricktracker_wishes"."set": set number as a string (e.g., "10255-1") # - "bricktracker_wishes"."name": set name # - "bricktracker_wishes"."year": set release year # - "bricktracker_wishes"."number_of_parts": set number of parts +# - "bricktracker_wishes"."theme_id": theme ID +# - "bricktracker_wishes"."rowid": insertion order (special column) # Default: "bricktracker_wishes"."rowid" DESC -# BK_WISHES_DEFAULT_ORDER="bricktracker_wishes"."set" DESC +# Examples: +# BK_WISHES_DEFAULT_ORDER="bricktracker_wishes"."year" DESC, "bricktracker_wishes"."name" ASC +# BK_WISHES_DEFAULT_ORDER="bricktracker_wishes"."number_of_parts" DESC +# BK_WISHES_DEFAULT_ORDER="bricktracker_wishes"."set" ASC + +# Optional: Show collection growth charts on the statistics page +# Default: true +# BK_STATISTICS_SHOW_CHARTS=false + +# Optional: Default state of statistics page sections (expanded or collapsed) +# When true, all sections start expanded. When false, all sections start collapsed. +# Default: true +# BK_STATISTICS_DEFAULT_EXPANDED=false + +# Optional: Enable dark mode by default +# When true, the application starts in dark mode. +# Default: false +# BK_DARK_MODE=true diff --git a/.gitignore b/.gitignore index 42357d7..fedff07 100644 --- a/.gitignore +++ b/.gitignore @@ -17,6 +17,7 @@ static/sets/ # IDE .vscode/ +*.code-workspace # Temporary *.csv @@ -33,3 +34,4 @@ vitepress/ # Local data offline/ +data/ diff --git a/CHANGELOG.md b/CHANGELOG.md index 3f318e5..ceb3264 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,8 +1,210 @@ # Changelog -## Unreleased +## 1.3 -### 1.2.4 +### Breaking Changes + +#### Data Folder Consolidation + +> **Warning** +> **BREAKING CHANGE**: Version 1.3 consolidates all user data into a single `data/` folder for easier backup and volume mapping. + +- **Path handling**: All relative paths are now resolved relative to the application root (`/app` in Docker) + - Example: `data/app.db` → `/app/data/app.db` + +- **New default paths** (automatically used for new installations): + - Database: `data/app.db` (was: `app.db` in root) + - Configuration: `data/.env` (was: `.env` in root) - *optional, backward compatible* + - CSV files: `data/*.csv` (was: `*.csv` in root) + - Images/PDFs: `data/{sets,parts,minifigures,instructions}/` (was: `static/*`) + +- **Configuration file (.env) location**: + - New recommended location: `data/.env` (included in data volume, settings persist) + - Backward compatible: `.env` in root still works (requires volume mount for admin panel persistence) + - Priority: `data/.env` > `.env` (automatic detection, no migration required) + +- **Migration options**: + 1. **Migrate to new structure** (recommended - single volume for all data including .env) + 2. **Keep current setup** (backward compatible - old paths continue to work) + +See [Migration Guide](docs/migration_guide.md) for detailed instructions + +#### Default Minifigures Folder Change + +> **Warning** +> **BREAKING CHANGE**: Default minifigures folder path changed from `minifigs` to `minifigures` + +- **Impact**: Users who relied on the default `BK_MINIFIGURES_FOLDER` value (without explicitly setting it) will need to either: + 1. Set `BK_MINIFIGURES_FOLDER=minifigs` in their environment to maintain existing behavior, or + 2. Rename their existing `minifigs` folder to `minifigures` +- **No impact**: Users who already have `BK_MINIFIGURES_FOLDER` explicitly configured +- Improved consistency across documentation and Docker configurations + +### New Features + +- **Live Settings changes** + - Added live environment variable configuration management system + - Configuration Management interface in admin panel with live preview and badge system + - **Live settings**: Can be changed without application restart (menu visibility, table display, pagination, features) + - **Static settings**: Require restart but can be edited and saved to .env file (authentication, server, database, API keys) + - Advanced badge system showing value status: True/False for booleans, Set/Default/Unset for other values, Changed indicator + - Live API endpoints: `/admin/api/config/update` for immediate changes, `/admin/api/config/update-static` for .env updates + - Form pre-population with current values and automatic page reload after successful live updates + - Fixed environment variable lock detection in admin configuration panel + - Resolved bug where all variables appeared "locked" after saving live settings + - Lock detection now correctly identifies only Docker environment variables set before .env loading + - Variables set via Docker's `environment:` directive remain properly locked + - Variables from data/.env or root .env are correctly shown as editable + - Added configuration persistence warning in admin panel + - Warning banner shows when using .env in root (non-persistent) + - Success banner shows when using data/.env (persistent) + - Provides migration instructions directly in the UI +- **Spare Parts** + - Added spare parts control options + - `BK_SKIP_SPARE_PARTS`: Skip importing spare parts when downloading sets from Rebrickable (parts not saved to database) + - `BK_HIDE_SPARE_PARTS`: Hide spare parts from all parts lists (parts must still be in database) + - Both options are live-changeable in admin configuration panel + - Options can be used independently or together for flexible spare parts management + - Affects all parts displays: /parts page, set details accordion, minifigure parts, and problem parts +- **Pagination** + - Added individual pagination control system per entity type + - `BK_SETS_SERVER_SIDE_PAGINATION`: Enable/disable pagination for sets + - `BK_PARTS_SERVER_SIDE_PAGINATION`: Enable/disable pagination for parts + - `BK_MINIFIGURES_SERVER_SIDE_PAGINATION`: Enable/disable pagination for minifigures + - Device-specific pagination sizes (desktop/mobile) for each entity type + - Supports search, filtering, and sorting in both server-side and client-side modes +- **Peeron Instructions** + - Added Peeron instructions integration + - Full image caching system with automatic thumbnail generation + - Optimized HTTP calls by downloading full images once and generating thumbnails locally + - Automatic cache cleanup after PDF generation to save disk space +- **Parts checkmark** + - Added parts checking/inventory system + - New "Checked" column in parts tables for tracking inventory progress + - Checkboxes to mark parts as verified during set walkthrough + - `BK_HIDE_TABLE_CHECKED_PARTS`: Environment variable to hide checked column +- **Set Consolidation** + - Added set consolidation/grouping functionality + - Automatic grouping of duplicate sets on main sets page + - Shows instance count with stack icon badge (e.g., "3 copies") + - Expandable drawer interface to view all set copies individually + - Full set cards for each instance with all badges, statuses, and functionality + - `BK_SETS_CONSOLIDATION`: Environment variable to enable/disable consolidation (default: false) + - Backwards compatible - when disabled, behaves exactly like original individual view + - Improved theme filtering: handles duplicate theme names correctly + - Fixed set number sorting: proper numeric sorting in both ascending and descending order + - Mixed status indicators for consolidated sets: three-state checkboxes (unchecked/partial/checked) with count badges + - Template logic handles three states: none (0/2), all (2/2), partial (1/2) with visual indicators + - Purple overlay styling for partial states, disabled checkboxes for read-only consolidated status display + - Individual sets maintain full interactive checkbox functionality +- **Statistics** + - Added comprehensive statistics system (#91) + - New Statistics page with collection analytics + - Financial overview: total cost, average price, price range, investment tracking + - Collection metrics: total sets, unique sets, parts count, minifigures count + - Theme distribution statistics with clickable drill-down to filtered sets + - Storage location statistics showing sets per location with value calculations + - Purchase location analytics with spending patterns and date ranges + - Problem tracking: missing and damaged parts statistics + - Clickable numbers throughout statistics that filter to relevant sets + - `BK_HIDE_STATISTICS`: Environment variable to hide statistics menu item + - Year-based analytics: Sets by release year and purchases by year + - Sets by Release Year: Shows collection distribution across LEGO release years + - Purchases by Year: Tracks spending patterns and acquisition timeline + - Year summary with peak collection/spending years and timeline insights + - Enhanced statistics interface and functionality + - Collapsible sections: All statistics sections have clickable headers to expand/collapse + - Collection growth charts: Line charts showing sets, parts, and minifigures over time + - Configuration options: `BK_STATISTICS_SHOW_CHARTS` and `BK_STATISTICS_DEFAULT_EXPANDED` environment variables +- **Admin Page Section Expansion** + - Added configurable admin page section expansion + - `BK_ADMIN_DEFAULT_EXPANDED_SECTIONS`: Environment variable to specify which sections expand by default + - Accepts comma-separated list of section names (e.g., "database,theme,instructions") + - Valid sections: authentication, instructions, image, theme, retired, metadata, owner, purchase_location, status, storage, tag, database + - URL parameters take priority over configuration (e.g., `?open_database=1`) + - Database section expanded by default to maintain original behavior + - Smart metadata handling: sub-section expansion automatically expands parent metadata section +- **Duplicate Sets filter** + - Added duplicate sets filter functionality + - New filter button on Sets page to show only duplicate/consolidated sets + - `BK_SHOW_SETS_DUPLICATE_FILTER`: Environment variable to show/hide the filter button (default: true) + - Works with both server-side and client-side pagination modes + - Consolidated mode: Shows sets that have multiple instances + - Non-consolidated mode: Shows sets that appear multiple times in collection +- **Bricklink Links** + - Added BrickLink links for sets + - BrickLink badge links now appear on set cards and set details pages alongside Rebrickable links + - `BK_BRICKLINK_LINK_SET_PATTERN`: New environment variable for BrickLink set URL pattern (default: https://www.bricklink.com/v2/catalog/catalogitem.page?S={set_num}) + - Controlled by existing `BK_BRICKLINK_LINKS` environment variable +- **Dark Mode** + - Added dark mode support + - `BK_DARK_MODE`: Environment variable to enable dark mode theme (default: false) + - Uses Bootstrap 5.3's native dark mode with `data-bs-theme` attribute + - Live-changeable via Admin > Live Settings + - Setting persists across sessions via .env file +- **Alphanumetic Set Number** + - Added alphanumeric set number support + - Database schema change: Set number column changed from INTEGER to TEXT + - Supports LEGO promotional and special edition sets with letters in their numbers + - Examples: "McDR6US-1", "COMCON035-1", "EG00021-1" + +### Improvements + +- Improved WebSocket/Socket.IO reliability for mobile devices + - Changed connection strategy to polling-first with automatic WebSocket upgrade + - Increased connection timeout to 30 seconds for slow mobile networks + - Added ping/pong keepalive settings (30s timeout, 25s interval) + - Improved server-side connection logging with user agent and transport details +- Fixed dynamic sort icons across all pages + - Sort icons now properly toggle between ascending/descending states +- Improved DataTable integration + - Disabled column header sorting when server-side pagination is enabled + - Prevents conflicting sort mechanisms between DataTable and server-side sorting +- Enhanced color dropdown functionality + - Automatic merging of duplicate color entries with same color_id + - Keeps entries with valid RGB data, removes entries with None/empty RGB + - Preserves selection state during dropdown consolidation + - Consistent search behavior (instant for client-side, Enter key for server-side) + - Mobile-friendly pagination navigation +- Added performance optimization + - SQLite WAL Mode: + - Increased cache size to 10,000 pages (~40MB) for faster query execution + - Set temp_store to memory for accelerated temporary operations + - Enabled foreign key constraints and optimized synchronous mode + - Added ANALYZE for improved query planning and statistics + - Database Indexes (Migration 0019): + - High-impact composite index for problem parts aggregation (`idx_bricktracker_parts_id_missing_damaged`) + - Parts lookup optimization (`idx_bricktracker_parts_part_color_spare`) + - Set storage filtering (`idx_bricktracker_sets_set_storage`) + - Search optimization with case-insensitive indexes (`idx_rebrickable_sets_name_lower`, `idx_rebrickable_parts_name_lower`) + - Year and theme filtering optimization (`idx_rebrickable_sets_year`, `idx_rebrickable_sets_theme_id`) + - Additional indexes for purchase dates, quantities, sorting, and minifigures aggregation + - Statistics Query Optimization: + - Replaced separate subqueries with efficient CTEs (Common Table Expressions) + - Consolidated aggregations for set, part, minifigure, and financial statistics + - Added default image handling for sets without images + - Sets with null/missing images from Rebrickable API now display placeholder image + - Automatic fallback to nil.png from parts folder for set previews + - Copy of nil placeholder saved as set image for consistent display across all routes + - Prevents errors when downloading sets that have no set_img_url in API response + - Fixed instructions download from Rebrickable + - Replaced cloudscraper with standard requests library + - Resolves 403 Forbidden errors when downloading instruction PDFs + - Fixed instructions display and URL generation + - Fixed "Open PDF" button links to use correct data route + - Corrected path resolution for data/instructions folder + - Fixed instruction listing page to scan correct folder location + - Fixed Peeron PDF creation to use correct data folder path + - Fixed foreign key constraint error when adding sets + - Rebrickable set is now inserted before BrickTracker set to satisfy FK constraints + - Resolves "FOREIGN KEY constraint failed" error when adding sets + - Fixed atomic transaction handling for set downloads + - All database operations during set addition now use deferred execution + - Ensures all-or-nothing behavior: if any part fails (set info, parts, minifigs), nothing is committed + - Prevents partial set additions that would leave the database in an inconsistent state + - Metadata updates (owners, tags) now defer until final commit + +## 1.2.4 > **Warning** > To use the new BrickLink color parameter in URLs, update your `.env` file: diff --git a/Dockerfile b/Dockerfile index 52cf033..871c23d 100644 --- a/Dockerfile +++ b/Dockerfile @@ -2,13 +2,16 @@ FROM python:3-slim WORKDIR /app +# Copy requirements first (so pip install can be cached) +COPY requirements.txt . + +# Python library requirements +RUN pip install --no-cache-dir -r requirements.txt + # Bricktracker COPY . . -# Fix line endings and set executable permissions for entrypoint script -RUN sed -i 's/\r$//' entrypoint.sh && chmod +x entrypoint.sh - -# Python library requirements -RUN pip --no-cache-dir install -r requirements.txt +# Set executable permissions for entrypoint script +RUN chmod +x entrypoint.sh ENTRYPOINT ["./entrypoint.sh"] diff --git a/README.md b/README.md index 58e4529..094ec23 100644 --- a/README.md +++ b/README.md @@ -1,3 +1,5 @@ + + # BrickTracker A web application for organizing and tracking LEGO sets, parts, and minifigures. Uses the Rebrickable API to fetch LEGO data and allows users to track missing pieces and collection status. @@ -18,17 +20,13 @@ A web application for organizing and tracking LEGO sets, parts, and minifigures. Use the provided [compose.yaml](compose.yaml) file. -See [Quickstart](docs/quickstart.md) to get up and running right away. +See [Quick Start](https://bricktracker.baerentsen.space/quick-start) to get up and running right away. -See [Setup](docs/setup.md) for a more setup guide. - -## Usage - -See [first steps](docs/first-steps.md). +See [Walk Through](https://bricktracker.baerentsen.space/tutorial-first-steps) for a more detailed guide. ## Documentation Most of the pages should be self explanatory to use. -However, you can find more specific documentation in the [documentation](docs/DOCS.md). +However, you can find more specific documentation in the [documentation](https://bricktracker.baerentsen.space/whatis). -You can find screenshots of the application in the [overview](docs/overview.md) documentation file. +You can find screenshots of the application in the [overview](https://bricktracker.baerentsen.space/overview) documentation. diff --git a/bricktracker/app.py b/bricktracker/app.py index b4aad9e..3d56b18 100644 --- a/bricktracker/app.py +++ b/bricktracker/app.py @@ -1,6 +1,8 @@ import logging +import os import sys import time +from pathlib import Path from zoneinfo import ZoneInfo from flask import current_app, Flask, g @@ -10,6 +12,7 @@ from bricktracker.configuration_list import BrickConfigurationList from bricktracker.login import LoginManager from bricktracker.navbar import Navbar from bricktracker.sql import close +from bricktracker.template_filters import replace_query_filter from bricktracker.version import __version__ from bricktracker.views.add import add_page from bricktracker.views.admin.admin import admin_page @@ -24,6 +27,7 @@ from bricktracker.views.admin.status import admin_status_page from bricktracker.views.admin.storage import admin_storage_page from bricktracker.views.admin.tag import admin_tag_page from bricktracker.views.admin.theme import admin_theme_page +from bricktracker.views.data import data_page from bricktracker.views.error import error_404 from bricktracker.views.index import index_page from bricktracker.views.instructions import instructions_page @@ -31,11 +35,65 @@ from bricktracker.views.login import login_page from bricktracker.views.minifigure import minifigure_page from bricktracker.views.part import part_page from bricktracker.views.set import set_page +from bricktracker.views.statistics import statistics_page from bricktracker.views.storage import storage_page from bricktracker.views.wish import wish_page +def load_env_file() -> None: + """Load .env file into os.environ with priority: data/.env > .env (root) + + Also stores which BK_ variables were set via Docker environment (before loading .env) + so we can detect locked variables in the admin panel. + """ + import json + + data_env = Path('data/.env') + root_env = Path('.env') + + # Store which BK_ variables were already in environment BEFORE loading .env + # These are "locked" (set via Docker's environment: directive) + docker_env_vars = {k: v for k, v in os.environ.items() if k.startswith('BK_')} + + # Store this in a way the admin panel can access it + # We'll use an environment variable to store the JSON list of locked var names + os.environ['_BK_DOCKER_ENV_VARS'] = json.dumps(list(docker_env_vars.keys())) + + env_file = None + if data_env.exists(): + env_file = data_env + logging.info(f"Loading environment from: {data_env}") + elif root_env.exists(): + env_file = root_env + logging.info(f"Loading environment from: {root_env} (consider migrating to data/.env)") + + if env_file: + # Simple .env parser (no external dependencies needed) + with open(env_file, 'r', encoding='utf-8') as f: + for line in f: + line = line.strip() + # Skip comments and empty lines + if not line or line.startswith('#'): + continue + # Parse key=value + if '=' in line: + key, value = line.split('=', 1) + key = key.strip() + value = value.strip() + # Remove quotes if present + if value.startswith('"') and value.endswith('"'): + value = value[1:-1] + elif value.startswith("'") and value.endswith("'"): + value = value[1:-1] + # Only set if not already in environment (environment variables take precedence) + if key not in os.environ: + os.environ[key] = value + + def setup_app(app: Flask) -> None: + # Load .env file before configuration (if not already loaded by Docker Compose) + load_env_file() + # Load the configuration BrickConfigurationList(app) @@ -59,7 +117,8 @@ def setup_app(app: Flask) -> None: # Setup the login manager LoginManager(app) - # I don't know :-) + # Configure proxy header handling for reverse proxy deployments (nginx, Apache, etc.) + # This ensures proper client IP detection and HTTPS scheme recognition app.wsgi_app = ProxyFix( app.wsgi_app, x_for=1, @@ -74,12 +133,14 @@ def setup_app(app: Flask) -> None: # Register app routes app.register_blueprint(add_page) + app.register_blueprint(data_page) app.register_blueprint(index_page) app.register_blueprint(instructions_page) app.register_blueprint(login_page) app.register_blueprint(minifigure_page) app.register_blueprint(part_page) app.register_blueprint(set_page) + app.register_blueprint(statistics_page) app.register_blueprint(storage_page) app.register_blueprint(wish_page) @@ -121,6 +182,9 @@ def setup_app(app: Flask) -> None: # Version g.version = __version__ + # Register custom Jinja2 filters + app.jinja_env.filters['replace_query'] = replace_query_filter + # Make sure all connections are closed at the end @app.teardown_request def teardown_request(_: BaseException | None) -> None: diff --git a/bricktracker/config.py b/bricktracker/config.py index 4f4d614..94e681c 100644 --- a/bricktracker/config.py +++ b/bricktracker/config.py @@ -11,8 +11,9 @@ CONFIG: Final[list[dict[str, Any]]] = [ {'n': 'AUTHENTICATION_PASSWORD', 'd': ''}, {'n': 'AUTHENTICATION_KEY', 'd': ''}, {'n': 'BRICKLINK_LINK_PART_PATTERN', 'd': 'https://www.bricklink.com/v2/catalog/catalogitem.page?P={part}&C={color}'}, # noqa: E501 + {'n': 'BRICKLINK_LINK_SET_PATTERN', 'd': 'https://www.bricklink.com/v2/catalog/catalogitem.page?S={set_num}'}, # noqa: E501 {'n': 'BRICKLINK_LINKS', 'c': bool}, - {'n': 'DATABASE_PATH', 'd': './app.db'}, + {'n': 'DATABASE_PATH', 'd': 'data/app.db'}, {'n': 'DATABASE_TIMESTAMP_FORMAT', 'd': '%Y-%m-%d-%H-%M-%S'}, {'n': 'DEBUG', 'c': bool}, {'n': 'DEFAULT_TABLE_PER_PAGE', 'd': 25, 'c': int}, @@ -21,25 +22,40 @@ CONFIG: Final[list[dict[str, Any]]] = [ {'n': 'HOST', 'd': '0.0.0.0'}, {'n': 'INDEPENDENT_ACCORDIONS', 'c': bool}, {'n': 'INSTRUCTIONS_ALLOWED_EXTENSIONS', 'd': ['.pdf'], 'c': list}, # noqa: E501 - {'n': 'INSTRUCTIONS_FOLDER', 'd': 'instructions', 's': True}, + {'n': 'INSTRUCTIONS_FOLDER', 'd': 'data/instructions'}, {'n': 'HIDE_ADD_SET', 'c': bool}, {'n': 'HIDE_ADD_BULK_SET', 'c': bool}, {'n': 'HIDE_ADMIN', 'c': bool}, + {'n': 'ADMIN_DEFAULT_EXPANDED_SECTIONS', 'd': ['database'], 'c': list}, {'n': 'HIDE_ALL_INSTRUCTIONS', 'c': bool}, {'n': 'HIDE_ALL_MINIFIGURES', 'c': bool}, {'n': 'HIDE_ALL_PARTS', 'c': bool}, {'n': 'HIDE_ALL_PROBLEMS_PARTS', 'e': 'BK_HIDE_MISSING_PARTS', 'c': bool}, {'n': 'HIDE_ALL_SETS', 'c': bool}, {'n': 'HIDE_ALL_STORAGES', 'c': bool}, + {'n': 'HIDE_STATISTICS', 'c': bool}, {'n': 'HIDE_SET_INSTRUCTIONS', 'c': bool}, {'n': 'HIDE_TABLE_DAMAGED_PARTS', 'c': bool}, {'n': 'HIDE_TABLE_MISSING_PARTS', 'c': bool}, + {'n': 'HIDE_TABLE_CHECKED_PARTS', 'c': bool}, {'n': 'HIDE_WISHES', 'c': bool}, {'n': 'MINIFIGURES_DEFAULT_ORDER', 'd': '"rebrickable_minifigures"."name" ASC'}, # noqa: E501 - {'n': 'MINIFIGURES_FOLDER', 'd': 'minifigs', 's': True}, + {'n': 'MINIFIGURES_FOLDER', 'd': 'data/minifigures'}, + {'n': 'MINIFIGURES_PAGINATION_SIZE_DESKTOP', 'd': 10, 'c': int}, + {'n': 'MINIFIGURES_PAGINATION_SIZE_MOBILE', 'd': 5, 'c': int}, + {'n': 'MINIFIGURES_SERVER_SIDE_PAGINATION', 'c': bool}, {'n': 'NO_THREADED_SOCKET', 'c': bool}, + {'n': 'PARTS_SERVER_SIDE_PAGINATION', 'c': bool}, + {'n': 'SETS_SERVER_SIDE_PAGINATION', 'c': bool}, {'n': 'PARTS_DEFAULT_ORDER', 'd': '"rebrickable_parts"."name" ASC, "rebrickable_parts"."color_name" ASC, "bricktracker_parts"."spare" ASC'}, # noqa: E501 - {'n': 'PARTS_FOLDER', 'd': 'parts', 's': True}, + {'n': 'PARTS_FOLDER', 'd': 'data/parts'}, + {'n': 'PARTS_PAGINATION_SIZE_DESKTOP', 'd': 10, 'c': int}, + {'n': 'PARTS_PAGINATION_SIZE_MOBILE', 'd': 5, 'c': int}, + {'n': 'PROBLEMS_PAGINATION_SIZE_DESKTOP', 'd': 10, 'c': int}, + {'n': 'PROBLEMS_PAGINATION_SIZE_MOBILE', 'd': 10, 'c': int}, + {'n': 'PROBLEMS_SERVER_SIDE_PAGINATION', 'c': bool}, + {'n': 'SETS_PAGINATION_SIZE_DESKTOP', 'd': 12, 'c': int}, + {'n': 'SETS_PAGINATION_SIZE_MOBILE', 'd': 4, 'c': int}, {'n': 'PORT', 'd': 3333, 'c': int}, {'n': 'PURCHASE_DATE_FORMAT', 'd': '%d/%m/%Y'}, {'n': 'PURCHASE_CURRENCY', 'd': '€'}, @@ -52,21 +68,33 @@ CONFIG: Final[list[dict[str, Any]]] = [ {'n': 'REBRICKABLE_LINK_PART_PATTERN', 'd': 'https://rebrickable.com/parts/{part}/_/{color}'}, # noqa: E501 {'n': 'REBRICKABLE_LINK_INSTRUCTIONS_PATTERN', 'd': 'https://rebrickable.com/instructions/{path}'}, # noqa: E501 {'n': 'REBRICKABLE_USER_AGENT', 'd': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/91.0.4472.124 Safari/537.36'}, # noqa: E501 + {'n': 'USER_AGENT', 'd': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/91.0.4472.124 Safari/537.36'}, # noqa: E501 + {'n': 'PEERON_DOWNLOAD_DELAY', 'd': 1000, 'c': int}, + {'n': 'PEERON_INSTRUCTION_PATTERN', 'd': 'http://peeron.com/scans/{set_number}-{version_number}'}, + {'n': 'PEERON_MIN_IMAGE_SIZE', 'd': 100, 'c': int}, + {'n': 'PEERON_SCAN_PATTERN', 'd': 'http://belay.peeron.com/scans/{set_number}-{version_number}/'}, + {'n': 'PEERON_THUMBNAIL_PATTERN', 'd': 'http://belay.peeron.com/thumbs/{set_number}-{version_number}/'}, {'n': 'REBRICKABLE_LINKS', 'e': 'LINKS', 'c': bool}, {'n': 'REBRICKABLE_PAGE_SIZE', 'd': 100, 'c': int}, {'n': 'RETIRED_SETS_FILE_URL', 'd': 'https://docs.google.com/spreadsheets/d/1rlYfEXtNKxUOZt2Mfv0H17DvK7bj6Pe0CuYwq6ay8WA/gviz/tq?tqx=out:csv&sheet=Sorted%20by%20Retirement%20Date'}, # noqa: E501 - {'n': 'RETIRED_SETS_PATH', 'd': './retired_sets.csv'}, + {'n': 'RETIRED_SETS_PATH', 'd': 'data/retired_sets.csv'}, {'n': 'SETS_DEFAULT_ORDER', 'd': '"rebrickable_sets"."number" DESC, "rebrickable_sets"."version" ASC'}, # noqa: E501 - {'n': 'SETS_FOLDER', 'd': 'sets', 's': True}, + {'n': 'SETS_FOLDER', 'd': 'data/sets'}, + {'n': 'SETS_CONSOLIDATION', 'd': False, 'c': bool}, {'n': 'SHOW_GRID_FILTERS', 'c': bool}, {'n': 'SHOW_GRID_SORT', 'c': bool}, + {'n': 'SHOW_SETS_DUPLICATE_FILTER', 'd': True, 'c': bool}, {'n': 'SKIP_SPARE_PARTS', 'c': bool}, + {'n': 'HIDE_SPARE_PARTS', 'c': bool}, {'n': 'SOCKET_NAMESPACE', 'd': 'bricksocket'}, {'n': 'SOCKET_PATH', 'd': '/bricksocket/'}, {'n': 'STORAGE_DEFAULT_ORDER', 'd': '"bricktracker_metadata_storages"."name" ASC'}, # noqa: E501 {'n': 'THEMES_FILE_URL', 'd': 'https://cdn.rebrickable.com/media/downloads/themes.csv.gz'}, # noqa: E501 - {'n': 'THEMES_PATH', 'd': './themes.csv'}, + {'n': 'THEMES_PATH', 'd': 'data/themes.csv'}, {'n': 'TIMEZONE', 'd': 'Etc/UTC'}, {'n': 'USE_REMOTE_IMAGES', 'c': bool}, {'n': 'WISHES_DEFAULT_ORDER', 'd': '"bricktracker_wishes"."rowid" DESC'}, + {'n': 'STATISTICS_SHOW_CHARTS', 'd': True, 'c': bool}, + {'n': 'STATISTICS_DEFAULT_EXPANDED', 'd': True, 'c': bool}, + {'n': 'DARK_MODE', 'c': bool}, ] diff --git a/bricktracker/config_manager.py b/bricktracker/config_manager.py new file mode 100644 index 0000000..d6985ec --- /dev/null +++ b/bricktracker/config_manager.py @@ -0,0 +1,331 @@ +import os +import logging +from typing import Any, Dict, Final, List, Optional +from pathlib import Path +from flask import current_app + +logger = logging.getLogger(__name__) + +# Environment variables that can be changed live without restart +LIVE_CHANGEABLE_VARS: Final[List[str]] = [ + 'BK_BRICKLINK_LINKS', + 'BK_DEFAULT_TABLE_PER_PAGE', + 'BK_INDEPENDENT_ACCORDIONS', + 'BK_HIDE_ADD_SET', + 'BK_HIDE_ADD_BULK_SET', + 'BK_HIDE_ADMIN', + 'BK_ADMIN_DEFAULT_EXPANDED_SECTIONS', + 'BK_HIDE_ALL_INSTRUCTIONS', + 'BK_HIDE_ALL_MINIFIGURES', + 'BK_HIDE_ALL_PARTS', + 'BK_HIDE_ALL_PROBLEMS_PARTS', + 'BK_HIDE_ALL_SETS', + 'BK_HIDE_ALL_STORAGES', + 'BK_HIDE_STATISTICS', + 'BK_HIDE_SET_INSTRUCTIONS', + 'BK_HIDE_TABLE_DAMAGED_PARTS', + 'BK_HIDE_TABLE_MISSING_PARTS', + 'BK_HIDE_TABLE_CHECKED_PARTS', + 'BK_HIDE_WISHES', + 'BK_MINIFIGURES_PAGINATION_SIZE_DESKTOP', + 'BK_MINIFIGURES_PAGINATION_SIZE_MOBILE', + 'BK_MINIFIGURES_SERVER_SIDE_PAGINATION', + 'BK_PARTS_PAGINATION_SIZE_DESKTOP', + 'BK_PARTS_PAGINATION_SIZE_MOBILE', + 'BK_PARTS_SERVER_SIDE_PAGINATION', + 'BK_SETS_SERVER_SIDE_PAGINATION', + 'BK_PROBLEMS_PAGINATION_SIZE_DESKTOP', + 'BK_PROBLEMS_PAGINATION_SIZE_MOBILE', + 'BK_PROBLEMS_SERVER_SIDE_PAGINATION', + 'BK_SETS_PAGINATION_SIZE_DESKTOP', + 'BK_SETS_PAGINATION_SIZE_MOBILE', + 'BK_SETS_CONSOLIDATION', + 'BK_RANDOM', + 'BK_REBRICKABLE_LINKS', + 'BK_SHOW_GRID_FILTERS', + 'BK_SHOW_GRID_SORT', + 'BK_SHOW_SETS_DUPLICATE_FILTER', + 'BK_SKIP_SPARE_PARTS', + 'BK_HIDE_SPARE_PARTS', + 'BK_USE_REMOTE_IMAGES', + 'BK_PEERON_DOWNLOAD_DELAY', + 'BK_PEERON_MIN_IMAGE_SIZE', + 'BK_REBRICKABLE_PAGE_SIZE', + 'BK_STATISTICS_SHOW_CHARTS', + 'BK_STATISTICS_DEFAULT_EXPANDED', + 'BK_DARK_MODE', + # Default ordering and formatting + 'BK_INSTRUCTIONS_ALLOWED_EXTENSIONS', + 'BK_MINIFIGURES_DEFAULT_ORDER', + 'BK_PARTS_DEFAULT_ORDER', + 'BK_SETS_DEFAULT_ORDER', + 'BK_PURCHASE_LOCATION_DEFAULT_ORDER', + 'BK_STORAGE_DEFAULT_ORDER', + 'BK_WISHES_DEFAULT_ORDER', + # URL and Pattern Variables + 'BK_BRICKLINK_LINK_PART_PATTERN', + 'BK_BRICKLINK_LINK_SET_PATTERN', + 'BK_REBRICKABLE_IMAGE_NIL', + 'BK_REBRICKABLE_IMAGE_NIL_MINIFIGURE', + 'BK_REBRICKABLE_LINK_MINIFIGURE_PATTERN', + 'BK_REBRICKABLE_LINK_PART_PATTERN', + 'BK_REBRICKABLE_LINK_INSTRUCTIONS_PATTERN', + 'BK_PEERON_INSTRUCTION_PATTERN', + 'BK_PEERON_SCAN_PATTERN', + 'BK_PEERON_THUMBNAIL_PATTERN', + 'BK_RETIRED_SETS_FILE_URL', + 'BK_RETIRED_SETS_PATH', + 'BK_THEMES_FILE_URL', + 'BK_THEMES_PATH' +] + +# Environment variables that require restart +RESTART_REQUIRED_VARS: Final[List[str]] = [ + 'BK_AUTHENTICATION_PASSWORD', + 'BK_AUTHENTICATION_KEY', + 'BK_DATABASE_PATH', + 'BK_DEBUG', + 'BK_DOMAIN_NAME', + 'BK_HOST', + 'BK_PORT', + 'BK_SOCKET_NAMESPACE', + 'BK_SOCKET_PATH', + 'BK_NO_THREADED_SOCKET', + 'BK_TIMEZONE', + 'BK_REBRICKABLE_API_KEY', + 'BK_INSTRUCTIONS_FOLDER', + 'BK_PARTS_FOLDER', + 'BK_SETS_FOLDER', + 'BK_MINIFIGURES_FOLDER', + 'BK_DATABASE_TIMESTAMP_FORMAT', + 'BK_FILE_DATETIME_FORMAT', + 'BK_PURCHASE_DATE_FORMAT', + 'BK_PURCHASE_CURRENCY', + 'BK_REBRICKABLE_USER_AGENT', + 'BK_USER_AGENT' +] + +class ConfigManager: + """Manages live configuration updates for BrickTracker""" + + def __init__(self): + # Check for .env in data folder first (v1.3+), fallback to root (backward compatibility) + data_env = Path('data/.env') + root_env = Path('.env') + + if data_env.exists(): + self.env_file_path = data_env + logger.info("Using configuration file: data/.env") + elif root_env.exists(): + self.env_file_path = root_env + logger.info("Using configuration file: .env (consider migrating to data/.env)") + else: + # Default to data/.env for new installations + self.env_file_path = data_env + logger.info("Configuration file will be created at: data/.env") + + def get_current_config(self) -> Dict[str, Any]: + """Get current configuration values for live-changeable variables""" + config = {} + for var in LIVE_CHANGEABLE_VARS: + # Get internal config name + internal_name = var.replace('BK_', '') + # Get current value from Flask config + if internal_name in current_app.config: + config[var] = current_app.config[internal_name] + else: + # Fallback to environment variable + config[var] = os.environ.get(var, '') + return config + + def get_restart_required_config(self) -> Dict[str, Any]: + """Get current configuration values for restart-required variables""" + config = {} + for var in RESTART_REQUIRED_VARS: + # Get internal config name + internal_name = var.replace('BK_', '') + # Get current value from Flask config + if internal_name in current_app.config: + config[var] = current_app.config[internal_name] + else: + # Fallback to environment variable + config[var] = os.environ.get(var, '') + return config + + def update_config(self, updates: Dict[str, Any]) -> Dict[str, str]: + """Update configuration values. Returns dict with status for each update""" + results = {} + for var_name, new_value in updates.items(): + if var_name not in LIVE_CHANGEABLE_VARS: + results[var_name] = f"Error: {var_name} requires restart to change" + continue + + try: + # Update environment variable + os.environ[var_name] = str(new_value) + # Update Flask config + internal_name = var_name.replace('BK_', '') + cast_value = self._cast_value(var_name, new_value) + current_app.config[internal_name] = cast_value + # Update .env file + self._update_env_file(var_name, new_value) + results[var_name] = "Updated successfully" + if current_app.debug: + logger.info(f"Config updated: {var_name}={new_value}") + except Exception as e: + results[var_name] = f"Error: {str(e)}" + logger.error(f"Failed to update {var_name}: {e}") + return results + + def _cast_value(self, var_name: str, value: Any) -> Any: + """Cast value to appropriate type based on variable name""" + # List variables (admin sections) - Check this FIRST before boolean check + if 'sections' in var_name.lower(): + if isinstance(value, str): + return [section.strip() for section in value.split(',') if section.strip()] + elif isinstance(value, list): + return value + else: + return [] + # Integer variables (pagination sizes, delays, etc.) - Check BEFORE boolean check + if any(keyword in var_name.lower() for keyword in ['_size', '_page', 'delay', 'min_', 'per_page', 'page_size']): + try: + return int(value) + except (ValueError, TypeError): + return 0 + # Boolean variables - More specific patterns to avoid conflicts + if any(keyword in var_name.lower() for keyword in ['hide_', 'server_side_pagination', '_links', 'random', 'skip_', 'show_', 'use_', '_consolidation', '_charts', '_expanded']): + if isinstance(value, str): + return value.lower() in ('true', '1', 'yes', 'on') + return bool(value) + # String variables (default) + return str(value) + + def _format_env_value(self, value: Any) -> str: + """Format value for .env file storage""" + if isinstance(value, bool): + return 'true' if value else 'false' + elif isinstance(value, (int, float)): + return str(value) + elif isinstance(value, list): + return ','.join(str(item) for item in value) + elif value is None: + return '' + else: + return str(value) + + def _update_env_file(self, var_name: str, value: Any) -> None: + """Update the .env file with new value""" + if not self.env_file_path.exists(): + # Ensure parent directory exists + self.env_file_path.parent.mkdir(parents=True, exist_ok=True) + self.env_file_path.touch() + + # Read current .env content + lines = [] + if self.env_file_path.exists(): + with open(self.env_file_path, 'r', encoding='utf-8') as f: + lines = f.readlines() + + # Format value for .env file + env_value = self._format_env_value(value) + + # Find and update the line, or add new line + updated = False + + # First pass: Look for existing active variable + for i, line in enumerate(lines): + if line.strip().startswith(f"{var_name}="): + lines[i] = f"{var_name}={env_value}\n" + updated = True + break + + # Second pass: If not found, look for commented-out variable + if not updated: + for i, line in enumerate(lines): + stripped = line.strip() + # Check for commented-out variable: # BK_VAR= or #BK_VAR= + if stripped.startswith('#') and var_name in stripped: + # Extract the part after #, handling optional space + comment_content = stripped[1:].strip() + if comment_content.startswith(f"{var_name}=") or comment_content.startswith(f"{var_name} ="): + # Uncomment and set new value, preserving any leading whitespace from original line + leading_whitespace = line[:len(line) - len(line.lstrip())] + lines[i] = f"{leading_whitespace}{var_name}={env_value}\n" + updated = True + logger.info(f"Uncommented and updated {var_name} in .env file") + break + + # Third pass: If still not found, append to end + if not updated: + lines.append(f"{var_name}={env_value}\n") + logger.info(f"Added new {var_name} to end of .env file") + + # Write back to file + with open(self.env_file_path, 'w', encoding='utf-8') as f: + f.writelines(lines) + + def validate_config(self) -> Dict[str, Any]: + """Validate current configuration""" + issues = [] + warnings = [] + + # Check if critical variables are set + if not os.environ.get('BK_REBRICKABLE_API_KEY'): + warnings.append("BK_REBRICKABLE_API_KEY not set - some features may not work") + + # Check for conflicting settings + if (os.environ.get('BK_PARTS_SERVER_SIDE_PAGINATION', '').lower() == 'false' and + int(os.environ.get('BK_PARTS_PAGINATION_SIZE_DESKTOP', '10')) > 100): + warnings.append("Large pagination size with client-side pagination may cause performance issues") + + # Check pagination sizes are reasonable + for var in ['BK_SETS_PAGINATION_SIZE_DESKTOP', 'BK_PARTS_PAGINATION_SIZE_DESKTOP', 'BK_MINIFIGURES_PAGINATION_SIZE_DESKTOP']: + try: + size = int(os.environ.get(var, '10')) + if size < 1: + issues.append(f"{var} must be at least 1") + elif size > 1000: + warnings.append(f"{var} is very large ({size}) - may cause performance issues") + except ValueError: + issues.append(f"{var} must be a valid integer") + + return { + 'issues': issues, + 'warnings': warnings, + 'status': 'valid' if not issues else 'has_issues' + } + + def get_variable_help(self, var_name: str) -> str: + """Get help text for a configuration variable""" + help_text = { + 'BK_BRICKLINK_LINKS': 'Show BrickLink links throughout the application', + 'BK_DEFAULT_TABLE_PER_PAGE': 'Default number of items per page in tables', + 'BK_INDEPENDENT_ACCORDIONS': 'Make accordion sections independent (can open multiple)', + 'BK_HIDE_ADD_SET': 'Hide the "Add Set" menu entry', + 'BK_HIDE_ADD_BULK_SET': 'Hide the "Add Bulk Set" menu entry', + 'BK_HIDE_ADMIN': 'Hide the "Admin" menu entry', + 'BK_ADMIN_DEFAULT_EXPANDED_SECTIONS': 'Admin sections to expand by default (comma-separated)', + 'BK_HIDE_ALL_INSTRUCTIONS': 'Hide the "Instructions" menu entry', + 'BK_HIDE_ALL_MINIFIGURES': 'Hide the "Minifigures" menu entry', + 'BK_HIDE_ALL_PARTS': 'Hide the "Parts" menu entry', + 'BK_HIDE_ALL_PROBLEMS_PARTS': 'Hide the "Problems" menu entry', + 'BK_HIDE_ALL_SETS': 'Hide the "Sets" menu entry', + 'BK_HIDE_ALL_STORAGES': 'Hide the "Storages" menu entry', + 'BK_HIDE_STATISTICS': 'Hide the "Statistics" menu entry', + 'BK_HIDE_SET_INSTRUCTIONS': 'Hide instructions section in set details', + 'BK_HIDE_TABLE_DAMAGED_PARTS': 'Hide the "Damaged" column in parts tables', + 'BK_HIDE_TABLE_MISSING_PARTS': 'Hide the "Missing" column in parts tables', + 'BK_HIDE_TABLE_CHECKED_PARTS': 'Hide the "Checked" column in parts tables', + 'BK_HIDE_WISHES': 'Hide the "Wishes" menu entry', + 'BK_SETS_CONSOLIDATION': 'Enable set consolidation/grouping functionality', + 'BK_SHOW_GRID_FILTERS': 'Show filter options on grids by default', + 'BK_SHOW_GRID_SORT': 'Show sort options on grids by default', + 'BK_SKIP_SPARE_PARTS': 'Skip importing spare parts when downloading sets from Rebrickable', + 'BK_HIDE_SPARE_PARTS': 'Hide spare parts from parts lists (spare parts must still be in database)', + 'BK_USE_REMOTE_IMAGES': 'Use remote images from Rebrickable CDN instead of local', + 'BK_STATISTICS_SHOW_CHARTS': 'Show collection growth charts on statistics page', + 'BK_STATISTICS_DEFAULT_EXPANDED': 'Expand all statistics sections by default', + 'BK_DARK_MODE': 'Enable dark mode theme' + } + return help_text.get(var_name, 'No help available for this variable') \ No newline at end of file diff --git a/bricktracker/configuration.py b/bricktracker/configuration.py index 6542b57..b0ef8f1 100644 --- a/bricktracker/configuration.py +++ b/bricktracker/configuration.py @@ -60,7 +60,7 @@ class BrickConfiguration(object): if self.cast == bool and isinstance(value, str): value = value.lower() in ('true', 'yes', '1') - # Static path fixup + # Static path fixup (legacy - only for paths with s: True flag) if self.static_path and isinstance(value, str): value = os.path.normpath(value) @@ -70,6 +70,10 @@ class BrickConfiguration(object): # Remove static prefix value = value.removeprefix('static/') + # Normalize regular paths (not marked as static) + elif not self.static_path and isinstance(value, str) and ('FOLDER' in self.name or 'PATH' in self.name): + value = os.path.normpath(value) + # Type casting if self.cast is not None: self.value = self.cast(value) diff --git a/bricktracker/instructions.py b/bricktracker/instructions.py index cde7734..a2ffb53 100644 --- a/bricktracker/instructions.py +++ b/bricktracker/instructions.py @@ -13,7 +13,6 @@ import requests from werkzeug.datastructures import FileStorage from werkzeug.utils import secure_filename import re -import cloudscraper from .exceptions import ErrorException, DownloadException if TYPE_CHECKING: @@ -101,16 +100,39 @@ class BrickInstructions(object): # Skip if we already have it if os.path.isfile(target): + pdf_url = self.url() return self.socket.complete( - message=f"File {self.filename} already exists, skipped" + message=f'File {self.filename} already exists, skipped - Open PDF' ) - # Fetch PDF via cloudscraper (to bypass Cloudflare) - scraper = cloudscraper.create_scraper() - scraper.headers.update({ - "User-Agent": current_app.config['REBRICKABLE_USER_AGENT'] + # Use plain requests instead of cloudscraper + session = requests.Session() + session.headers.update({ + 'User-Agent': current_app.config['REBRICKABLE_USER_AGENT'], + 'Accept': 'text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8', + 'Accept-Language': 'en-US,en;q=0.5', + 'DNT': '1', + 'Connection': 'keep-alive', + 'Upgrade-Insecure-Requests': '1', + 'Sec-Fetch-Dest': 'document', + 'Sec-Fetch-Mode': 'navigate', + 'Sec-Fetch-Site': 'same-origin', + 'Cache-Control': 'max-age=0' }) - resp = scraper.get(path, stream=True) + + # Visit the set's instructions listing page first to establish session cookies + set_number = None + if self.rebrickable: + set_number = self.rebrickable.fields.set + elif self.set: + set_number = self.set + + if set_number: + instructions_page = f"https://rebrickable.com/instructions/{set_number}/" + session.get(instructions_page) + session.headers.update({"Referer": instructions_page}) + + resp = session.get(path, stream=True, allow_redirects=True) if not resp.ok: raise DownloadException(f"Failed to download: HTTP {resp.status_code}") @@ -141,8 +163,9 @@ class BrickInstructions(object): # Done! logger.info(f"Downloaded {self.filename}") + pdf_url = self.url() self.socket.complete( - message=f"File {self.filename} downloaded ({self.human_size()})" + message=f'File {self.filename} downloaded ({self.human_size()}) - Open PDF' ) except Exception as e: @@ -170,11 +193,16 @@ class BrickInstructions(object): if filename is None: filename = self.filename - return os.path.join( - current_app.static_folder, # type: ignore - current_app.config['INSTRUCTIONS_FOLDER'], - filename - ) + folder = current_app.config['INSTRUCTIONS_FOLDER'] + + # If folder is absolute, use it directly + # Otherwise, make it relative to app root (not static folder) + if os.path.isabs(folder): + base_path = folder + else: + base_path = os.path.join(current_app.root_path, folder) + + return os.path.join(base_path, filename) # Rename an instructions file def rename(self, filename: str, /) -> None: @@ -215,10 +243,16 @@ class BrickInstructions(object): folder: str = current_app.config['INSTRUCTIONS_FOLDER'] - # Compute the path - path = os.path.join(folder, self.filename) - - return url_for('static', filename=path) + # Determine which route to use based on folder path + # If folder contains 'data' (new structure), use data route + # Otherwise use static route (legacy) + if 'data' in folder: + return url_for('data.serve_data_file', folder='instructions', filename=self.filename) + else: + # Legacy: folder is relative to static/ + folder_clean = folder.removeprefix('static/') + path = os.path.join(folder_clean, self.filename) + return url_for('static', filename=path) # Return the icon depending on the extension def icon(self, /) -> str: @@ -235,34 +269,49 @@ class BrickInstructions(object): @staticmethod def find_instructions(set: str, /) -> list[Tuple[str, str]]: """ - Scrape Rebrickable’s HTML and return a list of + Scrape Rebrickable's HTML and return a list of (filename_slug, download_url). Duplicate slugs get _1, _2, … """ page_url = f"https://rebrickable.com/instructions/{set}/" logger.debug(f"[find_instructions] fetching HTML from {page_url!r}") - # Solve Cloudflare’s challenge - scraper = cloudscraper.create_scraper() - scraper.headers.update({'User-Agent': current_app.config['REBRICKABLE_USER_AGENT']}) - resp = scraper.get(page_url) + # Use plain requests instead of cloudscraper + session = requests.Session() + session.headers.update({ + 'User-Agent': current_app.config['REBRICKABLE_USER_AGENT'], + 'Accept': 'text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8', + 'Accept-Language': 'en-US,en;q=0.5', + 'DNT': '1', + 'Connection': 'keep-alive', + 'Upgrade-Insecure-Requests': '1', + 'Sec-Fetch-Dest': 'document', + 'Sec-Fetch-Mode': 'navigate', + 'Sec-Fetch-Site': 'none', + 'Cache-Control': 'max-age=0' + }) + + resp = session.get(page_url) if not resp.ok: raise ErrorException(f'Failed to load instructions page for {set}. HTTP {resp.status_code}') soup = BeautifulSoup(resp.content, 'html.parser') + # Match download links with or without query parameters (e.g., ?cfe=timestamp&cfk=key) link_re = re.compile(r'^/instructions/\d+/.+/download/') raw: list[tuple[str, str]] = [] for a in soup.find_all('a', href=link_re): - img = a.find('img', alt=True) - if not img or set not in img['alt']: + img = a.find('img', alt=True) # type: ignore + if not img or set not in img['alt']: # type: ignore continue # Turn the alt text into a slug - alt_text = img['alt'].removeprefix('LEGO Building Instructions for ') + alt_text = img['alt'].removeprefix('LEGO Building Instructions for ') # type: ignore slug = re.sub(r'[^A-Za-z0-9]+', '-', alt_text).strip('-') - # Build the absolute download URL - download_url = urljoin('https://rebrickable.com', a['href']) + # Build the absolute download URL - this preserves query parameters + # BeautifulSoup's a['href'] includes the full href with ?cfe=...&cfk=... params + download_url = urljoin('https://rebrickable.com', a['href']) # type: ignore + logger.debug(f"[find_instructions] Found download link: {download_url}") raw.append((slug, download_url)) if not raw: diff --git a/bricktracker/instructions_list.py b/bricktracker/instructions_list.py index 6fc364d..259d1d7 100644 --- a/bricktracker/instructions_list.py +++ b/bricktracker/instructions_list.py @@ -36,11 +36,14 @@ class BrickInstructionsList(object): # Try to list the files in the instruction folder try: - # Make a folder relative to static - folder: str = os.path.join( - current_app.static_folder, # type: ignore - current_app.config['INSTRUCTIONS_FOLDER'], - ) + folder_config: str = current_app.config['INSTRUCTIONS_FOLDER'] + + # If folder is absolute, use it directly + # Otherwise, make it relative to app root (not static folder) + if os.path.isabs(folder_config): + folder = folder_config + else: + folder = os.path.join(current_app.root_path, folder_config) for file in os.scandir(folder): instruction = BrickInstructions(file) diff --git a/bricktracker/metadata.py b/bricktracker/metadata.py index 88d2623..a285bd1 100644 --- a/bricktracker/metadata.py +++ b/bricktracker/metadata.py @@ -191,19 +191,15 @@ class BrickMetadata(BrickRecord): parameters['set_id'] = brickset.fields.id parameters['state'] = state - rows, _ = BrickSQL().execute_and_commit( + rows, _ = BrickSQL().execute( self.update_set_state_query, parameters=parameters, + defer=True, name=self.as_column(), ) - if rows != 1: - raise DatabaseException('Could not update the {kind} "{name}" state for set {set} ({id})'.format( # noqa: E501 - kind=self.kind, - name=self.fields.name, - set=brickset.fields.set, - id=brickset.fields.id, - )) + # Note: rows will be -1 when deferred, so we can't validate here + # Validation will happen at final commit in set.py # Info logger.info('{kind} "{name}" state changed to "{state}" for set {set} ({id})'.format( # noqa: E501 diff --git a/bricktracker/metadata_list.py b/bricktracker/metadata_list.py index 13b9629..aa2eb67 100644 --- a/bricktracker/metadata_list.py +++ b/bricktracker/metadata_list.py @@ -111,6 +111,16 @@ class BrickMetadataList(BrickRecordList[T]): in new.filter(**kwargs) ]) + # Return the items as a dictionary mapping column names to UUIDs + @classmethod + def as_column_mapping(cls, /, **kwargs) -> dict: + new = cls.new() + + return { + record.as_column(): record.fields.id + for record in new.filter(**kwargs) + } + # Grab a specific status @classmethod def get(cls, id: str | None, /, *, allow_none: bool = False) -> T: diff --git a/bricktracker/minifigure_list.py b/bricktracker/minifigure_list.py index 40c13d0..f45a22b 100644 --- a/bricktracker/minifigure_list.py +++ b/bricktracker/minifigure_list.py @@ -43,6 +43,19 @@ class BrickMinifigureList(BrickRecordList[BrickMinifigure]): return self + # Load all minifigures with problems filter + def all_filtered(self, /, problems_filter: str = 'all', theme_id: str = 'all', year: str = 'all') -> Self: + context = {} + if problems_filter and problems_filter != 'all': + context['problems_filter'] = problems_filter + if theme_id and theme_id != 'all': + context['theme_id'] = theme_id + if year and year != 'all': + context['year'] = year + + self.list(override_query=self.all_query, **context) + return self + # Load all minifigures by owner def all_by_owner(self, owner_id: str | None = None, /) -> Self: # Save the owner_id parameter @@ -53,6 +66,78 @@ class BrickMinifigureList(BrickRecordList[BrickMinifigure]): return self + # Load all minifigures by owner with problems filter + def all_by_owner_filtered(self, /, owner_id: str | None = None, problems_filter: str = 'all', theme_id: str = 'all', year: str = 'all') -> Self: + # Save the owner_id parameter + self.fields.owner_id = owner_id + + context = {} + if problems_filter and problems_filter != 'all': + context['problems_filter'] = problems_filter + if theme_id and theme_id != 'all': + context['theme_id'] = theme_id + if year and year != 'all': + context['year'] = year + + # Load the minifigures from the database + self.list(override_query=self.all_by_owner_query, **context) + + return self + + # Load minifigures with pagination support + def all_filtered_paginated( + self, + owner_id: str | None = None, + problems_filter: str = 'all', + theme_id: str = 'all', + year: str = 'all', + search_query: str | None = None, + page: int = 1, + per_page: int = 50, + sort_field: str | None = None, + sort_order: str = 'asc' + ) -> tuple[Self, int]: + # Prepare filter context + filter_context = {} + if owner_id and owner_id != 'all': + filter_context['owner_id'] = owner_id + list_query = self.all_by_owner_query + else: + list_query = self.all_query + + if search_query: + filter_context['search_query'] = search_query + + if problems_filter and problems_filter != 'all': + filter_context['problems_filter'] = problems_filter + + if theme_id and theme_id != 'all': + filter_context['theme_id'] = theme_id + + if year and year != 'all': + filter_context['year'] = year + + # Field mapping for sorting + field_mapping = { + 'name': '"rebrickable_minifigures"."name"', + 'parts': '"rebrickable_minifigures"."number_of_parts"', + 'quantity': '"total_quantity"', + 'missing': '"total_missing"', + 'damaged': '"total_damaged"', + 'sets': '"total_sets"' + } + + # Use the base pagination method + return self.paginate( + page=page, + per_page=per_page, + sort_field=sort_field, + sort_order=sort_order, + list_query=list_query, + field_mapping=field_mapping, + **filter_context + ) + # Minifigures with a part damaged part def damaged_part(self, part: str, color: int, /) -> Self: # Save the parameters to the fields @@ -95,16 +180,19 @@ class BrickMinifigureList(BrickRecordList[BrickMinifigure]): brickset = None # Prepare template context for owner filtering - context = {} + context_vars = {} if hasattr(self.fields, 'owner_id') and self.fields.owner_id is not None: - context['owner_id'] = self.fields.owner_id + context_vars['owner_id'] = self.fields.owner_id + + # Merge with any additional context passed in + context_vars.update(context) # Load the sets from the database for record in super().select( override_query=override_query, order=order, limit=limit, - **context + **context_vars ): minifigure = BrickMinifigure(brickset=brickset, record=record) diff --git a/bricktracker/navbar.py b/bricktracker/navbar.py index 20a2b29..893dead 100644 --- a/bricktracker/navbar.py +++ b/bricktracker/navbar.py @@ -15,6 +15,7 @@ NAVBAR: Final[list[dict[str, Any]]] = [ {'e': 'minifigure.list', 't': 'Minifigures', 'i': 'group-line', 'f': 'HIDE_ALL_MINIFIGURES'}, # noqa: E501 {'e': 'instructions.list', 't': 'Instructions', 'i': 'file-line', 'f': 'HIDE_ALL_INSTRUCTIONS'}, # noqa: E501 {'e': 'storage.list', 't': 'Storages', 'i': 'archive-2-line', 'f': 'HIDE_ALL_STORAGES'}, # noqa: E501 + {'e': 'statistics.overview', 't': 'Statistics', 'i': 'bar-chart-line', 'f': 'HIDE_STATISTICS'}, # noqa: E501 {'e': 'wish.list', 't': 'Wishlist', 'i': 'gift-line', 'f': 'HIDE_WISHES'}, {'e': 'admin.admin', 't': 'Admin', 'i': 'settings-4-line', 'f': 'HIDE_ADMIN'}, # noqa: E501 ] diff --git a/bricktracker/pagination_helper.py b/bricktracker/pagination_helper.py new file mode 100644 index 0000000..183c25f --- /dev/null +++ b/bricktracker/pagination_helper.py @@ -0,0 +1,52 @@ +from flask import current_app, request +from typing import Any, Dict, Tuple + + +def get_pagination_config(entity_type: str) -> Tuple[int, bool]: + """Get pagination configuration for an entity type (sets, parts, minifigures)""" + # Check if pagination is enabled for this specific entity type + pagination_key = f'{entity_type.upper()}_SERVER_SIDE_PAGINATION' + use_pagination = current_app.config.get(pagination_key, False) + + if not use_pagination: + return 0, False + + # Determine page size based on device type and entity + user_agent = request.headers.get('User-Agent', '').lower() + is_mobile = any(device in user_agent for device in ['mobile', 'android', 'iphone', 'ipad']) + + # Get appropriate config keys based on entity type + entity_upper = entity_type.upper() + desktop_key = f'{entity_upper}_PAGINATION_SIZE_DESKTOP' + mobile_key = f'{entity_upper}_PAGINATION_SIZE_MOBILE' + + per_page = current_app.config[mobile_key] if is_mobile else current_app.config[desktop_key] + + return per_page, is_mobile + + +def build_pagination_context(page: int, per_page: int, total_count: int, is_mobile: bool) -> Dict[str, Any]: + """Build pagination context for templates""" + total_pages = (total_count + per_page - 1) // per_page if total_count > 0 else 1 + has_prev = page > 1 + has_next = page < total_pages + + return { + 'page': page, + 'per_page': per_page, + 'total_count': total_count, + 'total_pages': total_pages, + 'has_prev': has_prev, + 'has_next': has_next, + 'is_mobile': is_mobile + } + + +def get_request_params() -> Tuple[str, str, str, int]: + """Extract common request parameters for pagination""" + search_query = request.args.get('search', '').strip() + sort_field = request.args.get('sort', '') + sort_order = request.args.get('order', 'asc') + page = int(request.args.get('page', 1)) + + return search_query, sort_field, sort_order, page \ No newline at end of file diff --git a/bricktracker/parser.py b/bricktracker/parser.py index d3602e2..5764bfe 100644 --- a/bricktracker/parser.py +++ b/bricktracker/parser.py @@ -5,33 +5,29 @@ from .exceptions import ErrorException def parse_set(set: str, /) -> str: number, _, version = set.partition('-') - # Making sure both are integers + # Set number can be alphanumeric (e.g., "McDR6US", "10312", "COMCON035") + # Just validate it's not empty + if not number or number.strip() == '': + raise ErrorException('Set number cannot be empty') + + # Clean up the number (trim whitespace) + number = number.strip() + + # Version defaults to 1 if not provided if version == '': - version = 1 + version = '1' + # Version must be a positive integer try: - number = int(number) - except Exception: - raise ErrorException('Number "{number}" is not a number'.format( - number=number, - )) - - try: - version = int(version) + version_int = int(version) except Exception: raise ErrorException('Version "{version}" is not a number'.format( version=version, )) - # Make sure both are positive - if number < 0: - raise ErrorException('Number "{number}" should be positive'.format( - number=number, - )) - - if version < 0: - raise ErrorException('Version "{version}" should be positive'.format( # noqa: E501 + if version_int < 0: + raise ErrorException('Version "{version}" should be positive'.format( version=version, )) - return '{number}-{version}'.format(number=number, version=version) + return '{number}-{version}'.format(number=number, version=version_int) diff --git a/bricktracker/part.py b/bricktracker/part.py index 12eab28..ba1e427 100644 --- a/bricktracker/part.py +++ b/bricktracker/part.py @@ -159,6 +159,43 @@ class BrickPart(RebrickablePart): return self + # Update checked state for part walkthrough + def update_checked(self, json: Any | None, /) -> bool: + # Handle both direct 'checked' key and changer.js 'value' key format + if json: + checked = json.get('checked', json.get('value', False)) + else: + checked = False + + checked = bool(checked) + + # Update the field + self.fields.checked = checked + + BrickSQL().execute_and_commit( + 'part/update/checked', + parameters=self.sql_parameters() + ) + + return checked + + # Compute the url for updating checked state + def url_for_checked(self, /) -> str: + # Different URL for a minifigure part + if self.minifigure is not None: + figure = self.minifigure.fields.figure + else: + figure = None + + return url_for( + 'set.checked_part', + id=self.fields.id, + figure=figure, + part=self.fields.part, + color=self.fields.color, + spare=self.fields.spare, + ) + # Update a problematic part def update_problem(self, problem: str, json: Any | None, /) -> int: amount: str | int = json.get('value', '') # type: ignore diff --git a/bricktracker/part_list.py b/bricktracker/part_list.py index 96f8653..070750d 100644 --- a/bricktracker/part_list.py +++ b/bricktracker/part_list.py @@ -57,8 +57,8 @@ class BrickPartList(BrickRecordList[BrickPart]): return self - # Load all parts with filters (owner and/or color) - def all_filtered(self, owner_id: str | None = None, color_id: str | None = None, /) -> Self: + # Load all parts with filters (owner, color, theme, year) + def all_filtered(self, owner_id: str | None = None, color_id: str | None = None, theme_id: str | None = None, year: str | None = None, /) -> Self: # Save the filter parameters if owner_id is not None: self.fields.owner_id = owner_id @@ -71,11 +71,76 @@ class BrickPartList(BrickRecordList[BrickPart]): else: query = self.all_query + # Prepare context for query + context = {} + # Hide spare parts from display if configured + if current_app.config.get('HIDE_SPARE_PARTS', False): + context['skip_spare_parts'] = True + if theme_id and theme_id != 'all': + context['theme_id'] = theme_id + if year and year != 'all': + context['year'] = year + # Load the parts from the database - self.list(override_query=query) + self.list(override_query=query, **context) return self + # Load parts with pagination support + def all_filtered_paginated( + self, + owner_id: str | None = None, + color_id: str | None = None, + theme_id: str | None = None, + year: str | None = None, + search_query: str | None = None, + page: int = 1, + per_page: int = 50, + sort_field: str | None = None, + sort_order: str = 'asc' + ) -> tuple[Self, int]: + # Prepare filter context + filter_context = {} + if owner_id and owner_id != 'all': + filter_context['owner_id'] = owner_id + list_query = self.all_by_owner_query + else: + list_query = self.all_query + + if color_id and color_id != 'all': + filter_context['color_id'] = color_id + if theme_id and theme_id != 'all': + filter_context['theme_id'] = theme_id + if year and year != 'all': + filter_context['year'] = year + if search_query: + filter_context['search_query'] = search_query + # Hide spare parts from display if configured + if current_app.config.get('HIDE_SPARE_PARTS', False): + filter_context['skip_spare_parts'] = True + + # Field mapping for sorting + field_mapping = { + 'name': '"rebrickable_parts"."name"', + 'color': '"rebrickable_parts"."color_name"', + 'quantity': '"total_quantity"', + 'missing': '"total_missing"', + 'damaged': '"total_damaged"', + 'sets': '"total_sets"', + 'minifigures': '"total_minifigures"' + } + + # Use the base pagination method + return self.paginate( + page=page, + per_page=per_page, + sort_field=sort_field, + sort_order=sort_order, + list_query=list_query, + field_mapping=field_mapping, + **filter_context + ) + # Base part list def list( self, @@ -84,6 +149,7 @@ class BrickPartList(BrickRecordList[BrickPart]): override_query: str | None = None, order: str | None = None, limit: int | None = None, + offset: int | None = None, **context: Any, ) -> None: if order is None: @@ -105,12 +171,18 @@ class BrickPartList(BrickRecordList[BrickPart]): context_vars['owner_id'] = self.fields.owner_id if hasattr(self.fields, 'color_id') and self.fields.color_id is not None: context_vars['color_id'] = self.fields.color_id + if hasattr(self.fields, 'search_query') and self.fields.search_query: + context_vars['search_query'] = self.fields.search_query + + # Merge with any additional context passed in + context_vars.update(context) # Load the sets from the database for record in super().select( override_query=override_query, order=order, limit=limit, + offset=offset, **context_vars ): part = BrickPart( @@ -119,9 +191,6 @@ class BrickPartList(BrickRecordList[BrickPart]): record=record, ) - if current_app.config['SKIP_SPARE_PARTS'] and part.fields.spare: - continue - self.records.append(part) # List specific parts from a brickset or minifigure @@ -136,8 +205,13 @@ class BrickPartList(BrickRecordList[BrickPart]): self.brickset = brickset self.minifigure = minifigure + # Prepare context for hiding spare parts if configured + context = {} + if current_app.config.get('HIDE_SPARE_PARTS', False): + context['skip_spare_parts'] = True + # Load the parts from the database - self.list() + self.list(**context) return self @@ -150,8 +224,13 @@ class BrickPartList(BrickRecordList[BrickPart]): # Save the minifigure self.minifigure = minifigure + # Prepare context for hiding spare parts if configured + context = {} + if current_app.config.get('HIDE_SPARE_PARTS', False): + context['skip_spare_parts'] = True + # Load the parts from the database - self.list(override_query=self.minifigure_query) + self.list(override_query=self.minifigure_query, **context) return self @@ -181,6 +260,92 @@ class BrickPartList(BrickRecordList[BrickPart]): return self + def problem_filtered(self, owner_id: str | None = None, color_id: str | None = None, theme_id: str | None = None, year: str | None = None, storage_id: str | None = None, tag_id: str | None = None, /) -> Self: + # Save the filter parameters for client-side filtering + if owner_id is not None: + self.fields.owner_id = owner_id + if color_id is not None: + self.fields.color_id = color_id + + # Prepare context for query + context = {} + if owner_id and owner_id != 'all': + context['owner_id'] = owner_id + if color_id and color_id != 'all': + context['color_id'] = color_id + if theme_id and theme_id != 'all': + context['theme_id'] = theme_id + if year and year != 'all': + context['year'] = year + if storage_id and storage_id != 'all': + context['storage_id'] = storage_id + if tag_id and tag_id != 'all': + context['tag_id'] = tag_id + # Hide spare parts from display if configured + if current_app.config.get('HIDE_SPARE_PARTS', False): + context['skip_spare_parts'] = True + + # Load the problematic parts from the database + self.list(override_query=self.problem_query, **context) + + return self + + def problem_paginated( + self, + owner_id: str | None = None, + color_id: str | None = None, + theme_id: str | None = None, + year: str | None = None, + storage_id: str | None = None, + tag_id: str | None = None, + search_query: str | None = None, + page: int = 1, + per_page: int = 50, + sort_field: str | None = None, + sort_order: str = 'asc' + ) -> tuple[Self, int]: + # Prepare filter context + filter_context = {} + if owner_id and owner_id != 'all': + filter_context['owner_id'] = owner_id + if color_id and color_id != 'all': + filter_context['color_id'] = color_id + if theme_id and theme_id != 'all': + filter_context['theme_id'] = theme_id + if year and year != 'all': + filter_context['year'] = year + if storage_id and storage_id != 'all': + filter_context['storage_id'] = storage_id + if tag_id and tag_id != 'all': + filter_context['tag_id'] = tag_id + if search_query: + filter_context['search_query'] = search_query + # Hide spare parts from display if configured + if current_app.config.get('HIDE_SPARE_PARTS', False): + filter_context['skip_spare_parts'] = True + + # Field mapping for sorting + field_mapping = { + 'name': '"rebrickable_parts"."name"', + 'color': '"rebrickable_parts"."color_name"', + 'quantity': '"total_quantity"', + 'missing': '"total_missing"', + 'damaged': '"total_damaged"', + 'sets': '"total_sets"', + 'minifigures': '"total_minifigures"' + } + + # Use the base pagination method with problem query + return self.paginate( + page=page, + per_page=per_page, + sort_field=sort_field, + sort_order=sort_order, + list_query=self.problem_query, + field_mapping=field_mapping, + **filter_context + ) + # Return a dict with common SQL parameters for a parts list def sql_parameters(self, /) -> dict[str, Any]: parameters: dict[str, Any] = super().sql_parameters() @@ -256,7 +421,13 @@ class BrickPartList(BrickRecordList[BrickPart]): # Process each part number_of_parts: int = 0 + skip_spares = current_app.config.get('SKIP_SPARE_PARTS', False) + for part in inventory: + # Skip spare parts if configured + if skip_spares and part.fields.spare: + continue + # Count the number of parts for minifigures if minifigure is not None: number_of_parts += part.fields.quantity diff --git a/bricktracker/peeron_instructions.py b/bricktracker/peeron_instructions.py new file mode 100644 index 0000000..dc6d267 --- /dev/null +++ b/bricktracker/peeron_instructions.py @@ -0,0 +1,436 @@ +import hashlib +import logging +import os +from pathlib import Path +import time +from typing import Any, NamedTuple, TYPE_CHECKING +from urllib.parse import urljoin + +from bs4 import BeautifulSoup +from flask import current_app, url_for +import requests + +from .exceptions import ErrorException +if TYPE_CHECKING: + from .socket import BrickSocket + +logger = logging.getLogger(__name__) + + +def get_peeron_user_agent(): + """Get the User-Agent string for Peeron requests from config""" + return current_app.config.get('REBRICKABLE_USER_AGENT', + 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/91.0.4472.124 Safari/537.36') + + +def get_peeron_download_delay(): + """Get the delay in milliseconds between Peeron page downloads from config""" + return current_app.config.get('PEERON_DOWNLOAD_DELAY', 1000) + + +def get_min_image_size(): + """Get the minimum image size for valid Peeron instruction pages from config""" + return current_app.config.get('PEERON_MIN_IMAGE_SIZE', 100) + + +def get_peeron_instruction_url(set_number: str, version_number: str): + """Get the Peeron instruction page URL using the configured pattern""" + pattern = current_app.config.get('PEERON_INSTRUCTION_PATTERN', 'http://peeron.com/scans/{set_number}-{version_number}') + return pattern.format(set_number=set_number, version_number=version_number) + + +def get_peeron_thumbnail_url(set_number: str, version_number: str): + """Get the Peeron thumbnail base URL using the configured pattern""" + pattern = current_app.config.get('PEERON_THUMBNAIL_PATTERN', 'http://belay.peeron.com/thumbs/{set_number}-{version_number}/') + return pattern.format(set_number=set_number, version_number=version_number) + + +def get_peeron_scan_url(set_number: str, version_number: str): + """Get the Peeron scan base URL using the configured pattern""" + pattern = current_app.config.get('PEERON_SCAN_PATTERN', 'http://belay.peeron.com/scans/{set_number}-{version_number}/') + return pattern.format(set_number=set_number, version_number=version_number) + + +def create_peeron_scraper(): + """Create a requests session configured for Peeron""" + session = requests.Session() + session.headers.update({ + "User-Agent": get_peeron_user_agent() + }) + return session + + +def get_peeron_cache_dir(): + """Get the base directory for Peeron caching""" + static_dir = Path(current_app.static_folder) + cache_dir = static_dir / 'images' / 'peeron_cache' + cache_dir.mkdir(parents=True, exist_ok=True) + return cache_dir + + +def get_set_cache_dir(set_number: str, version_number: str) -> tuple[Path, Path]: + """Get cache directories for a specific set""" + base_cache_dir = get_peeron_cache_dir() + set_cache_key = f"{set_number}-{version_number}" + + full_cache_dir = base_cache_dir / 'full' / set_cache_key + thumb_cache_dir = base_cache_dir / 'thumbs' / set_cache_key + + full_cache_dir.mkdir(parents=True, exist_ok=True) + thumb_cache_dir.mkdir(parents=True, exist_ok=True) + + return full_cache_dir, thumb_cache_dir + + +def cache_full_image_and_generate_thumbnail(image_url: str, page_number: str, set_number: str, version_number: str, session=None) -> tuple[str | None, str | None]: + """ + Download and cache full-size image, then generate a thumbnail preview. + Uses the full-size scan URLs from Peeron. + Returns (cached_image_path, thumbnail_url) or (None, None) if caching fails. + """ + try: + full_cache_dir, thumb_cache_dir = get_set_cache_dir(set_number, version_number) + + full_filename = f"{page_number}.jpg" + thumb_filename = f"{page_number}.jpg" + full_cache_path = full_cache_dir / full_filename + thumb_cache_path = thumb_cache_dir / thumb_filename + + # Return existing cached files if they exist + if full_cache_path.exists() and thumb_cache_path.exists(): + set_cache_key = f"{set_number}-{version_number}" + thumbnail_url = url_for('static', filename=f'images/peeron_cache/thumbs/{set_cache_key}/{thumb_filename}') + return str(full_cache_path), thumbnail_url + + # Download the full-size image using provided session or create new one + if session is None: + session = create_peeron_scraper() + response = session.get(image_url, timeout=30) + + if response.status_code == 200 and len(response.content) > 0: + # Validate it's actually an image by checking minimum size + min_size = get_min_image_size() + if len(response.content) < min_size: + logger.warning(f"Image too small, skipping cache: {image_url}") + return None, None + + # Write full-size image to cache + with open(full_cache_path, 'wb') as f: + f.write(response.content) + + logger.debug(f"Cached full image: {image_url} -> {full_cache_path}") + + # Generate thumbnail from the cached full image + try: + from PIL import Image + with Image.open(full_cache_path) as img: + # Create thumbnail (max 150px on longest side to match template) + img.thumbnail((150, 150), Image.Resampling.LANCZOS) + img.save(thumb_cache_path, 'JPEG', quality=85) + + logger.debug(f"Generated thumbnail: {full_cache_path} -> {thumb_cache_path}") + + set_cache_key = f"{set_number}-{version_number}" + thumbnail_url = url_for('static', filename=f'images/peeron_cache/thumbs/{set_cache_key}/{thumb_filename}') + return str(full_cache_path), thumbnail_url + + except Exception as thumb_error: + logger.error(f"Failed to generate thumbnail for {page_number}: {thumb_error}") + # Clean up the full image if thumbnail generation failed + if full_cache_path.exists(): + full_cache_path.unlink() + return None, None + else: + logger.warning(f"Failed to download full image: {image_url}") + return None, None + + except Exception as e: + logger.error(f"Error caching full image {image_url}: {e}") + return None, None + + +def clear_set_cache(set_number: str, version_number: str) -> int: + """ + Clear all cached files for a specific set after PDF generation. + Returns the number of files deleted. + """ + try: + full_cache_dir, thumb_cache_dir = get_set_cache_dir(set_number, version_number) + deleted_count = 0 + + # Delete full images + if full_cache_dir.exists(): + for cache_file in full_cache_dir.glob('*.jpg'): + try: + cache_file.unlink() + deleted_count += 1 + logger.debug(f"Deleted cached full image: {cache_file}") + except OSError as e: + logger.warning(f"Failed to delete cache file {cache_file}: {e}") + + # Remove directory if empty + try: + full_cache_dir.rmdir() + except OSError: + pass # Directory not empty or other error + + # Delete thumbnails + if thumb_cache_dir.exists(): + for cache_file in thumb_cache_dir.glob('*.jpg'): + try: + cache_file.unlink() + deleted_count += 1 + logger.debug(f"Deleted cached thumbnail: {cache_file}") + except OSError as e: + logger.warning(f"Failed to delete cache file {cache_file}: {e}") + + # Remove directory if empty + try: + thumb_cache_dir.rmdir() + except OSError: + pass # Directory not empty or other error + + # Try to remove set directory if empty + try: + set_cache_key = f"{set_number}-{version_number}" + full_cache_dir.parent.rmdir() if full_cache_dir.parent.name == set_cache_key else None + thumb_cache_dir.parent.rmdir() if thumb_cache_dir.parent.name == set_cache_key else None + except OSError: + pass # Directory not empty or other error + + logger.info(f"Set cache cleanup completed for {set_number}-{version_number}: {deleted_count} files deleted") + return deleted_count + + except Exception as e: + logger.error(f"Error during set cache cleanup for {set_number}-{version_number}: {e}") + return 0 + + +def clear_old_cache(max_age_days: int = 7) -> int: + """ + Clear old cache files across all sets. + Returns the number of files deleted. + """ + try: + base_cache_dir = get_peeron_cache_dir() + if not base_cache_dir.exists(): + return 0 + + deleted_count = 0 + max_age_seconds = max_age_days * 24 * 60 * 60 + current_time = time.time() + + # Clean both full and thumbs directories + for cache_type in ['full', 'thumbs']: + cache_type_dir = base_cache_dir / cache_type + if cache_type_dir.exists(): + for set_dir in cache_type_dir.iterdir(): + if set_dir.is_dir(): + for cache_file in set_dir.glob('*.jpg'): + file_age = current_time - os.path.getmtime(cache_file) + if file_age > max_age_seconds: + try: + cache_file.unlink() + deleted_count += 1 + logger.debug(f"Deleted old cache file: {cache_file}") + except OSError as e: + logger.warning(f"Failed to delete cache file {cache_file}: {e}") + + # Remove empty directories + try: + if not any(set_dir.iterdir()): + set_dir.rmdir() + except OSError: + pass + + logger.info(f"Old cache cleanup completed: {deleted_count} files deleted") + return deleted_count + + except Exception as e: + logger.error(f"Error during old cache cleanup: {e}") + return 0 + + +class PeeronPage(NamedTuple): + """Represents a single instruction page from Peeron""" + page_number: str + original_image_url: str # Original Peeron full-size image URL + cached_full_image_path: str # Local full-size cached image path + cached_thumbnail_url: str # Local thumbnail URL for preview + alt_text: str + rotation: int = 0 # Rotation in degrees (0, 90, 180, 270) + + +# Peeron instruction scraper +class PeeronInstructions(object): + socket: 'BrickSocket | None' + set_number: str + version_number: str + pages: list[PeeronPage] + + def __init__( + self, + set_number: str, + version_number: str = '1', + /, + *, + socket: 'BrickSocket | None' = None, + ): + # Save the socket + self.socket = socket + + # Parse set number (handle both "4011" and "4011-1" formats) + if '-' in set_number: + parts = set_number.split('-', 1) + self.set_number = parts[0] + self.version_number = parts[1] if len(parts) > 1 else '1' + else: + self.set_number = set_number + self.version_number = version_number + + # Placeholder for pages + self.pages = [] + + # Check if instructions exist on Peeron (lightweight) + def exists(self, /) -> bool: + """Check if the set exists on Peeron without caching thumbnails""" + try: + base_url = get_peeron_instruction_url(self.set_number, self.version_number) + scraper = create_peeron_scraper() + response = scraper.get(base_url) + + if response.status_code != 200: + return False + + soup = BeautifulSoup(response.text, 'html.parser') + + # Check for "Browse instruction library" header (set not found) + if soup.find('h1', string="Browse instruction library"): + return False + + # Look for thumbnail images to confirm instructions exist + thumbnails = soup.select('table[cellspacing="5"] a img[src^="http://belay.peeron.com/thumbs/"]') + return len(thumbnails) > 0 + + except Exception: + return False + + # Find all available instruction pages on Peeron + def find_pages(self, /) -> list[PeeronPage]: + """ + Scrape Peeron's HTML and return a list of available instruction pages. + Similar to BrickInstructions.find_instructions() but for Peeron. + """ + base_url = get_peeron_instruction_url(self.set_number, self.version_number) + thumb_base_url = get_peeron_thumbnail_url(self.set_number, self.version_number) + scan_base_url = get_peeron_scan_url(self.set_number, self.version_number) + + logger.debug(f"[find_pages] fetching HTML from {base_url!r}") + + # Set up session with persistent cookies for Peeron (like working dl_peeron.py) + scraper = create_peeron_scraper() + + # Download the main HTML page to establish session and cookies + try: + logger.debug(f"[find_pages] Establishing session by visiting: {base_url}") + response = scraper.get(base_url) + logger.debug(f"[find_pages] Main page visit: HTTP {response.status_code}") + if response.status_code != 200: + raise ErrorException(f'Failed to load Peeron page for {self.set_number}-{self.version_number}. HTTP {response.status_code}') + except requests.exceptions.RequestException as e: + raise ErrorException(f'Failed to connect to Peeron: {e}') + + # Parse HTML to locate instruction pages + soup = BeautifulSoup(response.text, 'html.parser') + + # Check for "Browse instruction library" header (set not found) + if soup.find('h1', string="Browse instruction library"): + raise ErrorException(f'Set {self.set_number}-{self.version_number} not found on Peeron') + + # Locate all thumbnail images in the expected table structure + # Use the configured thumbnail pattern to build the expected URL prefix + thumb_base_url = get_peeron_thumbnail_url(self.set_number, self.version_number) + thumbnails = soup.select(f'table[cellspacing="5"] a img[src^="{thumb_base_url}"]') + + if not thumbnails: + raise ErrorException(f'No instruction pages found for {self.set_number}-{self.version_number} on Peeron') + + pages: list[PeeronPage] = [] + total_thumbnails = len(thumbnails) + + # Initialize progress if socket is available + if self.socket: + self.socket.progress_total = total_thumbnails + self.socket.progress_count = 0 + self.socket.progress(message=f"Starting to cache {total_thumbnails} full images") + + for idx, img in enumerate(thumbnails, 1): + thumb_url = img['src'] + + # Extract the page number from the thumbnail URL + page_number = thumb_url.split('/')[-2] + + # Build the full-size scan URL using the page number + full_size_url = f"{scan_base_url}{page_number}/" + + logger.debug(f"[find_pages] Page {page_number}: thumb={thumb_url}, full_size={full_size_url}") + + # Create alt text for the page + alt_text = f"LEGO Instructions {self.set_number}-{self.version_number} Page {page_number}" + + # Report progress if socket is available + if self.socket: + self.socket.progress_count = idx + self.socket.progress(message=f"Caching full image {idx} of {total_thumbnails}") + + # Cache the full-size image and generate thumbnail preview using established session + cached_full_path, cached_thumb_url = cache_full_image_and_generate_thumbnail( + full_size_url, page_number, self.set_number, self.version_number, session=scraper + ) + + # Skip this page if caching failed + if not cached_full_path or not cached_thumb_url: + logger.warning(f"[find_pages] Skipping page {page_number} due to caching failure") + continue + + page = PeeronPage( + page_number=page_number, + original_image_url=full_size_url, + cached_full_image_path=cached_full_path, + cached_thumbnail_url=cached_thumb_url, + alt_text=alt_text + ) + pages.append(page) + + # Cache the pages for later use + self.pages = pages + + logger.debug(f"[find_pages] found {len(pages)} pages for {self.set_number}-{self.version_number}") + return pages + + # Find instructions with fallback to Peeron + @staticmethod + def find_instructions_with_peeron_fallback(set: str, /) -> tuple[list[tuple[str, str]], list[PeeronPage] | None]: + """ + Enhanced version of BrickInstructions.find_instructions() that falls back to Peeron. + Returns (rebrickable_instructions, peeron_pages). + If rebrickable_instructions is empty, peeron_pages will contain Peeron data. + """ + from .instructions import BrickInstructions + + # First try Rebrickable + try: + rebrickable_instructions = BrickInstructions.find_instructions(set) + return rebrickable_instructions, None + except ErrorException as e: + logger.info(f"Rebrickable failed for {set}: {e}. Trying Peeron fallback...") + + # Fallback to Peeron + try: + peeron = PeeronInstructions(set) + peeron_pages = peeron.find_pages() + return [], peeron_pages + except ErrorException as peeron_error: + # Both failed, re-raise original Rebrickable error + logger.info(f"Peeron also failed for {set}: {peeron_error}") + raise e from peeron_error \ No newline at end of file diff --git a/bricktracker/peeron_pdf.py b/bricktracker/peeron_pdf.py new file mode 100644 index 0000000..77fd8db --- /dev/null +++ b/bricktracker/peeron_pdf.py @@ -0,0 +1,204 @@ +import logging +import os +import tempfile +import time +from typing import Any, TYPE_CHECKING + +from flask import current_app +from PIL import Image + +from .exceptions import DownloadException, ErrorException +from .instructions import BrickInstructions +from .peeron_instructions import PeeronPage, get_min_image_size, get_peeron_download_delay, get_peeron_instruction_url, create_peeron_scraper +if TYPE_CHECKING: + from .socket import BrickSocket + +logger = logging.getLogger(__name__) + + +# PDF generator for Peeron instruction pages +class PeeronPDF(object): + socket: 'BrickSocket' + set_number: str + version_number: str + pages: list[PeeronPage] + filename: str + + def __init__( + self, + set_number: str, + version_number: str, + pages: list[PeeronPage], + /, + *, + socket: 'BrickSocket', + ): + # Save the socket + self.socket = socket + + # Save set information + self.set_number = set_number + self.version_number = version_number + self.pages = pages + + # Generate filename following BrickTracker conventions + self.filename = f"{set_number}-{version_number}_peeron.pdf" + + # Download pages and create PDF + def create_pdf(self, /) -> None: + """ + Downloads selected Peeron pages and merges them into a PDF. + Uses progress updates via socket similar to BrickInstructions.download() + """ + try: + target_path = self._get_target_path() + + # Skip if we already have it + if os.path.isfile(target_path): + # Create BrickInstructions instance to get PDF URL + instructions = BrickInstructions(self.filename) + pdf_url = instructions.url() + return self.socket.complete( + message=f'File {self.filename} already exists, skipped - Open PDF' + ) + + # Set up progress tracking + total_pages = len(self.pages) + self.socket.update_total(total_pages) + self.socket.progress_count = 0 + self.socket.progress(message=f"Starting PDF creation from {total_pages} cached pages") + + # Use cached images directly - no downloads needed! + cached_files_with_rotation = [] + missing_pages = [] + + for i, page in enumerate(self.pages): + # Check if cached file exists + if os.path.isfile(page.cached_full_image_path): + cached_files_with_rotation.append((page.cached_full_image_path, page.rotation)) + + # Update progress + self.socket.progress_count += 1 + self.socket.progress( + message=f"Processing cached page {page.page_number} ({i + 1}/{total_pages})" + ) + else: + missing_pages.append(page.page_number) + logger.warning(f"Cached image missing for page {page.page_number}: {page.cached_full_image_path}") + + if not cached_files_with_rotation: + raise DownloadException(f"No cached images available for set {self.set_number}-{self.version_number}. Cache may have been cleared.") + + elif len(cached_files_with_rotation) < total_pages: + # Partial success + error_msg = f"Only found {len(cached_files_with_rotation)}/{total_pages} cached images." + if missing_pages: + error_msg += f" Missing pages: {', '.join(missing_pages)}." + logger.warning(error_msg) + + # Create PDF from cached images with rotation + self._create_pdf_from_images(cached_files_with_rotation, target_path) + + # Success + logger.info(f"Created PDF {self.filename} with {len(cached_files_with_rotation)} pages") + + # Create BrickInstructions instance to get PDF URL + instructions = BrickInstructions(self.filename) + pdf_url = instructions.url() + + self.socket.complete( + message=f'PDF {self.filename} created with {len(cached_files_with_rotation)} pages - Open PDF' + ) + + # Clean up set cache after successful PDF creation + try: + from .peeron_instructions import clear_set_cache + deleted_count = clear_set_cache(self.set_number, self.version_number) + if deleted_count > 0: + logger.info(f"[create_pdf] Cleaned up {deleted_count} cache files for set {self.set_number}-{self.version_number}") + except Exception as e: + logger.warning(f"[create_pdf] Failed to clean set cache: {e}") + + except Exception as e: + logger.error(f"Error creating PDF {self.filename}: {e}") + self.socket.fail( + message=f"Error creating PDF {self.filename}: {e}" + ) + + + # Create PDF from downloaded images + def _create_pdf_from_images(self, image_paths_and_rotations: list[tuple[str, int]], output_path: str, /) -> None: + """Create a PDF from a list of image files with their rotations""" + try: + # Import FPDF (should be available from requirements) + from fpdf import FPDF + except ImportError: + raise ErrorException("FPDF library not available. Install with: pip install fpdf2") + + pdf = FPDF() + + for i, (img_path, rotation) in enumerate(image_paths_and_rotations): + try: + # Open image and apply rotation if needed + with Image.open(img_path) as image: + # Apply rotation if specified + if rotation != 0: + # PIL rotation is counter-clockwise, so we negate for clockwise rotation + image = image.rotate(-rotation, expand=True) + + width, height = image.size + + # Add page with image dimensions (convert pixels to mm) + # 1 pixel = 0.264583 mm (assuming 96 DPI) + page_width = width * 0.264583 + page_height = height * 0.264583 + + pdf.add_page(format=(page_width, page_height)) + + # Save rotated image to temporary file for FPDF + temp_rotated_path = None + if rotation != 0: + import tempfile + temp_fd, temp_rotated_path = tempfile.mkstemp(suffix='.jpg', prefix=f'peeron_rotated_{i}_') + try: + os.close(temp_fd) # Close file descriptor, we'll use the path + image.save(temp_rotated_path, 'JPEG', quality=95) + pdf.image(temp_rotated_path, x=0, y=0, w=page_width, h=page_height) + finally: + # Clean up rotated temp file + if temp_rotated_path and os.path.exists(temp_rotated_path): + os.remove(temp_rotated_path) + else: + pdf.image(img_path, x=0, y=0, w=page_width, h=page_height) + + # Update progress + progress_msg = f"Processing page {i + 1}/{len(image_paths_and_rotations)} into PDF" + if rotation != 0: + progress_msg += f" (rotated {rotation}°)" + self.socket.progress(message=progress_msg) + + except Exception as e: + logger.warning(f"Failed to add image {img_path} to PDF: {e}") + continue + + # Save the PDF + pdf.output(output_path) + + # Get target file path + def _get_target_path(self, /) -> str: + """Get the full path where the PDF should be saved""" + folder = current_app.config['INSTRUCTIONS_FOLDER'] + + # If folder is absolute, use it directly + # Otherwise, make it relative to app root (not static folder) + if os.path.isabs(folder): + instructions_folder = folder + else: + instructions_folder = os.path.join(current_app.root_path, folder) + + return os.path.join(instructions_folder, self.filename) + + # Create BrickInstructions instance for the generated PDF + def get_instructions(self, /) -> BrickInstructions: + """Return a BrickInstructions instance for the generated PDF""" + return BrickInstructions(self.filename) \ No newline at end of file diff --git a/bricktracker/rebrickable_image.py b/bricktracker/rebrickable_image.py index 509e718..eed51d0 100644 --- a/bricktracker/rebrickable_image.py +++ b/bricktracker/rebrickable_image.py @@ -53,6 +53,23 @@ class RebrickableImage(object): if os.path.exists(path): return + # Check if the original image field is null - copy nil placeholder instead + if self.part is not None and self.part.fields.image is None: + return + if self.minifigure is not None and self.minifigure.fields.image is None: + return + if self.set.fields.image is None: + # Copy nil.png from parts folder to sets folder with set number as filename + parts_folder = current_app.config['PARTS_FOLDER'] + if not os.path.isabs(parts_folder): + parts_folder = os.path.join(current_app.root_path, parts_folder) + nil_source = os.path.join(parts_folder, f"{RebrickableImage.nil_name()}.{self.extension}") + + if os.path.exists(nil_source): + import shutil + shutil.copy2(nil_source, path) + return + url = self.url() if url is None: return @@ -96,9 +113,16 @@ class RebrickableImage(object): # Return the path depending on the objects provided def path(self, /) -> str: + folder = self.folder() + # If folder is an absolute path (starts with /), use it directly + # Otherwise, make it relative to app root (current_app.root_path) + if folder.startswith('/'): + base_path = folder + else: + base_path = os.path.join(current_app.root_path, folder) + return os.path.join( - current_app.static_folder, # type: ignore - self.folder(), + base_path, '{id}.{ext}'.format(id=self.id(), ext=self.extension), ) @@ -116,7 +140,11 @@ class RebrickableImage(object): else: return self.minifigure.fields.image - return self.set.fields.image + # Handle set images - use nil placeholder if image is null + if self.set.fields.image is None: + return current_app.config['REBRICKABLE_IMAGE_NIL'] + else: + return self.set.fields.image # Return the name of the nil image file @staticmethod @@ -152,10 +180,21 @@ class RebrickableImage(object): # _, extension = os.path.splitext(self.part_img_url) extension = '.jpg' - # Compute the path - path = os.path.join(folder, '{name}{ext}'.format( - name=name, - ext=extension, - )) - - return url_for('static', filename=path) + # Determine which route to use based on folder path + # If folder contains 'data' (new structure), use data route + # Otherwise use static route (legacy - relative paths like 'parts', 'sets') + if 'data' in folder: + # Extract the folder type from the folder_name config key + # E.g., 'PARTS_FOLDER' -> 'parts', 'SETS_FOLDER' -> 'sets' + folder_type = folder_name.replace('_FOLDER', '').lower() + filename = '{name}{ext}'.format(name=name, ext=extension) + return url_for('data.serve_data_file', folder=folder_type, filename=filename) + else: + # Legacy: folder is relative to static/ (e.g., 'parts' or 'static/parts') + # Strip 'static/' prefix if present to avoid double /static/ in URL + folder_clean = folder.removeprefix('static/') + path = os.path.join(folder_clean, '{name}{ext}'.format( + name=name, + ext=extension, + )) + return url_for('static', filename=path) diff --git a/bricktracker/rebrickable_part.py b/bricktracker/rebrickable_part.py index 91edd7b..09a3761 100644 --- a/bricktracker/rebrickable_part.py +++ b/bricktracker/rebrickable_part.py @@ -98,7 +98,7 @@ class RebrickablePart(BrickRecord): # Use BrickLink color ID if available and not None, otherwise fall back to Rebrickable color bricklink_color = getattr(self.fields, 'bricklink_color_id', None) color_param = bricklink_color if bricklink_color is not None else self.fields.color - print(f'BrickLink URL parameters: part={part_param}, color={color_param}') # Debugging line, can be removed later + # print(f'BrickLink URL parameters: part={part_param}, color={color_param}') # Debugging line, can be removed later return current_app.config['BRICKLINK_LINK_PART_PATTERN'].format( # noqa: E501 part=part_param, color=color_param, diff --git a/bricktracker/rebrickable_set.py b/bricktracker/rebrickable_set.py index 6beffc2..f98785f 100644 --- a/bricktracker/rebrickable_set.py +++ b/bricktracker/rebrickable_set.py @@ -155,9 +155,18 @@ class RebrickableSet(BrickRecord): # Return a short form of the Rebrickable set def short(self, /, *, from_download: bool = False) -> dict[str, Any]: + # Use nil image URL if set image is null + image_url = self.fields.image + if image_url is None: + # Return path to nil.png from parts folder + image_url = RebrickableImage.static_url( + RebrickableImage.nil_name(), + 'PARTS_FOLDER' + ) + return { 'download': from_download, - 'image': self.fields.image, + 'image': image_url, 'name': self.fields.name, 'set': self.fields.set, } @@ -179,6 +188,15 @@ class RebrickableSet(BrickRecord): return '' + # Compute the url for the bricklink page + def url_for_bricklink(self, /) -> str: + if current_app.config['BRICKLINK_LINKS']: + return current_app.config['BRICKLINK_LINK_SET_PATTERN'].format( + set_num=self.fields.set + ) + + return '' + # Compute the url for the refresh button def url_for_refresh(self, /) -> str: return url_for('set.refresh', set=self.fields.set) @@ -187,17 +205,18 @@ class RebrickableSet(BrickRecord): @staticmethod def from_rebrickable(data: dict[str, Any], /, **_) -> dict[str, Any]: # Extracting version and number + # Note: number can be alphanumeric (e.g., "McDR6US", "COMCON035") number, _, version = str(data['set_num']).partition('-') return { 'set': str(data['set_num']), - 'number': int(number), + 'number': str(number), # Keep as string to support alphanumeric sets 'version': int(version), 'name': str(data['name']), 'year': int(data['year']), 'theme_id': int(data['theme_id']), 'number_of_parts': int(data['num_parts']), - 'image': str(data['set_img_url']), + 'image': str(data['set_img_url']) if data['set_img_url'] is not None else None, 'url': str(data['set_url']), 'last_modified': str(data['last_modified_dt']), } diff --git a/bricktracker/rebrickable_set_list.py b/bricktracker/rebrickable_set_list.py index 0db84b7..6eabd5f 100644 --- a/bricktracker/rebrickable_set_list.py +++ b/bricktracker/rebrickable_set_list.py @@ -11,24 +11,19 @@ class RebrickableSetList(BrickRecordList[RebrickableSet]): select_query: str = 'rebrickable/set/list' refresh_query: str = 'rebrickable/set/need_refresh' - # All the sets - def all(self, /) -> Self: + # Implementation of abstract list method + def list(self, /, *, override_query: str | None = None, **context) -> None: # Load the sets from the database - for record in self.select(): + for record in self.select(override_query=override_query, **context): rebrickable_set = RebrickableSet(record=record) - self.records.append(rebrickable_set) + # All the sets + def all(self, /) -> Self: + self.list() return self # Sets needing refresh def need_refresh(self, /) -> Self: - # Load the sets from the database - for record in self.select( - override_query=self.refresh_query - ): - rebrickable_set = RebrickableSet(record=record) - - self.records.append(rebrickable_set) - + self.list(override_query=self.refresh_query) return self diff --git a/bricktracker/record_list.py b/bricktracker/record_list.py index 8d862ed..cc44877 100644 --- a/bricktracker/record_list.py +++ b/bricktracker/record_list.py @@ -1,5 +1,6 @@ +import re from sqlite3 import Row -from typing import Any, Generator, Generic, ItemsView, TypeVar, TYPE_CHECKING +from typing import Any, Generator, Generic, ItemsView, Self, TypeVar, TYPE_CHECKING from .fields import BrickRecordFields from .sql import BrickSQL @@ -72,6 +73,90 @@ class BrickRecordList(Generic[T]): **context ) + # Generic pagination method for all record lists + def paginate( + self, + page: int = 1, + per_page: int = 50, + sort_field: str | None = None, + sort_order: str = 'asc', + count_query: str | None = None, + list_query: str | None = None, + field_mapping: dict[str, str] | None = None, + **filter_context: Any + ) -> tuple['Self', int]: + """Generic pagination implementation for all record lists""" + from .sql import BrickSQL + + # Use provided queries or fall back to defaults + list_query = list_query or getattr(self, 'all_query', None) + if not list_query: + raise NotImplementedError("Subclass must define all_query") + + # Calculate offset + offset = (page - 1) * per_page + + # Get total count by wrapping the main query + if count_query: + # Use provided count query + count_result = BrickSQL().fetchone(count_query, **filter_context) + total_count = count_result['total_count'] if count_result else 0 + else: + # Generate count by wrapping the main query (without ORDER BY, LIMIT, OFFSET) + count_context = {k: v for k, v in filter_context.items() + if k not in ['order', 'limit', 'offset']} + + # Get the main query SQL without pagination clauses + main_sql = BrickSQL().load_query(list_query, **count_context) + + # Remove ORDER BY, LIMIT, OFFSET clauses for counting + # Remove ORDER BY clause and everything after it that's not part of subqueries + count_sql = re.sub(r'\s+ORDER\s+BY\s+[^)]*?(\s+LIMIT|\s+OFFSET|$)', r'\1', main_sql, flags=re.IGNORECASE) + # Remove LIMIT and OFFSET + count_sql = re.sub(r'\s+LIMIT\s+\d+', '', count_sql, flags=re.IGNORECASE) + count_sql = re.sub(r'\s+OFFSET\s+\d+', '', count_sql, flags=re.IGNORECASE) + + # Wrap in COUNT(*) + wrapped_sql = f"SELECT COUNT(*) as total_count FROM ({count_sql.strip()})" + + count_result = BrickSQL().raw_execute(wrapped_sql, {}).fetchone() + total_count = count_result['total_count'] if count_result else 0 + + # Prepare sort order + order_clause = None + if sort_field and field_mapping and sort_field in field_mapping: + sql_field = field_mapping[sort_field] + direction = 'DESC' if sort_order.lower() == 'desc' else 'ASC' + order_clause = f'{sql_field} {direction}' + + # Build pagination context + pagination_context = { + 'limit': per_page, + 'offset': offset, + 'order': order_clause or getattr(self, 'order', None), + **filter_context + } + + # Load paginated results using the existing list() method + # Check if this is a set list that needs do_theme parameter + if hasattr(self, 'themes'): # Only BrickSetList has this attribute + self.list(override_query=list_query, do_theme=True, **pagination_context) + else: + self.list(override_query=list_query, **pagination_context) + + return self, total_count + + # Base method that subclasses can override + def list( + self, + /, + *, + override_query: str | None = None, + **context: Any, + ) -> None: + """Load records from database - should be implemented by subclasses that use pagination""" + raise NotImplementedError("Subclass must implement list() method") + # Generic SQL parameters from fields def sql_parameters(self, /) -> dict[str, Any]: parameters: dict[str, Any] = {} diff --git a/bricktracker/set.py b/bricktracker/set.py index c397b13..cb67620 100644 --- a/bricktracker/set.py +++ b/bricktracker/set.py @@ -59,6 +59,10 @@ class BrickSet(RebrickableSet): # Generate an UUID for self self.fields.id = str(uuid4()) + # Insert the rebrickable set into database FIRST + # This must happen before inserting bricktracker_sets due to FK constraint + self.insert_rebrickable() + if not refresh: # Save the storage storage = BrickSetStorageList.get( @@ -74,7 +78,8 @@ class BrickSet(RebrickableSet): ) self.fields.purchase_location = purchase_location.fields.id - # Insert into database + # Insert into database (deferred - will execute at final commit) + # All operations are atomic - if anything fails, nothing is committed self.insert(commit=False) # Save the owners @@ -91,9 +96,6 @@ class BrickSet(RebrickableSet): tag = BrickSetTagList.get(id) tag.update_set_state(self, state=True) - # Insert the rebrickable set into database - self.insert_rebrickable() - # Load the inventory if not BrickPartList.download(socket, self, refresh=refresh): return False @@ -169,6 +171,20 @@ class BrickSet(RebrickableSet): else: return '' + # Purchase date max formatted for consolidated sets + def purchase_date_max_formatted(self, /, *, standard: bool = False) -> str: + if hasattr(self.fields, 'purchase_date_max') and self.fields.purchase_date_max is not None: + time = datetime.fromtimestamp(self.fields.purchase_date_max) + + if standard: + return time.strftime('%Y/%m/%d') + else: + return time.strftime( + current_app.config['PURCHASE_DATE_FORMAT'] + ) + else: + return '' + # Purchase price with currency def purchase_price(self, /) -> str: if self.fields.purchase_price is not None: diff --git a/bricktracker/set_list.py b/bricktracker/set_list.py index a8f5faa..77cb7ce 100644 --- a/bricktracker/set_list.py +++ b/bricktracker/set_list.py @@ -13,14 +13,19 @@ from .set_storage_list import BrickSetStorageList from .set_tag import BrickSetTag from .set_tag_list import BrickSetTagList from .set import BrickSet +from .theme_list import BrickThemeList +from .instructions_list import BrickInstructionsList # All the sets from the database class BrickSetList(BrickRecordList[BrickSet]): themes: list[str] + years: list[int] order: str # Queries + all_query: str = 'set/list/all' + consolidated_query: str = 'set/list/consolidated' damaged_minifigure_query: str = 'set/list/damaged_minifigure' damaged_part_query: str = 'set/list/damaged_part' generic_query: str = 'set/list/generic' @@ -37,17 +42,510 @@ class BrickSetList(BrickRecordList[BrickSet]): # Placeholders self.themes = [] + self.years = [] # Store the order for this list self.order = current_app.config['SETS_DEFAULT_ORDER'] # All the sets def all(self, /) -> Self: - # Load the sets from the database - self.list(do_theme=True) + # Load the sets from the database with metadata context for filtering + filter_context = { + 'owners': BrickSetOwnerList.as_columns(), + 'statuses': BrickSetStatusList.as_columns(), + 'tags': BrickSetTagList.as_columns(), + } + self.list(do_theme=True, **filter_context) return self + # All sets in consolidated/grouped view + def all_consolidated(self, /) -> Self: + # Load the sets from the database using consolidated query with metadata context + filter_context = { + 'owners_dict': BrickSetOwnerList.as_column_mapping(), + 'statuses_dict': BrickSetStatusList.as_column_mapping(), + 'tags_dict': BrickSetTagList.as_column_mapping(), + } + self.list(override_query=self.consolidated_query, do_theme=True, **filter_context) + + return self + + # All sets with pagination and filtering + def all_filtered_paginated( + self, + search_query: str | None = None, + page: int = 1, + per_page: int = 50, + sort_field: str | None = None, + sort_order: str = 'asc', + status_filter: str | None = None, + theme_filter: str | None = None, + owner_filter: str | None = None, + purchase_location_filter: str | None = None, + storage_filter: str | None = None, + tag_filter: str | None = None, + year_filter: str | None = None, + duplicate_filter: bool = False, + use_consolidated: bool = True + ) -> tuple[Self, int]: + # Convert theme name to theme ID for filtering + theme_id_filter = None + if theme_filter: + theme_id_filter = self._theme_name_to_id(theme_filter) + + # Check if any filters are applied + has_filters = any([status_filter, theme_id_filter, owner_filter, purchase_location_filter, storage_filter, tag_filter, year_filter, duplicate_filter]) + + # Prepare filter context + filter_context = { + 'search_query': search_query, + 'status_filter': status_filter, + 'theme_filter': theme_id_filter, # Use converted theme ID + 'owner_filter': owner_filter, + 'purchase_location_filter': purchase_location_filter, + 'storage_filter': storage_filter, + 'tag_filter': tag_filter, + 'year_filter': year_filter, + 'duplicate_filter': duplicate_filter, + 'owners': BrickSetOwnerList.as_columns(), + 'statuses': BrickSetStatusList.as_columns(), + 'tags': BrickSetTagList.as_columns(), + 'owners_dict': BrickSetOwnerList.as_column_mapping(), + 'statuses_dict': BrickSetStatusList.as_column_mapping(), + 'tags_dict': BrickSetTagList.as_column_mapping(), + } + + + + # Field mapping for sorting + if use_consolidated: + field_mapping = { + 'set': '"rebrickable_sets"."number", "rebrickable_sets"."version"', + 'name': '"rebrickable_sets"."name"', + 'year': '"rebrickable_sets"."year"', + 'parts': '"rebrickable_sets"."number_of_parts"', + 'theme': '"rebrickable_sets"."theme_id"', + 'minifigures': '"total_minifigures"', + 'missing': '"total_missing"', + 'damaged': '"total_damaged"', + 'instances': '"instance_count"', # New field for consolidated view + 'purchase-date': '"purchase_date"', # Use the MIN aggregated value + 'purchase-price': '"purchase_price"' # Use the MIN aggregated value + } + else: + field_mapping = { + 'set': '"rebrickable_sets"."number", "rebrickable_sets"."version"', + 'name': '"rebrickable_sets"."name"', + 'year': '"rebrickable_sets"."year"', + 'parts': '"rebrickable_sets"."number_of_parts"', + 'theme': '"rebrickable_sets"."theme_id"', + 'minifigures': '"total_minifigures"', # Use the alias from the SQL query + 'missing': '"total_missing"', # Use the alias from the SQL query + 'damaged': '"total_damaged"', # Use the alias from the SQL query + 'purchase-date': '"bricktracker_sets"."purchase_date"', + 'purchase-price': '"bricktracker_sets"."purchase_price"' + } + + # Choose query based on consolidation preference and filter complexity + # Owner/tag filters still need to fall back to non-consolidated for now + # due to complex aggregation requirements + complex_filters = [owner_filter, tag_filter] + if use_consolidated and not any(complex_filters): + query_to_use = self.consolidated_query + else: + # Use filtered query when consolidation is disabled or complex filters applied + query_to_use = 'set/list/all_filtered' + + # Handle instructions filtering + if status_filter in ['has-missing-instructions', '-has-missing-instructions']: + # For instructions filter, we need to load all sets first, then filter and paginate + return self._all_filtered_paginated_with_instructions( + search_query, page, per_page, sort_field, sort_order, + status_filter, theme_id_filter, owner_filter, + purchase_location_filter, storage_filter, tag_filter + ) + + # Handle special case for set sorting with multiple columns + if sort_field == 'set' and field_mapping: + # Create custom order clause for set sorting + direction = 'DESC' if sort_order.lower() == 'desc' else 'ASC' + custom_order = f'"rebrickable_sets"."number" {direction}, "rebrickable_sets"."version" {direction}' + filter_context['order'] = custom_order + # Remove set from field mapping to avoid double-processing + field_mapping_copy = field_mapping.copy() + field_mapping_copy.pop('set', None) + field_mapping = field_mapping_copy + sort_field = None # Disable automatic ORDER BY construction + + # Normal SQL-based filtering and pagination + result, total_count = self.paginate( + page=page, + per_page=per_page, + sort_field=sort_field, + sort_order=sort_order, + list_query=query_to_use, + field_mapping=field_mapping, + **filter_context + ) + + # Populate themes and years for filter dropdown from filtered dataset (not just current page) + # For themes dropdown, exclude theme_filter to show ALL available themes + themes_context = filter_context.copy() + themes_context.pop('theme_filter', None) + result._populate_themes_from_filtered_dataset( + query_to_use, + **themes_context + ) + # For years dropdown, exclude ALL filters to show ALL available years + years_context = { + 'search_query': filter_context.get('search_query'), + } + result._populate_years_from_filtered_dataset( + query_to_use, + **years_context + ) + + return result, total_count + + def _populate_themes(self) -> None: + """Populate themes list from the current records""" + themes = set() + for record in self.records: + if hasattr(record, 'theme') and hasattr(record.theme, 'name'): + themes.add(record.theme.name) + + self.themes = list(themes) + self.themes.sort() + + def _populate_years(self) -> None: + """Populate years list from the current records""" + years = set() + for record in self.records: + if hasattr(record, 'fields') and hasattr(record.fields, 'year') and record.fields.year: + years.add(record.fields.year) + + self.years = list(years) + self.years.sort(reverse=True) # Most recent years first + + def _theme_name_to_id(self, theme_name_or_id: str) -> str | None: + """Convert a theme name or ID to theme ID for filtering""" + try: + # Check if the input is already a numeric theme ID + if theme_name_or_id.isdigit(): + # Input is already a theme ID, validate it exists + theme_list = BrickThemeList() + theme_id = int(theme_name_or_id) + if theme_id in theme_list.themes: + return str(theme_id) + else: + return None + + # Input is a theme name, convert to ID + from .sql import BrickSQL + theme_list = BrickThemeList() + + # Find all theme IDs that match the name + matching_theme_ids = [] + for theme_id, theme in theme_list.themes.items(): + if theme.name.lower() == theme_name_or_id.lower(): + matching_theme_ids.append(str(theme_id)) + + if not matching_theme_ids: + return None + + # If only one match, return it + if len(matching_theme_ids) == 1: + return matching_theme_ids[0] + + # Multiple matches - check which theme ID actually has sets in the user's collection + sql = BrickSQL() + for theme_id in matching_theme_ids: + result = sql.fetchone( + 'set/check_theme_exists', + theme_id=theme_id + ) + count = result['count'] if result else 0 + if count > 0: + return theme_id + + # If none have sets, return the first match (fallback) + return matching_theme_ids[0] + + except Exception: + # If themes can't be loaded, return None to disable theme filtering + return None + + def _theme_id_to_name(self, theme_id: str) -> str | None: + """Convert a theme ID to theme name (lowercase) for dropdown display""" + try: + if not theme_id or not theme_id.isdigit(): + return None + + from .theme_list import BrickThemeList + theme_list = BrickThemeList() + theme_id_int = int(theme_id) + + if theme_id_int in theme_list.themes: + return theme_list.themes[theme_id_int].name.lower() + + return None + except Exception as e: + # For debugging - log the exception + import logging + logger = logging.getLogger(__name__) + logger.warning(f"Failed to convert theme ID {theme_id} to name: {e}") + return None + + def _all_filtered_paginated_with_instructions( + self, + search_query: str | None, + page: int, + per_page: int, + sort_field: str | None, + sort_order: str, + status_filter: str, + theme_id_filter: str | None, + owner_filter: str | None, + purchase_location_filter: str | None, + storage_filter: str | None, + tag_filter: str | None + ) -> tuple[Self, int]: + """Handle filtering when instructions filter is involved""" + try: + # Load all sets first (without pagination) with full metadata + all_sets = BrickSetList() + filter_context = { + 'owners': BrickSetOwnerList.as_columns(), + 'statuses': BrickSetStatusList.as_columns(), + 'tags': BrickSetTagList.as_columns(), + } + all_sets.list(do_theme=True, **filter_context) + + # Load instructions list + instructions_list = BrickInstructionsList() + instruction_sets = set(instructions_list.sets.keys()) + + # Apply all filters manually + filtered_records = [] + for record in all_sets.records: + # Apply instructions filter + set_id = record.fields.set + has_instructions = set_id in instruction_sets + + if status_filter == 'has-missing-instructions' and has_instructions: + continue # Skip sets that have instructions + elif status_filter == '-has-missing-instructions' and not has_instructions: + continue # Skip sets that don't have instructions + + # Apply other filters manually + if search_query and not self._matches_search(record, search_query): + continue + if theme_id_filter and not self._matches_theme(record, theme_id_filter): + continue + if owner_filter and not self._matches_owner(record, owner_filter): + continue + if purchase_location_filter and not self._matches_purchase_location(record, purchase_location_filter): + continue + if storage_filter and not self._matches_storage(record, storage_filter): + continue + if tag_filter and not self._matches_tag(record, tag_filter): + continue + + filtered_records.append(record) + + # Apply sorting + if sort_field: + filtered_records = self._sort_records(filtered_records, sort_field, sort_order) + + # Calculate pagination + total_count = len(filtered_records) + start_index = (page - 1) * per_page + end_index = start_index + per_page + paginated_records = filtered_records[start_index:end_index] + + # Create result + result = BrickSetList() + result.records = paginated_records + + # Copy themes and years from the source that has all sets + result.themes = all_sets.themes if hasattr(all_sets, 'themes') else [] + result.years = all_sets.years if hasattr(all_sets, 'years') else [] + + # If themes or years weren't populated, populate them from current records + if not result.themes: + result._populate_themes() + if not result.years: + result._populate_years() + + return result, total_count + + except Exception: + # Fall back to normal pagination without instructions filter + return self.all_filtered_paginated( + search_query, page, per_page, sort_field, sort_order, + None, theme_id_filter, owner_filter, + purchase_location_filter, storage_filter, tag_filter + ) + + def _populate_years_from_filtered_dataset(self, query_name: str, **filter_context) -> None: + """Populate years list from all available records in filtered dataset""" + try: + # Use a simplified query to get just distinct years + years_context = dict(filter_context) + years_context.pop('limit', None) + years_context.pop('offset', None) + + # Use a special lightweight query for years + year_records = super().select( + override_query='set/list/years_only', + **years_context + ) + + # Extract years from records + years = set() + for record in year_records: + year = record['year'] if 'year' in record.keys() else None + if year: + years.add(year) + + if years: + self.years = list(years) + self.years.sort(reverse=True) # Most recent years first + else: + import logging + logger = logging.getLogger(__name__) + logger.warning("No years found in filtered dataset, falling back to current page") + self._populate_years() + except Exception as e: + import logging + logger = logging.getLogger(__name__) + logger.error(f"Exception in _populate_years_from_filtered_dataset: {e}") + self._populate_years() + + def _populate_themes_from_filtered_dataset(self, query_name: str, **filter_context) -> None: + """Populate themes list from filtered dataset (all pages, not just current page)""" + try: + from .theme_list import BrickThemeList + + # Use a simplified query to get just distinct theme_ids + theme_context = dict(filter_context) + theme_context.pop('limit', None) + theme_context.pop('offset', None) + + # Use a special lightweight query for themes + theme_records = super().select( + override_query='set/list/themes_only', + **theme_context + ) + + # Convert to theme names + theme_list = BrickThemeList() + themes = set() + for record in theme_records: + theme_id = record.get('theme_id') + if theme_id: + theme = theme_list.get(theme_id) + if theme and hasattr(theme, 'name'): + themes.add(theme.name) + + self.themes = list(themes) + self.themes.sort() + + except Exception: + # Fall back to simpler approach: get themes from ALL sets (ignoring filters) + # This is better than showing only current page themes + try: + from .theme_list import BrickThemeList + all_sets = BrickSetList() + all_sets.list(do_theme=True) + + themes = set() + years = set() + for record in all_sets.records: + if hasattr(record, 'theme') and hasattr(record.theme, 'name'): + themes.add(record.theme.name) + if hasattr(record, 'fields') and hasattr(record.fields, 'year') and record.fields.year: + years.add(record.fields.year) + + self.themes = list(themes) + self.themes.sort() + self.years = list(years) + self.years.sort(reverse=True) + except Exception: + # Final fallback to current page themes + self._populate_themes() + self._populate_years() + + def _matches_search(self, record, search_query: str) -> bool: + """Check if record matches search query""" + search_lower = search_query.lower() + return (search_lower in record.fields.name.lower() or + search_lower in record.fields.set.lower()) + + def _matches_theme(self, record, theme_id: str) -> bool: + """Check if record matches theme filter""" + return str(record.fields.theme_id) == theme_id + + def _matches_owner(self, record, owner_filter: str) -> bool: + """Check if record matches owner filter""" + if not owner_filter.startswith('owner-'): + return True + + # Convert owner-uuid format to owner_uuid column name + owner_column = owner_filter.replace('-', '_') + + # Check if record has this owner attribute set to 1 + return hasattr(record.fields, owner_column) and getattr(record.fields, owner_column) == 1 + + def _matches_purchase_location(self, record, location_filter: str) -> bool: + """Check if record matches purchase location filter""" + return record.fields.purchase_location == location_filter + + def _matches_storage(self, record, storage_filter: str) -> bool: + """Check if record matches storage filter""" + return record.fields.storage == storage_filter + + def _matches_tag(self, record, tag_filter: str) -> bool: + """Check if record matches tag filter""" + if not tag_filter.startswith('tag-'): + return True + + # Convert tag-uuid format to tag_uuid column name + tag_column = tag_filter.replace('-', '_') + + # Check if record has this tag attribute set to 1 + return hasattr(record.fields, tag_column) and getattr(record.fields, tag_column) == 1 + + def _sort_records(self, records, sort_field: str, sort_order: str): + """Sort records manually""" + reverse = sort_order == 'desc' + + if sort_field == 'set': + return sorted(records, key=lambda r: self._set_sort_key(r.fields.set), reverse=reverse) + elif sort_field == 'name': + return sorted(records, key=lambda r: r.fields.name, reverse=reverse) + elif sort_field == 'year': + return sorted(records, key=lambda r: r.fields.year, reverse=reverse) + elif sort_field == 'parts': + return sorted(records, key=lambda r: r.fields.number_of_parts, reverse=reverse) + # Add more sort fields as needed + + return records + + + def _set_sort_key(self, set_number: str) -> tuple: + """Generate sort key for set numbers like '10121-1' -> (10121, 1)""" + try: + if '-' in set_number: + main_part, version_part = set_number.split('-', 1) + return (int(main_part), int(version_part)) + else: + return (int(set_number), 0) + except (ValueError, TypeError): + # Fallback to string sorting if parsing fails + return (float('inf'), set_number) + # Sets with a minifigure part damaged def damaged_minifigure(self, figure: str, /) -> Self: # Save the parameters to the fields @@ -93,6 +591,7 @@ class BrickSetList(BrickRecordList[BrickSet]): **context: Any, ) -> None: themes = set() + years = set() if order is None: order = self.order @@ -102,20 +601,22 @@ class BrickSetList(BrickRecordList[BrickSet]): override_query=override_query, order=order, limit=limit, - owners=BrickSetOwnerList.as_columns(), - statuses=BrickSetStatusList.as_columns(), - tags=BrickSetTagList.as_columns(), + **context ): brickset = BrickSet(record=record) self.records.append(brickset) if do_theme: themes.add(brickset.theme.name) + if hasattr(brickset, 'fields') and hasattr(brickset.fields, 'year') and brickset.fields.year: + years.add(brickset.fields.year) # Convert the set into a list and sort it if do_theme: self.themes = list(themes) self.themes.sort() + self.years = list(years) + self.years.sort(reverse=True) # Most recent years first # Sets missing a minifigure part def missing_minifigure(self, figure: str, /) -> Self: diff --git a/bricktracker/socket.py b/bricktracker/socket.py index 10c3358..86b8b5d 100644 --- a/bricktracker/socket.py +++ b/bricktracker/socket.py @@ -6,6 +6,8 @@ from flask_socketio import SocketIO from .instructions import BrickInstructions from .instructions_list import BrickInstructionsList +from .peeron_instructions import PeeronInstructions, PeeronPage +from .peeron_pdf import PeeronPDF from .set import BrickSet from .socket_decorator import authenticated_socket, rebrickable_socket from .sql import close as sql_close @@ -18,8 +20,10 @@ MESSAGES: Final[dict[str, str]] = { 'CONNECT': 'connect', 'DISCONNECT': 'disconnect', 'DOWNLOAD_INSTRUCTIONS': 'download_instructions', + 'DOWNLOAD_PEERON_PAGES': 'download_peeron_pages', 'FAIL': 'fail', 'IMPORT_SET': 'import_set', + 'LOAD_PEERON_PAGES': 'load_peeron_pages', 'LOAD_SET': 'load_set', 'PROGRESS': 'progress', 'SET_LOADED': 'set_loaded', @@ -71,6 +75,9 @@ class BrickSocket(object): **kwargs, path=app.config['SOCKET_PATH'], async_mode='gevent', + # Ping/pong settings for mobile network resilience + ping_timeout=30, # Wait 30s for pong response before disconnecting + ping_interval=25, # Send ping every 25s to keep connection alive ) # Store the socket in the app config @@ -82,9 +89,23 @@ class BrickSocket(object): self.connected() @self.socket.on(MESSAGES['DISCONNECT'], namespace=self.namespace) - def disconnect() -> None: + def disconnect(reason=None) -> None: self.disconnected() + @self.socket.on('connect_error', namespace=self.namespace) + def connect_error(data) -> None: + logger.error(f'Socket CONNECT_ERROR: {data}') + + @self.socket.on_error(namespace=self.namespace) + def error_handler(e) -> None: + logger.error(f'Socket ERROR: {e}') + try: + user_agent = request.headers.get('User-Agent', 'unknown') + remote_addr = request.remote_addr + logger.error(f'Socket ERROR details: ip={remote_addr}, ua={user_agent[:80]}...') + except Exception: + pass + @self.socket.on(MESSAGES['DOWNLOAD_INSTRUCTIONS'], namespace=self.namespace) # noqa: E501 @authenticated_socket(self) def download_instructions(data: dict[str, Any], /) -> None: @@ -106,6 +127,84 @@ class BrickSocket(object): BrickInstructionsList(force=True) + @self.socket.on(MESSAGES['LOAD_PEERON_PAGES'], namespace=self.namespace) # noqa: E501 + def load_peeron_pages(data: dict[str, Any], /) -> None: + logger.debug('Socket: LOAD_PEERON_PAGES={data} (from: {fr})'.format( + data=data, fr=request.remote_addr)) + + try: + set_number = data.get('set', '') + if not set_number: + self.fail(message="Set number is required") + return + + # Create Peeron instructions instance with socket for progress reporting + peeron = PeeronInstructions(set_number, socket=self) + + # Find pages (this will report progress for thumbnail caching) + pages = peeron.find_pages() + + # Complete the operation (JavaScript will handle redirect) + self.complete(message=f"Found {len(pages)} instruction pages on Peeron") + + except Exception as e: + logger.error(f"Error in load_peeron_pages: {e}") + self.fail(message=f"Error loading Peeron pages: {e}") + + @self.socket.on(MESSAGES['DOWNLOAD_PEERON_PAGES'], namespace=self.namespace) # noqa: E501 + @authenticated_socket(self) + def download_peeron_pages(data: dict[str, Any], /) -> None: + logger.debug('Socket: DOWNLOAD_PEERON_PAGES={data} (from: {fr})'.format( + data=data, + fr=request.sid, # type: ignore + )) + + try: + # Extract data from the request + set_number = data.get('set', '') + pages_data = data.get('pages', []) + + if not set_number: + raise ValueError("Set number is required") + + if not pages_data: + raise ValueError("No pages selected") + + # Parse set number + if '-' in set_number: + parts = set_number.split('-', 1) + set_num = parts[0] + version_num = parts[1] if len(parts) > 1 else '1' + else: + set_num = set_number + version_num = '1' + + # Convert page data to PeeronPage objects + pages = [] + for page_data in pages_data: + page = PeeronPage( + page_number=page_data.get('page_number', ''), + original_image_url=page_data.get('original_image_url', ''), + cached_full_image_path=page_data.get('cached_full_image_path', ''), + cached_thumbnail_url='', # Not needed for PDF generation + alt_text=page_data.get('alt_text', ''), + rotation=page_data.get('rotation', 0) + ) + pages.append(page) + + # Create PDF generator and start download + pdf_generator = PeeronPDF(set_num, version_num, pages, socket=self) + pdf_generator.create_pdf() + + # Note: Cache cleanup is handled automatically by pdf_generator.create_pdf() + + # Refresh instructions list to include new PDF + BrickInstructionsList(force=True) + + except Exception as e: + logger.error(f"Error in download_peeron_pages: {e}") + self.fail(message=f"Error downloading Peeron pages: {e}") + @self.socket.on(MESSAGES['IMPORT_SET'], namespace=self.namespace) @rebrickable_socket(self) def import_set(data: dict[str, Any], /) -> None: @@ -150,13 +249,32 @@ class BrickSocket(object): # Socket is connected def connected(self, /) -> Tuple[str, int]: - logger.debug('Socket: client connected') + # Get detailed connection info for debugging + try: + sid = request.sid # type: ignore + transport = request.environ.get('HTTP_UPGRADE', 'polling') + user_agent = request.headers.get('User-Agent', 'unknown') + remote_addr = request.remote_addr + + # Check if it's likely a mobile device + is_mobile = any(x in user_agent.lower() for x in ['iphone', 'ipad', 'android', 'mobile']) + + logger.info( + f'Socket CONNECTED: sid={sid}, transport={transport}, ' + f'ip={remote_addr}, mobile={is_mobile}, ua={user_agent[:80]}...' + ) + except Exception as e: + logger.warning(f'Socket connected but failed to get details: {e}') return '', 301 # Socket is disconnected def disconnected(self, /) -> None: - logger.debug('Socket: client disconnected') + try: + sid = request.sid # type: ignore + logger.info(f'Socket DISCONNECTED: sid={sid}') + except Exception as e: + logger.info(f'Socket disconnected (sid unavailable): {e}') # Emit a message through the socket def emit(self, name: str, *arg, all=False) -> None: diff --git a/bricktracker/sql.py b/bricktracker/sql.py index 18f7e99..fc02bd2 100644 --- a/bricktracker/sql.py +++ b/bricktracker/sql.py @@ -60,6 +60,29 @@ class BrickSQL(object): # Grab a cursor self.cursor = self.connection.cursor() + # SQLite Performance Optimizations + logger.debug('SQLite3: applying performance optimizations') + + # Enable WAL (Write-Ahead Logging) mode for better concurrency + # Allows multiple readers while writer is active + self.connection.execute('PRAGMA journal_mode=WAL') + + # Increase cache size for better query performance + # Default is 2000 pages, increase to 10000 pages (~40MB for 4KB pages) + self.connection.execute('PRAGMA cache_size=10000') + + # Store temporary tables and indices in memory for speed + self.connection.execute('PRAGMA temp_store=memory') + + # Enable foreign key constraints (good practice) + self.connection.execute('PRAGMA foreign_keys=ON') + + # Optimize for read performance (trade write speed for read speed) + self.connection.execute('PRAGMA synchronous=NORMAL') + + # Analyze database statistics for better query planning + self.connection.execute('ANALYZE') + # Grab the version and check try: version = self.fetchone('schema/get_version') diff --git a/bricktracker/sql/migrations/0018.sql b/bricktracker/sql/migrations/0018.sql new file mode 100644 index 0000000..9ff3d97 --- /dev/null +++ b/bricktracker/sql/migrations/0018.sql @@ -0,0 +1,9 @@ +-- description: Add checked field to bricktracker_parts table for part walkthrough tracking + +BEGIN TRANSACTION; + +-- Add checked field to the bricktracker_parts table +-- This allows users to track which parts they have checked during walkthroughs +ALTER TABLE "bricktracker_parts" ADD COLUMN "checked" BOOLEAN DEFAULT 0; + +COMMIT; \ No newline at end of file diff --git a/bricktracker/sql/migrations/0019.sql b/bricktracker/sql/migrations/0019.sql new file mode 100644 index 0000000..93220b0 --- /dev/null +++ b/bricktracker/sql/migrations/0019.sql @@ -0,0 +1,56 @@ +-- description: Performance optimization indexes + +-- High-impact composite index for problem parts aggregation +-- Used in set listings, statistics, and problem reports +CREATE INDEX IF NOT EXISTS idx_bricktracker_parts_id_missing_damaged +ON bricktracker_parts(id, missing, damaged); + +-- Composite index for parts lookup by part and color +-- Used in part listings and filtering operations +CREATE INDEX IF NOT EXISTS idx_bricktracker_parts_part_color_spare +ON bricktracker_parts(part, color, spare); + +-- Composite index for set storage filtering +-- Used in set listings filtered by storage location +CREATE INDEX IF NOT EXISTS idx_bricktracker_sets_set_storage +ON bricktracker_sets("set", storage); + +-- Search optimization index for set names +-- Improves text search performance on set listings +CREATE INDEX IF NOT EXISTS idx_rebrickable_sets_name_lower +ON rebrickable_sets(LOWER(name)); + +-- Search optimization index for part names +-- Improves text search performance on part listings +CREATE INDEX IF NOT EXISTS idx_rebrickable_parts_name_lower +ON rebrickable_parts(LOWER(name)); + +-- Additional indexes for common join patterns + +-- Set purchase filtering +CREATE INDEX IF NOT EXISTS idx_bricktracker_sets_purchase_location +ON bricktracker_sets(purchase_location); + +-- Parts quantity filtering +CREATE INDEX IF NOT EXISTS idx_bricktracker_parts_quantity +ON bricktracker_parts(quantity); + +-- Year-based filtering optimization +CREATE INDEX IF NOT EXISTS idx_rebrickable_sets_year +ON rebrickable_sets(year); + +-- Theme-based filtering optimization +CREATE INDEX IF NOT EXISTS idx_rebrickable_sets_theme_id +ON rebrickable_sets(theme_id); + +-- Rebrickable sets number and version for sorting +CREATE INDEX IF NOT EXISTS idx_rebrickable_sets_number_version +ON rebrickable_sets(number, version); + +-- Purchase date filtering and sorting +CREATE INDEX IF NOT EXISTS idx_bricktracker_sets_purchase_date +ON bricktracker_sets(purchase_date); + +-- Minifigures aggregation optimization +CREATE INDEX IF NOT EXISTS idx_bricktracker_minifigures_id_quantity +ON bricktracker_minifigures(id, quantity); \ No newline at end of file diff --git a/bricktracker/sql/migrations/0020.sql b/bricktracker/sql/migrations/0020.sql new file mode 100644 index 0000000..afd6d7d --- /dev/null +++ b/bricktracker/sql/migrations/0020.sql @@ -0,0 +1,58 @@ +-- description: Change set number column from INTEGER to TEXT to support alphanumeric set numbers + +-- Temporarily disable foreign key constraints for this migration +-- This is necessary because we're recreating a table that other tables reference +-- We verify integrity at the end to ensure safety +PRAGMA foreign_keys=OFF; + +BEGIN TRANSACTION; + +-- Create new table with TEXT number column +CREATE TABLE "rebrickable_sets_new" ( + "set" TEXT NOT NULL, + "number" TEXT NOT NULL, + "version" INTEGER NOT NULL, + "name" TEXT NOT NULL, + "year" INTEGER NOT NULL, + "theme_id" INTEGER NOT NULL, + "number_of_parts" INTEGER NOT NULL, + "image" TEXT, + "url" TEXT, + "last_modified" TEXT, + PRIMARY KEY("set") +); + +-- Copy all data from old table to new table +-- Cast INTEGER number to TEXT explicitly +INSERT INTO "rebrickable_sets_new" +SELECT + "set", + CAST("number" AS TEXT), + "version", + "name", + "year", + "theme_id", + "number_of_parts", + "image", + "url", + "last_modified" +FROM "rebrickable_sets"; + +-- Drop old table +DROP TABLE "rebrickable_sets"; + +-- Rename new table to original name +ALTER TABLE "rebrickable_sets_new" RENAME TO "rebrickable_sets"; + +-- Recreate the index +CREATE INDEX IF NOT EXISTS idx_rebrickable_sets_number_version +ON rebrickable_sets(number, version); + +-- Verify foreign key integrity before committing +-- This ensures we haven't broken any references +PRAGMA foreign_key_check; + +COMMIT; + +-- Re-enable foreign key constraints +PRAGMA foreign_keys=ON; diff --git a/bricktracker/sql/minifigure/base/base.sql b/bricktracker/sql/minifigure/base/base.sql index a3a30a7..76b816e 100644 --- a/bricktracker/sql/minifigure/base/base.sql +++ b/bricktracker/sql/minifigure/base/base.sql @@ -28,6 +28,8 @@ ON "bricktracker_minifigures"."figure" IS NOT DISTINCT FROM "rebrickable_minifig {% block group %}{% endblock %} +{% block having %}{% endblock %} + {% if order %} ORDER BY {{ order }} {% endif %} @@ -35,3 +37,7 @@ ORDER BY {{ order }} {% if limit %} LIMIT {{ limit }} {% endif %} + +{% if offset %} +OFFSET {{ offset }} +{% endif %} diff --git a/bricktracker/sql/minifigure/list/all.sql b/bricktracker/sql/minifigure/list/all.sql index 904e818..ebfdffc 100644 --- a/bricktracker/sql/minifigure/list/all.sql +++ b/bricktracker/sql/minifigure/list/all.sql @@ -17,6 +17,14 @@ IFNULL(COUNT("bricktracker_minifigures"."id"), 0) AS "total_sets" {% endblock %} {% block join %} +{% if theme_id or year %} +-- Join with sets for theme/year filtering +INNER JOIN "bricktracker_sets" AS "filter_sets" +ON "bricktracker_minifigures"."id" IS NOT DISTINCT FROM "filter_sets"."id" +INNER JOIN "rebrickable_sets" AS "filter_rs" +ON "filter_sets"."set" IS NOT DISTINCT FROM "filter_rs"."set" +{% endif %} + -- LEFT JOIN + SELECT to avoid messing the total LEFT JOIN ( SELECT @@ -34,6 +42,32 @@ ON "bricktracker_minifigures"."id" IS NOT DISTINCT FROM "problem_join"."id" AND "rebrickable_minifigures"."figure" IS NOT DISTINCT FROM "problem_join"."figure" {% endblock %} +{% block where %} +WHERE 1=1 +{% if theme_id and theme_id != 'all' %} +AND "filter_rs"."theme_id" = {{ theme_id }} +{% endif %} +{% if year and year != 'all' %} +AND "filter_rs"."year" = {{ year }} +{% endif %} +{% if search_query %} +AND (LOWER("rebrickable_minifigures"."name") LIKE LOWER('%{{ search_query }}%')) +{% endif %} +{% endblock %} + +{% block having %} +{% if problems_filter %} +HAVING 1=1 +{% if problems_filter == 'missing' %} +AND SUM(IFNULL("problem_join"."total_missing", 0)) > 0 +{% elif problems_filter == 'damaged' %} +AND SUM(IFNULL("problem_join"."total_damaged", 0)) > 0 +{% elif problems_filter == 'both' %} +AND SUM(IFNULL("problem_join"."total_missing", 0)) > 0 AND SUM(IFNULL("problem_join"."total_damaged", 0)) > 0 +{% endif %} +{% endif %} +{% endblock %} + {% block group %} GROUP BY "rebrickable_minifigures"."figure" diff --git a/bricktracker/sql/minifigure/list/all_by_owner.sql b/bricktracker/sql/minifigure/list/all_by_owner.sql index 6ae14ba..e9a1434 100644 --- a/bricktracker/sql/minifigure/list/all_by_owner.sql +++ b/bricktracker/sql/minifigure/list/all_by_owner.sql @@ -29,6 +29,10 @@ COUNT("bricktracker_minifigures"."id") AS "total_sets" INNER JOIN "bricktracker_sets" ON "bricktracker_minifigures"."id" IS NOT DISTINCT FROM "bricktracker_sets"."id" +-- Join with rebrickable sets for theme/year filtering +INNER JOIN "rebrickable_sets" +ON "bricktracker_sets"."set" IS NOT DISTINCT FROM "rebrickable_sets"."set" + -- Left join with set owners (using dynamic columns) LEFT JOIN "bricktracker_set_owners" ON "bricktracker_sets"."id" IS NOT DISTINCT FROM "bricktracker_set_owners"."id" @@ -60,12 +64,38 @@ AND "rebrickable_minifigures"."figure" IS NOT DISTINCT FROM "problem_join"."figu {% endblock %} {% block where %} +{% set conditions = [] %} {% if owner_id and owner_id != 'all' %} -WHERE "bricktracker_set_owners"."owner_{{ owner_id }}" = 1 + {% set _ = conditions.append('"bricktracker_set_owners"."owner_' ~ owner_id ~ '" = 1') %} +{% endif %} +{% if theme_id and theme_id != 'all' %} + {% set _ = conditions.append('"rebrickable_sets"."theme_id" = ' ~ theme_id) %} +{% endif %} +{% if year and year != 'all' %} + {% set _ = conditions.append('"rebrickable_sets"."year" = ' ~ year) %} +{% endif %} +{% if search_query %} + {% set _ = conditions.append('(LOWER("rebrickable_minifigures"."name") LIKE LOWER(\'%' ~ search_query ~ '%\'))') %} +{% endif %} +{% if conditions %} +WHERE {{ conditions | join(' AND ') }} {% endif %} {% endblock %} {% block group %} GROUP BY "rebrickable_minifigures"."figure" +{% endblock %} + +{% block having %} +{% if problems_filter %} +HAVING 1=1 +{% if problems_filter == 'missing' %} +AND SUM(IFNULL("problem_join"."total_missing", 0)) > 0 +{% elif problems_filter == 'damaged' %} +AND SUM(IFNULL("problem_join"."total_damaged", 0)) > 0 +{% elif problems_filter == 'both' %} +AND SUM(IFNULL("problem_join"."total_missing", 0)) > 0 AND SUM(IFNULL("problem_join"."total_damaged", 0)) > 0 +{% endif %} +{% endif %} {% endblock %} \ No newline at end of file diff --git a/bricktracker/sql/minifigure/themes/list.sql b/bricktracker/sql/minifigure/themes/list.sql new file mode 100644 index 0000000..6bd91f4 --- /dev/null +++ b/bricktracker/sql/minifigure/themes/list.sql @@ -0,0 +1,16 @@ +-- Get distinct themes from minifigures' sets +SELECT DISTINCT + "rebrickable_sets"."theme_id", + COUNT(DISTINCT "bricktracker_minifigures"."figure") as "minifigure_count" +FROM "bricktracker_minifigures" +INNER JOIN "bricktracker_sets" +ON "bricktracker_minifigures"."id" IS NOT DISTINCT FROM "bricktracker_sets"."id" +INNER JOIN "rebrickable_sets" +ON "bricktracker_sets"."set" IS NOT DISTINCT FROM "rebrickable_sets"."set" +{% if owner_id and owner_id != 'all' %} +INNER JOIN "bricktracker_set_owners" +ON "bricktracker_sets"."id" IS NOT DISTINCT FROM "bricktracker_set_owners"."id" +WHERE "bricktracker_set_owners"."owner_{{ owner_id }}" = 1 +{% endif %} +GROUP BY "rebrickable_sets"."theme_id" +ORDER BY "rebrickable_sets"."theme_id" ASC diff --git a/bricktracker/sql/minifigure/years/list.sql b/bricktracker/sql/minifigure/years/list.sql new file mode 100644 index 0000000..2349f29 --- /dev/null +++ b/bricktracker/sql/minifigure/years/list.sql @@ -0,0 +1,16 @@ +-- Get distinct years from minifigures' sets +SELECT DISTINCT + "rebrickable_sets"."year", + COUNT(DISTINCT "bricktracker_minifigures"."figure") as "minifigure_count" +FROM "bricktracker_minifigures" +INNER JOIN "bricktracker_sets" +ON "bricktracker_minifigures"."id" IS NOT DISTINCT FROM "bricktracker_sets"."id" +INNER JOIN "rebrickable_sets" +ON "bricktracker_sets"."set" IS NOT DISTINCT FROM "rebrickable_sets"."set" +{% if owner_id and owner_id != 'all' %} +INNER JOIN "bricktracker_set_owners" +ON "bricktracker_sets"."id" IS NOT DISTINCT FROM "bricktracker_set_owners"."id" +WHERE "bricktracker_set_owners"."owner_{{ owner_id }}" = 1 +{% endif %} +GROUP BY "rebrickable_sets"."year" +ORDER BY "rebrickable_sets"."year" DESC diff --git a/bricktracker/sql/part/base/base.sql b/bricktracker/sql/part/base/base.sql index d1b9c5d..890cf20 100644 --- a/bricktracker/sql/part/base/base.sql +++ b/bricktracker/sql/part/base/base.sql @@ -9,6 +9,7 @@ SELECT --"bricktracker_parts"."rebrickable_inventory", "bricktracker_parts"."missing", "bricktracker_parts"."damaged", + "bricktracker_parts"."checked", --"rebrickable_parts"."part", --"rebrickable_parts"."color_id", "rebrickable_parts"."color_name", @@ -60,3 +61,7 @@ ORDER BY {{ order }} {% if limit %} LIMIT {{ limit }} {% endif %} + +{% if offset %} +OFFSET {{ offset }} +{% endif %} diff --git a/bricktracker/sql/part/colors/list_problem.sql b/bricktracker/sql/part/colors/list_problem.sql new file mode 100644 index 0000000..9781b32 --- /dev/null +++ b/bricktracker/sql/part/colors/list_problem.sql @@ -0,0 +1,19 @@ +SELECT DISTINCT + "rebrickable_parts"."color_id" AS "color_id", + "rebrickable_parts"."color_name" AS "color_name", + "rebrickable_parts"."color_rgb" AS "color_rgb" +FROM "rebrickable_parts" +INNER JOIN "bricktracker_parts" +ON "bricktracker_parts"."part" IS NOT DISTINCT FROM "rebrickable_parts"."part" +AND "bricktracker_parts"."color" IS NOT DISTINCT FROM "rebrickable_parts"."color_id" +{% if owner_id and owner_id != 'all' %} +INNER JOIN "bricktracker_sets" +ON "bricktracker_parts"."id" IS NOT DISTINCT FROM "bricktracker_sets"."id" +INNER JOIN "bricktracker_set_owners" +ON "bricktracker_sets"."id" IS NOT DISTINCT FROM "bricktracker_set_owners"."id" +{% endif %} +WHERE ("bricktracker_parts"."missing" > 0 OR "bricktracker_parts"."damaged" > 0) +{% if owner_id and owner_id != 'all' %} +AND "bricktracker_set_owners"."owner_{{ owner_id }}" = 1 +{% endif %} +ORDER BY "rebrickable_parts"."color_name" ASC \ No newline at end of file diff --git a/bricktracker/sql/part/list/all.sql b/bricktracker/sql/part/list/all.sql index bcc729b..154ba29 100644 --- a/bricktracker/sql/part/list/all.sql +++ b/bricktracker/sql/part/list/all.sql @@ -24,11 +24,35 @@ SUM(IFNULL("bricktracker_minifigures"."quantity", 0)) AS "total_minifigures" LEFT JOIN "bricktracker_minifigures" ON "bricktracker_parts"."id" IS NOT DISTINCT FROM "bricktracker_minifigures"."id" AND "bricktracker_parts"."figure" IS NOT DISTINCT FROM "bricktracker_minifigures"."figure" + +{% if theme_id or year %} +INNER JOIN "bricktracker_sets" AS "filter_sets" +ON "bricktracker_parts"."id" IS NOT DISTINCT FROM "filter_sets"."id" +INNER JOIN "rebrickable_sets" AS "filter_rs" +ON "filter_sets"."set" IS NOT DISTINCT FROM "filter_rs"."set" +{% endif %} {% endblock %} {% block where %} +{% set conditions = [] %} {% if color_id and color_id != 'all' %} -WHERE "bricktracker_parts"."color" = {{ color_id }} + {% set _ = conditions.append('"bricktracker_parts"."color" = ' ~ color_id) %} +{% endif %} +{% if theme_id and theme_id != 'all' %} + {% set _ = conditions.append('"filter_rs"."theme_id" = ' ~ theme_id) %} +{% endif %} +{% if year and year != 'all' %} + {% set _ = conditions.append('"filter_rs"."year" = ' ~ year) %} +{% endif %} +{% if search_query %} + {% set search_condition = '(LOWER("rebrickable_parts"."name") LIKE LOWER(\'%' ~ search_query ~ '%\') OR LOWER("rebrickable_parts"."color_name") LIKE LOWER(\'%' ~ search_query ~ '%\') OR LOWER("bricktracker_parts"."part") LIKE LOWER(\'%' ~ search_query ~ '%\'))' %} + {% set _ = conditions.append(search_condition) %} +{% endif %} +{% if skip_spare_parts %} + {% set _ = conditions.append('"bricktracker_parts"."spare" = 0') %} +{% endif %} +{% if conditions %} +WHERE {{ conditions | join(' AND ') }} {% endif %} {% endblock %} diff --git a/bricktracker/sql/part/list/all_by_owner.sql b/bricktracker/sql/part/list/all_by_owner.sql index ccfc301..338f6e0 100644 --- a/bricktracker/sql/part/list/all_by_owner.sql +++ b/bricktracker/sql/part/list/all_by_owner.sql @@ -56,17 +56,22 @@ AND "bricktracker_parts"."figure" IS NOT DISTINCT FROM "bricktracker_minifigures {% endblock %} {% block where %} -{% set has_where = false %} +{% set conditions = [] %} {% if owner_id and owner_id != 'all' %} -WHERE "bricktracker_set_owners"."owner_{{ owner_id }}" = 1 -{% set has_where = true %} + {% set _ = conditions.append('"bricktracker_set_owners"."owner_' ~ owner_id ~ '" = 1') %} {% endif %} {% if color_id and color_id != 'all' %} -{% if has_where %} -AND "bricktracker_parts"."color" = {{ color_id }} -{% else %} -WHERE "bricktracker_parts"."color" = {{ color_id }} + {% set _ = conditions.append('"bricktracker_parts"."color" = ' ~ color_id) %} {% endif %} +{% if search_query %} + {% set search_condition = '(LOWER("rebrickable_parts"."name") LIKE LOWER(\'%' ~ search_query ~ '%\') OR LOWER("rebrickable_parts"."color_name") LIKE LOWER(\'%' ~ search_query ~ '%\') OR LOWER("bricktracker_parts"."part") LIKE LOWER(\'%' ~ search_query ~ '%\'))' %} + {% set _ = conditions.append(search_condition) %} +{% endif %} +{% if skip_spare_parts %} + {% set _ = conditions.append('"bricktracker_parts"."spare" = 0') %} +{% endif %} +{% if conditions %} +WHERE {{ conditions | join(' AND ') }} {% endif %} {% endblock %} diff --git a/bricktracker/sql/part/list/from_minifigure.sql b/bricktracker/sql/part/list/from_minifigure.sql index 115b791..a2b456f 100644 --- a/bricktracker/sql/part/list/from_minifigure.sql +++ b/bricktracker/sql/part/list/from_minifigure.sql @@ -10,7 +10,12 @@ SUM("bricktracker_parts"."damaged") AS "total_damaged", {% endblock %} {% block where %} -WHERE "bricktracker_parts"."figure" IS NOT DISTINCT FROM :figure +{% set conditions = [] %} +{% set _ = conditions.append('"bricktracker_parts"."figure" IS NOT DISTINCT FROM :figure') %} +{% if skip_spare_parts %} + {% set _ = conditions.append('"bricktracker_parts"."spare" = 0') %} +{% endif %} +WHERE {{ conditions | join(' AND ') }} {% endblock %} {% block group %} diff --git a/bricktracker/sql/part/list/problem.sql b/bricktracker/sql/part/list/problem.sql index 068b8d8..4d36df3 100644 --- a/bricktracker/sql/part/list/problem.sql +++ b/bricktracker/sql/part/list/problem.sql @@ -1,30 +1,100 @@ {% extends 'part/base/base.sql' %} {% block total_missing %} +{% if owner_id and owner_id != 'all' %} +SUM(CASE WHEN "bricktracker_set_owners"."owner_{{ owner_id }}" = 1 THEN "bricktracker_parts"."missing" ELSE 0 END) AS "total_missing", +{% else %} SUM("bricktracker_parts"."missing") AS "total_missing", +{% endif %} {% endblock %} {% block total_damaged %} +{% if owner_id and owner_id != 'all' %} +SUM(CASE WHEN "bricktracker_set_owners"."owner_{{ owner_id }}" = 1 THEN "bricktracker_parts"."damaged" ELSE 0 END) AS "total_damaged", +{% else %} SUM("bricktracker_parts"."damaged") AS "total_damaged", +{% endif %} +{% endblock %} + +{% block total_quantity %} +{% if owner_id and owner_id != 'all' %} +SUM(CASE WHEN "bricktracker_set_owners"."owner_{{ owner_id }}" = 1 THEN "bricktracker_parts"."quantity" * IFNULL("bricktracker_minifigures"."quantity", 1) ELSE 0 END) AS "total_quantity", +{% else %} +SUM("bricktracker_parts"."quantity" * IFNULL("bricktracker_minifigures"."quantity", 1)) AS "total_quantity", +{% endif %} {% endblock %} {% block total_sets %} -IFNULL(COUNT("bricktracker_parts"."id"), 0) - IFNULL(COUNT("bricktracker_parts"."figure"), 0) AS "total_sets", +{% if owner_id and owner_id != 'all' %} +COUNT(DISTINCT CASE WHEN "bricktracker_set_owners"."owner_{{ owner_id }}" = 1 THEN "bricktracker_parts"."id" ELSE NULL END) AS "total_sets", +{% else %} +COUNT(DISTINCT "bricktracker_parts"."id") AS "total_sets", +{% endif %} {% endblock %} {% block total_minifigures %} +{% if owner_id and owner_id != 'all' %} +SUM(CASE WHEN "bricktracker_set_owners"."owner_{{ owner_id }}" = 1 THEN IFNULL("bricktracker_minifigures"."quantity", 0) ELSE 0 END) AS "total_minifigures" +{% else %} SUM(IFNULL("bricktracker_minifigures"."quantity", 0)) AS "total_minifigures" +{% endif %} {% endblock %} {% block join %} +-- Join with sets to get owner information +INNER JOIN "bricktracker_sets" +ON "bricktracker_parts"."id" IS NOT DISTINCT FROM "bricktracker_sets"."id" + +-- Join with rebrickable sets for theme/year filtering +INNER JOIN "rebrickable_sets" +ON "bricktracker_sets"."set" IS NOT DISTINCT FROM "rebrickable_sets"."set" + +-- Left join with set owners (using dynamic columns) +LEFT JOIN "bricktracker_set_owners" +ON "bricktracker_sets"."id" IS NOT DISTINCT FROM "bricktracker_set_owners"."id" + +-- Left join with set tags (for tag filtering) +{% if tag_id and tag_id != 'all' %} +LEFT JOIN "bricktracker_set_tags" +ON "bricktracker_sets"."id" IS NOT DISTINCT FROM "bricktracker_set_tags"."id" +{% endif %} + +-- Left join with minifigures LEFT JOIN "bricktracker_minifigures" ON "bricktracker_parts"."id" IS NOT DISTINCT FROM "bricktracker_minifigures"."id" AND "bricktracker_parts"."figure" IS NOT DISTINCT FROM "bricktracker_minifigures"."figure" {% endblock %} {% block where %} -WHERE "bricktracker_parts"."missing" > 0 -OR "bricktracker_parts"."damaged" > 0 +{% set conditions = [] %} +-- Always filter for problematic parts +{% set _ = conditions.append('("bricktracker_parts"."missing" > 0 OR "bricktracker_parts"."damaged" > 0)') %} +{% if owner_id and owner_id != 'all' %} + {% set _ = conditions.append('"bricktracker_set_owners"."owner_' ~ owner_id ~ '" = 1') %} +{% endif %} +{% if color_id and color_id != 'all' %} + {% set _ = conditions.append('"bricktracker_parts"."color" = ' ~ color_id) %} +{% endif %} +{% if theme_id and theme_id != 'all' %} + {% set _ = conditions.append('"rebrickable_sets"."theme_id" = ' ~ theme_id) %} +{% endif %} +{% if year and year != 'all' %} + {% set _ = conditions.append('"rebrickable_sets"."year" = ' ~ year) %} +{% endif %} +{% if storage_id and storage_id != 'all' %} + {% set _ = conditions.append('"bricktracker_sets"."storage" = \'' ~ storage_id ~ '\'') %} +{% endif %} +{% if tag_id and tag_id != 'all' %} + {% set _ = conditions.append('"bricktracker_set_tags"."tag_' ~ tag_id ~ '" = 1') %} +{% endif %} +{% if search_query %} + {% set search_condition = '(LOWER("rebrickable_parts"."name") LIKE LOWER(\'%' ~ search_query ~ '%\') OR LOWER("rebrickable_parts"."color_name") LIKE LOWER(\'%' ~ search_query ~ '%\') OR LOWER("bricktracker_parts"."part") LIKE LOWER(\'%' ~ search_query ~ '%\'))' %} + {% set _ = conditions.append(search_condition) %} +{% endif %} +{% if skip_spare_parts %} + {% set _ = conditions.append('"bricktracker_parts"."spare" = 0') %} +{% endif %} +WHERE {{ conditions | join(' AND ') }} {% endblock %} {% block group %} diff --git a/bricktracker/sql/part/list/specific.sql b/bricktracker/sql/part/list/specific.sql index 7c62c68..996c127 100644 --- a/bricktracker/sql/part/list/specific.sql +++ b/bricktracker/sql/part/list/specific.sql @@ -10,6 +10,11 @@ IFNULL("bricktracker_parts"."damaged", 0) AS "total_damaged", {% endblock %} {% block where %} -WHERE "bricktracker_parts"."id" IS NOT DISTINCT FROM :id -AND "bricktracker_parts"."figure" IS NOT DISTINCT FROM :figure +{% set conditions = [] %} +{% set _ = conditions.append('"bricktracker_parts"."id" IS NOT DISTINCT FROM :id') %} +{% set _ = conditions.append('"bricktracker_parts"."figure" IS NOT DISTINCT FROM :figure') %} +{% if skip_spare_parts %} + {% set _ = conditions.append('"bricktracker_parts"."spare" = 0') %} +{% endif %} +WHERE {{ conditions | join(' AND ') }} {% endblock %} diff --git a/bricktracker/sql/part/storages/list_problem.sql b/bricktracker/sql/part/storages/list_problem.sql new file mode 100644 index 0000000..4df125d --- /dev/null +++ b/bricktracker/sql/part/storages/list_problem.sql @@ -0,0 +1,21 @@ +-- Get distinct storages from problem parts' sets +SELECT DISTINCT + "bricktracker_sets"."storage" AS "storage_id", + "bricktracker_metadata_storages"."name" AS "storage_name", + COUNT(DISTINCT "bricktracker_parts"."part") as "part_count" +FROM "bricktracker_parts" +INNER JOIN "bricktracker_sets" +ON "bricktracker_parts"."id" IS NOT DISTINCT FROM "bricktracker_sets"."id" +LEFT JOIN "bricktracker_metadata_storages" +ON "bricktracker_sets"."storage" IS NOT DISTINCT FROM "bricktracker_metadata_storages"."id" +{% if owner_id and owner_id != 'all' %} +INNER JOIN "bricktracker_set_owners" +ON "bricktracker_sets"."id" IS NOT DISTINCT FROM "bricktracker_set_owners"."id" +{% endif %} +WHERE ("bricktracker_parts"."missing" > 0 OR "bricktracker_parts"."damaged" > 0) +AND "bricktracker_sets"."storage" IS NOT NULL +{% if owner_id and owner_id != 'all' %} +AND "bricktracker_set_owners"."owner_{{ owner_id }}" = 1 +{% endif %} +GROUP BY "bricktracker_sets"."storage", "bricktracker_metadata_storages"."name" +ORDER BY "bricktracker_metadata_storages"."name" ASC diff --git a/bricktracker/sql/part/tags/list_problem.sql b/bricktracker/sql/part/tags/list_problem.sql new file mode 100644 index 0000000..ec45a6b --- /dev/null +++ b/bricktracker/sql/part/tags/list_problem.sql @@ -0,0 +1,7 @@ +-- Get list of all tags (simplified - filtering happens at application level) +-- Tags use dynamic columns in bricktracker_set_tags, making direct SQL filtering complex +SELECT + "bricktracker_metadata_tags"."id" AS "tag_id", + "bricktracker_metadata_tags"."name" AS "tag_name" +FROM "bricktracker_metadata_tags" +ORDER BY "bricktracker_metadata_tags"."name" ASC diff --git a/bricktracker/sql/part/themes/list.sql b/bricktracker/sql/part/themes/list.sql new file mode 100644 index 0000000..e5f8fa1 --- /dev/null +++ b/bricktracker/sql/part/themes/list.sql @@ -0,0 +1,16 @@ +-- Get distinct themes from parts' sets +SELECT DISTINCT + "rebrickable_sets"."theme_id", + COUNT(DISTINCT "bricktracker_parts"."part") as "part_count" +FROM "bricktracker_parts" +INNER JOIN "bricktracker_sets" +ON "bricktracker_parts"."id" IS NOT DISTINCT FROM "bricktracker_sets"."id" +INNER JOIN "rebrickable_sets" +ON "bricktracker_sets"."set" IS NOT DISTINCT FROM "rebrickable_sets"."set" +{% if owner_id and owner_id != 'all' %} +INNER JOIN "bricktracker_set_owners" +ON "bricktracker_sets"."id" IS NOT DISTINCT FROM "bricktracker_set_owners"."id" +WHERE "bricktracker_set_owners"."owner_{{ owner_id }}" = 1 +{% endif %} +GROUP BY "rebrickable_sets"."theme_id" +ORDER BY "rebrickable_sets"."theme_id" ASC diff --git a/bricktracker/sql/part/themes/list_problem.sql b/bricktracker/sql/part/themes/list_problem.sql new file mode 100644 index 0000000..b96f267 --- /dev/null +++ b/bricktracker/sql/part/themes/list_problem.sql @@ -0,0 +1,19 @@ +-- Get distinct themes from problem parts' sets +SELECT DISTINCT + "rebrickable_sets"."theme_id", + COUNT(DISTINCT "bricktracker_parts"."part") as "part_count" +FROM "bricktracker_parts" +INNER JOIN "bricktracker_sets" +ON "bricktracker_parts"."id" IS NOT DISTINCT FROM "bricktracker_sets"."id" +INNER JOIN "rebrickable_sets" +ON "bricktracker_sets"."set" IS NOT DISTINCT FROM "rebrickable_sets"."set" +{% if owner_id and owner_id != 'all' %} +INNER JOIN "bricktracker_set_owners" +ON "bricktracker_sets"."id" IS NOT DISTINCT FROM "bricktracker_set_owners"."id" +{% endif %} +WHERE ("bricktracker_parts"."missing" > 0 OR "bricktracker_parts"."damaged" > 0) +{% if owner_id and owner_id != 'all' %} +AND "bricktracker_set_owners"."owner_{{ owner_id }}" = 1 +{% endif %} +GROUP BY "rebrickable_sets"."theme_id" +ORDER BY "rebrickable_sets"."theme_id" ASC diff --git a/bricktracker/sql/part/update/checked.sql b/bricktracker/sql/part/update/checked.sql new file mode 100644 index 0000000..a29a3fc --- /dev/null +++ b/bricktracker/sql/part/update/checked.sql @@ -0,0 +1,7 @@ +UPDATE "bricktracker_parts" +SET "checked" = :checked +WHERE "bricktracker_parts"."id" IS NOT DISTINCT FROM :id +AND "bricktracker_parts"."figure" IS NOT DISTINCT FROM :figure +AND "bricktracker_parts"."part" IS NOT DISTINCT FROM :part +AND "bricktracker_parts"."color" IS NOT DISTINCT FROM :color +AND "bricktracker_parts"."spare" IS NOT DISTINCT FROM :spare \ No newline at end of file diff --git a/bricktracker/sql/part/years/list.sql b/bricktracker/sql/part/years/list.sql new file mode 100644 index 0000000..d2bb185 --- /dev/null +++ b/bricktracker/sql/part/years/list.sql @@ -0,0 +1,16 @@ +-- Get distinct years from parts' sets +SELECT DISTINCT + "rebrickable_sets"."year", + COUNT(DISTINCT "bricktracker_parts"."part") as "part_count" +FROM "bricktracker_parts" +INNER JOIN "bricktracker_sets" +ON "bricktracker_parts"."id" IS NOT DISTINCT FROM "bricktracker_sets"."id" +INNER JOIN "rebrickable_sets" +ON "bricktracker_sets"."set" IS NOT DISTINCT FROM "rebrickable_sets"."set" +{% if owner_id and owner_id != 'all' %} +INNER JOIN "bricktracker_set_owners" +ON "bricktracker_sets"."id" IS NOT DISTINCT FROM "bricktracker_set_owners"."id" +WHERE "bricktracker_set_owners"."owner_{{ owner_id }}" = 1 +{% endif %} +GROUP BY "rebrickable_sets"."year" +ORDER BY "rebrickable_sets"."year" DESC diff --git a/bricktracker/sql/part/years/list_problem.sql b/bricktracker/sql/part/years/list_problem.sql new file mode 100644 index 0000000..cc1af80 --- /dev/null +++ b/bricktracker/sql/part/years/list_problem.sql @@ -0,0 +1,19 @@ +-- Get distinct years from problem parts' sets +SELECT DISTINCT + "rebrickable_sets"."year", + COUNT(DISTINCT "bricktracker_parts"."part") as "part_count" +FROM "bricktracker_parts" +INNER JOIN "bricktracker_sets" +ON "bricktracker_parts"."id" IS NOT DISTINCT FROM "bricktracker_sets"."id" +INNER JOIN "rebrickable_sets" +ON "bricktracker_sets"."set" IS NOT DISTINCT FROM "rebrickable_sets"."set" +{% if owner_id and owner_id != 'all' %} +INNER JOIN "bricktracker_set_owners" +ON "bricktracker_sets"."id" IS NOT DISTINCT FROM "bricktracker_set_owners"."id" +{% endif %} +WHERE ("bricktracker_parts"."missing" > 0 OR "bricktracker_parts"."damaged" > 0) +{% if owner_id and owner_id != 'all' %} +AND "bricktracker_set_owners"."owner_{{ owner_id }}" = 1 +{% endif %} +GROUP BY "rebrickable_sets"."year" +ORDER BY "rebrickable_sets"."year" DESC diff --git a/bricktracker/sql/set/base/base.sql b/bricktracker/sql/set/base/base.sql index c19ffc8..5ecc2cf 100644 --- a/bricktracker/sql/set/base/base.sql +++ b/bricktracker/sql/set/base/base.sql @@ -49,3 +49,7 @@ ORDER BY {{ order }} {% if limit %} LIMIT {{ limit }} {% endif %} + +{% if offset %} +OFFSET {{ offset }} +{% endif %} diff --git a/bricktracker/sql/set/check_theme_exists.sql b/bricktracker/sql/set/check_theme_exists.sql new file mode 100644 index 0000000..93ff65d --- /dev/null +++ b/bricktracker/sql/set/check_theme_exists.sql @@ -0,0 +1,4 @@ +SELECT COUNT(*) as count +FROM "bricktracker_sets" +INNER JOIN "rebrickable_sets" ON "bricktracker_sets"."set" = "rebrickable_sets"."set" +WHERE "rebrickable_sets"."theme_id" = {{ theme_id }} \ No newline at end of file diff --git a/bricktracker/sql/set/delete/set.sql b/bricktracker/sql/set/delete/set.sql index 4eca845..7fbd678 100644 --- a/bricktracker/sql/set/delete/set.sql +++ b/bricktracker/sql/set/delete/set.sql @@ -3,22 +3,24 @@ BEGIN TRANSACTION; -DELETE FROM "bricktracker_sets" -WHERE "bricktracker_sets"."id" IS NOT DISTINCT FROM '{{ id }}'; - -DELETE FROM "bricktracker_set_owners" -WHERE "bricktracker_set_owners"."id" IS NOT DISTINCT FROM '{{ id }}'; - -DELETE FROM "bricktracker_set_statuses" -WHERE "bricktracker_set_statuses"."id" IS NOT DISTINCT FROM '{{ id }}'; - -DELETE FROM "bricktracker_set_tags" -WHERE "bricktracker_set_tags"."id" IS NOT DISTINCT FROM '{{ id }}'; +-- Delete child records first (those with foreign keys referencing bricktracker_sets) +DELETE FROM "bricktracker_parts" +WHERE "bricktracker_parts"."id" IS NOT DISTINCT FROM '{{ id }}'; DELETE FROM "bricktracker_minifigures" WHERE "bricktracker_minifigures"."id" IS NOT DISTINCT FROM '{{ id }}'; -DELETE FROM "bricktracker_parts" -WHERE "bricktracker_parts"."id" IS NOT DISTINCT FROM '{{ id }}'; +DELETE FROM "bricktracker_set_tags" +WHERE "bricktracker_set_tags"."id" IS NOT DISTINCT FROM '{{ id }}'; + +DELETE FROM "bricktracker_set_statuses" +WHERE "bricktracker_set_statuses"."id" IS NOT DISTINCT FROM '{{ id }}'; + +DELETE FROM "bricktracker_set_owners" +WHERE "bricktracker_set_owners"."id" IS NOT DISTINCT FROM '{{ id }}'; + +-- Delete the parent record last +DELETE FROM "bricktracker_sets" +WHERE "bricktracker_sets"."id" IS NOT DISTINCT FROM '{{ id }}'; COMMIT; \ No newline at end of file diff --git a/bricktracker/sql/set/list/all.sql b/bricktracker/sql/set/list/all.sql index 28629cc..c6d2ce4 100644 --- a/bricktracker/sql/set/list/all.sql +++ b/bricktracker/sql/set/list/all.sql @@ -1 +1,8 @@ {% extends 'set/base/full.sql' %} + +{% block where %} +{% if search_query %} +WHERE (LOWER("rebrickable_sets"."name") LIKE LOWER('%{{ search_query }}%') + OR LOWER("rebrickable_sets"."set") LIKE LOWER('%{{ search_query }}%')) +{% endif %} +{% endblock %} diff --git a/bricktracker/sql/set/list/all_filtered.sql b/bricktracker/sql/set/list/all_filtered.sql new file mode 100644 index 0000000..18fde15 --- /dev/null +++ b/bricktracker/sql/set/list/all_filtered.sql @@ -0,0 +1,81 @@ +{% extends 'set/base/full.sql' %} + +{% block where %} +WHERE 1=1 +{% if search_query %} +AND (LOWER("rebrickable_sets"."name") LIKE LOWER('%{{ search_query }}%') + OR LOWER("rebrickable_sets"."set") LIKE LOWER('%{{ search_query }}%')) +{% endif %} + +{% if theme_filter %} +AND "rebrickable_sets"."theme_id" = {{ theme_filter }} +{% endif %} + +{% if year_filter %} +AND "rebrickable_sets"."year" = {{ year_filter }} +{% endif %} + +{% if storage_filter %} +AND "bricktracker_sets"."storage" = '{{ storage_filter }}' +{% endif %} + +{% if purchase_location_filter %} +AND "bricktracker_sets"."purchase_location" = '{{ purchase_location_filter }}' +{% endif %} + +{% if status_filter %} +{% if status_filter == 'has-missing' %} +AND IFNULL("problem_join"."total_missing", 0) > 0 +{% elif status_filter == '-has-missing' %} +AND IFNULL("problem_join"."total_missing", 0) = 0 +{% elif status_filter == 'has-damaged' %} +AND IFNULL("problem_join"."total_damaged", 0) > 0 +{% elif status_filter == '-has-damaged' %} +AND IFNULL("problem_join"."total_damaged", 0) = 0 +{% elif status_filter == 'has-storage' %} +AND "bricktracker_sets"."storage" IS NOT NULL AND "bricktracker_sets"."storage" != '' +{% elif status_filter == '-has-storage' %} +AND ("bricktracker_sets"."storage" IS NULL OR "bricktracker_sets"."storage" = '') +{% elif status_filter.startswith('status-') %} +AND EXISTS ( + SELECT 1 FROM "bricktracker_set_statuses" + WHERE "bricktracker_set_statuses"."id" = "bricktracker_sets"."id" + AND "bricktracker_set_statuses"."{{ status_filter.replace('-', '_') }}" = 1 +) +{% elif status_filter.startswith('-status-') %} +AND NOT EXISTS ( + SELECT 1 FROM "bricktracker_set_statuses" + WHERE "bricktracker_set_statuses"."id" = "bricktracker_sets"."id" + AND "bricktracker_set_statuses"."{{ status_filter[1:].replace('-', '_') }}" = 1 +) +{% endif %} +{% endif %} + +{% if owner_filter %} +{% if owner_filter.startswith('owner-') %} +AND EXISTS ( + SELECT 1 FROM "bricktracker_set_owners" + WHERE "bricktracker_set_owners"."id" = "bricktracker_sets"."id" + AND "bricktracker_set_owners"."{{ owner_filter.replace('-', '_') }}" = 1 +) +{% endif %} +{% endif %} + +{% if tag_filter %} +{% if tag_filter.startswith('tag-') %} +AND EXISTS ( + SELECT 1 FROM "bricktracker_set_tags" + WHERE "bricktracker_set_tags"."id" = "bricktracker_sets"."id" + AND "bricktracker_set_tags"."{{ tag_filter.replace('-', '_') }}" = 1 +) +{% endif %} +{% endif %} + +{% if duplicate_filter %} +AND ( + SELECT COUNT(*) + FROM "bricktracker_sets" as "duplicate_check" + WHERE "duplicate_check"."set" = "bricktracker_sets"."set" +) > 1 +{% endif %} +{% endblock %} \ No newline at end of file diff --git a/bricktracker/sql/set/list/consolidated.sql b/bricktracker/sql/set/list/consolidated.sql new file mode 100644 index 0000000..e740015 --- /dev/null +++ b/bricktracker/sql/set/list/consolidated.sql @@ -0,0 +1,178 @@ +SELECT + (SELECT MIN("id") FROM "bricktracker_sets" WHERE "set" = "rebrickable_sets"."set") AS "id", + "rebrickable_sets"."set", + "rebrickable_sets"."number", + "rebrickable_sets"."version", + "rebrickable_sets"."name", + "rebrickable_sets"."year", + "rebrickable_sets"."theme_id", + "rebrickable_sets"."number_of_parts", + "rebrickable_sets"."image", + "rebrickable_sets"."url", + COUNT("bricktracker_sets"."id") AS "instance_count", + IFNULL(SUM("problem_join"."total_missing"), 0) AS "total_missing", + IFNULL(SUM("problem_join"."total_damaged"), 0) AS "total_damaged", + IFNULL(MAX("minifigures_join"."total"), 0) AS "total_minifigures", + -- Keep one representative instance for display purposes + GROUP_CONCAT("bricktracker_sets"."id", '|') AS "instance_ids", + REPLACE(GROUP_CONCAT(DISTINCT "bricktracker_sets"."storage"), ',', '|') AS "storage", + MIN("bricktracker_sets"."purchase_date") AS "purchase_date", + MAX("bricktracker_sets"."purchase_date") AS "purchase_date_max", + REPLACE(GROUP_CONCAT(DISTINCT "bricktracker_sets"."purchase_location"), ',', '|') AS "purchase_location", + ROUND(AVG("bricktracker_sets"."purchase_price"), 1) AS "purchase_price" + {% block owners %} + {% if owners_dict %} + {% for column, uuid in owners_dict.items() %} + , MAX("bricktracker_set_owners"."{{ column }}") AS "{{ column }}" + {% endfor %} + {% endif %} + {% endblock %} + {% block tags %} + {% if tags_dict %} + {% for column, uuid in tags_dict.items() %} + , MAX("bricktracker_set_tags"."{{ column }}") AS "{{ column }}" + {% endfor %} + {% endif %} + {% endblock %} + {% block statuses %} + {% if statuses_dict %} + {% for column, uuid in statuses_dict.items() %} + , MAX("bricktracker_set_statuses"."{{ column }}") AS "{{ column }}" + , IFNULL(SUM("bricktracker_set_statuses"."{{ column }}"), 0) AS "{{ column }}_count" + {% endfor %} + {% endif %} + {% endblock %} +FROM "bricktracker_sets" + +INNER JOIN "rebrickable_sets" +ON "bricktracker_sets"."set" IS NOT DISTINCT FROM "rebrickable_sets"."set" + +-- LEFT JOIN + SELECT to avoid messing the total +LEFT JOIN ( + SELECT + "bricktracker_parts"."id", + SUM("bricktracker_parts"."missing") AS "total_missing", + SUM("bricktracker_parts"."damaged") AS "total_damaged" + FROM "bricktracker_parts" + GROUP BY "bricktracker_parts"."id" +) "problem_join" +ON "bricktracker_sets"."id" IS NOT DISTINCT FROM "problem_join"."id" + +-- LEFT JOIN + SELECT to avoid messing the total +LEFT JOIN ( + SELECT + "bricktracker_minifigures"."id", + SUM("bricktracker_minifigures"."quantity") AS "total" + FROM "bricktracker_minifigures" + GROUP BY "bricktracker_minifigures"."id" +) "minifigures_join" +ON "bricktracker_sets"."id" IS NOT DISTINCT FROM "minifigures_join"."id" + +{% if owners_dict %} +LEFT JOIN "bricktracker_set_owners" +ON "bricktracker_sets"."id" IS NOT DISTINCT FROM "bricktracker_set_owners"."id" +{% endif %} + +{% if statuses_dict %} +LEFT JOIN "bricktracker_set_statuses" +ON "bricktracker_sets"."id" IS NOT DISTINCT FROM "bricktracker_set_statuses"."id" +{% endif %} + +{% if tags_dict %} +LEFT JOIN "bricktracker_set_tags" +ON "bricktracker_sets"."id" IS NOT DISTINCT FROM "bricktracker_set_tags"."id" +{% endif %} + +{% block where %} +WHERE 1=1 +{% if search_query %} +AND (LOWER("rebrickable_sets"."name") LIKE LOWER('%{{ search_query }}%') + OR LOWER("rebrickable_sets"."set") LIKE LOWER('%{{ search_query }}%')) +{% endif %} + +{% if theme_filter %} +AND "rebrickable_sets"."theme_id" = {{ theme_filter }} +{% endif %} + +{% if year_filter %} +AND "rebrickable_sets"."year" = {{ year_filter }} +{% endif %} + +{% if storage_filter %} +AND EXISTS ( + SELECT 1 FROM "bricktracker_sets" bs_filter + WHERE bs_filter."set" = "rebrickable_sets"."set" + AND bs_filter."storage" = '{{ storage_filter }}' +) +{% endif %} + +{% if purchase_location_filter %} +AND EXISTS ( + SELECT 1 FROM "bricktracker_sets" bs_filter + WHERE bs_filter."set" = "rebrickable_sets"."set" + AND bs_filter."purchase_location" = '{{ purchase_location_filter }}' +) +{% endif %} + +{% if status_filter %} +{% if status_filter == 'has-storage' %} +AND EXISTS ( + SELECT 1 FROM "bricktracker_sets" bs_filter + WHERE bs_filter."set" = "rebrickable_sets"."set" + AND bs_filter."storage" IS NOT NULL AND bs_filter."storage" != '' +) +{% elif status_filter == '-has-storage' %} +AND NOT EXISTS ( + SELECT 1 FROM "bricktracker_sets" bs_filter + WHERE bs_filter."set" = "rebrickable_sets"."set" + AND bs_filter."storage" IS NOT NULL AND bs_filter."storage" != '' +) +{% elif status_filter.startswith('status-') %} +AND EXISTS ( + SELECT 1 FROM "bricktracker_sets" bs_filter + JOIN "bricktracker_set_statuses" ON bs_filter."id" = "bricktracker_set_statuses"."id" + WHERE bs_filter."set" = "rebrickable_sets"."set" + AND "bricktracker_set_statuses"."{{ status_filter.replace('-', '_') }}" = 1 +) +{% elif status_filter.startswith('-status-') %} +AND NOT EXISTS ( + SELECT 1 FROM "bricktracker_sets" bs_filter + JOIN "bricktracker_set_statuses" ON bs_filter."id" = "bricktracker_set_statuses"."id" + WHERE bs_filter."set" = "rebrickable_sets"."set" + AND "bricktracker_set_statuses"."{{ status_filter[1:].replace('-', '_') }}" = 1 +) +{% endif %} +{% endif %} +{% endblock %} + +GROUP BY "rebrickable_sets"."set" + +{% if status_filter or duplicate_filter %} +HAVING 1=1 +{% if status_filter %} +{% if status_filter == 'has-missing' %} +AND IFNULL(SUM("problem_join"."total_missing"), 0) > 0 +{% elif status_filter == '-has-missing' %} +AND IFNULL(SUM("problem_join"."total_missing"), 0) = 0 +{% elif status_filter == 'has-damaged' %} +AND IFNULL(SUM("problem_join"."total_damaged"), 0) > 0 +{% elif status_filter == '-has-damaged' %} +AND IFNULL(SUM("problem_join"."total_damaged"), 0) = 0 +{% endif %} +{% endif %} +{% if duplicate_filter %} +AND COUNT("bricktracker_sets"."id") > 1 +{% endif %} +{% endif %} + +{% if order %} +ORDER BY {{ order }} +{% endif %} + +{% if limit %} +LIMIT {{ limit }} +{% endif %} + +{% if offset %} +OFFSET {{ offset }} +{% endif %} \ No newline at end of file diff --git a/bricktracker/sql/set/list/damaged_minifigure.sql b/bricktracker/sql/set/list/damaged_minifigure.sql index 51a615d..45868e6 100644 --- a/bricktracker/sql/set/list/damaged_minifigure.sql +++ b/bricktracker/sql/set/list/damaged_minifigure.sql @@ -5,7 +5,7 @@ WHERE "bricktracker_sets"."id" IN ( SELECT "bricktracker_parts"."id" FROM "bricktracker_parts" WHERE "bricktracker_parts"."figure" IS NOT DISTINCT FROM :figure - AND "bricktracker_parts"."missing" > 0 + AND "bricktracker_parts"."damaged" > 0 GROUP BY "bricktracker_parts"."id" ) {% endblock %} diff --git a/bricktracker/sql/set/list/themes_only.sql b/bricktracker/sql/set/list/themes_only.sql new file mode 100644 index 0000000..d0551f8 --- /dev/null +++ b/bricktracker/sql/set/list/themes_only.sql @@ -0,0 +1,87 @@ +SELECT DISTINCT "rebrickable_sets"."theme_id" +FROM "bricktracker_sets" + +INNER JOIN "rebrickable_sets" +ON "bricktracker_sets"."set" IS NOT DISTINCT FROM "rebrickable_sets"."set" + +{% block where %} +WHERE 1=1 +{% if search_query %} +AND (LOWER("rebrickable_sets"."name") LIKE LOWER('%{{ search_query }}%') + OR LOWER("rebrickable_sets"."set") LIKE LOWER('%{{ search_query }}%')) +{% endif %} + +{% if storage_filter %} +AND EXISTS ( + SELECT 1 FROM "bricktracker_sets" bs_filter + WHERE bs_filter."set" = "rebrickable_sets"."set" + AND bs_filter."storage" = '{{ storage_filter }}' +) +{% endif %} + +{% if purchase_location_filter %} +AND EXISTS ( + SELECT 1 FROM "bricktracker_sets" bs_filter + WHERE bs_filter."set" = "rebrickable_sets"."set" + AND bs_filter."purchase_location" = '{{ purchase_location_filter }}' +) +{% endif %} + +{% if status_filter %} +{% if status_filter == 'has-storage' %} +AND EXISTS ( + SELECT 1 FROM "bricktracker_sets" bs_filter + WHERE bs_filter."set" = "rebrickable_sets"."set" + AND bs_filter."storage" IS NOT NULL AND bs_filter."storage" != '' +) +{% elif status_filter == '-has-storage' %} +AND NOT EXISTS ( + SELECT 1 FROM "bricktracker_sets" bs_filter + WHERE bs_filter."set" = "rebrickable_sets"."set" + AND bs_filter."storage" IS NOT NULL AND bs_filter."storage" != '' +) +{% elif status_filter.startswith('status-') %} +AND EXISTS ( + SELECT 1 FROM "bricktracker_sets" bs_filter + JOIN "bricktracker_set_statuses" ON bs_filter."id" = "bricktracker_set_statuses"."id" + WHERE bs_filter."set" = "rebrickable_sets"."set" + AND "bricktracker_set_statuses"."{{ status_filter.replace('-', '_') }}" = 1 +) +{% elif status_filter.startswith('-status-') %} +AND NOT EXISTS ( + SELECT 1 FROM "bricktracker_sets" bs_filter + JOIN "bricktracker_set_statuses" ON bs_filter."id" = "bricktracker_set_statuses"."id" + WHERE bs_filter."set" = "rebrickable_sets"."set" + AND "bricktracker_set_statuses"."{{ status_filter[1:].replace('-', '_') }}" = 1 +) +{% elif status_filter == 'has-missing' %} +AND EXISTS ( + SELECT 1 FROM "bricktracker_sets" bs_filter + JOIN "bricktracker_parts" ON bs_filter."id" = "bricktracker_parts"."id" + WHERE bs_filter."set" = "rebrickable_sets"."set" + AND "bricktracker_parts"."missing" > 0 +) +{% elif status_filter == '-has-missing' %} +AND NOT EXISTS ( + SELECT 1 FROM "bricktracker_sets" bs_filter + JOIN "bricktracker_parts" ON bs_filter."id" = "bricktracker_parts"."id" + WHERE bs_filter."set" = "rebrickable_sets"."set" + AND "bricktracker_parts"."missing" > 0 +) +{% elif status_filter == 'has-damaged' %} +AND EXISTS ( + SELECT 1 FROM "bricktracker_sets" bs_filter + JOIN "bricktracker_parts" ON bs_filter."id" = "bricktracker_parts"."id" + WHERE bs_filter."set" = "rebrickable_sets"."set" + AND "bricktracker_parts"."damaged" > 0 +) +{% elif status_filter == '-has-damaged' %} +AND NOT EXISTS ( + SELECT 1 FROM "bricktracker_sets" bs_filter + JOIN "bricktracker_parts" ON bs_filter."id" = "bricktracker_parts"."id" + WHERE bs_filter."set" = "rebrickable_sets"."set" + AND "bricktracker_parts"."damaged" > 0 +) +{% endif %} +{% endif %} +{% endblock %} \ No newline at end of file diff --git a/bricktracker/sql/set/list/years_only.sql b/bricktracker/sql/set/list/years_only.sql new file mode 100644 index 0000000..ecd0abe --- /dev/null +++ b/bricktracker/sql/set/list/years_only.sql @@ -0,0 +1,13 @@ +SELECT DISTINCT "rebrickable_sets"."year" +FROM "bricktracker_sets" + +INNER JOIN "rebrickable_sets" +ON "bricktracker_sets"."set" IS NOT DISTINCT FROM "rebrickable_sets"."set" + +{% block where %} +WHERE 1=1 +{% if search_query %} +AND (LOWER("rebrickable_sets"."name") LIKE LOWER('%{{ search_query }}%') + OR LOWER("rebrickable_sets"."set") LIKE LOWER('%{{ search_query }}%')) +{% endif %} +{% endblock %} \ No newline at end of file diff --git a/bricktracker/sql/statistics/overview.sql b/bricktracker/sql/statistics/overview.sql new file mode 100644 index 0000000..4c19520 --- /dev/null +++ b/bricktracker/sql/statistics/overview.sql @@ -0,0 +1,83 @@ +-- Statistics Overview Query (Optimized with CTEs) +-- Provides comprehensive statistics for BrickTracker dashboard +-- Performance improved by consolidating subqueries into CTEs +-- Expected impact: 60-80% performance improvement for dashboard loading + +WITH +-- Set statistics aggregation +set_stats AS ( + SELECT + COUNT(*) AS total_sets, + COUNT(DISTINCT "set") AS unique_sets, + COUNT(CASE WHEN "purchase_price" IS NOT NULL THEN 1 END) AS sets_with_price, + ROUND(SUM("purchase_price"), 2) AS total_cost, + ROUND(AVG("purchase_price"), 2) AS average_cost, + ROUND(MIN("purchase_price"), 2) AS minimum_cost, + ROUND(MAX("purchase_price"), 2) AS maximum_cost, + COUNT(DISTINCT CASE WHEN "storage" IS NOT NULL THEN "storage" END) AS storage_locations_used, + COUNT(DISTINCT CASE WHEN "purchase_location" IS NOT NULL THEN "purchase_location" END) AS purchase_locations_used, + COUNT(CASE WHEN "storage" IS NOT NULL THEN 1 END) AS sets_with_storage, + COUNT(CASE WHEN "purchase_location" IS NOT NULL THEN 1 END) AS sets_with_purchase_location + FROM "bricktracker_sets" +), + +-- Part statistics aggregation +part_stats AS ( + SELECT + COUNT(*) AS total_part_instances, + SUM("quantity") AS total_parts_count, + COUNT(DISTINCT "part") AS unique_parts, + SUM("missing") AS total_missing_parts, + SUM("damaged") AS total_damaged_parts + FROM "bricktracker_parts" +), + +-- Minifigure statistics aggregation +minifig_stats AS ( + SELECT + COUNT(*) AS total_minifigure_instances, + SUM("quantity") AS total_minifigures_count, + COUNT(DISTINCT "figure") AS unique_minifigures + FROM "bricktracker_minifigures" +), + +-- Rebrickable sets count (for sets we actually own) +rebrickable_stats AS ( + SELECT COUNT(*) AS unique_rebrickable_sets + FROM "rebrickable_sets" + WHERE "set" IN (SELECT DISTINCT "set" FROM "bricktracker_sets") +) + +-- Final select combining all statistics +SELECT + -- Basic counts + set_stats.total_sets, + set_stats.unique_sets, + rebrickable_stats.unique_rebrickable_sets, + + -- Parts statistics + part_stats.total_part_instances, + part_stats.total_parts_count, + part_stats.unique_parts, + part_stats.total_missing_parts, + part_stats.total_damaged_parts, + + -- Minifigures statistics + minifig_stats.total_minifigure_instances, + minifig_stats.total_minifigures_count, + minifig_stats.unique_minifigures, + + -- Financial statistics + set_stats.sets_with_price, + set_stats.total_cost, + set_stats.average_cost, + set_stats.minimum_cost, + set_stats.maximum_cost, + + -- Storage and location statistics + set_stats.storage_locations_used, + set_stats.purchase_locations_used, + set_stats.sets_with_storage, + set_stats.sets_with_purchase_location + +FROM set_stats, part_stats, minifig_stats, rebrickable_stats \ No newline at end of file diff --git a/bricktracker/sql/statistics/purchase_locations.sql b/bricktracker/sql/statistics/purchase_locations.sql new file mode 100644 index 0000000..81db813 --- /dev/null +++ b/bricktracker/sql/statistics/purchase_locations.sql @@ -0,0 +1,45 @@ +-- Purchase Location Statistics +-- Shows statistics grouped by purchase location + +SELECT + "bricktracker_sets"."purchase_location" AS "location_id", + "bricktracker_metadata_purchase_locations"."name" AS "location_name", + COUNT("bricktracker_sets"."id") AS "set_count", + COUNT(DISTINCT "bricktracker_sets"."set") AS "unique_set_count", + SUM("rebrickable_sets"."number_of_parts") AS "total_parts", + ROUND(AVG("rebrickable_sets"."number_of_parts"), 0) AS "avg_parts_per_set", + -- Financial statistics per purchase location + COUNT(CASE WHEN "bricktracker_sets"."purchase_price" IS NOT NULL THEN 1 END) AS "sets_with_price", + ROUND(SUM("bricktracker_sets"."purchase_price"), 2) AS "total_spent", + ROUND(AVG("bricktracker_sets"."purchase_price"), 2) AS "avg_price", + ROUND(MIN("bricktracker_sets"."purchase_price"), 2) AS "min_price", + ROUND(MAX("bricktracker_sets"."purchase_price"), 2) AS "max_price", + -- Date range statistics + MIN("bricktracker_sets"."purchase_date") AS "first_purchase", + MAX("bricktracker_sets"."purchase_date") AS "latest_purchase", + -- Problem statistics per purchase location + COALESCE(SUM("problem_stats"."missing_parts"), 0) AS "missing_parts", + COALESCE(SUM("problem_stats"."damaged_parts"), 0) AS "damaged_parts", + -- Minifigure statistics per purchase location + COALESCE(SUM("minifigure_stats"."minifigure_count"), 0) AS "total_minifigures" +FROM "bricktracker_sets" +INNER JOIN "rebrickable_sets" ON "bricktracker_sets"."set" = "rebrickable_sets"."set" +LEFT JOIN "bricktracker_metadata_purchase_locations" ON "bricktracker_sets"."purchase_location" = "bricktracker_metadata_purchase_locations"."id" +LEFT JOIN ( + SELECT + "bricktracker_parts"."id", + SUM("bricktracker_parts"."missing") AS "missing_parts", + SUM("bricktracker_parts"."damaged") AS "damaged_parts" + FROM "bricktracker_parts" + GROUP BY "bricktracker_parts"."id" +) "problem_stats" ON "bricktracker_sets"."id" = "problem_stats"."id" +LEFT JOIN ( + SELECT + "bricktracker_minifigures"."id", + SUM("bricktracker_minifigures"."quantity") AS "minifigure_count" + FROM "bricktracker_minifigures" + GROUP BY "bricktracker_minifigures"."id" +) "minifigure_stats" ON "bricktracker_sets"."id" = "minifigure_stats"."id" +WHERE "bricktracker_sets"."purchase_location" IS NOT NULL +GROUP BY "bricktracker_sets"."purchase_location", "bricktracker_metadata_purchase_locations"."name" +ORDER BY "set_count" DESC, "location_name" ASC \ No newline at end of file diff --git a/bricktracker/sql/statistics/purchases_by_year.sql b/bricktracker/sql/statistics/purchases_by_year.sql new file mode 100644 index 0000000..d89eeba --- /dev/null +++ b/bricktracker/sql/statistics/purchases_by_year.sql @@ -0,0 +1,49 @@ +-- Purchases by Year Statistics +-- Shows statistics grouped by purchase year (when you bought the sets) + +SELECT + strftime('%Y', datetime("bricktracker_sets"."purchase_date", 'unixepoch')) AS "purchase_year", + COUNT("bricktracker_sets"."id") AS "total_sets", + COUNT(DISTINCT "bricktracker_sets"."set") AS "unique_sets", + SUM("rebrickable_sets"."number_of_parts") AS "total_parts", + ROUND(AVG("rebrickable_sets"."number_of_parts"), 0) AS "avg_parts_per_set", + -- Financial statistics per purchase year + COUNT(CASE WHEN "bricktracker_sets"."purchase_price" IS NOT NULL THEN 1 END) AS "sets_with_price", + ROUND(SUM("bricktracker_sets"."purchase_price"), 2) AS "total_spent", + ROUND(AVG("bricktracker_sets"."purchase_price"), 2) AS "avg_price_per_set", + ROUND(MIN("bricktracker_sets"."purchase_price"), 2) AS "min_price", + ROUND(MAX("bricktracker_sets"."purchase_price"), 2) AS "max_price", + -- Release year statistics for sets purchased in this year + MIN("rebrickable_sets"."year") AS "oldest_set_year", + MAX("rebrickable_sets"."year") AS "newest_set_year", + ROUND(AVG("rebrickable_sets"."year"), 0) AS "avg_set_release_year", + -- Problem statistics per purchase year + COALESCE(SUM("problem_stats"."missing_parts"), 0) AS "missing_parts", + COALESCE(SUM("problem_stats"."damaged_parts"), 0) AS "damaged_parts", + -- Minifigure statistics per purchase year + COALESCE(SUM("minifigure_stats"."minifigure_count"), 0) AS "total_minifigures", + -- Diversity statistics per purchase year + COUNT(DISTINCT "rebrickable_sets"."theme_id") AS "unique_themes", + COUNT(DISTINCT "bricktracker_sets"."purchase_location") AS "unique_purchase_locations", + -- Monthly statistics within the year + COUNT(DISTINCT strftime('%m', datetime("bricktracker_sets"."purchase_date", 'unixepoch'))) AS "months_with_purchases" +FROM "bricktracker_sets" +INNER JOIN "rebrickable_sets" ON "bricktracker_sets"."set" = "rebrickable_sets"."set" +LEFT JOIN ( + SELECT + "bricktracker_parts"."id", + SUM("bricktracker_parts"."missing") AS "missing_parts", + SUM("bricktracker_parts"."damaged") AS "damaged_parts" + FROM "bricktracker_parts" + GROUP BY "bricktracker_parts"."id" +) "problem_stats" ON "bricktracker_sets"."id" = "problem_stats"."id" +LEFT JOIN ( + SELECT + "bricktracker_minifigures"."id", + SUM("bricktracker_minifigures"."quantity") AS "minifigure_count" + FROM "bricktracker_minifigures" + GROUP BY "bricktracker_minifigures"."id" +) "minifigure_stats" ON "bricktracker_sets"."id" = "minifigure_stats"."id" +WHERE "bricktracker_sets"."purchase_date" IS NOT NULL +GROUP BY strftime('%Y', datetime("bricktracker_sets"."purchase_date", 'unixepoch')) +ORDER BY "purchase_year" DESC \ No newline at end of file diff --git a/bricktracker/sql/statistics/sets_by_year.sql b/bricktracker/sql/statistics/sets_by_year.sql new file mode 100644 index 0000000..292cd27 --- /dev/null +++ b/bricktracker/sql/statistics/sets_by_year.sql @@ -0,0 +1,44 @@ +-- Sets by Year Statistics +-- Shows statistics grouped by LEGO set release year + +SELECT + "rebrickable_sets"."year", + COUNT("bricktracker_sets"."id") AS "total_sets", + COUNT(DISTINCT "bricktracker_sets"."set") AS "unique_sets", + SUM("rebrickable_sets"."number_of_parts") AS "total_parts", + ROUND(AVG("rebrickable_sets"."number_of_parts"), 0) AS "avg_parts_per_set", + MIN("rebrickable_sets"."number_of_parts") AS "min_parts", + MAX("rebrickable_sets"."number_of_parts") AS "max_parts", + -- Financial statistics per year (release year) + COUNT(CASE WHEN "bricktracker_sets"."purchase_price" IS NOT NULL THEN 1 END) AS "sets_with_price", + ROUND(SUM("bricktracker_sets"."purchase_price"), 2) AS "total_spent", + ROUND(AVG("bricktracker_sets"."purchase_price"), 2) AS "avg_price_per_set", + ROUND(MIN("bricktracker_sets"."purchase_price"), 2) AS "min_price", + ROUND(MAX("bricktracker_sets"."purchase_price"), 2) AS "max_price", + -- Problem statistics per year + COALESCE(SUM("problem_stats"."missing_parts"), 0) AS "missing_parts", + COALESCE(SUM("problem_stats"."damaged_parts"), 0) AS "damaged_parts", + -- Minifigure statistics per year + COALESCE(SUM("minifigure_stats"."minifigure_count"), 0) AS "total_minifigures", + -- Theme diversity per year + COUNT(DISTINCT "rebrickable_sets"."theme_id") AS "unique_themes" +FROM "bricktracker_sets" +INNER JOIN "rebrickable_sets" ON "bricktracker_sets"."set" = "rebrickable_sets"."set" +LEFT JOIN ( + SELECT + "bricktracker_parts"."id", + SUM("bricktracker_parts"."missing") AS "missing_parts", + SUM("bricktracker_parts"."damaged") AS "damaged_parts" + FROM "bricktracker_parts" + GROUP BY "bricktracker_parts"."id" +) "problem_stats" ON "bricktracker_sets"."id" = "problem_stats"."id" +LEFT JOIN ( + SELECT + "bricktracker_minifigures"."id", + SUM("bricktracker_minifigures"."quantity") AS "minifigure_count" + FROM "bricktracker_minifigures" + GROUP BY "bricktracker_minifigures"."id" +) "minifigure_stats" ON "bricktracker_sets"."id" = "minifigure_stats"."id" +WHERE "rebrickable_sets"."year" IS NOT NULL +GROUP BY "rebrickable_sets"."year" +ORDER BY "rebrickable_sets"."year" DESC \ No newline at end of file diff --git a/bricktracker/sql/statistics/storage.sql b/bricktracker/sql/statistics/storage.sql new file mode 100644 index 0000000..2d78849 --- /dev/null +++ b/bricktracker/sql/statistics/storage.sql @@ -0,0 +1,40 @@ +-- Storage Location Statistics +-- Shows statistics grouped by storage location + +SELECT + "bricktracker_sets"."storage" AS "storage_id", + "bricktracker_metadata_storages"."name" AS "storage_name", + COUNT("bricktracker_sets"."id") AS "set_count", + COUNT(DISTINCT "bricktracker_sets"."set") AS "unique_set_count", + SUM("rebrickable_sets"."number_of_parts") AS "total_parts", + ROUND(AVG("rebrickable_sets"."number_of_parts"), 0) AS "avg_parts_per_set", + -- Financial statistics per storage + COUNT(CASE WHEN "bricktracker_sets"."purchase_price" IS NOT NULL THEN 1 END) AS "sets_with_price", + ROUND(SUM("bricktracker_sets"."purchase_price"), 2) AS "total_value", + ROUND(AVG("bricktracker_sets"."purchase_price"), 2) AS "avg_price", + -- Problem statistics per storage + COALESCE(SUM("problem_stats"."missing_parts"), 0) AS "missing_parts", + COALESCE(SUM("problem_stats"."damaged_parts"), 0) AS "damaged_parts", + -- Minifigure statistics per storage + COALESCE(SUM("minifigure_stats"."minifigure_count"), 0) AS "total_minifigures" +FROM "bricktracker_sets" +INNER JOIN "rebrickable_sets" ON "bricktracker_sets"."set" = "rebrickable_sets"."set" +LEFT JOIN "bricktracker_metadata_storages" ON "bricktracker_sets"."storage" = "bricktracker_metadata_storages"."id" +LEFT JOIN ( + SELECT + "bricktracker_parts"."id", + SUM("bricktracker_parts"."missing") AS "missing_parts", + SUM("bricktracker_parts"."damaged") AS "damaged_parts" + FROM "bricktracker_parts" + GROUP BY "bricktracker_parts"."id" +) "problem_stats" ON "bricktracker_sets"."id" = "problem_stats"."id" +LEFT JOIN ( + SELECT + "bricktracker_minifigures"."id", + SUM("bricktracker_minifigures"."quantity") AS "minifigure_count" + FROM "bricktracker_minifigures" + GROUP BY "bricktracker_minifigures"."id" +) "minifigure_stats" ON "bricktracker_sets"."id" = "minifigure_stats"."id" +WHERE "bricktracker_sets"."storage" IS NOT NULL +GROUP BY "bricktracker_sets"."storage", "bricktracker_metadata_storages"."name" +ORDER BY "set_count" DESC, "storage_name" ASC \ No newline at end of file diff --git a/bricktracker/sql/statistics/themes.sql b/bricktracker/sql/statistics/themes.sql new file mode 100644 index 0000000..082e38c --- /dev/null +++ b/bricktracker/sql/statistics/themes.sql @@ -0,0 +1,39 @@ +-- Theme Distribution Statistics +-- Shows statistics grouped by theme + +SELECT + "rebrickable_sets"."theme_id", + COUNT("bricktracker_sets"."id") AS "set_count", + COUNT(DISTINCT "bricktracker_sets"."set") AS "unique_set_count", + SUM("rebrickable_sets"."number_of_parts") AS "total_parts", + ROUND(AVG("rebrickable_sets"."number_of_parts"), 0) AS "avg_parts_per_set", + MIN("rebrickable_sets"."year") AS "earliest_year", + MAX("rebrickable_sets"."year") AS "latest_year", + -- Financial statistics per theme + COUNT(CASE WHEN "bricktracker_sets"."purchase_price" IS NOT NULL THEN 1 END) AS "sets_with_price", + ROUND(SUM("bricktracker_sets"."purchase_price"), 2) AS "total_spent", + ROUND(AVG("bricktracker_sets"."purchase_price"), 2) AS "avg_price", + -- Problem statistics per theme + COALESCE(SUM("problem_stats"."missing_parts"), 0) AS "missing_parts", + COALESCE(SUM("problem_stats"."damaged_parts"), 0) AS "damaged_parts", + -- Minifigure statistics per theme + COALESCE(SUM("minifigure_stats"."minifigure_count"), 0) AS "total_minifigures" +FROM "bricktracker_sets" +INNER JOIN "rebrickable_sets" ON "bricktracker_sets"."set" = "rebrickable_sets"."set" +LEFT JOIN ( + SELECT + "bricktracker_parts"."id", + SUM("bricktracker_parts"."missing") AS "missing_parts", + SUM("bricktracker_parts"."damaged") AS "damaged_parts" + FROM "bricktracker_parts" + GROUP BY "bricktracker_parts"."id" +) "problem_stats" ON "bricktracker_sets"."id" = "problem_stats"."id" +LEFT JOIN ( + SELECT + "bricktracker_minifigures"."id", + SUM("bricktracker_minifigures"."quantity") AS "minifigure_count" + FROM "bricktracker_minifigures" + GROUP BY "bricktracker_minifigures"."id" +) "minifigure_stats" ON "bricktracker_sets"."id" = "minifigure_stats"."id" +GROUP BY "rebrickable_sets"."theme_id" +ORDER BY "set_count" DESC, "rebrickable_sets"."theme_id" ASC \ No newline at end of file diff --git a/bricktracker/statistics.py b/bricktracker/statistics.py new file mode 100644 index 0000000..e446205 --- /dev/null +++ b/bricktracker/statistics.py @@ -0,0 +1,132 @@ +""" +Statistics module for BrickTracker +Provides statistics and analytics functionality +""" + +import logging +from typing import Any + +from .sql import BrickSQL +from .theme_list import BrickThemeList + +logger = logging.getLogger(__name__) + + +class BrickStatistics: + """Main statistics class providing overview and detailed statistics""" + + def __init__(self): + self.sql = BrickSQL() + + def get_overview(self) -> dict[str, Any]: + """Get overview statistics""" + result = self.sql.fetchone('statistics/overview') + if result: + return dict(result) + return {} + + def get_theme_statistics(self) -> list[dict[str, Any]]: + """Get statistics grouped by theme with theme names""" + results = self.sql.fetchall('statistics/themes') + + # Load theme list to get theme names + theme_list = BrickThemeList() + + statistics = [] + for row in results: + stat = dict(row) + # Add theme name from theme list + theme = theme_list.get(stat['theme_id']) + stat['theme_name'] = theme.name if theme else f"Theme {stat['theme_id']}" + statistics.append(stat) + + return statistics + + def get_storage_statistics(self) -> list[dict[str, Any]]: + """Get statistics grouped by storage location""" + results = self.sql.fetchall('statistics/storage') + return [dict(row) for row in results] + + def get_purchase_location_statistics(self) -> list[dict[str, Any]]: + """Get statistics grouped by purchase location""" + results = self.sql.fetchall('statistics/purchase_locations') + return [dict(row) for row in results] + + def get_financial_summary(self) -> dict[str, Any]: + """Get financial summary from overview statistics""" + overview = self.get_overview() + return { + 'total_cost': overview.get('total_cost') or 0, + 'average_cost': overview.get('average_cost') or 0, + 'minimum_cost': overview.get('minimum_cost') or 0, + 'maximum_cost': overview.get('maximum_cost') or 0, + 'sets_with_price': overview.get('sets_with_price') or 0, + 'total_sets': overview.get('total_sets') or 0, + 'percentage_with_price': round( + ((overview.get('sets_with_price') or 0) / max((overview.get('total_sets') or 0), 1)) * 100, 1 + ) + } + + def get_collection_summary(self) -> dict[str, Any]: + """Get collection summary from overview statistics""" + overview = self.get_overview() + return { + 'total_sets': overview.get('total_sets') or 0, + 'unique_sets': overview.get('unique_sets') or 0, + 'total_parts_count': overview.get('total_parts_count') or 0, + 'unique_parts': overview.get('unique_parts') or 0, + 'total_minifigures_count': overview.get('total_minifigures_count') or 0, + 'unique_minifigures': overview.get('unique_minifigures') or 0, + 'total_missing_parts': overview.get('total_missing_parts') or 0, + 'total_damaged_parts': overview.get('total_damaged_parts') or 0, + 'storage_locations_used': overview.get('storage_locations_used') or 0, + 'purchase_locations_used': overview.get('purchase_locations_used') or 0 + } + + def get_sets_by_year_statistics(self) -> list[dict[str, Any]]: + """Get statistics grouped by LEGO set release year""" + results = self.sql.fetchall('statistics/sets_by_year') + return [dict(row) for row in results] + + def get_purchases_by_year_statistics(self) -> list[dict[str, Any]]: + """Get statistics grouped by purchase year""" + results = self.sql.fetchall('statistics/purchases_by_year') + return [dict(row) for row in results] + + def get_year_summary(self) -> dict[str, Any]: + """Get year-based summary statistics""" + sets_by_year = self.get_sets_by_year_statistics() + purchases_by_year = self.get_purchases_by_year_statistics() + + # Calculate summary metrics + years_represented = len(sets_by_year) + years_with_purchases = len(purchases_by_year) + + # Find peak year for collection (by set count) + peak_collection_year = None + max_sets_in_year = 0 + if sets_by_year: + peak_year_data = max(sets_by_year, key=lambda x: x.get('total_sets') or 0) + peak_collection_year = peak_year_data.get('year') + max_sets_in_year = peak_year_data.get('total_sets') or 0 + + # Find peak spending year + peak_spending_year = None + max_spending = 0 + if purchases_by_year: + spending_years = [y for y in purchases_by_year if y.get('total_spent')] + if spending_years: + peak_spending_data = max(spending_years, key=lambda x: x.get('total_spent') or 0) + peak_spending_year = peak_spending_data.get('purchase_year') + max_spending = peak_spending_data.get('total_spent') or 0 + + return { + 'years_represented': years_represented, + 'years_with_purchases': years_with_purchases, + 'peak_collection_year': peak_collection_year, + 'max_sets_in_year': max_sets_in_year, + 'peak_spending_year': peak_spending_year, + 'max_spending': max_spending, + 'oldest_set_year': min([y['year'] for y in sets_by_year]) if sets_by_year else None, + 'newest_set_year': max([y['year'] for y in sets_by_year]) if sets_by_year else None + } \ No newline at end of file diff --git a/bricktracker/template_filters.py b/bricktracker/template_filters.py new file mode 100644 index 0000000..0d563bb --- /dev/null +++ b/bricktracker/template_filters.py @@ -0,0 +1,13 @@ +"""Custom Jinja2 template filters for BrickTracker.""" + +from urllib.parse import urlparse, parse_qs, urlencode, urlunparse + + +def replace_query_filter(url, key, value): + """Replace or add a query parameter in a URL""" + parsed = urlparse(url) + query_dict = parse_qs(parsed.query, keep_blank_values=True) + query_dict[key] = [str(value)] + + new_query = urlencode(query_dict, doseq=True) + return urlunparse((parsed.scheme, parsed.netloc, parsed.path, parsed.params, new_query, parsed.fragment)) \ No newline at end of file diff --git a/bricktracker/version.py b/bricktracker/version.py index 422f2c5..9b5d231 100644 --- a/bricktracker/version.py +++ b/bricktracker/version.py @@ -1,4 +1,4 @@ from typing import Final -__version__: Final[str] = '1.2.4' -__database_version__: Final[int] = 17 +__version__: Final[str] = '1.3.0' +__database_version__: Final[int] = 20 diff --git a/bricktracker/views/admin/admin.py b/bricktracker/views/admin/admin.py index 9835548..41a60bd 100644 --- a/bricktracker/views/admin/admin.py +++ b/bricktracker/views/admin/admin.py @@ -1,9 +1,12 @@ import logging +import os -from flask import Blueprint, request, render_template +from flask import Blueprint, request, render_template, current_app, jsonify from flask_login import login_required from ...configuration_list import BrickConfigurationList +from ...config_manager import ConfigManager +from ...config import CONFIG from ..exceptions import exception_handler from ...instructions_list import BrickInstructionsList from ...rebrickable_image import RebrickableImage @@ -27,6 +30,94 @@ logger = logging.getLogger(__name__) admin_page = Blueprint('admin', __name__, url_prefix='/admin') +def get_env_values(): + """Get current environment values, using defaults from config when not set""" + import json + from pathlib import Path + + env_values = {} + config_defaults = {} + env_explicit_values = {} # Track which values are explicitly set + env_locked_values = {} # Track which values are set via Docker environment (locked) + + # Get list of variables that were set via Docker environment (before .env was loaded) + # This was stored by load_env_file() in app.py + docker_env_vars = set() + if '_BK_DOCKER_ENV_VARS' in os.environ: + try: + docker_env_vars = set(json.loads(os.environ['_BK_DOCKER_ENV_VARS'])) + except (json.JSONDecodeError, TypeError): + pass + + # Read .env file if it exists (check both locations) + env_file = None + if Path('data/.env').exists(): + env_file = Path('data/.env') + elif Path('.env').exists(): + env_file = Path('.env') + + env_from_file = {} + if env_file: + with open(env_file, 'r', encoding='utf-8') as f: + for line in f: + line = line.strip() + if line and not line.startswith('#') and '=' in line: + key, value = line.split('=', 1) + # Strip quotes from value when reading + value = value.strip() + if value.startswith('"') and value.endswith('"'): + value = value[1:-1] + elif value.startswith("'") and value.endswith("'"): + value = value[1:-1] + env_from_file[key] = value + + # Process each config item + for config_item in CONFIG: + env_name = f"BK_{config_item['n']}" + + # Store default value (with casting applied) + default_value = config_item.get('d', '') + if 'c' in config_item and default_value is not None: + cast_type = config_item['c'] + if cast_type == bool and default_value == '': + default_value = False # Default for booleans is False only if no default specified + elif cast_type == list and isinstance(default_value, str): + default_value = [item.strip() for item in default_value.split(',') if item.strip()] + # For int/other types, keep the original default value + config_defaults[env_name] = default_value + + # Check if value is set via Docker environment directive + # A variable is "locked" if it was in os.environ BEFORE our app loaded .env + is_locked = env_name in docker_env_vars + env_locked_values[env_name] = is_locked + + # Check if value is explicitly set in .env file or environment + is_explicitly_set = env_name in env_from_file or env_name in os.environ + env_explicit_values[env_name] = is_explicitly_set + + # Get value from .env file, environment, or default + value = env_from_file.get(env_name) or os.environ.get(env_name) + if value is None: + value = default_value + else: + # Apply casting if specified + if 'c' in config_item and value is not None: + cast_type = config_item['c'] + if cast_type == bool and isinstance(value, str): + value = value.lower() in ('true', '1', 'yes', 'on') + elif cast_type == int and value != '': + try: + value = int(value) + except (ValueError, TypeError): + value = config_item.get('d', 0) + elif cast_type == list and isinstance(value, str): + value = [item.strip() for item in value.split(',') if item.strip()] + + env_values[env_name] = value + + return env_values, config_defaults, env_explicit_values, env_locked_values + + # Admin @admin_page.route('/', methods=['GET']) @login_required @@ -102,18 +193,72 @@ def admin() -> str: open_tag ) - open_database = ( - open_image is None and - open_instructions is None and - open_logout is None and - not open_metadata and - open_retired is None and - open_theme is None + # Get configurable default expanded sections + default_expanded_sections = current_app.config.get('ADMIN_DEFAULT_EXPANDED_SECTIONS', []) + + # Helper function to check if section should be expanded + def should_expand(section_name, url_param): + # URL parameter takes priority over default config + if url_param is not None: + return url_param + # Check if section is in default expanded list + return section_name in default_expanded_sections + + # Apply configurable default expansion logic + open_database = should_expand('database', request.args.get('open_database', None)) + open_image = should_expand('image', open_image) + open_instructions = should_expand('instructions', open_instructions) + open_logout = should_expand('authentication', open_logout) + open_retired = should_expand('retired', open_retired) + open_theme = should_expand('theme', open_theme) + + # Metadata sub-sections + open_owner = should_expand('owner', open_owner) + open_purchase_location = should_expand('purchase_location', open_purchase_location) + open_status = should_expand('status', open_status) + open_storage = should_expand('storage', open_storage) + open_tag = should_expand('tag', open_tag) + + # Recalculate metadata section based on sub-sections or direct config + open_metadata = ( + should_expand('metadata', open_metadata) or + open_owner or + open_purchase_location or + open_status or + open_storage or + open_tag ) + env_values, config_defaults, env_explicit_values, env_locked_values = get_env_values() + + # Check .env file location and set warnings + env_file_location = None + env_file_warning = False + env_file_missing = False + + if os.path.exists('data/.env'): + env_file_location = 'data/.env' + env_file_warning = False + env_file_missing = False + elif os.path.exists('.env'): + env_file_location = '.env' + env_file_warning = True # Warn: changes won't persist without volume mount + env_file_missing = False + else: + env_file_location = None + env_file_warning = False + env_file_missing = True # Warn: no .env file found + return render_template( 'admin.html', configuration=BrickConfigurationList.list(), + env_values=env_values, + config_defaults=config_defaults, + env_explicit_values=env_explicit_values, + env_locked_values=env_locked_values, + env_file_location=env_file_location, + env_file_warning=env_file_warning, + env_file_missing=env_file_missing, database_counters=database_counters, database_error=request.args.get('database_error'), database_exception=database_exception, @@ -149,3 +294,103 @@ def admin() -> str: tag_error=request.args.get('tag_error'), theme=BrickThemeList(), ) + + +# API Endpoints for Configuration Management + +@admin_page.route('/api/config/update', methods=['POST']) +@login_required +@exception_handler(__file__) +def update_config() -> str: + """Update live configuration variables""" + try: + data = request.get_json() + if not data: + return jsonify({ + 'status': 'error', + 'message': 'No JSON data provided' + }), 400 + + updates = data.get('updates', {}) + if not updates: + return jsonify({ + 'status': 'error', + 'message': 'No updates provided' + }), 400 + + # Use ConfigManager to update live configuration + config_manager = ConfigManager() + results = config_manager.update_config(updates) + + # Check if all updates were successful + successful_updates = {k: v for k, v in results.items() if "successfully" in v} + failed_updates = {k: v for k, v in results.items() if "successfully" not in v} + + logger.info(f"Configuration update: {len(successful_updates)} successful, {len(failed_updates)} failed") + + if failed_updates: + logger.warning(f"Failed updates: {failed_updates}") + + return jsonify({ + 'status': 'success' if not failed_updates else 'partial', + 'results': results, + 'successful_count': len(successful_updates), + 'failed_count': len(failed_updates) + }) + + except Exception as e: + logger.error(f"Error updating configuration: {e}") + return jsonify({ + 'status': 'error', + 'message': str(e) + }), 500 + + +@admin_page.route('/api/config/update-static', methods=['POST']) +@login_required +@exception_handler(__file__) +def update_static_config() -> str: + """Update static configuration variables (requires restart)""" + try: + data = request.get_json() + if not data: + return jsonify({ + 'status': 'error', + 'message': 'No JSON data provided' + }), 400 + + updates = data.get('updates', {}) + if not updates: + return jsonify({ + 'status': 'error', + 'message': 'No updates provided' + }), 400 + + # Use ConfigManager to update .env file + config_manager = ConfigManager() + + # Update each variable in the .env file + updated_count = 0 + for var_name, value in updates.items(): + try: + config_manager._update_env_file(var_name, value) + updated_count += 1 + logger.info(f"Updated static config: {var_name}") + except Exception as e: + logger.error(f"Failed to update static config {var_name}: {e}") + raise e + + logger.info(f"Updated {updated_count} static configuration variables") + + return jsonify({ + 'status': 'success', + 'message': f'Successfully updated {updated_count} static configuration variables to .env file', + 'updated_count': updated_count + }) + + except Exception as e: + logger.error(f"Error updating static configuration: {e}") + return jsonify({ + 'status': 'error', + 'message': str(e) + }), 500 diff --git a/bricktracker/views/data.py b/bricktracker/views/data.py new file mode 100644 index 0000000..9ff1c59 --- /dev/null +++ b/bricktracker/views/data.py @@ -0,0 +1,60 @@ +import os +import logging +from flask import Blueprint, current_app, send_from_directory, abort +from werkzeug.utils import secure_filename + +logger = logging.getLogger(__name__) + +data_page = Blueprint( + 'data', + __name__, + url_prefix='/data' +) + + +@data_page.route('//') +def serve_data_file(folder: str, filename: str): + """ + Serve files from the data folder (images, PDFs, etc.) + This replaces serving these files from static/ folder. + + Security: + - Only allows serving files from configured data folders + - Uses secure_filename to prevent path traversal + - Returns 404 if file doesn't exist or folder not allowed + """ + # Secure the filename to prevent path traversal attacks + safe_filename = secure_filename(filename) + + # Get the configured data folders + allowed_folders = { + 'sets': current_app.config.get('SETS_FOLDER', './data/sets'), + 'parts': current_app.config.get('PARTS_FOLDER', './data/parts'), + 'minifigures': current_app.config.get('MINIFIGURES_FOLDER', './data/minifigures'), + 'instructions': current_app.config.get('INSTRUCTIONS_FOLDER', './data/instructions'), + } + + # Check if the requested folder is allowed + if folder not in allowed_folders: + logger.warning(f"Attempt to access unauthorized folder: {folder}") + abort(404) + + # Get the actual folder path + folder_path = allowed_folders[folder] + + # If folder_path is relative (not absolute), make it relative to app root + if not os.path.isabs(folder_path): + folder_path = os.path.join(current_app.root_path, folder_path) + + # Check if file exists + file_path = os.path.join(folder_path, safe_filename) + if not os.path.isfile(file_path): + logger.debug(f"File not found: {file_path}") + abort(404) + + # Verify the resolved path is still within the allowed folder (security check) + if not os.path.abspath(file_path).startswith(os.path.abspath(folder_path)): + logger.warning(f"Path traversal attempt detected: {filename}") + abort(404) + + return send_from_directory(folder_path, safe_filename) diff --git a/bricktracker/views/instructions.py b/bricktracker/views/instructions.py index 2c2138a..9d41ac5 100644 --- a/bricktracker/views/instructions.py +++ b/bricktracker/views/instructions.py @@ -14,6 +14,7 @@ from .exceptions import exception_handler from ..instructions import BrickInstructions from ..instructions_list import BrickInstructionsList from ..parser import parse_set +from ..peeron_instructions import PeeronInstructions from ..socket import MESSAGES from .upload import upload_helper @@ -24,6 +25,22 @@ instructions_page = Blueprint( ) +def _render_peeron_select_page(set: str) -> str: + """Helper function to render the Peeron page selection interface with cached thumbnails.""" + peeron = PeeronInstructions(set) + peeron_pages = peeron.find_pages() # This will use the cached thumbnails + current_app.logger.debug(f"[peeron_loaded] Found {len(peeron_pages)} pages for {set}") + return render_template( + 'peeron_select.html', + download=True, + pages=peeron_pages, + set=set, + path=current_app.config['SOCKET_PATH'], + namespace=current_app.config['SOCKET_NAMESPACE'], + messages=MESSAGES + ) + + # Index @instructions_page.route('/', methods=['GET']) @exception_handler(__file__) @@ -141,6 +158,10 @@ def download() -> str: except Exception: set = '' + # Check if this is a redirect after Peeron pages were loaded + if request.args.get('peeron_loaded'): + return _render_peeron_select_page(set) + return render_template( 'instructions.html', download=True, @@ -160,12 +181,50 @@ def do_download() -> str: except Exception: set = '' - return render_template( - 'instructions.html', - download=True, - instructions=BrickInstructions.find_instructions(set), - set=set, - path=current_app.config['SOCKET_PATH'], - namespace=current_app.config['SOCKET_NAMESPACE'], - messages=MESSAGES - ) + # Check if this is a redirect after Peeron pages were loaded + if request.args.get('peeron_loaded'): + return _render_peeron_select_page(set) + + # Try Rebrickable first + try: + from .instructions import BrickInstructions + rebrickable_instructions = BrickInstructions.find_instructions(set) + # Standard Rebrickable instructions found + return render_template( + 'instructions.html', + download=True, + instructions=rebrickable_instructions, + set=set, + path=current_app.config['SOCKET_PATH'], + namespace=current_app.config['SOCKET_NAMESPACE'], + messages=MESSAGES + ) + except Exception: + # Rebrickable failed, check if Peeron has instructions (without caching thumbnails yet) + try: + peeron = PeeronInstructions(set) + # Just check if pages exist, don't cache thumbnails yet + if peeron.exists(): + # Peeron has instructions - show loading interface + return render_template( + 'peeron_select.html', + download=True, + loading_peeron=True, # Flag to show loading state + set=set, + path=current_app.config['SOCKET_PATH'], + namespace=current_app.config['SOCKET_NAMESPACE'], + messages=MESSAGES + ) + else: + raise Exception("Not found on Peeron either") + except Exception: + return render_template( + 'instructions.html', + download=True, + instructions=[], + set=set, + error='No instructions found on Rebrickable or Peeron', + path=current_app.config['SOCKET_PATH'], + namespace=current_app.config['SOCKET_NAMESPACE'], + messages=MESSAGES + ) diff --git a/bricktracker/views/minifigure.py b/bricktracker/views/minifigure.py index aeabe7e..0a290a2 100644 --- a/bricktracker/views/minifigure.py +++ b/bricktracker/views/minifigure.py @@ -1,8 +1,9 @@ -from flask import Blueprint, render_template, request +from flask import Blueprint, current_app, render_template, request from .exceptions import exception_handler from ..minifigure import BrickMinifigure from ..minifigure_list import BrickMinifigureList +from ..pagination_helper import get_pagination_config, build_pagination_context, get_request_params from ..set_list import BrickSetList, set_metadata_lists from ..set_owner_list import BrickSetOwnerList @@ -13,24 +14,84 @@ minifigure_page = Blueprint('minifigure', __name__, url_prefix='/minifigures') @minifigure_page.route('/', methods=['GET']) @exception_handler(__file__) def list() -> str: - # Get owner filter from request + # Get filter parameters from request owner_id = request.args.get('owner', 'all') + problems_filter = request.args.get('problems', 'all') + theme_id = request.args.get('theme', 'all') + year = request.args.get('year', 'all') + search_query, sort_field, sort_order, page = get_request_params() - # Get minifigures filtered by owner - if owner_id == 'all' or owner_id is None or owner_id == '': - minifigures = BrickMinifigureList().all() + # Get pagination configuration + per_page, is_mobile = get_pagination_config('minifigures') + use_pagination = per_page > 0 + + if use_pagination: + # PAGINATION MODE - Server-side pagination with search + minifigures, total_count = BrickMinifigureList().all_filtered_paginated( + owner_id=owner_id, + problems_filter=problems_filter, + theme_id=theme_id, + year=year, + search_query=search_query, + page=page, + per_page=per_page, + sort_field=sort_field, + sort_order=sort_order + ) + + pagination_context = build_pagination_context(page, per_page, total_count, is_mobile) else: - minifigures = BrickMinifigureList().all_by_owner(owner_id) + # ORIGINAL MODE - Single page with all data for client-side search + if owner_id == 'all' or owner_id is None or owner_id == '': + minifigures = BrickMinifigureList().all_filtered(problems_filter=problems_filter, theme_id=theme_id, year=year) + else: + minifigures = BrickMinifigureList().all_by_owner_filtered(owner_id=owner_id, problems_filter=problems_filter, theme_id=theme_id, year=year) + + pagination_context = None # Get list of owners for filter dropdown owners = BrickSetOwnerList.list() - return render_template( - 'minifigures.html', - table_collection=minifigures, - owners=owners, - selected_owner=owner_id, - ) + # Prepare context for dependent filters + filter_context = {} + if owner_id != 'all' and owner_id: + filter_context['owner_id'] = owner_id + + # Get list of themes for filter dropdown + from ..theme_list import BrickThemeList + from ..sql import BrickSQL + theme_list = BrickThemeList() + themes_data = BrickSQL().fetchall('minifigure/themes/list', **filter_context) + themes = [] + for theme_data in themes_data: + theme = theme_list.get(theme_data['theme_id']) + themes.append({ + 'theme_id': theme_data['theme_id'], + 'theme_name': theme.name if theme else f"Theme {theme_data['theme_id']}" + }) + + # Get list of years for filter dropdown + years = BrickSQL().fetchall('minifigure/years/list', **filter_context) + + template_context = { + 'table_collection': minifigures, + 'owners': owners, + 'selected_owner': owner_id, + 'selected_problems': problems_filter, + 'themes': themes, + 'selected_theme': theme_id, + 'years': years, + 'selected_year': year, + 'search_query': search_query, + 'use_pagination': use_pagination, + 'current_sort': sort_field, + 'current_order': sort_order + } + + if pagination_context: + template_context['pagination'] = pagination_context + + return render_template('minifigures.html', **template_context) # Minifigure details diff --git a/bricktracker/views/part.py b/bricktracker/views/part.py index f8e72f2..7450f48 100644 --- a/bricktracker/views/part.py +++ b/bricktracker/views/part.py @@ -2,6 +2,7 @@ from flask import Blueprint, render_template, request from .exceptions import exception_handler from ..minifigure_list import BrickMinifigureList +from ..pagination_helper import get_pagination_config, build_pagination_context, get_request_params from ..part import BrickPart from ..part_list import BrickPartList from ..set_list import BrickSetList, set_metadata_lists @@ -15,42 +16,177 @@ part_page = Blueprint('part', __name__, url_prefix='/parts') @part_page.route('/', methods=['GET']) @exception_handler(__file__) def list() -> str: - # Get filter parameters from request owner_id = request.args.get('owner', 'all') color_id = request.args.get('color', 'all') + theme_id = request.args.get('theme', 'all') + year = request.args.get('year', 'all') + search_query, sort_field, sort_order, page = get_request_params() - # Get parts with filters applied - parts = BrickPartList().all_filtered(owner_id, color_id) + # Get pagination configuration + per_page, is_mobile = get_pagination_config('parts') + use_pagination = per_page > 0 + + if use_pagination: + # PAGINATION MODE - Server-side pagination with search + parts, total_count = BrickPartList().all_filtered_paginated( + owner_id=owner_id, + color_id=color_id, + theme_id=theme_id, + year=year, + search_query=search_query, + page=page, + per_page=per_page, + sort_field=sort_field, + sort_order=sort_order + ) + + pagination_context = build_pagination_context(page, per_page, total_count, is_mobile) + else: + # ORIGINAL MODE - Single page with all data for client-side search + parts = BrickPartList().all_filtered(owner_id, color_id, theme_id, year) + pagination_context = None # Get list of owners for filter dropdown owners = BrickSetOwnerList.list() - # Get list of colors for filter dropdown - # Prepare context for color query (filter by owner if selected) - color_context = {} + # Prepare context for dependent filters + filter_context = {} if owner_id != 'all' and owner_id: - color_context['owner_id'] = owner_id + filter_context['owner_id'] = owner_id - colors = BrickSQL().fetchall('part/colors/list', **color_context) + # Get list of colors for filter dropdown + colors = BrickSQL().fetchall('part/colors/list', **filter_context) + + # Get list of themes for filter dropdown + from ..theme_list import BrickThemeList + theme_list = BrickThemeList() + themes_data = BrickSQL().fetchall('part/themes/list', **filter_context) + themes = [] + for theme_data in themes_data: + theme = theme_list.get(theme_data['theme_id']) + themes.append({ + 'theme_id': theme_data['theme_id'], + 'theme_name': theme.name if theme else f"Theme {theme_data['theme_id']}" + }) + + # Get list of years for filter dropdown + years = BrickSQL().fetchall('part/years/list', **filter_context) + + template_context = { + 'table_collection': parts, + 'owners': owners, + 'selected_owner': owner_id, + 'colors': colors, + 'selected_color': color_id, + 'themes': themes, + 'selected_theme': theme_id, + 'years': years, + 'selected_year': year, + 'search_query': search_query, + 'use_pagination': use_pagination, + 'current_sort': sort_field, + 'current_order': sort_order + } + + if pagination_context: + template_context['pagination'] = pagination_context + + return render_template('parts.html', **template_context) - return render_template( - 'parts.html', - table_collection=parts, - owners=owners, - selected_owner=owner_id, - colors=colors, - selected_color=color_id, - ) # Problem @part_page.route('/problem', methods=['GET']) @exception_handler(__file__) def problem() -> str: + # Get filter parameters from request + owner_id = request.args.get('owner', 'all') + color_id = request.args.get('color', 'all') + theme_id = request.args.get('theme', 'all') + year = request.args.get('year', 'all') + storage_id = request.args.get('storage', 'all') + tag_id = request.args.get('tag', 'all') + search_query, sort_field, sort_order, page = get_request_params() + + # Get pagination configuration + per_page, is_mobile = get_pagination_config('problems') + use_pagination = per_page > 0 + + if use_pagination: + # PAGINATION MODE - Server-side pagination with search and filters + parts, total_count = BrickPartList().problem_paginated( + owner_id=owner_id, + color_id=color_id, + theme_id=theme_id, + year=year, + storage_id=storage_id, + tag_id=tag_id, + search_query=search_query, + page=page, + per_page=per_page, + sort_field=sort_field, + sort_order=sort_order + ) + + pagination_context = build_pagination_context(page, per_page, total_count, is_mobile) + else: + # ORIGINAL MODE - Single page with all data for client-side search + parts = BrickPartList().problem_filtered(owner_id, color_id, theme_id, year, storage_id, tag_id) + pagination_context = None + + # Get list of owners for filter dropdown + owners = BrickSetOwnerList.list() + + # Prepare context for dependent filters + filter_context = {} + if owner_id != 'all' and owner_id: + filter_context['owner_id'] = owner_id + + # Get list of colors for filter dropdown (problem parts only) + colors = BrickSQL().fetchall('part/colors/list_problem', **filter_context) + + # Get list of themes for filter dropdown (problem parts only) + from ..theme_list import BrickThemeList + theme_list = BrickThemeList() + themes_data = BrickSQL().fetchall('part/themes/list_problem', **filter_context) + themes = [] + for theme_data in themes_data: + theme = theme_list.get(theme_data['theme_id']) + themes.append({ + 'theme_id': theme_data['theme_id'], + 'theme_name': theme.name if theme else f"Theme {theme_data['theme_id']}" + }) + + # Get list of years for filter dropdown (problem parts only) + years = BrickSQL().fetchall('part/years/list_problem', **filter_context) + + # Get list of storages for filter dropdown (problem parts only) + storages = BrickSQL().fetchall('part/storages/list_problem', **filter_context) + + # Get list of tags for filter dropdown (problem parts only) + tags = BrickSQL().fetchall('part/tags/list_problem', **filter_context) + return render_template( 'problem.html', - table_collection=BrickPartList().problem() + table_collection=parts, + pagination=pagination_context, + search_query=search_query, + sort_field=sort_field, + sort_order=sort_order, + use_pagination=use_pagination, + owners=owners, + colors=colors, + selected_owner=owner_id, + selected_color=color_id, + themes=themes, + selected_theme=theme_id, + years=years, + selected_year=year, + storages=storages, + selected_storage=storage_id, + tags=tags, + selected_tag=tag_id ) diff --git a/bricktracker/views/set.py b/bricktracker/views/set.py index 1ffec55..0ac0a82 100644 --- a/bricktracker/views/set.py +++ b/bricktracker/views/set.py @@ -15,6 +15,7 @@ from werkzeug.wrappers.response import Response from .exceptions import exception_handler from ..exceptions import ErrorException from ..minifigure import BrickMinifigure +from ..pagination_helper import get_pagination_config, build_pagination_context, get_request_params from ..part import BrickPart from ..rebrickable_set import RebrickableSet from ..set import BrickSet @@ -35,12 +36,83 @@ set_page = Blueprint('set', __name__, url_prefix='/sets') @set_page.route('/', methods=['GET']) @exception_handler(__file__) def list() -> str: - return render_template( - 'sets.html', - collection=BrickSetList().all(), - brickset_statuses=BrickSetStatusList.list(), + # Get filter parameters from request + search_query, sort_field, sort_order, page = get_request_params() + + # Get filter parameters + status_filter = request.args.get('status') + theme_filter = request.args.get('theme') + owner_filter = request.args.get('owner') + purchase_location_filter = request.args.get('purchase_location') + storage_filter = request.args.get('storage') + tag_filter = request.args.get('tag') + year_filter = request.args.get('year') + duplicate_filter = request.args.get('duplicate', '').lower() == 'true' + + # Get pagination configuration + per_page, is_mobile = get_pagination_config('sets') + use_pagination = per_page > 0 + + if use_pagination: + # PAGINATION MODE - Server-side pagination with search and filters + sets, total_count = BrickSetList().all_filtered_paginated( + search_query=search_query, + page=page, + per_page=per_page, + sort_field=sort_field, + sort_order=sort_order, + status_filter=status_filter, + theme_filter=theme_filter, + owner_filter=owner_filter, + purchase_location_filter=purchase_location_filter, + storage_filter=storage_filter, + tag_filter=tag_filter, + year_filter=year_filter, + duplicate_filter=duplicate_filter, + use_consolidated=current_app.config['SETS_CONSOLIDATION'] + ) + + pagination_context = build_pagination_context(page, per_page, total_count, is_mobile) + else: + # ORIGINAL MODE - Single page with all data for client-side search + if current_app.config['SETS_CONSOLIDATION']: + sets = BrickSetList().all_consolidated() + else: + sets = BrickSetList().all() + pagination_context = None + + # Convert theme ID to theme name for dropdown display if needed + display_theme_filter = theme_filter + if theme_filter and theme_filter.isdigit(): + # Theme filter is an ID, convert to name for dropdown + # Create a fresh BrickSetList instance for theme conversion + converter = BrickSetList() + theme_name = converter._theme_id_to_name(theme_filter) + if theme_name: + display_theme_filter = theme_name + + template_context = { + 'collection': sets, + 'search_query': search_query, + 'use_pagination': use_pagination, + 'current_sort': sort_field, + 'current_order': sort_order, + 'current_status_filter': status_filter, + 'current_theme_filter': display_theme_filter, + 'current_owner_filter': owner_filter, + 'current_purchase_location_filter': purchase_location_filter, + 'current_storage_filter': storage_filter, + 'current_tag_filter': tag_filter, + 'current_year_filter': year_filter, + 'current_duplicate_filter': duplicate_filter, + 'brickset_statuses': BrickSetStatusList.list(), **set_metadata_lists(as_class=True) - ) + } + + if pagination_context: + template_context['pagination'] = pagination_context + + return render_template('sets.html', **template_context) # Change the value of purchase date @@ -187,13 +259,44 @@ def deleted(*, id: str) -> str: @set_page.route('//details', methods=['GET']) @exception_handler(__file__) def details(*, id: str) -> str: - return render_template( - 'set.html', - item=BrickSet().select_specific(id), - open_instructions=request.args.get('open_instructions'), - brickset_statuses=BrickSetStatusList.list(all=True), - **set_metadata_lists(as_class=True) - ) + # Load the specific set + item = BrickSet().select_specific(id) + + # Check if there are multiple instances of this set + all_instances = BrickSetList() + # Load all sets with metadata context for tags, owners, etc. + filter_context = { + 'owners': BrickSetOwnerList.as_columns(), + 'statuses': BrickSetStatusList.as_columns(), + 'tags': BrickSetTagList.as_columns(), + } + all_instances.list(do_theme=True, **filter_context) + + # Find all instances with the same set number + same_set_instances = [ + record for record in all_instances.records + if record.fields.set == item.fields.set + ] + + # If consolidation is enabled and multiple instances exist, show consolidated view + if current_app.config['SETS_CONSOLIDATION'] and len(same_set_instances) > 1: + return render_template( + 'set.html', + item=item, + all_instances=same_set_instances, + open_instructions=request.args.get('open_instructions'), + brickset_statuses=BrickSetStatusList.list(all=True), + **set_metadata_lists(as_class=True) + ) + else: + # Single instance or consolidation disabled, show normal view + return render_template( + 'set.html', + item=item, + open_instructions=request.args.get('open_instructions'), + brickset_statuses=BrickSetStatusList.list(all=True), + **set_metadata_lists(as_class=True) + ) # Update problematic pieces of a set @@ -242,6 +345,50 @@ def problem_part( return jsonify({problem: amount}) +# Update checked state of parts during walkthrough +@set_page.route('//parts////checked', defaults={'figure': None}, methods=['POST']) # noqa: E501 +@set_page.route('//minifigures/
/parts////checked', methods=['POST']) # noqa: E501 +@login_required +@exception_handler(__file__, json=True) +def checked_part( + *, + id: str, + figure: str | None, + part: str, + color: int, + spare: int, +) -> Response: + brickset = BrickSet().select_specific(id) + + if figure is not None: + brickminifigure = BrickMinifigure().select_specific(brickset, figure) + else: + brickminifigure = None + + brickpart = BrickPart().select_specific( + brickset, + part, + color, + spare, + minifigure=brickminifigure, + ) + + checked = brickpart.update_checked(request.json) + + # Info + logger.info('Set {set} ({id}): updated part ({part} color: {color}, spare: {spare}, minifigure: {figure}) checked state to {checked}'.format( # noqa: E501 + set=brickset.fields.set, + id=brickset.fields.id, + figure=figure, + part=brickpart.fields.part, + color=brickpart.fields.color, + spare=brickpart.fields.spare, + checked=checked + )) + + return jsonify({'checked': checked}) + + # Refresh a set @set_page.route('/refresh//', methods=['GET']) @set_page.route('//refresh', methods=['GET']) diff --git a/bricktracker/views/statistics.py b/bricktracker/views/statistics.py new file mode 100644 index 0000000..3cf6a0a --- /dev/null +++ b/bricktracker/views/statistics.py @@ -0,0 +1,194 @@ +""" +Statistics views for BrickTracker +Provides statistics and analytics pages +""" + +import logging + +from flask import Blueprint, render_template, request, url_for, redirect, current_app +from flask_login import login_required +from werkzeug.wrappers.response import Response + +from .exceptions import exception_handler +from ..statistics import BrickStatistics + +logger = logging.getLogger(__name__) + +statistics_page = Blueprint('statistics', __name__, url_prefix='/statistics') + + +@statistics_page.route('/', methods=['GET']) +@login_required +@exception_handler(__file__) +def overview() -> str: + """Statistics overview page with metrics""" + + stats = BrickStatistics() + + # Get all statistics data + overview_stats = stats.get_overview() + theme_stats = stats.get_theme_statistics() + storage_stats = stats.get_storage_statistics() + purchase_location_stats = stats.get_purchase_location_statistics() + financial_summary = stats.get_financial_summary() + collection_summary = stats.get_collection_summary() + sets_by_year_stats = stats.get_sets_by_year_statistics() + purchases_by_year_stats = stats.get_purchases_by_year_statistics() + year_summary = stats.get_year_summary() + + # Prepare chart data for visualization (only if charts are enabled) + chart_data = {} + if current_app.config['STATISTICS_SHOW_CHARTS']: + chart_data = prepare_chart_data(sets_by_year_stats, purchases_by_year_stats) + + # Get filter parameters for clickable statistics + filter_type = request.args.get('filter_type') + filter_value = request.args.get('filter_value') + + # If a filter is applied, redirect to sets page with appropriate filters + if filter_type and filter_value: + return redirect_to_filtered_sets(filter_type, filter_value) + + return render_template( + 'statistics.html', + overview=overview_stats, + theme_statistics=theme_stats, + storage_statistics=storage_stats, + purchase_location_statistics=purchase_location_stats, + financial_summary=financial_summary, + collection_summary=collection_summary, + sets_by_year_statistics=sets_by_year_stats, + purchases_by_year_statistics=purchases_by_year_stats, + year_summary=year_summary, + chart_data=chart_data, + title="Statistics Overview" + ) + + +def redirect_to_filtered_sets(filter_type: str, filter_value: str) -> Response: + """Redirect to sets page with appropriate filters based on statistics click""" + + # Map filter types to sets page parameters + filter_mapping = { + 'theme': {'theme': filter_value}, + 'storage': {'storage': filter_value}, + 'purchase_location': {'purchase_location': filter_value}, + 'has_price': {'has_price': '1'} if filter_value == '1' else {}, + 'missing_parts': {'status': 'has-missing'}, + 'damaged_parts': {'status': 'has-damaged'}, + 'has_storage': {'status': 'has-storage'}, + 'no_storage': {'status': '-has-storage'}, + } + + # Get the appropriate filter parameters + filter_params = filter_mapping.get(filter_type, {}) + + if filter_params: + return redirect(url_for('set.list', **filter_params)) + else: + # Default fallback to sets page + return redirect(url_for('set.list')) + + +@statistics_page.route('/themes', methods=['GET']) +@login_required +@exception_handler(__file__) +def themes() -> str: + """Detailed theme statistics page""" + + stats = BrickStatistics() + theme_stats = stats.get_theme_statistics() + + return render_template( + 'statistics_themes.html', + theme_statistics=theme_stats, + title="Theme Statistics" + ) + + +@statistics_page.route('/storage', methods=['GET']) +@login_required +@exception_handler(__file__) +def storage() -> str: + """Detailed storage statistics page""" + + stats = BrickStatistics() + storage_stats = stats.get_storage_statistics() + + return render_template( + 'statistics_storage.html', + storage_statistics=storage_stats, + title="Storage Statistics" + ) + + +@statistics_page.route('/purchase-locations', methods=['GET']) +@login_required +@exception_handler(__file__) +def purchase_locations() -> str: + """Detailed purchase location statistics page""" + + stats = BrickStatistics() + purchase_stats = stats.get_purchase_location_statistics() + + return render_template( + 'statistics_purchase_locations.html', + purchase_location_statistics=purchase_stats, + title="Purchase Location Statistics" + ) + + +def prepare_chart_data(sets_by_year_stats, purchases_by_year_stats): + """Prepare data for Chart.js visualization""" + import json + + # Get all years from both datasets + all_years = set() + + # Add years from sets by year + if sets_by_year_stats: + for year_stat in sets_by_year_stats: + if 'year' in year_stat: + all_years.add(year_stat['year']) + + # Add years from purchases by year + if purchases_by_year_stats: + for year_stat in purchases_by_year_stats: + if 'purchase_year' in year_stat: + all_years.add(int(year_stat['purchase_year'])) + + # Create sorted list of years + years = sorted(list(all_years)) + + # Initialize data arrays + sets_data = [] + parts_data = [] + minifigs_data = [] + + # Create lookup dictionaries for quick access + sets_by_year_lookup = {} + if sets_by_year_stats: + for year_stat in sets_by_year_stats: + if 'year' in year_stat: + sets_by_year_lookup[year_stat['year']] = year_stat + + # Fill data arrays + for year in years: + # Get sets and parts data from sets_by_year + year_data = sets_by_year_lookup.get(year) + if year_data: + sets_data.append(year_data.get('total_sets') or 0) + parts_data.append(year_data.get('total_parts') or 0) + # Use actual minifigure count from the database + minifigs_data.append(year_data.get('total_minifigures') or 0) + else: + sets_data.append(0) + parts_data.append(0) + minifigs_data.append(0) + + return { + 'years': json.dumps(years), + 'sets_data': json.dumps(sets_data), + 'parts_data': json.dumps(parts_data), + 'minifigs_data': json.dumps(minifigs_data) + } \ No newline at end of file diff --git a/bricktracker/wish_list.py b/bricktracker/wish_list.py index d3038b8..e3a31ac 100644 --- a/bricktracker/wish_list.py +++ b/bricktracker/wish_list.py @@ -19,17 +19,24 @@ class BrickWishList(BrickRecordList[BrickWish]): # Queries select_query: str = 'wish/list/all' - # All the wished sets - def all(self, /) -> Self: + # Implementation of abstract list method + def list(self, /, *, override_query: str | None = None, **context) -> None: + # Use provided order or default + order = context.pop('order', current_app.config['WISHES_DEFAULT_ORDER']) + # Load the wished sets from the database for record in self.select( - order=current_app.config['WISHES_DEFAULT_ORDER'], + override_query=override_query, + order=order, owners=BrickWishOwnerList.as_columns(), + **context ): brickwish = BrickWish(record=record) - self.records.append(brickwish) + # All the wished sets + def all(self, /) -> Self: + self.list() return self # Add a set to the wishlist diff --git a/compose.local.yaml b/compose.local.yaml index bf7a8d6..dc4bfb4 100644 --- a/compose.local.yaml +++ b/compose.local.yaml @@ -2,21 +2,15 @@ services: bricktracker: container_name: BrickTracker restart: unless-stopped + # image: gitea.baerentsen.space/frederikbaerentsen/bricktracker:dev build: . ports: - "3334:3333" volumes: - - ./local:/local - - ./local/instructions:/app/static/instructions/ - - ./local/minifigures:/app/static/minifigures/ - - ./local/parts:/app/static/parts/ - - ./local/sets:/app/static/sets/ + - ./local:/app/data # Changed from ./local to ./data for consistency environment: - BK_DEBUG: true - BK_DATABASE_PATH: /local/app.db - BK_INSTRUCTIONS_FOLDER: instructions - BK_MINIFIGURES_FOLDER: minifigures - BK_PARTS_FOLDER: parts - BK_RETIRED_SETS_PATH: /local/retired_sets.csv - BK_SETS_FOLDER: sets - BK_THEMES_PATH: /local/themes.csv + - BK_DEBUG=true + # For local development, place .env in data/ folder + # The app automatically detects and uses data/.env (no env_file needed) + # Uncomment below only if you keep .env in root for backward compatibility + # env_file: .env diff --git a/compose.yaml b/compose.yaml index 535c64f..b3bfdbf 100644 --- a/compose.yaml +++ b/compose.yaml @@ -2,26 +2,14 @@ services: bricktracker: container_name: BrickTracker restart: unless-stopped - image: gitea.baerentsen.space/frederikbaerentsen/bricktracker:1.2.2 + image: gitea.baerentsen.space/frederikbaerentsen/bricktracker:latest ports: - "3333:3333" volumes: - - data:/data/ - - instructions:/app/static/instructions/ - - minifigures:/app/static/minifigures/ - - parts:/app/static/parts/ - - sets:/app/static/sets/ - # Or define those in your .env file - environment: - BK_DATABASE_PATH: /data/app.db - BK_MINIFIGURES_FOLDER: minifigures - BK_RETIRED_SETS_PATH: /data/retired_sets.csv - BK_THEMES_PATH: /data/themes.csv - env_file: ".env" + - ./data:/app/data/ + # Configuration can be done via .env file + # For new installations, place .env in data/ folder for persistence + # For backward compatibility, .env in root is also supported + # The app automatically detects and uses data/.env (priority) or .env (fallback) + # env_file: ".env" # Optional: Only needed if keeping .env in root for backward compatibility -volumes: - data: - instructions: - minifigures: - parts: - sets: diff --git a/docs/env.md b/docs/env.md index 4222ea3..b26a3f7 100644 --- a/docs/env.md +++ b/docs/env.md @@ -66,7 +66,7 @@ | Variable | Purpose | Default | Required | |----------|---------|----------|-----------| | `BK_INSTRUCTIONS_FOLDER` | Instructions storage path | `instructions` | No | -| `BK_MINIFIGURES_FOLDER` | Minifigures storage path | `minifigs` | No | +| `BK_MINIFIGURES_FOLDER` | Minifigures storage path | `minifigures` | No | | `BK_PARTS_FOLDER` | Parts storage path | `parts` | No | | `BK_SETS_FOLDER` | Sets storage path | `sets` | No | | `BK_INSTRUCTIONS_ALLOWED_EXTENSIONS` | Allowed instruction file types | `.pdf` | No | diff --git a/docs/images/1-3-new-01.png b/docs/images/1-3-new-01.png new file mode 100644 index 0000000..9166fb4 Binary files /dev/null and b/docs/images/1-3-new-01.png differ diff --git a/docs/images/1-3-new-02.png b/docs/images/1-3-new-02.png new file mode 100644 index 0000000..037045d Binary files /dev/null and b/docs/images/1-3-new-02.png differ diff --git a/docs/images/1-3-new-03.png b/docs/images/1-3-new-03.png new file mode 100644 index 0000000..8a536f2 Binary files /dev/null and b/docs/images/1-3-new-03.png differ diff --git a/docs/images/1-3-new-04.png b/docs/images/1-3-new-04.png new file mode 100644 index 0000000..c52c315 Binary files /dev/null and b/docs/images/1-3-new-04.png differ diff --git a/docs/images/1-3-new-05.png b/docs/images/1-3-new-05.png new file mode 100644 index 0000000..07a039e Binary files /dev/null and b/docs/images/1-3-new-05.png differ diff --git a/docs/images/1-3-new-06.png b/docs/images/1-3-new-06.png new file mode 100644 index 0000000..92ed64e Binary files /dev/null and b/docs/images/1-3-new-06.png differ diff --git a/docs/images/1-3-new-07.png b/docs/images/1-3-new-07.png new file mode 100644 index 0000000..2a8fd15 Binary files /dev/null and b/docs/images/1-3-new-07.png differ diff --git a/docs/images/1-3-new-08.png b/docs/images/1-3-new-08.png new file mode 100644 index 0000000..2f1fc04 Binary files /dev/null and b/docs/images/1-3-new-08.png differ diff --git a/docs/images/1-3-new-09.png b/docs/images/1-3-new-09.png new file mode 100644 index 0000000..d50f064 Binary files /dev/null and b/docs/images/1-3-new-09.png differ diff --git a/docs/images/1-3-new-10.png b/docs/images/1-3-new-10.png new file mode 100644 index 0000000..eb9868a Binary files /dev/null and b/docs/images/1-3-new-10.png differ diff --git a/docs/images/inst-dl-02.png b/docs/images/inst-dl-02.png new file mode 100644 index 0000000..e776daf Binary files /dev/null and b/docs/images/inst-dl-02.png differ diff --git a/docs/images/inst-dl.png b/docs/images/inst-dl.png new file mode 100644 index 0000000..ad38888 Binary files /dev/null and b/docs/images/inst-dl.png differ diff --git a/docs/migration_guide.md b/docs/migration_guide.md new file mode 100644 index 0000000..bdeef28 --- /dev/null +++ b/docs/migration_guide.md @@ -0,0 +1,353 @@ +# Data Folder Migration Guide (Docker Compose) + +## Overview + +Starting with version 1.3, BrickTracker consolidates all user data into a single `data/` folder for easier backup, persistence, and volume mapping. + +**This guide assumes you are running BrickTracker using Docker Compose with bind mounts.** + +> **Note:** If you're using Docker named volumes instead of bind mounts, you'll need to manually copy data between volumes. The commands below are specific to bind mount setups. + +**Backup your data before to making any changes!** + +## What Changed? + +### New Default Structure (v1.3+) + +**All relative paths are resolved relative to `/app` inside the container.** Previously all paths were relative to `/app/static`. + +For example: `data/app.db` → `/app/data/app.db` + +``` +Container (/app/): +├── data/ # NEW: Single volume mount for all user data +│ ├── .env # Configuration (recommended location) +│ ├── app.db # Database +│ ├── retired_sets.csv # Downloaded CSV files +│ ├── themes.csv +│ ├── sets/ # Set images +│ ├── parts/ # Part images +│ ├── minifigures/ # Minifigure images +│ └── instructions/ # PDF instructions +└── static/ # App assets + ├── brick.png + ├── styles.css + └── scripts/ +``` + +**Docker Compose volume:** Single mount `./data:/app/data/` + +### Previous Structure (v1.2 and earlier) + +``` +Container (/app/): +├── app.db # Mounted from ./data/ on host +├── retired_sets.csv # Mounted from ./data/ on host +├── themes.csv +└── static/ + ├── instructions/ # Separate bind mount + ├── minifigs/ # Separate bind mount + ├── parts/ # Separate bind mount + ├── sets/ # Separate bind mount +``` + +**Docker Compose bind mounts:** 5 separate mounts +```yaml +volumes: + - ./data:/app/ + - ./instructions:/app/static/instructions/ + - ./minifigs:/app/static/minifigs/ + - ./parts:/app/static/parts/ + - ./sets:/app/static/sets/ +``` + +## Migration Options + +> **Warning** +> Backup your data before making any changes! + +### Option 1: Migrate to New Data Folder Structure (Recommended) + +This is the recommended approach for cleaner backups and simpler bind mount management. + +1. **Stop the container:** + ```bash + docker compose down + ``` + +2. **Create new consolidated data directory on host:** + ```bash + mkdir -p ./bricktracker-data/{sets,parts,minifigures,instructions} + ``` + +3. **Move data from old bind mount locations to new structure:** + + Assuming your old `compose.yaml` had: + - `./data:/app/` (contains app.db, retired_sets.csv, themes.csv) + - `./instructions:/app/static/instructions/` + - `./minifigs:/app/static/minifigs/` + - `./parts:/app/static/parts/` + - `./sets:/app/static/sets/` + + > Default location for minifigures changed from `minifigs` to `minifigures` + + Run: + ```bash + # Move configuration file (optional but recommended) + mv .env ./bricktracker-data/.env + + # Move database and CSV files + mv ./data/app.db ./bricktracker-data/ + mv ./data/retired_sets.csv ./bricktracker-data/ + mv ./data/themes.csv ./bricktracker-data/ + + # Move image and instruction folders + mv ./instructions/* ./bricktracker-data/instructions/ + mv ./minifigs/* ./bricktracker-data/minifigures/ + mv ./parts/* ./bricktracker-data/parts/ + mv ./sets/* ./bricktracker-data/sets/ + ``` + +4. **Update `compose.yaml` to use single bind mount:** + ```yaml + services: + bricktracker: + volumes: + - ./bricktracker-data:/app/data/ + + # Remove old volume mounts and env_file (if .env was moved to data/) + ``` + +5. **Remove old path overrides from `.env` (if present):** + Delete any lines starting with: + - `BK_DATABASE_PATH=` + - `BK_INSTRUCTIONS_FOLDER=` + - `BK_MINIFIGURES_FOLDER=` + - `BK_PARTS_FOLDER=` + - `BK_SETS_FOLDER=` + - `BK_RETIRED_SETS_PATH=` + - `BK_THEMES_PATH=` + +6. **Start the container:** + ```bash + docker compose up -d + ``` + +7. **Verify everything works:** + ```bash + docker compose logs -f bricktracker + # Check the web interface to ensure images/data are loading + ``` + +8. **Clean up old directories (after verification):** + ```bash + rm -r ./data ./instructions ./minifigs ./parts ./sets + ``` + +### Option 2: Keep Current Setup (No Data Migration) + +If you want to keep your current volume structure without moving any files: + +1. **Add these environment variables to your `.env` file:** + + ```env + # Keep database and CSV files in /data volume (old location) + BK_DATABASE_PATH=app.db + BK_RETIRED_SETS_PATH=retired_sets.csv + BK_THEMES_PATH=themes.csv + + # Keep image/instruction folders in static/ (old location) + BK_INSTRUCTIONS_FOLDER=static/instructions + BK_MINIFIGURES_FOLDER=static/minifigures + BK_PARTS_FOLDER=static/parts + BK_SETS_FOLDER=static/sets + ``` + +2. **Keep your existing volume mounts in `compose.yaml`:** + + ```yaml + volumes: + - ./data:/app/ + - ./instructions:/app/static/instructions/ + - ./minifigures:/app/static/minifigures/ + - ./parts:/app/static/parts/ + - ./sets:/app/static/sets/ + ``` + +3. **Update to v1.3 and restart:** + ```bash + docker compose pull + docker compose up -d + ``` + +That's it! Your existing setup will continue to work. + +## Configuration File (.env) Location + +### New Behavior (v1.3+) + +BrickTracker now supports `.env` in two locations with automatic detection: + +1. **data/.env** (recommended - new location) + - Included in data volume backup + - Settings persist when changed via admin panel + - Priority location (checked first) + - **No `env_file` needed** - app reads it directly from `/app/data/.env` + +2. **.env** (backward compatibility - root) + - Continues to work for existing installations + - Requires `env_file: .env` in compose.yaml for Docker to load it at startup + - Not included in data volume (unless you add `.env` to `data/`) + +### Migration Steps for .env + +**Option A: Move to data folder (recommended)** + +```bash +# Move .env to data folder +mv .env data/.env + +# Update compose.yaml - remove or comment out env_file +# The app will automatically find and use /app/data/.env +``` + +**Option B: Keep in root (backward compatible)** + +```bash +# No changes needed +# Keep env_file: .env in compose.yaml +# App will use .env from root as fallback +``` + +**Note:** The application automatically detects which location has the .env file at runtime. No Docker Compose `env_file` directive is needed for `data/.env` because the app reads it directly from the mounted volume. + +## Configuration Reference + +### New Default Paths (v1.3+) + +All paths are relative to `/app` inside the container. + +| Config Variable | Default Value | Resolves To (Container) | Description | +|----------------|---------------|------------------------|-------------| +| `BK_DATABASE_PATH` | `data/app.db` | `/app/data/app.db` | Database file | +| `BK_RETIRED_SETS_PATH` | `data/retired_sets.csv` | `/app/data/retired_sets.csv` | Retired sets CSV | +| `BK_THEMES_PATH` | `data/themes.csv` | `/app/data/themes.csv` | Themes CSV | +| `BK_INSTRUCTIONS_FOLDER` | `data/instructions` | `/app/data/instructions` | PDF instructions | +| `BK_MINIFIGURES_FOLDER` | `data/minifigures` | `/app/data/minifigures` | Minifigure images | +| `BK_PARTS_FOLDER` | `data/parts` | `/app/data/parts` | Part images | +| `BK_SETS_FOLDER` | `data/sets` | `/app/data/sets` | Set images | + +**Docker Compose bind mount:** `./bricktracker-data:/app/data/` (single mount) + +### Old Paths (v1.2 and earlier) + +To preserve old volume structure without migration, add to `.env`: + +| Config Variable | Value to Preserve Old Behavior | Resolves To (Container) | +|----------------|-------------------------------|------------------------| +| `BK_DATABASE_PATH` | `app.db` | `/app/app.db` | +| `BK_RETIRED_SETS_PATH` | `retired_sets.csv` | `/app/retired_sets.csv` | +| `BK_THEMES_PATH` | `themes.csv` | `/app/themes.csv` | +| `BK_INSTRUCTIONS_FOLDER` | `static/instructions` | `/app/static/instructions` | +| `BK_MINIFIGURES_FOLDER` | `static/minifigs` | `/app/static/minifigs` | +| `BK_PARTS_FOLDER` | `static/parts` | `/app/static/parts` | +| `BK_SETS_FOLDER` | `static/sets` | `/app/static/sets` | + +## Benefits of New Structure + +1. **Single Bind Mount**: One `./bricktracker-data:/app/data/` mount instead of five separate mounts +2. **Easier Backups**: All user data in one location - just backup the `bricktracker-data` directory +3. **Cleaner Separation**: User data separated from application assets +4. **Better Portability**: Migrate between systems by copying/moving single directory + +## Troubleshooting + +### Images/Instructions Not Loading After Migration + +1. **Check if data was copied correctly:** + ```bash + docker compose exec bricktracker ls -la /app/data/ + docker compose exec bricktracker ls -la /app/data/sets/ + docker compose exec bricktracker ls -la /app/data/instructions/ + ``` + +2. **Verify bind mount:** + ```bash + docker compose config + # Should show: volumes: - ./bricktracker-data:/app/data/ + ``` + +3. **Check logs for path errors:** + ```bash + docker compose logs -f + ``` + +4. **Verify no old environment overrides:** + ```bash + cat .env | grep BK_ + ``` + +### Database Not Found + +1. **Check database file location in container:** + ```bash + docker compose exec bricktracker ls -la /app/data/app.db + ``` + +2. **If using old setup, verify environment variables:** + ```bash + docker compose exec bricktracker env | grep BK_DATABASE_PATH + ``` + +3. **Check host directory contains database:** + ```bash + ls -la ./bricktracker-data/ + # Should show: app.db, retired_sets.csv, themes.csv, and subdirectories + ``` + +### Settings Don't Persist After Restart + +**Error:** Admin panel changes revert after `docker compose restart` + +**Solution:** + +This happens when `.env` is not in a volume. Choose one: + +**Option A: Move .env to data folder** +```bash +mv .env data/.env +# Update compose.yaml - remove or comment out env_file +# The app will automatically find and use /app/data/.env +``` + +**Option B: Mount .env as volume** +```yaml +volumes: + - ./.env:/app/.env +``` + +### Permission Errors + +If you see permission errors after migration: + +```bash +# Fix permissions on bind-mounted directory +sudo chown -R $(id -u):$(id -g) ./bricktracker-data +``` + +**Permission denied writing .env:** + +If the admin panel shows an error when saving settings: + +1. Ensure .env file is writable by container user +2. If using volume mount, check host file permissions +3. In container: `docker exec BrickTracker ls -la /app/.env` or `/app/data/.env` + +### Reverting Migration + +If you need to revert to the old structure: + +1. Stop the container: `docker compose down` +2. Restore old `compose.yaml` with 5 volume mounts +3. Add old path environment variables to `.env` (see Option 1) +4. Start: `docker compose up -d` diff --git a/entrypoint.sh b/entrypoint.sh index 5570eee..92751e2 100755 --- a/entrypoint.sh +++ b/entrypoint.sh @@ -13,4 +13,4 @@ then fi # Execute the WSGI server -gunicorn --bind "${BK_HOST}:${BK_PORT}" "wsgi:application" --worker-class "gevent" --workers 1 "$@" +exec gunicorn --bind "${BK_HOST}:${BK_PORT}" "wsgi:application" --worker-class "gevent" --workers 1 "$@" diff --git a/requirements.txt b/requirements.txt index 7cd3644..aec142a 100644 --- a/requirements.txt +++ b/requirements.txt @@ -10,4 +10,5 @@ rebrick requests tzdata bs4 -cloudscraper +fpdf2 +pillow \ No newline at end of file diff --git a/static/brick.png b/static/brick.png index ee1b452..725e466 100644 Binary files a/static/brick.png and b/static/brick.png differ diff --git a/static/brick2.png b/static/brick2.png new file mode 100644 index 0000000..ee1b452 Binary files /dev/null and b/static/brick2.png differ diff --git a/static/scripts/admin_config.js b/static/scripts/admin_config.js new file mode 100644 index 0000000..640c5e9 --- /dev/null +++ b/static/scripts/admin_config.js @@ -0,0 +1,325 @@ +// Admin Configuration Management +// Handles live environment variable configuration interface + +// Initialize form values with current configuration +function initializeConfigValues() { + console.log('Initializing config values with:', window.CURRENT_CONFIG); + + Object.keys(window.CURRENT_CONFIG).forEach(varName => { + const value = window.CURRENT_CONFIG[varName]; + console.log(`Setting ${varName} = ${value}`); + + // Handle live settings (checkboxes and inputs) + const liveToggle = document.getElementById(varName); + if (liveToggle && liveToggle.type === 'checkbox') { + liveToggle.checked = value === true; + console.log(`Set checkbox ${varName} to ${value}`); + } + + const liveInputs = document.querySelectorAll(`input[data-var="${varName}"]:not(.config-static)`); + liveInputs.forEach(input => { + if (input.type !== 'checkbox') { + input.value = value !== null && value !== undefined ? value : ''; + console.log(`Set input ${varName} to ${input.value}`); + } + }); + + // Handle static settings + const staticToggle = document.getElementById(`static-${varName}`); + if (staticToggle && staticToggle.type === 'checkbox') { + staticToggle.checked = value === true; + console.log(`Set static checkbox ${varName} to ${value}`); + } + + const staticInputs = document.querySelectorAll(`input[data-var="${varName}"].config-static`); + staticInputs.forEach(input => { + if (input.type !== 'checkbox') { + input.value = value !== null && value !== undefined ? value : ''; + console.log(`Set static input ${varName} to ${input.value}`); + } + }); + }); +} + +// Handle config change events +function handleConfigChange(element) { + const varName = element.dataset.var; + let newValue; + + if (element.type === 'checkbox') { + newValue = element.checked; + } else if (element.type === 'number') { + newValue = parseInt(element.value) || 0; + } else { + newValue = element.value; + } + + // Update the badge display + updateConfigBadge(varName, newValue); + + // Note: Changes are only saved when "Save All Changes" button is clicked +} + +// Update badge display +function updateConfigBadge(varName, value) { + const defaultValue = window.DEFAULT_CONFIG[varName]; + const isChanged = JSON.stringify(value) !== JSON.stringify(defaultValue); + + // Remove existing badges but keep them inline + const existingBadges = document.querySelectorAll(`[data-badge-var="${varName}"]`); + existingBadges.forEach(badge => { + badge.remove(); + }); + + // Find the label where we should insert new badges + const label = document.querySelector(`label[for="${varName}"], label[for="static-${varName}"]`); + if (!label) return; + + // Find the description div (with .text-muted class) to insert badges before it + const descriptionDiv = label.querySelector('.text-muted'); + + // Create value badge based on new logic + let valueBadge; + if (value === true) { + valueBadge = document.createElement('span'); + valueBadge.className = 'badge rounded-pill text-bg-success ms-2'; + valueBadge.textContent = 'True'; + valueBadge.setAttribute('data-badge-var', varName); + valueBadge.setAttribute('data-badge-type', 'value'); + } else if (value === false) { + valueBadge = document.createElement('span'); + valueBadge.className = 'badge rounded-pill text-bg-danger ms-2'; + valueBadge.textContent = 'False'; + valueBadge.setAttribute('data-badge-var', varName); + valueBadge.setAttribute('data-badge-type', 'value'); + } else if (JSON.stringify(value) === JSON.stringify(defaultValue)) { + valueBadge = document.createElement('span'); + valueBadge.className = 'badge rounded-pill text-bg-light text-dark ms-2'; + valueBadge.textContent = `Default: ${defaultValue}`; + valueBadge.setAttribute('data-badge-var', varName); + valueBadge.setAttribute('data-badge-type', 'value'); + } else { + // For text/number fields that have been changed, show "Default: X" + valueBadge = document.createElement('span'); + valueBadge.className = 'badge rounded-pill text-bg-light text-dark ms-2'; + valueBadge.textContent = `Default: ${defaultValue}`; + valueBadge.setAttribute('data-badge-var', varName); + valueBadge.setAttribute('data-badge-type', 'value'); + } + + // Insert badge before the description div (to keep it on same line as title) + if (descriptionDiv) { + label.insertBefore(valueBadge, descriptionDiv); + } else { + label.appendChild(valueBadge); + } + + // Add changed badge if needed + if (isChanged) { + const changedBadge = document.createElement('span'); + changedBadge.className = 'badge rounded-pill text-bg-warning ms-1'; + changedBadge.textContent = 'Changed'; + changedBadge.setAttribute('data-badge-var', varName); + changedBadge.setAttribute('data-badge-type', 'changed'); + + // Insert changed badge after the value badge + if (descriptionDiv) { + label.insertBefore(changedBadge, descriptionDiv); + } else { + label.appendChild(changedBadge); + } + } +} + +// Handle static config save +function saveStaticConfig() { + const staticInputs = document.querySelectorAll('.config-static, .config-static-toggle'); + const updates = {}; + + staticInputs.forEach(input => { + const varName = input.dataset.var; + let value; + + if (input.type === 'checkbox') { + value = input.checked; + } else { + value = input.value; + } + + updates[varName] = value; + }); + + console.log('Saving static config:', updates); + + // Send to backend via fetch API + fetch('/admin/api/config/update-static', { + method: 'POST', + headers: { + 'Content-Type': 'application/json', + }, + body: JSON.stringify({ updates: updates }) + }) + .then(response => response.json()) + .then(data => { + const statusContainer = document.getElementById('config-status'); + if (statusContainer) { + if (data.status === 'success') { + statusContainer.innerHTML = '
Static configuration saved to .env file!
'; + setTimeout(() => { + statusContainer.innerHTML = ''; + }, 3000); + } else { + statusContainer.innerHTML = `
Error: ${data.message || 'Failed to save static configuration'}
`; + } + } + }) + .catch(error => { + console.error('Save static config error:', error); + const statusContainer = document.getElementById('config-status'); + if (statusContainer) { + statusContainer.innerHTML = '
Error: Failed to save static configuration
'; + } + }); +} + +// Handle button functionality +function setupButtonHandlers() { + // Save All Changes button + const saveAllBtn = document.getElementById('config-save-all'); + if (saveAllBtn) { + saveAllBtn.addEventListener('click', () => { + console.log('Save All Changes clicked'); + saveLiveConfiguration(); + }); + } + + // Refresh button + const refreshBtn = document.getElementById('config-refresh'); + if (refreshBtn) { + refreshBtn.addEventListener('click', () => { + console.log('Refresh clicked'); + location.reload(); + }); + } + + // Reset button + const resetBtn = document.getElementById('config-reset'); + if (resetBtn) { + resetBtn.addEventListener('click', () => { + console.log('Reset clicked'); + if (confirm('Are you sure you want to reset all settings to default values? This action cannot be undone.')) { + resetToDefaults(); + } + }); + } + + // Static config save button + const saveStaticBtn = document.getElementById('config-save-static'); + if (saveStaticBtn) { + saveStaticBtn.addEventListener('click', saveStaticConfig); + } +} + +// Save live configuration changes +function saveLiveConfiguration() { + const liveInputs = document.querySelectorAll('.config-toggle, .config-number, .config-text'); + const updates = {}; + + liveInputs.forEach(input => { + const varName = input.dataset.var; + let value; + + if (input.type === 'checkbox') { + value = input.checked; + } else if (input.type === 'number') { + value = parseInt(input.value) || 0; + } else { + value = input.value; + } + + updates[varName] = value; + }); + + console.log('Saving live configuration:', updates); + + // Show status message + const statusContainer = document.getElementById('config-status'); + if (statusContainer) { + statusContainer.innerHTML = '
Saving configuration...
'; + } + + // Send to backend via fetch API + fetch('/admin/api/config/update', { + method: 'POST', + headers: { + 'Content-Type': 'application/json', + }, + body: JSON.stringify({ updates: updates }) + }) + .then(response => response.json()) + .then(data => { + if (statusContainer) { + if (data.status === 'success') { + statusContainer.innerHTML = '
Configuration saved successfully! Reloading page...
'; + + // Reload the page after a short delay + setTimeout(() => { + location.reload(); + }, 1000); + } else { + statusContainer.innerHTML = `
Error: ${data.message || 'Failed to save configuration'}
`; + } + } + }) + .catch(error => { + console.error('Save error:', error); + if (statusContainer) { + statusContainer.innerHTML = '
Error: Failed to save configuration
'; + } + }); +} + +// Reset all settings to defaults +function resetToDefaults() { + console.log('Resetting to defaults'); + + // Reset all form inputs + document.querySelectorAll('.config-toggle, .config-number, .config-text').forEach(input => { + if (input.type === 'checkbox') { + input.checked = false; + } else { + input.value = ''; + } + }); + + // Update badges + Object.keys(window.CURRENT_CONFIG).forEach(varName => { + updateConfigBadge(varName, null); + }); + + // Show status message + const statusContainer = document.getElementById('config-status'); + if (statusContainer) { + statusContainer.innerHTML = '
Settings reset to defaults. Click "Save All Changes" to apply.
'; + } +} + +// Initialize when DOM is ready +document.addEventListener("DOMContentLoaded", () => { + console.log('DOM loaded, initializing configuration interface'); + + // Initialize form values + initializeConfigValues(); + + // Setup button handlers + setupButtonHandlers(); + + // Set up event listeners for form changes + document.addEventListener('change', (e) => { + if (e.target.matches('[data-var]')) { + handleConfigChange(e.target); + } + }); + + console.log('Configuration interface initialized - ready for API calls'); +}); \ No newline at end of file diff --git a/static/scripts/collapsible-state.js b/static/scripts/collapsible-state.js new file mode 100644 index 0000000..2f29d87 --- /dev/null +++ b/static/scripts/collapsible-state.js @@ -0,0 +1,515 @@ +/** + * Shared collapsible state management for filters and sort sections + * Handles BK_SHOW_GRID_FILTERS and BK_SHOW_GRID_SORT configuration with user preferences + */ + +// Generic state management for collapsible sections (filter and sort) +function initializeCollapsibleState(elementId, storageKey) { + const element = document.getElementById(elementId); + const toggleButton = document.querySelector(`[data-bs-target="#${elementId}"]`); + + if (!element || !toggleButton) return; + + // Restore state on page load + const savedState = sessionStorage.getItem(storageKey); + if (savedState === 'open') { + // User explicitly opened it + element.classList.add('show'); + toggleButton.setAttribute('aria-expanded', 'true'); + } else if (savedState === 'closed') { + // User explicitly closed it, override template state + element.classList.remove('show'); + toggleButton.setAttribute('aria-expanded', 'false'); + } + // If no saved state, keep the template state (respects BK_SHOW_GRID_FILTERS/BK_SHOW_GRID_SORT) + + // Listen for toggle events + element.addEventListener('show.bs.collapse', () => { + sessionStorage.setItem(storageKey, 'open'); + }); + + element.addEventListener('hide.bs.collapse', () => { + sessionStorage.setItem(storageKey, 'closed'); + }); +} + +// Initialize filter and sort states for a specific page +function initializePageCollapsibleStates(pagePrefix, filterElementId = 'table-filter', sortElementId = 'table-sort') { + initializeCollapsibleState(filterElementId, `${pagePrefix}-filter-state`); + initializeCollapsibleState(sortElementId, `${pagePrefix}-sort-state`); + + // Initialize sort icons based on current URL parameters (for all pages) + const urlParams = new URLSearchParams(window.location.search); + const currentSort = urlParams.get('sort'); + const currentOrder = urlParams.get('order'); + if (currentSort || currentOrder) { + updateSortIcon(currentOrder); + } +} + +// Shared function to preserve filter state during filter changes +function preserveCollapsibleStateOnChange(elementId, storageKey) { + const element = document.getElementById(elementId); + const wasOpen = element && element.classList.contains('show'); + + // Store the state to restore after page reload + if (wasOpen) { + sessionStorage.setItem(storageKey, 'open'); + } +} + +// Setup color dropdown with visual indicators (shared implementation) +function setupColorDropdown() { + const colorSelect = document.getElementById('filter-color'); + if (!colorSelect) return; + + // Merge duplicate color options where one has color_rgb and the other is None + const colorMap = new Map(); + const allOptions = colorSelect.querySelectorAll('option[data-color-id]'); + + // First pass: collect all options by color_id + allOptions.forEach(option => { + const colorId = option.dataset.colorId; + const colorRgb = option.dataset.colorRgb; + const colorName = option.textContent.trim(); + + if (!colorMap.has(colorId)) { + colorMap.set(colorId, []); + } + colorMap.get(colorId).push({ + element: option, + colorRgb: colorRgb, + colorName: colorName, + selected: option.selected + }); + }); + + // Second pass: merge duplicates, keeping the one with color_rgb + colorMap.forEach((options, colorId) => { + if (options.length > 1) { + // Find option with color_rgb (not empty/null/undefined) + const withRgb = options.find(opt => opt.colorRgb && opt.colorRgb !== 'None' && opt.colorRgb !== ''); + const withoutRgb = options.find(opt => !opt.colorRgb || opt.colorRgb === 'None' || opt.colorRgb === ''); + + if (withRgb && withoutRgb) { + // Keep the selected state from either option + const wasSelected = withRgb.selected || withoutRgb.selected; + + // Update the option with RGB to be selected if either was selected + if (wasSelected) { + withRgb.element.selected = true; + } + + // Remove the option without RGB + withoutRgb.element.remove(); + } + } + }); + + // Add color squares to remaining option text + const remainingOptions = colorSelect.querySelectorAll('option[data-color-rgb]'); + remainingOptions.forEach(option => { + const colorRgb = option.dataset.colorRgb; + const colorId = option.dataset.colorId; + const colorName = option.textContent.trim(); + + if (colorRgb && colorRgb !== 'None' && colorRgb !== '' && colorId !== '9999') { + // Create a visual indicator (using Unicode square) + option.textContent = `${colorName}`; //■ + //option.style.color = `#${colorRgb}`; + } + }); +} + +// Check if pagination mode is enabled for a specific table +function isPaginationModeForTable(tableId) { + const tableElement = document.querySelector(`#${tableId}`); + return tableElement && tableElement.getAttribute('data-table') === 'false'; +} + +// Update sort icon based on current sort direction +function updateSortIcon(sortDirection = null) { + // Find the main sort icon (could be in grid-sort or table-sort) + const sortIcon = document.querySelector('#grid-sort .ri-sort-asc, #grid-sort .ri-sort-desc, #table-sort .ri-sort-asc, #table-sort .ri-sort-desc'); + + if (!sortIcon) return; + + // Remove existing sort classes + sortIcon.classList.remove('ri-sort-asc', 'ri-sort-desc'); + + // Add appropriate class based on sort direction + if (sortDirection === 'desc') { + sortIcon.classList.add('ri-sort-desc'); + } else { + sortIcon.classList.add('ri-sort-asc'); + } +} + +// Initialize sort button states and icons for pagination mode +window.initializeSortButtonStates = function(currentSort, currentOrder) { + const sortButtons = document.querySelectorAll('[data-sort-attribute]'); + + // Update main sort icon + updateSortIcon(currentOrder); + + if (currentSort) { + sortButtons.forEach(btn => { + // Clear all buttons first + btn.classList.remove('btn-primary'); + btn.classList.add('btn-outline-primary'); + btn.removeAttribute('data-current-direction'); + + // Set active state for current sort + if (btn.dataset.sortAttribute === currentSort) { + btn.classList.remove('btn-outline-primary'); + btn.classList.add('btn-primary'); + btn.dataset.currentDirection = currentOrder || 'asc'; + } + }); + } +}; + +// Shared sort button setup function +window.setupSharedSortButtons = function(tableId, tableInstanceGlobal, columnMap) { + const sortButtons = document.querySelectorAll('[data-sort-attribute]'); + const clearButton = document.querySelector('[data-sort-clear]'); + const isPaginationMode = isPaginationModeForTable(tableId); + + sortButtons.forEach(button => { + button.addEventListener('click', () => { + const attribute = button.dataset.sortAttribute; + const isDesc = button.dataset.sortDesc === 'true'; + + if (isPaginationMode) { + // PAGINATION MODE - Server-side sorting via URL parameters + const currentUrl = new URL(window.location); + const currentSort = currentUrl.searchParams.get('sort'); + const currentOrder = currentUrl.searchParams.get('order'); + const isCurrentlyActive = currentSort === attribute; + + let newDirection; + if (isCurrentlyActive) { + // Toggle direction if same attribute + newDirection = currentOrder === 'asc' ? 'desc' : 'asc'; + } else { + // Use default direction for new attribute + newDirection = isDesc ? 'desc' : 'asc'; + } + + // Set sort parameters and reset to first page + currentUrl.searchParams.set('sort', attribute); + currentUrl.searchParams.set('order', newDirection); + currentUrl.searchParams.set('page', '1'); + + // Navigate to sorted results + window.location.href = currentUrl.toString(); + + } else { + // ORIGINAL MODE - Client-side sorting via Simple DataTables + const columnIndex = columnMap[attribute]; + const tableInstance = window[tableInstanceGlobal]; + + if (columnIndex !== undefined && tableInstance) { + // Determine sort direction + const isCurrentlyActive = button.classList.contains('btn-primary'); + const currentDirection = button.dataset.currentDirection || (isDesc ? 'desc' : 'asc'); + const newDirection = isCurrentlyActive ? + (currentDirection === 'asc' ? 'desc' : 'asc') : + (isDesc ? 'desc' : 'asc'); + + // Clear other active buttons + sortButtons.forEach(btn => { + btn.classList.remove('btn-primary'); + btn.classList.add('btn-outline-primary'); + btn.removeAttribute('data-current-direction'); + }); + + // Mark this button as active + button.classList.remove('btn-outline-primary'); + button.classList.add('btn-primary'); + button.dataset.currentDirection = newDirection; + + // Apply sort using Simple DataTables API + tableInstance.table.columns.sort(columnIndex, newDirection); + + // Update sort icon to reflect new direction + updateSortIcon(newDirection); + } + } + }); + }); + + if (clearButton) { + clearButton.addEventListener('click', () => { + if (isPaginationMode) { + // PAGINATION MODE - Clear server-side sorting via URL parameters + const currentUrl = new URL(window.location); + currentUrl.searchParams.delete('sort'); + currentUrl.searchParams.delete('order'); + currentUrl.searchParams.set('page', '1'); + window.location.href = currentUrl.toString(); + + } else { + // ORIGINAL MODE - Clear client-side sorting + // Clear all sort buttons + sortButtons.forEach(btn => { + btn.classList.remove('btn-primary'); + btn.classList.add('btn-outline-primary'); + btn.removeAttribute('data-current-direction'); + }); + + // Reset sort icon to default ascending + updateSortIcon('asc'); + + // Reset table sort - remove all sorting + const tableInstance = window[tableInstanceGlobal]; + if (tableInstance) { + const tableElement = document.querySelector(`#${tableId}`); + const currentPerPage = tableInstance.table.options.perPage; + tableInstance.table.destroy(); + + setTimeout(() => { + // Create new instance using the globally available BrickTable class + const newInstance = new window.BrickTable(tableElement, currentPerPage); + window[tableInstanceGlobal] = newInstance; + + // Re-enable search functionality + newInstance.table.searchable = true; + }, 50); + } + } + }); + } +}; + +// ================================================================= +// SHARED FUNCTIONS FOR PAGE-SPECIFIC OPERATIONS +// ================================================================= + +// Shared pagination mode detection (works for any table/grid ID) +window.isPaginationModeForPage = function(elementId, attributeName = 'data-table') { + const element = document.querySelector(`#${elementId}`); + return element && element.getAttribute(attributeName) === 'false'; +}; + +// Shared URL parameter update helper +window.updateUrlParams = function(params, resetPage = true) { + const currentUrl = new URL(window.location); + + // Apply parameter updates + Object.entries(params).forEach(([key, value]) => { + if (value === null || value === undefined || value === '' || value === 'all') { + currentUrl.searchParams.delete(key); + } else { + currentUrl.searchParams.set(key, value); + } + }); + + // Reset to page 1 if requested + if (resetPage) { + currentUrl.searchParams.set('page', '1'); + } + + // Navigate to updated URL + window.location.href = currentUrl.toString(); +}; + +// Shared filter application (supports owner, color, theme, year, storage, tag, and problems filters) +window.applyPageFilters = function(tableId) { + const ownerSelect = document.getElementById('filter-owner'); + const colorSelect = document.getElementById('filter-color'); + const themeSelect = document.getElementById('filter-theme'); + const yearSelect = document.getElementById('filter-year'); + const storageSelect = document.getElementById('filter-storage'); + const tagSelect = document.getElementById('filter-tag'); + const problemsSelect = document.getElementById('filter-problems'); + const params = {}; + + // Handle owner filter + if (ownerSelect) { + params.owner = ownerSelect.value; + } + + // Handle color filter + if (colorSelect) { + params.color = colorSelect.value; + } + + // Handle theme filter + if (themeSelect) { + params.theme = themeSelect.value; + } + + // Handle year filter + if (yearSelect) { + params.year = yearSelect.value; + } + + // Handle storage filter + if (storageSelect) { + params.storage = storageSelect.value; + } + + // Handle tag filter + if (tagSelect) { + params.tag = tagSelect.value; + } + + // Handle problems filter (for minifigures page) + if (problemsSelect) { + params.problems = problemsSelect.value; + } + + // Check if we're in pagination mode + const isPaginationMode = window.isPaginationModeForPage(tableId); + + // Update URL with new parameters + // Only reset to page 1 if in server-side pagination mode + window.updateUrlParams(params, isPaginationMode); +}; + +// Shared search setup for both pagination and client-side modes +window.setupPageSearch = function(tableId, searchInputId, clearButtonId, tableInstanceGlobal) { + const searchInput = document.getElementById(searchInputId); + const searchClear = document.getElementById(clearButtonId); + + if (!searchInput || !searchClear) return; + + const isPaginationMode = window.isPaginationModeForPage(tableId); + + if (isPaginationMode) { + // PAGINATION MODE - Server-side search with Enter key + searchInput.addEventListener('keypress', (e) => { + if (e.key === 'Enter') { + e.preventDefault(); + const searchValue = e.target.value.trim(); + window.updateUrlParams({ search: searchValue }, true); + } + }); + + // Clear search + searchClear.addEventListener('click', () => { + searchInput.value = ''; + window.updateUrlParams({ search: null }, true); + }); + + } else { + // ORIGINAL MODE - Client-side instant search via Simple DataTables + const setupClientSearch = () => { + const tableElement = document.querySelector(`table[data-table="true"]#${tableId}`); + const tableInstance = window[tableInstanceGlobal]; + + if (tableElement && tableInstance) { + // Enable search functionality + tableInstance.table.searchable = true; + + // Instant search as user types + searchInput.addEventListener('input', (e) => { + const searchValue = e.target.value.trim(); + tableInstance.table.search(searchValue); + }); + + // Clear search + searchClear.addEventListener('click', () => { + searchInput.value = ''; + tableInstance.table.search(''); + }); + } else { + // If table instance not ready, try again + setTimeout(setupClientSearch, 100); + } + }; + + setTimeout(setupClientSearch, 100); + } +}; + +// Shared function to preserve filter state and apply filters +window.applyFiltersAndKeepState = function(tableId, storageKey) { + const filterElement = document.getElementById('table-filter'); + const wasOpen = filterElement && filterElement.classList.contains('show'); + + // Apply the filters + window.applyPageFilters(tableId); + + // Store the state to restore after page reload + if (wasOpen) { + sessionStorage.setItem(storageKey, 'open'); + } +}; + +// Shared function to clear all filters for a page (works in both pagination and client-side modes) +window.clearPageFilters = function(tableId, filterParams) { + const isPaginationMode = window.isPaginationModeForPage(tableId); + + if (isPaginationMode) { + // SERVER-SIDE PAGINATION MODE: Remove all filter parameters and redirect to base URL + const currentUrl = new URL(window.location); + + // Remove all filter parameters + filterParams.forEach(param => { + currentUrl.searchParams.delete(param); + }); + + // Also remove page parameter to go back to clean base URL + currentUrl.searchParams.delete('page'); + + // Navigate to cleaned URL (will be just /xxx if no other params) + window.location.href = currentUrl.toString(); + } else { + // CLIENT-SIDE MODE: Reset all filter dropdowns to "all" + filterParams.forEach(param => { + const select = document.getElementById(`filter-${param}`); + if (select) { + select.value = 'all'; + } + }); + + // Remove page parameter from URL if present (without reloading) + const currentUrl = new URL(window.location); + if (currentUrl.searchParams.has('page')) { + currentUrl.searchParams.delete('page'); + window.history.replaceState({}, '', currentUrl.toString()); + } + + // Trigger filter application (will use existing filter logic) + window.applyPageFilters(tableId); + } +}; + +// Shared initialization for table pages (parts, problems, minifigures) +window.initializeTablePage = function(config) { + const { + pagePrefix, // e.g., 'parts', 'problems', 'minifigures' + tableId, // e.g., 'parts', 'problems', 'minifigures' + searchInputId = 'table-search', + clearButtonId = 'table-search-clear', + tableInstanceGlobal, // e.g., 'partsTableInstance', 'problemsTableInstance' + sortColumnMap, // Column mapping for sort buttons + hasColorDropdown = true + } = config; + + // Initialize collapsible states (filter and sort) + initializePageCollapsibleStates(pagePrefix); + + // Setup search functionality + window.setupPageSearch(tableId, searchInputId, clearButtonId, tableInstanceGlobal); + + // Setup color dropdown if needed + if (hasColorDropdown) { + setupColorDropdown(); + } + + // Setup sort buttons with shared functionality + if (sortColumnMap) { + window.setupSharedSortButtons(tableId, tableInstanceGlobal, sortColumnMap); + } + + // Initialize sort button states and icons for pagination mode + if (window.isPaginationModeForPage(tableId)) { + const urlParams = new URLSearchParams(window.location.search); + const currentSort = urlParams.get('sort'); + const currentOrder = urlParams.get('order'); + window.initializeSortButtonStates(currentSort, currentOrder); + } +}; \ No newline at end of file diff --git a/static/scripts/grid/filter.js b/static/scripts/grid/filter.js index 6de37a8..098e2b4 100644 --- a/static/scripts/grid/filter.js +++ b/static/scripts/grid/filter.js @@ -91,7 +91,7 @@ class BrickGridFilter { attribute: select.value, bool: true, value: "1" - }) + }); } break; } @@ -106,19 +106,48 @@ class BrickGridFilter { const attribute = current.getAttribute(`data-${filter.attribute}`); // Bool check - // Attribute not equal value, or undefined and value is truthy + // For boolean attributes (like owner/tag filtering) if (filter.bool) { - if ((attribute != null && attribute != filter.value) || (attribute == null && filter.value == "1")) { - current.parentElement.classList.add("d-none"); - return; + if (filter.value == "1") { + // Looking for sets WITH this attribute + // Hide if attribute is missing (null) or explicitly "0" + // For owner/tag attributes: missing = doesn't have this owner/tag + if (attribute == null || attribute == "0") { + current.parentElement.classList.add("d-none"); + return; + } + } else if (filter.value == "0") { + // Looking for sets WITHOUT this attribute + // Hide if attribute is present and "1" + // For owner/tag attributes: present and "1" = has this owner/tag + if (attribute == "1") { + current.parentElement.classList.add("d-none"); + return; + } + // Note: null (missing) is treated as "doesn't have" which is what we want for value="0" } } // Value check - // Attribute not equal value, or attribute undefined - else if ((attribute != null && attribute != filter.value) || attribute == null) { + // For consolidated cards, attributes may be comma or pipe-separated (e.g., "storage1,storage2" or "storage1|storage2") + else if (attribute == null) { + // Hide if attribute is missing current.parentElement.classList.add("d-none"); return; + } else if (attribute.includes(',') || attribute.includes('|')) { + // Handle comma or pipe-separated values (consolidated cards) + const separator = attribute.includes('|') ? '|' : ','; + const values = attribute.split(separator).map(v => v.trim()); + if (!values.includes(filter.value)) { + current.parentElement.classList.add("d-none"); + return; + } + } else { + // Handle single values (regular cards) + if (attribute != filter.value) { + current.parentElement.classList.add("d-none"); + return; + } } } @@ -151,7 +180,12 @@ class BrickGridFilter { } // If we passed all filters, we need to display it - current.parentElement.classList.remove("d-none"); + // But also check if it's hidden by duplicate filter + if (!current.parentElement.classList.contains("duplicate-filter-hidden")) { + current.parentElement.classList.remove("d-none"); + } else { + current.parentElement.classList.add("d-none"); + } }); } } diff --git a/static/scripts/grid/grid.js b/static/scripts/grid/grid.js index d0b9dab..1b8bd82 100644 --- a/static/scripts/grid/grid.js +++ b/static/scripts/grid/grid.js @@ -17,7 +17,13 @@ class BrickGrid { } } +// Global grid instances storage +window.gridInstances = {}; + // Helper to setup the grids const setup_grids = () => document.querySelectorAll('*[data-grid="true"]').forEach( - el => new BrickGrid(el) + el => { + const grid = new BrickGrid(el); + window.gridInstances[el.id] = grid; + } ); diff --git a/static/scripts/minifigures.js b/static/scripts/minifigures.js index 8da5e76..4984c8b 100644 --- a/static/scripts/minifigures.js +++ b/static/scripts/minifigures.js @@ -1,30 +1,88 @@ // Minifigures page functionality -function filterByOwner() { - const select = document.getElementById('filter-owner'); - const selectedOwner = select.value; + +// Check if we're in pagination mode (server-side) or original mode (client-side) +function isPaginationMode() { + const tableElement = document.querySelector('#minifigures'); + return tableElement && tableElement.getAttribute('data-table') === 'false'; +} + +function applyFilters() { + const ownerSelect = document.getElementById('filter-owner'); + const problemsSelect = document.getElementById('filter-problems'); + const themeSelect = document.getElementById('filter-theme'); + const yearSelect = document.getElementById('filter-year'); const currentUrl = new URL(window.location); - if (selectedOwner === 'all') { - currentUrl.searchParams.delete('owner'); - } else { - currentUrl.searchParams.set('owner', selectedOwner); + // Apply owner filter + if (ownerSelect) { + const selectedOwner = ownerSelect.value; + if (selectedOwner === 'all') { + currentUrl.searchParams.delete('owner'); + } else { + currentUrl.searchParams.set('owner', selectedOwner); + } } - window.location.href = currentUrl.toString(); + // Apply problems filter + if (problemsSelect) { + const selectedProblems = problemsSelect.value; + if (selectedProblems === 'all') { + currentUrl.searchParams.delete('problems'); + } else { + currentUrl.searchParams.set('problems', selectedProblems); + } + } + + // Apply theme filter + if (themeSelect) { + const selectedTheme = themeSelect.value; + if (selectedTheme === 'all') { + currentUrl.searchParams.delete('theme'); + } else { + currentUrl.searchParams.set('theme', selectedTheme); + } + } + + // Apply year filter + if (yearSelect) { + const selectedYear = yearSelect.value; + if (selectedYear === 'all') { + currentUrl.searchParams.delete('year'); + } else { + currentUrl.searchParams.set('year', selectedYear); + } + } + + // Only reset to page 1 when filtering in server-side pagination mode + if (isPaginationMode()) { + currentUrl.searchParams.set('page', '1'); + // Navigate to updated URL (server-side pagination) + window.location.href = currentUrl.toString(); + } else { + // Client-side mode: Update URL without page reload + window.history.replaceState({}, '', currentUrl.toString()); + } +} + +// Legacy function for compatibility +function filterByOwner() { + applyFilters(); +} + +// Initialize filter and sort states for minifigures page +function initializeCollapsibleStates() { + initializePageCollapsibleStates('minifigures'); } // Keep filters expanded after selection +function applyFiltersAndKeepOpen() { + preserveCollapsibleStateOnChange('table-filter', 'minifigures-filter-state'); + applyFilters(); +} + +// Legacy function for compatibility function filterByOwnerAndKeepOpen() { - // Remember if filters were open - const filterSection = document.getElementById('table-filter'); - const wasOpen = filterSection && filterSection.classList.contains('show'); - - filterByOwner(); - - // Store the state to restore after page reload - if (wasOpen) { - sessionStorage.setItem('keepFiltersOpen', 'true'); - } + applyFiltersAndKeepOpen(); } // Setup table search and sort functionality @@ -32,117 +90,104 @@ document.addEventListener("DOMContentLoaded", () => { const searchInput = document.getElementById('table-search'); const searchClear = document.getElementById('table-search-clear'); - // Restore filter state after page load - if (sessionStorage.getItem('keepFiltersOpen') === 'true') { - const filterSection = document.getElementById('table-filter'); - const filterButton = document.querySelector('[data-bs-target="#table-filter"]'); - - if (filterSection && filterButton) { - filterSection.classList.add('show'); - filterButton.setAttribute('aria-expanded', 'true'); - } - - sessionStorage.removeItem('keepFiltersOpen'); - } + // Initialize collapsible states (filter and sort) + initializeCollapsibleStates(); if (searchInput && searchClear) { - // Wait for table to be initialized by setup_tables - setTimeout(() => { - const tableElement = document.querySelector('table[data-table="true"]'); - if (tableElement && window.brickTableInstance) { - // Enable custom search for minifigures table - window.brickTableInstance.table.searchable = true; + if (isPaginationMode()) { + // PAGINATION MODE - Server-side search + const searchForm = document.createElement('form'); + searchForm.style.display = 'none'; + searchInput.parentNode.appendChild(searchForm); + searchForm.appendChild(searchInput.cloneNode(true)); - // Connect search input to table - searchInput.addEventListener('input', (e) => { - window.brickTableInstance.table.search(e.target.value); - }); + // Handle Enter key for search + searchInput.addEventListener('keydown', (e) => { + if (e.key === 'Enter') { + e.preventDefault(); + performServerSearch(); + } + }); - // Clear search - searchClear.addEventListener('click', () => { - searchInput.value = ''; - window.brickTableInstance.table.search(''); - }); - - // Setup sort buttons - setupSortButtons(); + // Handle search button click (if exists) + const searchButton = document.querySelector('[data-search-trigger]'); + if (searchButton) { + searchButton.addEventListener('click', performServerSearch); } - }, 100); + + // Clear search + searchClear.addEventListener('click', () => { + searchInput.value = ''; + performServerSearch(); + }); + + function performServerSearch() { + const currentUrl = new URL(window.location); + const searchQuery = searchInput.value.trim(); + + if (searchQuery) { + currentUrl.searchParams.set('search', searchQuery); + } else { + currentUrl.searchParams.delete('search'); + } + + // Reset to page 1 when searching + currentUrl.searchParams.set('page', '1'); + window.location.href = currentUrl.toString(); + } + + } else { + // ORIGINAL MODE - Client-side search with Simple DataTables + setTimeout(() => { + const tableElement = document.querySelector('table[data-table="true"]'); + if (tableElement && window.brickTableInstance) { + // Enable custom search for minifigures table + window.brickTableInstance.table.searchable = true; + + // Connect search input to table + searchInput.addEventListener('input', (e) => { + window.brickTableInstance.table.search(e.target.value); + }); + + // Clear search + searchClear.addEventListener('click', () => { + searchInput.value = ''; + window.brickTableInstance.table.search(''); + }); + } + }, 100); + } + } + + // Setup sort buttons for both modes + setupSortButtons(); + + // Initialize sort button states and icons for pagination mode + if (isPaginationMode()) { + const urlParams = new URLSearchParams(window.location.search); + const currentSort = urlParams.get('sort'); + const currentOrder = urlParams.get('order'); + window.initializeSortButtonStates(currentSort, currentOrder); + } + + // Initialize clear filters button + const clearButton = document.getElementById('table-filter-clear'); + if (clearButton) { + clearButton.addEventListener('click', () => { + window.clearPageFilters('minifigures', ['owner', 'problems', 'theme', 'year']); + }); } }); function setupSortButtons() { - // Sort button functionality - const sortButtons = document.querySelectorAll('[data-sort-attribute]'); - const clearButton = document.querySelector('[data-sort-clear]'); - - sortButtons.forEach(button => { - button.addEventListener('click', () => { - const attribute = button.dataset.sortAttribute; - const isDesc = button.dataset.sortDesc === 'true'; - - // Get column index based on attribute - const columnMap = { - 'name': 1, - 'parts': 2, - 'quantity': 3, - 'missing': 4, - 'damaged': 5, - 'sets': 6 - }; - - const columnIndex = columnMap[attribute]; - if (columnIndex !== undefined && window.brickTableInstance) { - // Determine sort direction - const isCurrentlyActive = button.classList.contains('btn-primary'); - const currentDirection = button.dataset.currentDirection || (isDesc ? 'desc' : 'asc'); - const newDirection = isCurrentlyActive ? - (currentDirection === 'asc' ? 'desc' : 'asc') : - (isDesc ? 'desc' : 'asc'); - - // Clear other active buttons - sortButtons.forEach(btn => { - btn.classList.remove('btn-primary'); - btn.classList.add('btn-outline-primary'); - btn.removeAttribute('data-current-direction'); - }); - - // Mark this button as active - button.classList.remove('btn-outline-primary'); - button.classList.add('btn-primary'); - button.dataset.currentDirection = newDirection; - - // Apply sort using Simple DataTables API - window.brickTableInstance.table.columns.sort(columnIndex, newDirection); - } - }); - }); - - if (clearButton) { - clearButton.addEventListener('click', () => { - // Clear all sort buttons - sortButtons.forEach(btn => { - btn.classList.remove('btn-primary'); - btn.classList.add('btn-outline-primary'); - btn.removeAttribute('data-current-direction'); - }); - - // Reset table sort - remove all sorting - if (window.brickTableInstance) { - // Destroy and recreate to clear sorting - const tableElement = document.querySelector('#minifigures'); - const currentPerPage = window.brickTableInstance.table.options.perPage; - window.brickTableInstance.table.destroy(); - - setTimeout(() => { - // Create new instance using the globally available BrickTable class - const newInstance = new window.BrickTable(tableElement, currentPerPage); - window.brickTableInstance = newInstance; - - // Re-enable search functionality - newInstance.table.searchable = true; - }, 50); - } - }); - } + const columnMap = { + 'name': 1, + 'parts': 2, + 'quantity': 3, + 'missing': 4, + 'damaged': 5, + 'sets': 6 + }; + // Use shared sort buttons setup from collapsible-state.js + window.setupSharedSortButtons('minifigures', 'brickTableInstance', columnMap); } \ No newline at end of file diff --git a/static/scripts/parts-bulk-operations.js b/static/scripts/parts-bulk-operations.js new file mode 100644 index 0000000..1a032bd --- /dev/null +++ b/static/scripts/parts-bulk-operations.js @@ -0,0 +1,182 @@ +// Bulk operations for parts in set details page +class PartsBulkOperations { + constructor(accordionId) { + this.accordionId = accordionId; + this.setupModal(); + this.setupEventListeners(); + } + + setupModal() { + // Create Bootstrap modal if it doesn't exist + if (!document.getElementById('partsConfirmModal')) { + const modalHTML = ` + + `; + document.body.insertAdjacentHTML('beforeend', modalHTML); + } + } + + setupEventListeners() { + // Mark all as missing (only if missing parts are not hidden) + const markAllMissingBtn = document.getElementById(`mark-all-missing-${this.accordionId}`); + if (markAllMissingBtn) { + markAllMissingBtn.addEventListener('click', (e) => { + e.preventDefault(); + this.confirmAndExecute( + 'Mark all parts as missing?', + 'This will set the missing count to the maximum quantity for all parts in this section.', + () => this.markAllMissing() + ); + }); + } + + // Clear all missing (only if missing parts are not hidden) + const clearAllMissingBtn = document.getElementById(`clear-all-missing-${this.accordionId}`); + if (clearAllMissingBtn) { + clearAllMissingBtn.addEventListener('click', (e) => { + e.preventDefault(); + this.confirmAndExecute( + 'Clear all missing parts?', + 'This will clear the missing field for all parts in this section.', + () => this.clearAllMissing() + ); + }); + } + + // Check all checkboxes (only if checked parts are not hidden) + const checkAllBtn = document.getElementById(`check-all-${this.accordionId}`); + if (checkAllBtn) { + checkAllBtn.addEventListener('click', (e) => { + e.preventDefault(); + this.checkAll(); + }); + } + + // Uncheck all checkboxes (only if checked parts are not hidden) + const uncheckAllBtn = document.getElementById(`uncheck-all-${this.accordionId}`); + if (uncheckAllBtn) { + uncheckAllBtn.addEventListener('click', (e) => { + e.preventDefault(); + this.uncheckAll(); + }); + } + } + + confirmAndExecute(title, message, callback) { + const modal = document.getElementById('partsConfirmModal'); + const modalTitle = document.getElementById('partsConfirmModalLabel'); + const modalMessage = document.getElementById('partsConfirmModalMessage'); + const confirmBtn = document.getElementById('partsConfirmModalConfirm'); + + // Set modal content + modalTitle.textContent = title; + modalMessage.textContent = message; + + // Remove any existing event listeners and add new one + const newConfirmBtn = confirmBtn.cloneNode(true); + confirmBtn.parentNode.replaceChild(newConfirmBtn, confirmBtn); + + newConfirmBtn.addEventListener('click', () => { + const modalInstance = bootstrap.Modal.getInstance(modal); + modalInstance.hide(); + callback(); + }); + + // Show modal + const modalInstance = new bootstrap.Modal(modal); + modalInstance.show(); + } + + markAllMissing() { + const accordionElement = document.getElementById(this.accordionId); + if (!accordionElement) return; + + // Find all rows in this accordion + const rows = accordionElement.querySelectorAll('tbody tr'); + rows.forEach(row => { + // Find the quantity cell (usually 4th column) + const quantityCell = row.cells[3]; // Index 3 for quantity column + const missingInput = row.querySelector('input[id*="-missing-"]'); + + if (quantityCell && missingInput) { + // Extract quantity from cell text content + const quantityText = quantityCell.textContent.trim(); + const quantity = parseInt(quantityText) || 1; // Default to 1 if can't parse + + if (missingInput.value !== quantity.toString()) { + missingInput.value = quantity.toString(); + // Trigger change event to activate BrickChanger + missingInput.dispatchEvent(new Event('change', { bubbles: true })); + } + } + }); + } + + clearAllMissing() { + const accordionElement = document.getElementById(this.accordionId); + if (!accordionElement) return; + + const missingInputs = accordionElement.querySelectorAll('input[id*="-missing-"]'); + missingInputs.forEach(input => { + if (input.value !== '') { + input.value = ''; + // Trigger change event to activate BrickChanger + input.dispatchEvent(new Event('change', { bubbles: true })); + } + }); + } + + checkAll() { + const accordionElement = document.getElementById(this.accordionId); + if (!accordionElement) return; + + const checkboxes = accordionElement.querySelectorAll('input[id*="-checked-"][type="checkbox"]'); + checkboxes.forEach(checkbox => { + if (!checkbox.checked) { + checkbox.checked = true; + // Trigger change event to activate BrickChanger + checkbox.dispatchEvent(new Event('change', { bubbles: true })); + } + }); + } + + uncheckAll() { + const accordionElement = document.getElementById(this.accordionId); + if (!accordionElement) return; + + const checkboxes = accordionElement.querySelectorAll('input[id*="-checked-"][type="checkbox"]'); + checkboxes.forEach(checkbox => { + if (checkbox.checked) { + checkbox.checked = false; + // Trigger change event to activate BrickChanger + checkbox.dispatchEvent(new Event('change', { bubbles: true })); + } + }); + } +} + +// Initialize bulk operations for all part accordions when DOM is ready +document.addEventListener('DOMContentLoaded', () => { + // Find all hamburger menus and initialize bulk operations + const hamburgerMenus = document.querySelectorAll('button[id^="hamburger-"]'); + hamburgerMenus.forEach(button => { + const accordionId = button.id.replace('hamburger-', ''); + new PartsBulkOperations(accordionId); + }); +}); \ No newline at end of file diff --git a/static/scripts/parts.js b/static/scripts/parts.js index 5c1393c..95d5cb9 100644 --- a/static/scripts/parts.js +++ b/static/scripts/parts.js @@ -1,190 +1,35 @@ -// Parts page functionality -function applyFilters() { - const ownerSelect = document.getElementById('filter-owner'); - const colorSelect = document.getElementById('filter-color'); - const currentUrl = new URL(window.location); - - // Handle owner filter - if (ownerSelect) { - const selectedOwner = ownerSelect.value; - if (selectedOwner === 'all') { - currentUrl.searchParams.delete('owner'); - } else { - currentUrl.searchParams.set('owner', selectedOwner); - } - } - - // Handle color filter - if (colorSelect) { - const selectedColor = colorSelect.value; - if (selectedColor === 'all') { - currentUrl.searchParams.delete('color'); - } else { - currentUrl.searchParams.set('color', selectedColor); - } - } - - window.location.href = currentUrl.toString(); -} - -function setupColorDropdown() { - const colorSelect = document.getElementById('filter-color'); - if (!colorSelect) return; - - // Add color squares to option text - const options = colorSelect.querySelectorAll('option[data-color-rgb]'); - options.forEach(option => { - const colorRgb = option.dataset.colorRgb; - const colorId = option.dataset.colorId; - const colorName = option.textContent.trim(); - - if (colorRgb && colorId !== '9999') { - // Create a visual indicator (using Unicode square) - option.textContent = `${colorName}`; //■ - //option.style.color = `#${colorRgb}`; - } - }); -} +// Parts page functionality - now uses shared functions // Keep filters expanded after selection function applyFiltersAndKeepOpen() { - // Remember if filters were open - const filterSection = document.getElementById('table-filter'); - const wasOpen = filterSection && filterSection.classList.contains('show'); - - applyFilters(); - - // Store the state to restore after page reload - if (wasOpen) { - sessionStorage.setItem('keepFiltersOpen', 'true'); - } + window.applyFiltersAndKeepState('parts', 'parts-filter-state'); } -function setupSortButtons() { - // Sort button functionality - const sortButtons = document.querySelectorAll('[data-sort-attribute]'); - const clearButton = document.querySelector('[data-sort-clear]'); - - sortButtons.forEach(button => { - button.addEventListener('click', () => { - const attribute = button.dataset.sortAttribute; - const isDesc = button.dataset.sortDesc === 'true'; - - // Get column index based on attribute - const columnMap = { - 'name': 1, - 'color': 2, - 'quantity': 3, - 'missing': 4, - 'damaged': 5, - 'sets': 6, - 'minifigures': 7 - }; - - const columnIndex = columnMap[attribute]; - if (columnIndex !== undefined && window.partsTableInstance) { - // Determine sort direction - const isCurrentlyActive = button.classList.contains('btn-primary'); - const currentDirection = button.dataset.currentDirection || (isDesc ? 'desc' : 'asc'); - const newDirection = isCurrentlyActive ? - (currentDirection === 'asc' ? 'desc' : 'asc') : - (isDesc ? 'desc' : 'asc'); - - // Clear other active buttons - sortButtons.forEach(btn => { - btn.classList.remove('btn-primary'); - btn.classList.add('btn-outline-primary'); - btn.removeAttribute('data-current-direction'); - }); - - // Mark this button as active - button.classList.remove('btn-outline-primary'); - button.classList.add('btn-primary'); - button.dataset.currentDirection = newDirection; - - // Apply sort using Simple DataTables API - window.partsTableInstance.table.columns.sort(columnIndex, newDirection); - } - }); +// Initialize parts page +document.addEventListener("DOMContentLoaded", () => { + // Use shared table page initialization + window.initializeTablePage({ + pagePrefix: 'parts', + tableId: 'parts', + tableInstanceGlobal: 'partsTableInstance', + sortColumnMap: { + 'name': 1, + 'color': 2, + 'quantity': 3, + 'missing': 4, + 'damaged': 5, + 'sets': 6, + 'minifigures': 7 + }, + hasColorDropdown: true }); + // Initialize clear filters button + const clearButton = document.getElementById('table-filter-clear'); if (clearButton) { clearButton.addEventListener('click', () => { - // Clear all sort buttons - sortButtons.forEach(btn => { - btn.classList.remove('btn-primary'); - btn.classList.add('btn-outline-primary'); - btn.removeAttribute('data-current-direction'); - }); - - // Reset table sort - remove all sorting - if (window.partsTableInstance) { - // Destroy and recreate to clear sorting - const tableElement = document.querySelector('#parts'); - const currentPerPage = window.partsTableInstance.table.options.perPage; - window.partsTableInstance.table.destroy(); - - setTimeout(() => { - // Create new instance using the globally available BrickTable class - const newInstance = new window.BrickTable(tableElement, currentPerPage); - window.partsTableInstance = newInstance; - - // Re-enable search functionality - newInstance.table.searchable = true; - }, 50); - } + window.clearPageFilters('parts', ['owner', 'color', 'theme', 'year']); }); } -} +}); -// Setup table search and sort functionality -document.addEventListener("DOMContentLoaded", () => { - const searchInput = document.getElementById('table-search'); - const searchClear = document.getElementById('table-search-clear'); - - // Setup color dropdown with color squares - setupColorDropdown(); - - // Restore filter state after page load - if (sessionStorage.getItem('keepFiltersOpen') === 'true') { - const filterSection = document.getElementById('table-filter'); - const filterButton = document.querySelector('[data-bs-target="#table-filter"]'); - - if (filterSection && filterButton) { - filterSection.classList.add('show'); - filterButton.setAttribute('aria-expanded', 'true'); - } - - sessionStorage.removeItem('keepFiltersOpen'); - } - - if (searchInput && searchClear) { - // Wait for table to be initialized by setup_tables - const setupSearch = () => { - const tableElement = document.querySelector('table[data-table="true"]'); - if (tableElement && window.partsTableInstance) { - // Enable custom search for parts table - window.partsTableInstance.table.searchable = true; - - // Connect search input to table - searchInput.addEventListener('input', (e) => { - window.partsTableInstance.table.search(e.target.value); - }); - - // Clear search - searchClear.addEventListener('click', () => { - searchInput.value = ''; - window.partsTableInstance.table.search(''); - }); - - // Setup sort buttons - setupSortButtons(); - } else { - // If table instance not ready, try again - setTimeout(setupSearch, 100); - } - }; - - setTimeout(setupSearch, 100); - } -}); \ No newline at end of file diff --git a/static/scripts/problems.js b/static/scripts/problems.js new file mode 100644 index 0000000..4e8397e --- /dev/null +++ b/static/scripts/problems.js @@ -0,0 +1,34 @@ +// Problems page functionality - now uses shared functions + +// Keep filters expanded after selection +function applyFiltersAndKeepOpen() { + window.applyFiltersAndKeepState('problems', 'problems-filter-state'); +} + +// Initialize problems page +document.addEventListener("DOMContentLoaded", () => { + // Use shared table page initialization + window.initializeTablePage({ + pagePrefix: 'problems', + tableId: 'problems', + tableInstanceGlobal: 'problemsTableInstance', + sortColumnMap: { + 'name': 1, + 'color': 2, + 'quantity': 3, + 'missing': 4, + 'damaged': 5, + 'sets': 6, + 'minifigures': 7 + }, + hasColorDropdown: true + }); + + // Initialize clear filters button + const clearButton = document.getElementById('table-filter-clear'); + if (clearButton) { + clearButton.addEventListener('click', () => { + window.clearPageFilters('problems', ['owner', 'color', 'theme', 'year', 'storage', 'tag']); + }); + } +}); \ No newline at end of file diff --git a/static/scripts/set-details.js b/static/scripts/set-details.js new file mode 100644 index 0000000..5265374 --- /dev/null +++ b/static/scripts/set-details.js @@ -0,0 +1,15 @@ +// Set details page functionality +document.addEventListener('DOMContentLoaded', function() { + const collapseElement = document.getElementById('all-instances'); + const toggleIcon = document.getElementById('copies-toggle-icon'); + + if (collapseElement && toggleIcon) { + collapseElement.addEventListener('shown.bs.collapse', function() { + toggleIcon.className = 'ri-arrow-up-s-line fs-4'; + }); + + collapseElement.addEventListener('hidden.bs.collapse', function() { + toggleIcon.className = 'ri-arrow-down-s-line fs-4'; + }); + } +}); \ No newline at end of file diff --git a/static/scripts/sets.js b/static/scripts/sets.js new file mode 100644 index 0000000..4e68564 --- /dev/null +++ b/static/scripts/sets.js @@ -0,0 +1,771 @@ +// Sets page functionality + +// Check if we're in pagination mode (server-side) or original mode (client-side) +function isPaginationMode() { + const gridElement = document.querySelector('#grid'); + return gridElement && gridElement.getAttribute('data-grid') === 'false'; +} + +// Initialize filter and sort states for sets page +function initializeCollapsibleStates() { + initializePageCollapsibleStates('sets', 'grid-filter', 'grid-sort'); +} + +// Setup page functionality +document.addEventListener("DOMContentLoaded", () => { + // Initialize collapsible states (filter and sort) + initializeCollapsibleStates(); + + const searchInput = document.getElementById('grid-search'); + const searchClear = document.getElementById('grid-search-clear'); + + // Initialize duplicate filter functionality + initializeDuplicateFilter(); + + // Initialize clear filters button + initializeClearFiltersButton(); + + if (searchInput && searchClear) { + if (isPaginationMode()) { + // PAGINATION MODE - Server-side search + const searchForm = document.createElement('form'); + searchForm.style.display = 'none'; + searchInput.parentNode.appendChild(searchForm); + searchForm.appendChild(searchInput.cloneNode(true)); + + // Handle Enter key for search + searchInput.addEventListener('keydown', (e) => { + if (e.key === 'Enter') { + e.preventDefault(); + performServerSearch(); + } + }); + + // Handle search button click (if exists) + const searchButton = document.querySelector('[data-search-trigger]'); + if (searchButton) { + searchButton.addEventListener('click', performServerSearch); + } + + // Clear search + searchClear.addEventListener('click', () => { + searchInput.value = ''; + performServerSearch(); + }); + + function performServerSearch() { + const currentUrl = new URL(window.location); + const searchQuery = searchInput.value.trim(); + + if (searchQuery) { + currentUrl.searchParams.set('search', searchQuery); + } else { + currentUrl.searchParams.delete('search'); + } + + // Reset to page 1 when searching + currentUrl.searchParams.set('page', '1'); + window.location.href = currentUrl.toString(); + } + + // Setup sort buttons for pagination mode + setupPaginationSortButtons(); + + // Setup filter dropdowns for pagination mode + setupPaginationFilterDropdowns(); + + // Initialize filter dropdowns from URL parameters + initializeFilterDropdowns(); + + // Initialize sort button states and icons for pagination mode + const urlParams = new URLSearchParams(window.location.search); + const currentSort = urlParams.get('sort'); + const currentOrder = urlParams.get('order'); + window.initializeSortButtonStates(currentSort, currentOrder); + + } else { + // ORIGINAL MODE - Client-side filtering with grid scripts + // Initialize filter dropdowns from URL parameters for client-side mode too + initializeClientSideFilterDropdowns(); + } + } +}); + +function setupPaginationSortButtons() { + // Sort button functionality for pagination mode + const sortButtons = document.querySelectorAll('[data-sort-attribute]'); + const clearButton = document.querySelector('[data-sort-clear]'); + + sortButtons.forEach(button => { + button.addEventListener('click', () => { + const attribute = button.dataset.sortAttribute; + const isDesc = button.dataset.sortDesc === 'true'; + + // PAGINATION MODE - Server-side sorting + const currentUrl = new URL(window.location); + const currentSort = currentUrl.searchParams.get('sort'); + const currentOrder = currentUrl.searchParams.get('order'); + + // Determine new sort direction + let newOrder = isDesc ? 'desc' : 'asc'; + if (currentSort === attribute) { + // Toggle direction if clicking the same column + newOrder = currentOrder === 'asc' ? 'desc' : 'asc'; + } + + currentUrl.searchParams.set('sort', attribute); + currentUrl.searchParams.set('order', newOrder); + + // Update sort icon immediately before navigation + updateSortIcon(newOrder); + + // Reset to page 1 when sorting + currentUrl.searchParams.set('page', '1'); + window.location.href = currentUrl.toString(); + }); + }); + + if (clearButton) { + clearButton.addEventListener('click', () => { + // PAGINATION MODE - Clear server-side sorting + const currentUrl = new URL(window.location); + currentUrl.searchParams.delete('sort'); + currentUrl.searchParams.delete('order'); + + // Reset sort icon to default ascending + updateSortIcon('asc'); + + currentUrl.searchParams.set('page', '1'); + window.location.href = currentUrl.toString(); + }); + } +} + +function initializeFilterDropdowns() { + // Set filter dropdown values from URL parameters + const urlParams = new URLSearchParams(window.location.search); + + // Set each filter dropdown value if the parameter exists + const yearParam = urlParams.get('year'); + if (yearParam) { + const yearDropdown = document.getElementById('grid-year'); + if (yearDropdown) { + yearDropdown.value = yearParam; + } + } + + const themeParam = urlParams.get('theme'); + if (themeParam) { + const themeDropdown = document.getElementById('grid-theme'); + if (themeDropdown) { + // Try to set the theme value directly first (for theme names) + themeDropdown.value = themeParam; + + // If that didn't work and the param is numeric (theme ID), + // try to find the corresponding theme name by looking at cards + if (themeDropdown.value !== themeParam && /^\d+$/.test(themeParam)) { + // Look for a card with this theme ID and get its theme name + const cardWithTheme = document.querySelector(`[data-theme-id="${themeParam}"]`); + if (cardWithTheme) { + const themeName = cardWithTheme.getAttribute('data-theme'); + if (themeName) { + themeDropdown.value = themeName; + } + } + } + } + } + + const statusParam = urlParams.get('status'); + if (statusParam) { + const statusDropdown = document.getElementById('grid-status'); + if (statusDropdown) { + statusDropdown.value = statusParam; + } + } + + const ownerParam = urlParams.get('owner'); + if (ownerParam) { + const ownerDropdown = document.getElementById('grid-owner'); + if (ownerDropdown) { + ownerDropdown.value = ownerParam; + } + } + + const purchaseLocationParam = urlParams.get('purchase_location'); + if (purchaseLocationParam) { + const purchaseLocationDropdown = document.getElementById('grid-purchase-location'); + if (purchaseLocationDropdown) { + purchaseLocationDropdown.value = purchaseLocationParam; + } + } + + const storageParam = urlParams.get('storage'); + if (storageParam) { + const storageDropdown = document.getElementById('grid-storage'); + if (storageDropdown) { + storageDropdown.value = storageParam; + } + } + + const tagParam = urlParams.get('tag'); + if (tagParam) { + const tagDropdown = document.getElementById('grid-tag'); + if (tagDropdown) { + tagDropdown.value = tagParam; + } + } +} + +function initializeClientSideFilterDropdowns() { + // Set filter dropdown values from URL parameters and trigger filtering for client-side mode + const urlParams = new URLSearchParams(window.location.search); + let needsFiltering = false; + + // Check if we have any filter parameters to avoid flash of all content + const hasFilterParams = urlParams.has('year') || urlParams.has('theme') || urlParams.has('storage') || urlParams.has('purchase_location'); + + // If we have filter parameters, temporarily hide the grid to prevent flash + if (hasFilterParams) { + const gridElement = document.querySelector('#grid'); + if (gridElement && gridElement.getAttribute('data-grid') === 'true') { + gridElement.style.opacity = '0'; + } + } + + // Set year filter if parameter exists + const yearParam = urlParams.get('year'); + if (yearParam) { + const yearDropdown = document.getElementById('grid-year'); + if (yearDropdown) { + yearDropdown.value = yearParam; + needsFiltering = true; + } + } + + // Set theme filter - handle both theme names and theme IDs + const themeParam = urlParams.get('theme'); + if (themeParam) { + const themeDropdown = document.getElementById('grid-theme'); + if (themeDropdown) { + if (/^\d+$/.test(themeParam)) { + // Theme parameter is an ID, need to convert to theme name by looking at cards + const themeNameFromId = findThemeNameById(themeParam); + if (themeNameFromId) { + themeDropdown.value = themeNameFromId; + needsFiltering = true; + } + } else { + // Theme parameter is already a name + themeDropdown.value = themeParam.toLowerCase(); + needsFiltering = true; + } + } + } + + // Set storage filter if parameter exists + const storageParam = urlParams.get('storage'); + if (storageParam) { + const storageDropdown = document.getElementById('grid-storage'); + if (storageDropdown) { + storageDropdown.value = storageParam; + needsFiltering = true; + } + } + + // Set purchase location filter if parameter exists + const purchaseLocationParam = urlParams.get('purchase_location'); + if (purchaseLocationParam) { + const purchaseLocationDropdown = document.getElementById('grid-purchase-location'); + if (purchaseLocationDropdown) { + purchaseLocationDropdown.value = purchaseLocationParam; + needsFiltering = true; + } + } + + // Trigger filtering if any parameters were set + if (needsFiltering) { + // Try to trigger filtering immediately + const tryToFilter = () => { + const gridElement = document.querySelector('#grid'); + if (gridElement && gridElement.getAttribute('data-grid') === 'true' && window.gridInstances) { + const gridInstance = window.gridInstances[gridElement.id]; + if (gridInstance && gridInstance.filter) { + // This is client-side mode, trigger the filter directly + gridInstance.filter.filter(); + + // Show the grid again after filtering + if (hasFilterParams) { + gridElement.style.opacity = '1'; + gridElement.style.transition = 'opacity 0.2s ease-in-out'; + } + + return true; + } + } + return false; + }; + + // Try filtering immediately + if (!tryToFilter()) { + // If not ready, try again with a shorter delay + setTimeout(() => { + if (!tryToFilter()) { + // Final attempt with longer delay + setTimeout(tryToFilter, 100); + } + }, 50); + } + } +} + +function findThemeNameById(themeId) { + // Look through all cards to find the theme name for this theme ID + const cards = document.querySelectorAll('.card[data-theme-id]'); + for (const card of cards) { + if (card.getAttribute('data-theme-id') === themeId) { + return card.getAttribute('data-theme'); + } + } + return null; +} + +function setupPaginationFilterDropdowns() { + // Filter dropdown functionality for pagination mode + const filterDropdowns = document.querySelectorAll('#grid-filter select'); + + filterDropdowns.forEach(dropdown => { + dropdown.addEventListener('change', () => { + performServerFilter(); + }); + }); + + function performServerFilter() { + const currentUrl = new URL(window.location); + + // Get all filter values + const statusFilter = document.getElementById('grid-status')?.value || ''; + const themeFilter = document.getElementById('grid-theme')?.value || ''; + const yearFilter = document.getElementById('grid-year')?.value || ''; + const ownerFilter = document.getElementById('grid-owner')?.value || ''; + const purchaseLocationFilter = document.getElementById('grid-purchase-location')?.value || ''; + const storageFilter = document.getElementById('grid-storage')?.value || ''; + const tagFilter = document.getElementById('grid-tag')?.value || ''; + + // Update URL parameters + if (statusFilter) { + currentUrl.searchParams.set('status', statusFilter); + } else { + currentUrl.searchParams.delete('status'); + } + + if (themeFilter) { + currentUrl.searchParams.set('theme', themeFilter); + } else { + currentUrl.searchParams.delete('theme'); + } + + if (yearFilter) { + currentUrl.searchParams.set('year', yearFilter); + } else { + currentUrl.searchParams.delete('year'); + } + + if (ownerFilter) { + currentUrl.searchParams.set('owner', ownerFilter); + } else { + currentUrl.searchParams.delete('owner'); + } + + if (purchaseLocationFilter) { + currentUrl.searchParams.set('purchase_location', purchaseLocationFilter); + } else { + currentUrl.searchParams.delete('purchase_location'); + } + + if (storageFilter) { + currentUrl.searchParams.set('storage', storageFilter); + } else { + currentUrl.searchParams.delete('storage'); + } + + if (tagFilter) { + currentUrl.searchParams.set('tag', tagFilter); + } else { + currentUrl.searchParams.delete('tag'); + } + + // Reset to page 1 when filtering + currentUrl.searchParams.set('page', '1'); + window.location.href = currentUrl.toString(); + } +} + +// Set grouping functionality +function initializeSetGrouping() { + const groupToggle = document.getElementById('group-identical-sets'); + if (!groupToggle) return; + + // Load saved state from localStorage + const savedState = localStorage.getItem('groupIdenticalSets') === 'true'; + groupToggle.checked = savedState; + + // Apply grouping on page load if enabled + if (savedState) { + applySetGrouping(); + } + + // Listen for toggle changes + groupToggle.addEventListener('change', function() { + // Save state to localStorage + localStorage.setItem('groupIdenticalSets', this.checked); + + if (this.checked) { + applySetGrouping(); + } else { + removeSetGrouping(); + } + }); +} + +function applySetGrouping() { + const grid = document.getElementById('grid'); + if (!grid) return; + + const setCards = Array.from(grid.children); + const groupedSets = {}; + + // Group sets by rebrickable_set_id + setCards.forEach(cardCol => { + const setCard = cardCol.querySelector('.card[data-set-id]'); + if (!setCard) return; + + const setId = setCard.getAttribute('data-set-id'); + const rebrickableId = setCard.getAttribute('data-rebrickable-id'); + + if (!rebrickableId) return; + + if (!groupedSets[rebrickableId]) { + groupedSets[rebrickableId] = []; + } + + groupedSets[rebrickableId].push({ + cardCol: cardCol, + setId: setId, + rebrickableId: rebrickableId + }); + }); + + // Process each group + Object.keys(groupedSets).forEach(rebrickableId => { + const group = groupedSets[rebrickableId]; + + if (group.length > 1) { + createGroupedSetDisplay(group); + } + }); +} + +function createGroupedSetDisplay(setGroup) { + const firstSet = setGroup[0]; + const firstCard = firstSet.cardCol.querySelector('.card'); + + if (!firstCard) return; + + // Calculate aggregate stats + let totalMissing = 0; + let totalDamaged = 0; + let allSetIds = []; + + setGroup.forEach(set => { + const card = set.cardCol.querySelector('.card'); + + // Get missing and damaged counts from existing data attributes + const missingCount = parseInt(card.getAttribute('data-missing') || '0'); + const damagedCount = parseInt(card.getAttribute('data-damaged') || '0'); + + totalMissing += missingCount; + totalDamaged += damagedCount; + allSetIds.push(set.setId); + }); + + // Create grouped card container + const groupContainer = document.createElement('div'); + groupContainer.className = firstSet.cardCol.className + ' set-group-container'; + groupContainer.innerHTML = ` +
+
+
+ + ${firstCard.querySelector('.card-title')?.textContent || 'Set'} + ${setGroup.length} sets +
+
+ ${totalMissing > 0 ? `${totalMissing} missing` : ''} + ${totalDamaged > 0 ? `${totalDamaged} damaged` : ''} +
+
+
+
+
+
+
+
+ `; + + // Add individual set cards to the group + const groupItems = groupContainer.querySelector('.set-group-items'); + setGroup.forEach(set => { + const itemContainer = document.createElement('div'); + itemContainer.className = 'col-12 mb-2'; + + // Clone the original card but make it smaller + const clonedCard = set.cardCol.querySelector('.card').cloneNode(true); + clonedCard.classList.add('set-group-item'); + + itemContainer.appendChild(clonedCard); + groupItems.appendChild(itemContainer); + + // Hide the original card + set.cardCol.style.display = 'none'; + set.cardCol.classList.add('grouped-set-hidden'); + }); + + // Insert the grouped container before the first hidden set + firstSet.cardCol.parentNode.insertBefore(groupContainer, firstSet.cardCol); + + // Add event listener to toggle arrow icon + const toggleBtn = groupContainer.querySelector('.group-toggle-btn'); + const collapseElement = groupContainer.querySelector('.collapse'); + + collapseElement.addEventListener('shown.bs.collapse', () => { + toggleBtn.querySelector('i').className = 'ri-arrow-down-line'; + }); + + collapseElement.addEventListener('hidden.bs.collapse', () => { + toggleBtn.querySelector('i').className = 'ri-arrow-right-line'; + }); +} + +function removeSetGrouping() { + // Show all hidden sets + const hiddenSets = document.querySelectorAll('.grouped-set-hidden'); + hiddenSets.forEach(setCol => { + setCol.style.display = ''; + setCol.classList.remove('grouped-set-hidden'); + }); + + // Remove all group containers + const groupContainers = document.querySelectorAll('.set-group-container'); + groupContainers.forEach(container => { + container.remove(); + }); +} + +// Initialize duplicate/consolidated filter functionality +function initializeDuplicateFilter() { + const duplicateFilterButton = document.getElementById('duplicate-filter-toggle'); + if (!duplicateFilterButton) return; + + // Check if the filter should be active from URL parameters + const urlParams = new URLSearchParams(window.location.search); + const isDuplicateFilterActive = urlParams.get('duplicate') === 'true'; + + // Set initial button state + if (isDuplicateFilterActive) { + duplicateFilterButton.classList.remove('btn-outline-secondary'); + duplicateFilterButton.classList.add('btn-secondary'); + } + + duplicateFilterButton.addEventListener('click', () => { + const isCurrentlyActive = duplicateFilterButton.classList.contains('btn-secondary'); + const newState = !isCurrentlyActive; + + // Update button appearance + if (newState) { + duplicateFilterButton.classList.remove('btn-outline-secondary'); + duplicateFilterButton.classList.add('btn-secondary'); + } else { + duplicateFilterButton.classList.remove('btn-secondary'); + duplicateFilterButton.classList.add('btn-outline-secondary'); + } + + if (isPaginationMode()) { + // SERVER-SIDE MODE - Update URL parameter + performDuplicateFilterServer(newState); + } else { + // CLIENT-SIDE MODE - Apply filtering directly + applyDuplicateFilter(newState); + } + }); +} + +// Server-side duplicate filter +function performDuplicateFilterServer(showOnlyDuplicates) { + const currentUrl = new URL(window.location); + + if (showOnlyDuplicates) { + currentUrl.searchParams.set('duplicate', 'true'); + } else { + currentUrl.searchParams.delete('duplicate'); + } + + // Reset to page 1 when filtering + currentUrl.searchParams.set('page', '1'); + window.location.href = currentUrl.toString(); +} + +// Apply duplicate/consolidated filter +function applyDuplicateFilter(showOnlyDuplicates) { + // Get the grid container and all column containers (not just the cards) + const gridContainer = document.getElementById('grid'); + if (!gridContainer) { + console.warn('Grid container not found'); + return; + } + + // Try multiple selectors to find column containers + let columnContainers = gridContainer.querySelectorAll('.col-md-6'); + if (columnContainers.length === 0) { + columnContainers = gridContainer.querySelectorAll('[class*="col-"]'); + } + + if (!showOnlyDuplicates) { + // Show all column containers by removing the duplicate-filter-hidden class + columnContainers.forEach(col => { + col.classList.remove('duplicate-filter-hidden'); + }); + // Trigger the existing grid filter to refresh + triggerGridRefresh(); + return; + } + + // Check if we're in consolidated mode by looking for data-instance-count + const consolidatedMode = document.querySelector('[data-instance-count]') !== null; + + if (consolidatedMode) { + // CONSOLIDATED MODE: Show only sets with instance count > 1 + columnContainers.forEach(col => { + const card = col.querySelector('[data-set-id]'); + if (card) { + const instanceCount = parseInt(card.dataset.instanceCount || '1'); + if (instanceCount > 1) { + col.classList.remove('duplicate-filter-hidden'); + } else { + col.classList.add('duplicate-filter-hidden'); + } + } + }); + } else { + // NON-CONSOLIDATED MODE: Show only sets that appear multiple times + const setByCounts = {}; + + // Count occurrences of each set + columnContainers.forEach(col => { + const card = col.querySelector('[data-set-id]'); + if (card) { + const rebrickableId = card.dataset.rebrickableId; + if (rebrickableId) { + setByCounts[rebrickableId] = (setByCounts[rebrickableId] || 0) + 1; + } + } + }); + + // Show/hide based on count + columnContainers.forEach(col => { + const card = col.querySelector('[data-set-id]'); + if (card) { + const rebrickableId = card.dataset.rebrickableId; + if (rebrickableId && setByCounts[rebrickableId] > 1) { + col.classList.remove('duplicate-filter-hidden'); + } else { + col.classList.add('duplicate-filter-hidden'); + } + } + }); + } + + // Trigger the existing grid filter to refresh and respect our duplicate filter + triggerGridRefresh(); +} + +// Helper function to trigger grid filter refresh +function triggerGridRefresh() { + // Check if we have a grid instance with filter capability + if (window.gridInstances) { + const gridElement = document.getElementById('grid'); + if (gridElement && window.gridInstances[gridElement.id]) { + const gridInstance = window.gridInstances[gridElement.id]; + if (gridInstance.filter) { + // Trigger the existing filter to refresh + gridInstance.filter.filter(); + } + } + } +} + +// Initialize clear filters button functionality +function initializeClearFiltersButton() { + const clearFiltersButton = document.getElementById('grid-filter-clear'); + if (!clearFiltersButton) return; + + clearFiltersButton.addEventListener('click', () => { + if (isPaginationMode()) { + // SERVER-SIDE PAGINATION MODE: Remove all filter parameters and redirect to base URL + const currentUrl = new URL(window.location); + + // Remove all filter parameters + const filterParams = ['status', 'theme', 'year', 'owner', 'purchase_location', 'storage', 'tag', 'duplicate']; + filterParams.forEach(param => { + currentUrl.searchParams.delete(param); + }); + + // Also remove page parameter to go back to clean base URL + currentUrl.searchParams.delete('page'); + + // Navigate to cleaned URL (will be just /sets if no other params) + window.location.href = currentUrl.toString(); + } else { + // CLIENT-SIDE MODE: Reset all filter dropdowns to empty string + const filterDropdowns = [ + 'grid-status', + 'grid-theme', + 'grid-year', + 'grid-owner', + 'grid-purchase-location', + 'grid-storage', + 'grid-tag' + ]; + + filterDropdowns.forEach(dropdownId => { + const dropdown = document.getElementById(dropdownId); + if (dropdown) { + dropdown.value = ''; + } + }); + + // Clear duplicate filter if active + const duplicateButton = document.getElementById('duplicate-filter-toggle'); + if (duplicateButton && duplicateButton.classList.contains('btn-secondary')) { + duplicateButton.classList.remove('btn-secondary'); + duplicateButton.classList.add('btn-outline-secondary'); + applyDuplicateFilter(false); + } + + // Remove page parameter from URL if present (without reloading) + const currentUrl = new URL(window.location); + if (currentUrl.searchParams.has('page')) { + currentUrl.searchParams.delete('page'); + window.history.replaceState({}, '', currentUrl.toString()); + } + + // Trigger filtering if grid instance exists + triggerGridRefresh(); + } + }); +} \ No newline at end of file diff --git a/static/scripts/socket/peeron.js b/static/scripts/socket/peeron.js new file mode 100644 index 0000000..9a94104 --- /dev/null +++ b/static/scripts/socket/peeron.js @@ -0,0 +1,206 @@ +// Peeron Socket class +class BrickPeeronSocket extends BrickSocket { + constructor(id, path, namespace, messages) { + super(id, path, namespace, messages, true); + + // Form elements (built based on the initial id) + this.html_button = document.getElementById(id); + this.html_files = document.getElementById(`${id}-files`); + + if (this.html_button) { + this.html_button.addEventListener("click", (e) => { + this.execute(); + }); + } + + // Setup select all button + this.setup_select_all_button(); + + // Setup rotation buttons + this.setup_rotation_buttons(); + + // Setup the socket + this.setup(); + } + + setup_select_all_button() { + const selectAllButton = document.getElementById('peeron-select-all'); + if (selectAllButton) { + selectAllButton.addEventListener('click', () => { + const checkboxes = this.get_files(); + const allChecked = checkboxes.every(cb => cb.checked); + checkboxes.forEach(cb => cb.checked = !allChecked); + + // Update button text and icon + if (allChecked) { + selectAllButton.innerHTML = ' Select All'; + } else { + selectAllButton.innerHTML = ' Deselect All'; + } + }); + } + } + + setup_rotation_buttons() { + document.querySelectorAll('.peeron-rotate-btn').forEach(button => { + button.addEventListener('click', (e) => { + e.preventDefault(); // Prevent label click + e.stopPropagation(); // Stop event bubbling + + const targetId = button.getAttribute('data-target'); + const checkboxId = button.getAttribute('data-checkbox'); + const targetImg = document.getElementById(targetId); + const checkbox = document.getElementById(checkboxId); + + if (targetImg && checkbox) { + let currentRotation = parseInt(button.getAttribute('data-rotation') || '0'); + currentRotation = (currentRotation + 90) % 360; + + // Update image rotation + targetImg.style.transform = `rotate(${currentRotation}deg)`; + button.setAttribute('data-rotation', currentRotation.toString()); + + // Store rotation in checkbox data for later use + checkbox.setAttribute('data-rotation', currentRotation.toString()); + + // Update the rotation icon to indicate current state + const icon = button.querySelector('i'); + if (icon) { + // Rotate the icon to match the image rotation + icon.style.transform = `rotate(${currentRotation}deg)`; + } + } + }); + }); + } + + // Upon receiving a complete message + complete(data) { + super.complete(data); + + // Clear progress display after completion + if (this.html_progress_message) { + this.html_progress_message.classList.add("d-none"); + this.html_progress_message.textContent = ""; + } + + if (this.html_count) { + this.html_count.classList.add("d-none"); + this.html_count.textContent = ""; + } + + // Ensure spinner is hidden + this.spinner(false); + + this.toggle(true); + } + + // Execute the action + execute() { + if (!this.disabled && this.socket !== undefined && this.socket.connected) { + this.toggle(false); + + this.download_peeron_pages(); + } + } + + // Get the list of checkboxes describing files + get_files(checked=false) { + let files = []; + + if (this.html_files) { + files = [...this.html_files.querySelectorAll('input[type="checkbox"]')]; + + if (checked) { + files = files.filter(file => file.checked); + } + } + + return files; + } + + // Download Peeron pages + download_peeron_pages() { + if (this.html_files) { + const selectedFiles = this.get_files(true); + + if (selectedFiles.length === 0) { + this.fail({message: "Please select at least one page to download."}); + this.toggle(true); + return; + } + + const pages = selectedFiles.map(checkbox => ({ + page_number: checkbox.getAttribute('data-page-number'), + original_image_url: checkbox.getAttribute('data-original-image-url'), + cached_full_image_path: checkbox.getAttribute('data-cached-full-image-path'), + alt_text: checkbox.getAttribute('data-alt-text'), + rotation: parseInt(checkbox.getAttribute('data-rotation') || '0') + })); + + this.clear(); + this.spinner(true); + + const setElement = document.querySelector('input[name="download-set"]'); + const set = setElement ? setElement.value : ''; + + this.socket.emit(this.messages.DOWNLOAD_PEERON_PAGES, { + set: set, + pages: pages, + total: pages.length, + current: 0 + }); + } else { + this.fail({message: "Could not find the list of pages to download"}); + } + } + + // Toggle clicking on the button, or sending events + toggle(enabled) { + super.toggle(enabled); + + if (this.html_files) { + this.get_files().forEach(el => el.disabled = !enabled); + } + + if (this.html_button) { + this.html_button.disabled = !enabled; + } + } +} + +// Simple Peeron page loader using standard socket pattern +class BrickPeeronPageLoader extends BrickSocket { + constructor(set, path, namespace, messages) { + // Use 'peeron-loader' as the ID for socket elements + super('peeron-loader', path, namespace, messages, false); + + this.set = set; + this.setup(); + + // Auto-start loading when connected + setTimeout(() => { + if (this.socket && this.socket.connected) { + this.loadPages(); + } else { + this.socket.on('connect', () => this.loadPages()); + } + }, 100); + } + + loadPages() { + this.socket.emit(this.messages.LOAD_PEERON_PAGES, { + set: this.set + }); + } + + // Override complete to redirect when done + complete(data) { + super.complete(data); + // Redirect to show the pages selection interface + const params = new URLSearchParams(); + params.set('set', this.set); + params.set('peeron_loaded', '1'); + window.location.href = `${window.location.pathname}?${params.toString()}`; + } +} \ No newline at end of file diff --git a/static/scripts/socket/socket.js b/static/scripts/socket/socket.js index cb08889..e67ab55 100644 --- a/static/scripts/socket/socket.js +++ b/static/scripts/socket/socket.js @@ -186,7 +186,15 @@ class BrickSocket { if (this.socket === undefined) { this.socket = io.connect(`${window.location.origin}/${this.namespace}`, { path: this.path, - transports: ["websocket"], + // Adding Polling as iOS is having issues with websockets + transports: ["polling", "websocket"], + // Increase timeout for slow mobile connections + timeout: 30000, + // Reconnection settings + reconnectionDelay: 1000, + reconnectionDelayMax: 5000, + // Allow upgrade from polling to websocket + upgrade: true, }); // Complete diff --git a/static/scripts/statistics.js b/static/scripts/statistics.js new file mode 100644 index 0000000..df2ca3a --- /dev/null +++ b/static/scripts/statistics.js @@ -0,0 +1,170 @@ +/** + * Statistics page chart functionality + */ + +document.addEventListener('DOMContentLoaded', function() { + // Check if charts are enabled and chart data exists + if (typeof window.chartData === 'undefined') { + return; + } + + // Debug: Log chart data to console + console.log('Chart Data:', window.chartData); + + // Get current theme colors + const isDarkMode = document.documentElement.getAttribute('data-bs-theme') === 'dark'; + const textColor = isDarkMode ? '#f8f9fa' : '#212529'; + const gridColor = isDarkMode ? 'rgba(255, 255, 255, 0.1)' : 'rgba(0, 0, 0, 0.1)'; + + // Common chart configuration + const commonConfig = { + type: 'line', + options: { + responsive: true, + maintainAspectRatio: false, + interaction: { + intersect: false, + mode: 'index' + }, + scales: { + x: { + title: { + display: true, + text: 'Release Year', + color: textColor + }, + ticks: { + color: textColor + }, + grid: { + display: false, + color: gridColor + } + }, + y: { + beginAtZero: true, + title: { + display: true, + text: 'Count', + color: textColor + }, + ticks: { + precision: 0, + color: textColor + }, + grid: { + color: gridColor + } + } + }, + plugins: { + legend: { + display: false + }, + tooltip: { + backgroundColor: 'rgba(0, 0, 0, 0.8)', + titleColor: 'white', + bodyColor: 'white', + cornerRadius: 4 + } + }, + elements: { + point: { + radius: 3, + hoverRadius: 5 + }, + line: { + borderWidth: 2 + } + } + } + }; + + // Sets Chart + const setsCanvas = document.getElementById('setsChart'); + if (setsCanvas) { + const setsCtx = setsCanvas.getContext('2d'); + new Chart(setsCtx, { + ...commonConfig, + data: { + labels: window.chartData.years, + datasets: [{ + label: 'Sets', + data: window.chartData.sets, + borderColor: '#0d6efd', + backgroundColor: 'rgba(13, 110, 253, 0.1)', + fill: true, + tension: 0.3 + }] + } + }); + } + + // Parts Chart + const partsCanvas = document.getElementById('partsChart'); + if (partsCanvas) { + const partsCtx = partsCanvas.getContext('2d'); + new Chart(partsCtx, { + ...commonConfig, + data: { + labels: window.chartData.years, + datasets: [{ + label: 'Parts', + data: window.chartData.parts, + borderColor: '#198754', + backgroundColor: 'rgba(25, 135, 84, 0.1)', + fill: true, + tension: 0.3 + }] + }, + options: { + ...commonConfig.options, + scales: { + ...commonConfig.options.scales, + y: { + ...commonConfig.options.scales.y, + title: { + display: true, + text: 'Parts Count', + color: textColor + } + } + } + } + }); + } + + // Minifigures Chart + const minifigsCanvas = document.getElementById('minifigsChart'); + if (minifigsCanvas) { + const minifigsCtx = minifigsCanvas.getContext('2d'); + new Chart(minifigsCtx, { + ...commonConfig, + data: { + labels: window.chartData.years, + datasets: [{ + label: 'Minifigures', + data: window.chartData.minifigs, + borderColor: '#fd7e14', + backgroundColor: 'rgba(253, 126, 20, 0.1)', + fill: true, + tension: 0.3 + }] + }, + options: { + ...commonConfig.options, + scales: { + ...commonConfig.options.scales, + y: { + ...commonConfig.options.scales.y, + title: { + display: true, + text: 'Minifigures Count', + color: textColor + } + } + } + } + }); + } +}); \ No newline at end of file diff --git a/static/scripts/table.js b/static/scripts/table.js index f9cf06f..089219e 100644 --- a/static/scripts/table.js +++ b/static/scripts/table.js @@ -36,14 +36,19 @@ window.BrickTable = class BrickTable { // Special configuration for tables with custom search/sort const isMinifiguresTable = table.id === 'minifigures'; const isPartsTable = table.id === 'parts'; - const hasCustomInterface = isMinifiguresTable || isPartsTable; + const isProblemsTable = table.id === 'problems'; + const isPartsTablePaginationMode = isPartsTable && table.getAttribute('data-table') === 'false'; + const isProblemsTablePaginationMode = isProblemsTable && table.getAttribute('data-table') === 'false'; + const hasCustomInterface = isMinifiguresTable || isPartsTablePaginationMode || isProblemsTablePaginationMode; + const hasCustomSearch = isPartsTable || isProblemsTable || isMinifiguresTable; // Parts, problems, and minifigures always have custom search this.table = new simpleDatatables.DataTable(`#${table.id}`, { columns: columns, + paging: !(isPartsTablePaginationMode || isProblemsTablePaginationMode), // Disable built-in pagination for tables in pagination mode pagerDelta: 1, perPage: per_page, perPageSelect: [10, 25, 50, 100, 500, 1000], - searchable: !hasCustomInterface, // Disable built-in search for tables with custom interface + searchable: !hasCustomSearch, // Disable built-in search for tables with custom search searchMethod: (table => (terms, cell, row, column, source) => table.search(terms, cell, row, column, source))(this), searchQuerySeparator: "", tableRender: () => { @@ -105,6 +110,8 @@ const setup_tables = (per_page) => document.querySelectorAll('table[data-table=" window.brickTableInstance = brickTable; } else if (el.id === 'parts') { window.partsTableInstance = brickTable; + } else if (el.id === 'problems') { + window.problemsTableInstance = brickTable; } } ); diff --git a/static/styles.css b/static/styles.css index 0651e20..857c3c6 100644 --- a/static/styles.css +++ b/static/styles.css @@ -50,12 +50,88 @@ max-width: 150px; } +/* Checkbox column width constraint */ +.table-td-input:has(.form-check-input[type="checkbox"]) { + width: 120px; + max-width: 120px; + min-width: 120px; +} + +/* Reserve space for status icon to prevent layout shift */ +.form-check-label i[id^="status-"] { + display: inline-block; + width: 1.2em; + text-align: center; + margin-left: 0.25rem; +} + +/* Hamburger menu styling */ +.table th .dropdown { + position: relative; +} + +.table th .dropdown-toggle { + border-radius: 0.375rem; + padding: 0.25rem 0.5rem; + font-size: 0.875rem; + line-height: 1.25rem; + border-color: #6c757d; +} + +.table th .dropdown-toggle:focus { + box-shadow: 0 0 0 0.25rem rgba(13, 110, 253, 0.25); +} + +.table th .dropdown-toggle:hover { + background-color: #f8f9fa; + border-color: #6c757d; +} + +[data-bs-theme="dark"] .table th .dropdown-toggle:hover { + background-color: #2b3035; + border-color: #6c757d; +} + +/* Style dropdown items */ +.dropdown-menu .dropdown-header { + font-size: 0.75rem; + font-weight: 600; + text-transform: uppercase; + color: #6c757d; + padding: 0.25rem 1rem; +} + +.dropdown-menu .dropdown-item { + font-size: 0.875rem; + padding: 0.5rem 1rem; +} + +.dropdown-menu .dropdown-item:hover { + background-color: #f8f9fa; + color: #212529; +} + +[data-bs-theme="dark"] .dropdown-menu .dropdown-item:hover { + background-color: #2b3035; + color: #fff; +} + +.dropdown-menu .dropdown-item i { + width: 1.25rem; + text-align: center; +} + /* Fixes for sortable.js */ .sortable { --th-color: #000 !important; --th-bg:#fff !important; } +[data-bs-theme="dark"] .sortable { + --th-color: #fff !important; + --th-bg: #212529 !important; +} + .sortable thead th { font-weight: bold !important; } @@ -77,4 +153,122 @@ linear-gradient(127deg, rgb(0 255 0 / 80%), rgb(0 255 0 / 0%) 70.71%), linear-gradient(336deg, rgb(0 0 255 / 80%), rgb(0 0 255 / 0%) 70.71%) ; -} \ No newline at end of file +} + +/* Mobile Pagination Fixes */ +.mobile-pagination { + display: flex !important; + width: 100% !important; + gap: 2px; +} + +.mobile-pagination .btn:first-child, +.mobile-pagination .btn:last-child { + flex: 0 0 auto; + min-width: 90px; + max-width: 120px; +} + +.mobile-pagination .btn:nth-child(2) { + flex: 1 1 auto; + white-space: nowrap; + overflow: hidden; + text-overflow: ellipsis; + min-width: 0; +} + +/* Partial Status Checkbox Styling */ +.partial-status { + position: relative; +} + +.partial-status::after { + content: ''; + position: absolute; + top: 2px; + left: 2px; + right: 2px; + bottom: 2px; + background: #6f42c1; + border-radius: 2px; + opacity: 0.8; + pointer-events: none; +}/* Duplicate filter support */ +.duplicate-filter-hidden { display: none !important; } + +/* Dark mode link and text fixes */ +[data-bs-theme="dark"] .text-reset { + color: var(--bs-body-color) !important; +} + +[data-bs-theme="dark"] a.text-reset:hover { + color: var(--bs-link-hover-color) !important; +} + +[data-bs-theme="dark"] .text-dark { + color: var(--bs-body-color) !important; +} + +/* Dark mode border fixes */ +[data-bs-theme="dark"] .border-black { + border-color: var(--bs-border-color) !important; +} + +/* Dark mode badge fixes */ +[data-bs-theme="dark"] .text-bg-light { + background-color: #343a40 !important; + color: #f8f9fa !important; +} + +[data-bs-theme="dark"] .bg-white { + background-color: #212529 !important; +} + +[data-bs-theme="dark"] .text-black { + color: #f8f9fa !important; +} + +/* Dark mode button fixes */ +[data-bs-theme="dark"] .btn-outline-dark { + border-color: #f8f9fa; + color: #f8f9fa; +} + +[data-bs-theme="dark"] .btn-outline-dark:hover { + background-color: #f8f9fa; + border-color: #f8f9fa; + color: #212529; +} + +[data-bs-theme="dark"] .btn-light { + background-color: #343a40; + border-color: #495057; + color: #f8f9fa; +} + +[data-bs-theme="dark"] .btn-light:hover { + background-color: #495057; + border-color: #6c757d; + color: #fff; +} + +/* Dark mode table fixes */ +[data-bs-theme="dark"] .table-light { + --bs-table-bg: #343a40; + --bs-table-color: #f8f9fa; +} + +/* Dark mode for card header links */ +[data-bs-theme="dark"] .text-decoration-none.text-dark { + color: var(--bs-body-color) !important; +} + +/* Dark mode table text fix - ensure all table text is visible */ +[data-bs-theme="dark"] .table { + --bs-table-color: var(--bs-body-color); +} + +[data-bs-theme="dark"] .table tbody td, +[data-bs-theme="dark"] .table tbody th { + color: var(--bs-body-color); +} diff --git a/templates/admin/configuration.html b/templates/admin/configuration.html index 9ef4496..25c3f78 100644 --- a/templates/admin/configuration.html +++ b/templates/admin/configuration.html @@ -1,29 +1,966 @@ {% import 'macro/accordion.html' as accordion %} -{{ accordion.header('Configuration variables', 'configuration', 'admin', icon='list-settings-line') }} -
    - {% for entry in configuration %} -
  • - {{ entry.name }}: - {% if entry.value == none or entry.value == '' %} - Unset - {% elif entry.value == true %} - True - {% elif entry.value == false %} - False - {% else %} - {% if entry.is_secret() %} - Set - {% else %} - {{ entry.value }} - {% endif %} - {% endif %} - Env: {{ entry.env_name }} - {% if entry.extra_name %}Env: {{ entry.extra_name }}{% endif %} - {% if entry.is_changed() %} - Changed - {% endif %} -
  • - {% endfor %} -
+ +{% macro is_locked(var_name) %}{{ 'disabled' if env_locked_values[var_name] else '' }}{% endmacro %} + + +{% macro config_badges(var_name) %} + {% set current_value = env_values[var_name] %} + {% set default_value = config_defaults[var_name] %} + {% set is_explicitly_set = env_explicit_values[var_name] %} + {% set is_locked = env_locked_values[var_name] %} + + + {% if is_locked %} + + Locked + + {% endif %} + + + {% if not is_explicitly_set %} + + Unset + Default Value + {% elif current_value is sameas true %} + True + {% elif current_value is sameas false %} + False + {% elif current_value == default_value %} + + Default: {{ default_value }} + {% else %} + + Default: {{ default_value }} + {% endif %} + + + {% if current_value != default_value and not is_locked %} + Changed + {% endif %} +{% endmacro %} + + +
+ +
+ + +{% if env_file_warning %} +
+
Configuration Persistence Warning
+

+ Configuration file location: {{ env_file_location }} +

+

+ + Settings changes will not persist across container restarts. + To persist configuration changes, either: +

    +
  • Move .env to data/.env folder (recommended), or
  • +
  • Add volume mount: - ./.env:/app/.env to your compose.yaml
  • +
+ See Migration Guide for details. + +

+
+{% elif env_file_missing %} +
+
No Configuration File Found
+

+ + No .env file was found at data/.env. Currently using default values and Docker environment variables. + If you save changes, a new configuration file will be created at data/.env (recommended location). + See Migration Guide for details. + +

+
+{% endif %} + + +
+
Badge Legend
+
+
+ + True Boolean setting enabled
+ False Boolean setting disabled
+ Set Custom value configured
+ Locked Set via Docker environment (cannot be changed) +
+
+
+ + Unset Not in .env file
+ Default Value Using default value
+ Changed Modified from default +
+
+
+
+ + +{{ accordion.header('Live Settings', 'live-settings', 'admin', expanded=true, icon='settings-4-line', class='p-3') }} +
+ Changes Applied On Save +
+ +
+ + + +
+ + + + +
Menu Visibility
+ +
+
+
+ + +
+ +
+ + +
+ +
+ + +
+ +
+ + +
+ +
+ + +
+ +
+ + +
+
+ +
+
+ + +
+ +
+ + +
+ +
+ + +
+ +
+ + +
+ +
+ + +
+
+
+ + +
Table Display
+ +
+
+
+ + +
+ +
+ + +
+ +
+ + +
+ +
+ + +
+
+ +
+
+ + +
+ +
+ + +
+ +
+ + +
+ +
+ + +
+ +
+ + +
+
+
+ + +
Pagination Settings
+ + +
+ +
+
Sets
+
+
+
+ + +
+
+
+ + +
+
+ + +
+
+
+ + +
+
Parts
+
+
+
+ + +
+
+
+ + +
+
+ + +
+
+
+
+ + +
+ +
+
Minifigures
+
+
+
+ + +
+
+
+ + +
+
+ + +
+
+
+ + +
+
Problems
+
+
+
+ + +
+
+
+ + +
+
+ + +
+
+
+
+ + +
Client side pagination
+
+
+ + +
+
+ + +
Features & External Services
+ +
+
+
+ + +
+ +
+ + +
+ +
+ + +
+ +
+ + +
+ +
+ + +
+
+ +
+
+ + +
+ +
+ + +
+ +
+ + +
+ +
+ + +
+
+
+ + +
Advanced Settings
+ +
+
+ + +
+ +
+ + +
+ +
+ + +
+ +
+ + +
+
+ + +
Default Ordering & Formatting
+ +
+
+ + +
+ +
+ + +
+ +
+ + +
+ +
+ + +
+ + +
+ + +
+ +
+ + +
+ +
+ + +
+
+ + +
URL Patterns & Links
+ +
+
+ + +
+ +
+ + +
+ +
+ + +
+ +
+ + +
+ +
+ + +
+ +
+ + +
+ +
+ + +
+
+ + +
Images & Resources
+ +
+
+ + +
+ +
+ + +
+ +
+ + +
+ +
+ + +
+ +
+ + +
+ +
+ + +
+
{{ accordion.footer() }} + + +{{ accordion.header('Static Settings', 'static-settings', 'admin', expanded=false, icon='database-2-line', class='p-3') }} +
+ Requires Restart +
+ + +
+
Restart Required
+ These settings require an application restart to take effect. Values can be edited here and will be saved to the .env file. +
+ + +
Authentication & Security
+
+
+ + +
+
+ + +
+
+ + +
Server Configuration
+
+
+ + +
+
+ + +
+
+
+ + +
+
+
+ + +
+
+ + +
+
+ + +
Database & Storage
+
+
+ + +
+
+ + +
+
+ + +
+
+ + +
+
+ + +
+
+ + +
API Configuration
+
+
+ + +
+
+ + +
+
+ + +
+
+ + +
Date & Currency Formats
+
+
+ + +
+
+ + +
+
+ + +
+
+ + +
+
+ + +
Socket Configuration
+
+
+ + +
+
+ + +
+
+
+ + +
+
+
+ + +
+ +
+{{ accordion.footer() }} + + + + + + diff --git a/templates/admin/database/upgrade.html b/templates/admin/database/upgrade.html index 29e59ef..8e62f4c 100644 --- a/templates/admin/database/upgrade.html +++ b/templates/admin/database/upgrade.html @@ -6,6 +6,10 @@ +
Upgrades
    diff --git a/templates/base.html b/templates/base.html index 6b4a4d2..968f6ca 100644 --- a/templates/base.html +++ b/templates/base.html @@ -1,5 +1,5 @@ - + @@ -12,6 +12,7 @@ +