Fix Wikipedia API with User-Agent (#822)
* refactor(serializers): remove unused gpxpy and geojson imports * fix(generate_description): improve error handling and response validation for Wikipedia API calls * Potential fix for code scanning alert no. 42: Information exposure through an exception Co-authored-by: Copilot Autofix powered by AI <62310815+github-advanced-security[bot]@users.noreply.github.com> * fix(generate_description): improve error logging for Wikipedia API data fetch failures * chore(deps): bump devalue (#823) Bumps the npm_and_yarn group with 1 update in the /frontend directory: [devalue](https://github.com/sveltejs/devalue). Updates `devalue` from 5.1.1 to 5.3.2 - [Release notes](https://github.com/sveltejs/devalue/releases) - [Changelog](https://github.com/sveltejs/devalue/blob/main/CHANGELOG.md) - [Commits](https://github.com/sveltejs/devalue/compare/v5.1.1...v5.3.2) --- updated-dependencies: - dependency-name: devalue dependency-version: 5.3.2 dependency-type: indirect dependency-group: npm_and_yarn ... Signed-off-by: dependabot[bot] <support@github.com> Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> Co-authored-by: Sean Morley <98704938+seanmorley15@users.noreply.github.com> * Refactor help documentation link in settings page - Updated the condition to display the help documentation link based on the `wandererEnabled` flag. - Removed the conditional rendering for staff users and Strava integration status. - Changed the documentation link to point to the Immich integration documentation. * fix(locations): update include_collections parameter handling for default behavior * Update backend/server/adventures/views/generate_description_view.py Co-authored-by: Copilot <175728472+Copilot@users.noreply.github.com> --------- Signed-off-by: dependabot[bot] <support@github.com> Co-authored-by: Copilot Autofix powered by AI <62310815+github-advanced-security[bot]@users.noreply.github.com> Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> Co-authored-by: Copilot <175728472+Copilot@users.noreply.github.com>
This commit is contained in:
@@ -7,8 +7,6 @@ from worldtravel.serializers import CountrySerializer, RegionSerializer, CitySer
|
|||||||
from geopy.distance import geodesic
|
from geopy.distance import geodesic
|
||||||
from integrations.models import ImmichIntegration
|
from integrations.models import ImmichIntegration
|
||||||
from adventures.utils.geojson import gpx_to_geojson
|
from adventures.utils.geojson import gpx_to_geojson
|
||||||
import gpxpy
|
|
||||||
import geojson
|
|
||||||
import logging
|
import logging
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
|||||||
@@ -3,42 +3,137 @@ from rest_framework.decorators import action
|
|||||||
from rest_framework.permissions import IsAuthenticated
|
from rest_framework.permissions import IsAuthenticated
|
||||||
from rest_framework.response import Response
|
from rest_framework.response import Response
|
||||||
import requests
|
import requests
|
||||||
|
from django.conf import settings
|
||||||
|
import urllib.parse
|
||||||
|
import logging
|
||||||
|
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
class GenerateDescription(viewsets.ViewSet):
|
class GenerateDescription(viewsets.ViewSet):
|
||||||
permission_classes = [IsAuthenticated]
|
permission_classes = [IsAuthenticated]
|
||||||
|
|
||||||
|
# User-Agent header required by Wikipedia API
|
||||||
|
HEADERS = {
|
||||||
|
'User-Agent': f'AdventureLog/{getattr(settings, "ADVENTURELOG_RELEASE_VERSION", "unknown")}'
|
||||||
|
}
|
||||||
|
|
||||||
@action(detail=False, methods=['get'],)
|
@action(detail=False, methods=['get'])
|
||||||
def desc(self, request):
|
def desc(self, request):
|
||||||
name = self.request.query_params.get('name', '')
|
name = self.request.query_params.get('name', '')
|
||||||
# un url encode the name
|
if not name:
|
||||||
name = name.replace('%20', ' ')
|
return Response({"error": "Name parameter is required"}, status=400)
|
||||||
name = self.get_search_term(name)
|
|
||||||
url = 'https://en.wikipedia.org/w/api.php?origin=*&action=query&prop=extracts&exintro&explaintext&format=json&titles=%s' % name
|
# Properly URL decode the name
|
||||||
response = requests.get(url)
|
name = urllib.parse.unquote(name)
|
||||||
data = response.json()
|
search_term = self.get_search_term(name)
|
||||||
data = response.json()
|
|
||||||
page_id = next(iter(data["query"]["pages"]))
|
if not search_term:
|
||||||
extract = data["query"]["pages"][page_id]
|
return Response({"error": "No matching Wikipedia article found"}, status=404)
|
||||||
if extract.get('extract') is None:
|
|
||||||
return Response({"error": "No description found"}, status=400)
|
# Properly URL encode the search term for the API
|
||||||
return Response(extract)
|
encoded_term = urllib.parse.quote(search_term)
|
||||||
@action(detail=False, methods=['get'],)
|
url = f'https://en.wikipedia.org/w/api.php?origin=*&action=query&prop=extracts&exintro&explaintext&format=json&titles={encoded_term}'
|
||||||
|
|
||||||
|
try:
|
||||||
|
response = requests.get(url, headers=self.HEADERS, timeout=10)
|
||||||
|
response.raise_for_status()
|
||||||
|
data = response.json()
|
||||||
|
|
||||||
|
pages = data.get("query", {}).get("pages", {})
|
||||||
|
if not pages:
|
||||||
|
return Response({"error": "No page data found"}, status=404)
|
||||||
|
|
||||||
|
page_id = next(iter(pages))
|
||||||
|
page_data = pages[page_id]
|
||||||
|
|
||||||
|
# Check if page exists (page_id of -1 means page doesn't exist)
|
||||||
|
if page_id == "-1":
|
||||||
|
return Response({"error": "Wikipedia page not found"}, status=404)
|
||||||
|
|
||||||
|
if not page_data.get('extract'):
|
||||||
|
return Response({"error": "No description found"}, status=404)
|
||||||
|
|
||||||
|
return Response(page_data)
|
||||||
|
|
||||||
|
except requests.exceptions.RequestException as e:
|
||||||
|
logger.exception("Failed to fetch data from Wikipedia")
|
||||||
|
return Response({"error": "Failed to fetch data from Wikipedia."}, status=500)
|
||||||
|
except ValueError as e: # JSON decode error
|
||||||
|
return Response({"error": "Invalid response from Wikipedia API"}, status=500)
|
||||||
|
|
||||||
|
@action(detail=False, methods=['get'])
|
||||||
def img(self, request):
|
def img(self, request):
|
||||||
name = self.request.query_params.get('name', '')
|
name = self.request.query_params.get('name', '')
|
||||||
# un url encode the name
|
if not name:
|
||||||
name = name.replace('%20', ' ')
|
return Response({"error": "Name parameter is required"}, status=400)
|
||||||
name = self.get_search_term(name)
|
|
||||||
url = 'https://en.wikipedia.org/w/api.php?origin=*&action=query&prop=pageimages&format=json&piprop=original&titles=%s' % name
|
# Properly URL decode the name
|
||||||
response = requests.get(url)
|
name = urllib.parse.unquote(name)
|
||||||
data = response.json()
|
search_term = self.get_search_term(name)
|
||||||
page_id = next(iter(data["query"]["pages"]))
|
|
||||||
extract = data["query"]["pages"][page_id]
|
if not search_term:
|
||||||
if extract.get('original') is None:
|
return Response({"error": "No matching Wikipedia article found"}, status=404)
|
||||||
return Response({"error": "No image found"}, status=400)
|
|
||||||
return Response(extract["original"])
|
# Properly URL encode the search term for the API
|
||||||
|
encoded_term = urllib.parse.quote(search_term)
|
||||||
|
url = f'https://en.wikipedia.org/w/api.php?origin=*&action=query&prop=pageimages&format=json&piprop=original&titles={encoded_term}'
|
||||||
|
|
||||||
|
try:
|
||||||
|
response = requests.get(url, headers=self.HEADERS, timeout=10)
|
||||||
|
response.raise_for_status()
|
||||||
|
data = response.json()
|
||||||
|
|
||||||
|
pages = data.get("query", {}).get("pages", {})
|
||||||
|
if not pages:
|
||||||
|
return Response({"error": "No page data found"}, status=404)
|
||||||
|
|
||||||
|
page_id = next(iter(pages))
|
||||||
|
page_data = pages[page_id]
|
||||||
|
|
||||||
|
# Check if page exists
|
||||||
|
if page_id == "-1":
|
||||||
|
return Response({"error": "Wikipedia page not found"}, status=404)
|
||||||
|
|
||||||
|
original_image = page_data.get('original')
|
||||||
|
if not original_image:
|
||||||
|
return Response({"error": "No image found"}, status=404)
|
||||||
|
|
||||||
|
return Response(original_image)
|
||||||
|
|
||||||
|
except requests.exceptions.RequestException as e:
|
||||||
|
logger.exception("Failed to fetch data from Wikipedia")
|
||||||
|
return Response({"error": "Failed to fetch data from Wikipedia."}, status=500)
|
||||||
|
except ValueError as e: # JSON decode error
|
||||||
|
return Response({"error": "Invalid response from Wikipedia API"}, status=500)
|
||||||
|
|
||||||
def get_search_term(self, term):
|
def get_search_term(self, term):
|
||||||
response = requests.get(f'https://en.wikipedia.org/w/api.php?action=opensearch&search={term}&limit=10&namespace=0&format=json')
|
if not term:
|
||||||
data = response.json()
|
return None
|
||||||
if data[1] and len(data[1]) > 0:
|
|
||||||
return data[1][0]
|
# Properly URL encode the search term
|
||||||
|
encoded_term = urllib.parse.quote(term)
|
||||||
|
url = f'https://en.wikipedia.org/w/api.php?action=opensearch&search={encoded_term}&limit=10&namespace=0&format=json'
|
||||||
|
|
||||||
|
try:
|
||||||
|
response = requests.get(url, headers=self.HEADERS, timeout=10)
|
||||||
|
response.raise_for_status()
|
||||||
|
|
||||||
|
# Check if response is empty
|
||||||
|
if not response.text.strip():
|
||||||
|
return None
|
||||||
|
|
||||||
|
data = response.json()
|
||||||
|
|
||||||
|
# OpenSearch API returns an array with 4 elements:
|
||||||
|
# [search_term, [titles], [descriptions], [urls]]
|
||||||
|
if len(data) >= 2 and data[1] and len(data[1]) > 0:
|
||||||
|
return data[1][0] # Return the first title match
|
||||||
|
|
||||||
|
return None
|
||||||
|
|
||||||
|
except requests.exceptions.RequestException:
|
||||||
|
# If search fails, return the original term as fallback
|
||||||
|
return term
|
||||||
|
except ValueError: # JSON decode error
|
||||||
|
# If JSON parsing fails, return the original term as fallback
|
||||||
|
return term
|
||||||
8
frontend/pnpm-lock.yaml
generated
8
frontend/pnpm-lock.yaml
generated
@@ -853,8 +853,8 @@ packages:
|
|||||||
resolution: {integrity: sha512-3UDv+G9CsCKO1WKMGw9fwq/SWJYbI0c5Y7LU1AXYoDdbhE2AHQ6N6Nb34sG8Fj7T5APy8qXDCKuuIHd1BR0tVA==}
|
resolution: {integrity: sha512-3UDv+G9CsCKO1WKMGw9fwq/SWJYbI0c5Y7LU1AXYoDdbhE2AHQ6N6Nb34sG8Fj7T5APy8qXDCKuuIHd1BR0tVA==}
|
||||||
engines: {node: '>=8'}
|
engines: {node: '>=8'}
|
||||||
|
|
||||||
devalue@5.1.1:
|
devalue@5.3.2:
|
||||||
resolution: {integrity: sha512-maua5KUiapvEwiEAe+XnlZ3Rh0GD+qI1J/nb9vrJc3muPXvcF/8gXYTWF76+5DAqHyDUtOIImEuo0YKE9mshVw==}
|
resolution: {integrity: sha512-UDsjUbpQn9kvm68slnrs+mfxwFkIflOhkanmyabZ8zOYk8SMEIbJ3TK+88g70hSIeytu4y18f0z/hYHMTrXIWw==}
|
||||||
|
|
||||||
didyoumean@1.2.2:
|
didyoumean@1.2.2:
|
||||||
resolution: {integrity: sha512-gxtyfqMg7GKyhQmb056K7M3xszy/myH8w+B4RT+QXBQsvAOdc3XymqDDPHx1BgPgsdAA5SIifona89YtRATDzw==}
|
resolution: {integrity: sha512-gxtyfqMg7GKyhQmb056K7M3xszy/myH8w+B4RT+QXBQsvAOdc3XymqDDPHx1BgPgsdAA5SIifona89YtRATDzw==}
|
||||||
@@ -2235,7 +2235,7 @@ snapshots:
|
|||||||
'@sveltejs/vite-plugin-svelte': 3.1.2(svelte@4.2.19)(vite@5.4.19(@types/node@22.15.2))
|
'@sveltejs/vite-plugin-svelte': 3.1.2(svelte@4.2.19)(vite@5.4.19(@types/node@22.15.2))
|
||||||
'@types/cookie': 0.6.0
|
'@types/cookie': 0.6.0
|
||||||
cookie: 0.6.0
|
cookie: 0.6.0
|
||||||
devalue: 5.1.1
|
devalue: 5.3.2
|
||||||
esm-env: 1.2.2
|
esm-env: 1.2.2
|
||||||
import-meta-resolve: 4.1.0
|
import-meta-resolve: 4.1.0
|
||||||
kleur: 4.1.5
|
kleur: 4.1.5
|
||||||
@@ -2535,7 +2535,7 @@ snapshots:
|
|||||||
|
|
||||||
detect-libc@2.0.4: {}
|
detect-libc@2.0.4: {}
|
||||||
|
|
||||||
devalue@5.1.1: {}
|
devalue@5.3.2: {}
|
||||||
|
|
||||||
didyoumean@1.2.2: {}
|
didyoumean@1.2.2: {}
|
||||||
|
|
||||||
|
|||||||
File diff suppressed because it is too large
Load Diff
@@ -22,7 +22,7 @@ export const load = (async (event) => {
|
|||||||
typeString = 'all';
|
typeString = 'all';
|
||||||
}
|
}
|
||||||
|
|
||||||
const include_collections = event.url.searchParams.get('include_collections') || 'false';
|
const include_collections = event.url.searchParams.get('include_collections') || 'true';
|
||||||
const order_by = event.url.searchParams.get('order_by') || 'updated_at';
|
const order_by = event.url.searchParams.get('order_by') || 'updated_at';
|
||||||
const order_direction = event.url.searchParams.get('order_direction') || 'asc';
|
const order_direction = event.url.searchParams.get('order_direction') || 'asc';
|
||||||
const page = event.url.searchParams.get('page') || '1';
|
const page = event.url.searchParams.get('page') || '1';
|
||||||
|
|||||||
@@ -114,10 +114,13 @@
|
|||||||
} else {
|
} else {
|
||||||
currentSort.visited = false;
|
currentSort.visited = false;
|
||||||
}
|
}
|
||||||
if (url.searchParams.get('include_collections') === 'on') {
|
if (url.searchParams.get('include_collections') === 'true') {
|
||||||
currentSort.includeCollections = true;
|
currentSort.includeCollections = true;
|
||||||
} else {
|
} else if (url.searchParams.get('include_collections') === 'false') {
|
||||||
currentSort.includeCollections = false;
|
currentSort.includeCollections = false;
|
||||||
|
} else {
|
||||||
|
// Default to true when no parameter is present (first visit)
|
||||||
|
currentSort.includeCollections = true;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (!currentSort.visited && !currentSort.planned) {
|
if (!currentSort.visited && !currentSort.planned) {
|
||||||
@@ -469,6 +472,18 @@
|
|||||||
id="include_collections"
|
id="include_collections"
|
||||||
class="checkbox checkbox-primary"
|
class="checkbox checkbox-primary"
|
||||||
checked={currentSort.includeCollections}
|
checked={currentSort.includeCollections}
|
||||||
|
on:change={(e) => {
|
||||||
|
const target = e.currentTarget;
|
||||||
|
currentSort.includeCollections = target.checked;
|
||||||
|
// Immediately update the URL to reflect the change
|
||||||
|
let url = new URL(window.location.href);
|
||||||
|
if (target.checked) {
|
||||||
|
url.searchParams.set('include_collections', 'true');
|
||||||
|
} else {
|
||||||
|
url.searchParams.set('include_collections', 'false');
|
||||||
|
}
|
||||||
|
goto(url.toString(), { invalidateAll: true, replaceState: true });
|
||||||
|
}}
|
||||||
/>
|
/>
|
||||||
<span class="label-text">{$t('adventures.collection_locations')}</span>
|
<span class="label-text">{$t('adventures.collection_locations')}</span>
|
||||||
</label>
|
</label>
|
||||||
|
|||||||
@@ -1226,22 +1226,17 @@
|
|||||||
{/if}
|
{/if}
|
||||||
|
|
||||||
<!-- Help documentation link -->
|
<!-- Help documentation link -->
|
||||||
{#if user.is_staff || !stravaGlobalEnabled}
|
|
||||||
|
{#if !wandererEnabled}
|
||||||
<div class="mt-4 p-4 bg-info/10 rounded-lg">
|
<div class="mt-4 p-4 bg-info/10 rounded-lg">
|
||||||
{#if user.is_staff}
|
<p class="text-sm">
|
||||||
<p class="text-sm">
|
📖 {$t('immich.need_help')}
|
||||||
📖 {$t('immich.need_help')}
|
<a
|
||||||
<a
|
class="link link-primary"
|
||||||
class="link link-primary"
|
href="https://adventurelog.app/docs/configuration/immich_integration.html"
|
||||||
href="https://adventurelog.app/docs/configuration/wanderer_integration.html"
|
target="_blank">{$t('navbar.documentation')}</a
|
||||||
target="_blank">{$t('navbar.documentation')}</a
|
>
|
||||||
>
|
</p>
|
||||||
</p>
|
|
||||||
{:else if !stravaGlobalEnabled}
|
|
||||||
<p class="text-sm">
|
|
||||||
ℹ️ {$t('google_maps.google_maps_integration_desc_no_staff')}
|
|
||||||
</p>
|
|
||||||
{/if}
|
|
||||||
</div>
|
</div>
|
||||||
{/if}
|
{/if}
|
||||||
</div>
|
</div>
|
||||||
|
|||||||
Reference in New Issue
Block a user