mirror of
https://github.com/umap-project/umap.git
synced 2025-04-29 11:52:38 +02:00
Compare commits
3 commits
Author | SHA1 | Date | |
---|---|---|---|
![]() |
78ee070969 | ||
![]() |
4c9bbb68aa | ||
![]() |
cdc900a9e8 |
6 changed files with 58 additions and 11 deletions
|
@ -1,5 +1,10 @@
|
||||||
# Changelog
|
# Changelog
|
||||||
|
|
||||||
|
## 2.7.2 - 2024-11-15
|
||||||
|
|
||||||
|
### Bug fixes
|
||||||
|
* use our fork of csv2geojson to be able to parse lat/lon with commas
|
||||||
|
|
||||||
## 2.7.1 - 2024-10-25
|
## 2.7.1 - 2024-10-25
|
||||||
|
|
||||||
### Bug fixes
|
### Bug fixes
|
||||||
|
|
|
@ -40,7 +40,7 @@
|
||||||
"@placemarkio/tokml": "0.3.4",
|
"@placemarkio/tokml": "0.3.4",
|
||||||
"@tmcw/togeojson": "^5.8.0",
|
"@tmcw/togeojson": "^5.8.0",
|
||||||
"colorbrewer": "1.5.7",
|
"colorbrewer": "1.5.7",
|
||||||
"csv2geojson": "5.1.2",
|
"csv2geojson": "github:umap-project/csv2geojson#patched",
|
||||||
"dompurify": "3.1.7",
|
"dompurify": "3.1.7",
|
||||||
"georsstogeojson": "^0.2.0",
|
"georsstogeojson": "^0.2.0",
|
||||||
"jsdom": "^24.0.0",
|
"jsdom": "^24.0.0",
|
||||||
|
|
|
@ -1 +1 @@
|
||||||
VERSION = "2.7.1"
|
VERSION = "2.7.2"
|
||||||
|
|
|
@ -81,6 +81,8 @@ export class Formatter {
|
||||||
{
|
{
|
||||||
delimiter: 'auto',
|
delimiter: 'auto',
|
||||||
includeLatLon: false,
|
includeLatLon: false,
|
||||||
|
sexagesimal: false,
|
||||||
|
parseLatLon: (raw) => Number.parseFloat(raw.toString().replace(',', '.')),
|
||||||
},
|
},
|
||||||
(err, result) => {
|
(err, result) => {
|
||||||
// csv2geojson fallback to null geometries when it cannot determine
|
// csv2geojson fallback to null geometries when it cannot determine
|
||||||
|
@ -115,7 +117,9 @@ export class Formatter {
|
||||||
}
|
}
|
||||||
|
|
||||||
async fromGeoRSS(str) {
|
async fromGeoRSS(str) {
|
||||||
const GeoRSSToGeoJSON = await import('../../vendors/georsstogeojson/GeoRSSToGeoJSON.js')
|
const GeoRSSToGeoJSON = await import(
|
||||||
|
'../../vendors/georsstogeojson/GeoRSSToGeoJSON.js'
|
||||||
|
)
|
||||||
return GeoRSSToGeoJSON.parse(this.toDom(str))
|
return GeoRSSToGeoJSON.parse(this.toDom(str))
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -81,6 +81,8 @@ function csv2geojson(x, options, callback) {
|
||||||
}
|
}
|
||||||
|
|
||||||
options.delimiter = options.delimiter || ',';
|
options.delimiter = options.delimiter || ',';
|
||||||
|
options.parseLatLon = options.parseLatLon || parseFloat;
|
||||||
|
options.sexagesimal = options.sexagesimal !== false;
|
||||||
|
|
||||||
var latfield = options.latfield || '',
|
var latfield = options.latfield || '',
|
||||||
lonfield = options.lonfield || '',
|
lonfield = options.lonfield || '',
|
||||||
|
@ -129,6 +131,7 @@ function csv2geojson(x, options, callback) {
|
||||||
|
|
||||||
if (!latfield) latfield = guessLatHeader(parsed[0]);
|
if (!latfield) latfield = guessLatHeader(parsed[0]);
|
||||||
if (!lonfield) lonfield = guessLonHeader(parsed[0]);
|
if (!lonfield) lonfield = guessLonHeader(parsed[0]);
|
||||||
|
|
||||||
var noGeometry = (!latfield || !lonfield);
|
var noGeometry = (!latfield || !lonfield);
|
||||||
|
|
||||||
if (noGeometry) {
|
if (noGeometry) {
|
||||||
|
@ -152,13 +155,15 @@ function csv2geojson(x, options, callback) {
|
||||||
lonf, latf,
|
lonf, latf,
|
||||||
a;
|
a;
|
||||||
|
|
||||||
a = sexagesimal(lonk, 'EW');
|
if (options.sexagesimal) {
|
||||||
if (a) lonk = a;
|
a = sexagesimal(lonk, 'EW');
|
||||||
a = sexagesimal(latk, 'NS');
|
if (a) lonk = a;
|
||||||
if (a) latk = a;
|
a = sexagesimal(latk, 'NS');
|
||||||
|
if (a) latk = a;
|
||||||
|
}
|
||||||
|
|
||||||
lonf = parseFloat(lonk);
|
lonf = options.parseLatLon(lonk);
|
||||||
latf = parseFloat(latk);
|
latf = options.parseLatLon(latk);
|
||||||
|
|
||||||
if (isNaN(lonf) ||
|
if (isNaN(lonf) ||
|
||||||
isNaN(latf)) {
|
isNaN(latf)) {
|
||||||
|
@ -179,8 +184,8 @@ function csv2geojson(x, options, callback) {
|
||||||
geometry: {
|
geometry: {
|
||||||
type: 'Point',
|
type: 'Point',
|
||||||
coordinates: [
|
coordinates: [
|
||||||
parseFloat(lonf),
|
lonf,
|
||||||
parseFloat(latf)
|
latf
|
||||||
]
|
]
|
||||||
}
|
}
|
||||||
});
|
});
|
||||||
|
|
|
@ -494,6 +494,39 @@ def test_import_csv_without_valid_latlon_headers(tilelayer, live_server, page):
|
||||||
expect(page.locator('umap-alert div[data-level="error"]')).to_be_visible()
|
expect(page.locator('umap-alert div[data-level="error"]')).to_be_visible()
|
||||||
|
|
||||||
|
|
||||||
|
def test_import_csv_with_commas_in_latlon(tilelayer, live_server, page, settings):
|
||||||
|
settings.UMAP_ALLOW_ANONYMOUS = True
|
||||||
|
page.goto(f"{live_server.url}/map/new/")
|
||||||
|
page.get_by_title("Open browser").click()
|
||||||
|
layers = page.locator(".umap-browser .datalayer")
|
||||||
|
markers = page.locator(".leaflet-marker-icon")
|
||||||
|
page.get_by_title("Import data").click()
|
||||||
|
textarea = page.locator(".umap-upload textarea")
|
||||||
|
textarea.fill("lat;lon;foobar\n12,24;48,34;mypoint\n12,23;48,35;mypoint2")
|
||||||
|
page.locator('select[name="format"]').select_option("csv")
|
||||||
|
page.get_by_role("button", name="Import data", exact=True).click()
|
||||||
|
expect(layers).to_have_count(1)
|
||||||
|
expect(markers).to_have_count(2)
|
||||||
|
with page.expect_response(re.compile(r".*/datalayer/create/.*")):
|
||||||
|
page.get_by_role("button", name="Save").click()
|
||||||
|
datalayer = DataLayer.objects.last()
|
||||||
|
saved_data = json.loads(Path(datalayer.geojson.path).read_text())
|
||||||
|
assert saved_data["features"][0]["geometry"] == {
|
||||||
|
"coordinates": [
|
||||||
|
48.35,
|
||||||
|
12.23,
|
||||||
|
],
|
||||||
|
"type": "Point",
|
||||||
|
}
|
||||||
|
assert saved_data["features"][1]["geometry"] == {
|
||||||
|
"coordinates": [
|
||||||
|
48.34,
|
||||||
|
12.24,
|
||||||
|
],
|
||||||
|
"type": "Point",
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
def test_create_remote_data(page, live_server, tilelayer):
|
def test_create_remote_data(page, live_server, tilelayer):
|
||||||
def handle(route):
|
def handle(route):
|
||||||
route.fulfill(
|
route.fulfill(
|
||||||
|
|
Loading…
Reference in a new issue