Summary: What if AI creates the Changeset Comments? We could send locations, tag types, and quantities to get an output. AI would have to be run locally with small models for cost and be validated by the user.
Problem 1: Time I assume that 1,000 users create 2 changes in 1 day. We assume that each change set takes 3.5 seconds. 1000 users *2 changes * 3.5 seconds per change = 7000 seconds. OSM Users spend about 1.9 hours per day.
Problem 2: Skill Outsourcing Users should spend time on the things AI can’t do.
Problem 3: Server Side Peer Review We have human generated changeset comments. We could create AI generated changeset comments. We could ask the AI, “are these 2 changeset comments so different that it looks malicious”?
General AI Inputs: 1. Location: Where did the user map? 2. Feature Types: What tags did the user use?
AI Prompt: “You are an AI system. A user made edits in OpenStreetMap, a collaborative mapping project. They mapped locations[Mappleville, MN, USA; Bobville, MN, USA] with tags[50xSidewalks, 20xMarkedCrossings, & 10xReligous Areas]. You will create a changeset comment that concisely tells human reviewers what this changeset was about in 3 sentences or less. Exact numbers are not important. Changesets describe changes, so don’t request anything. Don’t mention anything that is common across all changesets.”
AI Response (https://www.meta.ai/): “Added sidewalks, marked crossings, and religious areas in Mappleville and Bobville, MN. Improved pedestrian and accessibility mapping. Enhanced local community information.”
Specific AI Inputs for Locations: 1. Cities[1 to 5], States[1 to 5], Countries[1 to 5]. 2. Is this a place with unclear boundaries? (What if somebody maps the ocean) 3. What is the size of the bounding box for this edit in KM?
Specific AI Inputs for Feature Types: Tags[1 to 6] & corresponding Quantities
… 查看完整日记文章