1. Introduction
For decades, researchers have been expected to present their findings visually. Visualization transforms data, information, or ideas into images, diagrams, graphics, or other visual formats to enhance comprehension and interpretation [
1,
2]. Its primary goal is to represent complex or abstract concepts in a clear and illustrative manner, facilitating better understanding and analysis.
Scientific research findings are typically presented as numbers, tables, or text, which makes interpretation challenging even for experts. Therefore, it is crucial to depict results and theories visually. Scientific research should be accessible to as many people as possible; however, researchers are generally neither artists nor graphic designers. Consequently, their illustrations tend to be graphically simple, often consisting of basic line (vector) drawings. When presenting complex ecosystems or landscapes, overly simplistic illustrations may undermine the perceived rigor of the research. Aesthetic illustrations are generally considered more effective [
3], especially when shared on social media.
Landscape visualization employs a variety of software solutions tailored to different purposes [
4], enabling either realistic or stylized representations of the landscape [
5].
Three main types of visualizations are commonly used: (a) rendered still images, which display the landscape with high detail and realism but offer limited interactivity; (b) animations, which show changes and movements over time but lack full interactivity; and (c) real-time models, which allow users to move freely in virtual space, enhancing interactivity, though often at the cost of reduced detail due to computational limitations [
4].
Procedural landscape generation is a computer graphics technique that automatically creates landscapes using mathematical algorithms and rule-based systems [
14]. This method is particularly useful for generating large-scale or highly detailed virtual environments in video games, films, simulations, and cartography. Instead of manually designing landscapes, procedural generation creates them based on parameters such as elevation, terrain type, climate, and textures, often incorporating random elements to ensure uniqueness. Procedural techniques enable the efficient and scalable production of large terrains with high detail, eliminating the need for manual modeling. Some systems can also simulate real-world geological and meteorological processes, such as erosion, river formation, and vegetation distribution. Common applications include optimized terrain and height map generation, offering advanced erosion simulation and texture generation, or specializing in photorealistic landscape rendering commonly employed in film production and high-end visual effects. Overall, procedural landscape generation is a powerful tool that enables the rapid and automated creation of realistic or stylized terrains while giving users control over the final output through adjustable parameters.
Several of the visualization applications integrate AI-powered solutions. Luminar Neo utilizes AI to enhance photos by automatically adjusting lighting, colors, and details in landscapes. Its AI-driven tools include sky replacement, intelligent contrast adjustment, and object removal. Gaea [
9] is designed for procedural landscape generation, where AI-based algorithms assist in creating realistic terrain and textures. Adobe Photoshop’s Neural Filters feature enables automatic retouching, color correction, and style transformations through AI. Adobe Lightroom [
13] incorporates advanced AI-driven editing tools, such as sky correction and intelligent exposure adjustments. Google Earth Studio [
12], primarily used for cartographic animations. Procedural landscape generation software like World Machine and Terragen [
7,
8] can utilize machine learning-based algorithms, but are not exclusively AI-powered programs.
However, these applications are not widely used among scientists for landscape visualization, because they require in-depth software knowledge. The complexity of these tools exceeds the skill set of most researchers, and the workflow demands more research time (level of effort—LoE) than is typically allocated for visualization. One alternative is to commission studios or artists to create striking images or short films based on the researcher’s guidance, with some studios specializing in archaeological landscape reconstructions. Yet, the demand for illustrations and graphic designers surpasses what traditional means can supply. AI offers a cost-effective and rapid technology for visualization. Generating landscape images with the latest AI applications is not inherently difficult; even users with minimal experience can create visually impressive yet artificial ‘nature photographs’ within minutes. Social media is already flooded with AI-generated pseudo-landscapes, and some nature photography websites occasionally feature such images, which is considered an unethical practice, as they are not genuine nature photographs.
Researchers, however, aim to produce photorealistic and detailed scientific illustrations. Here, detail refers to the representation of intricate visual elements, while realism denotes how closely the visualization mirrors the actual landscape [
4]. Cost-effectiveness in this context means achieving the outlined workflow at a tiny fraction (about one hundredth) of the traditional cost of hiring a professional illustrator or animator.
But what makes a landscape reconstruction scientifically sound? Primarily, it depends on a solid scientific foundation, rather than merely knowledge of AI applications. Scientific accuracy is paramount, as the objective is not to depict an imaginary, ‘fairy tale’ landscape. Instead, it is necessary to study geographical, ecological, paleoenvironmental, and landscape history literature to precisely determine the intended outcome. Several scientific disciplines are involved in the historical reconstruction of landscapes and natural environments.
Historical geography investigates the spatial characteristics of social, economic, and natural conditions across different time periods [
15]. Its sources include old maps, charters, chronicles, and demographic records. This field enables the mapping of medieval settlement structures, changes in river courses, and the locations of economic centers over centuries. Closely related to this is landscape history, primarily studied by ecologists, which examines the formation and transformation of landscapes, uncovering the impacts of natural processes and human activities on vegetation development over various historical periods [
16,
17]. Paleoecology focuses on reconstructing past environments, studying changes in climate, vegetation, and fauna across geological epochs, relying mainly on fossils and sedimentary records, and is pursued by geologists and biologists [
18]. Archaeology increasingly emphasizes past landscape reconstruction: environmental archaeology examines the natural surroundings of archaeological sites and past human-environment interactions, analyzing the roles of climate, soil conditions, vegetation, and wildlife in shaping human lifestyles. Landscape archaeology investigates historical landscape changes, including natural and anthropogenic elements [
19], exploring how societies shaped their landscapes through land use and settlements. Ethnographic research also provides valuable insights, particularly regarding traditional ecological knowledge [
20], such as the expertise of shepherds in the Hortobágy ‘puszta’ region of Hungary about the historical conditions of the steppe [
21]. Furthermore, botanical and geomorphological knowledge is essential to distinguish which vegetation and landforms are ancient versus modern and natural versus man-made [
22].
The main goals of this publication are: (1) to demonstrate how different AI image generation techniques can be effectively used for landscape reconstruction; (2) to highlight common pitfalls of AI image generation and present simple solutions; and (3) to outline necessary directions for further developments. The approaches are illustrated using the Great Hungarian Plain, a landscape fundamentally altered in the 19th century.
3. Results and Evaluation
3.1. Writing an Effective Prompt
AI applications generate images based on a textual instruction known as prompts. In the case of landscapes, these prompts include details about topography, vegetation characteristics, and perspective. For example: ‘
Create an image depicting a wide, meandering river from a bird’s-eye view. The landscape is experiencing a flood, with the river overflowing its banks, submerging the surrounding fields and trees. At the center of the image, the river meanders, while the floodplain is covered with water. The water reflects the grayish sky, which is overcast with clouds. Trees stand partially submerged, and the grayish-brown hues of the landscape convey a misty, cool atmosphere’ (
Figure 2). Writing such detailed prompts requires practice, but AI can assist in refining them. By uploading an existing image, one can request a prompt in ChatGPT to generate a similar image in another application. For example: ‘
I uploaded an image for you, please create an accurate description of what is shown in the image. Then give me a prompt to generate the same image in Adobe Firefly.’
While this example uses descriptive natural language, many image generation tools also support compact, comma-separated prompts, often referred to as tag-based or fragmented prompting. These rely on keywords and phrases separated by commas to convey style, content, composition, and atmosphere concisely. Such prompt formats are popular in tools like Midjourney, Leonardo.AI, and ComfyUI due to their speed and modularity. The same visual scene described above can be transformed into a tagged prompt like this: “aerial view, wide meandering river, flooding landscape, submerged fields and trees, gray overcast sky, water reflecting sky, misty atmosphere, cool tones, gray-brown colors, partially submerged trees, dramatic natural disaster scene” This approach allows users to iteratively expand, modify, or refine individual visual elements while maintaining overall coherence.
Effective AI-generated landscape reconstructions require precise wording. Descriptions should specify topography, vegetation, and water bodies using clear terms such as ‘a gently rolling landscape with scattered oak forests and a meandering river’ instead of vague phrases. Employing scientific terminology, such as ‘alluvial plain’ or ‘karst formations,’ enhances accuracy, but may limit the number of suitable sample images. Specifying perspective—whether an ‘aerial view of a plain’ or a ‘low-angle perspective from within a dense forest’—prevents ambiguity. Concise prompts are more effective than overloaded descriptions. Contradictory phrases like ‘a barren desert with lush meadows’ should be avoided in favor of emphasizing a dominant landscape type or gradual transitions, such as ‘a dry steppe with patches of seasonal wetlands.’ Small objects should be described in a grouped context, for example, ‘a herd of sheep grazing on the misty hillside’, to improve AI interpretation. AI models analyze the contextual coherence and cannot process contradictory prompts, such as depicting a small object from a bird’s-eye view or including snow in a summer landscape.
Depth and realism are enhanced when foreground and background elements are harmoniously balanced, and atmospheric effects—such as ‘
a twilight sky with scattered cumulus clouds’—are incorporated. Since AI models are predominantly trained on English data, rephrasing rather than direct translation improves clarity. Highlighting distinctive features, for instance, ‘
an ancient oak forest with moss-covered trunks’ rather than the more general ‘
a temperate forest’ helps generate more specific and accurate outputs. By carefully structuring prompts with these considerations, AI-generated landscape reconstructions can achieve a higher level of realism and scientific accuracy while preserving the unique character of the studied environment (cf.
Table 1).
AI applications like Adobe Firefly often provide multiple image versions, allowing selection of the best one from a professional perspective and further improvement in subsequent workflow steps. While AI can theoretically generate almost anything, the primary limitation often lies in the precision of the prompt. Writing prompts in English are advisable, as this enables AI to access a broader range of reference images. However, this also exposes a fundamental issue with AI-generated images: users have limited control over the source images the AI draws upon for its visual output, raising copyright and ethical concerns. From a scientific standpoint, a major challenge is that AI-generated images tend to be generic and uninspired, often lacking the distinctive characteristics of the specific landscape under study—features that should make the research subject immediately recognizable. Moreover, errors in image generation are common, as AI does not fully comprehend natural laws, particularly those governing geomorphology and vegetation patterns. For instance, it may inaccurately depict river meanders or vegetation distribution in response to geomorphology. Training AI to understand these processes requires more than a well-crafted prompt.
3.2. Using Style References
The best results can be achieved by combining different techniques and AI applications. Photographs and drone footage of real landscapes can be uploaded as stylistic references for AI tools, enabling researchers to guide the AI by specifying desired topography, hydrography, and vegetation features. Based on these references, AI generates images that better reflect both the unique characteristics of the landscape and the researcher’s creative intent. However, caution is necessary when selecting reference images, as unwanted elements in the originals may inadvertently appear in the generated output. For example, if a paved road is visible in the reference image, the AI may incorporate a similar road into the final image, even if it was not intended to be part of the reconstructed landscape (
Figure 2). Therefore, selecting sample photo locations should be preceded by careful field research. It is advisable to use photos from pristine, protected areas that best represent the potential vegetation of the region. These areas are typically core zones of national parks, so consultation with nature conservation authorities is recommended.
One particularly useful tool in AI applications is generative fill, which allows users to modify images by selecting an area and instructing the program to replace it with new content. Currently, Adobe Firefly is the most effective program for this purpose. For example, arable fields can be transformed into grasslands, or a silted-up riverbed can be ‘restored’ to an active watercourse (
Figure 3). Additionally, a small, intact landscape spot can be expanded, effectively generating an entire landscape from a single real-world reference (
Figure 4). Generative fill can also be used to replace unwanted elements or to add decorative details to enhance realism. For instance, a farmhouse can be inserted into a steppe landscape, or a flock of sheep can be placed on a hillside. Adding such decorative elements is always recommended, as they help convey the spatial depth, making the image more visually engaging and scientifically informative (
Figure 5).
3.3. Using Compositional References
The highest level of landscape reconstruction is currently achieved by using historical maps or three-dimensional terrain models as compositional (structural) references. This approach allows for a high degree of spatial accuracy in depicting past landscapes. In Hungary, for example, the maps from the First Military Survey are particularly valuable for reconstructing the hydrography and vegetation patterns as they existed 200+ years ago (between 1782 and 1785) [
35]. By integrating such historical cartographic sources with AI-generated imagery, researchers can create reconstructions that are not only visually compelling but also scientifically well-founded. The accuracy of terrain features, such as river courses, wetlands, and forested areas, can be significantly improved by aligning AI-generated images with these historical spatial references.
The Arcanum Maps/MAPIRE source mentioned above is considered optimal in terms of area coverage by the published databases (most of Europe in the 18th and 19th centuries and parts of the United States mapped before 1926) and the large scale of the maps [
44]. However, georeferenced historical maps are also available on other portals. The largest of these is the David Rumsay Map Collection [
45], where the maps presented can be downloaded in georeferenced form, and multi-section maps (e.g., the early Cassini map of France) are available in unified, georeferenced mosaics. A similar technical approach is followed by the USGS Historical Topographic Maps project [
46]. An important resource is the OldMapsOnline server [
47], developed in the Czech Republic, which processes and presents maps from numerous archival collaborators. Some national archives and universities also publish maps of their main areas of interest independently, notably the National Library of Scotland—Historical Maps [
48], which publishes maps of the British Isles.
It is advisable to modify maps before AI image generation. Maps should be rotated for better perspective (
Figure 6), and in mountainous regions, they can be adjusted to fit a Digital Elevation Model (DEM). Any prominent labels, grids, or markings should be removed beforehand, as the AI may mistakenly interpret them as landscape features. Instead of using complex maps, creating a simple sketch with strong contrast to serve as a compositional guide would be beneficial. The use of DEM is not presented in this publication due to the flat topography of the landscape, but it would yield similar results to those achieved with maps.
3.4. Upscaling
The previous steps leave visible traces on the initial images, and the quality of the generative fill often differs noticeably from the original details. Another potential issue is that the resolution of the generated images may be insufficient for high-quality publications. To address this, upscaling the images is highly recommended. In this publication, all landscapes have been upscaled using the Krea application. This process not only enhances resolution but also ensures consistency in image quality and detail across different areas.
If only minor adjustments are needed, the AI strength can be set to a minimum. A common mistake is setting AI processing too high, which may distort certain objects (e.g., animals) or introduce elements that do not naturally fit the landscape. After this step, it may be necessary to sharpen the images, enhance contrast, or reduce noise. Adobe Lightroom is the easiest tool for finalizing these adjustments.
The final image resolution typically ranges between 10 and 20 megapixels, which is sufficient for most publications, making higher resolutions unnecessary. The images are saved in .png or .jpg format. In some cases, it was necessary to step back into the generative fill workflow to remove or add objects. Importantly, the entire workflow takes no more than 1–2 h, provided the selected input samples are suitable.
4. Discussion and Conclusions
Based on our experience, the most striking results can be achieved by restoring vanished landscape elements, such as the unregulated rivers and wetlands characteristic of the GHP. Representing specific, unique landscape features—such as traditional land-use patterns or folk architectural objects—poses certain challenges. Nevertheless, with appropriate methodologies and innovative approaches, these visualization issues can be progressively resolved.
An interesting question is whether the method described above can be verified to produce at least somewhat realistic images (cf.
Figure 7). It is clearly impossible to match the landscape of hundreds of years ago with drone photos that did not exist at that time. However, nothing prevents us from applying our method to today’s maps and comparing the AI-generated images with current images, although some minor aspects (see below) prevent full similarity.
The structural similarity between parts (a) and (c) of
Figure 8 is clear; the differences lie in the color scheme and generalization (neglect of minor details). We can therefore conclude that the “historical drone photographs” produced by our method show a similar generalization of the landscape. However, some factors cause differences in general appearance despite structural similarity: the “modern” map used is approximately 40 years old, surveyed in mid-1970s. Minor to medium-scale differences in land use exist, especially regarding the forest cover and the water-affected areas;
Active wetland restoration is ongoing in this area. The open water areas visible in the summer satellite image relate to this;
Even the visualization provided by Google Earth requires a certain level of expertise. In contrast, the generated AI image, although somewhat stylized and generalized, clearly illustrates the patterns of land use in a way that is understandable to everyone.
Figure 8.
Verification of the method with modern databases: (
a) AI-generated image of Farmos Nagy-Nádas (Central Hungary; zone (4) in
Figure 1), based on modern topographic map (
b); compared to the Google Earth image with satellite image from June 2021 (
c). The land cover types show some medium-level generalization, yet the similarity is evident.
Figure 8.
Verification of the method with modern databases: (
a) AI-generated image of Farmos Nagy-Nádas (Central Hungary; zone (4) in
Figure 1), based on modern topographic map (
b); compared to the Google Earth image with satellite image from June 2021 (
c). The land cover types show some medium-level generalization, yet the similarity is evident.
The biggest challenge in AI-generated images is accurately representing spatial relationships. Landscape elements are often misproportioned, with animals, people, and buildings appearing too large relative to their surroundings. Animals, in particular, tend to be distorted. Better results can be achieved by omitting animals from the initial image generation and adding them later using generative fill. Distortions commonly affect legs and horns, which often require correction through additional editing. Interestingly, some animals (e.g., sheep, wolves, pigs) are rendered more successfully than others. For optimal results, animals should be generated in groups—such as herds or packs—from a bird’s-eye view, where minor imperfections are less noticeable. However, AI’s tendency to repeat patterns is evident in these images.
One solution to address local specificity is to add similar elements to the original image using generative fill, e.g., increasing the number of animals (
Figure 4). Another technique is to use the name of a somewhat similar common object in the prompt. For example, instead of ‘
herd of Hungarian Grey cattle’, we enter ‘
group of white bulls’ works better, since from a bird’s eye view the difference is not noticeable (
Figure 5). Similarly, replacing the word ‘
swallow’, the ‘
long thatched farmhouse’ yields more successful results.
A future advancement in AI image generation could be the ability to modify specific parts of an image in real time using prompts, such as adjusting the size of certain objects. Currently, however, options for this functionality are limited. Another useful feature would be the ability to easily change the perspective of an image (e.g., height, direction).
For each new visualization method, it is important to evaluate its suitability for scientific applications. The content of visualizations significantly influences their effectiveness, with authenticity, relevance, and legitimacy being key factors [
4]. In
The Truthful Art, Alberto Cairo outlines five key quality criteria for effective data visualization [
49]:
Truthful—the visualization accurately and honestly represents the data, avoiding distortions or manipulations;
Functional—the design effectively communicates the intended message, making the data easy to understand for the audience;
Beautiful—aesthetics enhance the visualization without compromising clarity or accuracy;
Insightful—the visualization not only displays data but also reveals new patterns, trends, or relationships;
Enlightening—it educates the audience by providing context and helping them better understand the topic.
The first criterion—the truthfulness of generated images—can be questioned, as the outcome cannot be predicted with mathematical precision; the described method serves merely as a guideline. Unlike plotting a graph, the final result of image generation is not deterministic. Instead, a creative process is required, involving discussion and ultimately consensus among the research group members. Enhancing credibility requires visualizations to be based on reliable data sources and for creators to maintain transparency about their methods. The authors emphasize that complete objectivity is unattainable, as every visualization embodies a particular perspective. Consequently, involving research participants in the creation process is crucial to ensuring that the content accurately represents diverse viewpoints.
Using style and composition sample photos significantly increases the realistic appearance of the images, but the final result depends on the accuracy of the selected samples. It is also important to emphasize that ethical and copyright issues can be avoided by using one’s own photos.
Mastering AI applications will soon be essential for researchers, enabling them to generate images quickly and affordably. With well-chosen sample images, high-quality visuals can be created within an hour. As user interfaces become more interactive, AI tools will become a routine part of researchers’ skill sets. We believe this method will be most often used for creating graphic abstracts. The technology is advancing rapidly, and the era of AI-generated landscape videos is approaching, allowing experts to visualize landscape processes and phenomena dynamically.
However, striking landscape visualizations are not limited to scientific publications. This approach can also illustrate expected outcomes of habitat conservation reconstructions or depict the future impacts of global climate change. Such visualizations offer remarkable opportunities for communication campaigns as well.