3D capture, representation and manipulation of cuneiform tablets Sandra I. Woolley *, Nicholas J. Flowers , Theodoros N. Arvanitis , Alasdair Livingstone , Tom R. Davis and John Ellison The University of Birmingham, Edgbaston, Birmingham, England Harvard University, Cambridge, MA 02138, USA ABSTRACT This paper presents the digital imaging results of a collaborative research project working toward the generation of an online interactive digital image database of signs from ancient cuneiform tablets. An important aim of this project is the application of forensic analysis to the cuneiform symbols to identify scribal hands. Cuneiform tablets are amongst the earliest records of written communication, and could be considered as one of the original information technologies; an accessible, portable and robust medium for communication across distance and time. The earliest examples are up to 5,000 years old, and the writing technique remained in use for some 3,000 years. Unfortunately, only a small fraction of these tablets can be made available for display in museums and much important academic work has yet to be performed on the very large numbers of tablets to which there is necessarily restricted access. Our paper will describe the challenges encountered in the 2D image capture of a sample set of tablets held in the British Museum, explaining the motivation for attempting 3D imaging and the results of initial experiments scanning the smaller, more densely inscribed cuneiform tablets. We will also discuss the tractability of 3D digital capture, representation and manipulation, and investigate the requirements for scaleable data compression and transmission methods. Additional information can be found on the project website: www.cuneiform.net Keywords: 3D imaging, cuneiform tablets, laser scanning, forensic analysis, data archival, data compression. 1. CUNEIFORM TABLETS Cuneiform tablets1 are rectangular slabs of sun-dried or fired clay, moulded by hand and inscribed when wet with a wedgeshaped stylus. Cuneiform (from the Latin word cuneus, meaning wedge) was the most common system of writing in Western Asia from 3000 BC through to the middle of the first millennium BC when it was gradually replaced with various alphabetic systems of writing. The signs are typically 3 mm in height, with about 30 signs per line and 25 lines of writing both on the front and back of the tablet. The Scottish engineer James Nasmyth first came to live in London in 1829 and discovered the British Museum and their rapidly growing collection of cuneiform tablets, on which the work of decipherment was progressing apace. He later wrote in his biography 2, "I was specially impressed and interested with the so-called "Arrow-head" or "Cuneiform Inscriptions" in the Assyrian Department. ... I was particularly impressed with the precision and simple beauty of these cuneiform inscriptions". This interest motivated him to investigate the mechanics of cuneiform production including the reed stylus and the roller seal, now commonly referred to as a cylinder seal, as shown in Figure 1. The dimensions of a cuneiform tablet can range from those of a credit card to those of a palmtop computer, the so-called "monster tablets", and were inscribed by means of a square-profiled reed stylus that made triangular indentations, or wedges, with “tails”. * Correspondence: Email: [email protected]; WWW: http://www.eee.bham.ac.uk/woolleysi; Telephone: +44 121-414-7521; FAX: +44 121-414-4291 Three-Dimensional Image Capture and Applications IV, Brian D. Corner, Joseph H. Nurre, Roy P. Pargas, Editors, Proceedings of SPIE Vol. 4298 (2001) © 2001 SPIE. · 0277-786X/01/$15.00 103 Figure 1: James Nasmyth the 19th century engineer and his early descriptions of cuneiform production 2 An example of a neo-Babylonian tablet and script details are shown in Figure 2. As can be seen, each sign consists of a number of wedges. Along with ancient Egyptian and Chinese writings, these tablets are among the earliest records of written communication. Cuneiform writing itself developed a number of different varieties, and the tablets were also later inscribed with early forms of the alphabet that we still use today. The earliest examples are up to 5,000 years old, and the writing technique remained in use for some 3,000 years. The writing was deciphered in the last 150 years; hundreds of thousands of tablets have been discovered and are now preserved in museum archives. Unfortunately, only a small fraction of these tablets can be made available for display and much important academic work has yet to be performed on the very large numbers of tablets to which there is necessarily restricted access. b) a) c) Figure 2: Examples of cuneiform tablets a) neo-Babylonian b) detail of Neo-Babylonian script from tablet (a) and c) neo-Assyrian detail 104 Proc. SPIE Vol. 4298 2. THE CUNEIFORM DIGITAL FORENSIC PROJECT (CDFP) 2.1. Background The CDFP3 began in January 1999. It is a multidisciplinary project funded by The University of Birmingham, U.K. Our team includes an Assyriologist, a forensic handwriting expert and engineers specialising in database design and digital imaging; all are members of the University of Birmingham, UK. The main objective of the CDFP is the development of a digital cuneiform sign database4 requiring digitization of cuneiform tablets and signs, and the scientific forensic analysis of cuneiform sign configurations. Much of the cuneiform material is fragmentary; excavations performed in the 19th century lacked the rigor of modern scientific archaeology, with the result that much material is unprovenanced. Traditional cuneiform sign lists, both ancient and modern, have always recognised the existence of a variety of script forms pertaining to different areas and periods in the long history of the diverse civilisations that used the cuneiform script. In addition, scholars who have specialised in particular cultural complexes - rural, urban, temple, palace - with large numbers of tablets belonging to one area and period have frequently maintained that they can recognise the handwriting of individual scribes. The digitization of the cuneiform texts is an important first step in the development of a forensic database for scholarly research. It is hoped that our research activities will assist in enabling tablets to be assigned scientifically to the archives or contexts to which they belong, with implications for the documentary history of the areas concerned, thus providing objective support for expert knowledge. The digital imaging aspects of the work involve 2D and 3D imaging, and the delivery of a CD and www-accessible digital library for both research and teaching activities. 3. DATASETS FROM THE 1ST MILLENNIUM BC Two culturally specific areas have been targeted for concentrated work by the CDFP, an area where kings and courtesans rub shoulders with an elite of scholars on one hand, and a rather plebian ancient college of further education on the other. Each of these areas provides an extremely complex set of data. The first of these brings the project to the project to the palaces and royal halls of the mighty Assyrian Empire in the 7th century BC. It is known that a group of twenty or so learned men, the "inner circle", served as advisers to the kings Esarhaddon (680-669 BC) and Ashurbanipal (668-627 BC), while other groups of scribes and scholars at the court are also represented by copious cuneiform documentation in the form of letters and reports. The project has been using digital and forensic analysis to study the structure of the system of advisory personnel of the royal court. This investigation recently led to the discovery that a tablet found at Nineveh may in fact have been written personally by or for Ashurbanipal, probably while he was crown prince. It is well known that Ashurbanipal claimed literacy in a royal inscription and that as king he founded at Nineveh a massive library, some twenty-three thousand tablets and fragments of which are now in the British Museum. In addition, several crudely written tablets seem to be the work of Ashurbanipal himself, since the script shows similarities to the script of tablets signed by the scribe Balasi, who was appointed as Ashurbanipal's teacher. A letter to Ashurbanipal from his young brother (laku) survives. Perhaps most fascinating is a letter to Ashurbanipal's young wife Libbi-ali-sharrat ("O-city-she-is-queen!") from his elder sister Sherua-etirat ("Goddess-Sherua-is-the-one-who-saves!) concerning writing a tablet and reading it out. Sherua-etirat reminds Libbi-ali-sharrat that she is a "daughter of Akkad" and now a member of the ruling family in the "house of succession". By establishing a database of sign forms and individual cuneiform script types (hands) the project is contributing to the investigation of the cult literacy and the indigenous interest in cuneiform surrounding Ashurbanipal as crown prince of Assyria. The second area involves an archive of tablets which has recently been literally pieced together by I.L.Finkel of the Dept of Western Asiatic Antiquities of the British Museum. These tablets date to the reigns of the Persian kings of Babylon Darius and Xerxes, about two hundred years later than the material referred to above. Certain peculiarities of the documents such as spelling mistakes and multiple copies of the same documents in different hands show that what is involved here is a scribal school. The schoolmaster himself can be identified and it seems that he used various kinds of written material at his disposal in his instruction. The level is higher than that of a primary school and could be compared to a college. There are two varieties of instructional material, namely documents of a business nature such as contracts, and medical texts. This Proc. SPIE Vol. 4298 105 was then a Babylonian college of business and medicine. It provides an ideal dataset for the project aims of forensic handwriting analysis and database construction. 4. PHOTOGRAPHY The digital imaging of cuneiform tablets poses a number of challenges. Central to the challenges is the fact that many of the tablets contain densely compacted inscriptions on both sides, with text running over rounded edges and frequently continuing along each side edge, making the complete 2D capture of symbol sets problematic even with many exposures from different angles. Our initial experiments involved SLR photography at the British Museum, using a four-light bed with adjustable camera height and with tablets mounted on a sand tray to stabilise their orientation (Figure 3). Lighting proved problematic, not least because of repeated burns from the directional spotlights. The depth of the impressions also made the selection of correct lighting both difficult and time-consuming. North-west lighting is usually preferred for illumination of cuneiform script and did produce the best results. Soft, neon, north-west lighting was later adopted; however, many images, and occasionally whole reels of film were discarded on subsequent examination. Analogue photographic capture was ruled out as a tractable method for capture of the cuneiform datasets primarily because of the difficulty of preserving scale and context information, the inability to annotate (i.e. properly label) images, the continuous and complex adjustments to accommodate the tablet surfaces, and, importantly, the absence of direct visual feedback. 4.1. Image resolution and format In addition to cuneiform experts, others may also wish to view cuneiform tablets. For example, the optical recognition of characters presents a number of interesting challenges to researchers5, 6. Some viewers may only interested in the general shape and size of the tablets, and a simple, low-resolution model may suffice. Other scholars may be interested in reading the symbols and would therefore require more intelligible, higher resolution renditions. However, the highest level of detail is required by forensic analysts. Discussions with forensic experts indicate that a resolution of approximately 0.025mm is required to distinguish between different scribal hands. Our interest in both sign recognition and forensic analysis dictated high-resolution capture, which proved particularly difficult in the case of the smaller tablets. Conventional NTSC video cameras are convenient to use but only provide 480 lines of resolution, although many video capture devices offer higher resolution. This increased resolution is obtained by interpolation and smoothing, and although these techniques improve the visual quality of the image, the additional resolution is artificial and cannot be used for forensic analysis. The lossless 256-color GIF image format was selected as the most efficient format for 2D images because the highresolution color information of the baseline JPEG standard was not required. GIF was also preferred because of its universal support, in particular, its support in www browsers. New still digital cameras now provide significantly higher resolutions but have practical limitations, in particular, storage capacity and battery life. We used two digital image capture techniques; a simple video camera capture system and a 3.34 megapixel digital camera (The Nikon CoolPix 990), using both zip and CD disc media. The video camera capture resolution proved insufficient for forensic purposes. The digital camera produced images of much more acceptable resolutions, but both techniques required continuous lighting adjustments and frustrating power supply, laptop computer, software and media logistics. Other logistical problems included the regular transport of various pieces of equipment to the museum (much improved by the museum's kind co-operation in storing the light bed and camera), the precious nature of the objects which required 106 Proc. SPIE Vol. 4298 extreme care in handling at all times, the lack of on-site high-bandwidth networking, and, the need to continuously annotate captured details (requiring our expert Assyriologist). When reading the tablets, experts tend to rotate them constantly; grossly in order to present all the signs to inspection, but also subtly, in order to use light and shadow to bring out the indentations clearly. In a digitised tablet, this manipulation would ideally be enabled by high-resolution 3D rendering, rotation control and light source adjustment. Figure 3 : Image capture in the British Museum cuneiform archives 5. 3D SCANNING There are a number of different techniques for capturing 3D models, each offering different scanning areas and resolutions. The Digital Michelangelo project at Stanford University7 involved scanning larger objects, e.g., Michael Angelo's statue of David, at a resolution of up to 0.25mm using laser triangulation rangefinding. The 3D -Matic project8 at the Turing Institute and The University of Glasgow used ‘photogrammetry’ to give a resolution of 0.5mm. Commercial laser stripe triangulation systems9 can deliver resolutions of 0.1mm (potentially 0.05mm). Surface contact probes can achieve similar resolutions, but they make physical contact with the object obviating their application to precious or fragile objects and the scans are very slow; for example, it would take approximately three days to scan one face of a 50mm by 70mm cuneiform tablet. The best technique may be moiré, which theoretically offers very high resolution10 although commercially available portable moiré scanners offer resolutions of closer to 0.2mm. Table 1 summarizes the differences between the various scanning techniques. Technique Photogrammetry Laser Stripe Triangulation Contact Probe Morié (commercial) Morié (theoretical) Scan size (mm) 2000 2000 Resolution (mm) 0.5 0.05 Scan time for one side (s) <1 10 Equipment Cost ($) ? 100,000 300 200 n/a 0.05 0.2 0.001 260,000 (3 days) 1 n/a 2,000 n/a Table 1 : A comparison of 3D scanning methods Proc. SPIE Vol. 4298 107 Figure 4 : Experimental laser stripe triangulation cuneiform scans 5.1. File sizes We have obtained some preliminary scans using laser stripe triangulation at a resolution of 0.1mm; however, at this resolution the resulting 3D mesh model is over 100MBytes. It is not practical to transfer datasets of this size over the Internet, and is very wasteful of bandwidth considering that some scholars only require a low-resolution model. In order to satisfy the needs of various users, scalable delivery is desirable. A typical scenario would be this. The user first obtains a model of the entire tablet at a low resolution. The actual resolution will depend on the capabilities of the user’s viewing platform and on the available communication bandwidth. For example, if the user were viewing the tablet on a hand-held computer with a slow modem, a very low-resolution model would be transferred in a few seconds. If the user was using a workstation with a large high-resolution screen, connected to the Internet with a high-speed network connection, then a much more detailed model could be transferred. If the user chooses to zoom in on a specific area of the model, additional information could then be transferred to increase the resolution of the area of interest. The exact detail of the zoomed-in model would depend on the user’s computer capability. The important difference between this technique and standard delivery mechanisms is that the user would not need to download an entire high-resolution model in order to zoom in on a small area. There are emerging standards that may provide assistance with scalable delivery. The MPEG-4 standard11 allows for scalable rendering of 2D objects, but this is mainly aimed at multicast or broadcast data. It does this by delivering the entire dataset across the communication channel; the users' rendering device can then discard data if a low-resolution model is required. This technique is not suitable for delivery of specifically tuned data to individual users over narrow-bandwidth channels, which is what is required for our needs. The Synthetic/Natural Hybrid Coding group12 is currently investigating mechanisms that achieve interoperability and scalability of mixed media types suitable for different storage media and communication bandwidths. The group is also working on server-based view-dependent terrain or object rendering and view-scalable audio-video in 2D/3D environments. To date, the work on mesh coding has focused on 2D models, but there are plans to extend the work to 3D mesh coding in the future. 5.2. Object tagging It is very important that the 3D datasets are traceable throughout the entire capture, storage, delivery and manipulation process. There are many important pieces of information associated with the electronic representation of each tablet, for example the catalogue number of the cuneiform tablet and information on the capture process. Other information such as size, excavation details, storage location, ownership, and translations etc. may need to be included in the data tag. As the user manipulates the data and zooms in, information on the physical size and location of the subset of the dataset should be available. Since every user may have a different dataset, depending on their needs, this tagging must be embedded in the data so that it survives further manipulation. 108 Proc. SPIE Vol. 4298 It is also important that the data tags are available independently of the data itself and the tags should be accessible in an electronically searchable format. A typical use of this may be a user wanting to find all tablets excavated from a certain site, and they should be able to search the data tags and obtain a list of relevant tablets. The user could then select tablets of interest for further examination. 5.3. Issues of copyright A further requirement may be access control to certain tablets for copyright reasons. Some tablets may only be visible to authorized users, and this authorization should somehow be carried with the dataset representing the tablet. Another use of access control would be to limit the available resolution depending on user privileges. Public access could be given to low resolution data, and access to the high-resolution models limited to forensic experts. To prevent copyright infringement, it is important that this copyright information is closely coupled to the data itself. Techniques such as watermarking could be used, although there are no standards for 3D watermarking data at present. The MPEG-7 standard11 may offer a potential solution to this problem, but the standard is not yet complete and it is unclear if it offers the flexibility for our specific needs. 5.4. The virtual reality modeling language (VRML) VRML13 is a modeling language that can provide an environment for manipulation of 3D datasets. Standard VRML viewers work by downloading a complete dataset and rendering this on screen depending on the user’s chosen viewpoint. When used as a mechanism to view 3D models such as cuneiform tablets, the user can select which part of the tablet to look at (two-dimensional x and y co-ordinates) and can zoom in and out (z co-ordinate). Also, the user needs to be able to rotate the tablet in all three axes to be able to see all of the faces of the tablet. This gives six degrees of freedom; but most computers have only two-dimensional input devices (mouse, trackball or joystick). To provide the requisite six degrees of freedom, VRML viewers have different modes such as ‘pan’, ‘turn’ and ‘roll’ in order to translate movement of the input device into the required movement of the image. For example, the Cortona VRML viewer 14 has three major modes (‘walk’, ‘fly’ and ‘study’) and four sub-modes. Figure 5 shows the Cortona viewer displaying a sample 3D laser scanned cuneiform fragment. Standard VRML interfaces can be confusing, particularly to novice users, if, for example, left-right movement of the input device produces rotation of the tablet, when the user was expecting a lateral movement of the viewing window. An additional complexity is added when illumination is taken into account; for correct viewing of the cuneiform tablets the user must be able to alter the position of one or two light sources to create the shadows to provide the necessary depth information. This gives two or four extra parameters that need to be manipulated. An ideal solution would be a form of data glove that measures the position of the user’s hand (in all six degrees of freedom) and renders the image to correspond with the position of the hand. The other hand could be used to manipulate the position of the virtual light source using another data glove. If more than one light source is required, a more flexible human input device would be required. Figure 5 : Cortona VRML viewer Proc. SPIE Vol. 4298 109 5.5. Data compression of 3D sources In order to store and deliver the large datasets we require, efficient and scaleable compression techniques are essential. Scalable delivery of 2D data is incorporated in several image formats (e.g. JPEG 2000) and some of these use wavelet algorithms to generate good compression ratios. G. Taubin15 has developed techniques for 3D geometry compression and progressive transmission of polygonal models highly relevant to these objectives. 6. CONCLUSIONS AND FURTHER WORK Our experiments have indicated that sufficiently high-resolution scans are not easily realisable, and that in any case the resulting file sizes would prohibit remote access and real-time manipulation. Experimental scans were performed using laser stripe triangulation at resolutions of 10 lines per millimeter producing files in excess of 100MBytes. The current cost and complexity of the appropriate scanning processes makes remote capture, formatting, storage and communication even more challenging. The next phase of work will include the application of scaleable 3D compression methods and further investigation into robust and detailed data tagging systems. ACKNOWLEDGEMENTS Our sincere thanks to Ron Spencer of Mechanical and Manufacturing Engineering, The University of Birmingham, Professor Paul Ward of Kodak, UK and also to 3DScanners of Coventry. Also very special thanks to Graham Norrie for his very valuable photographic advise and to our many project students, in particular, in the School of Electronic and Electrical Engineering ; M. H. Yu (Ben), Matt Atkins, Peter Cooper. Last, but not least, our sincere thanks to Dr J. Curtis, Keeper of Western Asiatic Antiquities and his staff at the British Museum for their support and co-operation. REFERENCES 1. 2. 3. 4. 5. 6. 7. 8. 9. 10. 11. 12. 13. 14. 15. 110 C. B. F. Walker, Cuneiform (Reading the Past, Vol 3), Univ California Press; 1989, ISBN: 0520061152 J. Nasmyth, James Nasmyth Engineer - An Autobiography, John Murray Pub., London, 1885 T.N. Arvanitis, T.R. Davis, N. Hayward, A. Livingstone, J. Pinilla-Dutoit, S.I. Woolley, The Work of the Cuneiform Digital Forensic Project, VAST: Virtual Archaeology between Scientific Research and Territorial Marketing, Arezzo (Italy), 24-25 November 2000 A. Livingstone, S.I. Woolley, T.R. Davis, T.N. Arvanitis, XML and Digital Imaging Considerations for an Interactive Cuneiform Sign Database, Electronic Publication of Ancient Near Eastern Texts (Conference), The Oriental Institute of the University of Chicago, October 8-9 1999 N. Demoli, U. Dahms, H. Gruber, G. Wernicke, Optical pattern recognition in the analysis of ancient Babylonian cuneiform inscriptions. Holography and Correlation Optics II, SPIE-Proceedings 2647, pp. 138-144 N. Demoli, U. Dahms, H. Gruber, G. Wernicke, Cuneiform recognition experiments: Coherent optical methods and results. In: Optical techniques in the humanities, D. Dirksen, G. V. Bally (eds) Springer 1997, pp. 171-174 The Digital Michelangelo Project, Stanford University, USA http://graphics.stanford.edu/projects/mich/ 3D-Matic Research Laboratory, The University of Glasgow, United Kingdom http://www.faraday.gla.ac.uk 3D Scanners Ltd, http://www.3dscanners.co.uk X. Xie, J.T. Atkinson, M.J. Lalor, D.R. Burton, Three-Map Absolute Moiré Contouring, Applied Optics, Vol. 35, No. 35, 10 December 1996 The Moving Picture Expert Group, http://www.cselt.it/mpeg/ The Synthetic Natural Hybrid Coding Group, http://www.sait.samsung.co.kr/cgi-bin/snhc/src/index.html The Web3D Consortium, http://www.vrml.org/ The Cortona VRML client, Parallelgraphics, http://www.parallelgraphics.com G. Taubin, Visual and Geometric Computing Group, IBM T.J.Watson Research Center, http://www.research.ibm.com/people/t/taubin/index-ie4.html Proc. SPIE Vol. 4298
© Copyright 2026 Paperzz