.

Wednesday, July 31, 2019

The Dilemma of IELTS Test Takers at Preparatory Programs in the UAE: A Call for Lexical-Structural Syllabus Design

————————————————- Doctorate of Education Theoretical Perspectives on Materials and Syllabus Design in TESOL  (DED609) Title: The Dilemma of IELTS Test Takers at Preparatory Programmes in the UAE: A Call for Lexical-Structural Syllabus Design. By: Mohammed A. Molhim Module tutor: Dr. Yasmin yildiz No. of words 000 words Essay Outline * Introduction * Conceptual Framework * IELTS Marking Rubrics: Lexical Resources & Grammatical Range * Lexical Syllabus * Structural Syllabus Selection of Lexical and Grammatical Items in Lexical-structural Syllabus: * Structural Syllabus: Drawbacks * The Structural Syllabus and Grammar Teaching * Suggested Steps to design a Lexical-Structural Syllabus * Conclusion Introduction Standardized test preparation courses such as the International English Language Testing System (IELTS) in preparatory programmes at government and private universities in the United Arab Emirates (The UAE) prepare students to meet their exit score requirement.Curriculum planning has been driven towards these tests. Some research studies (Read and Hayes, 2003) warned against that type of curriculum in that it is restricted to those areas that are to be tested with no balance or integrated skills course design. Curriculum designers therefore depend heavily on materials geared to practice for the required exam. Moreover, Alderson and Hamp-Lyons (1996) argue that these type of curriculum are characterised by ‘teaching to the test’ and even text-book slaves.On the other hand, this essay will argue that designing and implementing curriculum that is supported by materials that reflect and meet the test question types and objectives can be beneficial to students preparing for standardised tests in the UAE. This paper will be outlined as follows: After introducing the conceptual framework of the essay, it will highlight the IELTS’ mark ing criteria of ‘Lexical Resource’ and ‘Grammatical Range’ in the hope to establish the argument that half of the IELTs marking criteria is based on lexical and grammatical items.Then, the type of lexical and structural syllabus that can be proposed to meet students’ needs in enhancing the aspect of English Language accuracy will be described. Some practical guidelines will then be provided for designing a lexical-structural syllabus. Conceptual Framework This essay falls within the conceptual framework of critical theory and critical pedagogy.From a social reconstructionism perspective, critical theory and critical pedagogy call for critically re-addressing status quo issues in education. Paolo Freire (1972), a critical theorist, believes that in most education systems there exists a hidden curriculum that underlines the status quo supported by the dominant culture and thus repressing minority subcultures. Curriculum planning should undergo a consta nt progressive change due to the progressive nature of the country and educational systems in the UAE.Curriculum change at the level of preparatory programmes is an inseparable part of this change since it is observed that research has provided enough evidence see published annually (IELTS, 2011) that in the UAE and other Arab countries where the majority of students and IELTS test takers struggle to reach a level of proficiency matching score band 5. 0 that enables them to proceed to their major.This judgment is supported by professional and formal data provided in Table 1 below which is adapted from the IELTS Annual Review of 2011. These records include a comparison of IELTS results by mother tongue candidates and by country of origin for the many countries using the IELTS tests. Table 1. Mean band score for the most frequent countries or regions of origin Country: The UAE| Listening| Reading| Writing| Speaking| Overall | Academic Module| 4. 9| 4. 8| 4. 7| 5. 4| 5. 0| General Modu le| 4. 3| 3. 5| 4. 3| 4. 9| 4. 3|It is shocking to learn that the UAE with all of its progressive trends in economy and education, coupled with billions of dollars supporting the educational system where ministries of Higher education and Education receive about 22 % of the total budget in the UAE (Ministry of Finance, 2011), is ranked at the bottom of the ranking list in IELTS scores (See Appendix A). This inexplicable fact urgently calls for critical pedagogy in education and in curriculum design in order to problematise the issue and arrive at proper and sustainable solutions.This is why this essay adopts a critical pedagogical approach. Within critical pedagogy paradigm, â€Å"curriculum perspective emphasizes the roles schools and learners can and should play in addressing social injustices and inequality. Curriculum development is not seen as a neutral process. † (Richards, 2001: 118). Similarly, Freire (1972) and Apple (1986) contend that schools do not present equal o pportunities so learners and teachers must be engaged to address important social and personal problems and seek learners’ empowerment.Morris (1995:10), furthermore, argues â€Å"Social injustices and inequality would be central issues in the curriculum. † Accordingly, curriculum designers in the UAE should study at the context in which learners are taught and start to constantly problematise current issues in order to improve their designs. In Applied linguistics, Pennycook (2001: 5) asserts that applied linguistics from a critical pedagogy perspective â€Å"is always concerned how the classroom, text, or conversation is related to broader social cultural and political relations,†.This paper is intended to problematise the issue of curriculum design in the UAE contest in that it seeks to establish that solid elements of lexical-structural syllabus can lead to a better success rate in IELTS. IELTS Marking Rubrics: Lexical Resources & Grammatical Range While ampl e IELTS research studies in university environments is available (see research reports in IELTS. org), there appears to be less research on the impact of Speaking and Writing assessment criteria or IELTS band descriptors on syllabus design.IELTS measures both test candidates’ abilities to produce different sentence structures, and the range and appropriateness of vocabulary. Consequently, lexical resources and grammatical range are regarded as significant constituents of the speaking and writing assessment, since they determine the level of vocabulary and grammar in speaking and writing the test candidates use. Test candidates are rated using detailed descriptors in IELTS. In the Speaking test, these describe spoken performance according to four different criteria: Fluency and Coherence, Lexical resources, Grammatical Range and Accuracy, and Pronunciation.In the Writing test, the descriptors assess candidates in terms of: Task achievement or response, Coherence and cohesion, Lexical Resources, and Grammatical Range and Accuracy (IELTS handbook, 2004: 18-23). It is clear from the speaking and writing assessment criteria that lexical resources and grammatical range constitute half of the assessment. They describe test takers’ level in grammar and vocabulary in the speaking and writing modules in detail. They should therefore occupy a big part of any curriculum design for IELTs preparation syllabus.Accordingly, this essay argues that a balanced lexical-structural syllabus can help in meeting the course objectives and students’ needs of any IELTS preparation course. Lexical Syllabus: Empowering students’ Lexical Resources in IELTS Historically, syllabuses were not lexical but structural. With the advent of the Communicative Approach, new ideas in syllabus design have been introduced with a focus on functions. In 1990, when Willis wrote The Lexical Syllabus, a large amount of attention to the Lexical syllabus surfaced. Willis (1990, P. v i)) identified three major aspects for lexis in lexical syllabus.First, the language to be learned should be somehow graded to make it easier for the learners to understand. Second, the language should be selected carefully without random exposure. Thirdly, the language syllabus should be itemised Over the past thirty years the Communicative Language Teaching (CLT), however, has not emphasized direct instruction of vocabulary or grammar (Laufer 2006). CLT should therefore be regarded as inappropriate for IELTS preparation exam courses and more explicit instruction of lexis and grammar could help students achieve their desired score in IELTS.In fact, according to Zahar, Cobb and Spada (2001), three periods of direct English instruction every week would lead to the incidental learning of only 70 new words a year. Therefore, acquisition of vocabulary through reading must be supplemented by direct instruction, which can be supported by incorporating a lexical syllabus. In addition, the lexical syllabus can make academic English vocabulary needed in IELTS more learnable to lower-level learners.There is also a well established connection between vocabulary knowledge and successful academic reading (Corson 1997; Nation 2001:144-149). Thus, the significance of incorporating lexical syllabus is significantly clear for large-scale high-stakes tests such as IELTS that assess the range of vocabulary exhibited in a student’s writing and speaking (See Appendix A). It should be admitted that producing satisfactory IELTS writing and speaking responses is not achieved by just learning individual vocabulary and grammatical structures.However, lexical items can lead to further enhancing students overall proficiency in IELTS. Some research ((Laufer, 2005; Nation 2001) has proven that lexical resources are the basis for comprehending grammar and lexical meanings. It is also the most important predictor of writing production (Read, 2002) and reading comprehension (Beck et. a l. 1987:147). In the IELTS speaking test, a similar lexical diversity would also be needed as suggested by Read (2005) who found that diverse lexical resources with some sophistication are rated higher.Regarding the adequate vocabulary range for academic success in international standardised tests, researchers (Laufer, 1992; Nation, 2001) argued that students, depending on the genre, need about 3,000 words in order for them to be able to read ungraded academic text and that the minimum word family level is the 3,000 word level needed for reading an unsimplified text. For IELTS reading texts, about 4,000 word families might be needed, 2,000 of General Service List (GSL) and about 570 words from the Academic Word List (AWL) (Coxhead, 1998) and about 1,000 technical words, proper nouns and low frequency vocabulary (Nation, 2001:147).Accordingly, commercially-produced syllabi such as Headway do not provide the wide range of lexical resources needed by the students (Horst, 2005; Lee, 200 7) in order for them to develop their lexis repertoire and to hence get the score required to pass IELTS. Moreover, there is enough suggestive research evidence (See Beaton, Grunederg and Ellis, 1995 – in Nation, 2001) that would indicate that the more vocabulary range manifested in speaking, the higher the score a student would get in writing.This shows that direct learning is in fact both effective and efficient and that the AWL can provide a reasonable target vocabulary for IELTS preparation courses and syllabi. In terms of the relation of grammar and vocabulary, David Wilkins (cited in Lewis, 2000: 8) noted that â€Å"Without grammar little can be conveyed, without vocabulary nothing can be conveyed†. This statement brings to focus the significance of incorporating lexical with a grammatical element when designing a syllabus aiming at enhancing students’ structural language system and accuracy and hence achieving a higher score in IELTS.The following section will shed some light on the significance of developing students’ accuracy through grammatical structures in structural syllabi. Structural Syllabus: Revisiting grammatical syllabus The structural syllabus comprises of a group of grammatical items that are selected and graded according to simplicity and complexity (Nunan, 1988). A structural-lexical syllabus’ principle objective is to help the learners acquire the grammatical structures and vocabulary of the language they are learning.The grammatical structures are usually presented in a certain sequence, for example, present followed by past, or in contrasting pairs, for example, simple present versus simple past (Long and Crookes, 1993). Wilkins (1976) describes this type of syllabus design approach as synthetic in which the various parts of language are separately taught and there is a â€Å"gradual accumulation of the parts until the whole structure of the language has been built up†. (p. 2) Wilkins suggests language is structured into smaller items and units in a structural syllabus and then it is presented discretely.Wilkins argues that this exposes learners to particular samples of language in which one grammatical feature dominates each lesson in the syllabus. This is mainly based on the hypothesis that grammatical and lexical rules are learned in an additive process, by which learners gain complete mastery of each item before a new one is taught (Nunan, 1988). It is hence the learner's task to re-synthesize the grammatical and lexical items that have been presented in a separate way (Wilkins, 1976). An important question raised here is the set of criteria teachers used to select which lexical and grammatical were to be used in the structural syllabus.This will be reviewed in the following sections. Selection of Lexical and Grammatical Items in Lexical-structural Syllabus: Mackey (1965) takes the stance that the lexical items in a structural syllabus are to be selected on the basis of criteria of frequency, coverage, availability, and learnability (White 1988:49-50). In the 1990s, the lexical approach in teaching and learning started to gain formal prominence among many researchers (Willis, 1990; Willis, 1996). As proposed by the lexical approach, a target reservoir of vocabulary is suggested to be learned at a specific level (Willis, 1990; Willis, 1996).These target vocabulary are derived from old corpus analysis of written and spoken English, but this could be regarded now as old-fashioned with the advancement of computerized corpus. However, one view, suggested by Richards (2001:154) is that lexical syllabuses were among the early types of syllabuses to be designed in language teaching. Regarding grammar structure, Mackey also (1965) proposes four criteria for the selection grammatical structures: simplicity, regularity, frequency, and contrastive difficulty.The last item, contrastive difficulty, means that learning some items at an early stage should rese mble the same item acquired in L1 (cited in Wilkins 1976: 6). These views of language selection have been challenged (e. g. , Ellis 2001). due to its lack of empirical experiments. Thus to determining the level of complexity, simplicity, regularity, etc. is almost entirely based on language teachers and syllabus designers’ common sense judgments. This dilemma, addressed in the following section, is one of the major shortcomings of the structural syllabus.Structural Syllabus: Drawbacks A number of limitations have been identified with the structural syllabus. The first shortcoming is drawn from Corder's (1967) â€Å"built-in syllabus† concept in which learners’ acquisition of different grammatical items is based on a natural order. This notion is supported by many researchers (for example, Hyltenstam and Pienemann, 1985). Within the framework of a built-in syllabus, grammatical complexity of a structure is not the reason for the difficulty of learning a structure of a language item.For instance, a common example of Arab Learners’ frequent mistakes is the acquisition of third person â€Å"s† morpheme (Hajjaj, 1999). Even though an average teacher can explain this item very well to Arabic learners, it always poses a challenge to them, which means difficulty or simplicity of language structure doesn’t always guarantee proper learning. This should be taken into serious consideration when designing a lexical-grammatical syllabus for IELTS preparation courses and thus adding meaningful activities that make the structures more memorable and relevant to learners’ needs and levels.The second main negative aspect of the structural syllabus is its lack of language functions. The structural syllabus is a reliable tool for assisting language learners to perfect common grammar rules. However, in terms of sociolinguistics, structural syllabus can prepare learners to be grammatically competent but communicatively incompetent (Jo hnson, 1982). A learner who replies ‘Yes, I do. ’ to ‘Do you mind if I open the window? ’ is a typical example of a learner whose command of English grammar might be ideal yet they are not capable of producing socially appropriate utterances.In other words, the structural syllabus helps learners produce instances of language usage rather than language use (Widdowson, 1978). Accordingly, syllabus designers who consider implementing lexical-structural syllabus need to promote both usage and use in order to avoid turning the language introduced into something meaningless. Socially contextualised and relevant language items introduced appropriately within interactive technique can bridge the gap between use and usage and make grammar teachable and learnable. This will be discussed in the following section. The Structural Syllabus and Grammar TeachingBefore advocating a lexical-grammatical syllabus, it is necessary to first establish the significance of explicit grammar teaching and learning. One debate in second language pedagogy has been about the teachability of grammar. Some scholars have argued against teaching grammar since it leads to only minimal achievement in the acquisition of linguistic competence in English (Krashen, 1982; Krashen and Terrel, 1983; Prabhu, 1987). Others have contended that explicitly teaching formal grammar plays a significant role in the development of the L2 learners' interlanguage system (Rutherford, 1987; Ellis, 1990, 1993).This paper supports the grammarians’ arguments in favor of formal grammar instruction that enhances IELTS’ UAE students and test takers’ accuracy. The first argument is based on studies on the emergence of immersion programs in Canada where many English L1 students study French and thus have been immersed in meaning- focused input in French (Swain, 1985; Swain and Lapkin, 1995). The results of research studies revealed that the majority of students have developed na tive-like receptive skills, but their productive ones are still far from native-like norms.This has therefore suggested that meaning-focused input instruction, which is devoid of any formal grammar teaching, is more likely to result in fossilization. The second argument defending formal grammar instruction derives from research (e. g. , Felix, 1985; White, 1987; Schachter, 1989) that suggests that adult L2 learners cannot have full access to the same acquisition resources as children with L1 do. This makes positive evidence such as formal instruction and corrective feedback.White (1987:105) finds that French learners of English as a second language tend to make sentences like ‘John drank slowly his coffee’ (John buvait lentement son cafe). These types of sentences are grammatically incorrect in English, but acceptable in their language. Those learners receiving positive evidence cannot notice that there are certain rules for adverb position within a sentence in English unless they are taught so by formal grammar instruction, which implies that lexical-structural syllabus can play a decisive role in some formal aspects of L2 learning and thus in IELTs preparation courses.The third argument has been proposed by Ellis (1990, 1993) who argues that formal grammar instruction develops explicit awareness of grammatical elements which hence helps learners acquire implicit knowledge. Ellis believes that the explicit knowledge of grammar instruction facilitates learning in three methods (1993:98). First, it encourages them to monitor their language before and after production. Terrell (1991:61) observes that â€Å"monitoring can apparently interact with acquisition, resulting in learners acquiring their own output† which shows the importance of monitoring in formal language instruction.Second, explicit teaching and learning support the factor of noticing certain elements in the input. Therefore, † if learners know that plural nouns have an â₠¬â€s, they are more likely to notice the —s on the ends of nouns they hear or read in input and also more likely to associate the —s morpheme with the meaning more than one†. (Ellis,1993:98). This example illustrates how specific features of grammar are best learned explicitly.Third, formal and explicit grammar knowledge assists learners in noticing some gaps in their language production. Thus if, for instance, Arab learners learn that verbs take an –s in the third person singular form, they are more likely to notice the gap when adding the third person singular –s or not, which leads to more accurate production later as they produce similar structures. Furthermore, Celce-Murica (1991: 467-468) suggests that formal lexical and grammar teaching can enhance meaning and social function.For instance, drawing attention to the different spatial denotation of the prepositions in and on the learner will â€Å"find it useful to know quite explicitly that à ¢â‚¬Ëœin’ favors the placement of objects in three-dimensional containers and ‘on’ favors the placement of objects on two-dimensional flat surfaces† if provided with formal instruction. Second, in an example of grammar enhancing social functions, Celee-Murcia discusses that learners need to be aware of the different modal auxiliaries uses in polite requests such as the difference between â€Å"Can you open the door? and â€Å"Could you open the door? â€Å". Celce-Murcia and Hilles (1988:4) believe that learners need to study grammar because many of them are to take part in international tests such as IELTS and that â€Å"Typically, a major component of such exams is grammar. Therefore, to give these students an incomplete grounding in grammar, regardless of one's conviction about teaching it, is to do them a great disservice. Students have to know and apply the rules of English grammar in order to do well on such tests. A relatively recent argument pro posed by Ellis (2001) argues that it is a mistake to suppose that learning foreign languages in schools is entirely to promote communication among speakers of different languages. Learning foreign languages has a more valuable end, i. e. promoting intellectual growth. (Ellis 2001:172) thus views learning a foreign language as a means to develop cognitive abilities and that grammar contains knowledge that contributes to learners’ cognitive skills.If one considers all the arguments discussed earlier in this essay claiming that formal teaching of lexical and grammatical items should be an indivisible part of any English classes, structural-lexical syllabus seems indispensible for students preparing for international examination such as IELTS and TOEFL. This is, however, only my conviction within my teaching context and does not entail that curriculum designers and teachers are to revert to old traditional language teaching methodologies such as grammar-translation method. I elie ve that adopting a merely lexical and/or structural syllabus has not survived in language education because they overemphasized the aspect of grammar teaching through many tedious structural drills. I believe that it is unfair to judge that grammar and lexical instruction should be eliminated simply because of insufficiencies of audio-lingual or grammar-translation methods. A well-balanced lexical-structural syllabus can be safely employed, especially in IELTS preparation courses, provided that it is supplemented by motivating and communicative tasksSuggested steps to design a Lexical-Structural Syllabus According to Stern (1983:339-340), there are major differences between English as a second language (ESL) and English as a foreign language (EFL). Accordingly, EFL and ESL students are positioned in different learning conditions. ESL students learn Target Language (TL) in a more supportive setting in that they are more exposed to TL in its natural environment EFL. Considering needs and motivation, ESL students are more required and motivated to communicate with foreigners than EFL students.Thus, ESL students learning needs are definitely not centered on one factor or need. As previously discussed, the majority of Arab students in the UAE are EFL learners who study English in order to pass national exams, such as IELTS. This educational structure of teaching applies in the EFL environment in the UAE. Based on my observations and experience, I found that students in Emirati high schools are also required to learn specific discrete items such as vocabulary and grammatical items without any exposure to communicative tasks that enable them to use these items meaningfully.Unfortunately, only those who wish to pursue their higher education strive to master the grammatical and vocabulary items of their course books in order for them to achieve the required score of the university entrance exams. When taking these conditions into serious consideration, I have come to r ealize the reason why lexical and grammar instruction can, and perhaps should, be an essential component of any curriculum in the Emirati EFL educational system in schools, and this can be possible through the lexical-structural syllabus proposed in this essay.It is argued that (Richards, 2001) old and traditional language teaching methods did not succeed not because they adhered to a structural or lexical syllabus but because the implementation process was too mechanical through dull drills of discrete language items. I believe that a well-designed lexical-structural syllabus proposed in this essay can survive through two types of tasks: consciousness-raising and communicative tasks. Consciousness-raising activities can develop explicit knowledge and communicative tasks improve the implicit knowledge of lexical and grammatical structures (Ellis, 2001).As defined by Ellis (1993:109), consciousness-raising is â€Å"a deliberate attempt on the part of the teacher to make the learners aware of specific features of the L2†. The following task in Table 2 (Ellis, 2001:173) is a straightforward sample of a consciousness-raising (CR) activity that is designed to help learners be conscious of the difference between the prepositions ‘for’ and ‘since’. Table 2. An Example of a CR Problem-solving Task: 1. Here is some information about when three people joined the company they now work for and how long they have been working there.Name Date Joined Length of Time Ms Regan 1945 45 yrs Mr Bush 1970 20 yrs Ms Thatcher 1989 9 mths Mr Baker 1990 (Feb) 10 days 2. Study these sentences about these people. When is â€Å"for† used and when is â€Å"since† used? a. Ms Regan has been working for her company for most of her life. b. Mr Bush has been working for his company since 1970. c. Ms Thatcher has been working for her company for 9 months. . Mr Baker has been working for his company since February. 3. Which of the following sentence s are ungrammatical? Why? a. Ms Regan has been working for her company for 1945. b. Mr Bush has been working for his company for 20 years. c. Ms Thatcher has been working for her company since 1989. d. Mr Baker has been working for his company since 10 days. 4. Try and make up a rule to explain when â€Å"for† and â€Å"since† are used. 5. Make up one sentence about when you started to learn English and one sentence about how long you have been studying English.Use â€Å"for† and â€Å"since†. | It is my argument that if this exercise is supported by a communicative task where learners interact in pairs or with the classroom teacher it would enable them to internalise the target structure and use it in real-life communication. Designing a lexical-structural based syllabus is not a very hard task. There are many resources based on well-established research that can help syllabus designers build up their course (Lewis, 1997a,1997b). This essay offers some s uggestions on how to build a lexical-structural activities and syllabus.First, syllabus designers need to settle a relevant set of structures, vocabulary and different types of collocations (Lewis,1997a) and organise them according to their increasing complexity, starting from simple structures to complex ones. Second, syllabus designers should choose categories of vocabulary items to be learned and sequence the categories according to situational and functional characteristics such as vocabulary related topics such as environment, health, internet, education and so on since IELTS module is commonly based on globally social topics.Third, syllabus designers should integrate the group of vocabulary and structures together into learning objectives to base the units of lexical-structural syllabus. Although this essay advocates the implementation of a lexical-structural syllabus in IELTS preparation course at preparatory programmes in the UAE, it is necessary to admit that an eclectic an d multi-focus syllabus which includes a wide range of components, specifically vocabulary, structures introduced through functions and, situations can allow for a two-sided approach.This can emphasise the importance of systematically learning specific vocabulary and structures and building up linguistic knowledge, and at the same time it can focus on meeting immediate language communicative needs and sociolinguistic and pragmatic competence. Conclusion In my essay, I only propose that lexical-structural syllabus can safely be employed in EFL settings, particularly in the UAE. Previously, it was argued that the majority of EFL students in the UAE must score reasonably high in their university entrance English exams. This condition requires them to learn many grammatical and lexical structures.It is my suggestion that a lexical-structural syllabus can efficiently serve this purpose. Nevertheless, this argument should not be interpreted as a call for a return to old approaches such as the grammar-translation method. What I have attempted to proposed in this essay is two-fold. First, designing a well-balanced external syllabus that adequately matches the learners’ internal syllabus and second, empowering this potential syllabus with innovative learning ideas and tasks discussed earlier, such as consciousness-raising and communicative tasks.I hope my recommendations will breathe fresh life into the body of the proposed lexical-structural syllabus. References Apple, L. 1986. Teachers and texts. New York: Routlege and Kegan Paul. Beck, I. L. , McKeown, M. G. , and Omanson, R. C. (1987). The effects and uses of diverse vocabulary instructional techniques. In M. McKeown and Curtis, M. E. (eds. ) (1987). The Nature of Vocabulary Acquisition. Lawrence Erlbaum Associates: 147 – 163. Celce-Murcia, M. (1991). Grammar pedagogy in second and foreign language teaching. TESOL Quarterly, 25, 45a-480. Celce-Murcia, M. , and Hilles, S. (1988). Techniques and resource s in teaching grammar.Oxford: Oxford University Press. Corder, S. P. (1967). The significance of learners' errors. IRAL, 5, 161-169. Corson, D. 1997. The learning and use of academic English words. Language Learning, 47 (4), 671 718. Coxhead, A. (1998). An Academic Word List. Occasional Publication Number 18, LALS, Victoria University of Wellington, New Zealand. Dixon, R. J. (1977). Modern American English: Book 2. New Edition. New York: Regents Publishing Company, Inc. Ellis, R. (1990). Instructed second language acquisition. Oxford: Basil Blackwell. Ellis, R. (1993). The structural syllabus and second language acquisition. TESOT Quarterly, 27, 91113.Ellis, R. (2001). Grammar teaching-practice or consciousness-raising? In J. C. Richards, and W. A. Renandya (Eds. ), Methodology in language teaching: An anthology of current practice (pp. I 67-174). Cambridge: Cambridge University Press. English Language Services. (1964). English 900. New York: Collier Macmillan. Felix, S. (1985). Mor e evidence on competing cognitive systems. Second Language Research, 1, 47-72. Freire, P. 1972. Pedagogy of the oppressed. New York: Herder and Herder. Hajjaj, A. (1999). Arab students` writing mistakes: Renewing the issue. Dirasat: Human and Social Sciences, University of Jordan, 26. : 621-633. Horst, M. (2005). Learning L2 vocabulary through extensive reading: A measurement study. Canadian Modern Language Review, 61, 355–382. Hyltenstam, K. , and Pienemann, M. (Eds. ). (1985). Modelling and assessing second language acquisition. Clevedon, Avon: Multilingual Matters. IELTS. (2007). IELTS Handbook 2007. Cambridge: USLES. Retrieved on 15/12/2012 from: < http://www. ielts. org/pdf/IELTS_Handbook_2007. pdf>. IELTS. (2011). IELTS Test taker performance 2011. Cambridge: USLES. Accessed 15/12/2012 from .Johnson, K. (1982). Communicative syllabus design and methodology. Oxford. Oxford University Press Krashen, S. (1982). Principles and practice in second language acquisition. Oxford : Pergamon. Krashen, S. , and Ten-el, T. (1983). The natural approach: language acquisition in the classroom. Oxford: Pergamon. Laufer, B. (2006). Comparing focus on form and focus on forms in second-language vocabulary learning. The Canadian Modern Language Review, 63 (1), 149-166. Laufer, B. (1992). How much lexis is necessary for comprehension? In P. Arnaud and H. Bejoint (eds. ) Vocabulary and Applied Linguistics. Macmillan; London. Lee, S-Y. 2007). Revelations from three consecutive studies on extensive reading. RELC Journal, 38, 150–170. Lewis, M. (1997a). Implementing the lexical approach: Putting theory into practice. Hove, England: Language Teaching Publications. Lewis, M. (1997b). Pedagogical implications of the lexical approach. In J. Coady & T. Huckin (Eds. ),  Second language vocabulary acquisition: A rationale for pedagogy  (pp. 255-270). Cambridge: Cambridge University Press. Long, M. , and Crookes, G. (1993). Units of analysis in syllabus design: The case for task. In G. Crookes, and S. M. Gass (Eds. ), Tasks in a pedagogical context (pp. 9-54).Clevedon, Avon: Multilingual Matters. Mackey, W. F. (1965). Language teaching analysis. London: Longman. Ministry of Finance, 2011. Budget Report. Accessed 15/12/2012 . Morris, P. (1995). The Hong Kong curriculum. Hong Kong: Hong Kong University Press Nation, I. S. P. (2001). Learning vocabulary in another language. Cambridge: Cambridge University Press. Nunan, D. (1988). Syllabus design. Oxford: Oxford University Press. Pennycook, A. (2001). Critical Applied Lingistics: A Critical Introduction. Mahwah, New Jersey and London: Lawrence Erlbaum Associates.Pienemann, M. (1985). Learnability and syllabus construction. In K. Hyltenstam, and M. Pienemann (Eds. ), Modelling and assessing second language acquisition. Clevedon, Avon: Multilingual Matters. Prabhu, N. S. (1987). Second language pedagogy. Oxford: Oxford University Press. Read, J. (2002). Assessing Vocabulary. Cambridge: Cambridge Universi ty Press. Read, J. (2005). Applying lexical statistics to the IELTS speaking test. Research Notes, May, 2005. Richards,  J. C. (2000). Curriculum development in language teaching. New York: Cambridge University Press. Rutherford, W. (1987). Second language grammar: Teaching and learning.London: Longman. Schachter, J. (1989). Testing a proposed universal. In S. Gass and J. Schachter (Eds. ), Linguistic perspectives on second language acquisition (pp. 73-88). Cambridge: Cambridge University Press. Soars, J. , and Soars, L. (1991). Headway: Pre- Intermediate. Oxford: Oxford University Press. Stern, H. H. (1983). Fundamental concepts of language teaching. Oxford: Oxford University Press. Swain, M. (1985). Communicative competence: Some roles of comprehensible input and comprehensible output in its development. In S. Gass and C. Madden (Eds. ), Input in Second language acquisition (pp. 235-253). Rowley, MA: Newbury House.Swain, M. and Lapkin, S. (1995). Problems in output and the cogni tive processes they generate: a step towards second language learning. Applied Linguistics, 16, 371-91. Terrell, T. D. (1991). The role of grammar instruction in a communicative approach. Modern Language Journal, 75, 52-63. Van Patten, B. (1992). Second language acquisition research and foreign language teaching, Part I. ADFL Bulletin, 23, 52-55. White, L. (1987). Against comprehensible input: The input hypothesis and the development of second language competence. Applied Linguistics, 8, 95- 110. White, R. V. (1988). The ELT curriculum. Design, innovation and management.Oxford: Basil Blackwell. Widdowson, H. G. (1978). Teaching language as communication. Oxford: Oxford University Press. Wilkins, D. (1976). Notional syllabuses. Oxford:Oxford University Press. Willis, D. (1990). The lexical syllabus: A new approach to language teaching. London: Collins. Willis. J. (1996). A framework for task-based learning. Oxford: Longman. Yalden, J. (1983). The communicative syllabus: Evolution, de sign and implementation. Oxford: Pergamon. Zahar, R. , Cobb, T. and Spada, N. 2001. Acquiring vocabulary through reading: Effects of frequency and contextual richness. The Canadian Modern Language Review, 57 (4), 541-572.Appendix A (Adapted from www. ielts. org) IELTS | Researchers – Test taker performance 2011 Band score information Place of Origin These figures show the mean overall and individual band scores achieved by 2011 Academic and General Training candidates from the top 40 places of origin. Mean band score for the most frequent countries or regions of origin (Academic) Academic| Listening| Reading| Writing| Speaking| OVERALL| Bangladesh| 5. 9| 5. 6| 5. 6| 5. 9| 5. 8| Brazil| 6. 8| 6. 9| 6. 2| 6. 9| 6. 7| China (People's Republic)| 5. 8| 5. 9| 5. 2| 5. 3| 5. 6| Colombia| 6. 2| 6. 4| 5. 7| 6. 4| 6. 2| Cyprus| 6. 4| 5. 9| 5. 7| 6. | 6. 1| Egypt| 6. 3| 6. 1| 5. 8| 6. 3| 6. 2| France| 6. 9| 6. 9| 6. 3| 6. 8| 6. 8| Germany| 7. 0| 6. 6| 6. 4| 7. 2| 6. 8| Ghana| 4. 8| 4. 3 | 5. 7| 6. 5| 5. 4| Hong Kong| 6. 7| 6. 4| 5. 9| 6. 2| 6. 4| India| 6. 4| 5. 6| 5. 8| 6. 2| 6. 1| Indonesia| 6. 6| 6. 3| 5. 9| 6. 3| 6. 4| Iran| 5. 8| 5. 4| 5. 8| 6. 3| 5. 9| Iraq| 5. 7| 5. 2| 5. 4| 6. 2| 5. 7| Italy| 6. 2| 6. 1| 5. 8| 6. 4| 6. 2| Japan| 6. 0| 5. 6| 5. 5| 5. 8| 5. 8| Jordan| 5. 9| 5. 5| 5. 5| 6. 3| 5. 9| Kenya| 6. 7| 6. 1| 6. 6| 7. 2| 6. 7| Korea, South| 6. 2| 6. 1| 5. 4| 5. 7| 5. 9| Kuwait| 5. 3| 4. 9| 4. 8| 5. 6| 5. 2| Libya| 5. 2| 5. 1| 5. 1| 5. 8| 5. 4| Malaysia| 7. | 7. 0| 6. 2| 6. 6| 6. 9| Mexico| 6. 7| 6. 8| 5. 9| 6. 5| 6. 6| Nepal| 6. 3| 5. 8| 5. 7| 6. 0| 6. 0| Nigeria| 6. 1| 6. 0| 6. 2| 7. 0| 6. 4| Oman| 5. 3| 5. 1| 5. 0| 5. 7| 5. 4| Pakistan| 5. 8| 5. 5| 5. 6| 6. 0| 5. 8| Philippines| 7. 0| 6. 6| 6. 2| 6. 8| 6. 7| Qatar| 4. 8| 4. 6| 4. 5| 5. 3| 4. 9| Romania| 7. 2| 7. 0| 6. 2| 6. 8| 6. 9| Russia| 6. 6| 6. 6| 5. 9| 6. 6| 6. 5| Saudi Arabia| 4. 9| 4. 8| 4. 7| 5. 6| 5. 1| Spain| 6. 7| 6. 9| 6. 0| 6. 5| 6. 6| Sri Lanka| 6. 6| 6. 0| 5. 9| 6. 5| 6. 3| Sudan| 5. 9| 5. 7| 5. 5| 6. 2| 5. 9| Taiwan| 5. 9| 6. 0| 5. 5| 5. 9| 5. 9| Thailand| 5. 9| 5. 9| 5. 3| 5. 7| 5. 8| Turkey| 6. 0| 5. | 5. 3| 5. 8| 5. 8| United Arab Emirates| 4. 9| 4. 8| 4. 7| 5. 4| 5. 0| Uzbekistan| 5. 7| 5. 6| 5. 0| 5. 5| 5. 5| Vietnam| 5. 9| 6. 1| 5. 6| 5. 7| 5. 9| Mean band score for the most frequent countries or regions of origin (General Training) General Training| Listening| Reading| Writing| Speaking| OVERALL| Bangladesh| 6. 1| 5. 4| 5. 8| 6. 2| 5. 9| Brazil| 6. 4| 6. 3| 6. 1| 6. 7| 6. 4| China (People's Republic)| 6. 3| 6. 0| 5. 7| 5. 8| 6. 0| Colombia| 5. 7| 5. 7| 5. 6| 6. 0| 5. 8| Egypt| 6. 3| 5. 9| 5. 9| 6. 4| 6. 2| France| 6. 9| 6. 8| 6. 3| 6. 8| 6. 8| Germany| 6. 9| 6. 6| 6. 4| 7. 1| 6. 8| Hong Kong SAR| 6. 7| 6. 4| 5. | 6. 2| 6. 4| India| 6. 3| 5. 7| 5. 8| 6. 2| 6. 1| Indonesia| 6. 7| 6. 2| 5. 9| 6. 3| 6. 3| Iran| 5. 7| 5. 4| 5. 8| 6. 3| 5. 9| Iraq| 5. 7| 5. 3| 5. 5| 6. 3| 5. 8| Italy| 6. 2| 6. 3| 5. 8| 6. 4| 6. 2| Japan| 6| 5. 6| 5. 5| 5. 9| 5. 8| Jordan| 6| 5 . 6| 5. 5| 6. 3| 5. 9| Kenya| 6. 8| 6. 4| 6. 7| 7. 3| 6. 9| Korea, South| 5. 8| 5. 4| 5. 3| 5. 4| 5. 5| Lebanon| 6. 3| 5. 7| 5. 9| 6. 6| 6. 2| Malaysia| 7. 3| 6. 9| 6. 5| 7. 0| 7. 0| Mauritius| 6. 5| 5. 9| 6. 2| 6. 7| 6. 4| Mexico| 6. 3| 6. 3| 6. 0| 6. 6| 6. 3| Nepal| 6. 3| 5. 6| 5. 9| 6. 2| 6. 1| Nigeria| 5. 9| 5. 9| 6. 6| 7. 1| 6. 4| Pakistan| 6. 2| 5. 5| 6. 0| 6. 4| 6. 1|Philippines| 6. 2| 5. 7| 5. 9| 6. 3| 6. 1| Romania| 6. 5| 6. 3| 5. 9| 6. 2| 6. 3| Russia| 6. 4| 6. 3| 6. 0| 6. 4| 6. 3| Saudi Arabia| 4. 6| 3. 8| 4. 4| 5. 0| 4. 5| Singapore| 7. 7| 7. 3| 6. 9| 7. 5| 7. 4| South Africa| 7. 4| 7. 0| 7. 3| 8. 4| 7. 6| Sri Lanka| 6. 3| 5. 7| 5. 8| 6. 3| 6. 1| Taiwan| 6. 3| 5. 8| 5. 8| 6. 2| 6. 1| Thailand| 5. 6| 5. 2| 5. 3| 5. 6| 5. 5| Turkey| 5. 8| 5. 5| 5. 5| 5. 8| 5. 7| Ukraine| 5. 8| 5. 6| 5. 5| 6. 0| 5. 8| United Arab Emirates| 4. 3| 3. 5| 4. 3| 4. 9| 4. 3| United States of America| 8. 0| 7. 6| 7. 7| 8. 6| 8. 0| Venezuela| 6. 3| 6. 3| 6. 1| 6. 4| 6. 3| Vietnam| 5. 8| 5. 6| 5. 7| 5. 7| 5. 8|

Tuesday, July 30, 2019

History of Digital Computer

The History of Digital Computers B. RANDELL Computing Laboratory, University of Newcastle upon Tyne This account describes the history of the development of digital computers, from the work of Charles Babbage to the earliest electronic stored program computers, It has been prepared for Volume 3 of â€Å"l’Histoire Generale des Techniques,† and is in the main based on the introductory text written by the author for the book â€Å"The Origins of Digital Computers: Selected Papers† (Springer Verlag, 1973). . Charles Babbage THE first electronic digital computers were completed in the late 1940’s. In most cases their developers were unaware that nearly all the important functional characteristics of these computers had been invented over a hundred years earlier by Charles Babbage. It was in 1821 that the English mathematician Charles Babbage became interested in the possibility of mechanising the computation and printing of mathematical tables.He successfully constructed a small machine, which he called a â€Å"difference engine,† capable of automatically generating successive values of simple algebraic functions by means of the method of finite differences. This encouraged him to plan a full-scale machine, and to seek financial backing from the British government. During the next 12 years both Babbage and the government poured considerable sums of money into the attempt at building his Difference Engine.However the project, which called for the construction of six interlinked adding mechanisms, each capable of adding two multiple-digit decimal numbers, together with an automatic printing mechanism, was considerably beyond the technological capabilities of the era – indeed it has been claimed that the efforts expended on the Difference Engine were more than justified simply by the improvements they generated in mechanical engineering equipment and practice.Although Babbage’s plans for a Difference Engine were somewha t premature, the basic scheme was vindicated when in 1843, inspired by their knowledge of his work, George and Edvard Scheutz successfully demonstrated a working prototype difference engine. A final version of this model was completed 10 years later, with financial assistance from the Swedish government. Several other difference engines ere constructed in the decades that followed, but such machines never achieved the importance of more conventional calculating machines, and when multi-register accounting machines became available in the 1920’s it was found that these could be used essentially as difference engines. However Babbage’s ideas soon progressed far beyond that of a special-purpose calculating machine – in fact almost as soon as he started work on his Difference Engine he became dissatisfied with its limitations.In particular he wished to avoid the need to have the highest order of difference constant, in order to be able to use the machine directly fo r transcendental as well as algebraic functions. In 1834 Babbage started active work on these matters, and on problems such as division and the need to speed up the part of the addition mechanism which dealt with the assimilation of carry digits. He developed several very ingenious methods of carry assimilation, but the time savings so obtainable would have been at the cost of a considerable amount of complex machinery.This led Babbage to realise the advantages of having a single centralised arithmetic mechanism, the â€Å"mill,† separate from the â€Å"figure axes,† i. e. , columns of discs which acted merely as storage locations rather than accumulators. Babbage’s first idea for controlling the sequencing of the various component mechanisms of the engine was to use â€Å"barrels,† i. e. , rotating pegged cylinders of the sort used in musical automata. He first planned to use a set of subsidiary barrels, with over-all control of the machine being specifi ed by a large central barrel with exchangeable pegs.However in June 1836 he took the major step of adopting a punched card mechanism, of the kind found in Jacquard looms, in place of the rather limited and cumbersome central barrel. He did so in the realisation that the â€Å"formulae† which specified the computation that the machine was to perform could therefore be of almost unbounded extent, and that it would be a simple matter to change from the use of one formula to another.Normally formula cards, each specifying an arithmetic operation to be performed, were to be read by the Jacquard mechanism in sequence, but Babbage also envisaged means whereby this sequence could be broken and then recommenced at an earlier or later card in the sequence. Moreover he allowed the choice of the next card which was to be used to be influenced by the partial results that the machine had obtained.These provisions allowed him to claim that computations of indefinite complexity could be perf ormed under the control of comparatively small sets of formula cards. Babbage talked at one time of having a store consisting of no less than 1000 figure axes, each capable of holding a signed 40-digit decimal number, and planned to provide for reading numbers from cards into the store, and for punching or printing the values of numbers held in the store.The movement of numbers between the mill and the store was to be controlled by a sequence of â€Å"variable cards,† each specifying which particular figure axis was involved. Therefore an arithmetic operation whose operands were to be obtained from the store and whose result was to be returned to the store would be specified by an operation card and several variable cards. He apparently intended these different kinds of control cards to be in separate sequences, read by separate Jacquard mechanisms.Thus in the space of perhaps 3 years Babbage had arrived at the concept of a general purpose digital computer consisting of a sto re, arithmetic unit, punched card input and output, and a card-controlled sequencing mechanism that provided iteration and conditional branching. Moreover although he continued to regard the machine, which he later came to call the Analytical Engine, as being principally for the construction of mathematical tables, he had a very clear grasp of the conceptual advances he had made.Basing his claim on the unbounded number of operation and variable cards that could be used to control the machine, the ease with which complicated conditional branches could be built from a sequence of simple ones, and the fact that automatic input and output, and multiple precision arithmetic, were provided, he stated that â€Å". . . it appears that the whole of the conditions which enable a finite machine to make calculations of unlimited extent are fulfilled in the Analytical Engine . . . . I have converted the infinity of space, which was required by the conditions of the problem, into the infinity of time. Because separate, but associated, sequences of cards were needed to control the Analytical Engine the concept of a program as we know it now does not appear very c1early in contemporary descriptions of the machine. However there is evidence that Babbage had realised the fact that the information punched on the cards which controlled the engine could itself have been manipulated by an automatic machine-for example he suggested the possibility of the Analytical Engine itself being used to assist in the preparation of lengthy sequences of control cards.Indeed in the description of the use of the Analytical Engine written by Lady Lovelace, in collaboration with Babbage, there are passages which would appear to indicate that it had been realised that an Analytical Engine was fully capable of manipulating symbolic as well as arithmetical quantities. Probably Babbage himself realised that the complete Analytical Engine was impractical to build, but he spent much of the rest of his l ife designing and redesigning mechanisms for the machine.The realisation of his dream had to await the development of a totally new technology, and an era when the considerable finances and facilities required for an automatic computer would be made available, the need at last being widely enough appreciated. He was a century ahead of his time, for as one of the pioneers of the modern electronic digital computer has written: â€Å"Babbage was moving in a world of logical design and system architecture, and was familiar with and had solutions for problems that were not to be discussed in the literature for another 100 years. †He died in 1871, leaving an immense collection of engineering drawings and documents, but merely a small portion of the Analytical Engine, consisting of an addition and a printing mechanism, whose assembly was completed by his son, Henry Babbage. This machine and Babbage’s engineering drawings are now in the Science Museum, London. 2. Babbageâ€⠄¢s direct successors Some years’ after Babbage’s death his son Henry Babbage recommenced work on the construction of a mechanical calculating machine, basing his efforts on the designs his father had made for the Mill of the Analytical Engine.This work was started in 1888 and carried on very intermittently. It was completed only in about 1910 when the Mill, which incorporated a printing mechanism, was demonstrated at a meeting of the Royal Astronomical Society. By this date however the work of a little-known successor to Charles Babbage, an Irish accountant named Percy Ludgate, was already well advanced. Ludgate started work in 1903 at the age of 20 on an entirely novel scheme for performing arithmetic on decimal numbers.Decimal digits were to be represented by the lateral position of a sliding metal rod, rather than the angular position of a geared disc. The basic operation provided was multiplication, which used a complicated mechanism for calculating the two-digit products resulting from multiplying pairs of decimal digits. together. The scheme involved first transforming the digits into a form of logarithm, adding the logarithms together, and then converting the result back into a two-digit sum.This scheme is quite unlike any known to have been used in earlier mechanical calculators, or for that matter since, although there had been several calculating machines constructed that used built-in multiplication tables to obtain two-digit products – the earliest known of these was that invented by Bollee in 1887. It is in fact difficult to see any advantages to Ludgate’s logarithmic scheme, although his form of number representation is reminiscent of that used in various mechanical calculating devices in the following decades.So striking are the differences between Ludgate’s and Babbage’s ideas for mechanical arithmetic that there is no reason to dispute Ludgate’s statement that he did not learn of Babbageâ€℠¢s prior work until the later stages of his own. It seems likely that Babbage was the eventual inspiration for Ludgate to investigate the provision of a sequence control mechanism. Here he made an advance over the rather awkward system that Babbage had planned, involving separate sets of operation and variable cards.Instead his machine was to have been controlled by a single perforated paper tape, each row of which represented an instruction consisting of an operation code and four address fields. Control transfers simply involved moving the tape the appropriate number of rows forwards or backwards. Moreover he also envisaged the provision of what we would now call subroutines, represented by sequences of perforations around the circumference of special cylinders-one such cylinder was to be provided The Institute of Mathematics and its Applications 2 for division.The machine was also to be controllable from a keyboard, a byproduct of whose operation would be a perforated tape which could then be used to enable the sequence of manually controlled operations to be repeated automatically. Ludgate estimated that his Analytical Machine would be capable of multiplying two twenty-digit numbers in about 10 seconds, and that, in considerable contrast to Babbage’s Analytical Engine, it would be portable. However there is no evidence that he ever tried to construct the machine, which he apparently worked on alone, in his spare time.He died in 1922, and even if at this time his plans for the Analytical Machine still existed there is now no trace of them, and our knowledge of the machine depends almost entirely on the one description of it that he published. The next person who is known to have followed in the footsteps of Babbage and to have worked on the problems of designing an analytical engine was Leonardo Torres y Quevedo. Torres was born in the province of Santander in Spain in 1852.Although qualified as a civil engineer he devoted his career to scientific re search, and in particular to the design and construction of an astonishing variety of calculating devices and automata. He gained great renown, particularly in France and in Spain, where he became President of the Academy of Sciences of Madrid, and where following his death in 1936 an institute for scientific research was named after him. Torres first worked on analog calculating devices, including equation solvers and integrators.In the early 1900’s he built various radio-controlled devices, including a torpedo and a boat which, according to the number of pulses it received, could select between various rudder positions and speeds, and cause a flag to be run up and down a mast. In 1911 he made and successfully demonstrated the first of two chess-playing automata for the end game of king and rook against king. The machine was fully automatic, with electrical sensing of the positions of the pieces on the board and a mechanical arm to move its own pieces. The second machine was built in 1922, and used magnets underneath the board to move the pieces. ) In all this work, he was deliberately exploiting the new facilities that electromechanical techniques offered, and challenging accepted ideas as to the limitations of machines. He picked on Babbage’s Analytical Engine as an important and interesting technical challenge, and in 1914 published a paper incorporating detailed schematic designs for a suitable set of electro-mechanical components.These included devices for storing, comparing and multiplying numbers, and were accompanied by a discussion of what is now called floating point number representation. He demonstrated the use of the devices in a design for a special-purpose program-controlled calculator. The program was to be represented by areas of conductive material placed on the surface of a rotating drum, and incorporated a means for specifying conditional branching. Torres clearly never intended to construct a machine to his design, but 6 yea rs later he built, and successfully demonstrated, a typewriter-controlled calculating machine primarily to demonstrate that an electromechanical analytical engine was completely feasible. He in fact never did build an analytical engine, although he designed, and in many cases built, various other digital devices including two more calculating machines, an automatic weighing machine, and a machine for playing a game somewhat like the game of Nim. However there seems little reason to doubt that, should the need have been sufficiently pressing, Torres would indeed have built a complete analytical engine.In the event, it was not until the 1939-1945 war that the desirability of largescale fully automatic calculating machines became so clear that the necessary environment was created for Babbage’s concept to become a reality. Before this occurred there is known to have been at least one further effort at designing an analytical engine. This was by a Frenchman, Louis Couffignal, who was motivated mainly by a desire to reduce the incidence of errors in numerical computations.He was familiar with the work of Babbage and Torres y Quevedo but, in contrast to their designs, proposed to use binary number representation. The binary digits of stored numbers were to be represented by the lateral position of a set of parallel bars controlled by electro-magnets. The various arithmetic operations were to be performed by relay networks, the whole machine being controlled by perforated tapes. Couffignal apparently had every intention of building this machine, in association with the Logabax Company, but presumably because of the war never did so.However after the war he was in charge of an electronic computer project for the Institut Blaise Pascal, the design study and construction of the machine being in the hands of the Logabax Company. With Couffignal’s pre-war plans, the line of direct succession to Babbage’s Analytical Engine seems to have come to an end. Most of the wartime computer projects were apparently carried out in ignorance of the extent to which many of the problems that had to be dealt with had been tackled by Babbage over a century earlier. However in some cases there is clear evidence that nowledge of Babbage’s work was an influence on the wartime pioneers, in particular Howard Aiken, originator of the Automatic Sequence Controlled Calculator, and William Phillips, an early proponent of binary calculation, and various other influential people, including Vannevar Bush and L. J. Comrie, were also well aware of his dream. 3. The contribution of the punched card industry An initially quite separate thread of activity leading to the development of the modern computer originated with the invention of the punched card tabulating system.The capabilities of Herman Hollerith’s equipment, first used on a large scale for the 1890 US National Census, were soon extended considerably. The original equipment allowed cards to hold binary information representing the answers to a Census questionnaire. These cards could be tabulated, one by one, using a machine which sensed the presence of holes in the card electrically and could be wired to count the number of cards processed in which particular holes or combinations of holes had been punched. A device could be attached to such a tabulator which assisted the manual sorting of cards into a number of separate sequences.Within 10 years automatic card handling mechanisms, which greatly increased the speed of machine operation, and addition units, which enabled card tabulators to sum decimal numbers punched on cards, had been provided. The system soon came into widespread use in the accounting departments of various commercial organisations, as well as being used for statistical tabulations in many countries of the world. After the 1900 US Census relations between Hollerith and the Census Bureau deteriorated, and the Bureau began to manufacture its own equ ipment for use in the 1910 Census.The person in charge of this work was James Powers who circumvented Hollerith’s patents by producing a mechanical card reading apparatus. He retained the patent rights to his inventions and formed his own company which eventually merged with Remington Rand in 1927. In 1911 Hollerith sold his own company, the Tabulating Machine Company, which he had formed in 1896, and it was shortly afterwards merged with two other companies to form the Computing-TabulatingRecording Company. This company which was under the direction of Thomas J.Watson from 1914 became the International Business Machines Corporation in 1924. During the 1920’s and 1930’s punched card systems developed steadily, aided no doubt by the stimulus of competition, not only in the USA but also in Britain, where the Hollerith and Powers-based systems continued to be marketed under the names of their original inventors, while in France a third manufacturer, Compagnie Machi nes Bull, was also active. Unfortunately the people involved in this work did not in general publish technical papers and their work has received little public recognition.Thus full appreciation of the contribution of IBM development engineers, such as J. W. Bryce, one of the most prolific inventors of his era, will probably have to await an analysis of the patent literature. One inventor whose work has, however, been documented is Gustav Tauschek, a self-taught Viennese engineer, with more than 200 patents in the computing field to his credit. While working for Rheinische Metallund Maschinenfabrik he designed and built a punched card electromechanical accounting machine.His other patents, many of which were filed whilst he was under contract to IBM during the 1930’s, also included a â€Å"reading-writing-calculating machine† which used photocells to compare printed input characters with templates held on photographic film, a number storage device using magnetised stee l plates, and an electromechanical accounting machine designed for use in small banks capable of storing the records of up to 10 000 accounts. By the 1930’s printing tabulators were available which worked at approximately 100 cards per minute, and there were sorters which worked at 400 cards per minute.The machines were controlled by fairly intricate plugboards, but arithmetic and logical computations involving sequences of operations of any great complexity were carried out by repeated processing of sets of cards, under the direction of operators. Various attempts were made to supplement the functional capabilities of punched card systems by linking together otherwise independent machines. One such system, the Synchro-Madas machine, incorporated a typewriter/accounting machine, an automatic calculating machine and an automatic card punch.These were linked together so that a single action by the operator sitting at the typewriter/accounting machine would control several opera tions on the different machines. One other system involving a set of inter-linked card machines, although very different in concept and scale from the Synchro-Madas machine, is worth mentioning. This is the Remote-control Accounting system which was experimented with in a Pittsburgh department store, also in the mid-1930’s. The system involved 250 terminals connected by telephone lines to 20 Powers card punch/tabulators and 15 on-line typewriters.The terminals transmitted data from punched merchandise tags which were used to produce punched sales record cards, later used for customer billing. The typewriter terminals were used for credit authorisation purposes. The intended peak transaction rate was 9000 per hour. Even during the 1920’s punched card systems were used not only for accounting and the compilation of statistics, but also for complex statistical calculations. However the first important scientific application of punched card systems was made by L.J. Comrie in 1929. Comrie was Superintendent of HM Nautical Almanac Office until 1936, and then founded the Scientific Computing Service. He made a speciality of putting commercial computing machinery to scientific use, and introduced Hollerith equipment to the Nautical Almanac Office. His calculations of the future positions of the Moon, which involved the punching of half a million cards, stimulated many other scientists to exploit the possibilities of punched card systems. One such scientist was Wallace J.Eckert, an astronomer at Columbia University, which already had been donated machines for a Statistical Laboratory by IBM in 1929, including the â€Å"Statistical Calculator,† a specially developed tabulator which was the forerunner of the IBM Type 600 series of multiplying punches, and of the mechanisms used in the Harvard Mark I machine. With assistance from IBM in 1934 Eckert set up a scientific computing laboratory in the Columbia Astronomy Department, a laboratory which was la ter to become the Thomas J.Watson Astronomical Computing Bureau. In order to facilitate the use of his punched card equipment Eckert developed a centralised control mechanism, linked to a numerical tabulator, a summary punch and a multiplying punch, so that a short cycle of different operations could be performed at high speed. The control mechanism which was based on a stepping switch enabled many calculations, even some solutions 4 The Institute of Mathematics and its Applications of differential equations, to be performed completely automatically.The potential of a system of inter-connected punched card machines, controlled by a fully general-purpose sequencing mechanism, and the essential similarity of such a system to Babbage’s plans for an Analytical Engine, were discussed in an article published by Vannevar Bush in 1936. Bush was at this time already renowned for his work on the first differential analyser, and during the war held the influential position of Director o f the US Office of Scientific Research and Development.In fact an attempt was made to build such a system of inter-connected punched card machines at the Institut fur Praktische Mathematik of the Technische Hochschule, Darmstadt, in Germany during the war. The plans called for the inter-connection of a standard Hollerith multiplier and tabulators, and specially constructed divider and function generators, using a punched tape sequence control mechanism. Work was abandoned on the project following a destructive air raid in September 1944. However, by this stage, in the United States much more ambitious efforts were being made to apply the expertise of punched card equipment designers.The efforts originated in 1937 with a proposal by Howard Aiken of Harvard University that a large-scale scientific calculator be constructed by inter-connecting a set of punched card machines via a master control panel. This would be plugged so as to govern the transmission of numerical operands and the sequencing of arithmetic operations. Through Dr. Shapley, director of the Harvard College Observatory, Aiken became acquainted with Wallace Eckert’s punched card installation at Columbia University.These contacts helped Aiken to persuade IBM to undertake the task of developing and building a machine to his basic design. For IBM, J. W. Bryce assigned C. D. Lake, F. E. Hamilton and B. M. Durfee to the task. Aiken later acknowledged these three engineers as co-inventors of the Automatic Sequence Controlled Calculator, or Harvard Mark I as it became known. The machine was built at the IBM development laboratories at Endicott and was demonstrated there in January 1943 before being shipped to Harvard, where it became operational in May 1944.In August of that year IBM, in the person of Thomas J. Watson, donated the machine to Harvard where it was used initially for classified work for the US Navy. The design of the Harvard Mark I followed the original proposals by Aiken fairly close ly, but it was built using a large number of the major components used in the various types of punched card machines then manufactured, rather than from a set of complete machines themselves. It incorporated 72 â€Å"storage counters† each of which served as both a storage location, and as a complete adding and subtracting machine.Each counter consisted of 24 electromechanical counter wheels and could store a signed 23digit decimal number. A special multiply/divide unit, and units for obtaining the value of previously computed functions held on perforated tape, and for performing interpolation, were provided together with input/output equipment such as card readers and punches, and typewriters. The various mechanisms and counter wheels were all driven and synchronised by a single gearconnected mechanical system extending along nearly the entire length of the calculator.A main sequence control mechanism incorporating a punched tape reader governed the operation of the machine. Each horizontal row on the tape had space for three groups of eight holes, known as the A, B and C groups. Together these specified a single instruction of the form â€Å"Take the number out of unit A, deliver it to unit B, and start operation C. † Somewhat surprisingly, in view of Aiken’s knowledge of Babbage’s work and writings, no provision was made originally for conditional branching.As it was, such provision was only made later when a subsidiary sequence control mechanism was built at Harvard and incorporated into the machine. The Harvard Mark I was a massive machine over 50 feet long, built on a lavish scale. Being largely mechanical its speed was somewhat limited – for example multiplication took 6 seconds – but it continued in active use at Harvard until 1959. It has an important place in the history of computers although the long-held belief that it was the world’s first operational programcontrolled computer was proved to be fals e, once the details of Zuse’s wartime work in Germany became known.It marked a major step by IBM towards full involvement in the design of general-purpose computers and, with ENIAC and the Bell Telephone Laboratories Series, represents the starting point of American computer developments. After completion of the Mark I, Aiken and IBM pursued independent paths. Aiken, still distrustful of the reliability of electronic components, moved to electromagnetic relays for the construction of the Harvard Mark II, another paper-tape-sequenced calculator.This machine had an internal store which could hold about 100 dccimal floating point numbers. One of the most interesting aspects of the machine was that it could be operated either as a single computer or as two separate ones. The complete system incorporated four of each type of input/output device, namely sequence tape readers, data tape readers and punches, numerical function tape readers and output printers. It also had multiple ar ithmetic facilities, including two adders and four multipliers (taking 0. 7 second) which could all be used simultaneously.Detailed design of the machine, which was intended for the US Naval Proving Ground, Dahlgren, Virginia, began at Harvard early in 1945, and the machine was completed in 1947. Afterwards Aiken and his colleagues went on to design the Mark III, an electronic computer with magnetic drum storage, completed in 1950, and the Mark IV, which incorporated 200 magnetic core shift registers, completed in 1952. The designers of IBM’s next machine, the Pluggable Sequence Relay Calculator, included two of the Harvard Mark I’s design team, namely C. D. Lake and B. M.Durfee, but the machine in fact had more in common with IBM’s earlier calculating punches than with the Mark I; like the punches it was controlled using plugboard-specified sequencing, rather than by a sequence control tape of essentially unlimited length. Its relay construction resulted in its basic operation speed being considerably faster than the Mark I, although it lacked the Mark I’s ease and flexibility of programming, demanding instead the kind of detailed design of parallel subsequencing that one sees nowadays at the microprogramming level of some computers.Great stress was raid by the designers on the efficient use of punched card input/output, and it was claimed that in many cases, where other machines’ internal storage capacity proved inadequate, the IBM relay calculators could outperform even the contemporary electronic computers. Several machines were built, the first of which was delivered in December 1944 to the Aberdeen Proving Ground, and two were installed at the Watson Scientific Computing Laboratory that IBM had set up at Columbia University under the directorship of Wallace Eckert.The Relay Calculator was followed by the giant IBM Selective Sequence Electronic Calculator, a machine which was very much in the tradition of the Mark I. Wal lace Eckert was responsible for the logical organisation of the machine, with Frank Hamilton being the chief engineer on the project. The design was a compromise between Eckert’s wish, for performance reasons, to use electronic components to the full, and Hamilton’s preference for electro-mechanical relays, on grounds of reliability. As a result vacuum tubes were used for the arithmetic unit, the control circuitry, and the 8 word high-speed store, relays being used elsewhere.In addition to the 8 word store there was a 150 word random access electro-magnetic store and storage for 20000 numbers in the form of punched tapes. Numbers would be read from the electro-magnetic store, or in sequence from the punched tape store, at the speed of the multiplier, i. e. , every 20 milliseconds. The design was started in 1945, and the machine was built in great secrecy at Endicott, before being moved to New York City, where it was publicly unveiled at an elaborate dedication ceremony in January 1948. The most important aspect of the SSEC, credited to R. R.Seeber, was that it could perform arithmetic on, and then execute, stored instructions – it was almost certainly the first operational machine with these capabilities. This led to IBM obtaining some very important patents, but the machine as a whole was soon regarded as somewhat anachronistic and was dismantled in 1952. It had however provided IBM with some valuable experience – for example, Hamilton and some of his engineering colleagues went on to design the highly successful IBM 650, and many of the SSEC programmers later became members of the IBM 701 programming group.Finally, mention should be made of one other machine manufactured by IBM which can be classed as a precursor to the modern electronic digital computer. This was the Card Programmed Calculator, a machine which along with its predecessors now tends to be overshadowed by the SSEC. Like the Pluggable Sequence Relay Calculator, the C PC can trace its origins to the IBM 600 series of multiplying punches. In 1946 IBM announced the Type 603, the first production electronic calculator. The IBM 603, which incorporated 300 valves, was developed from an experimental multiplier designed at Endicott under the direction of R.L. Palmer in 1942. One hundred machines were sold, and then IBM replaced it with the Type 604, a plugboardcontrolled electronic calculator, which provided conditional branching but, lacking backward jumps, no means of constructing program loops. Deliveries of the 604, which incorporated over 1400 valves, started in 1948 and within the next 10 years over 5000 were installed. In 1948 a 604 was coupled to a type 402 accounting machine by Northrop Aircraft Company, in order to provide the 604 with increased capacity and with printing facilities. This idea was taken up by IBM, and formed the basis of the CPC.Nearly 700 CPC’s were built, and this machine played a vital role in providing computing pow er to many installations in the USA until stored program electronic computers became commercially available on a reasonable scale. In the years that followed the introduction of the CPC, IBM continued to develop its range of electronic calculators and, starting in 1952 with the IBM 701, an electronic computer in the tradition of von Neumann’s IAS machine, took its first steps towards achieving its present dominant position amongst electronic computer manufacturers. . Konrad Zuse Konrad Zuse started to work on the development of mechanical aids to calculation as early as 1934, at the age of 24. He was studying civil engineering at the Technische Hochschule, Berlin-Charlottenburg, and sought some means of relief from the tedious calculations that had to be performed. His first idea had been to design special forms to facilitate ordinary manual calculation, but then he decided to try to mechanise the operation.Continuing to use the special layouts that he had designed for his fo rms, he investigated representing numerical data by means of perforations, and the use of a hand-held sensing device which could communicate the data over an electrical cable to an automatic calculating machine. The idea then arose of using a mechanical register rather than perforated cards, and, realising that the layout was irrelevant, Zuse started to develop a general purpose mechanical store, whose locations were addressed numerically.By 1936 he had the basic design of a floating point binary computer, controlled by a program tape consisting of a sequence of instructions, each of which specified an operation code, two operand addresses and a result address. Thus, apparently quite independently of earlier work by Babbage and his successors on analytical engines, Zuse had very quickly reached the point of having a design for a general-purpose program-controlled computer, although the idea of conditional branching was lacking.More importantly, even though the various basic The Inst itute of Mathematics and its Applications 6 ideas that his design incorporated had, it now turns out, been thought of earlier (i. e. , binary mechanical arithmetic (Leibniz), program control (Babbage), instruction formats with numerical storage addresses (Ludgate) and floating point number representations (Torres y Quevedo)), Zuse’s great achievement was to turn these ideas into reality. Zuse had considerable trouble finding sponsors willing to finance the building of his machine.Despite his financial difficulties his first machine, the Z1, which was of entirely mechanical construction was completed in 1938, but it proved unreliable in operation. He then started to construct a second, fixed-point binary, machine which incorporated the 16 word mechanical binary store of the Z1, but was otherwise built from second-hand telephone relays. Although the Z2 computer was completed it was inadequate for any practical use. However by this time a colleague, Helmut Schreyer, was already working with Zuse on the problem of producing an electronic version of the Z1.This led to the construction of a small 10 place binary arithmetic unit, with approximately 100 valves, but proposals that Schreyer and Zuse made to the German government for a 1500 valve electronic computer were rejected and the work was discontinued in 1942. Earlier, in 1939, Zuse was called up for military service, but managed to get released after about a year, and for the first time received significant government backing for his plans. This enabled him to build the Z3 computer, a binary machine with a 64 word store, all built out of telephone relays.This computer, since it was operational in 1941, is believed to have been the world’s first general-purpose program-controlled computer. It incorporated units for addition, subtraction, multiplication, division and square root, using a floating point number representation with a sign bit, a 7-bit exponent and a 14-bit mantissa. Input was via a manu al keyboard and output via a set of lights, in each case with automatic binary/decimal conversion, and the machine was controlled by a perforated tape carrying single address instructions, i. . , instructions specifying one operand, and an operation. In addition to his series of general-purpose computers, Zuse built two special-purpose computers, both used for calculations concerning aircraft wing profiles. The first of these was in use for 2 years at the Henschel Aircraft Works, before being destroyed through war damage. Both computers had fixed programs, wired on to rotary switches, and performed calculations involving addition, subtraction and multiplication by constant factors.Soon after completion of the Z3, the design of an improved version, the Z4, was started. This was mainly electro-mechanical but incorporated a purely mechanical binary store similar to that which had been used for the Zl and Z2 machines. The partially completed Z4 was the only one of Zuse’s machines to survive the war – indeed it eventually was completed and gave years of successful service at the Technische Hochschule, Zurich. The Z4 was inspected shortly after the war by R. C. Lyndon, whose report on the machine for the US Office f Naval Research was published in 1947. At this stage the Z4 had only manual input and output, and no means of conditional branching, although it was planned to add four tape readers and two tape punches, and facilities for repeating programs and for choosing between alternate subprograms. The machine was housed in the cellar of a farmhouse in the little village of Hopferau in Bavaria, and was not fully operational, but the mechanical store and various arithmetic operations and their automatic sequencing were successfully demonstrated to Lyndon.His report, although it gives a fairly full description of the Z4 (with the exception of the mechanical store, which he was not allowed to examine in detail), made virtually no mention of Zuse’s earlier work. Indeed it was many years before any other English language accounts of Zuse’s work were published, and Zuse’s rightful place in the chronology of computer development became at all widely appreciated. 5. Bell Telephone Laboratories The potentialities of telephone equipment for the construction of digital calculation devices were not realised for many years.The first automatic telephone exchange, which used the step-by-step or Strowger switch, was installed in 1892. As early as 1906 Molina devised a system for translating the pulses representing the dialled decimal digits into a more convenient number system. Exchanges based mainly on the use of electromechanical relays started to come into use at the turn of the century, the earliest successful centralised automatic exchanges dating from about 1914. However, from the late 1920’s various different calculating devices were developed using telephone equipment.Perhaps the most spectacular of these was the automatic totalisator. Totalisator, or â€Å"pari-mutuel,† betting became legal on British race courses in July 1929. Development of fully automatic totalisators consisting of ticket-issuing machines situated in various parts of the race course, a central calculating apparatus, and display boards which indicated the number and total value of bets made on each horse, and on the race as a whole, was already well under way.There were several rival systems. The Hamilton Totalisator and the totalisator produced by the British Automatic Totalisator Company were fully electrical, both as regards the calculations performed and the operation of the display boards, whereas the Lightning Totalisator used electrical impulses from remote ticket machines only to release steel balls which fell through tubes and actuated a mechanical adding apparatus.In January 1930 the Racecourse Betting Control Board demonstrated at Thirsk Racecourse a new standard electric totalisator supplied by Bri tish Thompson Houston, built from Strowger switches. This machine which was transportable from racecourse to racecourse could accumulate bets on up to six horses at a maximum rate of 12 000 per minute. The machine had in fact been designed in Baltimore, Maryland, in 1928 but the first complete machine to be used in the USA was installed by the American Totalisator Company at Arlington Park nly in 1933. In succeeding years much more sophisticated totalisators, involving hundreds of remote ticket-issuing machines, were used at racecourses all over USA, and it was not until many years after the advent of the electronic computer that one was used as a replacement for the central calculating apparatus of the totalisator. One early little-known design for a calculating machine to be built from telephone relays was that of Bernard Weiner in Czechoslovakia in 1923.Weiner, in association with the Vitkovice Iron Works, went on during the 1930’s to design a more powerful automatic calcu lator. He did not survive the war, and nothing is known about the results of this work. Other early work was done by Nicoladze who in 1928 designed a multiplier based on the principle of Genaille’s rods. (These were a non-mechanical aid to multiplication which enabled a person to read off the product of a multidigit number by a single digit number. Four years later Hamann described not only various different styles of relay-based multiplier, but also a device for solving sets of simultaneous linear equations, and shortly afterwards Weygandt demonstrated a prototype determinant evaluator, capable of dealing with 3 x 3 determinants. Undoubtedly in the years that followed many other digital calculating devices were developed based on telephone relay equipment, particularly during the war for such military applications as ballistics calculations and cryptanalysis – indeed, as mentioned earlier, some of Zuse’s machines made extensive use of telephone relays.It is per haps a little surprising that it was not until 1937 that Bell Telephone Laboratories investigated the design of calculating devices, although from about 1925 the possibility of using relay circuit techniques for such purposes was well accepted there. However, in 1937 George Stibitz started to experiment with relays, and drew up circuit designs for addition, multiplication and division. At first he concentrated on binary arithmetic, together with automatic decimal-binary and binarydecimal conversion, but later turned his attention to a binary-coded decimal number representation.The project became an official one when, prompted by T. C. Fry, Stibitz started to design a calculator capable of multiplying and dividing complex numbers, which was intended to fill a very practical need, namely to facilitate the solution of problems in the design of filter networks, and so started the very important Bell Telephone Laboratories Series of Relay Computers. In November 1938, S. B. Williams took over responsibility for the machine’s development and together with Stibitz refined the design of the calculator, whose construction was started in April and completed in October of 1939.The calculator, which became known as the â€Å"Complex Number Computer† (often shortened to â€Å"Complex Computer,† and as other calculators were built, the â€Å"Model I†), began routine operation in January 1940. Within a short time it was modified so as to provide facilities for the addition and subtraction of complex numbers, and was provided with a second, and then a third, teletype control, situated in remote locations. It remained in daily use at Bell Laboratories until 1949.The Complex Computer was publicly demonstrated for the first time in September 1940 by being operated in its New York City location from a teletypewriter installed in Hanover, New Hampshire, on the occasion of a meeting of the American Mathematical Society, a demonstration that both John Mauc hly and Norbert Wiener attended. During 1939 and 1940 Stibitz started work on the idea of automatic sequencing and on the use of error-detecting codes. These ideas were not pursued actively until, a year or so later, the onset of the war rovided a strong stimulus and the necessary financial climate. They then formed the basis of the second of the Bell Laboratories relay calculators, the â€Å"Relay Interpolator. † This was a special-purpose tape-controlled device, with selfchecking arithmetic, designed to solve fire control problems, and was built for the National Defense Research Council, to which Stibitz had been lent by Bell Laboratories. Although mainly used for interpolation it was also used for a few problems in harmonic analysis, calculation of roots of polynomials and solution of differential equations.It became operational in September 1943, and after the war it was handed over to the US Naval Research Laboratory, where it was in use until 1961. The Model III relay c alculator, the â€Å"Ballistic Computer,† work on which started in 1942, was a much more complete realisation of Stibitz’s early plans for an automatic computer, and although once again intended for fire control problems was much more versatile than the Model II. It was tape-controlled, and had a tenregister store, a built-in multiplier (designed by E. L.Vibbard), and devices for performing automatic look-up of tables held on perforated paper tape. Perhaps most impressive was the fact that the machine was 100 per cent. self-checked. The machine was completed in June 1944, and remained in use until 1958. The Model IV relay calculator was little different from the Model III, and the series culminated in the Model V, a truly general-purpose program-controlled computer, complete with convenient conditional branching facilities. (The final member of the series, Model VI, was essentially just a simplified version of the Model V. Two copies of the Model V were built, the firs t being delivered in 1946 to the National Advisory Committee on Aeronautics at Langley Field, Virginia, and the second in 1947 to the Ballistics Research Laboratory at Aberdeen, Maryland. With its multiple computing units, the Model V, which used floating point arithmetic, was what we would now call a multiprocessing system, and its â€Å"problem tapes† were the forerunners of the early simple batch-processing operating systems. Each of the two computing units comprising a complete system contained 15 storage registers.A single register could hold a floating point number consisting of a sign, a seven-decimal digit mantissa and a two-digit exponent. Decimal digits were stored in a bi-quinary form, using seven relays, and each register used a total of 62 relays. Each unit had independent provision for the addition, subtraction, multiplication and division and for 8 The Institute of Mathematics and its Applications taking the square root of floating point numbers, and for printi ng or punching its results.In addition a large set of tape readers, intended for tapes of input data, tabulated functions and programs, and for the problem tapes which controlled the running of series of separate programs, were shared by the two computer units. These units normally functioned as independent computers, but for large problems would be arranged to work cooperatively. Although somewhat slow in execution, the Model V set new standards for reliability, versatility and ease of switching from one task to another, and in so doing must surely have had an important influence on the designers of the earliest round of general-purpose electronic computers.In later years, quite a number of relay calculators were constructed, in both the USA and Europe, even after the first stored program electronic computers became operational, but the importance of their role in the history of computers hardly matches that of the Bell Laboratories Model V and its contemporaries. 6. The advent of electronic computers The earliest known electronic digital circuit, a â€Å"trigger relay,† which involved a pair of valves in a circuit with two stable states and was an early form of flip-flop, was described by Eccles and Jordan in 1919.The next development that we know of was the use by WynnWilliams at the Cavendish Laboratory, Cambridge, of thyratrons in counting circuits including, in 1932, a â€Å"scale-of-two† (binary) counter. By the end of the decade quite a few papers had been published on electronic counters intended for counting impulses from GeigerMuller tubes used in nuclear physics experiments. WynnWilliams’ work had a direct influence on the ideas of William Phillips, who apparently in 1935 attempted to patent a binary electronic computing machine.He built a mechanical model, which still exists, of the intended electronic multiplication unit but no other details are presently known of his planned machine. The first known attempt to build an elect ronic digital calculating machine was begun by John V. Atanasoff in the mid-1930’s at Iowa State College where there had been an active interest in statistical applications using punched card equipment since the early 1920’s. As an applied mathematician Atanasoff had many problems requiring generalisations of existing methods of approximating solutions of linear operational equations.He first explored the use of analog techniques and with Lynn Hannum, one of his graduate students, developed the â€Å"Laplaciometer,† a device for solving Laplace’s equation in two dimensions with various boundary conditions. By 1935 the realisation of the sharp limitations of analog computing forced Atanasoff to digital methods. The disadvantages of mechanical techniques and his knowledge of electronics and of the work of Eccles and Jordan then led him to consider an electronic approach.He soon found that in these circumstances a base two number system would have great adva ntages. In 19361937 Atanasoff abandoned the Eccles-Jordan approach and conceived a system employing memory and logic circuits, whose details were worked out in 1938. He received a grant from Iowa State in 1939, and was joined by Clifford E. Berry. With Berry’s assistance a prototype computing element was built and operating by the autumn of that year. They then undertook the design and construction of a large machine intended for the solution of up to 30 simultaneous linear equations.At the heart of the machine there was a pair of rotating cylinders around the surface of which a set of small electrical condensers was placed. Each condenser could, by the direction of its charge, represent a binary digit; although the charge would leak away slowly, it was arranged that as the cylinders rotated the charge on each condenser was detected and reinforced at 1 second time intervals so that information could be stored for as long as required.The condensers were arranged so as to provi de two sets of 30 binary words, each consisting of 50 bits, the condensers corresponding to a single word being arranged in a plane perpendicular to the axis of the cylinders. The results of intermediate steps of a computation were to be punched in binary form on cards, for later re-input to the machine. In order that card punching and reading should be fast enough to keep pace with the computation, special devices were designed that made and detected holes in cards by means of electrical sparks.Ordinary input and output was to be via conventional punched cards, with the machine providing automatic binary/decimal conversions. The machine, with binary addition, subtraction and shifting as its basic arithmetic facilities, was designed to solve sets of simultaneous linear equations by the method of successive elimination of unknowns. The electronic part of the computer was operational but the binary card reader was still unreliable when in 1942 Atanasoff and Bcrry left Iowa State for w artime jobs, so that the machine was abandoned, never having seen actual use.In the late 1930’s and early 1940’s several groups started to investigate the use of digital electronic circuits as replacements for mechanical or electro-mechanical calculating devices, including several of the American business machine manufacturers such as IBM, whose work was described briefly above. The earliest known efforts at applying electronics to a general-purpose program-controlled computer were those undertaken by Schreyer and Zuse, also mentioned earlier.The next development which should be mentioned is the still classified series of electronic cryptanalytic machines that were designed and built in Britain during the war. The machines that are of particular interest, with respect to the development of electronic computers are the Colossi, the first of which was operational in late 1943, while by the end of the war ten had been installed. Each Colossus incorporated approximately 20 00 valves, and processed a punched data tape that was read at a speed of 5000 characters per second.Preset patterns that were to be compared against the input data were generated from stored component patterns. These components were stored in ring registers made of thyratrons and could be manually set by plug-in pins. The Colossi were developed by a team led by M. H. A. Newman. Alan Turing, who had been one of the main people involved in the design of an electro-mechanical predecessor to the Colossi, was apparently not directly associated with the new design, but with others provided the requirements that the machines were to satisfy.The comparative lack of technical details about the design of these machines makes it unreasonable to attempt more than a preliminary, and somewhat hesitant, assessment of the Colossi with respect to the modern digital computer. It would appear that the arithmetical, as opposed to logical, capabilities were minimal, involving only counting rather than g eneral addition or other operations. They did, however, have a certain amount of electronic storage. Although fully automatic, even to the extent of producing printed output, they were very much special-purpose machines, but ithin their field of specialisation the facilities provided by plug-boards and banks of switches afforded a considerable degree of flexibility; in fact several of the people involved in the project have since characterised the machines as being â€Å"program-controlled. † Their importance as cryptanalytic machines, which must have been immense, can only be inferred from the number of machines that were made and the honours bestowed on various members of the team after the end of the war; however, their importance with respect to the development of computers was twofold.They demonstrated the practicality of largescale electronic digital equipment, just as ENIAC did, on an even grander scale, approximately 2 years later. Furthermore, they were also a major source of the designers of some of the first post-war British computers, namely the Manchester machine, the MOSAIC, and the ACE at the National Physical Laboratory. Fascinating though they are, none of the efforts described so far comes near to matching the importance of the work at the Moore School of Electrical Engineering, University of Pennsylvania, which led to the design of first the ENIAC and then the EDVAC computers.By 1942 the Moore School had, because of pressures of war, become closely associated with the Ballistic Research Laboratory of the US Army Ordnance Department, and the Moore School’s differential analyser was being used to supplement the work of the one at the Ballistic Research Laboratory on the production of ballistic tables. (The two analysers were identical and had been patterned on the original differential analyser invented by Vannevar Bush in 1930. ) One of the people who had worked with the analyser was John Mauchly, then an assistant professor at the Moore School.Mauchly was by this time well aware of what could be done with desk calculating machines and punched card equipment, although he was apparently unaware of the work Aiken was then doing on what became the Harvard Mark I, or of Babbage’s efforts 100 years earlier. He did however know of the work of Stibitz and had visited Iowa State in June 1941 in order to see Atanasoff’s special-purpose computer. Another person who worked on the Moore School differential analyser, and in fact made important improvements to it by replacing its mechanical amplifiers by partially electronic devices, was J. Presper Eckert, a research associate at the School.Eckert had met Mauchly in 1941, and it was their discussions about the possibility of surmounting the reliability problems of complex electronic devices that laid the groundwork for a memorandum that Mauchly wrote in August 1942. This proposed that an electronic digital computer be constructed for the purpose of solving numerical difference equations of the sort encountered in ballistics problems. Also at the Moore School, acting as a liaison officer for Colonel Paul N. Gillon of the office of the Chief of Ordnance, was Herman H. Goldstine, who before the war had been assistant professor of mathematics at the University of Michigan.In early 1943 Goldstine and Gillon became interested in the possibility of using an electronic calculating machine for the preparation of firing and bombing tables. By this time Mauchly’s 1942 memorandum had been mislaid, and it had to be recreated from his secretary’s notes. The second version of the memorandum, together with more detailed plans drawn up by Mauchly and Eckert, was included in a report dated April 1943 which formed the basis for a contract between the University of Pennsylvania and the US Government to develop an electronic computer.A large team was assembled at the Moore School in order to design and build the computer under the supervisi on of J. G. Brainerd, with Eckert as chief engineer and Mauchly as principal consultant. As the project progressed its aims broadened, so that the ENIAC, as it became known, turned out to be much more a general-purpose device than had been originally contemplated, and although programs were represented by plugged interconnecting wires, it provided full conditional branching facilities.It was an incredibly ambitious machine incorporating over 19 000 valves and consuming approximately 200 kilowatts of electric power! (The number of valves largely resulted from the use of them for high speed storage, and the choice of number representation, which can best be described as â€Å"unary-coded decimal. â€Å") The ENIAC incorporated 20 10-digit accumulators, which could be used for addition and subtraction, and for the temporary storage of numbers, a multiplier and a combination divider and square rooter.Addition took 200 microseconds, and multiplication of two 10-digit numbers approximat ely 3 milliseconds. Storage was provided for approximately 300 numerical constants in function tables, which could be set up by manual switches prior to commencing a computation. Input and output was via punched cards, using standard IBM devices. Early in its career the method of programming the machine was modified so that the program was represented by settings of the function tables without the need for changing the interconnecting cables.