context
stringclasses
269 values
id_string
stringlengths
15
16
answers
sequencelengths
5
5
label
int64
0
4
question
stringlengths
34
417
With their recognition of Maxine Hong Kingston as a mdetailing typical talk-story formsajor literary figure, some critics have suggested that her works have been produced almost ex nihilo, saying that they lack a large traceable body of direct literary antecedents especially within the Chinese American heritage in which her work is embedded. But these critics, who have examined only the development of written texts, the most visible signs of a culture's narrative production, have overlooked Kingston's connection to the long Chinese tradition of a highly developed genre of song and spoken narrative known as "talk-story" (gong gu tsai). Traditionally performed in the dialects of various ethnic enclaves, talk-story has been maintained within the confines of the family and has rarely surfaced into print. The tradition dates back to Sung dynasty (A.D. 970–1279) storytellers in China, and in the United States it is continually revitalized by an overlapping sequence of immigration from China. Thus, Chinese immigrants to the U.S. had a fully established, sophisticated oral culture, already ancient and capable of producing masterpieces, by the time they began arriving in the early nineteenth century. This transplanted oral heritage simply embraced new subject matter or new forms of Western discourse, as in the case of Kingston's adaptations written in English. Kingston herself believes that as a literary artist she is one in a long line of performers shaping a recalcitrant history into talk-story form. She distinguishes her "thematic" storytelling memory processes, which sift and reconstruct the essential elements of personally remembered stories, from the memory processes of a print-oriented culture that emphasizes the retention of precise sequences of words. Nor does the entry of print into the storytelling process substantially change her notion of the character of oral tradition. For Kingston, "writer" is synonymous with "singer" or "performer" in the ancient sense of privileged keeper, transmitter, and creator of stories whose current stage of development can be frozen in print, but which continue to grow both around and from that frozen text. Kingston's participation in the tradition of talk-story is evidenced in her book China Men, which utilizes forms typical of that genre and common to most oral cultures including: a fixed "grammar" of repetitive themes; a spectrum of stock characters; symmetrical structures, including balanced oppositions (verbal or physical contests, antithetical characters, dialectical discourse such as question-answer forms and riddles); and repetition. In China Men, Kingston also succeeds in investing idiomatic English with the allusive texture and oral-aural qualities of the Chinese language, a language rich in aural and visual puns, making her work a written form of talk-story.
200810_2-RC_3_21
[ "scholarly appreciation for its longstanding artistic sophistication", "mild disappointment that it has not distinguished itself from other oral traditions", "tentative approval of its resistance to critical evaluations", "clear respect for the diversity of its ancient sources and cultural derivations", "open admiration for the way it uses song to express narrative" ]
0
Which one of the following most accurately identifies the attitude shown by the author in the passage toward talk-story?
In economics, the term "speculative bubble" refers to a large upward move in an asset's price driven not by the asset's fundamentals—that is, by the earnings derivable from the asset—but rather by mere speculation that someone else will be willing to pay a higher price for it. The price increase is then followed by a dramatic decline in price, due to a loss in confidence that the price will continue to rise, and the "bubble" is said to have burst. According to Charles Mackay's classic nineteenth-century account, the seventeenth-century Dutch tulip market provides an example of a speculative bubble. But the economist Peter Garber challenges Mackay's view, arguing that there is no evidence that the Dutch tulip market really involved a speculative bubble. By the seventeenth century, the Netherlands had become a center of cultivation and development of new tulip varieties, and a market had developed in which rare varieties of bulbs sold at high prices. For example, a Semper Augustus bulb sold in 1625 for an amount of gold worth about U.S.$11,000 in 1999. Common bulb varieties, on the other hand, sold for very low prices. According to Mackay, by 1636 rapid price rises attracted speculators, and prices of many varieties surged upward from November 1636 through January 1637. Mackay further states that in February 1637 prices suddenly collapsed; bulbs could not be sold at 10 percent of their peak values. By 1739, the prices of all the most prized kinds of bulbs had fallen to no more than one two-hundredth of 1 percent of Semper Augustus's peak price. Garber acknowledges that bulb prices increased dramatically from 1636 to 1637 and eventually reached very low levels. But he argues that this episode should not be described as a speculative bubble, for the increase and eventual decline in bulb prices can be explained in terms of the fundamentals. Garber argues that a standard pricing pattern occurs for new varieties of flowers. When a particularly prized variety is developed, its original bulb sells for a high price. Thus, the dramatic rise in the price of some original tulip bulbs could have resulted as tulips in general, and certain varieties in particular, became fashionable. However, as the prized bulbs become more readily available through reproduction from the original bulb, their price falls rapidly; after less than 30 years, bulbs sell at reproduction cost. But this does not mean that the high prices of original bulbs are irrational, for earnings derivable from the millions of bulbs descendent from the original bulbs can be very high, even if each individual descendent bulb commands a very low price. Given that an original bulb can generate a reasonable return on investment even if the price of descendent bulbs decreases dramatically, a rapid rise and eventual fall of tulip bulb prices need not indicate a speculative bubble.
200810_2-RC_4_22
[ "The seventeenth-century Dutch tulip market is widely but mistakenly believed by economists to provide an example of a speculative bubble.", "Mackay did not accurately assess the earnings that could be derived from rare and expensive seventeenth-century Dutch tulip bulbs.", "A speculative bubble occurs whenever the price of an asset increases substantially followed by a rapid and dramatic decline.", "Garber argues that Mackay's classic account of the seventeenth-century Dutch tulip market as a speculative bubble is not supported by the evidence.", "A tulip bulb can generate a reasonable return on investment even if the price starts very high and decreases dramatically." ]
3
Which one of the following most accurately expresses the main point of the passage?
In economics, the term "speculative bubble" refers to a large upward move in an asset's price driven not by the asset's fundamentals—that is, by the earnings derivable from the asset—but rather by mere speculation that someone else will be willing to pay a higher price for it. The price increase is then followed by a dramatic decline in price, due to a loss in confidence that the price will continue to rise, and the "bubble" is said to have burst. According to Charles Mackay's classic nineteenth-century account, the seventeenth-century Dutch tulip market provides an example of a speculative bubble. But the economist Peter Garber challenges Mackay's view, arguing that there is no evidence that the Dutch tulip market really involved a speculative bubble. By the seventeenth century, the Netherlands had become a center of cultivation and development of new tulip varieties, and a market had developed in which rare varieties of bulbs sold at high prices. For example, a Semper Augustus bulb sold in 1625 for an amount of gold worth about U.S.$11,000 in 1999. Common bulb varieties, on the other hand, sold for very low prices. According to Mackay, by 1636 rapid price rises attracted speculators, and prices of many varieties surged upward from November 1636 through January 1637. Mackay further states that in February 1637 prices suddenly collapsed; bulbs could not be sold at 10 percent of their peak values. By 1739, the prices of all the most prized kinds of bulbs had fallen to no more than one two-hundredth of 1 percent of Semper Augustus's peak price. Garber acknowledges that bulb prices increased dramatically from 1636 to 1637 and eventually reached very low levels. But he argues that this episode should not be described as a speculative bubble, for the increase and eventual decline in bulb prices can be explained in terms of the fundamentals. Garber argues that a standard pricing pattern occurs for new varieties of flowers. When a particularly prized variety is developed, its original bulb sells for a high price. Thus, the dramatic rise in the price of some original tulip bulbs could have resulted as tulips in general, and certain varieties in particular, became fashionable. However, as the prized bulbs become more readily available through reproduction from the original bulb, their price falls rapidly; after less than 30 years, bulbs sell at reproduction cost. But this does not mean that the high prices of original bulbs are irrational, for earnings derivable from the millions of bulbs descendent from the original bulbs can be very high, even if each individual descendent bulb commands a very low price. Given that an original bulb can generate a reasonable return on investment even if the price of descendent bulbs decreases dramatically, a rapid rise and eventual fall of tulip bulb prices need not indicate a speculative bubble.
200810_2-RC_4_23
[ "someone who, after learning that many others had withdrawn their applications for a particular job, applied for the job in the belief that there would be less competition for it", "an art dealer who, after paying a very high price for a new painting, sells it at a very low price because it is now considered to be an inferior work", "someone who, after buying a box of rare motorcycle parts at a very high price, is forced to sell them at a much lower price because of the sudden availability of cheap substitute parts", "a publisher who pays an extremely high price for a new novel only to sell copies at a price affordable to nearly everyone", "an airline that, after selling most of the tickets for seats on a plane at a very high price, must sell the remaining tickets at a very low price" ]
3
Given Garber's account of the seventeenth-century Dutch tulip market, which one of the following is most analogous to someone who bought a tulip bulb of a certain variety in that market at a very high price, only to sell a bulb of that variety at a much lower price?
In economics, the term "speculative bubble" refers to a large upward move in an asset's price driven not by the asset's fundamentals—that is, by the earnings derivable from the asset—but rather by mere speculation that someone else will be willing to pay a higher price for it. The price increase is then followed by a dramatic decline in price, due to a loss in confidence that the price will continue to rise, and the "bubble" is said to have burst. According to Charles Mackay's classic nineteenth-century account, the seventeenth-century Dutch tulip market provides an example of a speculative bubble. But the economist Peter Garber challenges Mackay's view, arguing that there is no evidence that the Dutch tulip market really involved a speculative bubble. By the seventeenth century, the Netherlands had become a center of cultivation and development of new tulip varieties, and a market had developed in which rare varieties of bulbs sold at high prices. For example, a Semper Augustus bulb sold in 1625 for an amount of gold worth about U.S.$11,000 in 1999. Common bulb varieties, on the other hand, sold for very low prices. According to Mackay, by 1636 rapid price rises attracted speculators, and prices of many varieties surged upward from November 1636 through January 1637. Mackay further states that in February 1637 prices suddenly collapsed; bulbs could not be sold at 10 percent of their peak values. By 1739, the prices of all the most prized kinds of bulbs had fallen to no more than one two-hundredth of 1 percent of Semper Augustus's peak price. Garber acknowledges that bulb prices increased dramatically from 1636 to 1637 and eventually reached very low levels. But he argues that this episode should not be described as a speculative bubble, for the increase and eventual decline in bulb prices can be explained in terms of the fundamentals. Garber argues that a standard pricing pattern occurs for new varieties of flowers. When a particularly prized variety is developed, its original bulb sells for a high price. Thus, the dramatic rise in the price of some original tulip bulbs could have resulted as tulips in general, and certain varieties in particular, became fashionable. However, as the prized bulbs become more readily available through reproduction from the original bulb, their price falls rapidly; after less than 30 years, bulbs sell at reproduction cost. But this does not mean that the high prices of original bulbs are irrational, for earnings derivable from the millions of bulbs descendent from the original bulbs can be very high, even if each individual descendent bulb commands a very low price. Given that an original bulb can generate a reasonable return on investment even if the price of descendent bulbs decreases dramatically, a rapid rise and eventual fall of tulip bulb prices need not indicate a speculative bubble.
200810_2-RC_4_24
[ "If speculative bubbles occur at all, they occur very rarely.", "Many of the owners of high-priced original tulip bulbs could have expected to at least recoup their original investments from sales of the many bulbs propagated from the original bulbs.", "If there is not a speculative bubble in a market, then the level of prices in that market is not irrational.", "Most people who invested in Dutch tulip bulbs in the seventeenth century were generally rational in all their investments.", "Mackay mistakenly infers from the fact that tulip prices dropped rapidly that the very low prices that the bulbs eventually sold for were irrational." ]
1
The passage most strongly supports the inference that Garber would agree with which one of the following statements?
In economics, the term "speculative bubble" refers to a large upward move in an asset's price driven not by the asset's fundamentals—that is, by the earnings derivable from the asset—but rather by mere speculation that someone else will be willing to pay a higher price for it. The price increase is then followed by a dramatic decline in price, due to a loss in confidence that the price will continue to rise, and the "bubble" is said to have burst. According to Charles Mackay's classic nineteenth-century account, the seventeenth-century Dutch tulip market provides an example of a speculative bubble. But the economist Peter Garber challenges Mackay's view, arguing that there is no evidence that the Dutch tulip market really involved a speculative bubble. By the seventeenth century, the Netherlands had become a center of cultivation and development of new tulip varieties, and a market had developed in which rare varieties of bulbs sold at high prices. For example, a Semper Augustus bulb sold in 1625 for an amount of gold worth about U.S.$11,000 in 1999. Common bulb varieties, on the other hand, sold for very low prices. According to Mackay, by 1636 rapid price rises attracted speculators, and prices of many varieties surged upward from November 1636 through January 1637. Mackay further states that in February 1637 prices suddenly collapsed; bulbs could not be sold at 10 percent of their peak values. By 1739, the prices of all the most prized kinds of bulbs had fallen to no more than one two-hundredth of 1 percent of Semper Augustus's peak price. Garber acknowledges that bulb prices increased dramatically from 1636 to 1637 and eventually reached very low levels. But he argues that this episode should not be described as a speculative bubble, for the increase and eventual decline in bulb prices can be explained in terms of the fundamentals. Garber argues that a standard pricing pattern occurs for new varieties of flowers. When a particularly prized variety is developed, its original bulb sells for a high price. Thus, the dramatic rise in the price of some original tulip bulbs could have resulted as tulips in general, and certain varieties in particular, became fashionable. However, as the prized bulbs become more readily available through reproduction from the original bulb, their price falls rapidly; after less than 30 years, bulbs sell at reproduction cost. But this does not mean that the high prices of original bulbs are irrational, for earnings derivable from the millions of bulbs descendent from the original bulbs can be very high, even if each individual descendent bulb commands a very low price. Given that an original bulb can generate a reasonable return on investment even if the price of descendent bulbs decreases dramatically, a rapid rise and eventual fall of tulip bulb prices need not indicate a speculative bubble.
200810_2-RC_4_25
[ "The rapid rise in price of Dutch tulip bulbs was not due to the fashionability of the flowers they produced.", "The prices of certain varieties of Dutch tulip bulbs during the seventeenth century were, at least for a time, determined by speculation.", "The Netherlands was the only center of cultivation and development of new tulip varieties in the seventeenth century.", "The very high prices of bulbs in the seventeenth-century Dutch tulip market were not irrational.", "Buyers of rare and very expensive Dutch tulip bulbs were ultimately able to derive earnings from bulbs descendent from the original bulbs." ]
1
The passage states that Mackay claimed which one of the following?
In economics, the term "speculative bubble" refers to a large upward move in an asset's price driven not by the asset's fundamentals—that is, by the earnings derivable from the asset—but rather by mere speculation that someone else will be willing to pay a higher price for it. The price increase is then followed by a dramatic decline in price, due to a loss in confidence that the price will continue to rise, and the "bubble" is said to have burst. According to Charles Mackay's classic nineteenth-century account, the seventeenth-century Dutch tulip market provides an example of a speculative bubble. But the economist Peter Garber challenges Mackay's view, arguing that there is no evidence that the Dutch tulip market really involved a speculative bubble. By the seventeenth century, the Netherlands had become a center of cultivation and development of new tulip varieties, and a market had developed in which rare varieties of bulbs sold at high prices. For example, a Semper Augustus bulb sold in 1625 for an amount of gold worth about U.S.$11,000 in 1999. Common bulb varieties, on the other hand, sold for very low prices. According to Mackay, by 1636 rapid price rises attracted speculators, and prices of many varieties surged upward from November 1636 through January 1637. Mackay further states that in February 1637 prices suddenly collapsed; bulbs could not be sold at 10 percent of their peak values. By 1739, the prices of all the most prized kinds of bulbs had fallen to no more than one two-hundredth of 1 percent of Semper Augustus's peak price. Garber acknowledges that bulb prices increased dramatically from 1636 to 1637 and eventually reached very low levels. But he argues that this episode should not be described as a speculative bubble, for the increase and eventual decline in bulb prices can be explained in terms of the fundamentals. Garber argues that a standard pricing pattern occurs for new varieties of flowers. When a particularly prized variety is developed, its original bulb sells for a high price. Thus, the dramatic rise in the price of some original tulip bulbs could have resulted as tulips in general, and certain varieties in particular, became fashionable. However, as the prized bulbs become more readily available through reproduction from the original bulb, their price falls rapidly; after less than 30 years, bulbs sell at reproduction cost. But this does not mean that the high prices of original bulbs are irrational, for earnings derivable from the millions of bulbs descendent from the original bulbs can be very high, even if each individual descendent bulb commands a very low price. Given that an original bulb can generate a reasonable return on investment even if the price of descendent bulbs decreases dramatically, a rapid rise and eventual fall of tulip bulb prices need not indicate a speculative bubble.
200810_2-RC_4_26
[ "present the facts that are accepted by all experts in the field", "identify the mistake that one scholar alleges another scholar made", "explain the basis on which one scholar makes an inference with which another scholar disagrees", "undermine the case that one scholar makes for the claim with which another scholar disagrees", "outline the factual errors that led one scholar to draw the inference that he drew" ]
2
The main purpose of the second paragraph is to
In economics, the term "speculative bubble" refers to a large upward move in an asset's price driven not by the asset's fundamentals—that is, by the earnings derivable from the asset—but rather by mere speculation that someone else will be willing to pay a higher price for it. The price increase is then followed by a dramatic decline in price, due to a loss in confidence that the price will continue to rise, and the "bubble" is said to have burst. According to Charles Mackay's classic nineteenth-century account, the seventeenth-century Dutch tulip market provides an example of a speculative bubble. But the economist Peter Garber challenges Mackay's view, arguing that there is no evidence that the Dutch tulip market really involved a speculative bubble. By the seventeenth century, the Netherlands had become a center of cultivation and development of new tulip varieties, and a market had developed in which rare varieties of bulbs sold at high prices. For example, a Semper Augustus bulb sold in 1625 for an amount of gold worth about U.S.$11,000 in 1999. Common bulb varieties, on the other hand, sold for very low prices. According to Mackay, by 1636 rapid price rises attracted speculators, and prices of many varieties surged upward from November 1636 through January 1637. Mackay further states that in February 1637 prices suddenly collapsed; bulbs could not be sold at 10 percent of their peak values. By 1739, the prices of all the most prized kinds of bulbs had fallen to no more than one two-hundredth of 1 percent of Semper Augustus's peak price. Garber acknowledges that bulb prices increased dramatically from 1636 to 1637 and eventually reached very low levels. But he argues that this episode should not be described as a speculative bubble, for the increase and eventual decline in bulb prices can be explained in terms of the fundamentals. Garber argues that a standard pricing pattern occurs for new varieties of flowers. When a particularly prized variety is developed, its original bulb sells for a high price. Thus, the dramatic rise in the price of some original tulip bulbs could have resulted as tulips in general, and certain varieties in particular, became fashionable. However, as the prized bulbs become more readily available through reproduction from the original bulb, their price falls rapidly; after less than 30 years, bulbs sell at reproduction cost. But this does not mean that the high prices of original bulbs are irrational, for earnings derivable from the millions of bulbs descendent from the original bulbs can be very high, even if each individual descendent bulb commands a very low price. Given that an original bulb can generate a reasonable return on investment even if the price of descendent bulbs decreases dramatically, a rapid rise and eventual fall of tulip bulb prices need not indicate a speculative bubble.
200810_2-RC_4_27
[ "against which other pricing patterns are to be measured", "that conforms to a commonly agreed-upon criterion", "that is merely acceptable", "that regularly recurs in certain types of cases", "that serves as an exemplar" ]
3
The phrase "standard pricing pattern" as used in line 38 most nearly means a pricing pattern
With his first published works in the 1950s, Amos Tutuola became the first Nigerian writer to receive wide international recognition. Written in a mix of standard English, idiomatic Nigerian English, and literal translation of his native language, Yoruba, Tutuola's works were quick to be praised by many literary critics as fresh, inventive approaches to the form of the novel. Others, however, dismissed his works as simple retellings of local tales, full of unwelcome liberties taken with the details of the well-known story lines. However, to estimate properly Tutuola's rightful position in world literature, it is essential to be clear about the genre in which he wrote; literary critics have assumed too facilely that he wrote novels. No matter how flexible a definition of the novel one uses, establishing a set of criteria that enable Tutuola's works to be described as such applies to his works a body of assumptions the works are not designed to satisfy. Tutuola is not a novelist but a teller of folktales. Many of his critics are right to suggest that Tutuola's subjects are not strikingly original, but it is important to bear in mind that whereas realism and originality are expected of the novel, the teller of folktales is expected to derive subjects and frameworks from the corpus of traditional lore. The most useful approach to Tutuola's works, then, is one that regards him as working within the African oral tradition. Within this tradition, a folktale is common property, an expression of a people's culture and social circumstances. The teller of folktales knows that the basic story is already known to most listeners and, equally, that the teller's reputation depends on the inventiveness with which the tale is modified and embellished, for what the audience anticipates is not an accurate retelling of the story but effective improvisation and delivery. Thus, within the framework of the basic story, the teller is allowed considerable room to maneuver—in fact, the most brilliant tellers of folktales transform them into unique works. Tutuola's adherence to this tradition is clear: specific episodes, for example, are often repeated for emphasis, and he embellishes familiar tales with personal interpretations or by transferring them to modern settings. The blend of English with local idiom and Yoruba grammatical constructs, in which adjectives and verbs are often interchangeable, re-creates the folktales in singular ways. And, perhaps most revealingly, in the majority of Tutuola's works, the traditional accents and techniques of the teller of folktales are clearly discernible, for example in the adoption of an omniscient, summarizing voice at the end of his narratives, a device that is generally recognized as being employed to conclude most folktales.
200812_4-RC_1_1
[ "Amos Tutuola is an internationally acclaimed writer of folktales whose unique writing style blends together aspects of Yoruba, Nigerian English, and standard English.", "Amos Tutuola's literary works should be evaluated not as novels but as unique and inventively crafted retellings of folktales.", "Amos Tutuola is an important author because he is able to incorporate the traditions of an oral art form into his novels.", "Critics are divided as to whether Amos Tutuola's literary works should be regarded as novels or folktales.", "The folktale is a valuable African literary genre that finds singular expression in the works of Amos Tutuola." ]
1
Which one of the following most accurately expresses the main point of the passage?
With his first published works in the 1950s, Amos Tutuola became the first Nigerian writer to receive wide international recognition. Written in a mix of standard English, idiomatic Nigerian English, and literal translation of his native language, Yoruba, Tutuola's works were quick to be praised by many literary critics as fresh, inventive approaches to the form of the novel. Others, however, dismissed his works as simple retellings of local tales, full of unwelcome liberties taken with the details of the well-known story lines. However, to estimate properly Tutuola's rightful position in world literature, it is essential to be clear about the genre in which he wrote; literary critics have assumed too facilely that he wrote novels. No matter how flexible a definition of the novel one uses, establishing a set of criteria that enable Tutuola's works to be described as such applies to his works a body of assumptions the works are not designed to satisfy. Tutuola is not a novelist but a teller of folktales. Many of his critics are right to suggest that Tutuola's subjects are not strikingly original, but it is important to bear in mind that whereas realism and originality are expected of the novel, the teller of folktales is expected to derive subjects and frameworks from the corpus of traditional lore. The most useful approach to Tutuola's works, then, is one that regards him as working within the African oral tradition. Within this tradition, a folktale is common property, an expression of a people's culture and social circumstances. The teller of folktales knows that the basic story is already known to most listeners and, equally, that the teller's reputation depends on the inventiveness with which the tale is modified and embellished, for what the audience anticipates is not an accurate retelling of the story but effective improvisation and delivery. Thus, within the framework of the basic story, the teller is allowed considerable room to maneuver—in fact, the most brilliant tellers of folktales transform them into unique works. Tutuola's adherence to this tradition is clear: specific episodes, for example, are often repeated for emphasis, and he embellishes familiar tales with personal interpretations or by transferring them to modern settings. The blend of English with local idiom and Yoruba grammatical constructs, in which adjectives and verbs are often interchangeable, re-creates the folktales in singular ways. And, perhaps most revealingly, in the majority of Tutuola's works, the traditional accents and techniques of the teller of folktales are clearly discernible, for example in the adoption of an omniscient, summarizing voice at the end of his narratives, a device that is generally recognized as being employed to conclude most folktales.
200812_4-RC_1_2
[ "applied conventions of the modern novel to the retelling of Irish folktales", "re-created important elements of the Irish literary style within a purely oral art form", "combined characters from English and Irish folktales to tell a story of modern life", "transplanted traditional Irish folktales from their original setting to contemporary Irish life", "utilized an omniscient narrator in telling original stories about contemporary Irish life" ]
3
Tutuola's approach to writing folktales would be most clearly exemplified by a modern-day Irish author who
With his first published works in the 1950s, Amos Tutuola became the first Nigerian writer to receive wide international recognition. Written in a mix of standard English, idiomatic Nigerian English, and literal translation of his native language, Yoruba, Tutuola's works were quick to be praised by many literary critics as fresh, inventive approaches to the form of the novel. Others, however, dismissed his works as simple retellings of local tales, full of unwelcome liberties taken with the details of the well-known story lines. However, to estimate properly Tutuola's rightful position in world literature, it is essential to be clear about the genre in which he wrote; literary critics have assumed too facilely that he wrote novels. No matter how flexible a definition of the novel one uses, establishing a set of criteria that enable Tutuola's works to be described as such applies to his works a body of assumptions the works are not designed to satisfy. Tutuola is not a novelist but a teller of folktales. Many of his critics are right to suggest that Tutuola's subjects are not strikingly original, but it is important to bear in mind that whereas realism and originality are expected of the novel, the teller of folktales is expected to derive subjects and frameworks from the corpus of traditional lore. The most useful approach to Tutuola's works, then, is one that regards him as working within the African oral tradition. Within this tradition, a folktale is common property, an expression of a people's culture and social circumstances. The teller of folktales knows that the basic story is already known to most listeners and, equally, that the teller's reputation depends on the inventiveness with which the tale is modified and embellished, for what the audience anticipates is not an accurate retelling of the story but effective improvisation and delivery. Thus, within the framework of the basic story, the teller is allowed considerable room to maneuver—in fact, the most brilliant tellers of folktales transform them into unique works. Tutuola's adherence to this tradition is clear: specific episodes, for example, are often repeated for emphasis, and he embellishes familiar tales with personal interpretations or by transferring them to modern settings. The blend of English with local idiom and Yoruba grammatical constructs, in which adjectives and verbs are often interchangeable, re-creates the folktales in singular ways. And, perhaps most revealingly, in the majority of Tutuola's works, the traditional accents and techniques of the teller of folktales are clearly discernible, for example in the adoption of an omniscient, summarizing voice at the end of his narratives, a device that is generally recognized as being employed to conclude most folktales.
200812_4-RC_1_3
[ "convinced that Tutuola's works should be viewed within the context of the African oral tradition", "certain that Tutuola's works will generate a renewed interest in the study of oral traditions", "pleased at the reception that Tutuola's works have received from literary critics", "confident that the original integrity of Tutuola's works will be preserved despite numerous translations", "optimistic that Tutuola's works reflect what will become a growing new trend in literature" ]
0
Which one of the following most accurately characterizes the author's attitude toward Tutuola's position in world literature?
With his first published works in the 1950s, Amos Tutuola became the first Nigerian writer to receive wide international recognition. Written in a mix of standard English, idiomatic Nigerian English, and literal translation of his native language, Yoruba, Tutuola's works were quick to be praised by many literary critics as fresh, inventive approaches to the form of the novel. Others, however, dismissed his works as simple retellings of local tales, full of unwelcome liberties taken with the details of the well-known story lines. However, to estimate properly Tutuola's rightful position in world literature, it is essential to be clear about the genre in which he wrote; literary critics have assumed too facilely that he wrote novels. No matter how flexible a definition of the novel one uses, establishing a set of criteria that enable Tutuola's works to be described as such applies to his works a body of assumptions the works are not designed to satisfy. Tutuola is not a novelist but a teller of folktales. Many of his critics are right to suggest that Tutuola's subjects are not strikingly original, but it is important to bear in mind that whereas realism and originality are expected of the novel, the teller of folktales is expected to derive subjects and frameworks from the corpus of traditional lore. The most useful approach to Tutuola's works, then, is one that regards him as working within the African oral tradition. Within this tradition, a folktale is common property, an expression of a people's culture and social circumstances. The teller of folktales knows that the basic story is already known to most listeners and, equally, that the teller's reputation depends on the inventiveness with which the tale is modified and embellished, for what the audience anticipates is not an accurate retelling of the story but effective improvisation and delivery. Thus, within the framework of the basic story, the teller is allowed considerable room to maneuver—in fact, the most brilliant tellers of folktales transform them into unique works. Tutuola's adherence to this tradition is clear: specific episodes, for example, are often repeated for emphasis, and he embellishes familiar tales with personal interpretations or by transferring them to modern settings. The blend of English with local idiom and Yoruba grammatical constructs, in which adjectives and verbs are often interchangeable, re-creates the folktales in singular ways. And, perhaps most revealingly, in the majority of Tutuola's works, the traditional accents and techniques of the teller of folktales are clearly discernible, for example in the adoption of an omniscient, summarizing voice at the end of his narratives, a device that is generally recognized as being employed to conclude most folktales.
200812_4-RC_1_4
[ "his literary works do not exhibit enough similarities to the African oral tradition from which they are drawn", "his mixture of languages is not entirely effective as a vehicle for either traditional folktales or contemporary novels", "his attempt to fuse elements of traditional storytelling style with the format of the novel is detrimental to his artistic purposes", "his writing borrows substantially from well-known story lines and at the same time alters their details", "his unique works are not actually novels, even though he characterizes them as such" ]
3
According to the passage, some critics have criticized Tutuola's work on the ground that
With his first published works in the 1950s, Amos Tutuola became the first Nigerian writer to receive wide international recognition. Written in a mix of standard English, idiomatic Nigerian English, and literal translation of his native language, Yoruba, Tutuola's works were quick to be praised by many literary critics as fresh, inventive approaches to the form of the novel. Others, however, dismissed his works as simple retellings of local tales, full of unwelcome liberties taken with the details of the well-known story lines. However, to estimate properly Tutuola's rightful position in world literature, it is essential to be clear about the genre in which he wrote; literary critics have assumed too facilely that he wrote novels. No matter how flexible a definition of the novel one uses, establishing a set of criteria that enable Tutuola's works to be described as such applies to his works a body of assumptions the works are not designed to satisfy. Tutuola is not a novelist but a teller of folktales. Many of his critics are right to suggest that Tutuola's subjects are not strikingly original, but it is important to bear in mind that whereas realism and originality are expected of the novel, the teller of folktales is expected to derive subjects and frameworks from the corpus of traditional lore. The most useful approach to Tutuola's works, then, is one that regards him as working within the African oral tradition. Within this tradition, a folktale is common property, an expression of a people's culture and social circumstances. The teller of folktales knows that the basic story is already known to most listeners and, equally, that the teller's reputation depends on the inventiveness with which the tale is modified and embellished, for what the audience anticipates is not an accurate retelling of the story but effective improvisation and delivery. Thus, within the framework of the basic story, the teller is allowed considerable room to maneuver—in fact, the most brilliant tellers of folktales transform them into unique works. Tutuola's adherence to this tradition is clear: specific episodes, for example, are often repeated for emphasis, and he embellishes familiar tales with personal interpretations or by transferring them to modern settings. The blend of English with local idiom and Yoruba grammatical constructs, in which adjectives and verbs are often interchangeable, re-creates the folktales in singular ways. And, perhaps most revealingly, in the majority of Tutuola's works, the traditional accents and techniques of the teller of folktales are clearly discernible, for example in the adoption of an omniscient, summarizing voice at the end of his narratives, a device that is generally recognized as being employed to conclude most folktales.
200812_4-RC_1_5
[ "repetition of elements in his stories for emphasis", "relocation of traditional stories to modern settings", "attainment of international recognition", "use of an omniscient narrator in his works", "transformation of Yoruba folktales into modern novels" ]
4
The author attributes each of the following to Tutuola EXCEPT:
With his first published works in the 1950s, Amos Tutuola became the first Nigerian writer to receive wide international recognition. Written in a mix of standard English, idiomatic Nigerian English, and literal translation of his native language, Yoruba, Tutuola's works were quick to be praised by many literary critics as fresh, inventive approaches to the form of the novel. Others, however, dismissed his works as simple retellings of local tales, full of unwelcome liberties taken with the details of the well-known story lines. However, to estimate properly Tutuola's rightful position in world literature, it is essential to be clear about the genre in which he wrote; literary critics have assumed too facilely that he wrote novels. No matter how flexible a definition of the novel one uses, establishing a set of criteria that enable Tutuola's works to be described as such applies to his works a body of assumptions the works are not designed to satisfy. Tutuola is not a novelist but a teller of folktales. Many of his critics are right to suggest that Tutuola's subjects are not strikingly original, but it is important to bear in mind that whereas realism and originality are expected of the novel, the teller of folktales is expected to derive subjects and frameworks from the corpus of traditional lore. The most useful approach to Tutuola's works, then, is one that regards him as working within the African oral tradition. Within this tradition, a folktale is common property, an expression of a people's culture and social circumstances. The teller of folktales knows that the basic story is already known to most listeners and, equally, that the teller's reputation depends on the inventiveness with which the tale is modified and embellished, for what the audience anticipates is not an accurate retelling of the story but effective improvisation and delivery. Thus, within the framework of the basic story, the teller is allowed considerable room to maneuver—in fact, the most brilliant tellers of folktales transform them into unique works. Tutuola's adherence to this tradition is clear: specific episodes, for example, are often repeated for emphasis, and he embellishes familiar tales with personal interpretations or by transferring them to modern settings. The blend of English with local idiom and Yoruba grammatical constructs, in which adjectives and verbs are often interchangeable, re-creates the folktales in singular ways. And, perhaps most revealingly, in the majority of Tutuola's works, the traditional accents and techniques of the teller of folktales are clearly discernible, for example in the adoption of an omniscient, summarizing voice at the end of his narratives, a device that is generally recognized as being employed to conclude most folktales.
200812_4-RC_1_6
[ "distinguish expectations that apply to one literary genre from those that apply to another literary genre", "argue that two sharply differing literary genres are both equally valuable", "challenge critics who ascribe little merit to innovative ways of blending two distinct literary genres", "elucidate those characteristics of one literary genre that have direct counterparts in another, largely dissimilar genre", "argue for a new, more precise analysis of two literary genres whose distinguishing characteristics are poorly understood" ]
0
The author refers to the "corpus of traditional lore" (lines 26–27) as part of an attempt to
With his first published works in the 1950s, Amos Tutuola became the first Nigerian writer to receive wide international recognition. Written in a mix of standard English, idiomatic Nigerian English, and literal translation of his native language, Yoruba, Tutuola's works were quick to be praised by many literary critics as fresh, inventive approaches to the form of the novel. Others, however, dismissed his works as simple retellings of local tales, full of unwelcome liberties taken with the details of the well-known story lines. However, to estimate properly Tutuola's rightful position in world literature, it is essential to be clear about the genre in which he wrote; literary critics have assumed too facilely that he wrote novels. No matter how flexible a definition of the novel one uses, establishing a set of criteria that enable Tutuola's works to be described as such applies to his works a body of assumptions the works are not designed to satisfy. Tutuola is not a novelist but a teller of folktales. Many of his critics are right to suggest that Tutuola's subjects are not strikingly original, but it is important to bear in mind that whereas realism and originality are expected of the novel, the teller of folktales is expected to derive subjects and frameworks from the corpus of traditional lore. The most useful approach to Tutuola's works, then, is one that regards him as working within the African oral tradition. Within this tradition, a folktale is common property, an expression of a people's culture and social circumstances. The teller of folktales knows that the basic story is already known to most listeners and, equally, that the teller's reputation depends on the inventiveness with which the tale is modified and embellished, for what the audience anticipates is not an accurate retelling of the story but effective improvisation and delivery. Thus, within the framework of the basic story, the teller is allowed considerable room to maneuver—in fact, the most brilliant tellers of folktales transform them into unique works. Tutuola's adherence to this tradition is clear: specific episodes, for example, are often repeated for emphasis, and he embellishes familiar tales with personal interpretations or by transferring them to modern settings. The blend of English with local idiom and Yoruba grammatical constructs, in which adjectives and verbs are often interchangeable, re-creates the folktales in singular ways. And, perhaps most revealingly, in the majority of Tutuola's works, the traditional accents and techniques of the teller of folktales are clearly discernible, for example in the adoption of an omniscient, summarizing voice at the end of his narratives, a device that is generally recognized as being employed to conclude most folktales.
200812_4-RC_1_7
[ "illustrate the wide range of Tutuola's body of work", "explain the significance of the literary genre of the folktale and to defend it as a valid art form", "provide an account of Tutuola's body of work in order to help establish appropriate criteria for its evaluation", "distinguish accurately between the genre of the novel and that of the folktale", "summarize the disagreement among critics regarding Tutuola's place in world literature" ]
2
The primary purpose of the passage is to
Mechanisms for recognizing kin are found throughout the plant and animal kingdoms, regardless of an organism's social or mental complexity. Improvements in the general understanding of these mechanisms have turned some biologists' attention to the question of why kin recognition occurs at all. One response to this question is offered by the inclusive fitness theory, which was developed in the 1960s. The theory is based on the realization that an organism transmits its genetic attributes to succeeding generations not solely through its offspring, but more generally through all of its close relatives. Whereas the traditional view of evolution held that natural selection favors the continued genetic representation of individuals within a species that produce the greatest number of offspring, the inclusive fitness theory posits that natural selection similarly favors organisms that help their relatives, because doing so also increases their own total genetic representation. The theory has helped to explain previously mysterious phenomena, including the evolution of social insect species like the honeybee, most of whose members do not produce offspring and exist only to nurture relatives. Inclusive fitness theory has also been applied usefully to new findings concerning cannibalism within animal species. Based on the theory, cannibals should have evolved to avoid eating their own kin because of the obvious genetic costs of such a practice. Spadefoot toad tadpoles provide an illustration. Biologists have found that all tadpoles of that species begin life as omnivores, feeding mainly on organic debris in their soon-to-be-dry pool in the desert, but that occasionally one tadpole eats another or eats a freshwater shrimp. This event can trigger changes in the tadpole's physiology and dietary preference, causing the tadpole to become larger and exclusively carnivorous, feasting on other animals including members of its own species. Yet the cannibals have a procedure of discrimination whereby they nip at other tadpoles, eating nonsiblings but releasing siblings unharmed. This suggests that the inclusive fitness theory offers at least a partial answer to why kin recognition develops. Interestingly, a cannibal tadpole is less likely to avoid eating kin when it becomes very hungry, apparently putting its own unique genetic makeup ahead of its siblings'. But there may be other reasons why organisms recognize kin. For example, it has recently been found that tiger salamander larvae, also either omnivorous or cannibalistic, are plagued in nature by a deadly bacterium. Furthermore, it was determined that cannibal larvae are especially likely to be infected by eating diseased species members. The fact that this bacterium is more deadly when it comes from a close relative with a similar immune system suggests that natural selection may favor cannibals that avoid such pathogens by not eating kin. For tiger salamanders then, kin recognition can be explained simply as a means by which an organism preserves its own life, not as a means to aid in relatives' survival.
200812_4-RC_2_8
[ "Some findings support the hypothesis that kin recognition emerged through natural selection because it increased organisms' total genetic representation, but this hypothesis may not explain all instances of kin recognition.", "Current research supports the view that the mechanisms enabling the members of a species to recognize close relatives are as various as the purposes served by that ability.", "Recent research involving tiger salamanders undermines the hypothesis concerning the purpose of kin recognition that is espoused by traditional evolutionary theorists.", "New research involving tiger salamanders indicates that the traditional theory of natural selection is more strongly supported by the evidence than is thought by those who consider only the case of the spadefoot toad tadpole.", "While traditional evolutionary theory was unable to account for the phenomenon of kin recognition, this phenomenon is fully explained by the inclusive fitness theory." ]
0
Which one of the following most accurately expresses the main point of the passage?
Mechanisms for recognizing kin are found throughout the plant and animal kingdoms, regardless of an organism's social or mental complexity. Improvements in the general understanding of these mechanisms have turned some biologists' attention to the question of why kin recognition occurs at all. One response to this question is offered by the inclusive fitness theory, which was developed in the 1960s. The theory is based on the realization that an organism transmits its genetic attributes to succeeding generations not solely through its offspring, but more generally through all of its close relatives. Whereas the traditional view of evolution held that natural selection favors the continued genetic representation of individuals within a species that produce the greatest number of offspring, the inclusive fitness theory posits that natural selection similarly favors organisms that help their relatives, because doing so also increases their own total genetic representation. The theory has helped to explain previously mysterious phenomena, including the evolution of social insect species like the honeybee, most of whose members do not produce offspring and exist only to nurture relatives. Inclusive fitness theory has also been applied usefully to new findings concerning cannibalism within animal species. Based on the theory, cannibals should have evolved to avoid eating their own kin because of the obvious genetic costs of such a practice. Spadefoot toad tadpoles provide an illustration. Biologists have found that all tadpoles of that species begin life as omnivores, feeding mainly on organic debris in their soon-to-be-dry pool in the desert, but that occasionally one tadpole eats another or eats a freshwater shrimp. This event can trigger changes in the tadpole's physiology and dietary preference, causing the tadpole to become larger and exclusively carnivorous, feasting on other animals including members of its own species. Yet the cannibals have a procedure of discrimination whereby they nip at other tadpoles, eating nonsiblings but releasing siblings unharmed. This suggests that the inclusive fitness theory offers at least a partial answer to why kin recognition develops. Interestingly, a cannibal tadpole is less likely to avoid eating kin when it becomes very hungry, apparently putting its own unique genetic makeup ahead of its siblings'. But there may be other reasons why organisms recognize kin. For example, it has recently been found that tiger salamander larvae, also either omnivorous or cannibalistic, are plagued in nature by a deadly bacterium. Furthermore, it was determined that cannibal larvae are especially likely to be infected by eating diseased species members. The fact that this bacterium is more deadly when it comes from a close relative with a similar immune system suggests that natural selection may favor cannibals that avoid such pathogens by not eating kin. For tiger salamanders then, kin recognition can be explained simply as a means by which an organism preserves its own life, not as a means to aid in relatives' survival.
200812_4-RC_2_9
[ "They develop the ability to recognize fellow carnivores.", "They feed only upon omnivorous tadpoles.", "They change in body size when they become carnivores.", "Their carnivorousness constitutes an important piece of evidence that calls into question the inclusive fitness theory.", "Their carnivorousness would not occur unless it contributed in some way to the evolutionary success of the spadefoot toad species." ]
2
The passage states which one of the following about some spadefoot toad tadpoles?
Mechanisms for recognizing kin are found throughout the plant and animal kingdoms, regardless of an organism's social or mental complexity. Improvements in the general understanding of these mechanisms have turned some biologists' attention to the question of why kin recognition occurs at all. One response to this question is offered by the inclusive fitness theory, which was developed in the 1960s. The theory is based on the realization that an organism transmits its genetic attributes to succeeding generations not solely through its offspring, but more generally through all of its close relatives. Whereas the traditional view of evolution held that natural selection favors the continued genetic representation of individuals within a species that produce the greatest number of offspring, the inclusive fitness theory posits that natural selection similarly favors organisms that help their relatives, because doing so also increases their own total genetic representation. The theory has helped to explain previously mysterious phenomena, including the evolution of social insect species like the honeybee, most of whose members do not produce offspring and exist only to nurture relatives. Inclusive fitness theory has also been applied usefully to new findings concerning cannibalism within animal species. Based on the theory, cannibals should have evolved to avoid eating their own kin because of the obvious genetic costs of such a practice. Spadefoot toad tadpoles provide an illustration. Biologists have found that all tadpoles of that species begin life as omnivores, feeding mainly on organic debris in their soon-to-be-dry pool in the desert, but that occasionally one tadpole eats another or eats a freshwater shrimp. This event can trigger changes in the tadpole's physiology and dietary preference, causing the tadpole to become larger and exclusively carnivorous, feasting on other animals including members of its own species. Yet the cannibals have a procedure of discrimination whereby they nip at other tadpoles, eating nonsiblings but releasing siblings unharmed. This suggests that the inclusive fitness theory offers at least a partial answer to why kin recognition develops. Interestingly, a cannibal tadpole is less likely to avoid eating kin when it becomes very hungry, apparently putting its own unique genetic makeup ahead of its siblings'. But there may be other reasons why organisms recognize kin. For example, it has recently been found that tiger salamander larvae, also either omnivorous or cannibalistic, are plagued in nature by a deadly bacterium. Furthermore, it was determined that cannibal larvae are especially likely to be infected by eating diseased species members. The fact that this bacterium is more deadly when it comes from a close relative with a similar immune system suggests that natural selection may favor cannibals that avoid such pathogens by not eating kin. For tiger salamanders then, kin recognition can be explained simply as a means by which an organism preserves its own life, not as a means to aid in relatives' survival.
200812_4-RC_2_10
[ "It is impossible to understand the mechanisms underlying kin recognition until an evolutionary explanation of such recognition has been attained.", "Such explanations require no modifications to traditional evolutionary theory.", "For any such explanation to be fully adequate it should ignore the differences of social or mental complexity of the organisms whose abilities it is intended to explain.", "Kin recognition may have different evolutionary explanations in different species.", "No other evolutionary explanation can account for the wide diversity of unusual phenomena with the same success as the inclusive fitness theory." ]
3
Based on the passage, the author would be most likely to agree with which one of the following statements about evolutionary explanations of kin recognition?
Mechanisms for recognizing kin are found throughout the plant and animal kingdoms, regardless of an organism's social or mental complexity. Improvements in the general understanding of these mechanisms have turned some biologists' attention to the question of why kin recognition occurs at all. One response to this question is offered by the inclusive fitness theory, which was developed in the 1960s. The theory is based on the realization that an organism transmits its genetic attributes to succeeding generations not solely through its offspring, but more generally through all of its close relatives. Whereas the traditional view of evolution held that natural selection favors the continued genetic representation of individuals within a species that produce the greatest number of offspring, the inclusive fitness theory posits that natural selection similarly favors organisms that help their relatives, because doing so also increases their own total genetic representation. The theory has helped to explain previously mysterious phenomena, including the evolution of social insect species like the honeybee, most of whose members do not produce offspring and exist only to nurture relatives. Inclusive fitness theory has also been applied usefully to new findings concerning cannibalism within animal species. Based on the theory, cannibals should have evolved to avoid eating their own kin because of the obvious genetic costs of such a practice. Spadefoot toad tadpoles provide an illustration. Biologists have found that all tadpoles of that species begin life as omnivores, feeding mainly on organic debris in their soon-to-be-dry pool in the desert, but that occasionally one tadpole eats another or eats a freshwater shrimp. This event can trigger changes in the tadpole's physiology and dietary preference, causing the tadpole to become larger and exclusively carnivorous, feasting on other animals including members of its own species. Yet the cannibals have a procedure of discrimination whereby they nip at other tadpoles, eating nonsiblings but releasing siblings unharmed. This suggests that the inclusive fitness theory offers at least a partial answer to why kin recognition develops. Interestingly, a cannibal tadpole is less likely to avoid eating kin when it becomes very hungry, apparently putting its own unique genetic makeup ahead of its siblings'. But there may be other reasons why organisms recognize kin. For example, it has recently been found that tiger salamander larvae, also either omnivorous or cannibalistic, are plagued in nature by a deadly bacterium. Furthermore, it was determined that cannibal larvae are especially likely to be infected by eating diseased species members. The fact that this bacterium is more deadly when it comes from a close relative with a similar immune system suggests that natural selection may favor cannibals that avoid such pathogens by not eating kin. For tiger salamanders then, kin recognition can be explained simply as a means by which an organism preserves its own life, not as a means to aid in relatives' survival.
200812_4-RC_2_11
[ "to draw attention to behavior that further complicates the set of facts to be explained by any theory of natural selection that accounts for kin recognition", "to explain why cannibals in most species eat their kin less often than do cannibal spadefoot toad tadpoles", "to describe behavior that lends support to the account of kin recognition presented in the second paragraph", "to offer evidence that the behavior of cannibal spadefoot toad tadpoles is unexplainable", "to imply that the described behavior is more relevant to the issue at hand than is the immediately preceding material" ]
0
Which one of the following most accurately describes the function of the last sentence of the second paragraph?
Mechanisms for recognizing kin are found throughout the plant and animal kingdoms, regardless of an organism's social or mental complexity. Improvements in the general understanding of these mechanisms have turned some biologists' attention to the question of why kin recognition occurs at all. One response to this question is offered by the inclusive fitness theory, which was developed in the 1960s. The theory is based on the realization that an organism transmits its genetic attributes to succeeding generations not solely through its offspring, but more generally through all of its close relatives. Whereas the traditional view of evolution held that natural selection favors the continued genetic representation of individuals within a species that produce the greatest number of offspring, the inclusive fitness theory posits that natural selection similarly favors organisms that help their relatives, because doing so also increases their own total genetic representation. The theory has helped to explain previously mysterious phenomena, including the evolution of social insect species like the honeybee, most of whose members do not produce offspring and exist only to nurture relatives. Inclusive fitness theory has also been applied usefully to new findings concerning cannibalism within animal species. Based on the theory, cannibals should have evolved to avoid eating their own kin because of the obvious genetic costs of such a practice. Spadefoot toad tadpoles provide an illustration. Biologists have found that all tadpoles of that species begin life as omnivores, feeding mainly on organic debris in their soon-to-be-dry pool in the desert, but that occasionally one tadpole eats another or eats a freshwater shrimp. This event can trigger changes in the tadpole's physiology and dietary preference, causing the tadpole to become larger and exclusively carnivorous, feasting on other animals including members of its own species. Yet the cannibals have a procedure of discrimination whereby they nip at other tadpoles, eating nonsiblings but releasing siblings unharmed. This suggests that the inclusive fitness theory offers at least a partial answer to why kin recognition develops. Interestingly, a cannibal tadpole is less likely to avoid eating kin when it becomes very hungry, apparently putting its own unique genetic makeup ahead of its siblings'. But there may be other reasons why organisms recognize kin. For example, it has recently been found that tiger salamander larvae, also either omnivorous or cannibalistic, are plagued in nature by a deadly bacterium. Furthermore, it was determined that cannibal larvae are especially likely to be infected by eating diseased species members. The fact that this bacterium is more deadly when it comes from a close relative with a similar immune system suggests that natural selection may favor cannibals that avoid such pathogens by not eating kin. For tiger salamanders then, kin recognition can be explained simply as a means by which an organism preserves its own life, not as a means to aid in relatives' survival.
200812_4-RC_2_12
[ "It is not dependent solely on the use of visual cues.", "It is neither utilized nor possessed by those tadpoles that do not become cannibalistic.", "It does not always allow a tadpole to distinguish its siblings from tadpoles that are not siblings.", "It is rendered unnecessary by physiological changes accompanying the dietary shift from omnivorousness to carnivorousness.", "It could not have developed in a species in which all members are omnivorous." ]
0
The passage most strongly supports which one of the following statements about the mechanism by which cannibal spadefoot toad tadpoles recognize their kin?
Mechanisms for recognizing kin are found throughout the plant and animal kingdoms, regardless of an organism's social or mental complexity. Improvements in the general understanding of these mechanisms have turned some biologists' attention to the question of why kin recognition occurs at all. One response to this question is offered by the inclusive fitness theory, which was developed in the 1960s. The theory is based on the realization that an organism transmits its genetic attributes to succeeding generations not solely through its offspring, but more generally through all of its close relatives. Whereas the traditional view of evolution held that natural selection favors the continued genetic representation of individuals within a species that produce the greatest number of offspring, the inclusive fitness theory posits that natural selection similarly favors organisms that help their relatives, because doing so also increases their own total genetic representation. The theory has helped to explain previously mysterious phenomena, including the evolution of social insect species like the honeybee, most of whose members do not produce offspring and exist only to nurture relatives. Inclusive fitness theory has also been applied usefully to new findings concerning cannibalism within animal species. Based on the theory, cannibals should have evolved to avoid eating their own kin because of the obvious genetic costs of such a practice. Spadefoot toad tadpoles provide an illustration. Biologists have found that all tadpoles of that species begin life as omnivores, feeding mainly on organic debris in their soon-to-be-dry pool in the desert, but that occasionally one tadpole eats another or eats a freshwater shrimp. This event can trigger changes in the tadpole's physiology and dietary preference, causing the tadpole to become larger and exclusively carnivorous, feasting on other animals including members of its own species. Yet the cannibals have a procedure of discrimination whereby they nip at other tadpoles, eating nonsiblings but releasing siblings unharmed. This suggests that the inclusive fitness theory offers at least a partial answer to why kin recognition develops. Interestingly, a cannibal tadpole is less likely to avoid eating kin when it becomes very hungry, apparently putting its own unique genetic makeup ahead of its siblings'. But there may be other reasons why organisms recognize kin. For example, it has recently been found that tiger salamander larvae, also either omnivorous or cannibalistic, are plagued in nature by a deadly bacterium. Furthermore, it was determined that cannibal larvae are especially likely to be infected by eating diseased species members. The fact that this bacterium is more deadly when it comes from a close relative with a similar immune system suggests that natural selection may favor cannibals that avoid such pathogens by not eating kin. For tiger salamanders then, kin recognition can be explained simply as a means by which an organism preserves its own life, not as a means to aid in relatives' survival.
200812_4-RC_2_13
[ "The mechanisms are most easily explained if we assume that they have a similar purpose in all species regardless of the species' social or mental complexities.", "The mechanisms have become more clearly understood, prompting interest in the purpose they serve.", "The mechanisms have become the focus of theoretical attention only since the 1960s.", "The detailed workings of these mechanisms must be better understood before their purpose can be fully explained.", "The mechanisms operate differently in different species even when they serve exactly the same function." ]
1
The passage states which one of the following about the mechanisms that enable organisms to recognize their close genetic relatives?
Mechanisms for recognizing kin are found throughout the plant and animal kingdoms, regardless of an organism's social or mental complexity. Improvements in the general understanding of these mechanisms have turned some biologists' attention to the question of why kin recognition occurs at all. One response to this question is offered by the inclusive fitness theory, which was developed in the 1960s. The theory is based on the realization that an organism transmits its genetic attributes to succeeding generations not solely through its offspring, but more generally through all of its close relatives. Whereas the traditional view of evolution held that natural selection favors the continued genetic representation of individuals within a species that produce the greatest number of offspring, the inclusive fitness theory posits that natural selection similarly favors organisms that help their relatives, because doing so also increases their own total genetic representation. The theory has helped to explain previously mysterious phenomena, including the evolution of social insect species like the honeybee, most of whose members do not produce offspring and exist only to nurture relatives. Inclusive fitness theory has also been applied usefully to new findings concerning cannibalism within animal species. Based on the theory, cannibals should have evolved to avoid eating their own kin because of the obvious genetic costs of such a practice. Spadefoot toad tadpoles provide an illustration. Biologists have found that all tadpoles of that species begin life as omnivores, feeding mainly on organic debris in their soon-to-be-dry pool in the desert, but that occasionally one tadpole eats another or eats a freshwater shrimp. This event can trigger changes in the tadpole's physiology and dietary preference, causing the tadpole to become larger and exclusively carnivorous, feasting on other animals including members of its own species. Yet the cannibals have a procedure of discrimination whereby they nip at other tadpoles, eating nonsiblings but releasing siblings unharmed. This suggests that the inclusive fitness theory offers at least a partial answer to why kin recognition develops. Interestingly, a cannibal tadpole is less likely to avoid eating kin when it becomes very hungry, apparently putting its own unique genetic makeup ahead of its siblings'. But there may be other reasons why organisms recognize kin. For example, it has recently been found that tiger salamander larvae, also either omnivorous or cannibalistic, are plagued in nature by a deadly bacterium. Furthermore, it was determined that cannibal larvae are especially likely to be infected by eating diseased species members. The fact that this bacterium is more deadly when it comes from a close relative with a similar immune system suggests that natural selection may favor cannibals that avoid such pathogens by not eating kin. For tiger salamanders then, kin recognition can be explained simply as a means by which an organism preserves its own life, not as a means to aid in relatives' survival.
200812_4-RC_2_14
[ "was not known to be true before the 1960s", "can be explained only if we assume that these members are in turn nurtured by the relatives they nurture", "is what led most biologists to reject the traditional view of evolution", "calls into question the view that evolution proceeds by natural selection", "is difficult to explain without at least supplementing the traditional view of evolution with further explanatory hypotheses" ]
4
The information in the passage most strongly suggests that the fact that most honeybees exist only to nurture relatives
Mechanisms for recognizing kin are found throughout the plant and animal kingdoms, regardless of an organism's social or mental complexity. Improvements in the general understanding of these mechanisms have turned some biologists' attention to the question of why kin recognition occurs at all. One response to this question is offered by the inclusive fitness theory, which was developed in the 1960s. The theory is based on the realization that an organism transmits its genetic attributes to succeeding generations not solely through its offspring, but more generally through all of its close relatives. Whereas the traditional view of evolution held that natural selection favors the continued genetic representation of individuals within a species that produce the greatest number of offspring, the inclusive fitness theory posits that natural selection similarly favors organisms that help their relatives, because doing so also increases their own total genetic representation. The theory has helped to explain previously mysterious phenomena, including the evolution of social insect species like the honeybee, most of whose members do not produce offspring and exist only to nurture relatives. Inclusive fitness theory has also been applied usefully to new findings concerning cannibalism within animal species. Based on the theory, cannibals should have evolved to avoid eating their own kin because of the obvious genetic costs of such a practice. Spadefoot toad tadpoles provide an illustration. Biologists have found that all tadpoles of that species begin life as omnivores, feeding mainly on organic debris in their soon-to-be-dry pool in the desert, but that occasionally one tadpole eats another or eats a freshwater shrimp. This event can trigger changes in the tadpole's physiology and dietary preference, causing the tadpole to become larger and exclusively carnivorous, feasting on other animals including members of its own species. Yet the cannibals have a procedure of discrimination whereby they nip at other tadpoles, eating nonsiblings but releasing siblings unharmed. This suggests that the inclusive fitness theory offers at least a partial answer to why kin recognition develops. Interestingly, a cannibal tadpole is less likely to avoid eating kin when it becomes very hungry, apparently putting its own unique genetic makeup ahead of its siblings'. But there may be other reasons why organisms recognize kin. For example, it has recently been found that tiger salamander larvae, also either omnivorous or cannibalistic, are plagued in nature by a deadly bacterium. Furthermore, it was determined that cannibal larvae are especially likely to be infected by eating diseased species members. The fact that this bacterium is more deadly when it comes from a close relative with a similar immune system suggests that natural selection may favor cannibals that avoid such pathogens by not eating kin. For tiger salamanders then, kin recognition can be explained simply as a means by which an organism preserves its own life, not as a means to aid in relatives' survival.
200812_4-RC_2_15
[ "Many tiger salamander larvae infected by the deadly bacterium are not cannibalistic.", "The factor that determines which tiger salamander larvae are carnivorous and which are omnivorous is not contained in the genetic makeup of the larvae.", "Kin recognition helps tiger salamanders avoid inbreeding that may be life-threatening to their offspring.", "Noncannibalistic tiger salamanders tend to produce fewer offspring than cannibalistic tiger salamanders.", "Cannibalistic tiger salamanders are immune to certain diseases to which noncannibalistic salamanders are not." ]
2
Which one of the following would, if true, most help to undermine the author's evaluation in the last sentence of the passage?
Passage A There is no universally accepted definition within international law for the term "national minority." It is most commonly applied to (1) groups of persons—not necessarily citizens—under the jurisdiction of one country who have ethnic ties to another "homeland" country, or (2) groups of citizens of a country who have lasting ties to that country and have no such ties to any other country, but are distinguished from the majority of the population by ethnicity, religion, or language. The terms "people" and "nation" are also vaguely defined in international agreements. Documents that refer to a "nation" generally link the term to the concept of "nationalism," which is often associated with ties to land. It also connotes sovereignty, for which reason, perhaps, "people" is often used instead of "nation" for groups subject to a colonial power. While the lack of definition of the terms "minority," "people," and "nation" presents difficulties to numerous minority groups, this lack is particularly problematic for the Roma (Gypsies). The Roma are not a colonized people, they do not have a homeland, and many do not bear ties to any currently existing country. Some Roma are not even citizens of any country, in part because of their nomadic way of life, which developed in response to centuries of fleeing persecution. Instead, they have ethnic and linguistic ties to other groups of Roma that reside in other countries. Passage B Capotorti's definition of a minority includes four empirical criteria—a group's being numerically smaller than the rest of the population of the state; their being nondominant; their having distinctive ethnic, linguistic, or religious characteristics; and their desiring to preserve their own culture—and one legal criterion, that they be citizens of the state in question. This last element can be problematic, given the previous nomadic character of the Roma, that they still cross borders between European states to avoid persecution, and that some states have denied them citizenship, and thus minority status. Because this element essentially grants the state the arbitrary right to decide if the Roma constitute a minority without reference to empirical characteristics, it seems patently unfair that it should be included in the definition. However, the Roma easily fulfill the four objective elements of Capotorti's definition and should, therefore, be considered a minority in all major European states. Numerically, they are nowhere near a majority, though they number in the hundreds of thousands, even millions, in some states. Their nondominant position is evident—they are not even acknowledged as a minority in some states. The Roma have a number of distinctive linguistic, ethnic, and religious characteristics. For example, most speak Romani, an Indo-European language descended from Sanskrit. Roma groups also have their own distinctive legal and court systems, which are group oriented rather than individual-rights oriented. That they have preserved their language, customs, and identity through centuries of persecution is evidence enough of their desire to preserve their culture.
200812_4-RC_3_16
[ "Different definitions of certain key terms in international law conflict with one another in their application to the Roma.", "In at least some countries in which they live, the Roma are not generally considered a minority group.", "The lack of agreement regarding the definitions of such terms as \"minority,\" \"people,\" and \"nation\" is partly due to the unclear application of the terms to groups such as the Roma.", "Any attempt to define such concepts as people, nation, or minority group will probably fail to apply to certain borderline cases such as the Roma.", "The absence of a clear, generally agreed-upon understanding of what constitutes a people, nation, or minority group is a problem, especially in relation to the Roma." ]
4
Which one of the following most accurately expresses the main point of passage A?
Passage A There is no universally accepted definition within international law for the term "national minority." It is most commonly applied to (1) groups of persons—not necessarily citizens—under the jurisdiction of one country who have ethnic ties to another "homeland" country, or (2) groups of citizens of a country who have lasting ties to that country and have no such ties to any other country, but are distinguished from the majority of the population by ethnicity, religion, or language. The terms "people" and "nation" are also vaguely defined in international agreements. Documents that refer to a "nation" generally link the term to the concept of "nationalism," which is often associated with ties to land. It also connotes sovereignty, for which reason, perhaps, "people" is often used instead of "nation" for groups subject to a colonial power. While the lack of definition of the terms "minority," "people," and "nation" presents difficulties to numerous minority groups, this lack is particularly problematic for the Roma (Gypsies). The Roma are not a colonized people, they do not have a homeland, and many do not bear ties to any currently existing country. Some Roma are not even citizens of any country, in part because of their nomadic way of life, which developed in response to centuries of fleeing persecution. Instead, they have ethnic and linguistic ties to other groups of Roma that reside in other countries. Passage B Capotorti's definition of a minority includes four empirical criteria—a group's being numerically smaller than the rest of the population of the state; their being nondominant; their having distinctive ethnic, linguistic, or religious characteristics; and their desiring to preserve their own culture—and one legal criterion, that they be citizens of the state in question. This last element can be problematic, given the previous nomadic character of the Roma, that they still cross borders between European states to avoid persecution, and that some states have denied them citizenship, and thus minority status. Because this element essentially grants the state the arbitrary right to decide if the Roma constitute a minority without reference to empirical characteristics, it seems patently unfair that it should be included in the definition. However, the Roma easily fulfill the four objective elements of Capotorti's definition and should, therefore, be considered a minority in all major European states. Numerically, they are nowhere near a majority, though they number in the hundreds of thousands, even millions, in some states. Their nondominant position is evident—they are not even acknowledged as a minority in some states. The Roma have a number of distinctive linguistic, ethnic, and religious characteristics. For example, most speak Romani, an Indo-European language descended from Sanskrit. Roma groups also have their own distinctive legal and court systems, which are group oriented rather than individual-rights oriented. That they have preserved their language, customs, and identity through centuries of persecution is evidence enough of their desire to preserve their culture.
200812_4-RC_3_17
[ "giving rise to intense debate", "confusing and unclear", "resulting in difficulties", "difficult to solve", "theoretically incoherent" ]
2
The term "problematic" has which one of the following meanings in both passage A (line 19) and passage B (line 35)?
Passage A There is no universally accepted definition within international law for the term "national minority." It is most commonly applied to (1) groups of persons—not necessarily citizens—under the jurisdiction of one country who have ethnic ties to another "homeland" country, or (2) groups of citizens of a country who have lasting ties to that country and have no such ties to any other country, but are distinguished from the majority of the population by ethnicity, religion, or language. The terms "people" and "nation" are also vaguely defined in international agreements. Documents that refer to a "nation" generally link the term to the concept of "nationalism," which is often associated with ties to land. It also connotes sovereignty, for which reason, perhaps, "people" is often used instead of "nation" for groups subject to a colonial power. While the lack of definition of the terms "minority," "people," and "nation" presents difficulties to numerous minority groups, this lack is particularly problematic for the Roma (Gypsies). The Roma are not a colonized people, they do not have a homeland, and many do not bear ties to any currently existing country. Some Roma are not even citizens of any country, in part because of their nomadic way of life, which developed in response to centuries of fleeing persecution. Instead, they have ethnic and linguistic ties to other groups of Roma that reside in other countries. Passage B Capotorti's definition of a minority includes four empirical criteria—a group's being numerically smaller than the rest of the population of the state; their being nondominant; their having distinctive ethnic, linguistic, or religious characteristics; and their desiring to preserve their own culture—and one legal criterion, that they be citizens of the state in question. This last element can be problematic, given the previous nomadic character of the Roma, that they still cross borders between European states to avoid persecution, and that some states have denied them citizenship, and thus minority status. Because this element essentially grants the state the arbitrary right to decide if the Roma constitute a minority without reference to empirical characteristics, it seems patently unfair that it should be included in the definition. However, the Roma easily fulfill the four objective elements of Capotorti's definition and should, therefore, be considered a minority in all major European states. Numerically, they are nowhere near a majority, though they number in the hundreds of thousands, even millions, in some states. Their nondominant position is evident—they are not even acknowledged as a minority in some states. The Roma have a number of distinctive linguistic, ethnic, and religious characteristics. For example, most speak Romani, an Indo-European language descended from Sanskrit. Roma groups also have their own distinctive legal and court systems, which are group oriented rather than individual-rights oriented. That they have preserved their language, customs, and identity through centuries of persecution is evidence enough of their desire to preserve their culture.
200812_4-RC_3_18
[ "Those living in one country have ethnic ties to Roma in other countries.", "Some of them practice a nomadic way of life.", "They, as a people, have no recognizable homeland.", "In some countries, their population exceeds one million.", "The lack of a completely satisfactory definition of \"minority\" is a greater problem for them than for most." ]
3
Which one of the following claims about the Roma is NOT made in passage A?
Passage A There is no universally accepted definition within international law for the term "national minority." It is most commonly applied to (1) groups of persons—not necessarily citizens—under the jurisdiction of one country who have ethnic ties to another "homeland" country, or (2) groups of citizens of a country who have lasting ties to that country and have no such ties to any other country, but are distinguished from the majority of the population by ethnicity, religion, or language. The terms "people" and "nation" are also vaguely defined in international agreements. Documents that refer to a "nation" generally link the term to the concept of "nationalism," which is often associated with ties to land. It also connotes sovereignty, for which reason, perhaps, "people" is often used instead of "nation" for groups subject to a colonial power. While the lack of definition of the terms "minority," "people," and "nation" presents difficulties to numerous minority groups, this lack is particularly problematic for the Roma (Gypsies). The Roma are not a colonized people, they do not have a homeland, and many do not bear ties to any currently existing country. Some Roma are not even citizens of any country, in part because of their nomadic way of life, which developed in response to centuries of fleeing persecution. Instead, they have ethnic and linguistic ties to other groups of Roma that reside in other countries. Passage B Capotorti's definition of a minority includes four empirical criteria—a group's being numerically smaller than the rest of the population of the state; their being nondominant; their having distinctive ethnic, linguistic, or religious characteristics; and their desiring to preserve their own culture—and one legal criterion, that they be citizens of the state in question. This last element can be problematic, given the previous nomadic character of the Roma, that they still cross borders between European states to avoid persecution, and that some states have denied them citizenship, and thus minority status. Because this element essentially grants the state the arbitrary right to decide if the Roma constitute a minority without reference to empirical characteristics, it seems patently unfair that it should be included in the definition. However, the Roma easily fulfill the four objective elements of Capotorti's definition and should, therefore, be considered a minority in all major European states. Numerically, they are nowhere near a majority, though they number in the hundreds of thousands, even millions, in some states. Their nondominant position is evident—they are not even acknowledged as a minority in some states. The Roma have a number of distinctive linguistic, ethnic, and religious characteristics. For example, most speak Romani, an Indo-European language descended from Sanskrit. Roma groups also have their own distinctive legal and court systems, which are group oriented rather than individual-rights oriented. That they have preserved their language, customs, and identity through centuries of persecution is evidence enough of their desire to preserve their culture.
200812_4-RC_3_19
[ "The author of passage A, but not the author of passage B, disapproves of the latitude that international law allows individual states in determining their relations to nomadic Roma populations.", "The author of passage B, but not the author of passage A, considers the problems of the Roma to be a noteworthy example of how international law can be ineffective.", "The author of passage B, but not the author of passage A, considers the Roma to be a paradigmatic example of a people who do not constitute a nation.", "Both authors would prefer that the political issues involving the Roma be resolved on a case-by-case basis within each individual country rather than through international law.", "Both authors consider the problems that the Roma face in relation to international law to be anomalous and special." ]
4
The authors' views regarding the status of the Roma can most accurately be described in which one of the following ways?
Passage A There is no universally accepted definition within international law for the term "national minority." It is most commonly applied to (1) groups of persons—not necessarily citizens—under the jurisdiction of one country who have ethnic ties to another "homeland" country, or (2) groups of citizens of a country who have lasting ties to that country and have no such ties to any other country, but are distinguished from the majority of the population by ethnicity, religion, or language. The terms "people" and "nation" are also vaguely defined in international agreements. Documents that refer to a "nation" generally link the term to the concept of "nationalism," which is often associated with ties to land. It also connotes sovereignty, for which reason, perhaps, "people" is often used instead of "nation" for groups subject to a colonial power. While the lack of definition of the terms "minority," "people," and "nation" presents difficulties to numerous minority groups, this lack is particularly problematic for the Roma (Gypsies). The Roma are not a colonized people, they do not have a homeland, and many do not bear ties to any currently existing country. Some Roma are not even citizens of any country, in part because of their nomadic way of life, which developed in response to centuries of fleeing persecution. Instead, they have ethnic and linguistic ties to other groups of Roma that reside in other countries. Passage B Capotorti's definition of a minority includes four empirical criteria—a group's being numerically smaller than the rest of the population of the state; their being nondominant; their having distinctive ethnic, linguistic, or religious characteristics; and their desiring to preserve their own culture—and one legal criterion, that they be citizens of the state in question. This last element can be problematic, given the previous nomadic character of the Roma, that they still cross borders between European states to avoid persecution, and that some states have denied them citizenship, and thus minority status. Because this element essentially grants the state the arbitrary right to decide if the Roma constitute a minority without reference to empirical characteristics, it seems patently unfair that it should be included in the definition. However, the Roma easily fulfill the four objective elements of Capotorti's definition and should, therefore, be considered a minority in all major European states. Numerically, they are nowhere near a majority, though they number in the hundreds of thousands, even millions, in some states. Their nondominant position is evident—they are not even acknowledged as a minority in some states. The Roma have a number of distinctive linguistic, ethnic, and religious characteristics. For example, most speak Romani, an Indo-European language descended from Sanskrit. Roma groups also have their own distinctive legal and court systems, which are group oriented rather than individual-rights oriented. That they have preserved their language, customs, and identity through centuries of persecution is evidence enough of their desire to preserve their culture.
200812_4-RC_3_20
[ "\"The Lack of Clear-Cut Criteria for Classifying Jobs as Technical Causes Problems for Welders\" and \"A Point-by-Point Argument That Welding Fulfills the Union's Criteria for Classification of Jobs as 'Technical'\"", "\"Why the Current Criteria for Professional Competence in Welding Have Not Been Effectively Applied\" and \"A Review of the Essential Elements of Any Formal Statement of Professional Standards\"", "\"The Need for a Revised Definition of the Concept of Welding in Relation to Other Technical Jobs\" and \"An Enumeration and Description of the Essential Job Duties Usually Carried Out by Union Welders\"", "\"The Lack of Competent Welders in Our Company Can Be Attributed to a General Disregard for Professional and Technical Staff Recruitment\" and \"A Discussion of the Factors That Companies Should Consider in Recruiting Employees\"", "\"The Conceptual Links Between Professionalism and Technical Expertise\" and \"A Refutation of the Union's Position Regarding Which Types of Jobs Should Be Classified as Neither Professional nor Technical\"" ]
0
The relationship between which one of the following pairs of documents is most analogous to the relationship between passage A and passage B?
Passage A There is no universally accepted definition within international law for the term "national minority." It is most commonly applied to (1) groups of persons—not necessarily citizens—under the jurisdiction of one country who have ethnic ties to another "homeland" country, or (2) groups of citizens of a country who have lasting ties to that country and have no such ties to any other country, but are distinguished from the majority of the population by ethnicity, religion, or language. The terms "people" and "nation" are also vaguely defined in international agreements. Documents that refer to a "nation" generally link the term to the concept of "nationalism," which is often associated with ties to land. It also connotes sovereignty, for which reason, perhaps, "people" is often used instead of "nation" for groups subject to a colonial power. While the lack of definition of the terms "minority," "people," and "nation" presents difficulties to numerous minority groups, this lack is particularly problematic for the Roma (Gypsies). The Roma are not a colonized people, they do not have a homeland, and many do not bear ties to any currently existing country. Some Roma are not even citizens of any country, in part because of their nomadic way of life, which developed in response to centuries of fleeing persecution. Instead, they have ethnic and linguistic ties to other groups of Roma that reside in other countries. Passage B Capotorti's definition of a minority includes four empirical criteria—a group's being numerically smaller than the rest of the population of the state; their being nondominant; their having distinctive ethnic, linguistic, or religious characteristics; and their desiring to preserve their own culture—and one legal criterion, that they be citizens of the state in question. This last element can be problematic, given the previous nomadic character of the Roma, that they still cross borders between European states to avoid persecution, and that some states have denied them citizenship, and thus minority status. Because this element essentially grants the state the arbitrary right to decide if the Roma constitute a minority without reference to empirical characteristics, it seems patently unfair that it should be included in the definition. However, the Roma easily fulfill the four objective elements of Capotorti's definition and should, therefore, be considered a minority in all major European states. Numerically, they are nowhere near a majority, though they number in the hundreds of thousands, even millions, in some states. Their nondominant position is evident—they are not even acknowledged as a minority in some states. The Roma have a number of distinctive linguistic, ethnic, and religious characteristics. For example, most speak Romani, an Indo-European language descended from Sanskrit. Roma groups also have their own distinctive legal and court systems, which are group oriented rather than individual-rights oriented. That they have preserved their language, customs, and identity through centuries of persecution is evidence enough of their desire to preserve their culture.
200812_4-RC_3_21
[ "A definition that is vaguely formulated cannot serve as the basis for the provisions contained in a document of international law.", "A minority group's not being officially recognized as such by the government that has jurisdiction over it can be detrimental to the group's interests.", "Provisions in international law that apply only to minority groups should not be considered valid.", "Governments should recognize the legal and court systems used by minority populations within their jurisdictions.", "A group that often moves back and forth across a boundary between two countries can be legitimately considered citizens of both countries." ]
1
Which one of the following is a principle that can be most reasonably considered to underlie the reasoning in both of the passages?
During most of the nineteenth century, many French women continued to be educated according to models long established by custom and religious tradition. One recent observer has termed the failure to institute real and lasting educational reform at the end of the eighteenth century a "missed opportunity" —for in spite of the egalitarian and secular aims of the French Revolution in 1789, a truly nondiscriminatory education system for both women and men would not be established in the country until the 1880s. However, legislators had put forth many proposals for educational reform in the years just after the revolution; two in particular attempted to institute educational systems for women that were, to a great extent, egalitarian. The first of these proposals endeavored to replace the predominantly religious education that women originally received in convents and at home with reformed curricula. More importantly, the proposal insisted that, because education was a common good that should be offered to both sexes, instruction should be available to everyone. By the same token, teachers would be drawn from both sexes. Thus the proposal held it essential that schools for both men and women be established promptly throughout the country and that these schools be public, a tangible sign of the state's interest in all of its citizens. One limitation of this proposal, however, was that girls, unlike boys, were to leave school at age eight in order to be educated at home in the skills necessary for domestic life and for the raising of families. The second proposal took a more comprehensive approach. It advocated equal education for women and men on the grounds that women and men enjoy the same rights, and it was the only proposal of the time that called for coeducational schools, which were presented as a bulwark against the traditional gender roles enforced by religious tradition. In other respects, however, this proposal also continued to define women in terms of their roles in the domestic sphere and as mothers. That neither proposal was able to envision a system of education that was fully equal for women, and that neither was adopted into law even as such, bespeaks the immensity of the cultural and political obstacles to egalitarian education for women at the time. Nevertheless, the vision of egalitarian educational reform was not entirely lost. Nearly a century later, in the early 1880s, French legislators recalled the earlier proposals in their justification of new laws that founded public secondary schools for women, abolished fees for education, and established compulsory attendance for all students. In order to pass these reforms, the government needed to demonstrate that its new standards were rooted in a long philosophical, political, and pedagogical tradition. Various of the resulting institutions also made claim to revolutionary origin, as doing so allowed them to appropriate the legitimacy conferred by tradition and historical continuity.
200812_4-RC_4_22
[ "committed to removing education in the skills necessary for domestic life from the public school curriculum", "unaware of the difficulties that the earlier legislators faced when advocating similar legislation", "concerned with improving educational equality across economic strata as well as between the sexes", "more open to political compromise than were the legislators who introduced the previous proposals for reform", "more inclined to give religious authorities a role in education than were the legislators who introduced the previous proposals for reform" ]
2
It can be inferred from the passage that the French legislators who passed new educational laws in the early 1880s were
During most of the nineteenth century, many French women continued to be educated according to models long established by custom and religious tradition. One recent observer has termed the failure to institute real and lasting educational reform at the end of the eighteenth century a "missed opportunity" —for in spite of the egalitarian and secular aims of the French Revolution in 1789, a truly nondiscriminatory education system for both women and men would not be established in the country until the 1880s. However, legislators had put forth many proposals for educational reform in the years just after the revolution; two in particular attempted to institute educational systems for women that were, to a great extent, egalitarian. The first of these proposals endeavored to replace the predominantly religious education that women originally received in convents and at home with reformed curricula. More importantly, the proposal insisted that, because education was a common good that should be offered to both sexes, instruction should be available to everyone. By the same token, teachers would be drawn from both sexes. Thus the proposal held it essential that schools for both men and women be established promptly throughout the country and that these schools be public, a tangible sign of the state's interest in all of its citizens. One limitation of this proposal, however, was that girls, unlike boys, were to leave school at age eight in order to be educated at home in the skills necessary for domestic life and for the raising of families. The second proposal took a more comprehensive approach. It advocated equal education for women and men on the grounds that women and men enjoy the same rights, and it was the only proposal of the time that called for coeducational schools, which were presented as a bulwark against the traditional gender roles enforced by religious tradition. In other respects, however, this proposal also continued to define women in terms of their roles in the domestic sphere and as mothers. That neither proposal was able to envision a system of education that was fully equal for women, and that neither was adopted into law even as such, bespeaks the immensity of the cultural and political obstacles to egalitarian education for women at the time. Nevertheless, the vision of egalitarian educational reform was not entirely lost. Nearly a century later, in the early 1880s, French legislators recalled the earlier proposals in their justification of new laws that founded public secondary schools for women, abolished fees for education, and established compulsory attendance for all students. In order to pass these reforms, the government needed to demonstrate that its new standards were rooted in a long philosophical, political, and pedagogical tradition. Various of the resulting institutions also made claim to revolutionary origin, as doing so allowed them to appropriate the legitimacy conferred by tradition and historical continuity.
200812_4-RC_4_23
[ "Education in France during one historical period is described; two proposals that attempted to reform the educational system are presented; inconsistencies within each proposal are identified and lamented.", "The movement toward gender equality in France during one historical period is discussed; two proposals for educational reform are presented; the differences between the proposals and the educational system of that era are outlined.", "The traditional nature of French education for women is described; proposed breaks with tradition are discussed, followed by a discussion of why eventual change required less of a break with tradition.", "The egalitarian aims in France during one historical period are presented; proposals that foreshadowed eventual reform are described; the initial characterization of the aims is modified.", "The nature of education for women in France during one historical period is described; proposals for educational reform are presented; the relationship between the proposals and eventual reform is indicated." ]
4
Which one of the following most accurately describes the organization of the passage?
During most of the nineteenth century, many French women continued to be educated according to models long established by custom and religious tradition. One recent observer has termed the failure to institute real and lasting educational reform at the end of the eighteenth century a "missed opportunity" —for in spite of the egalitarian and secular aims of the French Revolution in 1789, a truly nondiscriminatory education system for both women and men would not be established in the country until the 1880s. However, legislators had put forth many proposals for educational reform in the years just after the revolution; two in particular attempted to institute educational systems for women that were, to a great extent, egalitarian. The first of these proposals endeavored to replace the predominantly religious education that women originally received in convents and at home with reformed curricula. More importantly, the proposal insisted that, because education was a common good that should be offered to both sexes, instruction should be available to everyone. By the same token, teachers would be drawn from both sexes. Thus the proposal held it essential that schools for both men and women be established promptly throughout the country and that these schools be public, a tangible sign of the state's interest in all of its citizens. One limitation of this proposal, however, was that girls, unlike boys, were to leave school at age eight in order to be educated at home in the skills necessary for domestic life and for the raising of families. The second proposal took a more comprehensive approach. It advocated equal education for women and men on the grounds that women and men enjoy the same rights, and it was the only proposal of the time that called for coeducational schools, which were presented as a bulwark against the traditional gender roles enforced by religious tradition. In other respects, however, this proposal also continued to define women in terms of their roles in the domestic sphere and as mothers. That neither proposal was able to envision a system of education that was fully equal for women, and that neither was adopted into law even as such, bespeaks the immensity of the cultural and political obstacles to egalitarian education for women at the time. Nevertheless, the vision of egalitarian educational reform was not entirely lost. Nearly a century later, in the early 1880s, French legislators recalled the earlier proposals in their justification of new laws that founded public secondary schools for women, abolished fees for education, and established compulsory attendance for all students. In order to pass these reforms, the government needed to demonstrate that its new standards were rooted in a long philosophical, political, and pedagogical tradition. Various of the resulting institutions also made claim to revolutionary origin, as doing so allowed them to appropriate the legitimacy conferred by tradition and historical continuity.
200812_4-RC_4_24
[ "\"Housing should be made available to all\" and \"Real estate practices should be nondiscriminatory\"", "\"Housing should be made available to all\" and \"The quality of housing should be improved\"", "\"There should be housing for all who can pay\" and \"Housing should be of uniform quality\"", "\"The quality of housing should be improved\" and \"Real estate practices should be nondiscriminatory\"", "\"Low-cost housing should be constructed\" and \"Housing should be of uniform quality\"" ]
0
Suppose that two proposals were put forward by lawmakers concerning housing reform today. Which one of the following pairs of proposals is most closely analogous to the pair of proposals discussed in the second paragraph of the passage?
During most of the nineteenth century, many French women continued to be educated according to models long established by custom and religious tradition. One recent observer has termed the failure to institute real and lasting educational reform at the end of the eighteenth century a "missed opportunity" —for in spite of the egalitarian and secular aims of the French Revolution in 1789, a truly nondiscriminatory education system for both women and men would not be established in the country until the 1880s. However, legislators had put forth many proposals for educational reform in the years just after the revolution; two in particular attempted to institute educational systems for women that were, to a great extent, egalitarian. The first of these proposals endeavored to replace the predominantly religious education that women originally received in convents and at home with reformed curricula. More importantly, the proposal insisted that, because education was a common good that should be offered to both sexes, instruction should be available to everyone. By the same token, teachers would be drawn from both sexes. Thus the proposal held it essential that schools for both men and women be established promptly throughout the country and that these schools be public, a tangible sign of the state's interest in all of its citizens. One limitation of this proposal, however, was that girls, unlike boys, were to leave school at age eight in order to be educated at home in the skills necessary for domestic life and for the raising of families. The second proposal took a more comprehensive approach. It advocated equal education for women and men on the grounds that women and men enjoy the same rights, and it was the only proposal of the time that called for coeducational schools, which were presented as a bulwark against the traditional gender roles enforced by religious tradition. In other respects, however, this proposal also continued to define women in terms of their roles in the domestic sphere and as mothers. That neither proposal was able to envision a system of education that was fully equal for women, and that neither was adopted into law even as such, bespeaks the immensity of the cultural and political obstacles to egalitarian education for women at the time. Nevertheless, the vision of egalitarian educational reform was not entirely lost. Nearly a century later, in the early 1880s, French legislators recalled the earlier proposals in their justification of new laws that founded public secondary schools for women, abolished fees for education, and established compulsory attendance for all students. In order to pass these reforms, the government needed to demonstrate that its new standards were rooted in a long philosophical, political, and pedagogical tradition. Various of the resulting institutions also made claim to revolutionary origin, as doing so allowed them to appropriate the legitimacy conferred by tradition and historical continuity.
200812_4-RC_4_25
[ "everyone should both learn and teach", "males and females should go to the same schools", "education should involve lifelong learning", "religious schools should be abolished", "education for girls should be both public and secular" ]
1
According to the passage, the second of the two proposals discussed was distinctive because it asserted that
During most of the nineteenth century, many French women continued to be educated according to models long established by custom and religious tradition. One recent observer has termed the failure to institute real and lasting educational reform at the end of the eighteenth century a "missed opportunity" —for in spite of the egalitarian and secular aims of the French Revolution in 1789, a truly nondiscriminatory education system for both women and men would not be established in the country until the 1880s. However, legislators had put forth many proposals for educational reform in the years just after the revolution; two in particular attempted to institute educational systems for women that were, to a great extent, egalitarian. The first of these proposals endeavored to replace the predominantly religious education that women originally received in convents and at home with reformed curricula. More importantly, the proposal insisted that, because education was a common good that should be offered to both sexes, instruction should be available to everyone. By the same token, teachers would be drawn from both sexes. Thus the proposal held it essential that schools for both men and women be established promptly throughout the country and that these schools be public, a tangible sign of the state's interest in all of its citizens. One limitation of this proposal, however, was that girls, unlike boys, were to leave school at age eight in order to be educated at home in the skills necessary for domestic life and for the raising of families. The second proposal took a more comprehensive approach. It advocated equal education for women and men on the grounds that women and men enjoy the same rights, and it was the only proposal of the time that called for coeducational schools, which were presented as a bulwark against the traditional gender roles enforced by religious tradition. In other respects, however, this proposal also continued to define women in terms of their roles in the domestic sphere and as mothers. That neither proposal was able to envision a system of education that was fully equal for women, and that neither was adopted into law even as such, bespeaks the immensity of the cultural and political obstacles to egalitarian education for women at the time. Nevertheless, the vision of egalitarian educational reform was not entirely lost. Nearly a century later, in the early 1880s, French legislators recalled the earlier proposals in their justification of new laws that founded public secondary schools for women, abolished fees for education, and established compulsory attendance for all students. In order to pass these reforms, the government needed to demonstrate that its new standards were rooted in a long philosophical, political, and pedagogical tradition. Various of the resulting institutions also made claim to revolutionary origin, as doing so allowed them to appropriate the legitimacy conferred by tradition and historical continuity.
200812_4-RC_4_26
[ "were a reaction to the excesses of the new government", "had their roots in a belief in the power of education", "had vast popular support within French society", "treated education for women as a prerequisite to the implementation of other reforms", "were influenced by egalitarian ideals" ]
4
Based on the passage, the fact that the proposed reforms were introduced shortly after the French Revolution most clearly suggests that the proposals
During most of the nineteenth century, many French women continued to be educated according to models long established by custom and religious tradition. One recent observer has termed the failure to institute real and lasting educational reform at the end of the eighteenth century a "missed opportunity" —for in spite of the egalitarian and secular aims of the French Revolution in 1789, a truly nondiscriminatory education system for both women and men would not be established in the country until the 1880s. However, legislators had put forth many proposals for educational reform in the years just after the revolution; two in particular attempted to institute educational systems for women that were, to a great extent, egalitarian. The first of these proposals endeavored to replace the predominantly religious education that women originally received in convents and at home with reformed curricula. More importantly, the proposal insisted that, because education was a common good that should be offered to both sexes, instruction should be available to everyone. By the same token, teachers would be drawn from both sexes. Thus the proposal held it essential that schools for both men and women be established promptly throughout the country and that these schools be public, a tangible sign of the state's interest in all of its citizens. One limitation of this proposal, however, was that girls, unlike boys, were to leave school at age eight in order to be educated at home in the skills necessary for domestic life and for the raising of families. The second proposal took a more comprehensive approach. It advocated equal education for women and men on the grounds that women and men enjoy the same rights, and it was the only proposal of the time that called for coeducational schools, which were presented as a bulwark against the traditional gender roles enforced by religious tradition. In other respects, however, this proposal also continued to define women in terms of their roles in the domestic sphere and as mothers. That neither proposal was able to envision a system of education that was fully equal for women, and that neither was adopted into law even as such, bespeaks the immensity of the cultural and political obstacles to egalitarian education for women at the time. Nevertheless, the vision of egalitarian educational reform was not entirely lost. Nearly a century later, in the early 1880s, French legislators recalled the earlier proposals in their justification of new laws that founded public secondary schools for women, abolished fees for education, and established compulsory attendance for all students. In order to pass these reforms, the government needed to demonstrate that its new standards were rooted in a long philosophical, political, and pedagogical tradition. Various of the resulting institutions also made claim to revolutionary origin, as doing so allowed them to appropriate the legitimacy conferred by tradition and historical continuity.
200812_4-RC_4_27
[ "They espoused reforms that were very modest by the standards of the day.", "They were fundamentally unethical due to their incomplete view of equality.", "They were well-meaning attempts to do as much as was feasible at the time.", "They were reasonable, and it is difficult to understand why they failed.", "They were not adopted because their aims were not fully comprehensive." ]
2
The author would most likely describe the proposals mentioned in the passage with which one of the following statements?
The United States government agency responsible for overseeing television and radio broadcasting, the Federal Communications Commission (FCC), had an early history of addressing only the concerns of parties with an economic interest in broadcasting—chiefly broadcasting companies. The rights of viewers and listeners were not recognized by the FCC, which regarded them merely as members of the public. Unless citizens' groups were applying for broadcasting licenses, citizens did not have the standing necessary to voice their views at an FCC hearing. Consequently, the FCC appeared to be exclusively at the service of the broadcasting industry. A landmark case changed the course of that history. In 1964, a local television station in Jackson, Mississippi was applying for a renewal of its broadcasting license. The United Church of Christ, representing Jackson's African American population, petitioned the FCC for a hearing about the broadcasting policies of that station. The church charged that the station advocated racial segregation to the point of excluding news and programs supporting integration. Arguing that the church lacked the level of economic interest required for a hearing, the FCC rejected the petition, though it attempted to mollify the church by granting only a short-term, probationary renewal to the station. Further, the FCC claimed that since it accepted the church's contentions with regard to misconduct on the part of the broadcasters, no hearing was necessary. However, that decision raised a question: If the contentions concerning the station were accepted, why was its license renewed at all? The real reason for denying the church a hearing was more likely the prospect that citizens' groups representing community preferences would begin to enter the closed worlds of government and industry. The church appealed the FCC's decision in court, and in 1967 was granted the right to a public hearing on the station's request for a long-term license. The hearing was to little avail: the FCC dismissed much of the public input and granted a full renewal to the station. The church appealed again, and this time the judge took the unprecedented step of revoking the station's license without remand to the FCC, ruling that the church members were performing a public service in voicing the legitimate concerns of the community and, as such, should be accorded the right to challenge the renewal of the station's broadcasting license. The case established a formidable precedent for opening up to the public the world of broadcasting. Subsequent rulings have supported the right of the public to question the performance of radio and television licensees before the FCC at renewal time every three years. Along with racial issues, a range of other matters—from the quality of children's programming and the portrayal of violence to equal time for opposing political viewpoints—are now discussed at licensing proceedings because of the church's intervention.
200906_4-RC_1_1
[ "Because of the efforts of a church group in challenging an FCC decision, public input is now considered in broadcast licensing proceedings.", "Court rulings have forced the FCC to abandon policies that appeared to encourage biased coverage of public issues.", "The history of the FCC is important because it explains why government agencies are now forced to respond to public input.", "Because it has begun to serve the interests of the public, the FCC is less responsive to the broadcasting industry.", "In response to pressure from citizens' groups, the FCC has decided to open its license renewal hearings to the public." ]
0
Which one of the following most accurately expresses the main point of the passage?
The United States government agency responsible for overseeing television and radio broadcasting, the Federal Communications Commission (FCC), had an early history of addressing only the concerns of parties with an economic interest in broadcasting—chiefly broadcasting companies. The rights of viewers and listeners were not recognized by the FCC, which regarded them merely as members of the public. Unless citizens' groups were applying for broadcasting licenses, citizens did not have the standing necessary to voice their views at an FCC hearing. Consequently, the FCC appeared to be exclusively at the service of the broadcasting industry. A landmark case changed the course of that history. In 1964, a local television station in Jackson, Mississippi was applying for a renewal of its broadcasting license. The United Church of Christ, representing Jackson's African American population, petitioned the FCC for a hearing about the broadcasting policies of that station. The church charged that the station advocated racial segregation to the point of excluding news and programs supporting integration. Arguing that the church lacked the level of economic interest required for a hearing, the FCC rejected the petition, though it attempted to mollify the church by granting only a short-term, probationary renewal to the station. Further, the FCC claimed that since it accepted the church's contentions with regard to misconduct on the part of the broadcasters, no hearing was necessary. However, that decision raised a question: If the contentions concerning the station were accepted, why was its license renewed at all? The real reason for denying the church a hearing was more likely the prospect that citizens' groups representing community preferences would begin to enter the closed worlds of government and industry. The church appealed the FCC's decision in court, and in 1967 was granted the right to a public hearing on the station's request for a long-term license. The hearing was to little avail: the FCC dismissed much of the public input and granted a full renewal to the station. The church appealed again, and this time the judge took the unprecedented step of revoking the station's license without remand to the FCC, ruling that the church members were performing a public service in voicing the legitimate concerns of the community and, as such, should be accorded the right to challenge the renewal of the station's broadcasting license. The case established a formidable precedent for opening up to the public the world of broadcasting. Subsequent rulings have supported the right of the public to question the performance of radio and television licensees before the FCC at renewal time every three years. Along with racial issues, a range of other matters—from the quality of children's programming and the portrayal of violence to equal time for opposing political viewpoints—are now discussed at licensing proceedings because of the church's intervention.
200906_4-RC_1_2
[ "support the author's claim that the case helped to open up to the public the world of broadcasting", "suggest the level of vigilance that citizens' groups must maintain with regard to broadcasters", "provide an explanation of why the public is allowed to question the performance of broadcasters on such a frequent basis", "illustrate other areas of misconduct with which the station discussed in the passage was charged", "demonstrate that the station discussed in the passage was not the only one to fall short of its obligation to the public" ]
0
The author mentions some additional topics now discussed at FCC hearings (lines 54–59) primarily in order to
The United States government agency responsible for overseeing television and radio broadcasting, the Federal Communications Commission (FCC), had an early history of addressing only the concerns of parties with an economic interest in broadcasting—chiefly broadcasting companies. The rights of viewers and listeners were not recognized by the FCC, which regarded them merely as members of the public. Unless citizens' groups were applying for broadcasting licenses, citizens did not have the standing necessary to voice their views at an FCC hearing. Consequently, the FCC appeared to be exclusively at the service of the broadcasting industry. A landmark case changed the course of that history. In 1964, a local television station in Jackson, Mississippi was applying for a renewal of its broadcasting license. The United Church of Christ, representing Jackson's African American population, petitioned the FCC for a hearing about the broadcasting policies of that station. The church charged that the station advocated racial segregation to the point of excluding news and programs supporting integration. Arguing that the church lacked the level of economic interest required for a hearing, the FCC rejected the petition, though it attempted to mollify the church by granting only a short-term, probationary renewal to the station. Further, the FCC claimed that since it accepted the church's contentions with regard to misconduct on the part of the broadcasters, no hearing was necessary. However, that decision raised a question: If the contentions concerning the station were accepted, why was its license renewed at all? The real reason for denying the church a hearing was more likely the prospect that citizens' groups representing community preferences would begin to enter the closed worlds of government and industry. The church appealed the FCC's decision in court, and in 1967 was granted the right to a public hearing on the station's request for a long-term license. The hearing was to little avail: the FCC dismissed much of the public input and granted a full renewal to the station. The church appealed again, and this time the judge took the unprecedented step of revoking the station's license without remand to the FCC, ruling that the church members were performing a public service in voicing the legitimate concerns of the community and, as such, should be accorded the right to challenge the renewal of the station's broadcasting license. The case established a formidable precedent for opening up to the public the world of broadcasting. Subsequent rulings have supported the right of the public to question the performance of radio and television licensees before the FCC at renewal time every three years. Along with racial issues, a range of other matters—from the quality of children's programming and the portrayal of violence to equal time for opposing political viewpoints—are now discussed at licensing proceedings because of the church's intervention.
200906_4-RC_1_3
[ "The broadcasting industry's economic goals can be met most easily by minimizing the attention given to the interests of viewers and listeners.", "The FCC was advised by broadcasters to bar groups with no economic interest in broadcasting from hearings concerning the broadcasting industry.", "The court ruled in the case brought by the United Church of Christ that the FCC had the ultimate authority to decide whether to renew a broadcaster's license.", "Before the United Church of Christ won its case, the FCC would not allow citizens' groups to speak as members of the public at FCC hearings.", "The case brought by the United Church of Christ represents the first time a citizens' group was successful in getting its concerns about government agencies addressed to its satisfaction." ]
3
Which one of the following statements is affirmed by the passage?
The United States government agency responsible for overseeing television and radio broadcasting, the Federal Communications Commission (FCC), had an early history of addressing only the concerns of parties with an economic interest in broadcasting—chiefly broadcasting companies. The rights of viewers and listeners were not recognized by the FCC, which regarded them merely as members of the public. Unless citizens' groups were applying for broadcasting licenses, citizens did not have the standing necessary to voice their views at an FCC hearing. Consequently, the FCC appeared to be exclusively at the service of the broadcasting industry. A landmark case changed the course of that history. In 1964, a local television station in Jackson, Mississippi was applying for a renewal of its broadcasting license. The United Church of Christ, representing Jackson's African American population, petitioned the FCC for a hearing about the broadcasting policies of that station. The church charged that the station advocated racial segregation to the point of excluding news and programs supporting integration. Arguing that the church lacked the level of economic interest required for a hearing, the FCC rejected the petition, though it attempted to mollify the church by granting only a short-term, probationary renewal to the station. Further, the FCC claimed that since it accepted the church's contentions with regard to misconduct on the part of the broadcasters, no hearing was necessary. However, that decision raised a question: If the contentions concerning the station were accepted, why was its license renewed at all? The real reason for denying the church a hearing was more likely the prospect that citizens' groups representing community preferences would begin to enter the closed worlds of government and industry. The church appealed the FCC's decision in court, and in 1967 was granted the right to a public hearing on the station's request for a long-term license. The hearing was to little avail: the FCC dismissed much of the public input and granted a full renewal to the station. The church appealed again, and this time the judge took the unprecedented step of revoking the station's license without remand to the FCC, ruling that the church members were performing a public service in voicing the legitimate concerns of the community and, as such, should be accorded the right to challenge the renewal of the station's broadcasting license. The case established a formidable precedent for opening up to the public the world of broadcasting. Subsequent rulings have supported the right of the public to question the performance of radio and television licensees before the FCC at renewal time every three years. Along with racial issues, a range of other matters—from the quality of children's programming and the portrayal of violence to equal time for opposing political viewpoints—are now discussed at licensing proceedings because of the church's intervention.
200906_4-RC_1_4
[ "If the United Church of Christ had not pursued its case, the FCC would not have been aware of the television station's broadcasting policies.", "By their very nature, industrial and business interests are opposed to public interests.", "The recourse of a citizens' group to the courts represents an effective means of protecting public interests.", "Governmental regulation cannot safeguard against individual businesses acting contrary to public interests.", "The government cannot be trusted to favor the rights of the public over broadcasters' economic interests." ]
2
Based on information presented in the passage, with which one of the following statements would the author be most likely to agree?
The United States government agency responsible for overseeing television and radio broadcasting, the Federal Communications Commission (FCC), had an early history of addressing only the concerns of parties with an economic interest in broadcasting—chiefly broadcasting companies. The rights of viewers and listeners were not recognized by the FCC, which regarded them merely as members of the public. Unless citizens' groups were applying for broadcasting licenses, citizens did not have the standing necessary to voice their views at an FCC hearing. Consequently, the FCC appeared to be exclusively at the service of the broadcasting industry. A landmark case changed the course of that history. In 1964, a local television station in Jackson, Mississippi was applying for a renewal of its broadcasting license. The United Church of Christ, representing Jackson's African American population, petitioned the FCC for a hearing about the broadcasting policies of that station. The church charged that the station advocated racial segregation to the point of excluding news and programs supporting integration. Arguing that the church lacked the level of economic interest required for a hearing, the FCC rejected the petition, though it attempted to mollify the church by granting only a short-term, probationary renewal to the station. Further, the FCC claimed that since it accepted the church's contentions with regard to misconduct on the part of the broadcasters, no hearing was necessary. However, that decision raised a question: If the contentions concerning the station were accepted, why was its license renewed at all? The real reason for denying the church a hearing was more likely the prospect that citizens' groups representing community preferences would begin to enter the closed worlds of government and industry. The church appealed the FCC's decision in court, and in 1967 was granted the right to a public hearing on the station's request for a long-term license. The hearing was to little avail: the FCC dismissed much of the public input and granted a full renewal to the station. The church appealed again, and this time the judge took the unprecedented step of revoking the station's license without remand to the FCC, ruling that the church members were performing a public service in voicing the legitimate concerns of the community and, as such, should be accorded the right to challenge the renewal of the station's broadcasting license. The case established a formidable precedent for opening up to the public the world of broadcasting. Subsequent rulings have supported the right of the public to question the performance of radio and television licensees before the FCC at renewal time every three years. Along with racial issues, a range of other matters—from the quality of children's programming and the portrayal of violence to equal time for opposing political viewpoints—are now discussed at licensing proceedings because of the church's intervention.
200906_4-RC_1_5
[ "Broadcasters are legally obligated to hold regular meetings at which the public can voice its concerns about broadcasting policies.", "Broadcasters are now required by the FCC to consult citizens' groups when making programming decisions.", "Except in cases involving clear misconduct by a broadcaster, the FCC need not seek public input in licensing hearings.", "When evaluating the performance of a broadcaster applying for a license renewal, the FCC must obtain information about the preferences of the public.", "In FCC licensing proceedings, parties representing community preferences should be granted standing along with those with an economic interest in broadcasting." ]
4
The passage suggests that which one of the following has been established by the case discussed in the third paragraph?
An effort should be made to dispel the misunderstandings that still prevent the much-needed synthesis and mutual supplementation of science and the humanities. This reconciliation should not be too difficult once it is recognized that the separation is primarily the result of a basic misunderstanding of the philosophical foundations of both science and the humanities. Some humanists still identify science with an absurd mechanistic reductionism. There are many who feel that the scientist is interested in nothing more than "bodies in motion," in the strictly mathematical, physical, and chemical laws that govern the material world. This is the caricature of science drawn by representatives of the humanities who are ignorant of the nature of modern science and also of the scientific outlook in philosophy. For example, it is claimed that science either ignores or explains away the most essential human values. Those who believe this also assert that there are aspects of the human mind, manifest especially in the domains of morality, religion, and the arts, that contain an irreducible spiritual element and for that reason can never be adequately explained by science. Some scientists, on the other hand, claim that the humanist is interested in nothing more than emotion and sentiment, exhibiting the vagrant fancies of an undisciplined mind. To such men and women the humanities are useless because they serve no immediate and technological function for the practical survival of human society in the material world. Such pragmatists believe that the areas of morality, religion, and the arts should have only a secondary importance in people's lives. Thus there are misconceptions among humanists and scientists alike that are in need of correction. This correction leads to a much more acceptable position that could be called "scientific humanism," attempting as it does to combine the common elements of both disciplines. Both science and the humanities attempt to describe and explain. It is true that they begin their descriptions and explanations at widely separated points, but the objectives remain the same: a clearer understanding of people and their world. In achieving this understanding, science in fact does not depend exclusively on measurable data, and the humanities in fact profit from attempts at controlled evaluation. Scientific humanism can combine the scientific attitude with an active interest in the whole scale of human values. If uninformed persons insist on viewing science as only materialistic and the humanities as only idealistic, a fruitful collaboration of both fields is unlikely. The combination of science and the humanities is, however, possible, even probable, if we begin by noting their common objectives, rather than seeing only their different means.
200906_4-RC_2_6
[ "Scientists' failure to understand humanists hinders collaborations between the two groups.", "The materialism of science and the idealism of the humanities have both been beneficial to modern society.", "Technological development will cease if science and the humanities remain at odds with each other.", "The current relationship between science and the humanities is less cooperative than their relationship once was.", "A synthesis of science and the humanities is possible and much-needed." ]
4
Which one of the following best describes the main idea of the passage?
An effort should be made to dispel the misunderstandings that still prevent the much-needed synthesis and mutual supplementation of science and the humanities. This reconciliation should not be too difficult once it is recognized that the separation is primarily the result of a basic misunderstanding of the philosophical foundations of both science and the humanities. Some humanists still identify science with an absurd mechanistic reductionism. There are many who feel that the scientist is interested in nothing more than "bodies in motion," in the strictly mathematical, physical, and chemical laws that govern the material world. This is the caricature of science drawn by representatives of the humanities who are ignorant of the nature of modern science and also of the scientific outlook in philosophy. For example, it is claimed that science either ignores or explains away the most essential human values. Those who believe this also assert that there are aspects of the human mind, manifest especially in the domains of morality, religion, and the arts, that contain an irreducible spiritual element and for that reason can never be adequately explained by science. Some scientists, on the other hand, claim that the humanist is interested in nothing more than emotion and sentiment, exhibiting the vagrant fancies of an undisciplined mind. To such men and women the humanities are useless because they serve no immediate and technological function for the practical survival of human society in the material world. Such pragmatists believe that the areas of morality, religion, and the arts should have only a secondary importance in people's lives. Thus there are misconceptions among humanists and scientists alike that are in need of correction. This correction leads to a much more acceptable position that could be called "scientific humanism," attempting as it does to combine the common elements of both disciplines. Both science and the humanities attempt to describe and explain. It is true that they begin their descriptions and explanations at widely separated points, but the objectives remain the same: a clearer understanding of people and their world. In achieving this understanding, science in fact does not depend exclusively on measurable data, and the humanities in fact profit from attempts at controlled evaluation. Scientific humanism can combine the scientific attitude with an active interest in the whole scale of human values. If uninformed persons insist on viewing science as only materialistic and the humanities as only idealistic, a fruitful collaboration of both fields is unlikely. The combination of science and the humanities is, however, possible, even probable, if we begin by noting their common objectives, rather than seeing only their different means.
200906_4-RC_2_7
[ "Science encourages the view that emotions are inexplicable.", "Science arises out of practical needs but serves other needs as well.", "Science depends exclusively on measurable data to support its claims.", "Science recognizes an irreducible spiritual element that makes the arts inexplicable.", "Science encourages the use of description in the study of human values." ]
2
Which one of the following would the author be most likely to characterize as an example of a misunderstanding of science by a humanist?
An effort should be made to dispel the misunderstandings that still prevent the much-needed synthesis and mutual supplementation of science and the humanities. This reconciliation should not be too difficult once it is recognized that the separation is primarily the result of a basic misunderstanding of the philosophical foundations of both science and the humanities. Some humanists still identify science with an absurd mechanistic reductionism. There are many who feel that the scientist is interested in nothing more than "bodies in motion," in the strictly mathematical, physical, and chemical laws that govern the material world. This is the caricature of science drawn by representatives of the humanities who are ignorant of the nature of modern science and also of the scientific outlook in philosophy. For example, it is claimed that science either ignores or explains away the most essential human values. Those who believe this also assert that there are aspects of the human mind, manifest especially in the domains of morality, religion, and the arts, that contain an irreducible spiritual element and for that reason can never be adequately explained by science. Some scientists, on the other hand, claim that the humanist is interested in nothing more than emotion and sentiment, exhibiting the vagrant fancies of an undisciplined mind. To such men and women the humanities are useless because they serve no immediate and technological function for the practical survival of human society in the material world. Such pragmatists believe that the areas of morality, religion, and the arts should have only a secondary importance in people's lives. Thus there are misconceptions among humanists and scientists alike that are in need of correction. This correction leads to a much more acceptable position that could be called "scientific humanism," attempting as it does to combine the common elements of both disciplines. Both science and the humanities attempt to describe and explain. It is true that they begin their descriptions and explanations at widely separated points, but the objectives remain the same: a clearer understanding of people and their world. In achieving this understanding, science in fact does not depend exclusively on measurable data, and the humanities in fact profit from attempts at controlled evaluation. Scientific humanism can combine the scientific attitude with an active interest in the whole scale of human values. If uninformed persons insist on viewing science as only materialistic and the humanities as only idealistic, a fruitful collaboration of both fields is unlikely. The combination of science and the humanities is, however, possible, even probable, if we begin by noting their common objectives, rather than seeing only their different means.
200906_4-RC_2_8
[ "Scientific humanism is characterized by the extension of description and explanation from science to the humanities.", "A clearer understanding of people is an objective of humanists that scientists have not yet come to share.", "Controlled measures of aesthetic experience are of little use in the study of the humanities.", "Humanists have profited from using methods generally considered useful primarily to scientists.", "Fruitful collaboration between scientists and humanists is unlikely to become more common." ]
3
It can be inferred from the passage that the author would be most likely to agree with which one of the following statements?
An effort should be made to dispel the misunderstandings that still prevent the much-needed synthesis and mutual supplementation of science and the humanities. This reconciliation should not be too difficult once it is recognized that the separation is primarily the result of a basic misunderstanding of the philosophical foundations of both science and the humanities. Some humanists still identify science with an absurd mechanistic reductionism. There are many who feel that the scientist is interested in nothing more than "bodies in motion," in the strictly mathematical, physical, and chemical laws that govern the material world. This is the caricature of science drawn by representatives of the humanities who are ignorant of the nature of modern science and also of the scientific outlook in philosophy. For example, it is claimed that science either ignores or explains away the most essential human values. Those who believe this also assert that there are aspects of the human mind, manifest especially in the domains of morality, religion, and the arts, that contain an irreducible spiritual element and for that reason can never be adequately explained by science. Some scientists, on the other hand, claim that the humanist is interested in nothing more than emotion and sentiment, exhibiting the vagrant fancies of an undisciplined mind. To such men and women the humanities are useless because they serve no immediate and technological function for the practical survival of human society in the material world. Such pragmatists believe that the areas of morality, religion, and the arts should have only a secondary importance in people's lives. Thus there are misconceptions among humanists and scientists alike that are in need of correction. This correction leads to a much more acceptable position that could be called "scientific humanism," attempting as it does to combine the common elements of both disciplines. Both science and the humanities attempt to describe and explain. It is true that they begin their descriptions and explanations at widely separated points, but the objectives remain the same: a clearer understanding of people and their world. In achieving this understanding, science in fact does not depend exclusively on measurable data, and the humanities in fact profit from attempts at controlled evaluation. Scientific humanism can combine the scientific attitude with an active interest in the whole scale of human values. If uninformed persons insist on viewing science as only materialistic and the humanities as only idealistic, a fruitful collaboration of both fields is unlikely. The combination of science and the humanities is, however, possible, even probable, if we begin by noting their common objectives, rather than seeing only their different means.
200906_4-RC_2_9
[ "inflammatory claims by scientists regarding the pragmatic value of the work of humanists", "misunderstandings of the philosophical foundations of each by the other", "the excessive influence of reductionism on both", "the predominance of a concern with mechanics in science", "the failure of humanists to develop rigorous methods" ]
1
According to the author, which one of the following is the primary cause of the existing separation between science and the humanities?
An effort should be made to dispel the misunderstandings that still prevent the much-needed synthesis and mutual supplementation of science and the humanities. This reconciliation should not be too difficult once it is recognized that the separation is primarily the result of a basic misunderstanding of the philosophical foundations of both science and the humanities. Some humanists still identify science with an absurd mechanistic reductionism. There are many who feel that the scientist is interested in nothing more than "bodies in motion," in the strictly mathematical, physical, and chemical laws that govern the material world. This is the caricature of science drawn by representatives of the humanities who are ignorant of the nature of modern science and also of the scientific outlook in philosophy. For example, it is claimed that science either ignores or explains away the most essential human values. Those who believe this also assert that there are aspects of the human mind, manifest especially in the domains of morality, religion, and the arts, that contain an irreducible spiritual element and for that reason can never be adequately explained by science. Some scientists, on the other hand, claim that the humanist is interested in nothing more than emotion and sentiment, exhibiting the vagrant fancies of an undisciplined mind. To such men and women the humanities are useless because they serve no immediate and technological function for the practical survival of human society in the material world. Such pragmatists believe that the areas of morality, religion, and the arts should have only a secondary importance in people's lives. Thus there are misconceptions among humanists and scientists alike that are in need of correction. This correction leads to a much more acceptable position that could be called "scientific humanism," attempting as it does to combine the common elements of both disciplines. Both science and the humanities attempt to describe and explain. It is true that they begin their descriptions and explanations at widely separated points, but the objectives remain the same: a clearer understanding of people and their world. In achieving this understanding, science in fact does not depend exclusively on measurable data, and the humanities in fact profit from attempts at controlled evaluation. Scientific humanism can combine the scientific attitude with an active interest in the whole scale of human values. If uninformed persons insist on viewing science as only materialistic and the humanities as only idealistic, a fruitful collaboration of both fields is unlikely. The combination of science and the humanities is, however, possible, even probable, if we begin by noting their common objectives, rather than seeing only their different means.
200906_4-RC_2_10
[ "to show that a proposal introduced in the first paragraph is implausible because of information presented in the second and third paragraphs", "to show that the views presented in the second and third paragraphs are correct but capable of reconciliation", "to present information supporting one of two opposing views presented in the second and third paragraphs", "to present an alternative to views presented in the second and third paragraphs", "to offer specific examples of the distinct views presented in the second and third paragraphs" ]
3
Which one of the following best describes one of the functions of the last paragraph in the passage?
An effort should be made to dispel the misunderstandings that still prevent the much-needed synthesis and mutual supplementation of science and the humanities. This reconciliation should not be too difficult once it is recognized that the separation is primarily the result of a basic misunderstanding of the philosophical foundations of both science and the humanities. Some humanists still identify science with an absurd mechanistic reductionism. There are many who feel that the scientist is interested in nothing more than "bodies in motion," in the strictly mathematical, physical, and chemical laws that govern the material world. This is the caricature of science drawn by representatives of the humanities who are ignorant of the nature of modern science and also of the scientific outlook in philosophy. For example, it is claimed that science either ignores or explains away the most essential human values. Those who believe this also assert that there are aspects of the human mind, manifest especially in the domains of morality, religion, and the arts, that contain an irreducible spiritual element and for that reason can never be adequately explained by science. Some scientists, on the other hand, claim that the humanist is interested in nothing more than emotion and sentiment, exhibiting the vagrant fancies of an undisciplined mind. To such men and women the humanities are useless because they serve no immediate and technological function for the practical survival of human society in the material world. Such pragmatists believe that the areas of morality, religion, and the arts should have only a secondary importance in people's lives. Thus there are misconceptions among humanists and scientists alike that are in need of correction. This correction leads to a much more acceptable position that could be called "scientific humanism," attempting as it does to combine the common elements of both disciplines. Both science and the humanities attempt to describe and explain. It is true that they begin their descriptions and explanations at widely separated points, but the objectives remain the same: a clearer understanding of people and their world. In achieving this understanding, science in fact does not depend exclusively on measurable data, and the humanities in fact profit from attempts at controlled evaluation. Scientific humanism can combine the scientific attitude with an active interest in the whole scale of human values. If uninformed persons insist on viewing science as only materialistic and the humanities as only idealistic, a fruitful collaboration of both fields is unlikely. The combination of science and the humanities is, however, possible, even probable, if we begin by noting their common objectives, rather than seeing only their different means.
200906_4-RC_2_11
[ "a realization that the scientist is less interested in describing \"bodies in motion\" than in constructing mathematical models of the material world", "an acknowledgement that there is a spiritual element in the arts that science does not account for", "an acceptance of the application of controlled evaluation to the examination of human values", "a less strident insistence on the primary importance of the arts in people's lives", "an emphasis on developing ways for showing how the humanities support the practical survival of mankind" ]
2
The passage suggests that the author would recommend that humanists accept which one of the following modifications of their point of view?
An effort should be made to dispel the misunderstandings that still prevent the much-needed synthesis and mutual supplementation of science and the humanities. This reconciliation should not be too difficult once it is recognized that the separation is primarily the result of a basic misunderstanding of the philosophical foundations of both science and the humanities. Some humanists still identify science with an absurd mechanistic reductionism. There are many who feel that the scientist is interested in nothing more than "bodies in motion," in the strictly mathematical, physical, and chemical laws that govern the material world. This is the caricature of science drawn by representatives of the humanities who are ignorant of the nature of modern science and also of the scientific outlook in philosophy. For example, it is claimed that science either ignores or explains away the most essential human values. Those who believe this also assert that there are aspects of the human mind, manifest especially in the domains of morality, religion, and the arts, that contain an irreducible spiritual element and for that reason can never be adequately explained by science. Some scientists, on the other hand, claim that the humanist is interested in nothing more than emotion and sentiment, exhibiting the vagrant fancies of an undisciplined mind. To such men and women the humanities are useless because they serve no immediate and technological function for the practical survival of human society in the material world. Such pragmatists believe that the areas of morality, religion, and the arts should have only a secondary importance in people's lives. Thus there are misconceptions among humanists and scientists alike that are in need of correction. This correction leads to a much more acceptable position that could be called "scientific humanism," attempting as it does to combine the common elements of both disciplines. Both science and the humanities attempt to describe and explain. It is true that they begin their descriptions and explanations at widely separated points, but the objectives remain the same: a clearer understanding of people and their world. In achieving this understanding, science in fact does not depend exclusively on measurable data, and the humanities in fact profit from attempts at controlled evaluation. Scientific humanism can combine the scientific attitude with an active interest in the whole scale of human values. If uninformed persons insist on viewing science as only materialistic and the humanities as only idealistic, a fruitful collaboration of both fields is unlikely. The combination of science and the humanities is, however, possible, even probable, if we begin by noting their common objectives, rather than seeing only their different means.
200906_4-RC_2_12
[ "wildly emotional", "excessively impractical", "unnecessarily intransigent", "justifiably optimistic", "logically inconsistent" ]
1
In using the phrase "vagrant fancies of an undisciplined mind" (lines 27–28), the author suggests that humanists are sometimes considered to be
Passage A When Cather gave examples of high quality in fiction, she cited Russian writers Ivan Turgenev or Leo Tolstoy or both. Indeed, Edmund Wilson noted in 1922 that Cather followed the manner of Turgenev, not depicting her characters' emotions directly but telling us how they behave and letting their "inner blaze of glory shine through the simple recital." Turgenev's method was to select details that described a character's appearance and actions without trying to explain them. A writer, he said, "must be a psychologist—but a secret one; he must know and feel the roots of phenomena, but only present the phenomena themselves." Similarly, he argued that a writer must have complete knowledge of a character so as to avoid overloading the work with unnecessary detail, concentrating instead on what is characteristic and typical. Here we have an impressionistic aesthetic that anticipates Cather's: what Turgenev referred to as secret knowledge Cather called "the thing not named." In one essay she writes that "whatever is felt upon the page without being specifically named there—that, one might say, is created." For both writers, there is the absolute importance of selection and simplification; for both, art is the fusing of the physical world of setting and actions with the emotional reality of the characters. What synthesizes all the elements of narrative for these writers is the establishment of a prevailing mood. Passage B In a famous 1927 letter, Cather writes of her novel Death Comes for the Archbishop, "Many [reviewers] assert vehemently that it is not a novel. Myself, I prefer to call it a narrative." Cather's preference anticipated an important reformulation of the criticism of fiction: the body of literary theory, called "narratology," articulated by French literary theorists in the 1960s. This approach broadens and simplifies the fundamental paradigms according to which we view fiction: they ask of narrative only that it be narrative, that it in fiction, she invariably cited Russian writers Ivan Turgenev or Leo Tolstoy or both. Indeed, Edmund Wilson noted in 1922 that Cather followed the manner of Turgenev, not depicting her characters' emotions directly but telling us how they behave and letting their "inner blaze of glory shine through the simple recital." Turgenev's method was to select details that described a character's appearance and actions without trying to explain them. A writer, he said, "must be a psychologist—but a secret one; he must know and feel the roots of phenomena, but only present the phenomena themselves." Similarly, he argued that a writer must have complete knowledge of a character so as to avoid overloading the work with unnecessary detail, concentrating instead on what is characteristic and typical. Here we have an impressionistic aesthetic that anticipates Cather's: what Turgenev referred to as secret knowledge Cather called "the thing not named." In one essay she writes that "whatever is felt upon the page without being specifically named there—that, one might say, is created." For both writers, there is the absolute importance of selection and simplification; for both, art is the fusing of the physical world of setting and actions with the emotional reality of the characters. What synthesizes all the elements of narrative for these writers is the establishment of a prevailing mood. Passage B In a famous 1927 letter, Cather writes of her novel Death Comes for the Archbishop, "Many [reviewers] assert vehemently that it is not a novel. Myself, I prefer to call it a narrative." Cather's preference anticipated an important reformulation of the criticism of fiction: the body of literary theory, called "narratology," articulated by French literary theorists in the 1960s. This approach broadens and simplifies the fundamental paradigms according to which we view fiction: they ask of narrative only that it be narrative, that it tell a story. Narratologists tend not to focus on the characteristics of narrative's dominant modern Western form, the "realistic novel" : direct psychological characterization, realistic treatment of time, causal plotting, logical closure. Such a model of criticism, which takes as its object "narrative" rather than the "novel," seems exactly appropriate to Cather's work. Indeed, her severest critics have always questioned precisely her capabilities as a novelist. Morton Zabel argued that "[Cather's] themes...could readily fail to find the structure and substance that might have given them life or redeemed them from the tenuity of a sketch" ; Leon Edel called one of her novels "two inconclusive fragments." These critics and others like them treat as failures some of the central features of Cather's impressionistic technique: unusual treatment of narrative time, unexpected focus, ambiguous conclusions, a preference for the bold, simple, and stylized in character as well as in landscape. These "non-novelistic" structures indirectly articulate the essential and conflicting forces of desire at work throughout Cather's fiction.
200906_4-RC_3_13
[ "Though Cather preferred to call Death Comes for the Archbishop a narrative rather than a novel, she would be unlikely to view most of her other novels in the same way.", "The critics who questioned Cather's abilities as a novelist focused mostly on her failed experiments and ignored her more aesthetically successful novels.", "A model of criticism that takes narrative rather than the novel as its object is likely to result in flawed interpretations of Cather's work.", "Critics who questioned Cather's abilities as a novelist fail to perceive the extent to which Cather actually embraced the conventions of the realistic novel.", "Cather's goal of representing the \"thing not named\" explains her preference for the bold, simple, and stylized in the presentation of character." ]
4
If the author of passage A were to read passage B, he or she would be most likely to agree with which one of the following?
Passage A When Cather gave examples of high quality in fiction, she cited Russian writers Ivan Turgenev or Leo Tolstoy or both. Indeed, Edmund Wilson noted in 1922 that Cather followed the manner of Turgenev, not depicting her characters' emotions directly but telling us how they behave and letting their "inner blaze of glory shine through the simple recital." Turgenev's method was to select details that described a character's appearance and actions without trying to explain them. A writer, he said, "must be a psychologist—but a secret one; he must know and feel the roots of phenomena, but only present the phenomena themselves." Similarly, he argued that a writer must have complete knowledge of a character so as to avoid overloading the work with unnecessary detail, concentrating instead on what is characteristic and typical. Here we have an impressionistic aesthetic that anticipates Cather's: what Turgenev referred to as secret knowledge Cather called "the thing not named." In one essay she writes that "whatever is felt upon the page without being specifically named there—that, one might say, is created." For both writers, there is the absolute importance of selection and simplification; for both, art is the fusing of the physical world of setting and actions with the emotional reality of the characters. What synthesizes all the elements of narrative for these writers is the establishment of a prevailing mood. Passage B In a famous 1927 letter, Cather writes of her novel Death Comes for the Archbishop, "Many [reviewers] assert vehemently that it is not a novel. Myself, I prefer to call it a narrative." Cather's preference anticipated an important reformulation of the criticism of fiction: the body of literary theory, called "narratology," articulated by French literary theorists in the 1960s. This approach broadens and simplifies the fundamental paradigms according to which we view fiction: they ask of narrative only that it be narrative, that it in fiction, she invariably cited Russian writers Ivan Turgenev or Leo Tolstoy or both. Indeed, Edmund Wilson noted in 1922 that Cather followed the manner of Turgenev, not depicting her characters' emotions directly but telling us how they behave and letting their "inner blaze of glory shine through the simple recital." Turgenev's method was to select details that described a character's appearance and actions without trying to explain them. A writer, he said, "must be a psychologist—but a secret one; he must know and feel the roots of phenomena, but only present the phenomena themselves." Similarly, he argued that a writer must have complete knowledge of a character so as to avoid overloading the work with unnecessary detail, concentrating instead on what is characteristic and typical. Here we have an impressionistic aesthetic that anticipates Cather's: what Turgenev referred to as secret knowledge Cather called "the thing not named." In one essay she writes that "whatever is felt upon the page without being specifically named there—that, one might say, is created." For both writers, there is the absolute importance of selection and simplification; for both, art is the fusing of the physical world of setting and actions with the emotional reality of the characters. What synthesizes all the elements of narrative for these writers is the establishment of a prevailing mood. Passage B In a famous 1927 letter, Cather writes of her novel Death Comes for the Archbishop, "Many [reviewers] assert vehemently that it is not a novel. Myself, I prefer to call it a narrative." Cather's preference anticipated an important reformulation of the criticism of fiction: the body of literary theory, called "narratology," articulated by French literary theorists in the 1960s. This approach broadens and simplifies the fundamental paradigms according to which we view fiction: they ask of narrative only that it be narrative, that it tell a story. Narratologists tend not to focus on the characteristics of narrative's dominant modern Western form, the "realistic novel" : direct psychological characterization, realistic treatment of time, causal plotting, logical closure. Such a model of criticism, which takes as its object "narrative" rather than the "novel," seems exactly appropriate to Cather's work. Indeed, her severest critics have always questioned precisely her capabilities as a novelist. Morton Zabel argued that "[Cather's] themes...could readily fail to find the structure and substance that might have given them life or redeemed them from the tenuity of a sketch" ; Leon Edel called one of her novels "two inconclusive fragments." These critics and others like them treat as failures some of the central features of Cather's impressionistic technique: unusual treatment of narrative time, unexpected focus, ambiguous conclusions, a preference for the bold, simple, and stylized in character as well as in landscape. These "non-novelistic" structures indirectly articulate the essential and conflicting forces of desire at work throughout Cather's fiction.
200906_4-RC_3_14
[ "Narratologists point to Cather's works as prime examples of pure narrative.", "Cather disliked the work of many of the novelists who preceded her.", "Cather regarded at least one of her works as not fitting straightforwardly into the category of the novel.", "Cather's unusual treatment of narrative time was influenced by the Russian writers Turgenev and Tolstoy.", "Cather's work was regarded as flawed by most contemporary critics." ]
2
Passage B indicates which one of the following?
Passage A When Cather gave examples of high quality in fiction, she cited Russian writers Ivan Turgenev or Leo Tolstoy or both. Indeed, Edmund Wilson noted in 1922 that Cather followed the manner of Turgenev, not depicting her characters' emotions directly but telling us how they behave and letting their "inner blaze of glory shine through the simple recital." Turgenev's method was to select details that described a character's appearance and actions without trying to explain them. A writer, he said, "must be a psychologist—but a secret one; he must know and feel the roots of phenomena, but only present the phenomena themselves." Similarly, he argued that a writer must have complete knowledge of a character so as to avoid overloading the work with unnecessary detail, concentrating instead on what is characteristic and typical. Here we have an impressionistic aesthetic that anticipates Cather's: what Turgenev referred to as secret knowledge Cather called "the thing not named." In one essay she writes that "whatever is felt upon the page without being specifically named there—that, one might say, is created." For both writers, there is the absolute importance of selection and simplification; for both, art is the fusing of the physical world of setting and actions with the emotional reality of the characters. What synthesizes all the elements of narrative for these writers is the establishment of a prevailing mood. Passage B In a famous 1927 letter, Cather writes of her novel Death Comes for the Archbishop, "Many [reviewers] assert vehemently that it is not a novel. Myself, I prefer to call it a narrative." Cather's preference anticipated an important reformulation of the criticism of fiction: the body of literary theory, called "narratology," articulated by French literary theorists in the 1960s. This approach broadens and simplifies the fundamental paradigms according to which we view fiction: they ask of narrative only that it be narrative, that it in fiction, she invariably cited Russian writers Ivan Turgenev or Leo Tolstoy or both. Indeed, Edmund Wilson noted in 1922 that Cather followed the manner of Turgenev, not depicting her characters' emotions directly but telling us how they behave and letting their "inner blaze of glory shine through the simple recital." Turgenev's method was to select details that described a character's appearance and actions without trying to explain them. A writer, he said, "must be a psychologist—but a secret one; he must know and feel the roots of phenomena, but only present the phenomena themselves." Similarly, he argued that a writer must have complete knowledge of a character so as to avoid overloading the work with unnecessary detail, concentrating instead on what is characteristic and typical. Here we have an impressionistic aesthetic that anticipates Cather's: what Turgenev referred to as secret knowledge Cather called "the thing not named." In one essay she writes that "whatever is felt upon the page without being specifically named there—that, one might say, is created." For both writers, there is the absolute importance of selection and simplification; for both, art is the fusing of the physical world of setting and actions with the emotional reality of the characters. What synthesizes all the elements of narrative for these writers is the establishment of a prevailing mood. Passage B In a famous 1927 letter, Cather writes of her novel Death Comes for the Archbishop, "Many [reviewers] assert vehemently that it is not a novel. Myself, I prefer to call it a narrative." Cather's preference anticipated an important reformulation of the criticism of fiction: the body of literary theory, called "narratology," articulated by French literary theorists in the 1960s. This approach broadens and simplifies the fundamental paradigms according to which we view fiction: they ask of narrative only that it be narrative, that it tell a story. Narratologists tend not to focus on the characteristics of narrative's dominant modern Western form, the "realistic novel" : direct psychological characterization, realistic treatment of time, causal plotting, logical closure. Such a model of criticism, which takes as its object "narrative" rather than the "novel," seems exactly appropriate to Cather's work. Indeed, her severest critics have always questioned precisely her capabilities as a novelist. Morton Zabel argued that "[Cather's] themes...could readily fail to find the structure and substance that might have given them life or redeemed them from the tenuity of a sketch" ; Leon Edel called one of her novels "two inconclusive fragments." These critics and others like them treat as failures some of the central features of Cather's impressionistic technique: unusual treatment of narrative time, unexpected focus, ambiguous conclusions, a preference for the bold, simple, and stylized in character as well as in landscape. These "non-novelistic" structures indirectly articulate the essential and conflicting forces of desire at work throughout Cather's fiction.
200906_4-RC_3_15
[ "A meticulous inventory of the elegant furniture and décor in a character's living room is used to indicate that the character is wealthy.", "An account of a character's emotional scars is used to explain the negative effects the character has on his family.", "A description of a slightly quivering drink in the hand of a character at a dinner party is used to suggest that the character is timid.", "A chronological summary of the events that spark a family conflict is used to supply the context for an in-depth narration of that conflict.", "A detailed narration of an unprovoked act of violence and the reprisals it triggers is used to portray the theme that violence begets violence." ]
2
It can be inferred that both authors would be most likely to regard which one of the following as exemplifying Cather's narrative technique?
Passage A When Cather gave examples of high quality in fiction, she cited Russian writers Ivan Turgenev or Leo Tolstoy or both. Indeed, Edmund Wilson noted in 1922 that Cather followed the manner of Turgenev, not depicting her characters' emotions directly but telling us how they behave and letting their "inner blaze of glory shine through the simple recital." Turgenev's method was to select details that described a character's appearance and actions without trying to explain them. A writer, he said, "must be a psychologist—but a secret one; he must know and feel the roots of phenomena, but only present the phenomena themselves." Similarly, he argued that a writer must have complete knowledge of a character so as to avoid overloading the work with unnecessary detail, concentrating instead on what is characteristic and typical. Here we have an impressionistic aesthetic that anticipates Cather's: what Turgenev referred to as secret knowledge Cather called "the thing not named." In one essay she writes that "whatever is felt upon the page without being specifically named there—that, one might say, is created." For both writers, there is the absolute importance of selection and simplification; for both, art is the fusing of the physical world of setting and actions with the emotional reality of the characters. What synthesizes all the elements of narrative for these writers is the establishment of a prevailing mood. Passage B In a famous 1927 letter, Cather writes of her novel Death Comes for the Archbishop, "Many [reviewers] assert vehemently that it is not a novel. Myself, I prefer to call it a narrative." Cather's preference anticipated an important reformulation of the criticism of fiction: the body of literary theory, called "narratology," articulated by French literary theorists in the 1960s. This approach broadens and simplifies the fundamental paradigms according to which we view fiction: they ask of narrative only that it be narrative, that it in fiction, she invariably cited Russian writers Ivan Turgenev or Leo Tolstoy or both. Indeed, Edmund Wilson noted in 1922 that Cather followed the manner of Turgenev, not depicting her characters' emotions directly but telling us how they behave and letting their "inner blaze of glory shine through the simple recital." Turgenev's method was to select details that described a character's appearance and actions without trying to explain them. A writer, he said, "must be a psychologist—but a secret one; he must know and feel the roots of phenomena, but only present the phenomena themselves." Similarly, he argued that a writer must have complete knowledge of a character so as to avoid overloading the work with unnecessary detail, concentrating instead on what is characteristic and typical. Here we have an impressionistic aesthetic that anticipates Cather's: what Turgenev referred to as secret knowledge Cather called "the thing not named." In one essay she writes that "whatever is felt upon the page without being specifically named there—that, one might say, is created." For both writers, there is the absolute importance of selection and simplification; for both, art is the fusing of the physical world of setting and actions with the emotional reality of the characters. What synthesizes all the elements of narrative for these writers is the establishment of a prevailing mood. Passage B In a famous 1927 letter, Cather writes of her novel Death Comes for the Archbishop, "Many [reviewers] assert vehemently that it is not a novel. Myself, I prefer to call it a narrative." Cather's preference anticipated an important reformulation of the criticism of fiction: the body of literary theory, called "narratology," articulated by French literary theorists in the 1960s. This approach broadens and simplifies the fundamental paradigms according to which we view fiction: they ask of narrative only that it be narrative, that it tell a story. Narratologists tend not to focus on the characteristics of narrative's dominant modern Western form, the "realistic novel" : direct psychological characterization, realistic treatment of time, causal plotting, logical closure. Such a model of criticism, which takes as its object "narrative" rather than the "novel," seems exactly appropriate to Cather's work. Indeed, her severest critics have always questioned precisely her capabilities as a novelist. Morton Zabel argued that "[Cather's] themes...could readily fail to find the structure and substance that might have given them life or redeemed them from the tenuity of a sketch" ; Leon Edel called one of her novels "two inconclusive fragments." These critics and others like them treat as failures some of the central features of Cather's impressionistic technique: unusual treatment of narrative time, unexpected focus, ambiguous conclusions, a preference for the bold, simple, and stylized in character as well as in landscape. These "non-novelistic" structures indirectly articulate the essential and conflicting forces of desire at work throughout Cather's fiction.
200906_4-RC_3_16
[ "Cather's fiction is best approached by focusing purely on narrative, rather than on the formal characteristics of the novel.", "Most commentators on Cather's novels have mistakenly treated her distinctive narrative techniques as aesthetic flaws.", "Cather intentionally avoided the realistic psychological characterization that is the central feature of the modern Western novel.", "Cather's impressionistic narratives served as an important impetus for the development of narratology in the 1960s.", "Cather rejected the narrative constraints of the realistic novel and instead concentrated on portraying her characters by sketching their inner lives." ]
0
Which one of the following most accurately states the main point of passage B?
Passage A When Cather gave examples of high quality in fiction, she cited Russian writers Ivan Turgenev or Leo Tolstoy or both. Indeed, Edmund Wilson noted in 1922 that Cather followed the manner of Turgenev, not depicting her characters' emotions directly but telling us how they behave and letting their "inner blaze of glory shine through the simple recital." Turgenev's method was to select details that described a character's appearance and actions without trying to explain them. A writer, he said, "must be a psychologist—but a secret one; he must know and feel the roots of phenomena, but only present the phenomena themselves." Similarly, he argued that a writer must have complete knowledge of a character so as to avoid overloading the work with unnecessary detail, concentrating instead on what is characteristic and typical. Here we have an impressionistic aesthetic that anticipates Cather's: what Turgenev referred to as secret knowledge Cather called "the thing not named." In one essay she writes that "whatever is felt upon the page without being specifically named there—that, one might say, is created." For both writers, there is the absolute importance of selection and simplification; for both, art is the fusing of the physical world of setting and actions with the emotional reality of the characters. What synthesizes all the elements of narrative for these writers is the establishment of a prevailing mood. Passage B In a famous 1927 letter, Cather writes of her novel Death Comes for the Archbishop, "Many [reviewers] assert vehemently that it is not a novel. Myself, I prefer to call it a narrative." Cather's preference anticipated an important reformulation of the criticism of fiction: the body of literary theory, called "narratology," articulated by French literary theorists in the 1960s. This approach broadens and simplifies the fundamental paradigms according to which we view fiction: they ask of narrative only that it be narrative, that it in fiction, she invariably cited Russian writers Ivan Turgenev or Leo Tolstoy or both. Indeed, Edmund Wilson noted in 1922 that Cather followed the manner of Turgenev, not depicting her characters' emotions directly but telling us how they behave and letting their "inner blaze of glory shine through the simple recital." Turgenev's method was to select details that described a character's appearance and actions without trying to explain them. A writer, he said, "must be a psychologist—but a secret one; he must know and feel the roots of phenomena, but only present the phenomena themselves." Similarly, he argued that a writer must have complete knowledge of a character so as to avoid overloading the work with unnecessary detail, concentrating instead on what is characteristic and typical. Here we have an impressionistic aesthetic that anticipates Cather's: what Turgenev referred to as secret knowledge Cather called "the thing not named." In one essay she writes that "whatever is felt upon the page without being specifically named there—that, one might say, is created." For both writers, there is the absolute importance of selection and simplification; for both, art is the fusing of the physical world of setting and actions with the emotional reality of the characters. What synthesizes all the elements of narrative for these writers is the establishment of a prevailing mood. Passage B In a famous 1927 letter, Cather writes of her novel Death Comes for the Archbishop, "Many [reviewers] assert vehemently that it is not a novel. Myself, I prefer to call it a narrative." Cather's preference anticipated an important reformulation of the criticism of fiction: the body of literary theory, called "narratology," articulated by French literary theorists in the 1960s. This approach broadens and simplifies the fundamental paradigms according to which we view fiction: they ask of narrative only that it be narrative, that it tell a story. Narratologists tend not to focus on the characteristics of narrative's dominant modern Western form, the "realistic novel" : direct psychological characterization, realistic treatment of time, causal plotting, logical closure. Such a model of criticism, which takes as its object "narrative" rather than the "novel," seems exactly appropriate to Cather's work. Indeed, her severest critics have always questioned precisely her capabilities as a novelist. Morton Zabel argued that "[Cather's] themes...could readily fail to find the structure and substance that might have given them life or redeemed them from the tenuity of a sketch" ; Leon Edel called one of her novels "two inconclusive fragments." These critics and others like them treat as failures some of the central features of Cather's impressionistic technique: unusual treatment of narrative time, unexpected focus, ambiguous conclusions, a preference for the bold, simple, and stylized in character as well as in landscape. These "non-novelistic" structures indirectly articulate the essential and conflicting forces of desire at work throughout Cather's fiction.
200906_4-RC_3_17
[ "More than her contemporaries, Cather used stream-of-consciousness narration to portray her characters.", "Cather's works were not intended as novels, but rather as narratives.", "Narratology is the most appropriate critical approach to Cather's work.", "Cather's technique of evoking the \"thing not named\" had a marked influence on later novelists.", "Cather used impressionistic narrative techniques to portray the psychology of her characters." ]
4
It is most likely that the authors of the two passages would both agree with which one of the following statements?
Passage A When Cather gave examples of high quality in fiction, she cited Russian writers Ivan Turgenev or Leo Tolstoy or both. Indeed, Edmund Wilson noted in 1922 that Cather followed the manner of Turgenev, not depicting her characters' emotions directly but telling us how they behave and letting their "inner blaze of glory shine through the simple recital." Turgenev's method was to select details that described a character's appearance and actions without trying to explain them. A writer, he said, "must be a psychologist—but a secret one; he must know and feel the roots of phenomena, but only present the phenomena themselves." Similarly, he argued that a writer must have complete knowledge of a character so as to avoid overloading the work with unnecessary detail, concentrating instead on what is characteristic and typical. Here we have an impressionistic aesthetic that anticipates Cather's: what Turgenev referred to as secret knowledge Cather called "the thing not named." In one essay she writes that "whatever is felt upon the page without being specifically named there—that, one might say, is created." For both writers, there is the absolute importance of selection and simplification; for both, art is the fusing of the physical world of setting and actions with the emotional reality of the characters. What synthesizes all the elements of narrative for these writers is the establishment of a prevailing mood. Passage B In a famous 1927 letter, Cather writes of her novel Death Comes for the Archbishop, "Many [reviewers] assert vehemently that it is not a novel. Myself, I prefer to call it a narrative." Cather's preference anticipated an important reformulation of the criticism of fiction: the body of literary theory, called "narratology," articulated by French literary theorists in the 1960s. This approach broadens and simplifies the fundamental paradigms according to which we view fiction: they ask of narrative only that it be narrative, that it in fiction, she invariably cited Russian writers Ivan Turgenev or Leo Tolstoy or both. Indeed, Edmund Wilson noted in 1922 that Cather followed the manner of Turgenev, not depicting her characters' emotions directly but telling us how they behave and letting their "inner blaze of glory shine through the simple recital." Turgenev's method was to select details that described a character's appearance and actions without trying to explain them. A writer, he said, "must be a psychologist—but a secret one; he must know and feel the roots of phenomena, but only present the phenomena themselves." Similarly, he argued that a writer must have complete knowledge of a character so as to avoid overloading the work with unnecessary detail, concentrating instead on what is characteristic and typical. Here we have an impressionistic aesthetic that anticipates Cather's: what Turgenev referred to as secret knowledge Cather called "the thing not named." In one essay she writes that "whatever is felt upon the page without being specifically named there—that, one might say, is created." For both writers, there is the absolute importance of selection and simplification; for both, art is the fusing of the physical world of setting and actions with the emotional reality of the characters. What synthesizes all the elements of narrative for these writers is the establishment of a prevailing mood. Passage B In a famous 1927 letter, Cather writes of her novel Death Comes for the Archbishop, "Many [reviewers] assert vehemently that it is not a novel. Myself, I prefer to call it a narrative." Cather's preference anticipated an important reformulation of the criticism of fiction: the body of literary theory, called "narratology," articulated by French literary theorists in the 1960s. This approach broadens and simplifies the fundamental paradigms according to which we view fiction: they ask of narrative only that it be narrative, that it tell a story. Narratologists tend not to focus on the characteristics of narrative's dominant modern Western form, the "realistic novel" : direct psychological characterization, realistic treatment of time, causal plotting, logical closure. Such a model of criticism, which takes as its object "narrative" rather than the "novel," seems exactly appropriate to Cather's work. Indeed, her severest critics have always questioned precisely her capabilities as a novelist. Morton Zabel argued that "[Cather's] themes...could readily fail to find the structure and substance that might have given them life or redeemed them from the tenuity of a sketch" ; Leon Edel called one of her novels "two inconclusive fragments." These critics and others like them treat as failures some of the central features of Cather's impressionistic technique: unusual treatment of narrative time, unexpected focus, ambiguous conclusions, a preference for the bold, simple, and stylized in character as well as in landscape. These "non-novelistic" structures indirectly articulate the essential and conflicting forces of desire at work throughout Cather's fiction.
200906_4-RC_3_18
[ "Description of the salient features of the setting, such as a chair in which a character often sits.", "A plot that does not follow chronological time, but rather moves frequently between the novel's past and present.", "Description of a character's physical appearance, dress, and facial expressions.", "Direct representation of dialogue between the novel's characters, using quotation marks to set off characters' words.", "A narration of a character's inner thoughts, including an account of the character's anxieties and wishes." ]
4
Both authors would be likely to agree that which one of the following, though typical of many novels, would NOT be found in Cather's work?
Passage A When Cather gave examples of high quality in fiction, she cited Russian writers Ivan Turgenev or Leo Tolstoy or both. Indeed, Edmund Wilson noted in 1922 that Cather followed the manner of Turgenev, not depicting her characters' emotions directly but telling us how they behave and letting their "inner blaze of glory shine through the simple recital." Turgenev's method was to select details that described a character's appearance and actions without trying to explain them. A writer, he said, "must be a psychologist—but a secret one; he must know and feel the roots of phenomena, but only present the phenomena themselves." Similarly, he argued that a writer must have complete knowledge of a character so as to avoid overloading the work with unnecessary detail, concentrating instead on what is characteristic and typical. Here we have an impressionistic aesthetic that anticipates Cather's: what Turgenev referred to as secret knowledge Cather called "the thing not named." In one essay she writes that "whatever is felt upon the page without being specifically named there—that, one might say, is created." For both writers, there is the absolute importance of selection and simplification; for both, art is the fusing of the physical world of setting and actions with the emotional reality of the characters. What synthesizes all the elements of narrative for these writers is the establishment of a prevailing mood. Passage B In a famous 1927 letter, Cather writes of her novel Death Comes for the Archbishop, "Many [reviewers] assert vehemently that it is not a novel. Myself, I prefer to call it a narrative." Cather's preference anticipated an important reformulation of the criticism of fiction: the body of literary theory, called "narratology," articulated by French literary theorists in the 1960s. This approach broadens and simplifies the fundamental paradigms according to which we view fiction: they ask of narrative only that it be narrative, that it in fiction, she invariably cited Russian writers Ivan Turgenev or Leo Tolstoy or both. Indeed, Edmund Wilson noted in 1922 that Cather followed the manner of Turgenev, not depicting her characters' emotions directly but telling us how they behave and letting their "inner blaze of glory shine through the simple recital." Turgenev's method was to select details that described a character's appearance and actions without trying to explain them. A writer, he said, "must be a psychologist—but a secret one; he must know and feel the roots of phenomena, but only present the phenomena themselves." Similarly, he argued that a writer must have complete knowledge of a character so as to avoid overloading the work with unnecessary detail, concentrating instead on what is characteristic and typical. Here we have an impressionistic aesthetic that anticipates Cather's: what Turgenev referred to as secret knowledge Cather called "the thing not named." In one essay she writes that "whatever is felt upon the page without being specifically named there—that, one might say, is created." For both writers, there is the absolute importance of selection and simplification; for both, art is the fusing of the physical world of setting and actions with the emotional reality of the characters. What synthesizes all the elements of narrative for these writers is the establishment of a prevailing mood. Passage B In a famous 1927 letter, Cather writes of her novel Death Comes for the Archbishop, "Many [reviewers] assert vehemently that it is not a novel. Myself, I prefer to call it a narrative." Cather's preference anticipated an important reformulation of the criticism of fiction: the body of literary theory, called "narratology," articulated by French literary theorists in the 1960s. This approach broadens and simplifies the fundamental paradigms according to which we view fiction: they ask of narrative only that it be narrative, that it tell a story. Narratologists tend not to focus on the characteristics of narrative's dominant modern Western form, the "realistic novel" : direct psychological characterization, realistic treatment of time, causal plotting, logical closure. Such a model of criticism, which takes as its object "narrative" rather than the "novel," seems exactly appropriate to Cather's work. Indeed, her severest critics have always questioned precisely her capabilities as a novelist. Morton Zabel argued that "[Cather's] themes...could readily fail to find the structure and substance that might have given them life or redeemed them from the tenuity of a sketch" ; Leon Edel called one of her novels "two inconclusive fragments." These critics and others like them treat as failures some of the central features of Cather's impressionistic technique: unusual treatment of narrative time, unexpected focus, ambiguous conclusions, a preference for the bold, simple, and stylized in character as well as in landscape. These "non-novelistic" structures indirectly articulate the essential and conflicting forces of desire at work throughout Cather's fiction.
200906_4-RC_3_19
[ "describe the primary influences on Cather's work", "identify some of the distinctive characteristics of Cather's work", "explain the critical reception Cather's work received in her lifetime", "compare Cather's novels to the archetypal form of the realistic novel", "examine the impact of European literature and literary theory on Cather's work" ]
1
A central purpose of each passage is to
Fractal geometry is a mathematical theory devoted to the study of complex shapes called fractals. Although an exact definition of fractals has not been established, fractals commonly exhibit the property of self-similarity: the reiteration of irregular details or patterns at progressively smaller scales so that each part, when magnified, looks basically like the object as a whole. The Koch curve is a significant fractal in mathematics and examining it provides some insight into fractal geometry. To generate the Koch curve, one begins with a straight line. The middle third of the line is removed and replaced with two line segments, each as long as the removed piece, which are positioned so as to meet and form the top of a triangle. At this stage, the curve consists of four connected segments of equal length that form a pointed protrusion in the middle. This process is repeated on the four segments so that all the protrusions are on the same side of the curve, and then the process is repeated indefinitely on the segments at each stage of the construction. Self-similarity is built into the construction process by treating segments at each stage the same way as the original segment was treated. Since the rules for getting from one stage to another are fully explicit and always the same, images of successive stages of the process can be generated by computer. Theoretically, the Koch curve is the result of infinitely many steps in the construction process, but the finest image approximating the Koch curve will be limited by the fact that eventually the segments will get too short to be drawn or displayed. However, using computer graphics to produce images of successive stages of the construction process dramatically illustrates a major attraction of fractal geometry: simple processes can be responsible for incredibly complex patterns. A worldwide public has become captivated by fractal geometry after viewing astonishing computer-generated images of fractals; enthusiastic practitioners in the field of fractal geometry consider it a new language for describing complex natural and mathematical forms. They anticipate that fractal geometry's significance will rival that of calculus and expect that proficiency in fractal geometry will allow mathematicians to describe the form of a cloud as easily and precisely as an architect can describe a house using the language of traditional geometry. Other mathematicians have reservations about the fractal geometers' preoccupation with computer-generated graphic images and their lack of interest in theory. These mathematicians point out that traditional mathematics consists of proving theorems, and while many theorems about fractals have already been proven using the notions of pre-fractal mathematics, fractal geometers have proven only a handful of theorems that could not have been proven with pre-fractal mathematics. According to these mathematicians, fractal geometry can attain a lasting role in mathematics only if it becomes a precise language supporting a system of theorems and proofs.
200906_4-RC_4_20
[ "Because of its unique forms, fractal geometry is especially adaptable to computer technology and is therefore likely to grow in importance and render pre-fractal mathematics obsolete.", "Though its use in the generation of extremely complex forms makes fractal geometry an intriguing new mathematical theory, it is not yet universally regarded as having attained the theoretical rigor of traditional mathematics.", "Fractal geometry is significant because of its use of self-similarity, a concept that has enabled geometers to generate extremely detailed computer images of natural forms.", "Using the Koch curve as a model, fractal geometers have developed a new mathematical language that is especially useful in technological contexts because it does not rely on theorems.", "Though fractal geometry has thus far been of great value for its capacity to define abstract mathematical shapes, it is not expected to be useful for the description of ordinary natural shapes." ]
1
Which one of the following most accurately expresses the main point of the passage?
Fractal geometry is a mathematical theory devoted to the study of complex shapes called fractals. Although an exact definition of fractals has not been established, fractals commonly exhibit the property of self-similarity: the reiteration of irregular details or patterns at progressively smaller scales so that each part, when magnified, looks basically like the object as a whole. The Koch curve is a significant fractal in mathematics and examining it provides some insight into fractal geometry. To generate the Koch curve, one begins with a straight line. The middle third of the line is removed and replaced with two line segments, each as long as the removed piece, which are positioned so as to meet and form the top of a triangle. At this stage, the curve consists of four connected segments of equal length that form a pointed protrusion in the middle. This process is repeated on the four segments so that all the protrusions are on the same side of the curve, and then the process is repeated indefinitely on the segments at each stage of the construction. Self-similarity is built into the construction process by treating segments at each stage the same way as the original segment was treated. Since the rules for getting from one stage to another are fully explicit and always the same, images of successive stages of the process can be generated by computer. Theoretically, the Koch curve is the result of infinitely many steps in the construction process, but the finest image approximating the Koch curve will be limited by the fact that eventually the segments will get too short to be drawn or displayed. However, using computer graphics to produce images of successive stages of the construction process dramatically illustrates a major attraction of fractal geometry: simple processes can be responsible for incredibly complex patterns. A worldwide public has become captivated by fractal geometry after viewing astonishing computer-generated images of fractals; enthusiastic practitioners in the field of fractal geometry consider it a new language for describing complex natural and mathematical forms. They anticipate that fractal geometry's significance will rival that of calculus and expect that proficiency in fractal geometry will allow mathematicians to describe the form of a cloud as easily and precisely as an architect can describe a house using the language of traditional geometry. Other mathematicians have reservations about the fractal geometers' preoccupation with computer-generated graphic images and their lack of interest in theory. These mathematicians point out that traditional mathematics consists of proving theorems, and while many theorems about fractals have already been proven using the notions of pre-fractal mathematics, fractal geometers have proven only a handful of theorems that could not have been proven with pre-fractal mathematics. According to these mathematicians, fractal geometry can attain a lasting role in mathematics only if it becomes a precise language supporting a system of theorems and proofs.
200906_4-RC_4_21
[ "illustrated by an example", "uncomplicated", "expressed unambiguously", "in need of lengthy computation", "agreed on by all" ]
2
Which one of the following is closest to the meaning of the phrase "fully explicit" as used in lines 24–25?
Fractal geometry is a mathematical theory devoted to the study of complex shapes called fractals. Although an exact definition of fractals has not been established, fractals commonly exhibit the property of self-similarity: the reiteration of irregular details or patterns at progressively smaller scales so that each part, when magnified, looks basically like the object as a whole. The Koch curve is a significant fractal in mathematics and examining it provides some insight into fractal geometry. To generate the Koch curve, one begins with a straight line. The middle third of the line is removed and replaced with two line segments, each as long as the removed piece, which are positioned so as to meet and form the top of a triangle. At this stage, the curve consists of four connected segments of equal length that form a pointed protrusion in the middle. This process is repeated on the four segments so that all the protrusions are on the same side of the curve, and then the process is repeated indefinitely on the segments at each stage of the construction. Self-similarity is built into the construction process by treating segments at each stage the same way as the original segment was treated. Since the rules for getting from one stage to another are fully explicit and always the same, images of successive stages of the process can be generated by computer. Theoretically, the Koch curve is the result of infinitely many steps in the construction process, but the finest image approximating the Koch curve will be limited by the fact that eventually the segments will get too short to be drawn or displayed. However, using computer graphics to produce images of successive stages of the construction process dramatically illustrates a major attraction of fractal geometry: simple processes can be responsible for incredibly complex patterns. A worldwide public has become captivated by fractal geometry after viewing astonishing computer-generated images of fractals; enthusiastic practitioners in the field of fractal geometry consider it a new language for describing complex natural and mathematical forms. They anticipate that fractal geometry's significance will rival that of calculus and expect that proficiency in fractal geometry will allow mathematicians to describe the form of a cloud as easily and precisely as an architect can describe a house using the language of traditional geometry. Other mathematicians have reservations about the fractal geometers' preoccupation with computer-generated graphic images and their lack of interest in theory. These mathematicians point out that traditional mathematics consists of proving theorems, and while many theorems about fractals have already been proven using the notions of pre-fractal mathematics, fractal geometers have proven only a handful of theorems that could not have been proven with pre-fractal mathematics. According to these mathematicians, fractal geometry can attain a lasting role in mathematics only if it becomes a precise language supporting a system of theorems and proofs.
200906_4-RC_4_22
[ "Any branch broken off a tree looks like the tree itself.", "Each portion of the intricately patterned frost on a window looks like the pattern as a whole.", "The pattern of blood vessels in each part of the human body is similar to the pattern of blood vessels in the entire body.", "The seeds of several subspecies of maple tree resemble one another in shape despite differences in size.", "The florets composing a cauliflower head resemble the entire cauliflower head." ]
3
According to the description in the passage, each one of the following illustrates the concept of self-similarity EXCEPT:
Fractal geometry is a mathematical theory devoted to the study of complex shapes called fractals. Although an exact definition of fractals has not been established, fractals commonly exhibit the property of self-similarity: the reiteration of irregular details or patterns at progressively smaller scales so that each part, when magnified, looks basically like the object as a whole. The Koch curve is a significant fractal in mathematics and examining it provides some insight into fractal geometry. To generate the Koch curve, one begins with a straight line. The middle third of the line is removed and replaced with two line segments, each as long as the removed piece, which are positioned so as to meet and form the top of a triangle. At this stage, the curve consists of four connected segments of equal length that form a pointed protrusion in the middle. This process is repeated on the four segments so that all the protrusions are on the same side of the curve, and then the process is repeated indefinitely on the segments at each stage of the construction. Self-similarity is built into the construction process by treating segments at each stage the same way as the original segment was treated. Since the rules for getting from one stage to another are fully explicit and always the same, images of successive stages of the process can be generated by computer. Theoretically, the Koch curve is the result of infinitely many steps in the construction process, but the finest image approximating the Koch curve will be limited by the fact that eventually the segments will get too short to be drawn or displayed. However, using computer graphics to produce images of successive stages of the construction process dramatically illustrates a major attraction of fractal geometry: simple processes can be responsible for incredibly complex patterns. A worldwide public has become captivated by fractal geometry after viewing astonishing computer-generated images of fractals; enthusiastic practitioners in the field of fractal geometry consider it a new language for describing complex natural and mathematical forms. They anticipate that fractal geometry's significance will rival that of calculus and expect that proficiency in fractal geometry will allow mathematicians to describe the form of a cloud as easily and precisely as an architect can describe a house using the language of traditional geometry. Other mathematicians have reservations about the fractal geometers' preoccupation with computer-generated graphic images and their lack of interest in theory. These mathematicians point out that traditional mathematics consists of proving theorems, and while many theorems about fractals have already been proven using the notions of pre-fractal mathematics, fractal geometers have proven only a handful of theorems that could not have been proven with pre-fractal mathematics. According to these mathematicians, fractal geometry can attain a lasting role in mathematics only if it becomes a precise language supporting a system of theorems and proofs.
200906_4-RC_4_23
[ "show how fractal geometry can be reduced to traditional geometry", "give an example of a natural form that can be described by fractal geometry", "anticipate the objection that fractal geometry is not a precise language", "illustrate the concept of self-similarity", "provide an exact definition of fractals" ]
3
The explanation of how a Koch curve is generated (lines 10–20) serves primarily to
Fractal geometry is a mathematical theory devoted to the study of complex shapes called fractals. Although an exact definition of fractals has not been established, fractals commonly exhibit the property of self-similarity: the reiteration of irregular details or patterns at progressively smaller scales so that each part, when magnified, looks basically like the object as a whole. The Koch curve is a significant fractal in mathematics and examining it provides some insight into fractal geometry. To generate the Koch curve, one begins with a straight line. The middle third of the line is removed and replaced with two line segments, each as long as the removed piece, which are positioned so as to meet and form the top of a triangle. At this stage, the curve consists of four connected segments of equal length that form a pointed protrusion in the middle. This process is repeated on the four segments so that all the protrusions are on the same side of the curve, and then the process is repeated indefinitely on the segments at each stage of the construction. Self-similarity is built into the construction process by treating segments at each stage the same way as the original segment was treated. Since the rules for getting from one stage to another are fully explicit and always the same, images of successive stages of the process can be generated by computer. Theoretically, the Koch curve is the result of infinitely many steps in the construction process, but the finest image approximating the Koch curve will be limited by the fact that eventually the segments will get too short to be drawn or displayed. However, using computer graphics to produce images of successive stages of the construction process dramatically illustrates a major attraction of fractal geometry: simple processes can be responsible for incredibly complex patterns. A worldwide public has become captivated by fractal geometry after viewing astonishing computer-generated images of fractals; enthusiastic practitioners in the field of fractal geometry consider it a new language for describing complex natural and mathematical forms. They anticipate that fractal geometry's significance will rival that of calculus and expect that proficiency in fractal geometry will allow mathematicians to describe the form of a cloud as easily and precisely as an architect can describe a house using the language of traditional geometry. Other mathematicians have reservations about the fractal geometers' preoccupation with computer-generated graphic images and their lack of interest in theory. These mathematicians point out that traditional mathematics consists of proving theorems, and while many theorems about fractals have already been proven using the notions of pre-fractal mathematics, fractal geometers have proven only a handful of theorems that could not have been proven with pre-fractal mathematics. According to these mathematicians, fractal geometry can attain a lasting role in mathematics only if it becomes a precise language supporting a system of theorems and proofs.
200906_4-RC_4_24
[ "It is potentially much more important than calculus.", "Its role in traditional mathematics will expand as computers become faster.", "It is the fastest-growing field of mathematics.", "It encourages the use of computer programs to prove mathematical theorems.", "It enables geometers to generate complex forms using simple processes." ]
4
Which one of the following does the author present as a characteristic of fractal geometry?
Fractal geometry is a mathematical theory devoted to the study of complex shapes called fractals. Although an exact definition of fractals has not been established, fractals commonly exhibit the property of self-similarity: the reiteration of irregular details or patterns at progressively smaller scales so that each part, when magnified, looks basically like the object as a whole. The Koch curve is a significant fractal in mathematics and examining it provides some insight into fractal geometry. To generate the Koch curve, one begins with a straight line. The middle third of the line is removed and replaced with two line segments, each as long as the removed piece, which are positioned so as to meet and form the top of a triangle. At this stage, the curve consists of four connected segments of equal length that form a pointed protrusion in the middle. This process is repeated on the four segments so that all the protrusions are on the same side of the curve, and then the process is repeated indefinitely on the segments at each stage of the construction. Self-similarity is built into the construction process by treating segments at each stage the same way as the original segment was treated. Since the rules for getting from one stage to another are fully explicit and always the same, images of successive stages of the process can be generated by computer. Theoretically, the Koch curve is the result of infinitely many steps in the construction process, but the finest image approximating the Koch curve will be limited by the fact that eventually the segments will get too short to be drawn or displayed. However, using computer graphics to produce images of successive stages of the construction process dramatically illustrates a major attraction of fractal geometry: simple processes can be responsible for incredibly complex patterns. A worldwide public has become captivated by fractal geometry after viewing astonishing computer-generated images of fractals; enthusiastic practitioners in the field of fractal geometry consider it a new language for describing complex natural and mathematical forms. They anticipate that fractal geometry's significance will rival that of calculus and expect that proficiency in fractal geometry will allow mathematicians to describe the form of a cloud as easily and precisely as an architect can describe a house using the language of traditional geometry. Other mathematicians have reservations about the fractal geometers' preoccupation with computer-generated graphic images and their lack of interest in theory. These mathematicians point out that traditional mathematics consists of proving theorems, and while many theorems about fractals have already been proven using the notions of pre-fractal mathematics, fractal geometers have proven only a handful of theorems that could not have been proven with pre-fractal mathematics. According to these mathematicians, fractal geometry can attain a lasting role in mathematics only if it becomes a precise language supporting a system of theorems and proofs.
200906_4-RC_4_25
[ "The total number of protrusions in the Koch curve at any stage of the construction depends on the length of the initial line chosen for the construction.", "The line segments at each successive stage of the construction of the Koch curve are shorter than the segments at the previous stage.", "Theoretically, as the Koch curve is constructed its line segments become infinitely small.", "At every stage of constructing the Koch curve, all the line segments composing it are of equal length.", "The length of the line segments in the Koch curve at any stage of its construction depends on the length of the initial line chosen for the construction." ]
0
Each of the following statements about the Koch curve can be properly deduced from the information given in the passage EXCEPT:
Fractal geometry is a mathematical theory devoted to the study of complex shapes called fractals. Although an exact definition of fractals has not been established, fractals commonly exhibit the property of self-similarity: the reiteration of irregular details or patterns at progressively smaller scales so that each part, when magnified, looks basically like the object as a whole. The Koch curve is a significant fractal in mathematics and examining it provides some insight into fractal geometry. To generate the Koch curve, one begins with a straight line. The middle third of the line is removed and replaced with two line segments, each as long as the removed piece, which are positioned so as to meet and form the top of a triangle. At this stage, the curve consists of four connected segments of equal length that form a pointed protrusion in the middle. This process is repeated on the four segments so that all the protrusions are on the same side of the curve, and then the process is repeated indefinitely on the segments at each stage of the construction. Self-similarity is built into the construction process by treating segments at each stage the same way as the original segment was treated. Since the rules for getting from one stage to another are fully explicit and always the same, images of successive stages of the process can be generated by computer. Theoretically, the Koch curve is the result of infinitely many steps in the construction process, but the finest image approximating the Koch curve will be limited by the fact that eventually the segments will get too short to be drawn or displayed. However, using computer graphics to produce images of successive stages of the construction process dramatically illustrates a major attraction of fractal geometry: simple processes can be responsible for incredibly complex patterns. A worldwide public has become captivated by fractal geometry after viewing astonishing computer-generated images of fractals; enthusiastic practitioners in the field of fractal geometry consider it a new language for describing complex natural and mathematical forms. They anticipate that fractal geometry's significance will rival that of calculus and expect that proficiency in fractal geometry will allow mathematicians to describe the form of a cloud as easily and precisely as an architect can describe a house using the language of traditional geometry. Other mathematicians have reservations about the fractal geometers' preoccupation with computer-generated graphic images and their lack of interest in theory. These mathematicians point out that traditional mathematics consists of proving theorems, and while many theorems about fractals have already been proven using the notions of pre-fractal mathematics, fractal geometers have proven only a handful of theorems that could not have been proven with pre-fractal mathematics. According to these mathematicians, fractal geometry can attain a lasting role in mathematics only if it becomes a precise language supporting a system of theorems and proofs.
200906_4-RC_4_26
[ "The Koch curve is the most easily generated, and therefore the most important, of the forms studied by fractal geometers.", "Fractal geometry will eventually be able to be used in the same applications for which traditional geometry is now used.", "The greatest importance of computer images of fractals is their ability to bring fractal geometry to the attention of a wider public.", "Studying self-similarity was impossible before the development of sophisticated computer technologies.", "Certain complex natural forms exhibit a type of self-similarity like that exhibited by fractals." ]
4
The enthusiastic practitioners of fractal geometry mentioned in lines 39–40 would be most likely to agree with which one of the following statements?
Fractal geometry is a mathematical theory devoted to the study of complex shapes called fractals. Although an exact definition of fractals has not been established, fractals commonly exhibit the property of self-similarity: the reiteration of irregular details or patterns at progressively smaller scales so that each part, when magnified, looks basically like the object as a whole. The Koch curve is a significant fractal in mathematics and examining it provides some insight into fractal geometry. To generate the Koch curve, one begins with a straight line. The middle third of the line is removed and replaced with two line segments, each as long as the removed piece, which are positioned so as to meet and form the top of a triangle. At this stage, the curve consists of four connected segments of equal length that form a pointed protrusion in the middle. This process is repeated on the four segments so that all the protrusions are on the same side of the curve, and then the process is repeated indefinitely on the segments at each stage of the construction. Self-similarity is built into the construction process by treating segments at each stage the same way as the original segment was treated. Since the rules for getting from one stage to another are fully explicit and always the same, images of successive stages of the process can be generated by computer. Theoretically, the Koch curve is the result of infinitely many steps in the construction process, but the finest image approximating the Koch curve will be limited by the fact that eventually the segments will get too short to be drawn or displayed. However, using computer graphics to produce images of successive stages of the construction process dramatically illustrates a major attraction of fractal geometry: simple processes can be responsible for incredibly complex patterns. A worldwide public has become captivated by fractal geometry after viewing astonishing computer-generated images of fractals; enthusiastic practitioners in the field of fractal geometry consider it a new language for describing complex natural and mathematical forms. They anticipate that fractal geometry's significance will rival that of calculus and expect that proficiency in fractal geometry will allow mathematicians to describe the form of a cloud as easily and precisely as an architect can describe a house using the language of traditional geometry. Other mathematicians have reservations about the fractal geometers' preoccupation with computer-generated graphic images and their lack of interest in theory. These mathematicians point out that traditional mathematics consists of proving theorems, and while many theorems about fractals have already been proven using the notions of pre-fractal mathematics, fractal geometers have proven only a handful of theorems that could not have been proven with pre-fractal mathematics. According to these mathematicians, fractal geometry can attain a lasting role in mathematics only if it becomes a precise language supporting a system of theorems and proofs.
200906_4-RC_4_27
[ "The appeal of a mathematical theory is limited to those individuals who can grasp the theorems and proofs produced in that theory.", "Most of the important recent breakthroughs in mathematical theory would not have been possible without the ability of computers to graphically represent complex shapes.", "Fractal geometry holds the potential to replace traditional geometry in most of its engineering applications.", "A mathematical theory can be developed and find applications even before it establishes a precise definition of its subject matter.", "Only a mathematical theory that supports a system of theorems and proofs will gain enthusiastic support among a significant number of mathematicians." ]
3
The information in the passage best supports which one of the following assertions?
Traditional sources of evidence about ancient history are archaeological remains and surviving texts. Those investigating the crafts practiced by women in ancient times, however, often derive little information from these sources, and the archaeological record is particularly unavailing for the study of ancient textile production, as researchers are thwarted by the perishable nature of cloth. What shreds persisted through millennia were, until recently, often discarded by excavators as useless, as were loom weights, which appeared to be nothing more than blobs of clay. Ancient texts, meanwhile, rarely mention the creation of textiles; moreover, those references that do exist use archaic, unrevealing terminology. Yet despite these obstacles, researchers have learned a great deal about ancient textiles and those who made them, and also about how to piece together a whole picture from many disparate sources of evidence. Technological advances in the analysis of archaeological remains provide much more information than was previously available, especially about minute remains. Successful modern methods include radiocarbon dating, infrared photography for seeing through dirt without removing it, isotope "fingerprinting" for tracing sources of raw materials, and thin-layer chromatography for analyzing dyes. As if in preparation for such advances, the field of archaeology has also undergone an important philosophical revolution in the past century. Once little more than a self-serving quest for artifacts to stock museums and private collections, the field has transformed itself into a scientific pursuit of knowledge about past cultures. As part of this process, archaeologists adopted the fundamental precept of preserving all objects, even those that have no immediately discernible value. Thus in the 1970s two researchers found the oldest known complete garment, a 5,000-year-old linen shirt, among a tumbled heap of dirty linens that had been preserved as part of the well-known Petrie collection decades before anyone began to study the history of textiles. The history of textiles and of the craftswomen who produced them has also advanced on a different front: recreating the actual production of cloth. Reconstructing and implementing ancient production methods provides a valuable way of generating and checking hypotheses. For example, these techniques made it possible to confirm that the excavated pieces of clay once considered useless in fact functioned as loom weights. Similarly, scholars have until recently been obliged to speculate as to which one of two statues of Athena, one large and one small, was adorned with a dress created by a group of Athenian women for a festival, as described in surviving texts. Because records show that it took nine months to produce the dress, scholars assumed it must have adorned the large statue. But by investigating the methods of production and the size of the looms used, researchers have ascertained that in fact a dress for the small statue would have taken nine months to produce.
200909_2-RC_1_1
[ "Archaeology is an expanding discipline that has transformed itself in response both to scientific advances and to changing cultural demands such as a recently increasing interest in women's history.", "A diversity of new approaches to the study of ancient textiles has enabled researchers to infer much about the history of textiles and their creators in the ancient world from the scant evidence that remains.", "Despite many obstacles, research into the textile production methods used by women in the ancient world has advanced over the past century to the point that archaeologists can now replicate ancient equipment and production techniques.", "Research into the history of textiles has spurred sweeping changes in the field of archaeology, from the application of advanced technology to the revaluation of ancient artifacts that were once deemed useless.", "Though researchers have verified certain theories about the history of textiles by using technological developments such as radiocarbon dating, most significant findings in this field have grown out of the reconstruction of ancient production techniques." ]
1
Which one of the following most accurately expresses the main point of the passage?
Traditional sources of evidence about ancient history are archaeological remains and surviving texts. Those investigating the crafts practiced by women in ancient times, however, often derive little information from these sources, and the archaeological record is particularly unavailing for the study of ancient textile production, as researchers are thwarted by the perishable nature of cloth. What shreds persisted through millennia were, until recently, often discarded by excavators as useless, as were loom weights, which appeared to be nothing more than blobs of clay. Ancient texts, meanwhile, rarely mention the creation of textiles; moreover, those references that do exist use archaic, unrevealing terminology. Yet despite these obstacles, researchers have learned a great deal about ancient textiles and those who made them, and also about how to piece together a whole picture from many disparate sources of evidence. Technological advances in the analysis of archaeological remains provide much more information than was previously available, especially about minute remains. Successful modern methods include radiocarbon dating, infrared photography for seeing through dirt without removing it, isotope "fingerprinting" for tracing sources of raw materials, and thin-layer chromatography for analyzing dyes. As if in preparation for such advances, the field of archaeology has also undergone an important philosophical revolution in the past century. Once little more than a self-serving quest for artifacts to stock museums and private collections, the field has transformed itself into a scientific pursuit of knowledge about past cultures. As part of this process, archaeologists adopted the fundamental precept of preserving all objects, even those that have no immediately discernible value. Thus in the 1970s two researchers found the oldest known complete garment, a 5,000-year-old linen shirt, among a tumbled heap of dirty linens that had been preserved as part of the well-known Petrie collection decades before anyone began to study the history of textiles. The history of textiles and of the craftswomen who produced them has also advanced on a different front: recreating the actual production of cloth. Reconstructing and implementing ancient production methods provides a valuable way of generating and checking hypotheses. For example, these techniques made it possible to confirm that the excavated pieces of clay once considered useless in fact functioned as loom weights. Similarly, scholars have until recently been obliged to speculate as to which one of two statues of Athena, one large and one small, was adorned with a dress created by a group of Athenian women for a festival, as described in surviving texts. Because records show that it took nine months to produce the dress, scholars assumed it must have adorned the large statue. But by investigating the methods of production and the size of the looms used, researchers have ascertained that in fact a dress for the small statue would have taken nine months to produce.
200909_2-RC_1_2
[ "skeptical regarding the validity of some of the new hypotheses proposed by researchers", "doubtful that any additional useful knowledge can be generated given the nature of the evidence available", "impatient about the pace of research in light of the resources available", "optimistic that recent scholarly advances will attract increasing numbers of researchers", "satisfied that considerable progress is being made in this field" ]
4
The author's attitude concerning the history of ancient textile production can most accurately be described as
Traditional sources of evidence about ancient history are archaeological remains and surviving texts. Those investigating the crafts practiced by women in ancient times, however, often derive little information from these sources, and the archaeological record is particularly unavailing for the study of ancient textile production, as researchers are thwarted by the perishable nature of cloth. What shreds persisted through millennia were, until recently, often discarded by excavators as useless, as were loom weights, which appeared to be nothing more than blobs of clay. Ancient texts, meanwhile, rarely mention the creation of textiles; moreover, those references that do exist use archaic, unrevealing terminology. Yet despite these obstacles, researchers have learned a great deal about ancient textiles and those who made them, and also about how to piece together a whole picture from many disparate sources of evidence. Technological advances in the analysis of archaeological remains provide much more information than was previously available, especially about minute remains. Successful modern methods include radiocarbon dating, infrared photography for seeing through dirt without removing it, isotope "fingerprinting" for tracing sources of raw materials, and thin-layer chromatography for analyzing dyes. As if in preparation for such advances, the field of archaeology has also undergone an important philosophical revolution in the past century. Once little more than a self-serving quest for artifacts to stock museums and private collections, the field has transformed itself into a scientific pursuit of knowledge about past cultures. As part of this process, archaeologists adopted the fundamental precept of preserving all objects, even those that have no immediately discernible value. Thus in the 1970s two researchers found the oldest known complete garment, a 5,000-year-old linen shirt, among a tumbled heap of dirty linens that had been preserved as part of the well-known Petrie collection decades before anyone began to study the history of textiles. The history of textiles and of the craftswomen who produced them has also advanced on a different front: recreating the actual production of cloth. Reconstructing and implementing ancient production methods provides a valuable way of generating and checking hypotheses. For example, these techniques made it possible to confirm that the excavated pieces of clay once considered useless in fact functioned as loom weights. Similarly, scholars have until recently been obliged to speculate as to which one of two statues of Athena, one large and one small, was adorned with a dress created by a group of Athenian women for a festival, as described in surviving texts. Because records show that it took nine months to produce the dress, scholars assumed it must have adorned the large statue. But by investigating the methods of production and the size of the looms used, researchers have ascertained that in fact a dress for the small statue would have taken nine months to produce.
200909_2-RC_1_3
[ "investigating the meanings of certain previously unintelligible technical terms in ancient texts", "tracing the sources of raw materials used in the production of certain fabrics", "constructing certain public museum displays concerning cloth-making", "verifying that a particular 5,000-year-old cloth was indeed a shirt", "exploring the issue of which of two statues of Athena was clothed with a particular garment" ]
4
The passage indicates that the re-creation of ancient techniques was used in which one of the following?
Traditional sources of evidence about ancient history are archaeological remains and surviving texts. Those investigating the crafts practiced by women in ancient times, however, often derive little information from these sources, and the archaeological record is particularly unavailing for the study of ancient textile production, as researchers are thwarted by the perishable nature of cloth. What shreds persisted through millennia were, until recently, often discarded by excavators as useless, as were loom weights, which appeared to be nothing more than blobs of clay. Ancient texts, meanwhile, rarely mention the creation of textiles; moreover, those references that do exist use archaic, unrevealing terminology. Yet despite these obstacles, researchers have learned a great deal about ancient textiles and those who made them, and also about how to piece together a whole picture from many disparate sources of evidence. Technological advances in the analysis of archaeological remains provide much more information than was previously available, especially about minute remains. Successful modern methods include radiocarbon dating, infrared photography for seeing through dirt without removing it, isotope "fingerprinting" for tracing sources of raw materials, and thin-layer chromatography for analyzing dyes. As if in preparation for such advances, the field of archaeology has also undergone an important philosophical revolution in the past century. Once little more than a self-serving quest for artifacts to stock museums and private collections, the field has transformed itself into a scientific pursuit of knowledge about past cultures. As part of this process, archaeologists adopted the fundamental precept of preserving all objects, even those that have no immediately discernible value. Thus in the 1970s two researchers found the oldest known complete garment, a 5,000-year-old linen shirt, among a tumbled heap of dirty linens that had been preserved as part of the well-known Petrie collection decades before anyone began to study the history of textiles. The history of textiles and of the craftswomen who produced them has also advanced on a different front: recreating the actual production of cloth. Reconstructing and implementing ancient production methods provides a valuable way of generating and checking hypotheses. For example, these techniques made it possible to confirm that the excavated pieces of clay once considered useless in fact functioned as loom weights. Similarly, scholars have until recently been obliged to speculate as to which one of two statues of Athena, one large and one small, was adorned with a dress created by a group of Athenian women for a festival, as described in surviving texts. Because records show that it took nine months to produce the dress, scholars assumed it must have adorned the large statue. But by investigating the methods of production and the size of the looms used, researchers have ascertained that in fact a dress for the small statue would have taken nine months to produce.
200909_2-RC_1_4
[ "ancient clay objects that cannot be identified as pieces of pottery by the researchers who unearth them", "historically significant pieces of cloth discovered in the course of an excavation", "the oldest known complete garment, which was found among other pieces of cloth in a collection", "re-creations of looms from which inferences about ancient weaving techniques can be made", "ancient accounts of the adornment of a statue of Athena with a dress made by Athenian women" ]
3
The author intends the term "traditional sources" (line 1) to exclude which one of the following?
Traditional sources of evidence about ancient history are archaeological remains and surviving texts. Those investigating the crafts practiced by women in ancient times, however, often derive little information from these sources, and the archaeological record is particularly unavailing for the study of ancient textile production, as researchers are thwarted by the perishable nature of cloth. What shreds persisted through millennia were, until recently, often discarded by excavators as useless, as were loom weights, which appeared to be nothing more than blobs of clay. Ancient texts, meanwhile, rarely mention the creation of textiles; moreover, those references that do exist use archaic, unrevealing terminology. Yet despite these obstacles, researchers have learned a great deal about ancient textiles and those who made them, and also about how to piece together a whole picture from many disparate sources of evidence. Technological advances in the analysis of archaeological remains provide much more information than was previously available, especially about minute remains. Successful modern methods include radiocarbon dating, infrared photography for seeing through dirt without removing it, isotope "fingerprinting" for tracing sources of raw materials, and thin-layer chromatography for analyzing dyes. As if in preparation for such advances, the field of archaeology has also undergone an important philosophical revolution in the past century. Once little more than a self-serving quest for artifacts to stock museums and private collections, the field has transformed itself into a scientific pursuit of knowledge about past cultures. As part of this process, archaeologists adopted the fundamental precept of preserving all objects, even those that have no immediately discernible value. Thus in the 1970s two researchers found the oldest known complete garment, a 5,000-year-old linen shirt, among a tumbled heap of dirty linens that had been preserved as part of the well-known Petrie collection decades before anyone began to study the history of textiles. The history of textiles and of the craftswomen who produced them has also advanced on a different front: recreating the actual production of cloth. Reconstructing and implementing ancient production methods provides a valuable way of generating and checking hypotheses. For example, these techniques made it possible to confirm that the excavated pieces of clay once considered useless in fact functioned as loom weights. Similarly, scholars have until recently been obliged to speculate as to which one of two statues of Athena, one large and one small, was adorned with a dress created by a group of Athenian women for a festival, as described in surviving texts. Because records show that it took nine months to produce the dress, scholars assumed it must have adorned the large statue. But by investigating the methods of production and the size of the looms used, researchers have ascertained that in fact a dress for the small statue would have taken nine months to produce.
200909_2-RC_1_5
[ "a defense of the controversial methods adopted by certain researchers in a particular discipline", "a set of recommendations to guide future activities in a particular field of inquiry", "an account of how a particular branch of research has successfully coped with certain difficulties", "a rejection of some commonly held views about the methodologies of a certain discipline", "a summary of the hypotheses advanced by researchers who have used innovative methods of investigation" ]
2
The passage as a whole functions primarily as
Traditional sources of evidence about ancient history are archaeological remains and surviving texts. Those investigating the crafts practiced by women in ancient times, however, often derive little information from these sources, and the archaeological record is particularly unavailing for the study of ancient textile production, as researchers are thwarted by the perishable nature of cloth. What shreds persisted through millennia were, until recently, often discarded by excavators as useless, as were loom weights, which appeared to be nothing more than blobs of clay. Ancient texts, meanwhile, rarely mention the creation of textiles; moreover, those references that do exist use archaic, unrevealing terminology. Yet despite these obstacles, researchers have learned a great deal about ancient textiles and those who made them, and also about how to piece together a whole picture from many disparate sources of evidence. Technological advances in the analysis of archaeological remains provide much more information than was previously available, especially about minute remains. Successful modern methods include radiocarbon dating, infrared photography for seeing through dirt without removing it, isotope "fingerprinting" for tracing sources of raw materials, and thin-layer chromatography for analyzing dyes. As if in preparation for such advances, the field of archaeology has also undergone an important philosophical revolution in the past century. Once little more than a self-serving quest for artifacts to stock museums and private collections, the field has transformed itself into a scientific pursuit of knowledge about past cultures. As part of this process, archaeologists adopted the fundamental precept of preserving all objects, even those that have no immediately discernible value. Thus in the 1970s two researchers found the oldest known complete garment, a 5,000-year-old linen shirt, among a tumbled heap of dirty linens that had been preserved as part of the well-known Petrie collection decades before anyone began to study the history of textiles. The history of textiles and of the craftswomen who produced them has also advanced on a different front: recreating the actual production of cloth. Reconstructing and implementing ancient production methods provides a valuable way of generating and checking hypotheses. For example, these techniques made it possible to confirm that the excavated pieces of clay once considered useless in fact functioned as loom weights. Similarly, scholars have until recently been obliged to speculate as to which one of two statues of Athena, one large and one small, was adorned with a dress created by a group of Athenian women for a festival, as described in surviving texts. Because records show that it took nine months to produce the dress, scholars assumed it must have adorned the large statue. But by investigating the methods of production and the size of the looms used, researchers have ascertained that in fact a dress for the small statue would have taken nine months to produce.
200909_2-RC_1_6
[ "an increased interest in the crafts practiced in the ancient world", "some archaeologists' adoption of textile conservation experts' preservation techniques", "innovative methods of restoring damaged artifacts", "the discovery of the oldest known complete garment", "archaeologists' policy of not discarding ancient objects that have no readily identifiable value" ]
4
According to the passage, which one of the following was an element in the transformation of archaeology in the past century?
Traditional sources of evidence about ancient history are archaeological remains and surviving texts. Those investigating the crafts practiced by women in ancient times, however, often derive little information from these sources, and the archaeological record is particularly unavailing for the study of ancient textile production, as researchers are thwarted by the perishable nature of cloth. What shreds persisted through millennia were, until recently, often discarded by excavators as useless, as were loom weights, which appeared to be nothing more than blobs of clay. Ancient texts, meanwhile, rarely mention the creation of textiles; moreover, those references that do exist use archaic, unrevealing terminology. Yet despite these obstacles, researchers have learned a great deal about ancient textiles and those who made them, and also about how to piece together a whole picture from many disparate sources of evidence. Technological advances in the analysis of archaeological remains provide much more information than was previously available, especially about minute remains. Successful modern methods include radiocarbon dating, infrared photography for seeing through dirt without removing it, isotope "fingerprinting" for tracing sources of raw materials, and thin-layer chromatography for analyzing dyes. As if in preparation for such advances, the field of archaeology has also undergone an important philosophical revolution in the past century. Once little more than a self-serving quest for artifacts to stock museums and private collections, the field has transformed itself into a scientific pursuit of knowledge about past cultures. As part of this process, archaeologists adopted the fundamental precept of preserving all objects, even those that have no immediately discernible value. Thus in the 1970s two researchers found the oldest known complete garment, a 5,000-year-old linen shirt, among a tumbled heap of dirty linens that had been preserved as part of the well-known Petrie collection decades before anyone began to study the history of textiles. The history of textiles and of the craftswomen who produced them has also advanced on a different front: recreating the actual production of cloth. Reconstructing and implementing ancient production methods provides a valuable way of generating and checking hypotheses. For example, these techniques made it possible to confirm that the excavated pieces of clay once considered useless in fact functioned as loom weights. Similarly, scholars have until recently been obliged to speculate as to which one of two statues of Athena, one large and one small, was adorned with a dress created by a group of Athenian women for a festival, as described in surviving texts. Because records show that it took nine months to produce the dress, scholars assumed it must have adorned the large statue. But by investigating the methods of production and the size of the looms used, researchers have ascertained that in fact a dress for the small statue would have taken nine months to produce.
200909_2-RC_1_7
[ "A particularly difficult archaeological problem is described in order to underscore the significance of new methods used to resolve that problem, which are described in the following paragraphs.", "A previously neglected body of archaeological evidence is described in order to cast doubt on received views regarding ancient cultures developed from conventional sources of evidence, as described in the following paragraphs.", "The fruitfulness of new technologically based methods of analysis is described in order to support the subsequent argument that apparently insignificant archaeological remains ought to be preserved for possible future research.", "The findings of recent archaeological research are outlined as the foundation for a claim advanced in the following paragraphs that the role of women in ancient cultures has been underestimated by archaeologists.", "A recently developed branch of archaeological research is described as evidence for the subsequent argument that other, more established branches of archaeology should take advantage of new technologies in their research." ]
0
Which one of the following most accurately describes the function of the first paragraph in relation to the rest of the passage?
This passage was adapted from articles published in the 1990s. The success that Nigerian-born computer scientist Philip Emeagwali (b.1954) has had in designing computers that solve real-world problems has been fueled by his willingness to reach beyond established paradigms and draw inspiration for his designs from nature. In the 1980s, Emeagwali achieved breakthroughs in the design of parallel computer systems. Whereas single computers work sequentially, making one calculation at a time, computers connected in parallel can process calculations simultaneously. In 1989, Emeagwali pioneered the use of massively parallel computers that used a network of thousands of smaller computers to solve what is considered one of the most computationally difficult problems: predicting the flow of oil through the subterranean geologic formations that make up oil fields. Until that time, supercomputers had been used for oil field calculations, but because these supercomputers worked sequentially, they were too slow and inefficient to accurately predict such extremely complex movements. To model oil field flow using a computer requires the simulation of the distribution of the oil at tens of thousands of locations throughout the field. At each location, hundreds of simultaneous calculations must be made at regular time intervals relating to such variables as temperature, direction of oil flow, viscosity, and pressure, as well as geologic properties of the basin holding the oil. In order to solve this problem, Emeagwali designed a massively parallel computer by using the Internet to connect to more than 65,000 smaller computers. One of the great difficulties of parallel computing is dividing up the tasks among the separate smaller computers so that they do not interfere with each other, and it was here that Emeagwali turned to natural processes for ideas, noting that tree species that survive today are those that, over the course of hundreds of millions of years, have developed branching patterns that have maximized the amount of sunlight gathered and the quantity of water and sap delivered. Emeagwali demonstrated that, for modeling certain phenomena such as subterranean oil flow, a network design based on the mathematical principle that underlies the branching structures of trees will enable a massively parallel computer to gather and broadcast the largest quantity of messages to its processing points in the shortest time. In 1996 Emeagwali had another breakthrough when he presented the design for a massively parallel computer that he claims will be powerful enough to predict global weather patterns a century in advance. The computer's design is based on the geometry of bees'honeycombs, which use an extremely efficient three-dimensional spacing. Emeagwali believes that computer scientists in the future will increasingly look to nature for elegant solutions to complex technical problems. This paradigm shift, he asserts, will enable us to better understand the systems evolved by nature and, thereby, to facilitate the evolution of human technology.
200909_2-RC_2_8
[ "Emeagwali's establishment of new computational paradigms has enabled parallel computer systems to solve a wide array of real-world problems that supercomputers cannot solve.", "Emeagwali has shown that scientists' allegiance to established paradigms has until now prevented the solution of many real-world computational problems that could otherwise have been solved with little difficulty.", "Emeagwali's discovery of the basic mathematical principles underlying natural systems has led to a growing use of parallel computer systems to solve complex real-world computational problems.", "Emeagwali has designed parallel computer systems that are modeled on natural systems and that are aimed at solving real-world computational problems that would be difficult to solve with more traditional designs.", "The paradigm shift initiated by Emeagwali's computer designs has made it more likely that scientists will in the future look to systems evolved by nature to facilitate the evolution of human technology." ]
3
Which one of the following most accurately expresses the main point of the passage?
This passage was adapted from articles published in the 1990s. The success that Nigerian-born computer scientist Philip Emeagwali (b.1954) has had in designing computers that solve real-world problems has been fueled by his willingness to reach beyond established paradigms and draw inspiration for his designs from nature. In the 1980s, Emeagwali achieved breakthroughs in the design of parallel computer systems. Whereas single computers work sequentially, making one calculation at a time, computers connected in parallel can process calculations simultaneously. In 1989, Emeagwali pioneered the use of massively parallel computers that used a network of thousands of smaller computers to solve what is considered one of the most computationally difficult problems: predicting the flow of oil through the subterranean geologic formations that make up oil fields. Until that time, supercomputers had been used for oil field calculations, but because these supercomputers worked sequentially, they were too slow and inefficient to accurately predict such extremely complex movements. To model oil field flow using a computer requires the simulation of the distribution of the oil at tens of thousands of locations throughout the field. At each location, hundreds of simultaneous calculations must be made at regular time intervals relating to such variables as temperature, direction of oil flow, viscosity, and pressure, as well as geologic properties of the basin holding the oil. In order to solve this problem, Emeagwali designed a massively parallel computer by using the Internet to connect to more than 65,000 smaller computers. One of the great difficulties of parallel computing is dividing up the tasks among the separate smaller computers so that they do not interfere with each other, and it was here that Emeagwali turned to natural processes for ideas, noting that tree species that survive today are those that, over the course of hundreds of millions of years, have developed branching patterns that have maximized the amount of sunlight gathered and the quantity of water and sap delivered. Emeagwali demonstrated that, for modeling certain phenomena such as subterranean oil flow, a network design based on the mathematical principle that underlies the branching structures of trees will enable a massively parallel computer to gather and broadcast the largest quantity of messages to its processing points in the shortest time. In 1996 Emeagwali had another breakthrough when he presented the design for a massively parallel computer that he claims will be powerful enough to predict global weather patterns a century in advance. The computer's design is based on the geometry of bees'honeycombs, which use an extremely efficient three-dimensional spacing. Emeagwali believes that computer scientists in the future will increasingly look to nature for elegant solutions to complex technical problems. This paradigm shift, he asserts, will enable us to better understand the systems evolved by nature and, thereby, to facilitate the evolution of human technology.
200909_2-RC_2_9
[ "Emeagwali's breakthroughs in computer design have begun to make computers that work sequentially obsolete.", "Emeagwali's first breakthrough in computer design came in response to a request by an oil company.", "Emeagwali was the first to use a massively parallel computer to predict the flow of oil in oil fields.", "Emeagwali was the first computer scientist to use nature as a model for human technology.", "Emeagwali was the first to apply parallel processing to solving real-world problems." ]
2
According to the passage, which one of the following is true?
This passage was adapted from articles published in the 1990s. The success that Nigerian-born computer scientist Philip Emeagwali (b.1954) has had in designing computers that solve real-world problems has been fueled by his willingness to reach beyond established paradigms and draw inspiration for his designs from nature. In the 1980s, Emeagwali achieved breakthroughs in the design of parallel computer systems. Whereas single computers work sequentially, making one calculation at a time, computers connected in parallel can process calculations simultaneously. In 1989, Emeagwali pioneered the use of massively parallel computers that used a network of thousands of smaller computers to solve what is considered one of the most computationally difficult problems: predicting the flow of oil through the subterranean geologic formations that make up oil fields. Until that time, supercomputers had been used for oil field calculations, but because these supercomputers worked sequentially, they were too slow and inefficient to accurately predict such extremely complex movements. To model oil field flow using a computer requires the simulation of the distribution of the oil at tens of thousands of locations throughout the field. At each location, hundreds of simultaneous calculations must be made at regular time intervals relating to such variables as temperature, direction of oil flow, viscosity, and pressure, as well as geologic properties of the basin holding the oil. In order to solve this problem, Emeagwali designed a massively parallel computer by using the Internet to connect to more than 65,000 smaller computers. One of the great difficulties of parallel computing is dividing up the tasks among the separate smaller computers so that they do not interfere with each other, and it was here that Emeagwali turned to natural processes for ideas, noting that tree species that survive today are those that, over the course of hundreds of millions of years, have developed branching patterns that have maximized the amount of sunlight gathered and the quantity of water and sap delivered. Emeagwali demonstrated that, for modeling certain phenomena such as subterranean oil flow, a network design based on the mathematical principle that underlies the branching structures of trees will enable a massively parallel computer to gather and broadcast the largest quantity of messages to its processing points in the shortest time. In 1996 Emeagwali had another breakthrough when he presented the design for a massively parallel computer that he claims will be powerful enough to predict global weather patterns a century in advance. The computer's design is based on the geometry of bees'honeycombs, which use an extremely efficient three-dimensional spacing. Emeagwali believes that computer scientists in the future will increasingly look to nature for elegant solutions to complex technical problems. This paradigm shift, he asserts, will enable us to better understand the systems evolved by nature and, thereby, to facilitate the evolution of human technology.
200909_2-RC_2_10
[ "Some natural systems have arrived at efficient solutions to problems that are analogous in significant ways to technical problems faced by computer scientists.", "Global weather is likely too complicated to be accurately predictable more than a few decades in advance.", "Most computer designs will in the future be inspired by natural systems.", "Massively parallel computers will eventually be practical enough to warrant their use even in relatively mundane computing tasks.", "The mathematical structure of branching trees is useful primarily for designing computer systems to predict the flow of oil through oil fields." ]
0
The passage most strongly suggests that Emeagwali holds which one of the following views?
This passage was adapted from articles published in the 1990s. The success that Nigerian-born computer scientist Philip Emeagwali (b.1954) has had in designing computers that solve real-world problems has been fueled by his willingness to reach beyond established paradigms and draw inspiration for his designs from nature. In the 1980s, Emeagwali achieved breakthroughs in the design of parallel computer systems. Whereas single computers work sequentially, making one calculation at a time, computers connected in parallel can process calculations simultaneously. In 1989, Emeagwali pioneered the use of massively parallel computers that used a network of thousands of smaller computers to solve what is considered one of the most computationally difficult problems: predicting the flow of oil through the subterranean geologic formations that make up oil fields. Until that time, supercomputers had been used for oil field calculations, but because these supercomputers worked sequentially, they were too slow and inefficient to accurately predict such extremely complex movements. To model oil field flow using a computer requires the simulation of the distribution of the oil at tens of thousands of locations throughout the field. At each location, hundreds of simultaneous calculations must be made at regular time intervals relating to such variables as temperature, direction of oil flow, viscosity, and pressure, as well as geologic properties of the basin holding the oil. In order to solve this problem, Emeagwali designed a massively parallel computer by using the Internet to connect to more than 65,000 smaller computers. One of the great difficulties of parallel computing is dividing up the tasks among the separate smaller computers so that they do not interfere with each other, and it was here that Emeagwali turned to natural processes for ideas, noting that tree species that survive today are those that, over the course of hundreds of millions of years, have developed branching patterns that have maximized the amount of sunlight gathered and the quantity of water and sap delivered. Emeagwali demonstrated that, for modeling certain phenomena such as subterranean oil flow, a network design based on the mathematical principle that underlies the branching structures of trees will enable a massively parallel computer to gather and broadcast the largest quantity of messages to its processing points in the shortest time. In 1996 Emeagwali had another breakthrough when he presented the design for a massively parallel computer that he claims will be powerful enough to predict global weather patterns a century in advance. The computer's design is based on the geometry of bees'honeycombs, which use an extremely efficient three-dimensional spacing. Emeagwali believes that computer scientists in the future will increasingly look to nature for elegant solutions to complex technical problems. This paradigm shift, he asserts, will enable us to better understand the systems evolved by nature and, thereby, to facilitate the evolution of human technology.
200909_2-RC_2_11
[ "They provide an example of an established paradigm that Emeagwali's work has challenged.", "They help explain why supercomputers are unable to accurately predict the movements of oil through underground geologic formations.", "They provide examples of a network design based on the mathematical principles underlying the branching structures of trees.", "They describe a mathematical model that Emeagwali used in order to understand a natural system.", "They provide specific examples of a paradigm shift that will help scientists understand certain systems evolved by nature." ]
1
Which one of the following most accurately describes the function of the first two sentences of the second paragraph?
This passage was adapted from articles published in the 1990s. The success that Nigerian-born computer scientist Philip Emeagwali (b.1954) has had in designing computers that solve real-world problems has been fueled by his willingness to reach beyond established paradigms and draw inspiration for his designs from nature. In the 1980s, Emeagwali achieved breakthroughs in the design of parallel computer systems. Whereas single computers work sequentially, making one calculation at a time, computers connected in parallel can process calculations simultaneously. In 1989, Emeagwali pioneered the use of massively parallel computers that used a network of thousands of smaller computers to solve what is considered one of the most computationally difficult problems: predicting the flow of oil through the subterranean geologic formations that make up oil fields. Until that time, supercomputers had been used for oil field calculations, but because these supercomputers worked sequentially, they were too slow and inefficient to accurately predict such extremely complex movements. To model oil field flow using a computer requires the simulation of the distribution of the oil at tens of thousands of locations throughout the field. At each location, hundreds of simultaneous calculations must be made at regular time intervals relating to such variables as temperature, direction of oil flow, viscosity, and pressure, as well as geologic properties of the basin holding the oil. In order to solve this problem, Emeagwali designed a massively parallel computer by using the Internet to connect to more than 65,000 smaller computers. One of the great difficulties of parallel computing is dividing up the tasks among the separate smaller computers so that they do not interfere with each other, and it was here that Emeagwali turned to natural processes for ideas, noting that tree species that survive today are those that, over the course of hundreds of millions of years, have developed branching patterns that have maximized the amount of sunlight gathered and the quantity of water and sap delivered. Emeagwali demonstrated that, for modeling certain phenomena such as subterranean oil flow, a network design based on the mathematical principle that underlies the branching structures of trees will enable a massively parallel computer to gather and broadcast the largest quantity of messages to its processing points in the shortest time. In 1996 Emeagwali had another breakthrough when he presented the design for a massively parallel computer that he claims will be powerful enough to predict global weather patterns a century in advance. The computer's design is based on the geometry of bees'honeycombs, which use an extremely efficient three-dimensional spacing. Emeagwali believes that computer scientists in the future will increasingly look to nature for elegant solutions to complex technical problems. This paradigm shift, he asserts, will enable us to better understand the systems evolved by nature and, thereby, to facilitate the evolution of human technology.
200909_2-RC_2_12
[ "Until recently, computer scientists have had very limited awareness of many of the mathematical principles that have been shown to underlie a wide variety of natural processes.", "Some of the variables affecting global weather patterns have yet to be discovered by scientists who study these patterns.", "Computer designs for the prediction of natural phenomena tend to be more successful when those phenomena are not affected by human activities.", "Some of the mathematical principles underlying Emeagwali's model of oil field flow also underlie his designs for other massively parallel computer systems.", "Underlying the designs for many traditional technologies are mathematical principles of which the designers of those technologies were not explicitly aware." ]
0
Which one of the following, if true, would provide the most support for Emeagwali's prediction mentioned in lines 55–58?
This passage was adapted from articles published in the 1990s. The success that Nigerian-born computer scientist Philip Emeagwali (b.1954) has had in designing computers that solve real-world problems has been fueled by his willingness to reach beyond established paradigms and draw inspiration for his designs from nature. In the 1980s, Emeagwali achieved breakthroughs in the design of parallel computer systems. Whereas single computers work sequentially, making one calculation at a time, computers connected in parallel can process calculations simultaneously. In 1989, Emeagwali pioneered the use of massively parallel computers that used a network of thousands of smaller computers to solve what is considered one of the most computationally difficult problems: predicting the flow of oil through the subterranean geologic formations that make up oil fields. Until that time, supercomputers had been used for oil field calculations, but because these supercomputers worked sequentially, they were too slow and inefficient to accurately predict such extremely complex movements. To model oil field flow using a computer requires the simulation of the distribution of the oil at tens of thousands of locations throughout the field. At each location, hundreds of simultaneous calculations must be made at regular time intervals relating to such variables as temperature, direction of oil flow, viscosity, and pressure, as well as geologic properties of the basin holding the oil. In order to solve this problem, Emeagwali designed a massively parallel computer by using the Internet to connect to more than 65,000 smaller computers. One of the great difficulties of parallel computing is dividing up the tasks among the separate smaller computers so that they do not interfere with each other, and it was here that Emeagwali turned to natural processes for ideas, noting that tree species that survive today are those that, over the course of hundreds of millions of years, have developed branching patterns that have maximized the amount of sunlight gathered and the quantity of water and sap delivered. Emeagwali demonstrated that, for modeling certain phenomena such as subterranean oil flow, a network design based on the mathematical principle that underlies the branching structures of trees will enable a massively parallel computer to gather and broadcast the largest quantity of messages to its processing points in the shortest time. In 1996 Emeagwali had another breakthrough when he presented the design for a massively parallel computer that he claims will be powerful enough to predict global weather patterns a century in advance. The computer's design is based on the geometry of bees'honeycombs, which use an extremely efficient three-dimensional spacing. Emeagwali believes that computer scientists in the future will increasingly look to nature for elegant solutions to complex technical problems. This paradigm shift, he asserts, will enable us to better understand the systems evolved by nature and, thereby, to facilitate the evolution of human technology.
200909_2-RC_2_13
[ "supercomputers are sufficiently powerful to handle most computational problems, including most problems arising from oil production", "the possibility of using a network of smaller computers to solve computationally difficult problems had not yet been considered", "the general public was not yet aware of the existence or vast capabilities of the Internet", "oil companies had not yet perceived the need for modeling the flow of oil in subterranean fields", "smaller computers can interfere with one another when they are connected together in parallel to solve a computationally difficult problem" ]
4
It can be inferred from the passage that one of the reasons massively parallel computers had not been used to model oil field flow prior to 1989 is that
Proponents of the tangible-object theory of copyright argue that copyright and similar intellectual-property rights can be explained as logical extensions of the right to own concrete, tangible objects. This view depends on the claim that every copyrightable work can be manifested in some physical form, such as a manuscript or a videotape. It also accepts the premise that ownership of an object confers a number of rights on the owner, who may essentially do whatever he or she pleases with the object to the extent that this does not violate other people's rights. One may, for example, hide or display the object, copy it, or destroy it. One may also transfer ownership of it to another. In creating a new and original object from materials that one owns, one becomes the owner of that object and thereby acquires all of the rights that ownership entails. But if the owner transfers ownership of the object, the full complement of rights is not necessarily transferred to the new owner; instead, the original owner may retain one or more of these rights. This notion of retained rights is common in many areas of law; for example, the seller of a piece of land may retain certain rights to the land in the form of easements or building restrictions. Applying the notion of retained rights to the domain of intellectual property, theorists argue that copyrighting a work secures official recognition of one's intention to retain certain rights to that work. Among the rights typically retained by the original producer of an object such as a literary manuscript or a musical score would be the right to copy the object for profit and the right to use it as a guide for the production of similar or analogous things—for example, a public performance of a musical score. According to proponents of the tangible-object theory, its chief advantage is that it justifies intellectual property rights without recourse to the widely accepted but problematic supposition that one can own abstract, intangible things such as ideas. But while this account seems plausible for copyrightable entities that do, in fact, have enduring tangible forms, it cannot accommodate the standard assumption that such evanescent things as live broadcasts of sporting events can be copyrighted. More importantly, it does not acknowledge that in many cases the work of conceiving ideas is more crucial and more valuable than that of putting them into tangible form. Suppose that a poet dictates a new poem to a friend, who writes it down on paper that the friend has supplied. The creator of the tangible object in this case is not the poet but the friend, and there would seem to be no ground for the poet's claiming copyright unless the poet can be said to already own the ideas expressed in the work.
200909_2-RC_3_14
[ "Copyright and other intellectual-property rights can be explained as logical extensions of the right to own concrete objects.", "Attempts to explain copyright and similar intellectual-property rights purely in terms of rights to ownership of physical objects are ultimately misguided.", "Copyrighting a work amounts to securing official recognition of one's intention to retain certain rights to that work.", "Explanations of copyright and other intellectual-property rights in terms of rights to ownership of tangible objects fail to consider the argument that ideas should be allowed to circulate freely.", "Under the tangible-object theory of intellectual property, rights of ownership are straightforwardly applicable to both ideas and physical objects." ]
1
Which one of the following most accurately expresses the main point of the passage?
Proponents of the tangible-object theory of copyright argue that copyright and similar intellectual-property rights can be explained as logical extensions of the right to own concrete, tangible objects. This view depends on the claim that every copyrightable work can be manifested in some physical form, such as a manuscript or a videotape. It also accepts the premise that ownership of an object confers a number of rights on the owner, who may essentially do whatever he or she pleases with the object to the extent that this does not violate other people's rights. One may, for example, hide or display the object, copy it, or destroy it. One may also transfer ownership of it to another. In creating a new and original object from materials that one owns, one becomes the owner of that object and thereby acquires all of the rights that ownership entails. But if the owner transfers ownership of the object, the full complement of rights is not necessarily transferred to the new owner; instead, the original owner may retain one or more of these rights. This notion of retained rights is common in many areas of law; for example, the seller of a piece of land may retain certain rights to the land in the form of easements or building restrictions. Applying the notion of retained rights to the domain of intellectual property, theorists argue that copyrighting a work secures official recognition of one's intention to retain certain rights to that work. Among the rights typically retained by the original producer of an object such as a literary manuscript or a musical score would be the right to copy the object for profit and the right to use it as a guide for the production of similar or analogous things—for example, a public performance of a musical score. According to proponents of the tangible-object theory, its chief advantage is that it justifies intellectual property rights without recourse to the widely accepted but problematic supposition that one can own abstract, intangible things such as ideas. But while this account seems plausible for copyrightable entities that do, in fact, have enduring tangible forms, it cannot accommodate the standard assumption that such evanescent things as live broadcasts of sporting events can be copyrighted. More importantly, it does not acknowledge that in many cases the work of conceiving ideas is more crucial and more valuable than that of putting them into tangible form. Suppose that a poet dictates a new poem to a friend, who writes it down on paper that the friend has supplied. The creator of the tangible object in this case is not the poet but the friend, and there would seem to be no ground for the poet's claiming copyright unless the poet can be said to already own the ideas expressed in the work.
200909_2-RC_3_15
[ "any work entitled to intellectual-property protection can be expressed in physical form", "only the original creator of an intellectual work can hold the copyright for that work", "the work of putting ideas into tangible form is more crucial and more valuable than the work of conceiving those ideas", "in a few cases, it is necessary to recognize the right to own abstract, intangible things", "the owner of an item of intellectual property may legally destroy it" ]
0
According to the passage, the theory that copyright and other intellectual-property rights can be construed as logical extensions of the right to own concrete, tangible objects depends on the claim that
Proponents of the tangible-object theory of copyright argue that copyright and similar intellectual-property rights can be explained as logical extensions of the right to own concrete, tangible objects. This view depends on the claim that every copyrightable work can be manifested in some physical form, such as a manuscript or a videotape. It also accepts the premise that ownership of an object confers a number of rights on the owner, who may essentially do whatever he or she pleases with the object to the extent that this does not violate other people's rights. One may, for example, hide or display the object, copy it, or destroy it. One may also transfer ownership of it to another. In creating a new and original object from materials that one owns, one becomes the owner of that object and thereby acquires all of the rights that ownership entails. But if the owner transfers ownership of the object, the full complement of rights is not necessarily transferred to the new owner; instead, the original owner may retain one or more of these rights. This notion of retained rights is common in many areas of law; for example, the seller of a piece of land may retain certain rights to the land in the form of easements or building restrictions. Applying the notion of retained rights to the domain of intellectual property, theorists argue that copyrighting a work secures official recognition of one's intention to retain certain rights to that work. Among the rights typically retained by the original producer of an object such as a literary manuscript or a musical score would be the right to copy the object for profit and the right to use it as a guide for the production of similar or analogous things—for example, a public performance of a musical score. According to proponents of the tangible-object theory, its chief advantage is that it justifies intellectual property rights without recourse to the widely accepted but problematic supposition that one can own abstract, intangible things such as ideas. But while this account seems plausible for copyrightable entities that do, in fact, have enduring tangible forms, it cannot accommodate the standard assumption that such evanescent things as live broadcasts of sporting events can be copyrighted. More importantly, it does not acknowledge that in many cases the work of conceiving ideas is more crucial and more valuable than that of putting them into tangible form. Suppose that a poet dictates a new poem to a friend, who writes it down on paper that the friend has supplied. The creator of the tangible object in this case is not the poet but the friend, and there would seem to be no ground for the poet's claiming copyright unless the poet can be said to already own the ideas expressed in the work.
200909_2-RC_3_16
[ "Do proponents of the tangible-object theory of intellectual property advocate any changes in existing laws relating to copyright?", "Do proponents of the tangible-object theory of intellectual property hold that ownership of anything besides real estate can involve retained rights?", "Has the tangible-object theory of intellectual property influenced the ways in which copyright cases or other cases involving issues of intellectual property are decided in the courts?", "Does existing copyright law provide protection against unauthorized copying of manuscripts and musical scores in cases in which their creators have not officially applied for copyright protection?", "Are there standard procedures governing the transfer of intellectual property that are common to most legal systems?" ]
1
The passage most directly answers which one of the following questions?
Proponents of the tangible-object theory of copyright argue that copyright and similar intellectual-property rights can be explained as logical extensions of the right to own concrete, tangible objects. This view depends on the claim that every copyrightable work can be manifested in some physical form, such as a manuscript or a videotape. It also accepts the premise that ownership of an object confers a number of rights on the owner, who may essentially do whatever he or she pleases with the object to the extent that this does not violate other people's rights. One may, for example, hide or display the object, copy it, or destroy it. One may also transfer ownership of it to another. In creating a new and original object from materials that one owns, one becomes the owner of that object and thereby acquires all of the rights that ownership entails. But if the owner transfers ownership of the object, the full complement of rights is not necessarily transferred to the new owner; instead, the original owner may retain one or more of these rights. This notion of retained rights is common in many areas of law; for example, the seller of a piece of land may retain certain rights to the land in the form of easements or building restrictions. Applying the notion of retained rights to the domain of intellectual property, theorists argue that copyrighting a work secures official recognition of one's intention to retain certain rights to that work. Among the rights typically retained by the original producer of an object such as a literary manuscript or a musical score would be the right to copy the object for profit and the right to use it as a guide for the production of similar or analogous things—for example, a public performance of a musical score. According to proponents of the tangible-object theory, its chief advantage is that it justifies intellectual property rights without recourse to the widely accepted but problematic supposition that one can own abstract, intangible things such as ideas. But while this account seems plausible for copyrightable entities that do, in fact, have enduring tangible forms, it cannot accommodate the standard assumption that such evanescent things as live broadcasts of sporting events can be copyrighted. More importantly, it does not acknowledge that in many cases the work of conceiving ideas is more crucial and more valuable than that of putting them into tangible form. Suppose that a poet dictates a new poem to a friend, who writes it down on paper that the friend has supplied. The creator of the tangible object in this case is not the poet but the friend, and there would seem to be no ground for the poet's claiming copyright unless the poet can be said to already own the ideas expressed in the work.
200909_2-RC_3_17
[ "Only the engineer is entitled to claim the invention as intellectual property.", "Only the inventor is entitled to claim the invention as intellectual property.", "The inventor and the engineer are equally entitled to claim the invention as intellectual property.", "The engineer is entitled to claim the invention as intellectual property, but only if the inventor retains the right to all profits generated by the invention.", "The inventor is entitled to claim the invention as intellectual property, but only if the engineer retains the right to all profits generated by the invention." ]
0
Suppose an inventor describes an innovative idea for an invention to an engineer, who volunteers to draft specifications for a prototype and then produces the prototype using the engineer's own materials. Which one of the following statements would apply to this case under the tangible-object theory of intellectual property, as the author describes that theory?
Proponents of the tangible-object theory of copyright argue that copyright and similar intellectual-property rights can be explained as logical extensions of the right to own concrete, tangible objects. This view depends on the claim that every copyrightable work can be manifested in some physical form, such as a manuscript or a videotape. It also accepts the premise that ownership of an object confers a number of rights on the owner, who may essentially do whatever he or she pleases with the object to the extent that this does not violate other people's rights. One may, for example, hide or display the object, copy it, or destroy it. One may also transfer ownership of it to another. In creating a new and original object from materials that one owns, one becomes the owner of that object and thereby acquires all of the rights that ownership entails. But if the owner transfers ownership of the object, the full complement of rights is not necessarily transferred to the new owner; instead, the original owner may retain one or more of these rights. This notion of retained rights is common in many areas of law; for example, the seller of a piece of land may retain certain rights to the land in the form of easements or building restrictions. Applying the notion of retained rights to the domain of intellectual property, theorists argue that copyrighting a work secures official recognition of one's intention to retain certain rights to that work. Among the rights typically retained by the original producer of an object such as a literary manuscript or a musical score would be the right to copy the object for profit and the right to use it as a guide for the production of similar or analogous things—for example, a public performance of a musical score. According to proponents of the tangible-object theory, its chief advantage is that it justifies intellectual property rights without recourse to the widely accepted but problematic supposition that one can own abstract, intangible things such as ideas. But while this account seems plausible for copyrightable entities that do, in fact, have enduring tangible forms, it cannot accommodate the standard assumption that such evanescent things as live broadcasts of sporting events can be copyrighted. More importantly, it does not acknowledge that in many cases the work of conceiving ideas is more crucial and more valuable than that of putting them into tangible form. Suppose that a poet dictates a new poem to a friend, who writes it down on paper that the friend has supplied. The creator of the tangible object in this case is not the poet but the friend, and there would seem to be no ground for the poet's claiming copyright unless the poet can be said to already own the ideas expressed in the work.
200909_2-RC_3_18
[ "A literary work cannot receive copyright protection unless it exists in an edition produced by an established publisher.", "Most legal systems explicitly rely on the tangible-object theory of intellectual property in order to avoid asserting that one can own abstract things.", "Copyright protects the right to copy for profit, but not the right to copy for other reasons.", "Some works deserving of copyright protection simply cannot be manifested as concrete, tangible objects.", "To afford patent protection for inventions, the law need not invoke the notion of inventors' ownership of abstract ideas." ]
4
Legal theorists supporting the tangible-object theory of intellectual property are most likely to believe which one of the following?
Proponents of the tangible-object theory of copyright argue that copyright and similar intellectual-property rights can be explained as logical extensions of the right to own concrete, tangible objects. This view depends on the claim that every copyrightable work can be manifested in some physical form, such as a manuscript or a videotape. It also accepts the premise that ownership of an object confers a number of rights on the owner, who may essentially do whatever he or she pleases with the object to the extent that this does not violate other people's rights. One may, for example, hide or display the object, copy it, or destroy it. One may also transfer ownership of it to another. In creating a new and original object from materials that one owns, one becomes the owner of that object and thereby acquires all of the rights that ownership entails. But if the owner transfers ownership of the object, the full complement of rights is not necessarily transferred to the new owner; instead, the original owner may retain one or more of these rights. This notion of retained rights is common in many areas of law; for example, the seller of a piece of land may retain certain rights to the land in the form of easements or building restrictions. Applying the notion of retained rights to the domain of intellectual property, theorists argue that copyrighting a work secures official recognition of one's intention to retain certain rights to that work. Among the rights typically retained by the original producer of an object such as a literary manuscript or a musical score would be the right to copy the object for profit and the right to use it as a guide for the production of similar or analogous things—for example, a public performance of a musical score. According to proponents of the tangible-object theory, its chief advantage is that it justifies intellectual property rights without recourse to the widely accepted but problematic supposition that one can own abstract, intangible things such as ideas. But while this account seems plausible for copyrightable entities that do, in fact, have enduring tangible forms, it cannot accommodate the standard assumption that such evanescent things as live broadcasts of sporting events can be copyrighted. More importantly, it does not acknowledge that in many cases the work of conceiving ideas is more crucial and more valuable than that of putting them into tangible form. Suppose that a poet dictates a new poem to a friend, who writes it down on paper that the friend has supplied. The creator of the tangible object in this case is not the poet but the friend, and there would seem to be no ground for the poet's claiming copyright unless the poet can be said to already own the ideas expressed in the work.
200909_2-RC_3_19
[ "In most transactions involving the transfer of non-intellectual property, at least some rights of ownership are retained by the seller.", "The notion of retained rights of ownership is currently applied to only those areas of law that do not involve intellectual property.", "The idea that ownership of the right to copy an item for profit can be transferred is compatible with a tangible-object theory of intellectual property.", "Ownership of intellectual property is sufficiently protected by the provisions that, under many legal systems, apply to ownership of material things such as land.", "Protection of computer programs under intellectual-property law is justifiable only if the programs are likely to be used as a guide for the production of similar or analogous programs." ]
2
The passage provides the most support for inferring which one of the following statements?
Proponents of the tangible-object theory of copyright argue that copyright and similar intellectual-property rights can be explained as logical extensions of the right to own concrete, tangible objects. This view depends on the claim that every copyrightable work can be manifested in some physical form, such as a manuscript or a videotape. It also accepts the premise that ownership of an object confers a number of rights on the owner, who may essentially do whatever he or she pleases with the object to the extent that this does not violate other people's rights. One may, for example, hide or display the object, copy it, or destroy it. One may also transfer ownership of it to another. In creating a new and original object from materials that one owns, one becomes the owner of that object and thereby acquires all of the rights that ownership entails. But if the owner transfers ownership of the object, the full complement of rights is not necessarily transferred to the new owner; instead, the original owner may retain one or more of these rights. This notion of retained rights is common in many areas of law; for example, the seller of a piece of land may retain certain rights to the land in the form of easements or building restrictions. Applying the notion of retained rights to the domain of intellectual property, theorists argue that copyrighting a work secures official recognition of one's intention to retain certain rights to that work. Among the rights typically retained by the original producer of an object such as a literary manuscript or a musical score would be the right to copy the object for profit and the right to use it as a guide for the production of similar or analogous things—for example, a public performance of a musical score. According to proponents of the tangible-object theory, its chief advantage is that it justifies intellectual property rights without recourse to the widely accepted but problematic supposition that one can own abstract, intangible things such as ideas. But while this account seems plausible for copyrightable entities that do, in fact, have enduring tangible forms, it cannot accommodate the standard assumption that such evanescent things as live broadcasts of sporting events can be copyrighted. More importantly, it does not acknowledge that in many cases the work of conceiving ideas is more crucial and more valuable than that of putting them into tangible form. Suppose that a poet dictates a new poem to a friend, who writes it down on paper that the friend has supplied. The creator of the tangible object in this case is not the poet but the friend, and there would seem to be no ground for the poet's claiming copyright unless the poet can be said to already own the ideas expressed in the work.
200909_2-RC_3_20
[ "Theorists who suggest that the notion of retained rights is applicable to intellectual property do not fully understand what it means to transfer ownership of property.", "If a work does not exist in a concrete, tangible form, there is no valid theoretical basis for claiming that it should have copyright protection.", "Under existing statutes, creators of original tangible works that have intellectual or artistic significance generally do not have the legal right to own the abstract ideas embodied in those works.", "An adequate theoretical justification of copyright would likely presuppose that a work's creator originally owns the ideas embodied in that work.", "It is common, but incorrect, to assume that such evanescent things as live broadcasts of sporting events can be copyrighted." ]
3
It can be inferred that the author of the passage is most likely to believe which one of the following?
Passage AIn music, a certain complexity of sounds can be expected to have a positive effect on the listener. A single, pure tone is not that interesting to explore; a measure of intricacy is required to excite human curiosity. Sounds that are too complex or disorganized, however, tend to be overwhelming. We prefer some sort of coherence, a principle that connects the various sounds and makes them comprehensible. In this respect, music is like human language. Single sounds are in most cases not sufficient to convey meaning in speech, whereas when put together in a sequence they form words and sentences. Likewise, if the tones in music are not perceived to be tied together sequentially or rhythmically—for example, in what is commonly called melody— listeners are less likely to feel any emotional connection or to show appreciation. Certain music can also have a relaxing effect. The fact that such music tends to be continuous and rhythmical suggests a possible explanation for this effect. In a natural environment, danger tends to be accompanied by sudden, unexpected sounds. Thus, a background of constant noise suggests peaceful conditions; discontinuous sounds demand more attention. Even soft discontinuous sounds that we consciously realize do not signal danger can be disturbing—for example, the erratic dripping of a leaky tap. A continuous sound, particularly one that is judged to be safe, relaxes the brain. Passage B There are certain elements within music, such as a change of melodic line or rhythm, that create expectations about the future development of the music. The expectation the listener has about the further course of musical events is a key determinant for the experience of "musical emotions." Music creates expectations that, if not immediately satisfied, create tension. Emotion is experienced in relation to the buildup and release of tension. The more elaborate the buildup of tension, the more intense the emotions that will be experienon of the expected musical course, depending on one's personal involvement, causes the search for an explanation. This results from a "mismatch" between one's musical expectation and the actual course of the music. Negative emotions will be the result of an extreme mismatch between expectations and experience. Positive emotions result if the converse happens. When we listen to music, we take into account factors such as the complexity and novelty of the music. The degree to which the music sounds familiar determines whether the music is experienced as pleasurable or uncomfortable. The pleasure experienced is minimal when the music is entirely new to the listener, increases with increasing familiarity, and decreases again when the music is totally known. Musical preference is based on one's desire to maintain a constant level of certain preferable emotions. As such, a trained listener will have a greater preference for complex melodies than will a naive listener, as the threshold for experiencing emotion is higher.
200909_2-RC_4_21
[ "continuous sound", "tension", "language", "improvisation", "complexity" ]
4
Which one of the following concepts is linked to positive musical experiences in both passages?
Passage AIn music, a certain complexity of sounds can be expected to have a positive effect on the listener. A single, pure tone is not that interesting to explore; a measure of intricacy is required to excite human curiosity. Sounds that are too complex or disorganized, however, tend to be overwhelming. We prefer some sort of coherence, a principle that connects the various sounds and makes them comprehensible. In this respect, music is like human language. Single sounds are in most cases not sufficient to convey meaning in speech, whereas when put together in a sequence they form words and sentences. Likewise, if the tones in music are not perceived to be tied together sequentially or rhythmically—for example, in what is commonly called melody— listeners are less likely to feel any emotional connection or to show appreciation. Certain music can also have a relaxing effect. The fact that such music tends to be continuous and rhythmical suggests a possible explanation for this effect. In a natural environment, danger tends to be accompanied by sudden, unexpected sounds. Thus, a background of constant noise suggests peaceful conditions; discontinuous sounds demand more attention. Even soft discontinuous sounds that we consciously realize do not signal danger can be disturbing—for example, the erratic dripping of a leaky tap. A continuous sound, particularly one that is judged to be safe, relaxes the brain. Passage B There are certain elements within music, such as a change of melodic line or rhythm, that create expectations about the future development of the music. The expectation the listener has about the further course of musical events is a key determinant for the experience of "musical emotions." Music creates expectations that, if not immediately satisfied, create tension. Emotion is experienced in relation to the buildup and release of tension. The more elaborate the buildup of tension, the more intense the emotions that will be experienon of the expected musical course, depending on one's personal involvement, causes the search for an explanation. This results from a "mismatch" between one's musical expectation and the actual course of the music. Negative emotions will be the result of an extreme mismatch between expectations and experience. Positive emotions result if the converse happens. When we listen to music, we take into account factors such as the complexity and novelty of the music. The degree to which the music sounds familiar determines whether the music is experienced as pleasurable or uncomfortable. The pleasure experienced is minimal when the music is entirely new to the listener, increases with increasing familiarity, and decreases again when the music is totally known. Musical preference is based on one's desire to maintain a constant level of certain preferable emotions. As such, a trained listener will have a greater preference for complex melodies than will a naive listener, as the threshold for experiencing emotion is higher.
200909_2-RC_4_22
[ "the theoretical underpinnings of how music is composed", "the nature of the conceptual difference between music and discontinuous sound", "the impact music can have on human emotional states", "the most effective techniques for teaching novices to appreciate complex music", "the influence music has had on the development of spoken language" ]
2
The passages most strongly suggest that both are targeting an audience that is interested in which one of the following?
Passage AIn music, a certain complexity of sounds can be expected to have a positive effect on the listener. A single, pure tone is not that interesting to explore; a measure of intricacy is required to excite human curiosity. Sounds that are too complex or disorganized, however, tend to be overwhelming. We prefer some sort of coherence, a principle that connects the various sounds and makes them comprehensible. In this respect, music is like human language. Single sounds are in most cases not sufficient to convey meaning in speech, whereas when put together in a sequence they form words and sentences. Likewise, if the tones in music are not perceived to be tied together sequentially or rhythmically—for example, in what is commonly called melody— listeners are less likely to feel any emotional connection or to show appreciation. Certain music can also have a relaxing effect. The fact that such music tends to be continuous and rhythmical suggests a possible explanation for this effect. In a natural environment, danger tends to be accompanied by sudden, unexpected sounds. Thus, a background of constant noise suggests peaceful conditions; discontinuous sounds demand more attention. Even soft discontinuous sounds that we consciously realize do not signal danger can be disturbing—for example, the erratic dripping of a leaky tap. A continuous sound, particularly one that is judged to be safe, relaxes the brain. Passage B There are certain elements within music, such as a change of melodic line or rhythm, that create expectations about the future development of the music. The expectation the listener has about the further course of musical events is a key determinant for the experience of "musical emotions." Music creates expectations that, if not immediately satisfied, create tension. Emotion is experienced in relation to the buildup and release of tension. The more elaborate the buildup of tension, the more intense the emotions that will be experienon of the expected musical course, depending on one's personal involvement, causes the search for an explanation. This results from a "mismatch" between one's musical expectation and the actual course of the music. Negative emotions will be the result of an extreme mismatch between expectations and experience. Positive emotions result if the converse happens. When we listen to music, we take into account factors such as the complexity and novelty of the music. The degree to which the music sounds familiar determines whether the music is experienced as pleasurable or uncomfortable. The pleasure experienced is minimal when the music is entirely new to the listener, increases with increasing familiarity, and decreases again when the music is totally known. Musical preference is based on one's desire to maintain a constant level of certain preferable emotions. As such, a trained listener will have a greater preference for complex melodies than will a naive listener, as the threshold for experiencing emotion is higher.
200909_2-RC_4_23
[ "the preference of some people for falling asleep to white noise, such as the sound of an electric fan", "the preference of many moviegoers for movies with plots that are clear and easy to follow", "the preference of many diners for restaurants that serve large portions", "the preference of many young listeners for fast music over slower music", "the preference of most children for sweet foods over bitter foods" ]
1
Which one of the following describes a preference that is most analogous to the preference mentioned in the first paragraph of passage A?
Passage AIn music, a certain complexity of sounds can be expected to have a positive effect on the listener. A single, pure tone is not that interesting to explore; a measure of intricacy is required to excite human curiosity. Sounds that are too complex or disorganized, however, tend to be overwhelming. We prefer some sort of coherence, a principle that connects the various sounds and makes them comprehensible. In this respect, music is like human language. Single sounds are in most cases not sufficient to convey meaning in speech, whereas when put together in a sequence they form words and sentences. Likewise, if the tones in music are not perceived to be tied together sequentially or rhythmically—for example, in what is commonly called melody— listeners are less likely to feel any emotional connection or to show appreciation. Certain music can also have a relaxing effect. The fact that such music tends to be continuous and rhythmical suggests a possible explanation for this effect. In a natural environment, danger tends to be accompanied by sudden, unexpected sounds. Thus, a background of constant noise suggests peaceful conditions; discontinuous sounds demand more attention. Even soft discontinuous sounds that we consciously realize do not signal danger can be disturbing—for example, the erratic dripping of a leaky tap. A continuous sound, particularly one that is judged to be safe, relaxes the brain. Passage B There are certain elements within music, such as a change of melodic line or rhythm, that create expectations about the future development of the music. The expectation the listener has about the further course of musical events is a key determinant for the experience of "musical emotions." Music creates expectations that, if not immediately satisfied, create tension. Emotion is experienced in relation to the buildup and release of tension. The more elaborate the buildup of tension, the more intense the emotions that will be experienon of the expected musical course, depending on one's personal involvement, causes the search for an explanation. This results from a "mismatch" between one's musical expectation and the actual course of the music. Negative emotions will be the result of an extreme mismatch between expectations and experience. Positive emotions result if the converse happens. When we listen to music, we take into account factors such as the complexity and novelty of the music. The degree to which the music sounds familiar determines whether the music is experienced as pleasurable or uncomfortable. The pleasure experienced is minimal when the music is entirely new to the listener, increases with increasing familiarity, and decreases again when the music is totally known. Musical preference is based on one's desire to maintain a constant level of certain preferable emotions. As such, a trained listener will have a greater preference for complex melodies than will a naive listener, as the threshold for experiencing emotion is higher.
200909_2-RC_4_24
[ "The type of musical emotion experienced by a listener is determined by the level to which the listener's expectations are satisfied.", "Trained listeners are more able to consciously manipulate their own emotional experiences of complex music than are naive listeners.", "If the development of a piece of music is greatly at odds with the listener's musical expectations, then the listener will experience negative emotions.", "Listeners can learn to appreciate changes in melodic line and other musical complexities.", "Music that is experienced by listeners as relaxing usually produces a buildup and release of tension in those listeners." ]
0
Which one of the following most accurately expresses the main point of passage B?
Passage AIn music, a certain complexity of sounds can be expected to have a positive effect on the listener. A single, pure tone is not that interesting to explore; a measure of intricacy is required to excite human curiosity. Sounds that are too complex or disorganized, however, tend to be overwhelming. We prefer some sort of coherence, a principle that connects the various sounds and makes them comprehensible. In this respect, music is like human language. Single sounds are in most cases not sufficient to convey meaning in speech, whereas when put together in a sequence they form words and sentences. Likewise, if the tones in music are not perceived to be tied together sequentially or rhythmically—for example, in what is commonly called melody— listeners are less likely to feel any emotional connection or to show appreciation. Certain music can also have a relaxing effect. The fact that such music tends to be continuous and rhythmical suggests a possible explanation for this effect. In a natural environment, danger tends to be accompanied by sudden, unexpected sounds. Thus, a background of constant noise suggests peaceful conditions; discontinuous sounds demand more attention. Even soft discontinuous sounds that we consciously realize do not signal danger can be disturbing—for example, the erratic dripping of a leaky tap. A continuous sound, particularly one that is judged to be safe, relaxes the brain. Passage B There are certain elements within music, such as a change of melodic line or rhythm, that create expectations about the future development of the music. The expectation the listener has about the further course of musical events is a key determinant for the experience of "musical emotions." Music creates expectations that, if not immediately satisfied, create tension. Emotion is experienced in relation to the buildup and release of tension. The more elaborate the buildup of tension, the more intense the emotions that will be experienon of the expected musical course, depending on one's personal involvement, causes the search for an explanation. This results from a "mismatch" between one's musical expectation and the actual course of the music. Negative emotions will be the result of an extreme mismatch between expectations and experience. Positive emotions result if the converse happens. When we listen to music, we take into account factors such as the complexity and novelty of the music. The degree to which the music sounds familiar determines whether the music is experienced as pleasurable or uncomfortable. The pleasure experienced is minimal when the music is entirely new to the listener, increases with increasing familiarity, and decreases again when the music is totally known. Musical preference is based on one's desire to maintain a constant level of certain preferable emotions. As such, a trained listener will have a greater preference for complex melodies than will a naive listener, as the threshold for experiencing emotion is higher.
200909_2-RC_4_25
[ "The musical traditions of different cultures vary greatly in terms of the complexity of the rhythms they employ.", "The rhythmic structure of a language is determined in part by the pattern of stressed syllables in the words and sentences of the language.", "Many people find the steady and rhythmic sound of a rocking chair to be very unnerving.", "The sudden interruption of the expected development of a melody tends to interfere with listeners' perception of the melody as coherent.", "Some of the most admired contemporary composers write music that is notably simpler than is most of the music written in previous centuries." ]
2
Which one of the following most undermines the explanation provided in passage A for the relaxing effect that some music has on listeners?
Passage AIn music, a certain complexity of sounds can be expected to have a positive effect on the listener. A single, pure tone is not that interesting to explore; a measure of intricacy is required to excite human curiosity. Sounds that are too complex or disorganized, however, tend to be overwhelming. We prefer some sort of coherence, a principle that connects the various sounds and makes them comprehensible. In this respect, music is like human language. Single sounds are in most cases not sufficient to convey meaning in speech, whereas when put together in a sequence they form words and sentences. Likewise, if the tones in music are not perceived to be tied together sequentially or rhythmically—for example, in what is commonly called melody— listeners are less likely to feel any emotional connection or to show appreciation. Certain music can also have a relaxing effect. The fact that such music tends to be continuous and rhythmical suggests a possible explanation for this effect. In a natural environment, danger tends to be accompanied by sudden, unexpected sounds. Thus, a background of constant noise suggests peaceful conditions; discontinuous sounds demand more attention. Even soft discontinuous sounds that we consciously realize do not signal danger can be disturbing—for example, the erratic dripping of a leaky tap. A continuous sound, particularly one that is judged to be safe, relaxes the brain. Passage B There are certain elements within music, such as a change of melodic line or rhythm, that create expectations about the future development of the music. The expectation the listener has about the further course of musical events is a key determinant for the experience of "musical emotions." Music creates expectations that, if not immediately satisfied, create tension. Emotion is experienced in relation to the buildup and release of tension. The more elaborate the buildup of tension, the more intense the emotions that will be experienon of the expected musical course, depending on one's personal involvement, causes the search for an explanation. This results from a "mismatch" between one's musical expectation and the actual course of the music. Negative emotions will be the result of an extreme mismatch between expectations and experience. Positive emotions result if the converse happens. When we listen to music, we take into account factors such as the complexity and novelty of the music. The degree to which the music sounds familiar determines whether the music is experienced as pleasurable or uncomfortable. The pleasure experienced is minimal when the music is entirely new to the listener, increases with increasing familiarity, and decreases again when the music is totally known. Musical preference is based on one's desire to maintain a constant level of certain preferable emotions. As such, a trained listener will have a greater preference for complex melodies than will a naive listener, as the threshold for experiencing emotion is higher.
200909_2-RC_4_26
[ "\"The Biological Underpinnings of Musical Emotions\"", "\"The Psychology of Listener Response to Music\"", "\"How Music Differs from Other Art Forms\"", "\"Cultural Patterns in Listeners' Responses to Music\"", "\"How Composers Convey Meaning Through Music\"" ]
1
Which one of the following would be most appropriate as a title for each of the passages?
Passage AIn music, a certain complexity of sounds can be expected to have a positive effect on the listener. A single, pure tone is not that interesting to explore; a measure of intricacy is required to excite human curiosity. Sounds that are too complex or disorganized, however, tend to be overwhelming. We prefer some sort of coherence, a principle that connects the various sounds and makes them comprehensible. In this respect, music is like human language. Single sounds are in most cases not sufficient to convey meaning in speech, whereas when put together in a sequence they form words and sentences. Likewise, if the tones in music are not perceived to be tied together sequentially or rhythmically—for example, in what is commonly called melody— listeners are less likely to feel any emotional connection or to show appreciation. Certain music can also have a relaxing effect. The fact that such music tends to be continuous and rhythmical suggests a possible explanation for this effect. In a natural environment, danger tends to be accompanied by sudden, unexpected sounds. Thus, a background of constant noise suggests peaceful conditions; discontinuous sounds demand more attention. Even soft discontinuous sounds that we consciously realize do not signal danger can be disturbing—for example, the erratic dripping of a leaky tap. A continuous sound, particularly one that is judged to be safe, relaxes the brain. Passage B There are certain elements within music, such as a change of melodic line or rhythm, that create expectations about the future development of the music. The expectation the listener has about the further course of musical events is a key determinant for the experience of "musical emotions." Music creates expectations that, if not immediately satisfied, create tension. Emotion is experienced in relation to the buildup and release of tension. The more elaborate the buildup of tension, the more intense the emotions that will be experienon of the expected musical course, depending on one's personal involvement, causes the search for an explanation. This results from a "mismatch" between one's musical expectation and the actual course of the music. Negative emotions will be the result of an extreme mismatch between expectations and experience. Positive emotions result if the converse happens. When we listen to music, we take into account factors such as the complexity and novelty of the music. The degree to which the music sounds familiar determines whether the music is experienced as pleasurable or uncomfortable. The pleasure experienced is minimal when the music is entirely new to the listener, increases with increasing familiarity, and decreases again when the music is totally known. Musical preference is based on one's desire to maintain a constant level of certain preferable emotions. As such, a trained listener will have a greater preference for complex melodies than will a naive listener, as the threshold for experiencing emotion is higher.
200909_2-RC_4_27
[ "The more complex a piece of music, the more it is likely to be enjoyed by most listeners.", "More knowledgeable listeners tend to prefer music that is discontinuous and unpredictable.", "The capacity of music to elicit strong emotional responses from listeners is the central determinant of its artistic value.", "Music that lacks a predictable course is unlikely to cause a listener to feel relaxed.", "Music that changes from soft to loud is perceived as disturbing and unpleasant by most listeners." ]
3
It can be inferred that both authors would be likely to agree with which one of the following statements?
Passage A Recent studies have shown that sophisticated computer models of the oceans and atmosphere are capable of simulating large-scale climate trends with remarkable accuracy. But these models make use of large numbers of variables, many of which have wide ranges of possible values. Because even small differences in those values can have a significant impact on what the simulations predict, it is important to determine the impact when values differ even slightly. Since the interactions between the many variables in climate simulations are highly complex, there is no alternative to a "brute force" exploration of all possible combinations of their values if predictions are to be reliable. This method requires very large numbers of calculations and simulation runs. For example, exhaustive examination of five values for each of only nine variables would require 2 million calculation-intensive simulation runs. Currently available individual computers are completely inadequate for such a task. However, the continuing increase in computing capacity of the average desktop computer means that climate simulations can now be run on privately owned desktop machines connected to one another via the Internet. The calculations are divided among the individual desktop computers, which work simultaneously on their share of the overall problem. Some public resource computing projects of this kind have already been successful, although only when they captured the public's interest sufficiently to secure widespread participation. Passage B Researchers are now learning that many problems in nature, human society, science, and engineering are naturally "parallel" ; that is, that they can be effectively solved by using methods that work simultaneously in parallel. These problems share the common characteristic of involving a large number of similar elements such as molecules, animals, even people, whose individual actions are governed by simple rules but, taken collectively, function as a highly complex system. An example is the method used by ants to forage for food. As Lewis Thomas observed, a solitary ant is little more than a few neurons strung together by fibers. Its behavior follows a few simple rules. But when one sees a dense mass of thousands of ants, crowded together around their anthill retrieving food or repelling an intruder, a more complex pictur emerges; it is as if the whole is thinking, planning, calculating. It is an intelligence, a kind of live computer, with crawling bits for wits. We are now living through a great paradigm shift in the field of computing, a shift from sequential computing (performing one calculation at a time) to massive parallel computing, which employs thousands of computers working simultaneously to solve one computation-intensive problem. Since many computation-intensive problems are inherently parallel, it only makes sense to use a computing model that exploits that parallelism. A computing model that resembles the inherently parallel problem it is trying to solve will perform best. The old paradigm, in contrast, is subject to the speed limits imposed by purely sequential computing.
200912_4-RC_1_1
[ "Many difficult problems in computing are naturally parallel.", "Sequential computing is no longer useful because of the speed limits it imposes.", "There is currently a paradigm shift occurring in the field of computing toward parallel computing.", "Complex biological and social systems are the next frontier in the field of computer simulation.", "Inherently parallel computing problems are best solved by means of computers modeled on the human mind." ]
2
Which one of the following most accurately expresses the main point of passage B?
Passage A Recent studies have shown that sophisticated computer models of the oceans and atmosphere are capable of simulating large-scale climate trends with remarkable accuracy. But these models make use of large numbers of variables, many of which have wide ranges of possible values. Because even small differences in those values can have a significant impact on what the simulations predict, it is important to determine the impact when values differ even slightly. Since the interactions between the many variables in climate simulations are highly complex, there is no alternative to a "brute force" exploration of all possible combinations of their values if predictions are to be reliable. This method requires very large numbers of calculations and simulation runs. For example, exhaustive examination of five values for each of only nine variables would require 2 million calculation-intensive simulation runs. Currently available individual computers are completely inadequate for such a task. However, the continuing increase in computing capacity of the average desktop computer means that climate simulations can now be run on privately owned desktop machines connected to one another via the Internet. The calculations are divided among the individual desktop computers, which work simultaneously on their share of the overall problem. Some public resource computing projects of this kind have already been successful, although only when they captured the public's interest sufficiently to secure widespread participation. Passage B Researchers are now learning that many problems in nature, human society, science, and engineering are naturally "parallel" ; that is, that they can be effectively solved by using methods that work simultaneously in parallel. These problems share the common characteristic of involving a large number of similar elements such as molecules, animals, even people, whose individual actions are governed by simple rules but, taken collectively, function as a highly complex system. An example is the method used by ants to forage for food. As Lewis Thomas observed, a solitary ant is little more than a few neurons strung together by fibers. Its behavior follows a few simple rules. But when one sees a dense mass of thousands of ants, crowded together around their anthill retrieving food or repelling an intruder, a more complex pictur emerges; it is as if the whole is thinking, planning, calculating. It is an intelligence, a kind of live computer, with crawling bits for wits. We are now living through a great paradigm shift in the field of computing, a shift from sequential computing (performing one calculation at a time) to massive parallel computing, which employs thousands of computers working simultaneously to solve one computation-intensive problem. Since many computation-intensive problems are inherently parallel, it only makes sense to use a computing model that exploits that parallelism. A computing model that resembles the inherently parallel problem it is trying to solve will perform best. The old paradigm, in contrast, is subject to the speed limits imposed by purely sequential computing.
200912_4-RC_1_2
[ "the thousands of computers working simultaneously to solve a calculation-intensive problem", "the simple rules that shape the behavior of a single ant", "the highly complex behavior of a dense mass of thousands of ants", "the paradigm shift from sequential to parallel computing", "the speed limits imposed by computing purely sequentially" ]
2
The large-scale climate trends discussed in passage A are most analogous to which one of the following elements in passage B?
Passage A Recent studies have shown that sophisticated computer models of the oceans and atmosphere are capable of simulating large-scale climate trends with remarkable accuracy. But these models make use of large numbers of variables, many of which have wide ranges of possible values. Because even small differences in those values can have a significant impact on what the simulations predict, it is important to determine the impact when values differ even slightly. Since the interactions between the many variables in climate simulations are highly complex, there is no alternative to a "brute force" exploration of all possible combinations of their values if predictions are to be reliable. This method requires very large numbers of calculations and simulation runs. For example, exhaustive examination of five values for each of only nine variables would require 2 million calculation-intensive simulation runs. Currently available individual computers are completely inadequate for such a task. However, the continuing increase in computing capacity of the average desktop computer means that climate simulations can now be run on privately owned desktop machines connected to one another via the Internet. The calculations are divided among the individual desktop computers, which work simultaneously on their share of the overall problem. Some public resource computing projects of this kind have already been successful, although only when they captured the public's interest sufficiently to secure widespread participation. Passage B Researchers are now learning that many problems in nature, human society, science, and engineering are naturally "parallel" ; that is, that they can be effectively solved by using methods that work simultaneously in parallel. These problems share the common characteristic of involving a large number of similar elements such as molecules, animals, even people, whose individual actions are governed by simple rules but, taken collectively, function as a highly complex system. An example is the method used by ants to forage for food. As Lewis Thomas observed, a solitary ant is little more than a few neurons strung together by fibers. Its behavior follows a few simple rules. But when one sees a dense mass of thousands of ants, crowded together around their anthill retrieving food or repelling an intruder, a more complex pictur emerges; it is as if the whole is thinking, planning, calculating. It is an intelligence, a kind of live computer, with crawling bits for wits. We are now living through a great paradigm shift in the field of computing, a shift from sequential computing (performing one calculation at a time) to massive parallel computing, which employs thousands of computers working simultaneously to solve one computation-intensive problem. Since many computation-intensive problems are inherently parallel, it only makes sense to use a computing model that exploits that parallelism. A computing model that resembles the inherently parallel problem it is trying to solve will perform best. The old paradigm, in contrast, is subject to the speed limits imposed by purely sequential computing.
200912_4-RC_1_3
[ "Massive, parallel computing systems are able to solve complex computation-intensive problems without having to resort to \"brute force.\"", "Computer models are not capable of simulating the behavior of very large biological populations such as insect colonies.", "Parallel computing systems that link privately owned desktop computers via the Internet are not feasible because they rely too heavily on public participation.", "Currently available computers are not well-suited to running simulations, even if the simulated problems are relatively simple.", "Parallel computing systems employing multiple computers are the best means for simulating large-scale climate trends." ]
4
It can be inferred that the authors of the two passages would be most likely to agree on which one of the following statements concerning computing systems?
Passage A Recent studies have shown that sophisticated computer models of the oceans and atmosphere are capable of simulating large-scale climate trends with remarkable accuracy. But these models make use of large numbers of variables, many of which have wide ranges of possible values. Because even small differences in those values can have a significant impact on what the simulations predict, it is important to determine the impact when values differ even slightly. Since the interactions between the many variables in climate simulations are highly complex, there is no alternative to a "brute force" exploration of all possible combinations of their values if predictions are to be reliable. This method requires very large numbers of calculations and simulation runs. For example, exhaustive examination of five values for each of only nine variables would require 2 million calculation-intensive simulation runs. Currently available individual computers are completely inadequate for such a task. However, the continuing increase in computing capacity of the average desktop computer means that climate simulations can now be run on privately owned desktop machines connected to one another via the Internet. The calculations are divided among the individual desktop computers, which work simultaneously on their share of the overall problem. Some public resource computing projects of this kind have already been successful, although only when they captured the public's interest sufficiently to secure widespread participation. Passage B Researchers are now learning that many problems in nature, human society, science, and engineering are naturally "parallel" ; that is, that they can be effectively solved by using methods that work simultaneously in parallel. These problems share the common characteristic of involving a large number of similar elements such as molecules, animals, even people, whose individual actions are governed by simple rules but, taken collectively, function as a highly complex system. An example is the method used by ants to forage for food. As Lewis Thomas observed, a solitary ant is little more than a few neurons strung together by fibers. Its behavior follows a few simple rules. But when one sees a dense mass of thousands of ants, crowded together around their anthill retrieving food or repelling an intruder, a more complex pictur emerges; it is as if the whole is thinking, planning, calculating. It is an intelligence, a kind of live computer, with crawling bits for wits. We are now living through a great paradigm shift in the field of computing, a shift from sequential computing (performing one calculation at a time) to massive parallel computing, which employs thousands of computers working simultaneously to solve one computation-intensive problem. Since many computation-intensive problems are inherently parallel, it only makes sense to use a computing model that exploits that parallelism. A computing model that resembles the inherently parallel problem it is trying to solve will perform best. The old paradigm, in contrast, is subject to the speed limits imposed by purely sequential computing.
200912_4-RC_1_4
[ "encourage public engagement in the sort of computing model discussed in the passage", "identify a factor affecting the feasibility of the computing model advocated in the passage", "indicate that government support of large-scale computing efforts is needed", "demonstrate that adequate support for the type of approach described in the passage already exists", "suggest that a computing model like that proposed in the passage is infeasible because of forces beyond the designers' control" ]
1
The author of passage A mentions public participation (lines 30–32) primarily in order to
Passage A Recent studies have shown that sophisticated computer models of the oceans and atmosphere are capable of simulating large-scale climate trends with remarkable accuracy. But these models make use of large numbers of variables, many of which have wide ranges of possible values. Because even small differences in those values can have a significant impact on what the simulations predict, it is important to determine the impact when values differ even slightly. Since the interactions between the many variables in climate simulations are highly complex, there is no alternative to a "brute force" exploration of all possible combinations of their values if predictions are to be reliable. This method requires very large numbers of calculations and simulation runs. For example, exhaustive examination of five values for each of only nine variables would require 2 million calculation-intensive simulation runs. Currently available individual computers are completely inadequate for such a task. However, the continuing increase in computing capacity of the average desktop computer means that climate simulations can now be run on privately owned desktop machines connected to one another via the Internet. The calculations are divided among the individual desktop computers, which work simultaneously on their share of the overall problem. Some public resource computing projects of this kind have already been successful, although only when they captured the public's interest sufficiently to secure widespread participation. Passage B Researchers are now learning that many problems in nature, human society, science, and engineering are naturally "parallel" ; that is, that they can be effectively solved by using methods that work simultaneously in parallel. These problems share the common characteristic of involving a large number of similar elements such as molecules, animals, even people, whose individual actions are governed by simple rules but, taken collectively, function as a highly complex system. An example is the method used by ants to forage for food. As Lewis Thomas observed, a solitary ant is little more than a few neurons strung together by fibers. Its behavior follows a few simple rules. But when one sees a dense mass of thousands of ants, crowded together around their anthill retrieving food or repelling an intruder, a more complex pictur emerges; it is as if the whole is thinking, planning, calculating. It is an intelligence, a kind of live computer, with crawling bits for wits. We are now living through a great paradigm shift in the field of computing, a shift from sequential computing (performing one calculation at a time) to massive parallel computing, which employs thousands of computers working simultaneously to solve one computation-intensive problem. Since many computation-intensive problems are inherently parallel, it only makes sense to use a computing model that exploits that parallelism. A computing model that resembles the inherently parallel problem it is trying to solve will perform best. The old paradigm, in contrast, is subject to the speed limits imposed by purely sequential computing.
200912_4-RC_1_5
[ "The argument in passage B has little bearing on the issues discussed in passage A.", "The explanation offered in passage B shows why the plan proposed in passage A is unlikely to be implemented.", "The ideas advanced in passage B provide a rationale for the solution proposed in passage A.", "The example given in passage B illustrates the need for the \"brute force\" exploration mentioned in passage A.", "The discussion in passage B conflicts with the assumptions about individual computers made in passage A." ]
2
Passage B relates to passage A in which one of the following ways?
Passage A Recent studies have shown that sophisticated computer models of the oceans and atmosphere are capable of simulating large-scale climate trends with remarkable accuracy. But these models make use of large numbers of variables, many of which have wide ranges of possible values. Because even small differences in those values can have a significant impact on what the simulations predict, it is important to determine the impact when values differ even slightly. Since the interactions between the many variables in climate simulations are highly complex, there is no alternative to a "brute force" exploration of all possible combinations of their values if predictions are to be reliable. This method requires very large numbers of calculations and simulation runs. For example, exhaustive examination of five values for each of only nine variables would require 2 million calculation-intensive simulation runs. Currently available individual computers are completely inadequate for such a task. However, the continuing increase in computing capacity of the average desktop computer means that climate simulations can now be run on privately owned desktop machines connected to one another via the Internet. The calculations are divided among the individual desktop computers, which work simultaneously on their share of the overall problem. Some public resource computing projects of this kind have already been successful, although only when they captured the public's interest sufficiently to secure widespread participation. Passage B Researchers are now learning that many problems in nature, human society, science, and engineering are naturally "parallel" ; that is, that they can be effectively solved by using methods that work simultaneously in parallel. These problems share the common characteristic of involving a large number of similar elements such as molecules, animals, even people, whose individual actions are governed by simple rules but, taken collectively, function as a highly complex system. An example is the method used by ants to forage for food. As Lewis Thomas observed, a solitary ant is little more than a few neurons strung together by fibers. Its behavior follows a few simple rules. But when one sees a dense mass of thousands of ants, crowded together around their anthill retrieving food or repelling an intruder, a more complex pictur emerges; it is as if the whole is thinking, planning, calculating. It is an intelligence, a kind of live computer, with crawling bits for wits. We are now living through a great paradigm shift in the field of computing, a shift from sequential computing (performing one calculation at a time) to massive parallel computing, which employs thousands of computers working simultaneously to solve one computation-intensive problem. Since many computation-intensive problems are inherently parallel, it only makes sense to use a computing model that exploits that parallelism. A computing model that resembles the inherently parallel problem it is trying to solve will perform best. The old paradigm, in contrast, is subject to the speed limits imposed by purely sequential computing.
200912_4-RC_1_6
[ "to show that the traditional paradigm in computing is ineffective for many common computing tasks", "to argue that a new approach to computing is an effective way to solve a difficult type of problem", "to convince skeptics of the usefulness of desktop computers for calculation-intensive problems", "to demonstrate that a new computing paradigm has supplanted the traditional paradigm for most large-scale computing problems", "to describe complex and as yet unsolved problems that have recently arisen in computing" ]
1
The passages share which one of the following as their primary purpose?
Passage A Recent studies have shown that sophisticated computer models of the oceans and atmosphere are capable of simulating large-scale climate trends with remarkable accuracy. But these models make use of large numbers of variables, many of which have wide ranges of possible values. Because even small differences in those values can have a significant impact on what the simulations predict, it is important to determine the impact when values differ even slightly. Since the interactions between the many variables in climate simulations are highly complex, there is no alternative to a "brute force" exploration of all possible combinations of their values if predictions are to be reliable. This method requires very large numbers of calculations and simulation runs. For example, exhaustive examination of five values for each of only nine variables would require 2 million calculation-intensive simulation runs. Currently available individual computers are completely inadequate for such a task. However, the continuing increase in computing capacity of the average desktop computer means that climate simulations can now be run on privately owned desktop machines connected to one another via the Internet. The calculations are divided among the individual desktop computers, which work simultaneously on their share of the overall problem. Some public resource computing projects of this kind have already been successful, although only when they captured the public's interest sufficiently to secure widespread participation. Passage B Researchers are now learning that many problems in nature, human society, science, and engineering are naturally "parallel" ; that is, that they can be effectively solved by using methods that work simultaneously in parallel. These problems share the common characteristic of involving a large number of similar elements such as molecules, animals, even people, whose individual actions are governed by simple rules but, taken collectively, function as a highly complex system. An example is the method used by ants to forage for food. As Lewis Thomas observed, a solitary ant is little more than a few neurons strung together by fibers. Its behavior follows a few simple rules. But when one sees a dense mass of thousands of ants, crowded together around their anthill retrieving food or repelling an intruder, a more complex pictur emerges; it is as if the whole is thinking, planning, calculating. It is an intelligence, a kind of live computer, with crawling bits for wits. We are now living through a great paradigm shift in the field of computing, a shift from sequential computing (performing one calculation at a time) to massive parallel computing, which employs thousands of computers working simultaneously to solve one computation-intensive problem. Since many computation-intensive problems are inherently parallel, it only makes sense to use a computing model that exploits that parallelism. A computing model that resembles the inherently parallel problem it is trying to solve will perform best. The old paradigm, in contrast, is subject to the speed limits imposed by purely sequential computing.
200912_4-RC_1_7
[ "the behavior of the colony of ants functions as a complex, organized whole", "the paradigm shift taking place in computing was inspired by observations of living systems", "computers are agglomerations of elements that can be viewed as being alive in a metaphorical sense", "computer simulations can simulate the behavior of large biological populations with great accuracy", "the simple rules that govern the behavior of individual ants have been adapted for use in computer simulations" ]
0
In calling a population of ants "an intelligence, a kind of live computer" (lines 51–52) the author of passage B most likely means that
Passage A Recent studies have shown that sophisticated computer models of the oceans and atmosphere are capable of simulating large-scale climate trends with remarkable accuracy. But these models make use of large numbers of variables, many of which have wide ranges of possible values. Because even small differences in those values can have a significant impact on what the simulations predict, it is important to determine the impact when values differ even slightly. Since the interactions between the many variables in climate simulations are highly complex, there is no alternative to a "brute force" exploration of all possible combinations of their values if predictions are to be reliable. This method requires very large numbers of calculations and simulation runs. For example, exhaustive examination of five values for each of only nine variables would require 2 million calculation-intensive simulation runs. Currently available individual computers are completely inadequate for such a task. However, the continuing increase in computing capacity of the average desktop computer means that climate simulations can now be run on privately owned desktop machines connected to one another via the Internet. The calculations are divided among the individual desktop computers, which work simultaneously on their share of the overall problem. Some public resource computing projects of this kind have already been successful, although only when they captured the public's interest sufficiently to secure widespread participation. Passage B Researchers are now learning that many problems in nature, human society, science, and engineering are naturally "parallel" ; that is, that they can be effectively solved by using methods that work simultaneously in parallel. These problems share the common characteristic of involving a large number of similar elements such as molecules, animals, even people, whose individual actions are governed by simple rules but, taken collectively, function as a highly complex system. An example is the method used by ants to forage for food. As Lewis Thomas observed, a solitary ant is little more than a few neurons strung together by fibers. Its behavior follows a few simple rules. But when one sees a dense mass of thousands of ants, crowded together around their anthill retrieving food or repelling an intruder, a more complex pictur emerges; it is as if the whole is thinking, planning, calculating. It is an intelligence, a kind of live computer, with crawling bits for wits. We are now living through a great paradigm shift in the field of computing, a shift from sequential computing (performing one calculation at a time) to massive parallel computing, which employs thousands of computers working simultaneously to solve one computation-intensive problem. Since many computation-intensive problems are inherently parallel, it only makes sense to use a computing model that exploits that parallelism. A computing model that resembles the inherently parallel problem it is trying to solve will perform best. The old paradigm, in contrast, is subject to the speed limits imposed by purely sequential computing.
200912_4-RC_1_8
[ "It would be a kind of live computer.", "It would be completely inadequate for simulating large-scale climate trends.", "It would impose strict limitations on the number of variables that could be used in any simulation it runs.", "It would be likely to secure widespread public participation.", "It would solve calculation-intensive problems faster than a traditional sequential computer would." ]
4
The author of passage B would be most likely to agree with which one of the following statements regarding the computing system proposed in the last paragraph of passage A?
A proficiency in understanding, applying, and even formulating statutes—the actual texts of laws enacted by legislative bodies—is a vital aspect of the practice of law, but statutory law is often given too little attention by law schools. Much of legal education, with its focus on judicial decisions and analysis of cases, can give a law student the impression that the practice of law consists mainly in analyzing past cases to determine their relevance to a client's situation and arriving at a speculative interpretation of the law relevant to the client's legal problem. Lawyers discover fairly soon, however, that much of their practice does not depend on the kind of painstaking analysis of cases that is performed in law school. For example, a lawyer representing the owner of a business can often find an explicit answer as to what the client should do about a certain tax-related issue by consulting the relevant statutes. In such a case the facts are clear and the statutes' relation to them transparent, so that the client's question can be answered by direct reference to the wording of the statutes. But statutes' meanings and their applicability to relevant situations are not always so obvious, and that is one reason that the ability to interpret them accurately is an essential skill for law students to learn. Another skill that teaching statutory law would improve is synthesis. Law professors work hard at developing their students' ability to analyze individual cases, but in so doing they favor the ability to apply the law in particular cases over the ability to understand the interrelations among laws. In contrast, the study of all the statutes of a legal system in a certain small area of the law would enable the student to see how these laws form a coherent whole. Students would then be able to apply this ability to synthesize in other areas of statutory law that they encounter in their study or practice. This is especially important because most students intend to specialize in a chosen area, or areas, of the law. One possible argument against including training in statutory law as a standard part of law school curricula is that many statutes vary from region to region within a nation, so that the mastery of a set of statutes would usually not be generally applicable. There is some truth to this objection; law schools that currently provide some training in statutes generally intend it as a preparation for practice in their particular region, but for schools that are nationally oriented, this could seem to be an inappropriate investment of time and resources. But while the knowledge of a particular region's statutory law is not generally transferable to other regions, the skills acquired in mastering a particular set of statutes are, making the study of statutory law an important undertaking even for law schools with a national orientation.
200912_4-RC_2_9
[ "In spite of the reservations that nationally oriented law schools can be expected to have, law schools can serve the overall needs of law students better by implementing a standard national curriculum in statutory law.", "Since the skills promoted by the study of statutory law are ultimately more important than those promoted by case analysis, the relative emphasis that law schools place on these two areas should be reversed.", "Although statutes typically vary from region to region, law schools should provide training in statutory law in order to develop students' ability to synthesize legal information and interpret individual statutes.", "In the theoretical world of law school training, as opposed to the actual practice of law, a proficiency in case law is often one of the most important assets that students can have.", "Law schools generally are deficient in their attention to statutory law training and therefore fail to impart the skills necessary for the analysis of legal information." ]
2
Which one of the following most accurately expresses the main point of the passage?
A proficiency in understanding, applying, and even formulating statutes—the actual texts of laws enacted by legislative bodies—is a vital aspect of the practice of law, but statutory law is often given too little attention by law schools. Much of legal education, with its focus on judicial decisions and analysis of cases, can give a law student the impression that the practice of law consists mainly in analyzing past cases to determine their relevance to a client's situation and arriving at a speculative interpretation of the law relevant to the client's legal problem. Lawyers discover fairly soon, however, that much of their practice does not depend on the kind of painstaking analysis of cases that is performed in law school. For example, a lawyer representing the owner of a business can often find an explicit answer as to what the client should do about a certain tax-related issue by consulting the relevant statutes. In such a case the facts are clear and the statutes' relation to them transparent, so that the client's question can be answered by direct reference to the wording of the statutes. But statutes' meanings and their applicability to relevant situations are not always so obvious, and that is one reason that the ability to interpret them accurately is an essential skill for law students to learn. Another skill that teaching statutory law would improve is synthesis. Law professors work hard at developing their students' ability to analyze individual cases, but in so doing they favor the ability to apply the law in particular cases over the ability to understand the interrelations among laws. In contrast, the study of all the statutes of a legal system in a certain small area of the law would enable the student to see how these laws form a coherent whole. Students would then be able to apply this ability to synthesize in other areas of statutory law that they encounter in their study or practice. This is especially important because most students intend to specialize in a chosen area, or areas, of the law. One possible argument against including training in statutory law as a standard part of law school curricula is that many statutes vary from region to region within a nation, so that the mastery of a set of statutes would usually not be generally applicable. There is some truth to this objection; law schools that currently provide some training in statutes generally intend it as a preparation for practice in their particular region, but for schools that are nationally oriented, this could seem to be an inappropriate investment of time and resources. But while the knowledge of a particular region's statutory law is not generally transferable to other regions, the skills acquired in mastering a particular set of statutes are, making the study of statutory law an important undertaking even for law schools with a national orientation.
200912_4-RC_2_10
[ "Such training would divert resources away from the far more important development of the ability to analyze cases.", "Such training is not essentially different from what is already provided in the core areas of law school education.", "The goals of such training can better be achieved by other means, most of which are more directly related to the actual practice of law.", "Such training would be irrelevant for those students who do not plan to specialize.", "The lack of geographic uniformity among statutory laws makes expertise in the statutes of any particular region generally nontransferable." ]
4
Which one of the following is cited in the passage as a reason that might be given for not including statutory law training in law school curricula?
A proficiency in understanding, applying, and even formulating statutes—the actual texts of laws enacted by legislative bodies—is a vital aspect of the practice of law, but statutory law is often given too little attention by law schools. Much of legal education, with its focus on judicial decisions and analysis of cases, can give a law student the impression that the practice of law consists mainly in analyzing past cases to determine their relevance to a client's situation and arriving at a speculative interpretation of the law relevant to the client's legal problem. Lawyers discover fairly soon, however, that much of their practice does not depend on the kind of painstaking analysis of cases that is performed in law school. For example, a lawyer representing the owner of a business can often find an explicit answer as to what the client should do about a certain tax-related issue by consulting the relevant statutes. In such a case the facts are clear and the statutes' relation to them transparent, so that the client's question can be answered by direct reference to the wording of the statutes. But statutes' meanings and their applicability to relevant situations are not always so obvious, and that is one reason that the ability to interpret them accurately is an essential skill for law students to learn. Another skill that teaching statutory law would improve is synthesis. Law professors work hard at developing their students' ability to analyze individual cases, but in so doing they favor the ability to apply the law in particular cases over the ability to understand the interrelations among laws. In contrast, the study of all the statutes of a legal system in a certain small area of the law would enable the student to see how these laws form a coherent whole. Students would then be able to apply this ability to synthesize in other areas of statutory law that they encounter in their study or practice. This is especially important because most students intend to specialize in a chosen area, or areas, of the law. One possible argument against including training in statutory law as a standard part of law school curricula is that many statutes vary from region to region within a nation, so that the mastery of a set of statutes would usually not be generally applicable. There is some truth to this objection; law schools that currently provide some training in statutes generally intend it as a preparation for practice in their particular region, but for schools that are nationally oriented, this could seem to be an inappropriate investment of time and resources. But while the knowledge of a particular region's statutory law is not generally transferable to other regions, the skills acquired in mastering a particular set of statutes are, making the study of statutory law an important undertaking even for law schools with a national orientation.
200912_4-RC_2_11
[ "Many law school administrators recommend the inclusion of statutory law training in the curricula of their schools.", "Most lawyers easily and quickly develop proficiency in statutory law through their work experiences after law school.", "Most lawyers do not practice law in the same geographic area in which they attended law school.", "The curricula of many regionally oriented law schools rely primarily on analysis of cases.", "Most lawyers who have undergone training in statutory law are thoroughly familiar with only a narrow range of statutes." ]
1
Which one of the following would, if true, most weaken the author's argument as expressed in the passage?
A proficiency in understanding, applying, and even formulating statutes—the actual texts of laws enacted by legislative bodies—is a vital aspect of the practice of law, but statutory law is often given too little attention by law schools. Much of legal education, with its focus on judicial decisions and analysis of cases, can give a law student the impression that the practice of law consists mainly in analyzing past cases to determine their relevance to a client's situation and arriving at a speculative interpretation of the law relevant to the client's legal problem. Lawyers discover fairly soon, however, that much of their practice does not depend on the kind of painstaking analysis of cases that is performed in law school. For example, a lawyer representing the owner of a business can often find an explicit answer as to what the client should do about a certain tax-related issue by consulting the relevant statutes. In such a case the facts are clear and the statutes' relation to them transparent, so that the client's question can be answered by direct reference to the wording of the statutes. But statutes' meanings and their applicability to relevant situations are not always so obvious, and that is one reason that the ability to interpret them accurately is an essential skill for law students to learn. Another skill that teaching statutory law would improve is synthesis. Law professors work hard at developing their students' ability to analyze individual cases, but in so doing they favor the ability to apply the law in particular cases over the ability to understand the interrelations among laws. In contrast, the study of all the statutes of a legal system in a certain small area of the law would enable the student to see how these laws form a coherent whole. Students would then be able to apply this ability to synthesize in other areas of statutory law that they encounter in their study or practice. This is especially important because most students intend to specialize in a chosen area, or areas, of the law. One possible argument against including training in statutory law as a standard part of law school curricula is that many statutes vary from region to region within a nation, so that the mastery of a set of statutes would usually not be generally applicable. There is some truth to this objection; law schools that currently provide some training in statutes generally intend it as a preparation for practice in their particular region, but for schools that are nationally oriented, this could seem to be an inappropriate investment of time and resources. But while the knowledge of a particular region's statutory law is not generally transferable to other regions, the skills acquired in mastering a particular set of statutes are, making the study of statutory law an important undertaking even for law schools with a national orientation.
200912_4-RC_2_12
[ "identify and describe one of the benefits that the author says would result from the change that is advocated in the passage", "indicate that law schools currently value certain other skills over this skill and explain why this is so", "argue for the greater importance of this skill as compared with certain others that are discussed earlier in the passage", "explain why this skill is necessary for the study of statutory law", "provide an example of the type of problem typically encountered in the practice of law" ]
0
The author discusses the skill of synthesis in the third paragraph primarily in order to