context
stringclasses
269 values
id_string
stringlengths
15
16
answers
sequencelengths
5
5
label
int64
0
4
question
stringlengths
34
417
Years after the movement to obtain civil rights for black people in the United States made its most important gains, scholars are reaching for a theoretical perspective capable of clarifying its momentous developments. New theories of social movements are being discussed, not just among social psychologists, but also among political theorists. Of the many competing formulations of the "classical" social psychological theory of social movement, three are prominent in the literature on the civil rights movement: "rising expectations," "relative deprivation," and "J-curve." Each conforms to a causal sequence characteristic of classical social movement theory, linking some unusual condition, or "system strain," to the generation of unrest. When these versions of the classical theory are applied to the civil rights movement, the source of strain is identified as a change in black socioeconomic status that occurred shortly before the widespread protest activity of the movement. For example, the theory of rising expectations asserts that protest activity was a response to psychological tensions generated by gains experienced immediately prior to the civil rights movement. Advancement did not satisfy ambition, but created the desire for further advancement. Only slightly different is the theory of relative deprivation. Here the impetus to protest is identified as gains achieved during the premovement period, coupled with simultaneous failure to make any appreciable headway relative to the dominant group. The J-curve theory argues that the movement occurred because a prolonged period of rising expectations and gratification was followed by a sharp reversal. Political theorists have been dismissive of these applications of classical theory to the civil rights movement. Their arguments rest on the conviction that, implicitly, the classical theory trivializes the political ends of movement participants, focusing rather on presumed psychological dysfunctions; reduction of complex social situations to simple paradigms of stimulus and response obviates the relevance of all but the shortest-term analysis. Furthermore, the theories lack predictive value: "strain" is always present to some degree, but social movement is not. How can we know which strain will provoke upheaval? These very legitimate complaints having frequently been made, it remains to find a means of testing the strength of the theories. Problematically, while proponents of the various theories have contradictory interpretations of socioeconomic conditions leading to the civil rights movement, examination of various statistical records regarding the material status of black Americans yields ample evidence to support any of the three theories. The steady rise in median black family income supports the rising expectations hypothesis; the stability of the economic position of black vis-à-vis white Americans lends credence to the relative deprivation interpretation; unemployment data are consistent with the J-curve theory. A better test is the comparison of each of these economic indicators with the frequency of movement-initiated events reported in the press; unsurprisingly, none correlates significantly with the pace of reports about movement activity.
199402_3-RC_4_27
[ "persuade historians of the indispensability of a theoretical framework for understanding recent history", "present a new model of social movement", "account for a shift in a theoretical debate", "show the unity underlying the diverse classical models of social movement", "discuss the reasoning behind and shortcomings of certain social psychological theories" ]
4
The main purpose of the passage is to
Nearly every writer on the philosophy of civil rights activist Martin Luther King, Jr., makes a connection between King and Henry David Thoreau, usually via Thoreau's famous essay, "Civil Disobedience" (1849). In his book Stride Toward Freedom (1958), King himself stated that Thoreau's essay was his first intellectual contact with the theory of passive resistance to governmental laws that are perceived as morally unjust. However, this emphasis on Thoreau's influence on King is unfortunate: first, King would not have agreed with many other aspects of Thoreau's philosophy, including Thoreau's ultimate acceptance of violence as a form of protest; second, an overemphasis on the influence of one essay has kept historians from noting other correspondences between King's philosophy and transcendentalism. "Civil Disobedience" was the only example of transcendentalist writing with which King was familiar, and in many other transcendentalist writings, including works by Ralph Waldo Emerson and Margaret Fuller, King would have found ideas more nearly akin to his own. The kind of civil disobedience King had in mind was, in fact, quite different from Thoreau's view of civil disobedience. Thoreau, like most other transcendentalists, was primarily interested in reform of the individual, whereas King was primarily interested in reform of society. As a protest against the Mexican War, Thoreau refused to pay taxes, but he did not hope by his action to force a change in national policy. While he encouraged others to adopt similar protests, he did not attempt to mount any mass protest action against unjust laws. In contrast to Thoreau, King began to advocate the use of mass civil disobedience to effect revolutionary changes within the social system. However, King's writings suggest that, without realizing it, he was an incipient transcendentalist. Most transcendentalists subscribed to the concept of "higher law" and included civil disobedience to unjust laws as part of their strategy. They often invoked the concept of higher law to justify their opposition to slavery and to advocate disobedience to the strengthened Fugitive Slave Law of 1850. In his second major book, King's discussion of just and unjust laws and the responsibility of the individual is very similar to the transcendentalists' discussion of higher law. In reference to how one can advocate breaking some laws and obeying others, King notes that there are two types of laws, just and unjust; he describes a just law as a "code that squares with the moral law" and an unjust law as a "code that is out of harmony with the moral law." Thus, King's opposition to the injustice of legalized segregation in the twentieth century is philosophically akin to the transcendentalists' opposition to the Fugitive Slave Law in the nineteenth century.
199406_3-RC_1_1
[ "King's philosophy was more influenced by Thoreau's essay on civil disobedience than by any other writing of the transcendentalists.", "While historians may have overestimated Thoreau's influence on King, King was greatly influenced by a number of the transcendentalist philosophers.", "Thoreau's and King's views on civil disobedience differed in that King was more concerned with the social reform than with the economic reform of society.", "Although historians have overemphasized Thoreau's influence on King, there are parallels between King's philosophy and transcendentalism that have not been fully appreciated.", "King's ideas about law and civil disobedience were influenced by transcendentalism in general and Thoreau's essays in particular." ]
3
Which one of the following best states the main idea of the passage?
Nearly every writer on the philosophy of civil rights activist Martin Luther King, Jr., makes a connection between King and Henry David Thoreau, usually via Thoreau's famous essay, "Civil Disobedience" (1849). In his book Stride Toward Freedom (1958), King himself stated that Thoreau's essay was his first intellectual contact with the theory of passive resistance to governmental laws that are perceived as morally unjust. However, this emphasis on Thoreau's influence on King is unfortunate: first, King would not have agreed with many other aspects of Thoreau's philosophy, including Thoreau's ultimate acceptance of violence as a form of protest; second, an overemphasis on the influence of one essay has kept historians from noting other correspondences between King's philosophy and transcendentalism. "Civil Disobedience" was the only example of transcendentalist writing with which King was familiar, and in many other transcendentalist writings, including works by Ralph Waldo Emerson and Margaret Fuller, King would have found ideas more nearly akin to his own. The kind of civil disobedience King had in mind was, in fact, quite different from Thoreau's view of civil disobedience. Thoreau, like most other transcendentalists, was primarily interested in reform of the individual, whereas King was primarily interested in reform of society. As a protest against the Mexican War, Thoreau refused to pay taxes, but he did not hope by his action to force a change in national policy. While he encouraged others to adopt similar protests, he did not attempt to mount any mass protest action against unjust laws. In contrast to Thoreau, King began to advocate the use of mass civil disobedience to effect revolutionary changes within the social system. However, King's writings suggest that, without realizing it, he was an incipient transcendentalist. Most transcendentalists subscribed to the concept of "higher law" and included civil disobedience to unjust laws as part of their strategy. They often invoked the concept of higher law to justify their opposition to slavery and to advocate disobedience to the strengthened Fugitive Slave Law of 1850. In his second major book, King's discussion of just and unjust laws and the responsibility of the individual is very similar to the transcendentalists' discussion of higher law. In reference to how one can advocate breaking some laws and obeying others, King notes that there are two types of laws, just and unjust; he describes a just law as a "code that squares with the moral law" and an unjust law as a "code that is out of harmony with the moral law." Thus, King's opposition to the injustice of legalized segregation in the twentieth century is philosophically akin to the transcendentalists' opposition to the Fugitive Slave Law in the nineteenth century.
199406_3-RC_1_2
[ "It was not King's first contact with the concept of passive resistance to unjust laws.", "It was one of many examples of transcendentalist writing with which King was familiar.", "It provided King with a model for using passive resistance to effect social change.", "It contains a number of ideas with which other transcendentalists strongly disagreed.", "It influenced King's philosophy on passive resistance to unjust laws." ]
4
Which one of the following statements about "Civil Disobedience" would the author consider most accurate?
Nearly every writer on the philosophy of civil rights activist Martin Luther King, Jr., makes a connection between King and Henry David Thoreau, usually via Thoreau's famous essay, "Civil Disobedience" (1849). In his book Stride Toward Freedom (1958), King himself stated that Thoreau's essay was his first intellectual contact with the theory of passive resistance to governmental laws that are perceived as morally unjust. However, this emphasis on Thoreau's influence on King is unfortunate: first, King would not have agreed with many other aspects of Thoreau's philosophy, including Thoreau's ultimate acceptance of violence as a form of protest; second, an overemphasis on the influence of one essay has kept historians from noting other correspondences between King's philosophy and transcendentalism. "Civil Disobedience" was the only example of transcendentalist writing with which King was familiar, and in many other transcendentalist writings, including works by Ralph Waldo Emerson and Margaret Fuller, King would have found ideas more nearly akin to his own. The kind of civil disobedience King had in mind was, in fact, quite different from Thoreau's view of civil disobedience. Thoreau, like most other transcendentalists, was primarily interested in reform of the individual, whereas King was primarily interested in reform of society. As a protest against the Mexican War, Thoreau refused to pay taxes, but he did not hope by his action to force a change in national policy. While he encouraged others to adopt similar protests, he did not attempt to mount any mass protest action against unjust laws. In contrast to Thoreau, King began to advocate the use of mass civil disobedience to effect revolutionary changes within the social system. However, King's writings suggest that, without realizing it, he was an incipient transcendentalist. Most transcendentalists subscribed to the concept of "higher law" and included civil disobedience to unjust laws as part of their strategy. They often invoked the concept of higher law to justify their opposition to slavery and to advocate disobedience to the strengthened Fugitive Slave Law of 1850. In his second major book, King's discussion of just and unjust laws and the responsibility of the individual is very similar to the transcendentalists' discussion of higher law. In reference to how one can advocate breaking some laws and obeying others, King notes that there are two types of laws, just and unjust; he describes a just law as a "code that squares with the moral law" and an unjust law as a "code that is out of harmony with the moral law." Thus, King's opposition to the injustice of legalized segregation in the twentieth century is philosophically akin to the transcendentalists' opposition to the Fugitive Slave Law in the nineteenth century.
199406_3-RC_1_3
[ "chronicling the development of King's philosophy on passive resistance to unjust law", "suggesting that a common emphasis on one influence on King's philosophy has been misleading", "providing new information about the influence of twentieth-century philosophers on King's work", "summarizing the work of historians on the most important influences on King's philosophy", "providing background information about nineteenth-century transcendentalist philosophers" ]
1
In the first paragraph, the author is primarily concerned with
Nearly every writer on the philosophy of civil rights activist Martin Luther King, Jr., makes a connection between King and Henry David Thoreau, usually via Thoreau's famous essay, "Civil Disobedience" (1849). In his book Stride Toward Freedom (1958), King himself stated that Thoreau's essay was his first intellectual contact with the theory of passive resistance to governmental laws that are perceived as morally unjust. However, this emphasis on Thoreau's influence on King is unfortunate: first, King would not have agreed with many other aspects of Thoreau's philosophy, including Thoreau's ultimate acceptance of violence as a form of protest; second, an overemphasis on the influence of one essay has kept historians from noting other correspondences between King's philosophy and transcendentalism. "Civil Disobedience" was the only example of transcendentalist writing with which King was familiar, and in many other transcendentalist writings, including works by Ralph Waldo Emerson and Margaret Fuller, King would have found ideas more nearly akin to his own. The kind of civil disobedience King had in mind was, in fact, quite different from Thoreau's view of civil disobedience. Thoreau, like most other transcendentalists, was primarily interested in reform of the individual, whereas King was primarily interested in reform of society. As a protest against the Mexican War, Thoreau refused to pay taxes, but he did not hope by his action to force a change in national policy. While he encouraged others to adopt similar protests, he did not attempt to mount any mass protest action against unjust laws. In contrast to Thoreau, King began to advocate the use of mass civil disobedience to effect revolutionary changes within the social system. However, King's writings suggest that, without realizing it, he was an incipient transcendentalist. Most transcendentalists subscribed to the concept of "higher law" and included civil disobedience to unjust laws as part of their strategy. They often invoked the concept of higher law to justify their opposition to slavery and to advocate disobedience to the strengthened Fugitive Slave Law of 1850. In his second major book, King's discussion of just and unjust laws and the responsibility of the individual is very similar to the transcendentalists' discussion of higher law. In reference to how one can advocate breaking some laws and obeying others, King notes that there are two types of laws, just and unjust; he describes a just law as a "code that squares with the moral law" and an unjust law as a "code that is out of harmony with the moral law." Thus, King's opposition to the injustice of legalized segregation in the twentieth century is philosophically akin to the transcendentalists' opposition to the Fugitive Slave Law in the nineteenth century.
199406_3-RC_1_4
[ "Some of their ideas were less typical of transcendentalism than were some of Thoreau's ideas.", "They were more concerned with the reform of society than with the reform of the individual.", "They would have been more likely than Thoreau to agree with King on the necessity of mass protest in civil disobedience.", "Their ideas about civil disobedience and unjust laws are as well known as Thoreau's are.", "Some of their ideas were more similar to King's than were some of Thoreau's." ]
4
According to the passage, which one of the following is true of Emerson and Fuller?
Nearly every writer on the philosophy of civil rights activist Martin Luther King, Jr., makes a connection between King and Henry David Thoreau, usually via Thoreau's famous essay, "Civil Disobedience" (1849). In his book Stride Toward Freedom (1958), King himself stated that Thoreau's essay was his first intellectual contact with the theory of passive resistance to governmental laws that are perceived as morally unjust. However, this emphasis on Thoreau's influence on King is unfortunate: first, King would not have agreed with many other aspects of Thoreau's philosophy, including Thoreau's ultimate acceptance of violence as a form of protest; second, an overemphasis on the influence of one essay has kept historians from noting other correspondences between King's philosophy and transcendentalism. "Civil Disobedience" was the only example of transcendentalist writing with which King was familiar, and in many other transcendentalist writings, including works by Ralph Waldo Emerson and Margaret Fuller, King would have found ideas more nearly akin to his own. The kind of civil disobedience King had in mind was, in fact, quite different from Thoreau's view of civil disobedience. Thoreau, like most other transcendentalists, was primarily interested in reform of the individual, whereas King was primarily interested in reform of society. As a protest against the Mexican War, Thoreau refused to pay taxes, but he did not hope by his action to force a change in national policy. While he encouraged others to adopt similar protests, he did not attempt to mount any mass protest action against unjust laws. In contrast to Thoreau, King began to advocate the use of mass civil disobedience to effect revolutionary changes within the social system. However, King's writings suggest that, without realizing it, he was an incipient transcendentalist. Most transcendentalists subscribed to the concept of "higher law" and included civil disobedience to unjust laws as part of their strategy. They often invoked the concept of higher law to justify their opposition to slavery and to advocate disobedience to the strengthened Fugitive Slave Law of 1850. In his second major book, King's discussion of just and unjust laws and the responsibility of the individual is very similar to the transcendentalists' discussion of higher law. In reference to how one can advocate breaking some laws and obeying others, King notes that there are two types of laws, just and unjust; he describes a just law as a "code that squares with the moral law" and an unjust law as a "code that is out of harmony with the moral law." Thus, King's opposition to the injustice of legalized segregation in the twentieth century is philosophically akin to the transcendentalists' opposition to the Fugitive Slave Law in the nineteenth century.
199406_3-RC_1_5
[ "opposed violence as a form of civil protest", "opposed war as an instrument of foreign policy under any circumstances", "believed that just laws had an inherent moral value", "was more interested in reforming society than in reforming the individual", "protested social and legal injustice in United States society rather than United States foreign policy" ]
3
According to the passage, King differed from most transcendentalists in that he
Nearly every writer on the philosophy of civil rights activist Martin Luther King, Jr., makes a connection between King and Henry David Thoreau, usually via Thoreau's famous essay, "Civil Disobedience" (1849). In his book Stride Toward Freedom (1958), King himself stated that Thoreau's essay was his first intellectual contact with the theory of passive resistance to governmental laws that are perceived as morally unjust. However, this emphasis on Thoreau's influence on King is unfortunate: first, King would not have agreed with many other aspects of Thoreau's philosophy, including Thoreau's ultimate acceptance of violence as a form of protest; second, an overemphasis on the influence of one essay has kept historians from noting other correspondences between King's philosophy and transcendentalism. "Civil Disobedience" was the only example of transcendentalist writing with which King was familiar, and in many other transcendentalist writings, including works by Ralph Waldo Emerson and Margaret Fuller, King would have found ideas more nearly akin to his own. The kind of civil disobedience King had in mind was, in fact, quite different from Thoreau's view of civil disobedience. Thoreau, like most other transcendentalists, was primarily interested in reform of the individual, whereas King was primarily interested in reform of society. As a protest against the Mexican War, Thoreau refused to pay taxes, but he did not hope by his action to force a change in national policy. While he encouraged others to adopt similar protests, he did not attempt to mount any mass protest action against unjust laws. In contrast to Thoreau, King began to advocate the use of mass civil disobedience to effect revolutionary changes within the social system. However, King's writings suggest that, without realizing it, he was an incipient transcendentalist. Most transcendentalists subscribed to the concept of "higher law" and included civil disobedience to unjust laws as part of their strategy. They often invoked the concept of higher law to justify their opposition to slavery and to advocate disobedience to the strengthened Fugitive Slave Law of 1850. In his second major book, King's discussion of just and unjust laws and the responsibility of the individual is very similar to the transcendentalists' discussion of higher law. In reference to how one can advocate breaking some laws and obeying others, King notes that there are two types of laws, just and unjust; he describes a just law as a "code that squares with the moral law" and an unjust law as a "code that is out of harmony with the moral law." Thus, King's opposition to the injustice of legalized segregation in the twentieth century is philosophically akin to the transcendentalists' opposition to the Fugitive Slave Law in the nineteenth century.
199406_3-RC_1_6
[ "He was the first to develop fully the theory of civil disobedience.", "His work has had a greater influence on contemporary thinkers than has the work of Emerson and Fuller.", "His philosophy does not contain all of the same elements as the philosophies of the other transcendentalists.", "He advocated using civil disobedience to force the federal government to change its policies on war.", "He is better known for his ideas on social and legal reform than for his ideas on individual reform." ]
2
The passage suggests which one of the following about Thoreau?
Nearly every writer on the philosophy of civil rights activist Martin Luther King, Jr., makes a connection between King and Henry David Thoreau, usually via Thoreau's famous essay, "Civil Disobedience" (1849). In his book Stride Toward Freedom (1958), King himself stated that Thoreau's essay was his first intellectual contact with the theory of passive resistance to governmental laws that are perceived as morally unjust. However, this emphasis on Thoreau's influence on King is unfortunate: first, King would not have agreed with many other aspects of Thoreau's philosophy, including Thoreau's ultimate acceptance of violence as a form of protest; second, an overemphasis on the influence of one essay has kept historians from noting other correspondences between King's philosophy and transcendentalism. "Civil Disobedience" was the only example of transcendentalist writing with which King was familiar, and in many other transcendentalist writings, including works by Ralph Waldo Emerson and Margaret Fuller, King would have found ideas more nearly akin to his own. The kind of civil disobedience King had in mind was, in fact, quite different from Thoreau's view of civil disobedience. Thoreau, like most other transcendentalists, was primarily interested in reform of the individual, whereas King was primarily interested in reform of society. As a protest against the Mexican War, Thoreau refused to pay taxes, but he did not hope by his action to force a change in national policy. While he encouraged others to adopt similar protests, he did not attempt to mount any mass protest action against unjust laws. In contrast to Thoreau, King began to advocate the use of mass civil disobedience to effect revolutionary changes within the social system. However, King's writings suggest that, without realizing it, he was an incipient transcendentalist. Most transcendentalists subscribed to the concept of "higher law" and included civil disobedience to unjust laws as part of their strategy. They often invoked the concept of higher law to justify their opposition to slavery and to advocate disobedience to the strengthened Fugitive Slave Law of 1850. In his second major book, King's discussion of just and unjust laws and the responsibility of the individual is very similar to the transcendentalists' discussion of higher law. In reference to how one can advocate breaking some laws and obeying others, King notes that there are two types of laws, just and unjust; he describes a just law as a "code that squares with the moral law" and an unjust law as a "code that is out of harmony with the moral law." Thus, King's opposition to the injustice of legalized segregation in the twentieth century is philosophically akin to the transcendentalists' opposition to the Fugitive Slave Law in the nineteenth century.
199406_3-RC_1_7
[ "They are an example of a way in which King's ideas differed from Thoreau's but were similar to the ideas of other transcendentalists.", "They provide evidence that proves that King's philosophy was affected by transcendentalist thought.", "They suggest that King, like the transcendentalists, judged human laws by ethical standards.", "They suggest a theoretical basis for King's philosophy of government.", "They provide a paraphrase of Thoreau's position on just and unjust laws." ]
2
The passage provides support for which one of the following statements about the quotations in lines 52–55 ?
In Democracy and its Critics, Robert Dahl defends both democratic values and pluralist democracies, or polyarchies (a rough shorthand term for Western political systems). Dahl argues convincingly that the idea of democracy rests on political equality—the equal capacity of all citizens to determine or influence collective decisions. Of course, as Dahl recognizes, if hierarchical ordering is inevitable in any structure of government, and if no society can guarantee perfect equality in the resources that may give rise to political influence, the democratic principle of political equality is incapable of full realization. So actual systems can be deemed democratic only as approximations to the ideal. It is on these grounds that Dahl defends polyarchy. As a representative system in which elected officials both determine government policy and are accountable to a broad-based electorate, polyarchy reinforces a diffusion of power away from any single center and toward a variety of individuals, groups, and organizations. It is this centrifugal characteristic, Dahl argues, that makes polyarchy the nearest possible approximation to the democratic ideal. Polyarchy achieves this diffusion of power through party competition and the operation of pressure groups. Competing for votes, parties seek to offer different sections of the electorate what they most want; they do not ask what the majority thinks of an issue, but what policy commitments will sway the electoral decisions of particular groups. Equally, groups that have strong feelings about an issue can organize in pressure groups to influence public policy. During the 1960s and 1970s, criticism of the theory of pluralist democracy was vigorous. Many critics pointed to a gap between the model and the reality of Western political systems. They argued that the distribution of power resources other than the vote was so uneven that the political order systematically gave added weight to those who were already richer or organizationally more powerful. So the power of some groups to exclude issues altogether from the political agenda effectively countered any diffusion of influence on decision-making. Although such criticism became subdued during the 1980s, Dahl himself seems to support some of the earlier criticism. Although he regrets that some Western intellectuals demand more democracy from polyarchies than is possible, and is cautious about the possibility of further democratization, he nevertheless ends his book by asking what changes in structures and consciousness might make political life more democratic in present polyarchies. One answer, he suggests, is to look at the economic order of polyarchies from the point of view of the citizen as well as from that of producers and consumers. This would require a critical examination of both the distribution of those economic resources that are at the same time political resources, and the relationship between political structures and economic enterprises.
199406_3-RC_2_8
[ "way in which political power is decentralized in a polyarchy", "central role of power resources in a polyarchy", "kind of concentrated power that political parties generate in a polyarchy", "dynamic balance that exists between economic enterprises and elected officials in a polyarchy", "dynamic balance that exists between voters and elected officials in a polyarchy" ]
0
The characterization of polyarchies as "centrifugal" (line 22) emphasizes the
In Democracy and its Critics, Robert Dahl defends both democratic values and pluralist democracies, or polyarchies (a rough shorthand term for Western political systems). Dahl argues convincingly that the idea of democracy rests on political equality—the equal capacity of all citizens to determine or influence collective decisions. Of course, as Dahl recognizes, if hierarchical ordering is inevitable in any structure of government, and if no society can guarantee perfect equality in the resources that may give rise to political influence, the democratic principle of political equality is incapable of full realization. So actual systems can be deemed democratic only as approximations to the ideal. It is on these grounds that Dahl defends polyarchy. As a representative system in which elected officials both determine government policy and are accountable to a broad-based electorate, polyarchy reinforces a diffusion of power away from any single center and toward a variety of individuals, groups, and organizations. It is this centrifugal characteristic, Dahl argues, that makes polyarchy the nearest possible approximation to the democratic ideal. Polyarchy achieves this diffusion of power through party competition and the operation of pressure groups. Competing for votes, parties seek to offer different sections of the electorate what they most want; they do not ask what the majority thinks of an issue, but what policy commitments will sway the electoral decisions of particular groups. Equally, groups that have strong feelings about an issue can organize in pressure groups to influence public policy. During the 1960s and 1970s, criticism of the theory of pluralist democracy was vigorous. Many critics pointed to a gap between the model and the reality of Western political systems. They argued that the distribution of power resources other than the vote was so uneven that the political order systematically gave added weight to those who were already richer or organizationally more powerful. So the power of some groups to exclude issues altogether from the political agenda effectively countered any diffusion of influence on decision-making. Although such criticism became subdued during the 1980s, Dahl himself seems to support some of the earlier criticism. Although he regrets that some Western intellectuals demand more democracy from polyarchies than is possible, and is cautious about the possibility of further democratization, he nevertheless ends his book by asking what changes in structures and consciousness might make political life more democratic in present polyarchies. One answer, he suggests, is to look at the economic order of polyarchies from the point of view of the citizen as well as from that of producers and consumers. This would require a critical examination of both the distribution of those economic resources that are at the same time political resources, and the relationship between political structures and economic enterprises.
199406_3-RC_2_9
[ "refute Dahl's statement that Western intellectuals expect more democracy from polyarchies than is possible", "advocate the need for rethinking the basic principles on which the theory of democracy rests", "suggest that the structure of government within pluralist democracies should be changed", "point out the flaw in Dahl's argument that the principle of equality cannot be fully realized", "point out an objection to Dahl's defense of polyarchy" ]
4
In the third paragraph, the author of the passage refers to criticism of the theory of pluralist democracy primarily in order to
In Democracy and its Critics, Robert Dahl defends both democratic values and pluralist democracies, or polyarchies (a rough shorthand term for Western political systems). Dahl argues convincingly that the idea of democracy rests on political equality—the equal capacity of all citizens to determine or influence collective decisions. Of course, as Dahl recognizes, if hierarchical ordering is inevitable in any structure of government, and if no society can guarantee perfect equality in the resources that may give rise to political influence, the democratic principle of political equality is incapable of full realization. So actual systems can be deemed democratic only as approximations to the ideal. It is on these grounds that Dahl defends polyarchy. As a representative system in which elected officials both determine government policy and are accountable to a broad-based electorate, polyarchy reinforces a diffusion of power away from any single center and toward a variety of individuals, groups, and organizations. It is this centrifugal characteristic, Dahl argues, that makes polyarchy the nearest possible approximation to the democratic ideal. Polyarchy achieves this diffusion of power through party competition and the operation of pressure groups. Competing for votes, parties seek to offer different sections of the electorate what they most want; they do not ask what the majority thinks of an issue, but what policy commitments will sway the electoral decisions of particular groups. Equally, groups that have strong feelings about an issue can organize in pressure groups to influence public policy. During the 1960s and 1970s, criticism of the theory of pluralist democracy was vigorous. Many critics pointed to a gap between the model and the reality of Western political systems. They argued that the distribution of power resources other than the vote was so uneven that the political order systematically gave added weight to those who were already richer or organizationally more powerful. So the power of some groups to exclude issues altogether from the political agenda effectively countered any diffusion of influence on decision-making. Although such criticism became subdued during the 1980s, Dahl himself seems to support some of the earlier criticism. Although he regrets that some Western intellectuals demand more democracy from polyarchies than is possible, and is cautious about the possibility of further democratization, he nevertheless ends his book by asking what changes in structures and consciousness might make political life more democratic in present polyarchies. One answer, he suggests, is to look at the economic order of polyarchies from the point of view of the citizen as well as from that of producers and consumers. This would require a critical examination of both the distribution of those economic resources that are at the same time political resources, and the relationship between political structures and economic enterprises.
199406_3-RC_2_10
[ "determine what the position of the majority of voters is on a particular issue", "determine what position on an issue will earn the support of particular groups of voters", "organize voters into pressure groups in order to influence public policy on a particular issue", "ensure that elected officials accurately represent the position of the party on specific issues", "ensure that elected officials accurately represent the position of the electorate on specific issues" ]
1
According to the passage, the aim of a political party in a polyarchy is to do which one of the following?
In Democracy and its Critics, Robert Dahl defends both democratic values and pluralist democracies, or polyarchies (a rough shorthand term for Western political systems). Dahl argues convincingly that the idea of democracy rests on political equality—the equal capacity of all citizens to determine or influence collective decisions. Of course, as Dahl recognizes, if hierarchical ordering is inevitable in any structure of government, and if no society can guarantee perfect equality in the resources that may give rise to political influence, the democratic principle of political equality is incapable of full realization. So actual systems can be deemed democratic only as approximations to the ideal. It is on these grounds that Dahl defends polyarchy. As a representative system in which elected officials both determine government policy and are accountable to a broad-based electorate, polyarchy reinforces a diffusion of power away from any single center and toward a variety of individuals, groups, and organizations. It is this centrifugal characteristic, Dahl argues, that makes polyarchy the nearest possible approximation to the democratic ideal. Polyarchy achieves this diffusion of power through party competition and the operation of pressure groups. Competing for votes, parties seek to offer different sections of the electorate what they most want; they do not ask what the majority thinks of an issue, but what policy commitments will sway the electoral decisions of particular groups. Equally, groups that have strong feelings about an issue can organize in pressure groups to influence public policy. During the 1960s and 1970s, criticism of the theory of pluralist democracy was vigorous. Many critics pointed to a gap between the model and the reality of Western political systems. They argued that the distribution of power resources other than the vote was so uneven that the political order systematically gave added weight to those who were already richer or organizationally more powerful. So the power of some groups to exclude issues altogether from the political agenda effectively countered any diffusion of influence on decision-making. Although such criticism became subdued during the 1980s, Dahl himself seems to support some of the earlier criticism. Although he regrets that some Western intellectuals demand more democracy from polyarchies than is possible, and is cautious about the possibility of further democratization, he nevertheless ends his book by asking what changes in structures and consciousness might make political life more democratic in present polyarchies. One answer, he suggests, is to look at the economic order of polyarchies from the point of view of the citizen as well as from that of producers and consumers. This would require a critical examination of both the distribution of those economic resources that are at the same time political resources, and the relationship between political structures and economic enterprises.
199406_3-RC_2_11
[ "Polyarchies are limited in the extent to which they can embody the idea of democracy.", "The structure of polyarchical governments is free of hierarchical ordering.", "The citizens of a polyarchy have equal access to the resources that provide political influence.", "Polyarchy is the best political system to foster the growth of political parties.", "Polyarchy is a form of government that is not influenced by the interests of economic enterprises." ]
0
It can be inferred from the passage that Dahl assumes which one of the following in his defense of polyarchies?
In Democracy and its Critics, Robert Dahl defends both democratic values and pluralist democracies, or polyarchies (a rough shorthand term for Western political systems). Dahl argues convincingly that the idea of democracy rests on political equality—the equal capacity of all citizens to determine or influence collective decisions. Of course, as Dahl recognizes, if hierarchical ordering is inevitable in any structure of government, and if no society can guarantee perfect equality in the resources that may give rise to political influence, the democratic principle of political equality is incapable of full realization. So actual systems can be deemed democratic only as approximations to the ideal. It is on these grounds that Dahl defends polyarchy. As a representative system in which elected officials both determine government policy and are accountable to a broad-based electorate, polyarchy reinforces a diffusion of power away from any single center and toward a variety of individuals, groups, and organizations. It is this centrifugal characteristic, Dahl argues, that makes polyarchy the nearest possible approximation to the democratic ideal. Polyarchy achieves this diffusion of power through party competition and the operation of pressure groups. Competing for votes, parties seek to offer different sections of the electorate what they most want; they do not ask what the majority thinks of an issue, but what policy commitments will sway the electoral decisions of particular groups. Equally, groups that have strong feelings about an issue can organize in pressure groups to influence public policy. During the 1960s and 1970s, criticism of the theory of pluralist democracy was vigorous. Many critics pointed to a gap between the model and the reality of Western political systems. They argued that the distribution of power resources other than the vote was so uneven that the political order systematically gave added weight to those who were already richer or organizationally more powerful. So the power of some groups to exclude issues altogether from the political agenda effectively countered any diffusion of influence on decision-making. Although such criticism became subdued during the 1980s, Dahl himself seems to support some of the earlier criticism. Although he regrets that some Western intellectuals demand more democracy from polyarchies than is possible, and is cautious about the possibility of further democratization, he nevertheless ends his book by asking what changes in structures and consciousness might make political life more democratic in present polyarchies. One answer, he suggests, is to look at the economic order of polyarchies from the point of view of the citizen as well as from that of producers and consumers. This would require a critical examination of both the distribution of those economic resources that are at the same time political resources, and the relationship between political structures and economic enterprises.
199406_3-RC_2_12
[ "an exact copy of an ancient artifact that is on display in a museum", "a performance of a musical score whose range of tonality cannot be completely captured by any actual instruments", "a lecture by a former astronaut to a class of young students who would like to be astronauts", "the commemoration of a historical event each year by a historian presenting a lecture on a topic related to the event", "the mold from which a number of identical castings of a sculpture are made" ]
1
Which one of the following is most closely analogous to pluralist democracies as they are described in relation to the democratic principle of political equality?
In Democracy and its Critics, Robert Dahl defends both democratic values and pluralist democracies, or polyarchies (a rough shorthand term for Western political systems). Dahl argues convincingly that the idea of democracy rests on political equality—the equal capacity of all citizens to determine or influence collective decisions. Of course, as Dahl recognizes, if hierarchical ordering is inevitable in any structure of government, and if no society can guarantee perfect equality in the resources that may give rise to political influence, the democratic principle of political equality is incapable of full realization. So actual systems can be deemed democratic only as approximations to the ideal. It is on these grounds that Dahl defends polyarchy. As a representative system in which elected officials both determine government policy and are accountable to a broad-based electorate, polyarchy reinforces a diffusion of power away from any single center and toward a variety of individuals, groups, and organizations. It is this centrifugal characteristic, Dahl argues, that makes polyarchy the nearest possible approximation to the democratic ideal. Polyarchy achieves this diffusion of power through party competition and the operation of pressure groups. Competing for votes, parties seek to offer different sections of the electorate what they most want; they do not ask what the majority thinks of an issue, but what policy commitments will sway the electoral decisions of particular groups. Equally, groups that have strong feelings about an issue can organize in pressure groups to influence public policy. During the 1960s and 1970s, criticism of the theory of pluralist democracy was vigorous. Many critics pointed to a gap between the model and the reality of Western political systems. They argued that the distribution of power resources other than the vote was so uneven that the political order systematically gave added weight to those who were already richer or organizationally more powerful. So the power of some groups to exclude issues altogether from the political agenda effectively countered any diffusion of influence on decision-making. Although such criticism became subdued during the 1980s, Dahl himself seems to support some of the earlier criticism. Although he regrets that some Western intellectuals demand more democracy from polyarchies than is possible, and is cautious about the possibility of further democratization, he nevertheless ends his book by asking what changes in structures and consciousness might make political life more democratic in present polyarchies. One answer, he suggests, is to look at the economic order of polyarchies from the point of view of the citizen as well as from that of producers and consumers. This would require a critical examination of both the distribution of those economic resources that are at the same time political resources, and the relationship between political structures and economic enterprises.
199406_3-RC_2_13
[ "The political agenda in a polyarchy is strongly influenced by how power resources other than the vote are distributed.", "The outcome of elections is more often determined by the financial resources candidates are able to spend during campaigns than by their stands on political issues.", "Public policy in a polyarchy is primarily determined by decision-makers who are not accountable to elected officials.", "Political parties in a polyarchy help concentrate political power in the central government.", "Small and diverse pressure groups are able to exert as much influence on public policy in a polyarchy as are large and powerful groups." ]
4
Which one of the following, if true, would most strengthen Dahl's defense of polyarchy?
In Democracy and its Critics, Robert Dahl defends both democratic values and pluralist democracies, or polyarchies (a rough shorthand term for Western political systems). Dahl argues convincingly that the idea of democracy rests on political equality—the equal capacity of all citizens to determine or influence collective decisions. Of course, as Dahl recognizes, if hierarchical ordering is inevitable in any structure of government, and if no society can guarantee perfect equality in the resources that may give rise to political influence, the democratic principle of political equality is incapable of full realization. So actual systems can be deemed democratic only as approximations to the ideal. It is on these grounds that Dahl defends polyarchy. As a representative system in which elected officials both determine government policy and are accountable to a broad-based electorate, polyarchy reinforces a diffusion of power away from any single center and toward a variety of individuals, groups, and organizations. It is this centrifugal characteristic, Dahl argues, that makes polyarchy the nearest possible approximation to the democratic ideal. Polyarchy achieves this diffusion of power through party competition and the operation of pressure groups. Competing for votes, parties seek to offer different sections of the electorate what they most want; they do not ask what the majority thinks of an issue, but what policy commitments will sway the electoral decisions of particular groups. Equally, groups that have strong feelings about an issue can organize in pressure groups to influence public policy. During the 1960s and 1970s, criticism of the theory of pluralist democracy was vigorous. Many critics pointed to a gap between the model and the reality of Western political systems. They argued that the distribution of power resources other than the vote was so uneven that the political order systematically gave added weight to those who were already richer or organizationally more powerful. So the power of some groups to exclude issues altogether from the political agenda effectively countered any diffusion of influence on decision-making. Although such criticism became subdued during the 1980s, Dahl himself seems to support some of the earlier criticism. Although he regrets that some Western intellectuals demand more democracy from polyarchies than is possible, and is cautious about the possibility of further democratization, he nevertheless ends his book by asking what changes in structures and consciousness might make political life more democratic in present polyarchies. One answer, he suggests, is to look at the economic order of polyarchies from the point of view of the citizen as well as from that of producers and consumers. This would require a critical examination of both the distribution of those economic resources that are at the same time political resources, and the relationship between political structures and economic enterprises.
199406_3-RC_2_14
[ "an inquiry into how present-day polyarchies can be made more democratic", "a commentary on the means pressure groups employ to exert influence within polyarchies", "a description of the relationship between polyarchies and economic enterprises", "a discussion of the strengths and weaknesses of polyarchy as a form of democracy", "an overview of the similarities between political parties and pressure groups in a polyarchy" ]
3
The passage can best be described as
The old belief that climatic stability accounts for the high level of species diversity in the Amazon River basin of South America emerged, strangely enough, from observations of the deep sea. Sanders discovered high diversity among the mud-dwelling animals of the deep ocean. He argued that such diversity could be attributed to the absence of significant fluctuations in climate and physical conditions, without which the extinction of species should be rare. In the course of time new species would continue to evolve, and so the rate of speciation would be greater than the rate of extinction, resulting in the accumulation of great diversity. Sanders argued that the Amazon tropical rain forest is analogous to the deep sea: because the rain forest has a stable climate, extinction should be rare. Evidence that some species of rain forest trees have persisted for some 30 million years in the Amazon basin, added to the absence of winter and glaciation, supports this view. Recently, however, several observations have cast doubt on the validity of the stability hypothesis and suggest that the climate of the Amazon basin has fluctuated significantly in the past. Haffer noted that different species of birds inhabit different corners of the basin in spite of the fact that essentially unbroken green forest spreads from the western edge to the eastern edge of the region. This pattern presented a puzzle to biologists studying the distributions of plants and animals: why would different species inhabit different parts of the forest if the habitat in which they lived had a stable climate? Haffer proposed a compelling explanation for the distribution of species. Observing that species found on high ground are different from those on low ground, and knowing that in the Amazon lowlands are drier than uplands, he proposed that during the ice ages the Amazon lowlands became a near-desert arid plain; meanwhile, the more elevated regions became islands of moisture and hence served as refuges for the fauna and flora of the rain forest. Populations that were once continuous diverged and became permanently separated. Haffer's hypothesis appears to explain the distribution of species as well as the unusual species diversity. The ice-age refuges would have protected existing species from extinction. But the periodic geographic isolation of related populations (there have been an estimated 13 ice ages to date) would have facilitated the development of new species as existing species on the lowlands adapted to changing climates. Although no conclusive proof has yet been found to support Haffer's hypothesis, it has led other researchers to gauge the effects of climatic changes, such as storms and flooding, on species diversity in the Amazon basin. Their research suggests that climatic disturbances help account for the splendid diversity of the Amazon rain forest today.
199406_3-RC_3_15
[ "Both the Amazon basin and the deep sea support an unusually high rate of speciation.", "Both the rain forest trees in the Amazon basin and the mud-dwelling animals in the deep sea have survived for 30 million years.", "Both the deep sea and the Amazon basin have not experienced dramatic changes in climate or physical conditions.", "A dependable supply of water to the Amazon basin and the deep sea has moderated the rate of extinction in both habitats.", "The rate of speciation in the Amazon basin is equivalent to the rate of speciation in the deep sea." ]
2
As discussed in the first paragraph of the passage Sanders' analogy between the deep sea and the Amazon basin involves which one of the following assumptions?
The old belief that climatic stability accounts for the high level of species diversity in the Amazon River basin of South America emerged, strangely enough, from observations of the deep sea. Sanders discovered high diversity among the mud-dwelling animals of the deep ocean. He argued that such diversity could be attributed to the absence of significant fluctuations in climate and physical conditions, without which the extinction of species should be rare. In the course of time new species would continue to evolve, and so the rate of speciation would be greater than the rate of extinction, resulting in the accumulation of great diversity. Sanders argued that the Amazon tropical rain forest is analogous to the deep sea: because the rain forest has a stable climate, extinction should be rare. Evidence that some species of rain forest trees have persisted for some 30 million years in the Amazon basin, added to the absence of winter and glaciation, supports this view. Recently, however, several observations have cast doubt on the validity of the stability hypothesis and suggest that the climate of the Amazon basin has fluctuated significantly in the past. Haffer noted that different species of birds inhabit different corners of the basin in spite of the fact that essentially unbroken green forest spreads from the western edge to the eastern edge of the region. This pattern presented a puzzle to biologists studying the distributions of plants and animals: why would different species inhabit different parts of the forest if the habitat in which they lived had a stable climate? Haffer proposed a compelling explanation for the distribution of species. Observing that species found on high ground are different from those on low ground, and knowing that in the Amazon lowlands are drier than uplands, he proposed that during the ice ages the Amazon lowlands became a near-desert arid plain; meanwhile, the more elevated regions became islands of moisture and hence served as refuges for the fauna and flora of the rain forest. Populations that were once continuous diverged and became permanently separated. Haffer's hypothesis appears to explain the distribution of species as well as the unusual species diversity. The ice-age refuges would have protected existing species from extinction. But the periodic geographic isolation of related populations (there have been an estimated 13 ice ages to date) would have facilitated the development of new species as existing species on the lowlands adapted to changing climates. Although no conclusive proof has yet been found to support Haffer's hypothesis, it has led other researchers to gauge the effects of climatic changes, such as storms and flooding, on species diversity in the Amazon basin. Their research suggests that climatic disturbances help account for the splendid diversity of the Amazon rain forest today.
199406_3-RC_3_16
[ "It provides an intriguing and complete explanation for the high rate of species diversity in the Amazon basin.", "It is partially correct in that a number of climatic disturbances account for species diversity in the Amazon basin.", "It has not yet been verified, but it has had an influential effect on current research on species diversity in the Amazon basin.", "It is better than Sanders' theory in accounting for the low rate of species extinction in the Amazon basin.", "It provides a compelling explanation for the distribution of species in the Amazon basin but does not account for the high species diversity." ]
2
The author of the passage would most likely agree with which one of the following statements about Haffer's hypothesis?
The old belief that climatic stability accounts for the high level of species diversity in the Amazon River basin of South America emerged, strangely enough, from observations of the deep sea. Sanders discovered high diversity among the mud-dwelling animals of the deep ocean. He argued that such diversity could be attributed to the absence of significant fluctuations in climate and physical conditions, without which the extinction of species should be rare. In the course of time new species would continue to evolve, and so the rate of speciation would be greater than the rate of extinction, resulting in the accumulation of great diversity. Sanders argued that the Amazon tropical rain forest is analogous to the deep sea: because the rain forest has a stable climate, extinction should be rare. Evidence that some species of rain forest trees have persisted for some 30 million years in the Amazon basin, added to the absence of winter and glaciation, supports this view. Recently, however, several observations have cast doubt on the validity of the stability hypothesis and suggest that the climate of the Amazon basin has fluctuated significantly in the past. Haffer noted that different species of birds inhabit different corners of the basin in spite of the fact that essentially unbroken green forest spreads from the western edge to the eastern edge of the region. This pattern presented a puzzle to biologists studying the distributions of plants and animals: why would different species inhabit different parts of the forest if the habitat in which they lived had a stable climate? Haffer proposed a compelling explanation for the distribution of species. Observing that species found on high ground are different from those on low ground, and knowing that in the Amazon lowlands are drier than uplands, he proposed that during the ice ages the Amazon lowlands became a near-desert arid plain; meanwhile, the more elevated regions became islands of moisture and hence served as refuges for the fauna and flora of the rain forest. Populations that were once continuous diverged and became permanently separated. Haffer's hypothesis appears to explain the distribution of species as well as the unusual species diversity. The ice-age refuges would have protected existing species from extinction. But the periodic geographic isolation of related populations (there have been an estimated 13 ice ages to date) would have facilitated the development of new species as existing species on the lowlands adapted to changing climates. Although no conclusive proof has yet been found to support Haffer's hypothesis, it has led other researchers to gauge the effects of climatic changes, such as storms and flooding, on species diversity in the Amazon basin. Their research suggests that climatic disturbances help account for the splendid diversity of the Amazon rain forest today.
199406_3-RC_3_17
[ "Lowlands are desertlike, whereas uplands are lush.", "Lowlands are less vulnerable to glaciation during the ice ages than are uplands.", "Uplands support a greater diversity of species than do lowlands.", "Uplands are wetter than are lowlands.", "Uplands are more densely populated than are lowlands." ]
3
According to the passage, lowlands in the Amazon basin currently differ from uplands in which one of the following respects?
The old belief that climatic stability accounts for the high level of species diversity in the Amazon River basin of South America emerged, strangely enough, from observations of the deep sea. Sanders discovered high diversity among the mud-dwelling animals of the deep ocean. He argued that such diversity could be attributed to the absence of significant fluctuations in climate and physical conditions, without which the extinction of species should be rare. In the course of time new species would continue to evolve, and so the rate of speciation would be greater than the rate of extinction, resulting in the accumulation of great diversity. Sanders argued that the Amazon tropical rain forest is analogous to the deep sea: because the rain forest has a stable climate, extinction should be rare. Evidence that some species of rain forest trees have persisted for some 30 million years in the Amazon basin, added to the absence of winter and glaciation, supports this view. Recently, however, several observations have cast doubt on the validity of the stability hypothesis and suggest that the climate of the Amazon basin has fluctuated significantly in the past. Haffer noted that different species of birds inhabit different corners of the basin in spite of the fact that essentially unbroken green forest spreads from the western edge to the eastern edge of the region. This pattern presented a puzzle to biologists studying the distributions of plants and animals: why would different species inhabit different parts of the forest if the habitat in which they lived had a stable climate? Haffer proposed a compelling explanation for the distribution of species. Observing that species found on high ground are different from those on low ground, and knowing that in the Amazon lowlands are drier than uplands, he proposed that during the ice ages the Amazon lowlands became a near-desert arid plain; meanwhile, the more elevated regions became islands of moisture and hence served as refuges for the fauna and flora of the rain forest. Populations that were once continuous diverged and became permanently separated. Haffer's hypothesis appears to explain the distribution of species as well as the unusual species diversity. The ice-age refuges would have protected existing species from extinction. But the periodic geographic isolation of related populations (there have been an estimated 13 ice ages to date) would have facilitated the development of new species as existing species on the lowlands adapted to changing climates. Although no conclusive proof has yet been found to support Haffer's hypothesis, it has led other researchers to gauge the effects of climatic changes, such as storms and flooding, on species diversity in the Amazon basin. Their research suggests that climatic disturbances help account for the splendid diversity of the Amazon rain forest today.
199406_3-RC_3_18
[ "A hypothesis is discussed, evidence that undercuts that hypothesis is presented, and a new hypothesis that may account for the evidence is described.", "A recently observed phenomenon is described, an explanation for that phenomenon is discussed, and the explanation is evaluated in light of previous research findings.", "Several hypotheses that may account for a puzzling phenomenon are described and discounted, and a more promising hypothesis is presented.", "A hypothesis and the assumptions on which it is based are described, and evidence is provided to suggest that the hypothesis is only partially correct.", "Two alternative explanations for a phenomenon are presented and compared, and experiments designed to test each theory are described." ]
0
Which one of the following best describes the organization of the passage?
The old belief that climatic stability accounts for the high level of species diversity in the Amazon River basin of South America emerged, strangely enough, from observations of the deep sea. Sanders discovered high diversity among the mud-dwelling animals of the deep ocean. He argued that such diversity could be attributed to the absence of significant fluctuations in climate and physical conditions, without which the extinction of species should be rare. In the course of time new species would continue to evolve, and so the rate of speciation would be greater than the rate of extinction, resulting in the accumulation of great diversity. Sanders argued that the Amazon tropical rain forest is analogous to the deep sea: because the rain forest has a stable climate, extinction should be rare. Evidence that some species of rain forest trees have persisted for some 30 million years in the Amazon basin, added to the absence of winter and glaciation, supports this view. Recently, however, several observations have cast doubt on the validity of the stability hypothesis and suggest that the climate of the Amazon basin has fluctuated significantly in the past. Haffer noted that different species of birds inhabit different corners of the basin in spite of the fact that essentially unbroken green forest spreads from the western edge to the eastern edge of the region. This pattern presented a puzzle to biologists studying the distributions of plants and animals: why would different species inhabit different parts of the forest if the habitat in which they lived had a stable climate? Haffer proposed a compelling explanation for the distribution of species. Observing that species found on high ground are different from those on low ground, and knowing that in the Amazon lowlands are drier than uplands, he proposed that during the ice ages the Amazon lowlands became a near-desert arid plain; meanwhile, the more elevated regions became islands of moisture and hence served as refuges for the fauna and flora of the rain forest. Populations that were once continuous diverged and became permanently separated. Haffer's hypothesis appears to explain the distribution of species as well as the unusual species diversity. The ice-age refuges would have protected existing species from extinction. But the periodic geographic isolation of related populations (there have been an estimated 13 ice ages to date) would have facilitated the development of new species as existing species on the lowlands adapted to changing climates. Although no conclusive proof has yet been found to support Haffer's hypothesis, it has led other researchers to gauge the effects of climatic changes, such as storms and flooding, on species diversity in the Amazon basin. Their research suggests that climatic disturbances help account for the splendid diversity of the Amazon rain forest today.
199406_3-RC_3_19
[ "provide proof that cooler and drier temperatures are primarily responsible for the distribution of species in the Amazon", "explain how populations of species were protected from extinction in the Amazon basin", "explain how most existing species were able to survive periodic climatic disturbances in the Amazon basin", "suggest that certain kinds of climatic disturbances cause more species diversity than do other kinds of climatic disturbances", "suggest that geographic isolation may have occurred often enough to cause high species diversity in the Amazon basin" ]
4
The author of the passage mentions the number of ice ages in the third paragraph most probably in order to
The old belief that climatic stability accounts for the high level of species diversity in the Amazon River basin of South America emerged, strangely enough, from observations of the deep sea. Sanders discovered high diversity among the mud-dwelling animals of the deep ocean. He argued that such diversity could be attributed to the absence of significant fluctuations in climate and physical conditions, without which the extinction of species should be rare. In the course of time new species would continue to evolve, and so the rate of speciation would be greater than the rate of extinction, resulting in the accumulation of great diversity. Sanders argued that the Amazon tropical rain forest is analogous to the deep sea: because the rain forest has a stable climate, extinction should be rare. Evidence that some species of rain forest trees have persisted for some 30 million years in the Amazon basin, added to the absence of winter and glaciation, supports this view. Recently, however, several observations have cast doubt on the validity of the stability hypothesis and suggest that the climate of the Amazon basin has fluctuated significantly in the past. Haffer noted that different species of birds inhabit different corners of the basin in spite of the fact that essentially unbroken green forest spreads from the western edge to the eastern edge of the region. This pattern presented a puzzle to biologists studying the distributions of plants and animals: why would different species inhabit different parts of the forest if the habitat in which they lived had a stable climate? Haffer proposed a compelling explanation for the distribution of species. Observing that species found on high ground are different from those on low ground, and knowing that in the Amazon lowlands are drier than uplands, he proposed that during the ice ages the Amazon lowlands became a near-desert arid plain; meanwhile, the more elevated regions became islands of moisture and hence served as refuges for the fauna and flora of the rain forest. Populations that were once continuous diverged and became permanently separated. Haffer's hypothesis appears to explain the distribution of species as well as the unusual species diversity. The ice-age refuges would have protected existing species from extinction. But the periodic geographic isolation of related populations (there have been an estimated 13 ice ages to date) would have facilitated the development of new species as existing species on the lowlands adapted to changing climates. Although no conclusive proof has yet been found to support Haffer's hypothesis, it has led other researchers to gauge the effects of climatic changes, such as storms and flooding, on species diversity in the Amazon basin. Their research suggests that climatic disturbances help account for the splendid diversity of the Amazon rain forest today.
199406_3-RC_3_20
[ "He underestimated the effects of winter and glaciation in the Amazon basin on the tropical rain forest.", "He failed to recognize the similarity in physical conditions of the Amazon lowlands and the Amazon uplands.", "He failed to take into account the relatively high rate of extinction during the ice ages in the Amazon basin.", "He overestimated the length of time that species have survived in the Amazon basin.", "He failed to account for the distribution of species in the Amazon basin." ]
4
The passage suggests that which one of the following is true of Sanders' hypothesis?
The old belief that climatic stability accounts for the high level of species diversity in the Amazon River basin of South America emerged, strangely enough, from observations of the deep sea. Sanders discovered high diversity among the mud-dwelling animals of the deep ocean. He argued that such diversity could be attributed to the absence of significant fluctuations in climate and physical conditions, without which the extinction of species should be rare. In the course of time new species would continue to evolve, and so the rate of speciation would be greater than the rate of extinction, resulting in the accumulation of great diversity. Sanders argued that the Amazon tropical rain forest is analogous to the deep sea: because the rain forest has a stable climate, extinction should be rare. Evidence that some species of rain forest trees have persisted for some 30 million years in the Amazon basin, added to the absence of winter and glaciation, supports this view. Recently, however, several observations have cast doubt on the validity of the stability hypothesis and suggest that the climate of the Amazon basin has fluctuated significantly in the past. Haffer noted that different species of birds inhabit different corners of the basin in spite of the fact that essentially unbroken green forest spreads from the western edge to the eastern edge of the region. This pattern presented a puzzle to biologists studying the distributions of plants and animals: why would different species inhabit different parts of the forest if the habitat in which they lived had a stable climate? Haffer proposed a compelling explanation for the distribution of species. Observing that species found on high ground are different from those on low ground, and knowing that in the Amazon lowlands are drier than uplands, he proposed that during the ice ages the Amazon lowlands became a near-desert arid plain; meanwhile, the more elevated regions became islands of moisture and hence served as refuges for the fauna and flora of the rain forest. Populations that were once continuous diverged and became permanently separated. Haffer's hypothesis appears to explain the distribution of species as well as the unusual species diversity. The ice-age refuges would have protected existing species from extinction. But the periodic geographic isolation of related populations (there have been an estimated 13 ice ages to date) would have facilitated the development of new species as existing species on the lowlands adapted to changing climates. Although no conclusive proof has yet been found to support Haffer's hypothesis, it has led other researchers to gauge the effects of climatic changes, such as storms and flooding, on species diversity in the Amazon basin. Their research suggests that climatic disturbances help account for the splendid diversity of the Amazon rain forest today.
199406_3-RC_3_21
[ "Accurately dated sediment cores from a freshwater lake in the Amazon indicate that the lake's water level rose significantly during the last ice age.", "Data based on radiocarbon dating of fossils suggest that the Amazon uplands were too cold to support rain forests during the last ice age.", "Computer models of climate during global ice ages predict only insignificant reductions of monsoon rains in tropical areas such as the Amazon.", "Fossils preserved in the Amazon uplands during the last ice age are found together with minerals that are the products of an arid landscape.", "Fossilized pollen from the Amazon lowlands indicates that during the last ice age the Amazon lowlands supported vegetation that needs little water rather than the rain forests they support today." ]
4
Which one of the following is evidence that would contribute to the "proof" mentioned in line 54?
Although surveys of medieval legislation, guild organization, and terminology used to designate different medical practitioners have demonstrated that numerous medical specialties were recognized in Europe during the Middle Ages, most historians continue to equate the term "woman medical practitioner," wherever they encounter it in medieval records, with "midwife." This common practice obscures the fact that, although women were not represented on all levels of medicine equally, they were represented in a variety of specialties throughout the broad medical community. A reliable study by Wickersheimer and Jacquart documents that, of 7,647 medical practitioners in France during the twelfth through fifteenth centuries, 121 were women; of these, only 44 were identified as midwives, while the rest practiced as physicians, surgeons, apothecaries, barbers, and other healers. While preserving terminological distinctions somewhat increases the quality of the information extracted from medieval documents concerning women medical practitioners, scholars must also reopen the whole question of why documentary evidence for women medical practitioners comprises such a tiny fraction of the evidence historians of medieval medicine usually present. Is this due to the limitations of the historical record, as has been claimed, or does it also result from the methods historians use? Granted, apart from medical licenses, the principal sources of information regarding medical practitioners available to researchers are wills, property transfers, court records, and similar documents, all of which typically underrepresent women because of restrictive medieval legal traditions. Nonetheless, the parameters researchers choose when they define their investigations may contribute to the problem. Studies focusing on the upper echelons of "learned" medicine, for example, tend to exclude healers on the legal and social fringes of medical practice, where most women would have been found. The advantages of broadening the scope of such studies is immediately apparent in Pelling and Webster's study of sixteenth-century London. Instead of focusing solely on officially recognized and licensed practitioners, the researchers defined a medical practitioner as "any individual whose occupation is basically concerned with the care of the sick." Using this definition, they found primary source information suggesting that there were 60 women medical practitioners in the city of London in 1560. Although this figure may be slightly exaggerated, the evidence contrasts strikingly with that of Gottfried, whose earlier survey identified only 28 women medical practitioners in all of England between 1330 and 1530. Finally, such studies provide only statistical information about the variety and prevalence of women's medical practice in medieval Europe. Future studies might also make profitable use of analyses developed in other areas of women's history as a basis for exploring the social context of women's medical practice. Information about economic rivalry in medicine, women's literacy, and the control of medical knowledge could add much to our growing understanding of women medical practitioners' role in medieval society.
199406_3-RC_4_22
[ "Recent studies demonstrate that women medical practitioners were more common in England than in the rest of Western Europe during the Middle Ages.", "The quantity and quality of the information historians uncover concerning women's medical practice in medieval Europe would be improved if they changed their methods of study.", "The sparse evidence for women medical practitioners in studies dealing with the Middle Ages is due primarily to the limitations of the historical record.", "Knowledge about the social issues that influenced the role women medical practitioners played in medieval society has been enhanced by several recent studies.", "Analyses developed in other areas of women's history could probably be used to provide more information about the social context of women's medical practice during the Middle Ages." ]
1
Which one of the following best expresses the main point of the passage?
Although surveys of medieval legislation, guild organization, and terminology used to designate different medical practitioners have demonstrated that numerous medical specialties were recognized in Europe during the Middle Ages, most historians continue to equate the term "woman medical practitioner," wherever they encounter it in medieval records, with "midwife." This common practice obscures the fact that, although women were not represented on all levels of medicine equally, they were represented in a variety of specialties throughout the broad medical community. A reliable study by Wickersheimer and Jacquart documents that, of 7,647 medical practitioners in France during the twelfth through fifteenth centuries, 121 were women; of these, only 44 were identified as midwives, while the rest practiced as physicians, surgeons, apothecaries, barbers, and other healers. While preserving terminological distinctions somewhat increases the quality of the information extracted from medieval documents concerning women medical practitioners, scholars must also reopen the whole question of why documentary evidence for women medical practitioners comprises such a tiny fraction of the evidence historians of medieval medicine usually present. Is this due to the limitations of the historical record, as has been claimed, or does it also result from the methods historians use? Granted, apart from medical licenses, the principal sources of information regarding medical practitioners available to researchers are wills, property transfers, court records, and similar documents, all of which typically underrepresent women because of restrictive medieval legal traditions. Nonetheless, the parameters researchers choose when they define their investigations may contribute to the problem. Studies focusing on the upper echelons of "learned" medicine, for example, tend to exclude healers on the legal and social fringes of medical practice, where most women would have been found. The advantages of broadening the scope of such studies is immediately apparent in Pelling and Webster's study of sixteenth-century London. Instead of focusing solely on officially recognized and licensed practitioners, the researchers defined a medical practitioner as "any individual whose occupation is basically concerned with the care of the sick." Using this definition, they found primary source information suggesting that there were 60 women medical practitioners in the city of London in 1560. Although this figure may be slightly exaggerated, the evidence contrasts strikingly with that of Gottfried, whose earlier survey identified only 28 women medical practitioners in all of England between 1330 and 1530. Finally, such studies provide only statistical information about the variety and prevalence of women's medical practice in medieval Europe. Future studies might also make profitable use of analyses developed in other areas of women's history as a basis for exploring the social context of women's medical practice. Information about economic rivalry in medicine, women's literacy, and the control of medical knowledge could add much to our growing understanding of women medical practitioners' role in medieval society.
199406_3-RC_4_23
[ "equating pear with apple", "equating science with biology", "equating supervisor with subordinate", "equating member with nonmember", "equating instructor with trainee" ]
1
Which one of the following is most closely analogous to the error the author believes historians make when they equate the term "woman medical practitioner" with "midwife" ?
Although surveys of medieval legislation, guild organization, and terminology used to designate different medical practitioners have demonstrated that numerous medical specialties were recognized in Europe during the Middle Ages, most historians continue to equate the term "woman medical practitioner," wherever they encounter it in medieval records, with "midwife." This common practice obscures the fact that, although women were not represented on all levels of medicine equally, they were represented in a variety of specialties throughout the broad medical community. A reliable study by Wickersheimer and Jacquart documents that, of 7,647 medical practitioners in France during the twelfth through fifteenth centuries, 121 were women; of these, only 44 were identified as midwives, while the rest practiced as physicians, surgeons, apothecaries, barbers, and other healers. While preserving terminological distinctions somewhat increases the quality of the information extracted from medieval documents concerning women medical practitioners, scholars must also reopen the whole question of why documentary evidence for women medical practitioners comprises such a tiny fraction of the evidence historians of medieval medicine usually present. Is this due to the limitations of the historical record, as has been claimed, or does it also result from the methods historians use? Granted, apart from medical licenses, the principal sources of information regarding medical practitioners available to researchers are wills, property transfers, court records, and similar documents, all of which typically underrepresent women because of restrictive medieval legal traditions. Nonetheless, the parameters researchers choose when they define their investigations may contribute to the problem. Studies focusing on the upper echelons of "learned" medicine, for example, tend to exclude healers on the legal and social fringes of medical practice, where most women would have been found. The advantages of broadening the scope of such studies is immediately apparent in Pelling and Webster's study of sixteenth-century London. Instead of focusing solely on officially recognized and licensed practitioners, the researchers defined a medical practitioner as "any individual whose occupation is basically concerned with the care of the sick." Using this definition, they found primary source information suggesting that there were 60 women medical practitioners in the city of London in 1560. Although this figure may be slightly exaggerated, the evidence contrasts strikingly with that of Gottfried, whose earlier survey identified only 28 women medical practitioners in all of England between 1330 and 1530. Finally, such studies provide only statistical information about the variety and prevalence of women's medical practice in medieval Europe. Future studies might also make profitable use of analyses developed in other areas of women's history as a basis for exploring the social context of women's medical practice. Information about economic rivalry in medicine, women's literacy, and the control of medical knowledge could add much to our growing understanding of women medical practitioners' role in medieval society.
199406_3-RC_4_24
[ "Gottfried's study would have recorded a much larger number of women medical practitioners if the time frame covered by the study had included the late sixteenth century.", "The small number of women medical practitioners identified in Gottfried's study is due primarily to problems caused by inaccurate sources.", "The small number of women medical practitioners identified in Gottfried's study is due primarily to the loss of many medieval documents.", "The results of Gottfried's study need to be considered in light of the social changes occurring in Western Europe during the fourteenth and fifteenth centuries.", "In setting the parameters for his study, Gottfried appears to have defined the term \"medical practitioner\" very narrowly." ]
4
It can be inferred from the passage that the author would be most likely to agree with which one of the following assertions regarding Gottfried's study?
Although surveys of medieval legislation, guild organization, and terminology used to designate different medical practitioners have demonstrated that numerous medical specialties were recognized in Europe during the Middle Ages, most historians continue to equate the term "woman medical practitioner," wherever they encounter it in medieval records, with "midwife." This common practice obscures the fact that, although women were not represented on all levels of medicine equally, they were represented in a variety of specialties throughout the broad medical community. A reliable study by Wickersheimer and Jacquart documents that, of 7,647 medical practitioners in France during the twelfth through fifteenth centuries, 121 were women; of these, only 44 were identified as midwives, while the rest practiced as physicians, surgeons, apothecaries, barbers, and other healers. While preserving terminological distinctions somewhat increases the quality of the information extracted from medieval documents concerning women medical practitioners, scholars must also reopen the whole question of why documentary evidence for women medical practitioners comprises such a tiny fraction of the evidence historians of medieval medicine usually present. Is this due to the limitations of the historical record, as has been claimed, or does it also result from the methods historians use? Granted, apart from medical licenses, the principal sources of information regarding medical practitioners available to researchers are wills, property transfers, court records, and similar documents, all of which typically underrepresent women because of restrictive medieval legal traditions. Nonetheless, the parameters researchers choose when they define their investigations may contribute to the problem. Studies focusing on the upper echelons of "learned" medicine, for example, tend to exclude healers on the legal and social fringes of medical practice, where most women would have been found. The advantages of broadening the scope of such studies is immediately apparent in Pelling and Webster's study of sixteenth-century London. Instead of focusing solely on officially recognized and licensed practitioners, the researchers defined a medical practitioner as "any individual whose occupation is basically concerned with the care of the sick." Using this definition, they found primary source information suggesting that there were 60 women medical practitioners in the city of London in 1560. Although this figure may be slightly exaggerated, the evidence contrasts strikingly with that of Gottfried, whose earlier survey identified only 28 women medical practitioners in all of England between 1330 and 1530. Finally, such studies provide only statistical information about the variety and prevalence of women's medical practice in medieval Europe. Future studies might also make profitable use of analyses developed in other areas of women's history as a basis for exploring the social context of women's medical practice. Information about economic rivalry in medicine, women's literacy, and the control of medical knowledge could add much to our growing understanding of women medical practitioners' role in medieval society.
199406_3-RC_4_25
[ "the effect of social change on the political and economic structure of medieval society", "the effect of social constraints on medieval women's access to a medical education", "the types of medical specialties that developed during the Middle Ages", "the reasons why medieval historians tend to equate the term \"woman medical practitioner\" with midwife", "the historical developments responsible for the medieval legal tradition's restrictions on women" ]
1
The passage suggests that a future study that would be more informative about medieval women medical practitioners might focus on which one of the following?
Although surveys of medieval legislation, guild organization, and terminology used to designate different medical practitioners have demonstrated that numerous medical specialties were recognized in Europe during the Middle Ages, most historians continue to equate the term "woman medical practitioner," wherever they encounter it in medieval records, with "midwife." This common practice obscures the fact that, although women were not represented on all levels of medicine equally, they were represented in a variety of specialties throughout the broad medical community. A reliable study by Wickersheimer and Jacquart documents that, of 7,647 medical practitioners in France during the twelfth through fifteenth centuries, 121 were women; of these, only 44 were identified as midwives, while the rest practiced as physicians, surgeons, apothecaries, barbers, and other healers. While preserving terminological distinctions somewhat increases the quality of the information extracted from medieval documents concerning women medical practitioners, scholars must also reopen the whole question of why documentary evidence for women medical practitioners comprises such a tiny fraction of the evidence historians of medieval medicine usually present. Is this due to the limitations of the historical record, as has been claimed, or does it also result from the methods historians use? Granted, apart from medical licenses, the principal sources of information regarding medical practitioners available to researchers are wills, property transfers, court records, and similar documents, all of which typically underrepresent women because of restrictive medieval legal traditions. Nonetheless, the parameters researchers choose when they define their investigations may contribute to the problem. Studies focusing on the upper echelons of "learned" medicine, for example, tend to exclude healers on the legal and social fringes of medical practice, where most women would have been found. The advantages of broadening the scope of such studies is immediately apparent in Pelling and Webster's study of sixteenth-century London. Instead of focusing solely on officially recognized and licensed practitioners, the researchers defined a medical practitioner as "any individual whose occupation is basically concerned with the care of the sick." Using this definition, they found primary source information suggesting that there were 60 women medical practitioners in the city of London in 1560. Although this figure may be slightly exaggerated, the evidence contrasts strikingly with that of Gottfried, whose earlier survey identified only 28 women medical practitioners in all of England between 1330 and 1530. Finally, such studies provide only statistical information about the variety and prevalence of women's medical practice in medieval Europe. Future studies might also make profitable use of analyses developed in other areas of women's history as a basis for exploring the social context of women's medical practice. Information about economic rivalry in medicine, women's literacy, and the control of medical knowledge could add much to our growing understanding of women medical practitioners' role in medieval society.
199406_3-RC_4_26
[ "demonstrate that numerous medical specialties were recognized in Western Europe during the Middle Ages", "demonstrate that women are often underrepresented in studies of medieval medical practitioners", "prove that midwives were officially recognized as members of the medical community during the Middle Ages", "prove that midwives were only a part of a larger community of women medical practitioners during the Middle Ages", "prove that the existence of midwives can be documented in Western Europe as early as the twelfth century" ]
3
The author refers to the study by Wickersheimer and Jacquart in order to
Although surveys of medieval legislation, guild organization, and terminology used to designate different medical practitioners have demonstrated that numerous medical specialties were recognized in Europe during the Middle Ages, most historians continue to equate the term "woman medical practitioner," wherever they encounter it in medieval records, with "midwife." This common practice obscures the fact that, although women were not represented on all levels of medicine equally, they were represented in a variety of specialties throughout the broad medical community. A reliable study by Wickersheimer and Jacquart documents that, of 7,647 medical practitioners in France during the twelfth through fifteenth centuries, 121 were women; of these, only 44 were identified as midwives, while the rest practiced as physicians, surgeons, apothecaries, barbers, and other healers. While preserving terminological distinctions somewhat increases the quality of the information extracted from medieval documents concerning women medical practitioners, scholars must also reopen the whole question of why documentary evidence for women medical practitioners comprises such a tiny fraction of the evidence historians of medieval medicine usually present. Is this due to the limitations of the historical record, as has been claimed, or does it also result from the methods historians use? Granted, apart from medical licenses, the principal sources of information regarding medical practitioners available to researchers are wills, property transfers, court records, and similar documents, all of which typically underrepresent women because of restrictive medieval legal traditions. Nonetheless, the parameters researchers choose when they define their investigations may contribute to the problem. Studies focusing on the upper echelons of "learned" medicine, for example, tend to exclude healers on the legal and social fringes of medical practice, where most women would have been found. The advantages of broadening the scope of such studies is immediately apparent in Pelling and Webster's study of sixteenth-century London. Instead of focusing solely on officially recognized and licensed practitioners, the researchers defined a medical practitioner as "any individual whose occupation is basically concerned with the care of the sick." Using this definition, they found primary source information suggesting that there were 60 women medical practitioners in the city of London in 1560. Although this figure may be slightly exaggerated, the evidence contrasts strikingly with that of Gottfried, whose earlier survey identified only 28 women medical practitioners in all of England between 1330 and 1530. Finally, such studies provide only statistical information about the variety and prevalence of women's medical practice in medieval Europe. Future studies might also make profitable use of analyses developed in other areas of women's history as a basis for exploring the social context of women's medical practice. Information about economic rivalry in medicine, women's literacy, and the control of medical knowledge could add much to our growing understanding of women medical practitioners' role in medieval society.
199406_3-RC_4_27
[ "describing new methodological approaches", "revising the definitions of certain concepts", "comparing two different analyses", "arguing in favor of changes in method", "chronicling certain historical developments" ]
3
In the passage, the author is primarily concerned with doing which one of the following?
Modern architecture has been criticized for emphasizing practical and technical issues at the expense of aesthetic concerns. The high-rise buildings constructed throughout the industrialized world in the 1960s and 1970s provide ample evidence that cost-efficiency and utility have become the overriding concerns of the modern architect. However, Otto Wagner's seminal text on modern architecture, first published in Germany in 1896, indicates that the failures of modern architecture cannot be blamed on the ideals of its founders. Wagner's Modern Architecture called for a new style based on modern technologies and modes of construction. He insisted that there could be no return to traditional, preindustrial models; only by accepting wholeheartedly the political and technological revolutions of the nineteenth century could the architect establish the forms appropriate to a modern, urban society. "All modern creations," Wagner wrote, "must correspond to the new materials and demands of the present…must illustrate our own better, democratic, self-confident, ideal nature," and must incorporate the new "colossal technical and scientific achievements" of the age. This would indeed seem to be the basis of a purely materialist definition of architecture, a prototype for the simplistic form-follows-function dogma that opponents have identified as the intellectual basis of modern architecture. But the picture was more complex, for Wagner was always careful to distinguish between art and engineering. Ultimately, he envisaged the architect developing the skills of the engineer without losing the powers of aesthetic judgment that Wagner felt were unique to the artist. "Since the engineer is seldom a born artist and the architect must learn as a rule to be an engineer, architects will in time succeed in extending their influence into the realm occupied by the engineers, so that legitimate aesthetic demands can be met in a satisfactory way." In this symbiotic relationship essential to Modernism, art was to exercise the controlling influence. No other prospect was imaginable for Wagner, who was firmly rooted as a designer and, indeed, as a teacher in the Classical tradition. The apparent inconsistency of a confessed Classicist advising against the mechanical imitation of historical models and arguing for new forms appropriate to the modern age created exactly the tension that made Wagner's writings and buildings so interesting. While he justified, for example, the choice of a circular ground plan for churches in terms of optimal sight-lines and the technology of the gasometer, the true inspiration was derived from the centralized churches of the Italian Renaissance. He acknowledged as a rationalist that there was no way back to the social and technological conditions that had produced the work of Michelangelo or Fischer von Erlach, but he recognized his emotional attachment to the great works of the Italian Renaissance and Austrian Baroque.
199410_3-RC_1_1
[ "Modern architecture has been criticized for emphasizing practical and technical issues and for failing to focus on aesthetic concerns.", "Critics have failed to take into account the technological innovations and aesthetic features that architects have incorporated into modern buildings.", "Wagner's Modern Architecture provides architects with a chronicle of the origins of modern architecture.", "Wagner's Modern Architecture indicates that the founders of modern architecture did not believe that practical issues should supersede the aesthetic concerns of the past.", "Wagner's seminal text, Modern Architecture, provides the intellectual basis for the purely materialistic definition of modern architecture." ]
3
Which one of the following best expresses the main idea of the passage?
Modern architecture has been criticized for emphasizing practical and technical issues at the expense of aesthetic concerns. The high-rise buildings constructed throughout the industrialized world in the 1960s and 1970s provide ample evidence that cost-efficiency and utility have become the overriding concerns of the modern architect. However, Otto Wagner's seminal text on modern architecture, first published in Germany in 1896, indicates that the failures of modern architecture cannot be blamed on the ideals of its founders. Wagner's Modern Architecture called for a new style based on modern technologies and modes of construction. He insisted that there could be no return to traditional, preindustrial models; only by accepting wholeheartedly the political and technological revolutions of the nineteenth century could the architect establish the forms appropriate to a modern, urban society. "All modern creations," Wagner wrote, "must correspond to the new materials and demands of the present…must illustrate our own better, democratic, self-confident, ideal nature," and must incorporate the new "colossal technical and scientific achievements" of the age. This would indeed seem to be the basis of a purely materialist definition of architecture, a prototype for the simplistic form-follows-function dogma that opponents have identified as the intellectual basis of modern architecture. But the picture was more complex, for Wagner was always careful to distinguish between art and engineering. Ultimately, he envisaged the architect developing the skills of the engineer without losing the powers of aesthetic judgment that Wagner felt were unique to the artist. "Since the engineer is seldom a born artist and the architect must learn as a rule to be an engineer, architects will in time succeed in extending their influence into the realm occupied by the engineers, so that legitimate aesthetic demands can be met in a satisfactory way." In this symbiotic relationship essential to Modernism, art was to exercise the controlling influence. No other prospect was imaginable for Wagner, who was firmly rooted as a designer and, indeed, as a teacher in the Classical tradition. The apparent inconsistency of a confessed Classicist advising against the mechanical imitation of historical models and arguing for new forms appropriate to the modern age created exactly the tension that made Wagner's writings and buildings so interesting. While he justified, for example, the choice of a circular ground plan for churches in terms of optimal sight-lines and the technology of the gasometer, the true inspiration was derived from the centralized churches of the Italian Renaissance. He acknowledged as a rationalist that there was no way back to the social and technological conditions that had produced the work of Michelangelo or Fischer von Erlach, but he recognized his emotional attachment to the great works of the Italian Renaissance and Austrian Baroque.
199410_3-RC_1_2
[ "The architect should make decisions about aesthetic issues and leave decisions about technical matters to the engineer.", "The engineer has often developed the powers of aesthetic judgment previously thought to be unique to the architect.", "The judgment of the engineer should be as important as the judgment of the architect when decisions are made about aesthetic issues.", "The technical judgment of the engineer should prevail over the aesthetic judgment of the architect in the design of modern buildings.", "The architect should acquire the knowledge of technical matters typically held by the engineer." ]
4
According to the passage, Wagner asserts which one of the following about the roles of architect and engineer?
Modern architecture has been criticized for emphasizing practical and technical issues at the expense of aesthetic concerns. The high-rise buildings constructed throughout the industrialized world in the 1960s and 1970s provide ample evidence that cost-efficiency and utility have become the overriding concerns of the modern architect. However, Otto Wagner's seminal text on modern architecture, first published in Germany in 1896, indicates that the failures of modern architecture cannot be blamed on the ideals of its founders. Wagner's Modern Architecture called for a new style based on modern technologies and modes of construction. He insisted that there could be no return to traditional, preindustrial models; only by accepting wholeheartedly the political and technological revolutions of the nineteenth century could the architect establish the forms appropriate to a modern, urban society. "All modern creations," Wagner wrote, "must correspond to the new materials and demands of the present…must illustrate our own better, democratic, self-confident, ideal nature," and must incorporate the new "colossal technical and scientific achievements" of the age. This would indeed seem to be the basis of a purely materialist definition of architecture, a prototype for the simplistic form-follows-function dogma that opponents have identified as the intellectual basis of modern architecture. But the picture was more complex, for Wagner was always careful to distinguish between art and engineering. Ultimately, he envisaged the architect developing the skills of the engineer without losing the powers of aesthetic judgment that Wagner felt were unique to the artist. "Since the engineer is seldom a born artist and the architect must learn as a rule to be an engineer, architects will in time succeed in extending their influence into the realm occupied by the engineers, so that legitimate aesthetic demands can be met in a satisfactory way." In this symbiotic relationship essential to Modernism, art was to exercise the controlling influence. No other prospect was imaginable for Wagner, who was firmly rooted as a designer and, indeed, as a teacher in the Classical tradition. The apparent inconsistency of a confessed Classicist advising against the mechanical imitation of historical models and arguing for new forms appropriate to the modern age created exactly the tension that made Wagner's writings and buildings so interesting. While he justified, for example, the choice of a circular ground plan for churches in terms of optimal sight-lines and the technology of the gasometer, the true inspiration was derived from the centralized churches of the Italian Renaissance. He acknowledged as a rationalist that there was no way back to the social and technological conditions that had produced the work of Michelangelo or Fischer von Erlach, but he recognized his emotional attachment to the great works of the Italian Renaissance and Austrian Baroque.
199410_3-RC_1_3
[ "The modern architect should avoid the mechanical imitation of the models of the Italian Renaissance and Austrian Baroque.", "The modern architect cannot design buildings appropriate to a modern, urban society and still retain emotional attachments to the forms of the Italian Renaissance and Austrian Baroque.", "The modern architect should possess knowledge of engineering as well as of the architecture of the past.", "The modern architect should not base designs on the technological conditions that underlay the designs of the models of the Italian Renaissance and Austrian Baroque.", "The designs of modern architects should reflect political ideals different from those reflected in the designs of classical architecture." ]
1
The passage suggests that Wagner would be LEAST likely to agree with which one of the following statements about classical architecture and the modern architect?
Modern architecture has been criticized for emphasizing practical and technical issues at the expense of aesthetic concerns. The high-rise buildings constructed throughout the industrialized world in the 1960s and 1970s provide ample evidence that cost-efficiency and utility have become the overriding concerns of the modern architect. However, Otto Wagner's seminal text on modern architecture, first published in Germany in 1896, indicates that the failures of modern architecture cannot be blamed on the ideals of its founders. Wagner's Modern Architecture called for a new style based on modern technologies and modes of construction. He insisted that there could be no return to traditional, preindustrial models; only by accepting wholeheartedly the political and technological revolutions of the nineteenth century could the architect establish the forms appropriate to a modern, urban society. "All modern creations," Wagner wrote, "must correspond to the new materials and demands of the present…must illustrate our own better, democratic, self-confident, ideal nature," and must incorporate the new "colossal technical and scientific achievements" of the age. This would indeed seem to be the basis of a purely materialist definition of architecture, a prototype for the simplistic form-follows-function dogma that opponents have identified as the intellectual basis of modern architecture. But the picture was more complex, for Wagner was always careful to distinguish between art and engineering. Ultimately, he envisaged the architect developing the skills of the engineer without losing the powers of aesthetic judgment that Wagner felt were unique to the artist. "Since the engineer is seldom a born artist and the architect must learn as a rule to be an engineer, architects will in time succeed in extending their influence into the realm occupied by the engineers, so that legitimate aesthetic demands can be met in a satisfactory way." In this symbiotic relationship essential to Modernism, art was to exercise the controlling influence. No other prospect was imaginable for Wagner, who was firmly rooted as a designer and, indeed, as a teacher in the Classical tradition. The apparent inconsistency of a confessed Classicist advising against the mechanical imitation of historical models and arguing for new forms appropriate to the modern age created exactly the tension that made Wagner's writings and buildings so interesting. While he justified, for example, the choice of a circular ground plan for churches in terms of optimal sight-lines and the technology of the gasometer, the true inspiration was derived from the centralized churches of the Italian Renaissance. He acknowledged as a rationalist that there was no way back to the social and technological conditions that had produced the work of Michelangelo or Fischer von Erlach, but he recognized his emotional attachment to the great works of the Italian Renaissance and Austrian Baroque.
199410_3-RC_1_4
[ "They represent the part of Wagner's work that has had the least influence on the architects who designed the high-rise buildings of the 1960s and 1970s.", "They describe the part of Wagner's work that is most often evoked by proponents of Wagner's ideas on art and technology.", "They do not adequately reflect the complexity of Wagner's ideas on the use of modern technology in architecture.", "They reflect Wagner's active participation in the political revolutions of the nineteenth century.", "They provide an overview of Wagner's ideas on the relationship between art and technology." ]
2
The passage suggests which one of the following about the quotations from Modern Architecture cited in the second paragraph?
Modern architecture has been criticized for emphasizing practical and technical issues at the expense of aesthetic concerns. The high-rise buildings constructed throughout the industrialized world in the 1960s and 1970s provide ample evidence that cost-efficiency and utility have become the overriding concerns of the modern architect. However, Otto Wagner's seminal text on modern architecture, first published in Germany in 1896, indicates that the failures of modern architecture cannot be blamed on the ideals of its founders. Wagner's Modern Architecture called for a new style based on modern technologies and modes of construction. He insisted that there could be no return to traditional, preindustrial models; only by accepting wholeheartedly the political and technological revolutions of the nineteenth century could the architect establish the forms appropriate to a modern, urban society. "All modern creations," Wagner wrote, "must correspond to the new materials and demands of the present…must illustrate our own better, democratic, self-confident, ideal nature," and must incorporate the new "colossal technical and scientific achievements" of the age. This would indeed seem to be the basis of a purely materialist definition of architecture, a prototype for the simplistic form-follows-function dogma that opponents have identified as the intellectual basis of modern architecture. But the picture was more complex, for Wagner was always careful to distinguish between art and engineering. Ultimately, he envisaged the architect developing the skills of the engineer without losing the powers of aesthetic judgment that Wagner felt were unique to the artist. "Since the engineer is seldom a born artist and the architect must learn as a rule to be an engineer, architects will in time succeed in extending their influence into the realm occupied by the engineers, so that legitimate aesthetic demands can be met in a satisfactory way." In this symbiotic relationship essential to Modernism, art was to exercise the controlling influence. No other prospect was imaginable for Wagner, who was firmly rooted as a designer and, indeed, as a teacher in the Classical tradition. The apparent inconsistency of a confessed Classicist advising against the mechanical imitation of historical models and arguing for new forms appropriate to the modern age created exactly the tension that made Wagner's writings and buildings so interesting. While he justified, for example, the choice of a circular ground plan for churches in terms of optimal sight-lines and the technology of the gasometer, the true inspiration was derived from the centralized churches of the Italian Renaissance. He acknowledged as a rationalist that there was no way back to the social and technological conditions that had produced the work of Michelangelo or Fischer von Erlach, but he recognized his emotional attachment to the great works of the Italian Renaissance and Austrian Baroque.
199410_3-RC_1_5
[ "Cost-efficiency, utility, and aesthetic demands are the primary concerns of the modern architect.", "Practical issues supersede aesthetic concerns in the design of many modern buildings.", "Cost-efficiency is more important to the modern architect than are other practical concerns.", "The design of many new buildings suggests that modern architects are still inspired by architectural forms of the past.", "Many modern architects use current technology to design modern buildings that are aesthetically pleasing." ]
1
The author of the passage states which one of the following about the concerns of modern architecture?
Modern architecture has been criticized for emphasizing practical and technical issues at the expense of aesthetic concerns. The high-rise buildings constructed throughout the industrialized world in the 1960s and 1970s provide ample evidence that cost-efficiency and utility have become the overriding concerns of the modern architect. However, Otto Wagner's seminal text on modern architecture, first published in Germany in 1896, indicates that the failures of modern architecture cannot be blamed on the ideals of its founders. Wagner's Modern Architecture called for a new style based on modern technologies and modes of construction. He insisted that there could be no return to traditional, preindustrial models; only by accepting wholeheartedly the political and technological revolutions of the nineteenth century could the architect establish the forms appropriate to a modern, urban society. "All modern creations," Wagner wrote, "must correspond to the new materials and demands of the present…must illustrate our own better, democratic, self-confident, ideal nature," and must incorporate the new "colossal technical and scientific achievements" of the age. This would indeed seem to be the basis of a purely materialist definition of architecture, a prototype for the simplistic form-follows-function dogma that opponents have identified as the intellectual basis of modern architecture. But the picture was more complex, for Wagner was always careful to distinguish between art and engineering. Ultimately, he envisaged the architect developing the skills of the engineer without losing the powers of aesthetic judgment that Wagner felt were unique to the artist. "Since the engineer is seldom a born artist and the architect must learn as a rule to be an engineer, architects will in time succeed in extending their influence into the realm occupied by the engineers, so that legitimate aesthetic demands can be met in a satisfactory way." In this symbiotic relationship essential to Modernism, art was to exercise the controlling influence. No other prospect was imaginable for Wagner, who was firmly rooted as a designer and, indeed, as a teacher in the Classical tradition. The apparent inconsistency of a confessed Classicist advising against the mechanical imitation of historical models and arguing for new forms appropriate to the modern age created exactly the tension that made Wagner's writings and buildings so interesting. While he justified, for example, the choice of a circular ground plan for churches in terms of optimal sight-lines and the technology of the gasometer, the true inspiration was derived from the centralized churches of the Italian Renaissance. He acknowledged as a rationalist that there was no way back to the social and technological conditions that had produced the work of Michelangelo or Fischer von Erlach, but he recognized his emotional attachment to the great works of the Italian Renaissance and Austrian Baroque.
199410_3-RC_1_6
[ "provide an example of the kinds of technological innovations Wagner introduced into modern architecture", "provide an example of Wagner's dismissal of historical forms from the Italian Renaissance", "provide an example of a modern building where technological issues were much less significant than aesthetic demands", "provide evidence of Wagner's tendency to imitate Italian Renaissance and Austrian Baroque models", "provide evidence of the tension between Wagner's commitment to modern technology and to the Classical tradition" ]
4
The author mentions Wagner's choice of a "circular ground plan for churches" (line 54) most likely in order to
Modern architecture has been criticized for emphasizing practical and technical issues at the expense of aesthetic concerns. The high-rise buildings constructed throughout the industrialized world in the 1960s and 1970s provide ample evidence that cost-efficiency and utility have become the overriding concerns of the modern architect. However, Otto Wagner's seminal text on modern architecture, first published in Germany in 1896, indicates that the failures of modern architecture cannot be blamed on the ideals of its founders. Wagner's Modern Architecture called for a new style based on modern technologies and modes of construction. He insisted that there could be no return to traditional, preindustrial models; only by accepting wholeheartedly the political and technological revolutions of the nineteenth century could the architect establish the forms appropriate to a modern, urban society. "All modern creations," Wagner wrote, "must correspond to the new materials and demands of the present…must illustrate our own better, democratic, self-confident, ideal nature," and must incorporate the new "colossal technical and scientific achievements" of the age. This would indeed seem to be the basis of a purely materialist definition of architecture, a prototype for the simplistic form-follows-function dogma that opponents have identified as the intellectual basis of modern architecture. But the picture was more complex, for Wagner was always careful to distinguish between art and engineering. Ultimately, he envisaged the architect developing the skills of the engineer without losing the powers of aesthetic judgment that Wagner felt were unique to the artist. "Since the engineer is seldom a born artist and the architect must learn as a rule to be an engineer, architects will in time succeed in extending their influence into the realm occupied by the engineers, so that legitimate aesthetic demands can be met in a satisfactory way." In this symbiotic relationship essential to Modernism, art was to exercise the controlling influence. No other prospect was imaginable for Wagner, who was firmly rooted as a designer and, indeed, as a teacher in the Classical tradition. The apparent inconsistency of a confessed Classicist advising against the mechanical imitation of historical models and arguing for new forms appropriate to the modern age created exactly the tension that made Wagner's writings and buildings so interesting. While he justified, for example, the choice of a circular ground plan for churches in terms of optimal sight-lines and the technology of the gasometer, the true inspiration was derived from the centralized churches of the Italian Renaissance. He acknowledged as a rationalist that there was no way back to the social and technological conditions that had produced the work of Michelangelo or Fischer von Erlach, but he recognized his emotional attachment to the great works of the Italian Renaissance and Austrian Baroque.
199410_3-RC_1_7
[ "summarizing the history of a debate", "explaining a traditional argument", "describing and evaluating a recent approach", "justifying a recent criticism by presenting new evidence", "supporting an assertion by discussing an important work" ]
4
The passage is primarily concerned with
In order to explain the socioeconomic achievement, in the face of disadvantages due to racial discrimination, of Chinese and Japanese immigrants to the United States and their descendants, sociologists have typically applied either culturally based or structurally based theories—but never both together. To use an economic metaphor, culturally based explanations assert the importance of the supply side of the labor market, emphasizing the qualities immigrant groups bring with them for competition in the United States labor market. Such explanations reflect a human-capital perspective in which status attainment is seen as a result of individuals' ability to generate resources. Structurally based explanations, on the other hand, examine the market condition of the immigrants' host society, particularly its discriminatory practices and their impact on the status attainment process of immigrant groups. In the economic metaphor, structural explanations assert the importance of the demand side of the labor market. In order to understand the socioeconomic mobility of Chinese and Japanese immigrants and their descendants, only an analysis of supply-side and demand-side factors together, in the context of historical events, will suffice. On the cultural or supply side, differences in immigration pattern and family formation resulted in different rates of socioeconomic achievement for Chinese and Japanese immigrants. For various reasons, Chinese immigrants remained sojourners and did not (except for urban merchants) establish families. They were also hampered by ethnic conflict in the labor market. Japanese immigrants, on the other hand, were less constrained, made the transition from sojourner to settler within the first two decades of immigration, and left low-wage labor to establish small businesses based on a household mode of production. Chinese sojourners without families were more vulnerable to demoralization, whereas Japanese immigrants faced societal hostility with the emotional resources provided by a stable family life. Once Chinese immigrants began to establish nuclear families and produce a second generation, instituting household production similar to that established by Japanese immigrants, their socioeconomic attainment soon paralleled that of Japanese immigrants and their descendants. On the structural or demand side, changes in institutional constraints, immigration laws, labor markets, and societal hostility were rooted in the dynamics of capitalist economic development. Early capitalist development generated a demand for low-wage labor that could not be fulfilled. Early Chinese and Japanese immigration was a response to this demand. In an advanced capitalist economy, the demand for immigrant labor is more differentiated: skilled professional and technical labor fills empty positions in the primary labor market and, with the traditional unskilled low-wage labor, creates two immigrant streams. The high levels of education attained by the descendants of Chinese and Japanese immigrants and their concentration in strategic states such as California paved the way for the movement of the second generation into the expanding primary labor market in the advanced capitalist economy that existed after the Second World War.
199410_3-RC_2_8
[ "The socioeconomic achievement of Chinese and Japanese immigrants and their descendants is best explained by a historical examination of the economic structures prevalent in the United States when such immigrant groups arrived.", "The socioeconomic achievement of Chinese and Japanese immigrants and their descendants is best explained by an examination of their cultural backgrounds, in particular their level of educational attainment.", "The socioeconomic achievement of Chinese and Japanese immigrants and their descendants has taken place in the context of a culturally based emphasis on the economic welfare of the nuclear family.", "Only the market structure of the capitalist economy of the United States in which supply has historically been regulated by demand can account for the socioeconomic achievement of Chinese and Japanese immigrants and their descendants.", "Only an analysis that combines an examination of the culture of Chinese and Japanese immigrant groups and the socioeconomic structure of the host country can adequately explain the socioeconomic achievement of Chinese and Japanese immigrants and their descendants." ]
4
Which one of the following best expresses the main idea of the passage?
In order to explain the socioeconomic achievement, in the face of disadvantages due to racial discrimination, of Chinese and Japanese immigrants to the United States and their descendants, sociologists have typically applied either culturally based or structurally based theories—but never both together. To use an economic metaphor, culturally based explanations assert the importance of the supply side of the labor market, emphasizing the qualities immigrant groups bring with them for competition in the United States labor market. Such explanations reflect a human-capital perspective in which status attainment is seen as a result of individuals' ability to generate resources. Structurally based explanations, on the other hand, examine the market condition of the immigrants' host society, particularly its discriminatory practices and their impact on the status attainment process of immigrant groups. In the economic metaphor, structural explanations assert the importance of the demand side of the labor market. In order to understand the socioeconomic mobility of Chinese and Japanese immigrants and their descendants, only an analysis of supply-side and demand-side factors together, in the context of historical events, will suffice. On the cultural or supply side, differences in immigration pattern and family formation resulted in different rates of socioeconomic achievement for Chinese and Japanese immigrants. For various reasons, Chinese immigrants remained sojourners and did not (except for urban merchants) establish families. They were also hampered by ethnic conflict in the labor market. Japanese immigrants, on the other hand, were less constrained, made the transition from sojourner to settler within the first two decades of immigration, and left low-wage labor to establish small businesses based on a household mode of production. Chinese sojourners without families were more vulnerable to demoralization, whereas Japanese immigrants faced societal hostility with the emotional resources provided by a stable family life. Once Chinese immigrants began to establish nuclear families and produce a second generation, instituting household production similar to that established by Japanese immigrants, their socioeconomic attainment soon paralleled that of Japanese immigrants and their descendants. On the structural or demand side, changes in institutional constraints, immigration laws, labor markets, and societal hostility were rooted in the dynamics of capitalist economic development. Early capitalist development generated a demand for low-wage labor that could not be fulfilled. Early Chinese and Japanese immigration was a response to this demand. In an advanced capitalist economy, the demand for immigrant labor is more differentiated: skilled professional and technical labor fills empty positions in the primary labor market and, with the traditional unskilled low-wage labor, creates two immigrant streams. The high levels of education attained by the descendants of Chinese and Japanese immigrants and their concentration in strategic states such as California paved the way for the movement of the second generation into the expanding primary labor market in the advanced capitalist economy that existed after the Second World War.
199410_3-RC_2_9
[ "concentration of small businesses in a given geographical area", "need for workers with varying degrees of skill", "high value placed by immigrants on work", "expansion of the primary labor market", "development of an advanced capitalist economy" ]
2
Which one of the following can best be described as a supply-side element in the labor market, as such elements are explained in the passage?
In order to explain the socioeconomic achievement, in the face of disadvantages due to racial discrimination, of Chinese and Japanese immigrants to the United States and their descendants, sociologists have typically applied either culturally based or structurally based theories—but never both together. To use an economic metaphor, culturally based explanations assert the importance of the supply side of the labor market, emphasizing the qualities immigrant groups bring with them for competition in the United States labor market. Such explanations reflect a human-capital perspective in which status attainment is seen as a result of individuals' ability to generate resources. Structurally based explanations, on the other hand, examine the market condition of the immigrants' host society, particularly its discriminatory practices and their impact on the status attainment process of immigrant groups. In the economic metaphor, structural explanations assert the importance of the demand side of the labor market. In order to understand the socioeconomic mobility of Chinese and Japanese immigrants and their descendants, only an analysis of supply-side and demand-side factors together, in the context of historical events, will suffice. On the cultural or supply side, differences in immigration pattern and family formation resulted in different rates of socioeconomic achievement for Chinese and Japanese immigrants. For various reasons, Chinese immigrants remained sojourners and did not (except for urban merchants) establish families. They were also hampered by ethnic conflict in the labor market. Japanese immigrants, on the other hand, were less constrained, made the transition from sojourner to settler within the first two decades of immigration, and left low-wage labor to establish small businesses based on a household mode of production. Chinese sojourners without families were more vulnerable to demoralization, whereas Japanese immigrants faced societal hostility with the emotional resources provided by a stable family life. Once Chinese immigrants began to establish nuclear families and produce a second generation, instituting household production similar to that established by Japanese immigrants, their socioeconomic attainment soon paralleled that of Japanese immigrants and their descendants. On the structural or demand side, changes in institutional constraints, immigration laws, labor markets, and societal hostility were rooted in the dynamics of capitalist economic development. Early capitalist development generated a demand for low-wage labor that could not be fulfilled. Early Chinese and Japanese immigration was a response to this demand. In an advanced capitalist economy, the demand for immigrant labor is more differentiated: skilled professional and technical labor fills empty positions in the primary labor market and, with the traditional unskilled low-wage labor, creates two immigrant streams. The high levels of education attained by the descendants of Chinese and Japanese immigrants and their concentration in strategic states such as California paved the way for the movement of the second generation into the expanding primary labor market in the advanced capitalist economy that existed after the Second World War.
199410_3-RC_2_10
[ "It demonstrates the effects of changes in human capital.", "It illustrates the operation of the primary labor market.", "It explains the nature of early Chinese and Japanese immigration.", "It characterizes the result of changing demand-side factors.", "It underscores an influence on the labor market." ]
3
Which one of the following best states the function of the author's mention of "two immigrant streams" (line 62)?
In order to explain the socioeconomic achievement, in the face of disadvantages due to racial discrimination, of Chinese and Japanese immigrants to the United States and their descendants, sociologists have typically applied either culturally based or structurally based theories—but never both together. To use an economic metaphor, culturally based explanations assert the importance of the supply side of the labor market, emphasizing the qualities immigrant groups bring with them for competition in the United States labor market. Such explanations reflect a human-capital perspective in which status attainment is seen as a result of individuals' ability to generate resources. Structurally based explanations, on the other hand, examine the market condition of the immigrants' host society, particularly its discriminatory practices and their impact on the status attainment process of immigrant groups. In the economic metaphor, structural explanations assert the importance of the demand side of the labor market. In order to understand the socioeconomic mobility of Chinese and Japanese immigrants and their descendants, only an analysis of supply-side and demand-side factors together, in the context of historical events, will suffice. On the cultural or supply side, differences in immigration pattern and family formation resulted in different rates of socioeconomic achievement for Chinese and Japanese immigrants. For various reasons, Chinese immigrants remained sojourners and did not (except for urban merchants) establish families. They were also hampered by ethnic conflict in the labor market. Japanese immigrants, on the other hand, were less constrained, made the transition from sojourner to settler within the first two decades of immigration, and left low-wage labor to establish small businesses based on a household mode of production. Chinese sojourners without families were more vulnerable to demoralization, whereas Japanese immigrants faced societal hostility with the emotional resources provided by a stable family life. Once Chinese immigrants began to establish nuclear families and produce a second generation, instituting household production similar to that established by Japanese immigrants, their socioeconomic attainment soon paralleled that of Japanese immigrants and their descendants. On the structural or demand side, changes in institutional constraints, immigration laws, labor markets, and societal hostility were rooted in the dynamics of capitalist economic development. Early capitalist development generated a demand for low-wage labor that could not be fulfilled. Early Chinese and Japanese immigration was a response to this demand. In an advanced capitalist economy, the demand for immigrant labor is more differentiated: skilled professional and technical labor fills empty positions in the primary labor market and, with the traditional unskilled low-wage labor, creates two immigrant streams. The high levels of education attained by the descendants of Chinese and Japanese immigrants and their concentration in strategic states such as California paved the way for the movement of the second generation into the expanding primary labor market in the advanced capitalist economy that existed after the Second World War.
199410_3-RC_2_11
[ "address the effects of the interaction of causal factors", "exclude the factor of a developing capitalist economy", "do not apply an economic metaphor", "emphasize the disadvantageous effects of racial discrimination", "focus on a single type of theoretical explanation" ]
4
It can be inferred that the author's analysis of the socioeconomic achievement of Chinese and Japanese immigrants and their descendants differs from that of most sociologists primarily in that most sociologists
In order to explain the socioeconomic achievement, in the face of disadvantages due to racial discrimination, of Chinese and Japanese immigrants to the United States and their descendants, sociologists have typically applied either culturally based or structurally based theories—but never both together. To use an economic metaphor, culturally based explanations assert the importance of the supply side of the labor market, emphasizing the qualities immigrant groups bring with them for competition in the United States labor market. Such explanations reflect a human-capital perspective in which status attainment is seen as a result of individuals' ability to generate resources. Structurally based explanations, on the other hand, examine the market condition of the immigrants' host society, particularly its discriminatory practices and their impact on the status attainment process of immigrant groups. In the economic metaphor, structural explanations assert the importance of the demand side of the labor market. In order to understand the socioeconomic mobility of Chinese and Japanese immigrants and their descendants, only an analysis of supply-side and demand-side factors together, in the context of historical events, will suffice. On the cultural or supply side, differences in immigration pattern and family formation resulted in different rates of socioeconomic achievement for Chinese and Japanese immigrants. For various reasons, Chinese immigrants remained sojourners and did not (except for urban merchants) establish families. They were also hampered by ethnic conflict in the labor market. Japanese immigrants, on the other hand, were less constrained, made the transition from sojourner to settler within the first two decades of immigration, and left low-wage labor to establish small businesses based on a household mode of production. Chinese sojourners without families were more vulnerable to demoralization, whereas Japanese immigrants faced societal hostility with the emotional resources provided by a stable family life. Once Chinese immigrants began to establish nuclear families and produce a second generation, instituting household production similar to that established by Japanese immigrants, their socioeconomic attainment soon paralleled that of Japanese immigrants and their descendants. On the structural or demand side, changes in institutional constraints, immigration laws, labor markets, and societal hostility were rooted in the dynamics of capitalist economic development. Early capitalist development generated a demand for low-wage labor that could not be fulfilled. Early Chinese and Japanese immigration was a response to this demand. In an advanced capitalist economy, the demand for immigrant labor is more differentiated: skilled professional and technical labor fills empty positions in the primary labor market and, with the traditional unskilled low-wage labor, creates two immigrant streams. The high levels of education attained by the descendants of Chinese and Japanese immigrants and their concentration in strategic states such as California paved the way for the movement of the second generation into the expanding primary labor market in the advanced capitalist economy that existed after the Second World War.
199410_3-RC_2_12
[ "initial status as sojourners", "slow accumulation of capital", "quick transition from laborer to manager", "rapid establishment of nuclear families", "rapid acquisition of technical skills" ]
0
It can be inferred that which one of the following was an element of the experience of both Chinese and Japanese immigrants in the United States?
In order to explain the socioeconomic achievement, in the face of disadvantages due to racial discrimination, of Chinese and Japanese immigrants to the United States and their descendants, sociologists have typically applied either culturally based or structurally based theories—but never both together. To use an economic metaphor, culturally based explanations assert the importance of the supply side of the labor market, emphasizing the qualities immigrant groups bring with them for competition in the United States labor market. Such explanations reflect a human-capital perspective in which status attainment is seen as a result of individuals' ability to generate resources. Structurally based explanations, on the other hand, examine the market condition of the immigrants' host society, particularly its discriminatory practices and their impact on the status attainment process of immigrant groups. In the economic metaphor, structural explanations assert the importance of the demand side of the labor market. In order to understand the socioeconomic mobility of Chinese and Japanese immigrants and their descendants, only an analysis of supply-side and demand-side factors together, in the context of historical events, will suffice. On the cultural or supply side, differences in immigration pattern and family formation resulted in different rates of socioeconomic achievement for Chinese and Japanese immigrants. For various reasons, Chinese immigrants remained sojourners and did not (except for urban merchants) establish families. They were also hampered by ethnic conflict in the labor market. Japanese immigrants, on the other hand, were less constrained, made the transition from sojourner to settler within the first two decades of immigration, and left low-wage labor to establish small businesses based on a household mode of production. Chinese sojourners without families were more vulnerable to demoralization, whereas Japanese immigrants faced societal hostility with the emotional resources provided by a stable family life. Once Chinese immigrants began to establish nuclear families and produce a second generation, instituting household production similar to that established by Japanese immigrants, their socioeconomic attainment soon paralleled that of Japanese immigrants and their descendants. On the structural or demand side, changes in institutional constraints, immigration laws, labor markets, and societal hostility were rooted in the dynamics of capitalist economic development. Early capitalist development generated a demand for low-wage labor that could not be fulfilled. Early Chinese and Japanese immigration was a response to this demand. In an advanced capitalist economy, the demand for immigrant labor is more differentiated: skilled professional and technical labor fills empty positions in the primary labor market and, with the traditional unskilled low-wage labor, creates two immigrant streams. The high levels of education attained by the descendants of Chinese and Japanese immigrants and their concentration in strategic states such as California paved the way for the movement of the second generation into the expanding primary labor market in the advanced capitalist economy that existed after the Second World War.
199410_3-RC_2_13
[ "advancing a synthesis of approaches to an issue", "challenging a tentative answer to a question", "evaluating the soundness of theories", "resolving the differences between schools of thought", "outlining the achievements of a group" ]
0
The author is primarily concerned with
Although the legal systems of England and the United States are superficially similar, they differ profoundly in their approaches to and uses of legal reasons: substantive reasons are more common than formal reasons in the United States, whereas in England the reverse is true. This distinction reflects a difference in the visions of law that prevail in the two countries. In England the law has traditionally been viewed as a system of rules; the United States favors a vision of law as an outward expression of the community's sense of right and justice. Substantive reasons, as applied to law, are based on moral, economic, political, and other considerations. These reasons are found both "in the law" and "outside the law," so to speak. Substantive reasons inform the content of a large part of the law: constitutions, statutes, contracts verdicts, and the like. Consider, for example, a statute providing that "no vehicles shall be taken into public parks." Suppose that no specific rationales or purposes were explicitly written into this statute, but that it was clear (from its legislative history) that the substantive purpose of the statute was to ensure quiet and safety in the park. Now suppose that a veterans' group mounts a World War II jeep (in running order but without a battery) as a war memorial on a concrete slab in the park, and charges are brought against its members. Most judges in the United States would find the defendants not guilty because what they did had no adverse effect on park quiet and safety. Formal reasons are different in that they frequently prevent substantive reasons from coming into play, even when substantive reasons are explicitly incorporated into the law at hand. For example, when a document fails to comply with stipulated requirements, the court may render the document legally ineffective. A will requiring written witness may be declared null and void and, therefore, unenforceable for the formal reason that the requirement was not observed. Once the legal rule—that a will is invalid for lack of proper witnessing—has been clearly established, and the legality of the rule is not in question, application of that rule precludes from consideration substantive arguments in favor of the will's validity or enforcement. Legal scholars in England and the United States have long bemused themselves with extreme examples of formal and substantive reasoning. On the one hand, formal reasoning in England has led to wooden interpretations of statutes and an unwillingness to develop the common law through judicial activism. On the other hand, freewheeling substantive reasoning in the United States has resulted in statutory interpretations so liberal that the texts of some statutes have been ignored altogether.
199410_3-RC_3_14
[ "an analysis of similarities and differences between the legal systems of England and the United States", "a reevaluation of two legal systems with the use of examples", "a contrast between the types of reasons embodied in the United States and English legal systems", "an explanation of how two distinct visions of the law shaped the development of legal reasoning", "a presentation of two types of legal reasons that shows the characteristics they have in common" ]
2
Which one of the following best describes the content of the passage as a whole?
Although the legal systems of England and the United States are superficially similar, they differ profoundly in their approaches to and uses of legal reasons: substantive reasons are more common than formal reasons in the United States, whereas in England the reverse is true. This distinction reflects a difference in the visions of law that prevail in the two countries. In England the law has traditionally been viewed as a system of rules; the United States favors a vision of law as an outward expression of the community's sense of right and justice. Substantive reasons, as applied to law, are based on moral, economic, political, and other considerations. These reasons are found both "in the law" and "outside the law," so to speak. Substantive reasons inform the content of a large part of the law: constitutions, statutes, contracts verdicts, and the like. Consider, for example, a statute providing that "no vehicles shall be taken into public parks." Suppose that no specific rationales or purposes were explicitly written into this statute, but that it was clear (from its legislative history) that the substantive purpose of the statute was to ensure quiet and safety in the park. Now suppose that a veterans' group mounts a World War II jeep (in running order but without a battery) as a war memorial on a concrete slab in the park, and charges are brought against its members. Most judges in the United States would find the defendants not guilty because what they did had no adverse effect on park quiet and safety. Formal reasons are different in that they frequently prevent substantive reasons from coming into play, even when substantive reasons are explicitly incorporated into the law at hand. For example, when a document fails to comply with stipulated requirements, the court may render the document legally ineffective. A will requiring written witness may be declared null and void and, therefore, unenforceable for the formal reason that the requirement was not observed. Once the legal rule—that a will is invalid for lack of proper witnessing—has been clearly established, and the legality of the rule is not in question, application of that rule precludes from consideration substantive arguments in favor of the will's validity or enforcement. Legal scholars in England and the United States have long bemused themselves with extreme examples of formal and substantive reasoning. On the one hand, formal reasoning in England has led to wooden interpretations of statutes and an unwillingness to develop the common law through judicial activism. On the other hand, freewheeling substantive reasoning in the United States has resulted in statutory interpretations so liberal that the texts of some statutes have been ignored altogether.
199410_3-RC_3_15
[ "not to do so would encourage others to act as the group did", "not to do so would be to violate the substantive reasons underlying the law", "the veterans failed to comply with the substantive purpose of the statute", "the veterans failed to demonstrate that their activities had no adverse effect on the public", "the veterans failed to comply with the stipulated requirements of the statute" ]
4
It can be inferred from the passage that English judges would be likely to find the veterans' group discussed in the second paragraph guilty of violating the statute because
Although the legal systems of England and the United States are superficially similar, they differ profoundly in their approaches to and uses of legal reasons: substantive reasons are more common than formal reasons in the United States, whereas in England the reverse is true. This distinction reflects a difference in the visions of law that prevail in the two countries. In England the law has traditionally been viewed as a system of rules; the United States favors a vision of law as an outward expression of the community's sense of right and justice. Substantive reasons, as applied to law, are based on moral, economic, political, and other considerations. These reasons are found both "in the law" and "outside the law," so to speak. Substantive reasons inform the content of a large part of the law: constitutions, statutes, contracts verdicts, and the like. Consider, for example, a statute providing that "no vehicles shall be taken into public parks." Suppose that no specific rationales or purposes were explicitly written into this statute, but that it was clear (from its legislative history) that the substantive purpose of the statute was to ensure quiet and safety in the park. Now suppose that a veterans' group mounts a World War II jeep (in running order but without a battery) as a war memorial on a concrete slab in the park, and charges are brought against its members. Most judges in the United States would find the defendants not guilty because what they did had no adverse effect on park quiet and safety. Formal reasons are different in that they frequently prevent substantive reasons from coming into play, even when substantive reasons are explicitly incorporated into the law at hand. For example, when a document fails to comply with stipulated requirements, the court may render the document legally ineffective. A will requiring written witness may be declared null and void and, therefore, unenforceable for the formal reason that the requirement was not observed. Once the legal rule—that a will is invalid for lack of proper witnessing—has been clearly established, and the legality of the rule is not in question, application of that rule precludes from consideration substantive arguments in favor of the will's validity or enforcement. Legal scholars in England and the United States have long bemused themselves with extreme examples of formal and substantive reasoning. On the one hand, formal reasoning in England has led to wooden interpretations of statutes and an unwillingness to develop the common law through judicial activism. On the other hand, freewheeling substantive reasoning in the United States has resulted in statutory interpretations so liberal that the texts of some statutes have been ignored altogether.
199410_3-RC_3_16
[ "The legal rule requiring that a will be witnessed in writing does not stipulate the format of the will.", "The legal rule requiring that a will be witnessed stipulates that the will must be witnessed in writing by two people.", "The legal ruling requiring that a will be witnessed in writing stipulates that the witnessing must be done in the presence of a judge.", "A judge rules that the law requires a will to be witnessed in writing regardless of extenuating circumstances.", "A judge rules that the law can be interpreted to allow for a verbal witness to a will in a case involving a medical emergency." ]
4
From the discussion of wills in the third paragraph it can be inferred that substantive arguments as to the validity of a will might be considered under which one of the following circumstances?
Although the legal systems of England and the United States are superficially similar, they differ profoundly in their approaches to and uses of legal reasons: substantive reasons are more common than formal reasons in the United States, whereas in England the reverse is true. This distinction reflects a difference in the visions of law that prevail in the two countries. In England the law has traditionally been viewed as a system of rules; the United States favors a vision of law as an outward expression of the community's sense of right and justice. Substantive reasons, as applied to law, are based on moral, economic, political, and other considerations. These reasons are found both "in the law" and "outside the law," so to speak. Substantive reasons inform the content of a large part of the law: constitutions, statutes, contracts verdicts, and the like. Consider, for example, a statute providing that "no vehicles shall be taken into public parks." Suppose that no specific rationales or purposes were explicitly written into this statute, but that it was clear (from its legislative history) that the substantive purpose of the statute was to ensure quiet and safety in the park. Now suppose that a veterans' group mounts a World War II jeep (in running order but without a battery) as a war memorial on a concrete slab in the park, and charges are brought against its members. Most judges in the United States would find the defendants not guilty because what they did had no adverse effect on park quiet and safety. Formal reasons are different in that they frequently prevent substantive reasons from coming into play, even when substantive reasons are explicitly incorporated into the law at hand. For example, when a document fails to comply with stipulated requirements, the court may render the document legally ineffective. A will requiring written witness may be declared null and void and, therefore, unenforceable for the formal reason that the requirement was not observed. Once the legal rule—that a will is invalid for lack of proper witnessing—has been clearly established, and the legality of the rule is not in question, application of that rule precludes from consideration substantive arguments in favor of the will's validity or enforcement. Legal scholars in England and the United States have long bemused themselves with extreme examples of formal and substantive reasoning. On the one hand, formal reasoning in England has led to wooden interpretations of statutes and an unwillingness to develop the common law through judicial activism. On the other hand, freewheeling substantive reasoning in the United States has resulted in statutory interpretations so liberal that the texts of some statutes have been ignored altogether.
199410_3-RC_3_17
[ "comparison and contrast", "generalization", "explication of terms", "a chronology of historical developments", "a hypothetical case" ]
3
The author of the passage makes use of all of the following in presenting the discussion of the English and the United States legal systems EXCEPT
Although the legal systems of England and the United States are superficially similar, they differ profoundly in their approaches to and uses of legal reasons: substantive reasons are more common than formal reasons in the United States, whereas in England the reverse is true. This distinction reflects a difference in the visions of law that prevail in the two countries. In England the law has traditionally been viewed as a system of rules; the United States favors a vision of law as an outward expression of the community's sense of right and justice. Substantive reasons, as applied to law, are based on moral, economic, political, and other considerations. These reasons are found both "in the law" and "outside the law," so to speak. Substantive reasons inform the content of a large part of the law: constitutions, statutes, contracts verdicts, and the like. Consider, for example, a statute providing that "no vehicles shall be taken into public parks." Suppose that no specific rationales or purposes were explicitly written into this statute, but that it was clear (from its legislative history) that the substantive purpose of the statute was to ensure quiet and safety in the park. Now suppose that a veterans' group mounts a World War II jeep (in running order but without a battery) as a war memorial on a concrete slab in the park, and charges are brought against its members. Most judges in the United States would find the defendants not guilty because what they did had no adverse effect on park quiet and safety. Formal reasons are different in that they frequently prevent substantive reasons from coming into play, even when substantive reasons are explicitly incorporated into the law at hand. For example, when a document fails to comply with stipulated requirements, the court may render the document legally ineffective. A will requiring written witness may be declared null and void and, therefore, unenforceable for the formal reason that the requirement was not observed. Once the legal rule—that a will is invalid for lack of proper witnessing—has been clearly established, and the legality of the rule is not in question, application of that rule precludes from consideration substantive arguments in favor of the will's validity or enforcement. Legal scholars in England and the United States have long bemused themselves with extreme examples of formal and substantive reasoning. On the one hand, formal reasoning in England has led to wooden interpretations of statutes and an unwillingness to develop the common law through judicial activism. On the other hand, freewheeling substantive reasoning in the United States has resulted in statutory interpretations so liberal that the texts of some statutes have been ignored altogether.
199410_3-RC_3_18
[ "It presents the consequences of extreme interpretations of the two types of legal reasons discussed by the author.", "It shows how legal scholars can incorrectly use extreme examples to support their views.", "It corrects inaccuracies in legal scholars' views of the nature of the two types of legal systems.", "It suggests how characterizations of the two types of legal reasons can become convoluted and inaccurate.", "It presents scholars' characterizations of both legal systems that are only partially correct." ]
0
Which one of the following best describes the function of the last paragraph of the passage?
Although the legal systems of England and the United States are superficially similar, they differ profoundly in their approaches to and uses of legal reasons: substantive reasons are more common than formal reasons in the United States, whereas in England the reverse is true. This distinction reflects a difference in the visions of law that prevail in the two countries. In England the law has traditionally been viewed as a system of rules; the United States favors a vision of law as an outward expression of the community's sense of right and justice. Substantive reasons, as applied to law, are based on moral, economic, political, and other considerations. These reasons are found both "in the law" and "outside the law," so to speak. Substantive reasons inform the content of a large part of the law: constitutions, statutes, contracts verdicts, and the like. Consider, for example, a statute providing that "no vehicles shall be taken into public parks." Suppose that no specific rationales or purposes were explicitly written into this statute, but that it was clear (from its legislative history) that the substantive purpose of the statute was to ensure quiet and safety in the park. Now suppose that a veterans' group mounts a World War II jeep (in running order but without a battery) as a war memorial on a concrete slab in the park, and charges are brought against its members. Most judges in the United States would find the defendants not guilty because what they did had no adverse effect on park quiet and safety. Formal reasons are different in that they frequently prevent substantive reasons from coming into play, even when substantive reasons are explicitly incorporated into the law at hand. For example, when a document fails to comply with stipulated requirements, the court may render the document legally ineffective. A will requiring written witness may be declared null and void and, therefore, unenforceable for the formal reason that the requirement was not observed. Once the legal rule—that a will is invalid for lack of proper witnessing—has been clearly established, and the legality of the rule is not in question, application of that rule precludes from consideration substantive arguments in favor of the will's validity or enforcement. Legal scholars in England and the United States have long bemused themselves with extreme examples of formal and substantive reasoning. On the one hand, formal reasoning in England has led to wooden interpretations of statutes and an unwillingness to develop the common law through judicial activism. On the other hand, freewheeling substantive reasoning in the United States has resulted in statutory interpretations so liberal that the texts of some statutes have been ignored altogether.
199410_3-RC_3_19
[ "Social conditions have changed to the extent that to continue to enforce the rule would be to decide contrary to present-day social norms.", "The composition of the legislature has changed to the extent that to enforce the rule would be contrary to the views of the majority in the present legislative assembly.", "The legality of the rule is in question and its enforcement is open to judicial interpretation.", "Individuals who have violated the legal rule argue that application of the rule would lead to unfair judicial interpretations.", "Superior court judges have consistently ruled in decisions regarding the interpretation of the legal rule." ]
2
The author of the passage suggests that in English law a substantive interpretation of a legal rule might be warranted under which one of the following circumstances?
Although the legal systems of England and the United States are superficially similar, they differ profoundly in their approaches to and uses of legal reasons: substantive reasons are more common than formal reasons in the United States, whereas in England the reverse is true. This distinction reflects a difference in the visions of law that prevail in the two countries. In England the law has traditionally been viewed as a system of rules; the United States favors a vision of law as an outward expression of the community's sense of right and justice. Substantive reasons, as applied to law, are based on moral, economic, political, and other considerations. These reasons are found both "in the law" and "outside the law," so to speak. Substantive reasons inform the content of a large part of the law: constitutions, statutes, contracts verdicts, and the like. Consider, for example, a statute providing that "no vehicles shall be taken into public parks." Suppose that no specific rationales or purposes were explicitly written into this statute, but that it was clear (from its legislative history) that the substantive purpose of the statute was to ensure quiet and safety in the park. Now suppose that a veterans' group mounts a World War II jeep (in running order but without a battery) as a war memorial on a concrete slab in the park, and charges are brought against its members. Most judges in the United States would find the defendants not guilty because what they did had no adverse effect on park quiet and safety. Formal reasons are different in that they frequently prevent substantive reasons from coming into play, even when substantive reasons are explicitly incorporated into the law at hand. For example, when a document fails to comply with stipulated requirements, the court may render the document legally ineffective. A will requiring written witness may be declared null and void and, therefore, unenforceable for the formal reason that the requirement was not observed. Once the legal rule—that a will is invalid for lack of proper witnessing—has been clearly established, and the legality of the rule is not in question, application of that rule precludes from consideration substantive arguments in favor of the will's validity or enforcement. Legal scholars in England and the United States have long bemused themselves with extreme examples of formal and substantive reasoning. On the one hand, formal reasoning in England has led to wooden interpretations of statutes and an unwillingness to develop the common law through judicial activism. On the other hand, freewheeling substantive reasoning in the United States has resulted in statutory interpretations so liberal that the texts of some statutes have been ignored altogether.
199410_3-RC_3_20
[ "They may be written into laws, but they may also exert an external influence on the law.", "They must be explicitly written into the law in order to be relevant to the application of the law.", "They are legal in nature and determine particular applications of most laws.", "They often provide judges with specific rationales for disregarding the laws of the land.", "They are peripheral to the law, whereas formal reasons are central to the law." ]
0
According to the passage, which one of the following statements about substantive reasons is true?
How does the brain know when carbohydrates have been or should be consumed? The answer to this question is not known, but one element in the explanation seems to be the neurotransmitter serotonin, one of a class of chemical mediators that may be released from a presynaptic neuron and that cause the transmission of a nerve impulse across a synapse to an adjacent postsynaptic neuron. In general, it has been found that drugs that selectively facilitate serotonin-mediated neurotransmission tend to cause weight loss, whereas drugs that block serotonin-mediated transmission often have the opposite effect: they often induce carbohydrate craving and consequent weight gain. Serotonin is a derivative of tryptophan, an amino acid that is normally present at low levels in the bloodstream. The rate of conversion is affected by the proportion of carbohydrates in an individual's diet: carbohydrates stimulate the secretion of insulin, which facilitates the uptake of most amino acids into peripheral tissues, such as muscles. Blood tryptophan levels, however, are unaffected by insulin, so the proportion of tryptophan in the blood relative to the other amino acids increases when carbohydrates are consumed. Since tryptophan competes with other amino acids for transport across the blood-brain barrier into the brain, insulin secretion indirectly speeds tryptophan's entry into the central nervous system where, in a special cluster of neurons, it is converted into serotonin. The level of serotonin in the brain in turn affects the amount of carbohydrate an individual chooses to eat. Rats that are allowed to choose among synthetic foods containing different proportions of carbohydrate and protein will normally alternate between foods containing mostly protein and those containing mostly carbohydrate. However, if rats are given drugs that enhance the effect of serotonin, the rats' carbohydrate intake is reduced. On the other hand, when rats are given drugs that interrupt serotonin-mediated neurotransmission, their brains fail to respond when carbohydrates are eaten, so the desire for them persists. In human beings a serotoninlike drug, d-fenfluramine (which releases serotonin into brain synapses and then prolongs its action by blocking its reabsorption into the presynaptic neuron), selectively suppresses carbohydrate snacking (and its associated weight gain) in people who crave carbohydrates. In contrast, drugs that block serotonin-mediated transmission or that interact with neurotransmitters other than serotonin have the opposite effect: they often induce carbohydrate craving and subsequent weight gain. People who crave carbohydrates report feeling refreshed and invigorated after eating a carbohydrate-rich meal (which would be expected to increase brain serotonin levels). In contrast, those who do not crave carbohydrates become sleepy following a high-carbohydrate meal. These findings suggest that serotonin has other effects that may be useful indicators of serotonin levels in human beings.
199410_3-RC_4_21
[ "The body's need for carbohydrates varies with the level of serotonin in the blood.", "The body's use of carbohydrates can be regulated by the administration of serotoninlike drugs.", "The role of serotonin in regulating the consumption of carbohydrates is similar in rats and in humans.", "The body's desire for carbohydrates can be influenced by serotonin or serotoninlike drugs.", "Tryptophan initiates a chain of events that regulates the body's use of carbohydrates." ]
3
Which one of the following best states the main idea of the passage?
How does the brain know when carbohydrates have been or should be consumed? The answer to this question is not known, but one element in the explanation seems to be the neurotransmitter serotonin, one of a class of chemical mediators that may be released from a presynaptic neuron and that cause the transmission of a nerve impulse across a synapse to an adjacent postsynaptic neuron. In general, it has been found that drugs that selectively facilitate serotonin-mediated neurotransmission tend to cause weight loss, whereas drugs that block serotonin-mediated transmission often have the opposite effect: they often induce carbohydrate craving and consequent weight gain. Serotonin is a derivative of tryptophan, an amino acid that is normally present at low levels in the bloodstream. The rate of conversion is affected by the proportion of carbohydrates in an individual's diet: carbohydrates stimulate the secretion of insulin, which facilitates the uptake of most amino acids into peripheral tissues, such as muscles. Blood tryptophan levels, however, are unaffected by insulin, so the proportion of tryptophan in the blood relative to the other amino acids increases when carbohydrates are consumed. Since tryptophan competes with other amino acids for transport across the blood-brain barrier into the brain, insulin secretion indirectly speeds tryptophan's entry into the central nervous system where, in a special cluster of neurons, it is converted into serotonin. The level of serotonin in the brain in turn affects the amount of carbohydrate an individual chooses to eat. Rats that are allowed to choose among synthetic foods containing different proportions of carbohydrate and protein will normally alternate between foods containing mostly protein and those containing mostly carbohydrate. However, if rats are given drugs that enhance the effect of serotonin, the rats' carbohydrate intake is reduced. On the other hand, when rats are given drugs that interrupt serotonin-mediated neurotransmission, their brains fail to respond when carbohydrates are eaten, so the desire for them persists. In human beings a serotoninlike drug, d-fenfluramine (which releases serotonin into brain synapses and then prolongs its action by blocking its reabsorption into the presynaptic neuron), selectively suppresses carbohydrate snacking (and its associated weight gain) in people who crave carbohydrates. In contrast, drugs that block serotonin-mediated transmission or that interact with neurotransmitters other than serotonin have the opposite effect: they often induce carbohydrate craving and subsequent weight gain. People who crave carbohydrates report feeling refreshed and invigorated after eating a carbohydrate-rich meal (which would be expected to increase brain serotonin levels). In contrast, those who do not crave carbohydrates become sleepy following a high-carbohydrate meal. These findings suggest that serotonin has other effects that may be useful indicators of serotonin levels in human beings.
199410_3-RC_4_22
[ "serotonin is produced from tryptophan", "carbohydrates are taken into the body", "carbohydrates stimulate the secretion of insulin", "insulin facilitates the uptake of amino acids into peripheral tissues", "tryptophan enters the bloodstream" ]
0
The term "rate" (line 17) refers to the rate at which
How does the brain know when carbohydrates have been or should be consumed? The answer to this question is not known, but one element in the explanation seems to be the neurotransmitter serotonin, one of a class of chemical mediators that may be released from a presynaptic neuron and that cause the transmission of a nerve impulse across a synapse to an adjacent postsynaptic neuron. In general, it has been found that drugs that selectively facilitate serotonin-mediated neurotransmission tend to cause weight loss, whereas drugs that block serotonin-mediated transmission often have the opposite effect: they often induce carbohydrate craving and consequent weight gain. Serotonin is a derivative of tryptophan, an amino acid that is normally present at low levels in the bloodstream. The rate of conversion is affected by the proportion of carbohydrates in an individual's diet: carbohydrates stimulate the secretion of insulin, which facilitates the uptake of most amino acids into peripheral tissues, such as muscles. Blood tryptophan levels, however, are unaffected by insulin, so the proportion of tryptophan in the blood relative to the other amino acids increases when carbohydrates are consumed. Since tryptophan competes with other amino acids for transport across the blood-brain barrier into the brain, insulin secretion indirectly speeds tryptophan's entry into the central nervous system where, in a special cluster of neurons, it is converted into serotonin. The level of serotonin in the brain in turn affects the amount of carbohydrate an individual chooses to eat. Rats that are allowed to choose among synthetic foods containing different proportions of carbohydrate and protein will normally alternate between foods containing mostly protein and those containing mostly carbohydrate. However, if rats are given drugs that enhance the effect of serotonin, the rats' carbohydrate intake is reduced. On the other hand, when rats are given drugs that interrupt serotonin-mediated neurotransmission, their brains fail to respond when carbohydrates are eaten, so the desire for them persists. In human beings a serotoninlike drug, d-fenfluramine (which releases serotonin into brain synapses and then prolongs its action by blocking its reabsorption into the presynaptic neuron), selectively suppresses carbohydrate snacking (and its associated weight gain) in people who crave carbohydrates. In contrast, drugs that block serotonin-mediated transmission or that interact with neurotransmitters other than serotonin have the opposite effect: they often induce carbohydrate craving and subsequent weight gain. People who crave carbohydrates report feeling refreshed and invigorated after eating a carbohydrate-rich meal (which would be expected to increase brain serotonin levels). In contrast, those who do not crave carbohydrates become sleepy following a high-carbohydrate meal. These findings suggest that serotonin has other effects that may be useful indicators of serotonin levels in human beings.
199410_3-RC_4_23
[ "the amount of insulin produced is too high", "the amount of serotonin in the brain is too low", "more tryptophan than usual crosses the blood-brain barrier", "neurotransmission by neurotransmitters other than serotonin is interrupted", "amino acids other than tryptophan are taken up by peripheral tissues" ]
1
It can be inferred that a person is likely to crave carbohydrates when
How does the brain know when carbohydrates have been or should be consumed? The answer to this question is not known, but one element in the explanation seems to be the neurotransmitter serotonin, one of a class of chemical mediators that may be released from a presynaptic neuron and that cause the transmission of a nerve impulse across a synapse to an adjacent postsynaptic neuron. In general, it has been found that drugs that selectively facilitate serotonin-mediated neurotransmission tend to cause weight loss, whereas drugs that block serotonin-mediated transmission often have the opposite effect: they often induce carbohydrate craving and consequent weight gain. Serotonin is a derivative of tryptophan, an amino acid that is normally present at low levels in the bloodstream. The rate of conversion is affected by the proportion of carbohydrates in an individual's diet: carbohydrates stimulate the secretion of insulin, which facilitates the uptake of most amino acids into peripheral tissues, such as muscles. Blood tryptophan levels, however, are unaffected by insulin, so the proportion of tryptophan in the blood relative to the other amino acids increases when carbohydrates are consumed. Since tryptophan competes with other amino acids for transport across the blood-brain barrier into the brain, insulin secretion indirectly speeds tryptophan's entry into the central nervous system where, in a special cluster of neurons, it is converted into serotonin. The level of serotonin in the brain in turn affects the amount of carbohydrate an individual chooses to eat. Rats that are allowed to choose among synthetic foods containing different proportions of carbohydrate and protein will normally alternate between foods containing mostly protein and those containing mostly carbohydrate. However, if rats are given drugs that enhance the effect of serotonin, the rats' carbohydrate intake is reduced. On the other hand, when rats are given drugs that interrupt serotonin-mediated neurotransmission, their brains fail to respond when carbohydrates are eaten, so the desire for them persists. In human beings a serotoninlike drug, d-fenfluramine (which releases serotonin into brain synapses and then prolongs its action by blocking its reabsorption into the presynaptic neuron), selectively suppresses carbohydrate snacking (and its associated weight gain) in people who crave carbohydrates. In contrast, drugs that block serotonin-mediated transmission or that interact with neurotransmitters other than serotonin have the opposite effect: they often induce carbohydrate craving and subsequent weight gain. People who crave carbohydrates report feeling refreshed and invigorated after eating a carbohydrate-rich meal (which would be expected to increase brain serotonin levels). In contrast, those who do not crave carbohydrates become sleepy following a high-carbohydrate meal. These findings suggest that serotonin has other effects that may be useful indicators of serotonin levels in human beings.
199410_3-RC_4_24
[ "Subjects would probably show a preference for carbohydrate-rich snacks rather than protein-rich snacks.", "Subjects would probably become sleepy after eating a carbohydrate-rich meal.", "Subjects would be more likely to lose weight than before they took the drug.", "Subjects' blood tryptophan levels would probably increase.", "Subjects' desire for both carbohydrates and proteins would increase." ]
0
The information in the passage indicates that if human beings were given a drug that inhibits the action of serotonin, which one of the following might be expected to occur?
How does the brain know when carbohydrates have been or should be consumed? The answer to this question is not known, but one element in the explanation seems to be the neurotransmitter serotonin, one of a class of chemical mediators that may be released from a presynaptic neuron and that cause the transmission of a nerve impulse across a synapse to an adjacent postsynaptic neuron. In general, it has been found that drugs that selectively facilitate serotonin-mediated neurotransmission tend to cause weight loss, whereas drugs that block serotonin-mediated transmission often have the opposite effect: they often induce carbohydrate craving and consequent weight gain. Serotonin is a derivative of tryptophan, an amino acid that is normally present at low levels in the bloodstream. The rate of conversion is affected by the proportion of carbohydrates in an individual's diet: carbohydrates stimulate the secretion of insulin, which facilitates the uptake of most amino acids into peripheral tissues, such as muscles. Blood tryptophan levels, however, are unaffected by insulin, so the proportion of tryptophan in the blood relative to the other amino acids increases when carbohydrates are consumed. Since tryptophan competes with other amino acids for transport across the blood-brain barrier into the brain, insulin secretion indirectly speeds tryptophan's entry into the central nervous system where, in a special cluster of neurons, it is converted into serotonin. The level of serotonin in the brain in turn affects the amount of carbohydrate an individual chooses to eat. Rats that are allowed to choose among synthetic foods containing different proportions of carbohydrate and protein will normally alternate between foods containing mostly protein and those containing mostly carbohydrate. However, if rats are given drugs that enhance the effect of serotonin, the rats' carbohydrate intake is reduced. On the other hand, when rats are given drugs that interrupt serotonin-mediated neurotransmission, their brains fail to respond when carbohydrates are eaten, so the desire for them persists. In human beings a serotoninlike drug, d-fenfluramine (which releases serotonin into brain synapses and then prolongs its action by blocking its reabsorption into the presynaptic neuron), selectively suppresses carbohydrate snacking (and its associated weight gain) in people who crave carbohydrates. In contrast, drugs that block serotonin-mediated transmission or that interact with neurotransmitters other than serotonin have the opposite effect: they often induce carbohydrate craving and subsequent weight gain. People who crave carbohydrates report feeling refreshed and invigorated after eating a carbohydrate-rich meal (which would be expected to increase brain serotonin levels). In contrast, those who do not crave carbohydrates become sleepy following a high-carbohydrate meal. These findings suggest that serotonin has other effects that may be useful indicators of serotonin levels in human beings.
199410_3-RC_4_25
[ "provide an overview of current research concerning the effect of serotonin on carbohydrate consumption", "contrast the role of tryptophan in the body with that of serotonin", "discuss the role of serotonin in the transmission of neural impulses", "explain how the brain knows that carbohydrates should be consumed", "establish a connection between carbohydrate intake and the production of serotonin" ]
4
The primary purpose of the second paragraph in the passage is to
How does the brain know when carbohydrates have been or should be consumed? The answer to this question is not known, but one element in the explanation seems to be the neurotransmitter serotonin, one of a class of chemical mediators that may be released from a presynaptic neuron and that cause the transmission of a nerve impulse across a synapse to an adjacent postsynaptic neuron. In general, it has been found that drugs that selectively facilitate serotonin-mediated neurotransmission tend to cause weight loss, whereas drugs that block serotonin-mediated transmission often have the opposite effect: they often induce carbohydrate craving and consequent weight gain. Serotonin is a derivative of tryptophan, an amino acid that is normally present at low levels in the bloodstream. The rate of conversion is affected by the proportion of carbohydrates in an individual's diet: carbohydrates stimulate the secretion of insulin, which facilitates the uptake of most amino acids into peripheral tissues, such as muscles. Blood tryptophan levels, however, are unaffected by insulin, so the proportion of tryptophan in the blood relative to the other amino acids increases when carbohydrates are consumed. Since tryptophan competes with other amino acids for transport across the blood-brain barrier into the brain, insulin secretion indirectly speeds tryptophan's entry into the central nervous system where, in a special cluster of neurons, it is converted into serotonin. The level of serotonin in the brain in turn affects the amount of carbohydrate an individual chooses to eat. Rats that are allowed to choose among synthetic foods containing different proportions of carbohydrate and protein will normally alternate between foods containing mostly protein and those containing mostly carbohydrate. However, if rats are given drugs that enhance the effect of serotonin, the rats' carbohydrate intake is reduced. On the other hand, when rats are given drugs that interrupt serotonin-mediated neurotransmission, their brains fail to respond when carbohydrates are eaten, so the desire for them persists. In human beings a serotoninlike drug, d-fenfluramine (which releases serotonin into brain synapses and then prolongs its action by blocking its reabsorption into the presynaptic neuron), selectively suppresses carbohydrate snacking (and its associated weight gain) in people who crave carbohydrates. In contrast, drugs that block serotonin-mediated transmission or that interact with neurotransmitters other than serotonin have the opposite effect: they often induce carbohydrate craving and subsequent weight gain. People who crave carbohydrates report feeling refreshed and invigorated after eating a carbohydrate-rich meal (which would be expected to increase brain serotonin levels). In contrast, those who do not crave carbohydrates become sleepy following a high-carbohydrate meal. These findings suggest that serotonin has other effects that may be useful indicators of serotonin levels in human beings.
199410_3-RC_4_26
[ "inclined to gain weight", "sleepy much of the time", "unlikely to crave carbohydrates", "unable to sleep as much as usual", "likely to secrete more insulin than usual" ]
2
It can be inferred that after a person has taken d-fenfluramine, he or she will probably be
How does the brain know when carbohydrates have been or should be consumed? The answer to this question is not known, but one element in the explanation seems to be the neurotransmitter serotonin, one of a class of chemical mediators that may be released from a presynaptic neuron and that cause the transmission of a nerve impulse across a synapse to an adjacent postsynaptic neuron. In general, it has been found that drugs that selectively facilitate serotonin-mediated neurotransmission tend to cause weight loss, whereas drugs that block serotonin-mediated transmission often have the opposite effect: they often induce carbohydrate craving and consequent weight gain. Serotonin is a derivative of tryptophan, an amino acid that is normally present at low levels in the bloodstream. The rate of conversion is affected by the proportion of carbohydrates in an individual's diet: carbohydrates stimulate the secretion of insulin, which facilitates the uptake of most amino acids into peripheral tissues, such as muscles. Blood tryptophan levels, however, are unaffected by insulin, so the proportion of tryptophan in the blood relative to the other amino acids increases when carbohydrates are consumed. Since tryptophan competes with other amino acids for transport across the blood-brain barrier into the brain, insulin secretion indirectly speeds tryptophan's entry into the central nervous system where, in a special cluster of neurons, it is converted into serotonin. The level of serotonin in the brain in turn affects the amount of carbohydrate an individual chooses to eat. Rats that are allowed to choose among synthetic foods containing different proportions of carbohydrate and protein will normally alternate between foods containing mostly protein and those containing mostly carbohydrate. However, if rats are given drugs that enhance the effect of serotonin, the rats' carbohydrate intake is reduced. On the other hand, when rats are given drugs that interrupt serotonin-mediated neurotransmission, their brains fail to respond when carbohydrates are eaten, so the desire for them persists. In human beings a serotoninlike drug, d-fenfluramine (which releases serotonin into brain synapses and then prolongs its action by blocking its reabsorption into the presynaptic neuron), selectively suppresses carbohydrate snacking (and its associated weight gain) in people who crave carbohydrates. In contrast, drugs that block serotonin-mediated transmission or that interact with neurotransmitters other than serotonin have the opposite effect: they often induce carbohydrate craving and subsequent weight gain. People who crave carbohydrates report feeling refreshed and invigorated after eating a carbohydrate-rich meal (which would be expected to increase brain serotonin levels). In contrast, those who do not crave carbohydrates become sleepy following a high-carbohydrate meal. These findings suggest that serotonin has other effects that may be useful indicators of serotonin levels in human beings.
199410_3-RC_4_27
[ "defend a point of view", "correct a misconception", "assess conflicting evidence", "suggest new directions for investigation", "provide information that helps explain a phenomenon" ]
4
The author's primary purpose is to
A major tenet of the neurosciences has been that all neurons (nerve cells) in the brains of vertebrate animals are formed early in development. An adult vertebrate, it was believed, must make do with a fixed number of neurons: those lost through disease or injury are not replaced, and adult learning takes place not through generation of new cells but through modification of connections among existing ones. However, new evidence for neurogenesis (the birth of new neurons) has come from the study of canary song. Young canaries and other songbirds learn to sing much as humans learn to speak, by imitating models provided by their elders. Several weeks after birth, a young bird produces its first rudimentary attempts at singing; over the next few months the song becomes more structured and stable, reaching a fully developed state by the time the bird approaches its first breeding season. But this repertoire of song is not permanently learned. After each breeding season, during late summer and fall, the bird loses mastery of its developed "vocabulary," and its song becomes as unstable as that of a juvenile bird. During the following winter and spring, however, the canary acquires new songs, and by the next breeding season it has developed an entirely new repertoire. Recent neurological research into this learning and relearning process has shown that the two most important regions of the canary's brain related to the learning of songs actually vary in size at different times of the year. In the spring, when the bird's song is highly developed and uniform, the regions are roughly twice as large as they are in the fall. Further experiments tracing individual nerve cells within these regions have shown that the number of neurons drops by about 38 percent after the breeding season, but by the following breeding season, new ones have been generated to replace them. A possible explanation for this continual replacement of nerve cells may have to do with the canary's relatively long life span and the requirements of flight. Its brain would have to be substantially larger and heavier than might be feasible for flying if it had to carry all the brain cells needed to process and retain all the information gathered over a lifetime. Although the idea of neurogenesis in the adult mammalian brain is still not generally accepted, these findings might help uncover a mechanism that would enable the human brain to repair itself through neurogenesis. Whether such replacement of neurons would disrupt complex learning processes or long-term memory is not known, but songbird research challenges scientists to identify the genes or hormones that orchestrate neurogenesis in the young human brain and to learn how to activate them in the adult brain.
199412_3-RC_1_1
[ "New evidence of neurogenesis in canaries challenges an established neurological theory concerning brain cells in vertebrates and suggests the possibility that human brains may repair themselves.", "The brains of canaries differ from the brains of other vertebrate animals in that the brains of adult canaries are able to generate neurons.", "Recent studies of neurogenesis in canaries, building on established theories of vertebrate neurology, provide important clues as to why researchers are not likely to discover neurogenesis in adult humans.", "Recent research into neurogenesis in canaries refutes a long-held belief about the limited supply of brain cells and provides new information about neurogenesis in the adult human brain.", "New information about neurogenesis in canaries challenges older hypotheses and clarifies the importance of the yearly cycle in learning processes and neurological replacement among vertebrates." ]
0
Which one of the following best expresses the main idea of the passage?
A major tenet of the neurosciences has been that all neurons (nerve cells) in the brains of vertebrate animals are formed early in development. An adult vertebrate, it was believed, must make do with a fixed number of neurons: those lost through disease or injury are not replaced, and adult learning takes place not through generation of new cells but through modification of connections among existing ones. However, new evidence for neurogenesis (the birth of new neurons) has come from the study of canary song. Young canaries and other songbirds learn to sing much as humans learn to speak, by imitating models provided by their elders. Several weeks after birth, a young bird produces its first rudimentary attempts at singing; over the next few months the song becomes more structured and stable, reaching a fully developed state by the time the bird approaches its first breeding season. But this repertoire of song is not permanently learned. After each breeding season, during late summer and fall, the bird loses mastery of its developed "vocabulary," and its song becomes as unstable as that of a juvenile bird. During the following winter and spring, however, the canary acquires new songs, and by the next breeding season it has developed an entirely new repertoire. Recent neurological research into this learning and relearning process has shown that the two most important regions of the canary's brain related to the learning of songs actually vary in size at different times of the year. In the spring, when the bird's song is highly developed and uniform, the regions are roughly twice as large as they are in the fall. Further experiments tracing individual nerve cells within these regions have shown that the number of neurons drops by about 38 percent after the breeding season, but by the following breeding season, new ones have been generated to replace them. A possible explanation for this continual replacement of nerve cells may have to do with the canary's relatively long life span and the requirements of flight. Its brain would have to be substantially larger and heavier than might be feasible for flying if it had to carry all the brain cells needed to process and retain all the information gathered over a lifetime. Although the idea of neurogenesis in the adult mammalian brain is still not generally accepted, these findings might help uncover a mechanism that would enable the human brain to repair itself through neurogenesis. Whether such replacement of neurons would disrupt complex learning processes or long-term memory is not known, but songbird research challenges scientists to identify the genes or hormones that orchestrate neurogenesis in the young human brain and to learn how to activate them in the adult brain.
199412_3-RC_1_2
[ "The canary's song repertoire takes on a fully structured and stable quality.", "A process of neurogenesis replaces the song-learning neurons that were lost during the preceding months.", "The canary begins to learn an entirely new repertoire of songs based on the models of other canaries.", "The regions in the canary's brain that are central to the learning of song decrease in size.", "The canary performs slightly modified versions of the songs it learned during the preceding breeding season." ]
3
According to the passage, which one of the following is true of the typical adult canary during the late summer and fall?
A major tenet of the neurosciences has been that all neurons (nerve cells) in the brains of vertebrate animals are formed early in development. An adult vertebrate, it was believed, must make do with a fixed number of neurons: those lost through disease or injury are not replaced, and adult learning takes place not through generation of new cells but through modification of connections among existing ones. However, new evidence for neurogenesis (the birth of new neurons) has come from the study of canary song. Young canaries and other songbirds learn to sing much as humans learn to speak, by imitating models provided by their elders. Several weeks after birth, a young bird produces its first rudimentary attempts at singing; over the next few months the song becomes more structured and stable, reaching a fully developed state by the time the bird approaches its first breeding season. But this repertoire of song is not permanently learned. After each breeding season, during late summer and fall, the bird loses mastery of its developed "vocabulary," and its song becomes as unstable as that of a juvenile bird. During the following winter and spring, however, the canary acquires new songs, and by the next breeding season it has developed an entirely new repertoire. Recent neurological research into this learning and relearning process has shown that the two most important regions of the canary's brain related to the learning of songs actually vary in size at different times of the year. In the spring, when the bird's song is highly developed and uniform, the regions are roughly twice as large as they are in the fall. Further experiments tracing individual nerve cells within these regions have shown that the number of neurons drops by about 38 percent after the breeding season, but by the following breeding season, new ones have been generated to replace them. A possible explanation for this continual replacement of nerve cells may have to do with the canary's relatively long life span and the requirements of flight. Its brain would have to be substantially larger and heavier than might be feasible for flying if it had to carry all the brain cells needed to process and retain all the information gathered over a lifetime. Although the idea of neurogenesis in the adult mammalian brain is still not generally accepted, these findings might help uncover a mechanism that would enable the human brain to repair itself through neurogenesis. Whether such replacement of neurons would disrupt complex learning processes or long-term memory is not known, but songbird research challenges scientists to identify the genes or hormones that orchestrate neurogenesis in the young human brain and to learn how to activate them in the adult brain.
199412_3-RC_1_3
[ "research on possible similarities between the neurological structures of humans and canaries", "studies that compare the ratio of brain weight to body weight in canaries to that in humans", "neurological research on the genes or hormones that activate neurogenesis in the brain of human infants", "studies about the ways in which long-term memory functions in the human brain", "research concerning the processes by which humans learn complicated tasks" ]
1
Information in the passage suggests that the author would most likely regard which one of the following as LEAST important in future research on neurogenesis in humans?
A major tenet of the neurosciences has been that all neurons (nerve cells) in the brains of vertebrate animals are formed early in development. An adult vertebrate, it was believed, must make do with a fixed number of neurons: those lost through disease or injury are not replaced, and adult learning takes place not through generation of new cells but through modification of connections among existing ones. However, new evidence for neurogenesis (the birth of new neurons) has come from the study of canary song. Young canaries and other songbirds learn to sing much as humans learn to speak, by imitating models provided by their elders. Several weeks after birth, a young bird produces its first rudimentary attempts at singing; over the next few months the song becomes more structured and stable, reaching a fully developed state by the time the bird approaches its first breeding season. But this repertoire of song is not permanently learned. After each breeding season, during late summer and fall, the bird loses mastery of its developed "vocabulary," and its song becomes as unstable as that of a juvenile bird. During the following winter and spring, however, the canary acquires new songs, and by the next breeding season it has developed an entirely new repertoire. Recent neurological research into this learning and relearning process has shown that the two most important regions of the canary's brain related to the learning of songs actually vary in size at different times of the year. In the spring, when the bird's song is highly developed and uniform, the regions are roughly twice as large as they are in the fall. Further experiments tracing individual nerve cells within these regions have shown that the number of neurons drops by about 38 percent after the breeding season, but by the following breeding season, new ones have been generated to replace them. A possible explanation for this continual replacement of nerve cells may have to do with the canary's relatively long life span and the requirements of flight. Its brain would have to be substantially larger and heavier than might be feasible for flying if it had to carry all the brain cells needed to process and retain all the information gathered over a lifetime. Although the idea of neurogenesis in the adult mammalian brain is still not generally accepted, these findings might help uncover a mechanism that would enable the human brain to repair itself through neurogenesis. Whether such replacement of neurons would disrupt complex learning processes or long-term memory is not known, but songbird research challenges scientists to identify the genes or hormones that orchestrate neurogenesis in the young human brain and to learn how to activate them in the adult brain.
199412_3-RC_1_4
[ "A number of songbird species related to the canary have a shorter life span than the canary and do not experience neurogenesis.", "The brain size of several types of airborne birds with life spans similar to those of canaries has been shown to vary according to a two-year cycle of neurogenesis.", "Several species of airborne birds similar to canaries in size are known to have brains that are substantially heavier than the canary's brain.", "Individual canaries that have larger-than-average repertoires of songs tend to have better developed muscles for flying.", "Individual canaries with smaller and lighter brains than the average tend to retain a smaller-than-average repertoire of songs." ]
2
Which one of the following, if true, would most seriously undermine the explanation proposed by the author in the third paragraph?
A major tenet of the neurosciences has been that all neurons (nerve cells) in the brains of vertebrate animals are formed early in development. An adult vertebrate, it was believed, must make do with a fixed number of neurons: those lost through disease or injury are not replaced, and adult learning takes place not through generation of new cells but through modification of connections among existing ones. However, new evidence for neurogenesis (the birth of new neurons) has come from the study of canary song. Young canaries and other songbirds learn to sing much as humans learn to speak, by imitating models provided by their elders. Several weeks after birth, a young bird produces its first rudimentary attempts at singing; over the next few months the song becomes more structured and stable, reaching a fully developed state by the time the bird approaches its first breeding season. But this repertoire of song is not permanently learned. After each breeding season, during late summer and fall, the bird loses mastery of its developed "vocabulary," and its song becomes as unstable as that of a juvenile bird. During the following winter and spring, however, the canary acquires new songs, and by the next breeding season it has developed an entirely new repertoire. Recent neurological research into this learning and relearning process has shown that the two most important regions of the canary's brain related to the learning of songs actually vary in size at different times of the year. In the spring, when the bird's song is highly developed and uniform, the regions are roughly twice as large as they are in the fall. Further experiments tracing individual nerve cells within these regions have shown that the number of neurons drops by about 38 percent after the breeding season, but by the following breeding season, new ones have been generated to replace them. A possible explanation for this continual replacement of nerve cells may have to do with the canary's relatively long life span and the requirements of flight. Its brain would have to be substantially larger and heavier than might be feasible for flying if it had to carry all the brain cells needed to process and retain all the information gathered over a lifetime. Although the idea of neurogenesis in the adult mammalian brain is still not generally accepted, these findings might help uncover a mechanism that would enable the human brain to repair itself through neurogenesis. Whether such replacement of neurons would disrupt complex learning processes or long-term memory is not known, but songbird research challenges scientists to identify the genes or hormones that orchestrate neurogenesis in the young human brain and to learn how to activate them in the adult brain.
199412_3-RC_1_5
[ "demonstrate the presence of a rudimentary grammatical structure in canary song", "point out a similarity between the patterned groupings of sounds in a canary's song and the syllabic structures of words", "stress the stability and uniformity of the canary's song throughout its lifetime", "suggest a similarity between the possession of a repertoire of words among humans and a repertoire of songs among canaries", "imply that the complexity of the canary's song repertoire is equal to that of human language" ]
3
The use of the word "vocabulary" (line 23) serves primarily to
A major tenet of the neurosciences has been that all neurons (nerve cells) in the brains of vertebrate animals are formed early in development. An adult vertebrate, it was believed, must make do with a fixed number of neurons: those lost through disease or injury are not replaced, and adult learning takes place not through generation of new cells but through modification of connections among existing ones. However, new evidence for neurogenesis (the birth of new neurons) has come from the study of canary song. Young canaries and other songbirds learn to sing much as humans learn to speak, by imitating models provided by their elders. Several weeks after birth, a young bird produces its first rudimentary attempts at singing; over the next few months the song becomes more structured and stable, reaching a fully developed state by the time the bird approaches its first breeding season. But this repertoire of song is not permanently learned. After each breeding season, during late summer and fall, the bird loses mastery of its developed "vocabulary," and its song becomes as unstable as that of a juvenile bird. During the following winter and spring, however, the canary acquires new songs, and by the next breeding season it has developed an entirely new repertoire. Recent neurological research into this learning and relearning process has shown that the two most important regions of the canary's brain related to the learning of songs actually vary in size at different times of the year. In the spring, when the bird's song is highly developed and uniform, the regions are roughly twice as large as they are in the fall. Further experiments tracing individual nerve cells within these regions have shown that the number of neurons drops by about 38 percent after the breeding season, but by the following breeding season, new ones have been generated to replace them. A possible explanation for this continual replacement of nerve cells may have to do with the canary's relatively long life span and the requirements of flight. Its brain would have to be substantially larger and heavier than might be feasible for flying if it had to carry all the brain cells needed to process and retain all the information gathered over a lifetime. Although the idea of neurogenesis in the adult mammalian brain is still not generally accepted, these findings might help uncover a mechanism that would enable the human brain to repair itself through neurogenesis. Whether such replacement of neurons would disrupt complex learning processes or long-term memory is not known, but songbird research challenges scientists to identify the genes or hormones that orchestrate neurogenesis in the young human brain and to learn how to activate them in the adult brain.
199412_3-RC_1_6
[ "the life span of the average canary", "the process by which canaries learn songs", "the frequency of canary breeding seasons", "the number of regions in the canary brain related to song learning", "the amount of time an average canary needs to learn a repertoire of songs" ]
0
According to the passage, which one of the following factors may help account for the occurrence of neurogenesis in canaries?
A major tenet of the neurosciences has been that all neurons (nerve cells) in the brains of vertebrate animals are formed early in development. An adult vertebrate, it was believed, must make do with a fixed number of neurons: those lost through disease or injury are not replaced, and adult learning takes place not through generation of new cells but through modification of connections among existing ones. However, new evidence for neurogenesis (the birth of new neurons) has come from the study of canary song. Young canaries and other songbirds learn to sing much as humans learn to speak, by imitating models provided by their elders. Several weeks after birth, a young bird produces its first rudimentary attempts at singing; over the next few months the song becomes more structured and stable, reaching a fully developed state by the time the bird approaches its first breeding season. But this repertoire of song is not permanently learned. After each breeding season, during late summer and fall, the bird loses mastery of its developed "vocabulary," and its song becomes as unstable as that of a juvenile bird. During the following winter and spring, however, the canary acquires new songs, and by the next breeding season it has developed an entirely new repertoire. Recent neurological research into this learning and relearning process has shown that the two most important regions of the canary's brain related to the learning of songs actually vary in size at different times of the year. In the spring, when the bird's song is highly developed and uniform, the regions are roughly twice as large as they are in the fall. Further experiments tracing individual nerve cells within these regions have shown that the number of neurons drops by about 38 percent after the breeding season, but by the following breeding season, new ones have been generated to replace them. A possible explanation for this continual replacement of nerve cells may have to do with the canary's relatively long life span and the requirements of flight. Its brain would have to be substantially larger and heavier than might be feasible for flying if it had to carry all the brain cells needed to process and retain all the information gathered over a lifetime. Although the idea of neurogenesis in the adult mammalian brain is still not generally accepted, these findings might help uncover a mechanism that would enable the human brain to repair itself through neurogenesis. Whether such replacement of neurons would disrupt complex learning processes or long-term memory is not known, but songbird research challenges scientists to identify the genes or hormones that orchestrate neurogenesis in the young human brain and to learn how to activate them in the adult brain.
199412_3-RC_1_7
[ "A theory is presented, analyzed, and modified, and a justification for the modification is offered.", "Research results are advanced and reconciled with results from other studies, and a shared principle is described.", "Research results are presented, further details are provided, and a hypothesis is offered to explain the results.", "Research findings are described, their implications are explained, and an application to a related field is proposed.", "Research results are reported, their significance is clarified, and they are reconciled with previously established neurological tenets." ]
2
Which one of the following best describes the organization of the third paragraph?
A major tenet of the neurosciences has been that all neurons (nerve cells) in the brains of vertebrate animals are formed early in development. An adult vertebrate, it was believed, must make do with a fixed number of neurons: those lost through disease or injury are not replaced, and adult learning takes place not through generation of new cells but through modification of connections among existing ones. However, new evidence for neurogenesis (the birth of new neurons) has come from the study of canary song. Young canaries and other songbirds learn to sing much as humans learn to speak, by imitating models provided by their elders. Several weeks after birth, a young bird produces its first rudimentary attempts at singing; over the next few months the song becomes more structured and stable, reaching a fully developed state by the time the bird approaches its first breeding season. But this repertoire of song is not permanently learned. After each breeding season, during late summer and fall, the bird loses mastery of its developed "vocabulary," and its song becomes as unstable as that of a juvenile bird. During the following winter and spring, however, the canary acquires new songs, and by the next breeding season it has developed an entirely new repertoire. Recent neurological research into this learning and relearning process has shown that the two most important regions of the canary's brain related to the learning of songs actually vary in size at different times of the year. In the spring, when the bird's song is highly developed and uniform, the regions are roughly twice as large as they are in the fall. Further experiments tracing individual nerve cells within these regions have shown that the number of neurons drops by about 38 percent after the breeding season, but by the following breeding season, new ones have been generated to replace them. A possible explanation for this continual replacement of nerve cells may have to do with the canary's relatively long life span and the requirements of flight. Its brain would have to be substantially larger and heavier than might be feasible for flying if it had to carry all the brain cells needed to process and retain all the information gathered over a lifetime. Although the idea of neurogenesis in the adult mammalian brain is still not generally accepted, these findings might help uncover a mechanism that would enable the human brain to repair itself through neurogenesis. Whether such replacement of neurons would disrupt complex learning processes or long-term memory is not known, but songbird research challenges scientists to identify the genes or hormones that orchestrate neurogenesis in the young human brain and to learn how to activate them in the adult brain.
199412_3-RC_1_8
[ "exhaustive", "progressive", "incomplete", "antiquated", "incorrect" ]
2
It can be inferred from the passage that the author would most likely describe the current understanding of neurogenesis as
For too many years scholars of African American history focused on the harm done by slaveholders and by the institution of slavery, rather than on what Africans in the United States were able to accomplish despite the effects of that institution. In Myne Owne Ground, T. H. Breen and Stephen Innes contribute significantly to a recent, welcome shift from a white-centered to a black-centered inquiry into the role of African Americans in the American colonial period. Breen and Innes focus not on slaves, but on a small group of freed indentured servants in Northampton County (in the Chesapeake Bay region of Virginia) who, according to the authors, maintained their freedom, secured property, and interacted with persons of different races and economic standing from 1620 through the 1670s. African Americans living on the Chesapeake were to some extent disadvantaged, say Breen and Innes, but this did not preclude the attainment of status roughly equal to that of certain white planters of the area. Continuously acting within black social networks, and forming economic relationships with white planters, local Native Americans, indentured servants, and white settlers outside the gentry class, the free African Americans of Northampton County held their own in the rough-hewn world of Chesapeake Bay. The authors emphasize that in this early period, when the percentage of African Americans in any given Chesapeake county was still no more than 10 percent of the population, very little was predetermined so far as racial status or race relations were concerned. By schooling themselves in the local legal process and by working prodigiously on the land, African Americans acquired property, established families, and warded off contentious white neighbors. Breen and Innes do acknowledge that political power on the Chesapeake was asymmetrically distributed among black and white residents. However, they underemphasize much evidence that customary law, only gradually embodied in statutory law, was closing in on free African Americans well before the 1670s: during the 1660s, when the proportion of African Americans in Virginia increased dramatically, Virginia tightened a law regulating interracial relations (1662) and enacted a statute prohibiting baptism from altering slave status (1667). Anthony Johnson, a leader in the community of free African Americans in the Chesapeake Bay region, sold the land he had cultivated for more than twenty years and moved north with his family around 1665, an action that the authors attribute to a search for "fresh, more productive land." But the answer to why the Johnsons left that area where they had labored so long may lie in their realization that their white neighbors were already beginning the transition from a largely white indentured labor force to reliance on a largely black slave labor force, and that the institution of slavery was threatening their descendants' chances for freedom and success in Virginia.
199412_3-RC_2_9
[ "Their emphases have been on statutory law rather than on customary law.", "They have ignored specific historical situations and personages in favor of broad interpretations.", "They have focused on the least eventful periods in African American history.", "They have underemphasized the economic system that was the basis of the institution of slavery.", "They have failed to focus to a sufficient extent on the achievements of African Americans." ]
4
The author of the passage objects to many scholarly studies of African American history for which one of the following reasons?
For too many years scholars of African American history focused on the harm done by slaveholders and by the institution of slavery, rather than on what Africans in the United States were able to accomplish despite the effects of that institution. In Myne Owne Ground, T. H. Breen and Stephen Innes contribute significantly to a recent, welcome shift from a white-centered to a black-centered inquiry into the role of African Americans in the American colonial period. Breen and Innes focus not on slaves, but on a small group of freed indentured servants in Northampton County (in the Chesapeake Bay region of Virginia) who, according to the authors, maintained their freedom, secured property, and interacted with persons of different races and economic standing from 1620 through the 1670s. African Americans living on the Chesapeake were to some extent disadvantaged, say Breen and Innes, but this did not preclude the attainment of status roughly equal to that of certain white planters of the area. Continuously acting within black social networks, and forming economic relationships with white planters, local Native Americans, indentured servants, and white settlers outside the gentry class, the free African Americans of Northampton County held their own in the rough-hewn world of Chesapeake Bay. The authors emphasize that in this early period, when the percentage of African Americans in any given Chesapeake county was still no more than 10 percent of the population, very little was predetermined so far as racial status or race relations were concerned. By schooling themselves in the local legal process and by working prodigiously on the land, African Americans acquired property, established families, and warded off contentious white neighbors. Breen and Innes do acknowledge that political power on the Chesapeake was asymmetrically distributed among black and white residents. However, they underemphasize much evidence that customary law, only gradually embodied in statutory law, was closing in on free African Americans well before the 1670s: during the 1660s, when the proportion of African Americans in Virginia increased dramatically, Virginia tightened a law regulating interracial relations (1662) and enacted a statute prohibiting baptism from altering slave status (1667). Anthony Johnson, a leader in the community of free African Americans in the Chesapeake Bay region, sold the land he had cultivated for more than twenty years and moved north with his family around 1665, an action that the authors attribute to a search for "fresh, more productive land." But the answer to why the Johnsons left that area where they had labored so long may lie in their realization that their white neighbors were already beginning the transition from a largely white indentured labor force to reliance on a largely black slave labor force, and that the institution of slavery was threatening their descendants' chances for freedom and success in Virginia.
199412_3-RC_2_10
[ "The laws affecting black citizens were embodied in statutes much more gradually than were laws affecting white citizens.", "As the percentage of black citizens in the population grew, the legal restrictions placed on them also increased.", "Because of discriminatory laws, black farmers suffered more economic setbacks than did white farmers.", "Because of legal constraints on hiring indentured servants, black farmers faced a chronic labor shortage on their farms.", "The adherence to customary law was more rigid in regions with relatively large numbers of free black citizens." ]
1
Which one of the following can be inferred from the passage concerning the relationship between the African American population and the law in the Chesapeake Bay region of Virginia between 1650 and 1670?
For too many years scholars of African American history focused on the harm done by slaveholders and by the institution of slavery, rather than on what Africans in the United States were able to accomplish despite the effects of that institution. In Myne Owne Ground, T. H. Breen and Stephen Innes contribute significantly to a recent, welcome shift from a white-centered to a black-centered inquiry into the role of African Americans in the American colonial period. Breen and Innes focus not on slaves, but on a small group of freed indentured servants in Northampton County (in the Chesapeake Bay region of Virginia) who, according to the authors, maintained their freedom, secured property, and interacted with persons of different races and economic standing from 1620 through the 1670s. African Americans living on the Chesapeake were to some extent disadvantaged, say Breen and Innes, but this did not preclude the attainment of status roughly equal to that of certain white planters of the area. Continuously acting within black social networks, and forming economic relationships with white planters, local Native Americans, indentured servants, and white settlers outside the gentry class, the free African Americans of Northampton County held their own in the rough-hewn world of Chesapeake Bay. The authors emphasize that in this early period, when the percentage of African Americans in any given Chesapeake county was still no more than 10 percent of the population, very little was predetermined so far as racial status or race relations were concerned. By schooling themselves in the local legal process and by working prodigiously on the land, African Americans acquired property, established families, and warded off contentious white neighbors. Breen and Innes do acknowledge that political power on the Chesapeake was asymmetrically distributed among black and white residents. However, they underemphasize much evidence that customary law, only gradually embodied in statutory law, was closing in on free African Americans well before the 1670s: during the 1660s, when the proportion of African Americans in Virginia increased dramatically, Virginia tightened a law regulating interracial relations (1662) and enacted a statute prohibiting baptism from altering slave status (1667). Anthony Johnson, a leader in the community of free African Americans in the Chesapeake Bay region, sold the land he had cultivated for more than twenty years and moved north with his family around 1665, an action that the authors attribute to a search for "fresh, more productive land." But the answer to why the Johnsons left that area where they had labored so long may lie in their realization that their white neighbors were already beginning the transition from a largely white indentured labor force to reliance on a largely black slave labor force, and that the institution of slavery was threatening their descendants' chances for freedom and success in Virginia.
199412_3-RC_2_11
[ "provide a specific example of the potential shortcomings of Breen and Innes's interpretation of historical events", "provide a specific example of relevant data overlooked by Breen and Innes in their discussion of historical events", "provide a specific example of data that Breen and Innes might profitably have used in proving their thesis", "argue that the standard interpretation of historical events is superior to Breen and Innes's revisionist interpretation", "argue that a new historiographical method is needed to provide a full and coherent reading of historical events" ]
0
The author of the passage most probably refers to Anthony Johnson and his family in order to
For too many years scholars of African American history focused on the harm done by slaveholders and by the institution of slavery, rather than on what Africans in the United States were able to accomplish despite the effects of that institution. In Myne Owne Ground, T. H. Breen and Stephen Innes contribute significantly to a recent, welcome shift from a white-centered to a black-centered inquiry into the role of African Americans in the American colonial period. Breen and Innes focus not on slaves, but on a small group of freed indentured servants in Northampton County (in the Chesapeake Bay region of Virginia) who, according to the authors, maintained their freedom, secured property, and interacted with persons of different races and economic standing from 1620 through the 1670s. African Americans living on the Chesapeake were to some extent disadvantaged, say Breen and Innes, but this did not preclude the attainment of status roughly equal to that of certain white planters of the area. Continuously acting within black social networks, and forming economic relationships with white planters, local Native Americans, indentured servants, and white settlers outside the gentry class, the free African Americans of Northampton County held their own in the rough-hewn world of Chesapeake Bay. The authors emphasize that in this early period, when the percentage of African Americans in any given Chesapeake county was still no more than 10 percent of the population, very little was predetermined so far as racial status or race relations were concerned. By schooling themselves in the local legal process and by working prodigiously on the land, African Americans acquired property, established families, and warded off contentious white neighbors. Breen and Innes do acknowledge that political power on the Chesapeake was asymmetrically distributed among black and white residents. However, they underemphasize much evidence that customary law, only gradually embodied in statutory law, was closing in on free African Americans well before the 1670s: during the 1660s, when the proportion of African Americans in Virginia increased dramatically, Virginia tightened a law regulating interracial relations (1662) and enacted a statute prohibiting baptism from altering slave status (1667). Anthony Johnson, a leader in the community of free African Americans in the Chesapeake Bay region, sold the land he had cultivated for more than twenty years and moved north with his family around 1665, an action that the authors attribute to a search for "fresh, more productive land." But the answer to why the Johnsons left that area where they had labored so long may lie in their realization that their white neighbors were already beginning the transition from a largely white indentured labor force to reliance on a largely black slave labor force, and that the institution of slavery was threatening their descendants' chances for freedom and success in Virginia.
199412_3-RC_2_12
[ "condescending dismissal", "wholehearted acceptance", "contentious challenge", "qualified approval", "sincere puzzlement" ]
3
The attitude of the author of the passage toward Breen and Innes's study can best be described as one of
For too many years scholars of African American history focused on the harm done by slaveholders and by the institution of slavery, rather than on what Africans in the United States were able to accomplish despite the effects of that institution. In Myne Owne Ground, T. H. Breen and Stephen Innes contribute significantly to a recent, welcome shift from a white-centered to a black-centered inquiry into the role of African Americans in the American colonial period. Breen and Innes focus not on slaves, but on a small group of freed indentured servants in Northampton County (in the Chesapeake Bay region of Virginia) who, according to the authors, maintained their freedom, secured property, and interacted with persons of different races and economic standing from 1620 through the 1670s. African Americans living on the Chesapeake were to some extent disadvantaged, say Breen and Innes, but this did not preclude the attainment of status roughly equal to that of certain white planters of the area. Continuously acting within black social networks, and forming economic relationships with white planters, local Native Americans, indentured servants, and white settlers outside the gentry class, the free African Americans of Northampton County held their own in the rough-hewn world of Chesapeake Bay. The authors emphasize that in this early period, when the percentage of African Americans in any given Chesapeake county was still no more than 10 percent of the population, very little was predetermined so far as racial status or race relations were concerned. By schooling themselves in the local legal process and by working prodigiously on the land, African Americans acquired property, established families, and warded off contentious white neighbors. Breen and Innes do acknowledge that political power on the Chesapeake was asymmetrically distributed among black and white residents. However, they underemphasize much evidence that customary law, only gradually embodied in statutory law, was closing in on free African Americans well before the 1670s: during the 1660s, when the proportion of African Americans in Virginia increased dramatically, Virginia tightened a law regulating interracial relations (1662) and enacted a statute prohibiting baptism from altering slave status (1667). Anthony Johnson, a leader in the community of free African Americans in the Chesapeake Bay region, sold the land he had cultivated for more than twenty years and moved north with his family around 1665, an action that the authors attribute to a search for "fresh, more productive land." But the answer to why the Johnsons left that area where they had labored so long may lie in their realization that their white neighbors were already beginning the transition from a largely white indentured labor force to reliance on a largely black slave labor force, and that the institution of slavery was threatening their descendants' chances for freedom and success in Virginia.
199412_3-RC_2_13
[ "summarize previous interpretations", "advocate a new approach", "propose and then illustrate a thesis", "present and evaluate an interpretation", "describe a historical event" ]
3
The primary purpose of the passage is to
Late-nineteenth-century books about the French artist Watteau (1684–1721) betray a curious blind spot: more than any single artist before or since, Watteau provided his age with an influential image of itself, and nineteenth-century writers accepted this image as genuine. This was largely due to the enterprise of Watteau's friends who, soon after his death, organized the printing of engraved reproductions of the great bulk of his work—both his paintings and his drawings—so that Watteau's total artistic output became and continued to be more accessible than that of any other artist until the twentieth-century advent of art monographs illustrated with photographs. These engravings presented aristocratic (and would-be aristocratic) eighteenth-century French society with an image of itself that was highly acceptable and widely imitated by other artists, however little relationship that image bore to reality. By 1884, the bicentenary of Watteau's birth, it was standard practice for biographers to refer to him as "the personification of the witty and amiable eighteenth century." In fact, Watteau saw little enough of that "witty and amiable" century for which so much nostalgia was generally felt between about 1870 and 1920, a period during which enthusiasm for the artist reached its peak. The eighteenth century's first decades, the period of his artistic activity, were fairly calamitous ones. During his short life, France was almost continually at war: his native region was overrun with foreign troops, and Paris was threatened by siege and by a rampaging army rabble. The dreadful winter of 1709, the year of Watteau's first Paris successes, was marked by military defeat and a disastrous famine. Most of Watteau's nineteenth-century admirers simply ignored the grim background of the works they found so lyrical and charming. Those who took the inconvenient historical facts into consideration did so only in order to refute the widely held deterministic view that the content and style of an artist's work were absolutely dictated by heredity and environment. (For Watteau admirers, such determinism was unthinkable: the artist was born in a Flemish town only six years after it first became part of France, yet Watteau was quintessentially French. As one patriotic French biographer put it, "In Dresden, Potsdam, and Berlin I have never come across a Watteau without feeling refreshed by a breath of native air." ) Even such writers, however, persisted in according Watteau's canvases a privileged status as representative "personifications" of the eighteenth century. The discrepancy between historical fact and artistic vision, useful in refuting the extreme deterministic position, merely forced these writers to seek a new formula that allowed them to preserve the desired identity between image and reality, this time a rather suspiciously psychic one: Watteau did not record the society he knew, but rather "foresaw" a society that developed shortly after his death.
199412_3-RC_3_14
[ "A particular phenomenon is discussed, the reasons that it is atypical are put forward, and these reasons are evaluated and refined.", "An assumption is made, results deriving from it are compared with what is known to be true, and the assumption is finally rejected as counterfactual.", "A point of view is described, one hypothesis accounting for it is introduced and rejected, and a better hypothesis is offered for consideration.", "A general characterization is offered, examples supporting it are introduced, and its special applicability to a particular group is asserted.", "A particular viewpoint is explained, its shortcomings are discussed, and its persistence in the face of these is noted." ]
4
Which one of the following best describes the overall organization of the passage?
Late-nineteenth-century books about the French artist Watteau (1684–1721) betray a curious blind spot: more than any single artist before or since, Watteau provided his age with an influential image of itself, and nineteenth-century writers accepted this image as genuine. This was largely due to the enterprise of Watteau's friends who, soon after his death, organized the printing of engraved reproductions of the great bulk of his work—both his paintings and his drawings—so that Watteau's total artistic output became and continued to be more accessible than that of any other artist until the twentieth-century advent of art monographs illustrated with photographs. These engravings presented aristocratic (and would-be aristocratic) eighteenth-century French society with an image of itself that was highly acceptable and widely imitated by other artists, however little relationship that image bore to reality. By 1884, the bicentenary of Watteau's birth, it was standard practice for biographers to refer to him as "the personification of the witty and amiable eighteenth century." In fact, Watteau saw little enough of that "witty and amiable" century for which so much nostalgia was generally felt between about 1870 and 1920, a period during which enthusiasm for the artist reached its peak. The eighteenth century's first decades, the period of his artistic activity, were fairly calamitous ones. During his short life, France was almost continually at war: his native region was overrun with foreign troops, and Paris was threatened by siege and by a rampaging army rabble. The dreadful winter of 1709, the year of Watteau's first Paris successes, was marked by military defeat and a disastrous famine. Most of Watteau's nineteenth-century admirers simply ignored the grim background of the works they found so lyrical and charming. Those who took the inconvenient historical facts into consideration did so only in order to refute the widely held deterministic view that the content and style of an artist's work were absolutely dictated by heredity and environment. (For Watteau admirers, such determinism was unthinkable: the artist was born in a Flemish town only six years after it first became part of France, yet Watteau was quintessentially French. As one patriotic French biographer put it, "In Dresden, Potsdam, and Berlin I have never come across a Watteau without feeling refreshed by a breath of native air." ) Even such writers, however, persisted in according Watteau's canvases a privileged status as representative "personifications" of the eighteenth century. The discrepancy between historical fact and artistic vision, useful in refuting the extreme deterministic position, merely forced these writers to seek a new formula that allowed them to preserve the desired identity between image and reality, this time a rather suspiciously psychic one: Watteau did not record the society he knew, but rather "foresaw" a society that developed shortly after his death.
199412_3-RC_3_15
[ "what they saw as Watteau's typical eighteenth-century talent for transcending reality through art", "their opposition to the determinism that dominated late-nineteenth-century French thought", "a lack of access to historical source material concerning the early eighteenth century in France", "the nature of the image conveyed by the works of Watteau and his many imitators", "their political bias in favor of aristocratic regimes and societies" ]
3
The passage suggests that late-nineteenth-century biographers of Watteau considered the eighteenth century to be "witty and amiable" in large part because of
Late-nineteenth-century books about the French artist Watteau (1684–1721) betray a curious blind spot: more than any single artist before or since, Watteau provided his age with an influential image of itself, and nineteenth-century writers accepted this image as genuine. This was largely due to the enterprise of Watteau's friends who, soon after his death, organized the printing of engraved reproductions of the great bulk of his work—both his paintings and his drawings—so that Watteau's total artistic output became and continued to be more accessible than that of any other artist until the twentieth-century advent of art monographs illustrated with photographs. These engravings presented aristocratic (and would-be aristocratic) eighteenth-century French society with an image of itself that was highly acceptable and widely imitated by other artists, however little relationship that image bore to reality. By 1884, the bicentenary of Watteau's birth, it was standard practice for biographers to refer to him as "the personification of the witty and amiable eighteenth century." In fact, Watteau saw little enough of that "witty and amiable" century for which so much nostalgia was generally felt between about 1870 and 1920, a period during which enthusiasm for the artist reached its peak. The eighteenth century's first decades, the period of his artistic activity, were fairly calamitous ones. During his short life, France was almost continually at war: his native region was overrun with foreign troops, and Paris was threatened by siege and by a rampaging army rabble. The dreadful winter of 1709, the year of Watteau's first Paris successes, was marked by military defeat and a disastrous famine. Most of Watteau's nineteenth-century admirers simply ignored the grim background of the works they found so lyrical and charming. Those who took the inconvenient historical facts into consideration did so only in order to refute the widely held deterministic view that the content and style of an artist's work were absolutely dictated by heredity and environment. (For Watteau admirers, such determinism was unthinkable: the artist was born in a Flemish town only six years after it first became part of France, yet Watteau was quintessentially French. As one patriotic French biographer put it, "In Dresden, Potsdam, and Berlin I have never come across a Watteau without feeling refreshed by a breath of native air." ) Even such writers, however, persisted in according Watteau's canvases a privileged status as representative "personifications" of the eighteenth century. The discrepancy between historical fact and artistic vision, useful in refuting the extreme deterministic position, merely forced these writers to seek a new formula that allowed them to preserve the desired identity between image and reality, this time a rather suspiciously psychic one: Watteau did not record the society he knew, but rather "foresaw" a society that developed shortly after his death.
199412_3-RC_3_16
[ "If such explanations were widely accepted, too many people who would otherwise have admired Watteau would cease to appreciate Watteau's works.", "If such explanations were adopted, they would make it difficult for Watteau admirers to explain why Watteau's works were purchased and admired by foreigners.", "If such explanations were correct, many artists who, like Watteau, considered themselves French would have to be excluded from histories of French art.", "If such simple explanations were offered, other more complex arguments concerning what made Watteau's works especially charming would go unexplored.", "If such explanations were true, Watteau's works would reflect a \"Flemish\" sensibility rather than the especially \"French\" one these admirers saw in them." ]
4
According to the passage, explanations of artistic production based on determinism were unthinkable to Watteau admirers for which one of the following reasons?
Late-nineteenth-century books about the French artist Watteau (1684–1721) betray a curious blind spot: more than any single artist before or since, Watteau provided his age with an influential image of itself, and nineteenth-century writers accepted this image as genuine. This was largely due to the enterprise of Watteau's friends who, soon after his death, organized the printing of engraved reproductions of the great bulk of his work—both his paintings and his drawings—so that Watteau's total artistic output became and continued to be more accessible than that of any other artist until the twentieth-century advent of art monographs illustrated with photographs. These engravings presented aristocratic (and would-be aristocratic) eighteenth-century French society with an image of itself that was highly acceptable and widely imitated by other artists, however little relationship that image bore to reality. By 1884, the bicentenary of Watteau's birth, it was standard practice for biographers to refer to him as "the personification of the witty and amiable eighteenth century." In fact, Watteau saw little enough of that "witty and amiable" century for which so much nostalgia was generally felt between about 1870 and 1920, a period during which enthusiasm for the artist reached its peak. The eighteenth century's first decades, the period of his artistic activity, were fairly calamitous ones. During his short life, France was almost continually at war: his native region was overrun with foreign troops, and Paris was threatened by siege and by a rampaging army rabble. The dreadful winter of 1709, the year of Watteau's first Paris successes, was marked by military defeat and a disastrous famine. Most of Watteau's nineteenth-century admirers simply ignored the grim background of the works they found so lyrical and charming. Those who took the inconvenient historical facts into consideration did so only in order to refute the widely held deterministic view that the content and style of an artist's work were absolutely dictated by heredity and environment. (For Watteau admirers, such determinism was unthinkable: the artist was born in a Flemish town only six years after it first became part of France, yet Watteau was quintessentially French. As one patriotic French biographer put it, "In Dresden, Potsdam, and Berlin I have never come across a Watteau without feeling refreshed by a breath of native air." ) Even such writers, however, persisted in according Watteau's canvases a privileged status as representative "personifications" of the eighteenth century. The discrepancy between historical fact and artistic vision, useful in refuting the extreme deterministic position, merely forced these writers to seek a new formula that allowed them to preserve the desired identity between image and reality, this time a rather suspiciously psychic one: Watteau did not record the society he knew, but rather "foresaw" a society that developed shortly after his death.
199412_3-RC_3_17
[ "some biographers' persistent inability to appreciate what the author considers a particularly admirable quality", "certain writers' surprising lack of awareness of what the author considers an obvious discrepancy", "some writers' willful refusal to evaluate properly what the author considers a valuable source of information about the past", "an inexplicable tendency on the part of some writers to undervalue an artist whom the author considers extremely influential", "a marked bias in favor of a certain painter and a concomitant prejudice against contemporaries the author considers equally talented" ]
1
The phrase "curious blind spot" (lines 2–3) can best be interpreted as referring to which one of the following?
Late-nineteenth-century books about the French artist Watteau (1684–1721) betray a curious blind spot: more than any single artist before or since, Watteau provided his age with an influential image of itself, and nineteenth-century writers accepted this image as genuine. This was largely due to the enterprise of Watteau's friends who, soon after his death, organized the printing of engraved reproductions of the great bulk of his work—both his paintings and his drawings—so that Watteau's total artistic output became and continued to be more accessible than that of any other artist until the twentieth-century advent of art monographs illustrated with photographs. These engravings presented aristocratic (and would-be aristocratic) eighteenth-century French society with an image of itself that was highly acceptable and widely imitated by other artists, however little relationship that image bore to reality. By 1884, the bicentenary of Watteau's birth, it was standard practice for biographers to refer to him as "the personification of the witty and amiable eighteenth century." In fact, Watteau saw little enough of that "witty and amiable" century for which so much nostalgia was generally felt between about 1870 and 1920, a period during which enthusiasm for the artist reached its peak. The eighteenth century's first decades, the period of his artistic activity, were fairly calamitous ones. During his short life, France was almost continually at war: his native region was overrun with foreign troops, and Paris was threatened by siege and by a rampaging army rabble. The dreadful winter of 1709, the year of Watteau's first Paris successes, was marked by military defeat and a disastrous famine. Most of Watteau's nineteenth-century admirers simply ignored the grim background of the works they found so lyrical and charming. Those who took the inconvenient historical facts into consideration did so only in order to refute the widely held deterministic view that the content and style of an artist's work were absolutely dictated by heredity and environment. (For Watteau admirers, such determinism was unthinkable: the artist was born in a Flemish town only six years after it first became part of France, yet Watteau was quintessentially French. As one patriotic French biographer put it, "In Dresden, Potsdam, and Berlin I have never come across a Watteau without feeling refreshed by a breath of native air." ) Even such writers, however, persisted in according Watteau's canvases a privileged status as representative "personifications" of the eighteenth century. The discrepancy between historical fact and artistic vision, useful in refuting the extreme deterministic position, merely forced these writers to seek a new formula that allowed them to preserve the desired identity between image and reality, this time a rather suspiciously psychic one: Watteau did not record the society he knew, but rather "foresaw" a society that developed shortly after his death.
199412_3-RC_3_18
[ "Unlike most late-nineteenth-century Watteau admirers, the author appreciates the importance of Watteau's artistic accomplishment.", "The author finds Watteau's works to be much less lyrical and charming than did most late-nineteenth-century admirers of the works.", "In contrast to most late-nineteenth-century Watteau admirers, the author finds it misleading to see Watteau's works as accurately reflecting social reality.", "The author is much more willing to entertain deterministic explanations of the origins of Watteau's works than were most late-nineteenth-century Watteau admirers.", "Unlike most late-nineteenth-century admirers of Watteau, the author considers it impossible for any work of art to personify or represent a particular historical period." ]
2
It can be inferred from the passage that the author's view of Watteau's works differs most significantly from that of most late-nineteenth-century Watteau admirers in which one of the following ways?
Late-nineteenth-century books about the French artist Watteau (1684–1721) betray a curious blind spot: more than any single artist before or since, Watteau provided his age with an influential image of itself, and nineteenth-century writers accepted this image as genuine. This was largely due to the enterprise of Watteau's friends who, soon after his death, organized the printing of engraved reproductions of the great bulk of his work—both his paintings and his drawings—so that Watteau's total artistic output became and continued to be more accessible than that of any other artist until the twentieth-century advent of art monographs illustrated with photographs. These engravings presented aristocratic (and would-be aristocratic) eighteenth-century French society with an image of itself that was highly acceptable and widely imitated by other artists, however little relationship that image bore to reality. By 1884, the bicentenary of Watteau's birth, it was standard practice for biographers to refer to him as "the personification of the witty and amiable eighteenth century." In fact, Watteau saw little enough of that "witty and amiable" century for which so much nostalgia was generally felt between about 1870 and 1920, a period during which enthusiasm for the artist reached its peak. The eighteenth century's first decades, the period of his artistic activity, were fairly calamitous ones. During his short life, France was almost continually at war: his native region was overrun with foreign troops, and Paris was threatened by siege and by a rampaging army rabble. The dreadful winter of 1709, the year of Watteau's first Paris successes, was marked by military defeat and a disastrous famine. Most of Watteau's nineteenth-century admirers simply ignored the grim background of the works they found so lyrical and charming. Those who took the inconvenient historical facts into consideration did so only in order to refute the widely held deterministic view that the content and style of an artist's work were absolutely dictated by heredity and environment. (For Watteau admirers, such determinism was unthinkable: the artist was born in a Flemish town only six years after it first became part of France, yet Watteau was quintessentially French. As one patriotic French biographer put it, "In Dresden, Potsdam, and Berlin I have never come across a Watteau without feeling refreshed by a breath of native air." ) Even such writers, however, persisted in according Watteau's canvases a privileged status as representative "personifications" of the eighteenth century. The discrepancy between historical fact and artistic vision, useful in refuting the extreme deterministic position, merely forced these writers to seek a new formula that allowed them to preserve the desired identity between image and reality, this time a rather suspiciously psychic one: Watteau did not record the society he knew, but rather "foresaw" a society that developed shortly after his death.
199412_3-RC_3_19
[ "widespread social upheaval caused by war", "a pervasive sense of nostalgia for an idealized past", "increased domination of public affairs by a powerful aristocracy", "rapid adoption by the middle classes of aristocratic manners and life-styles", "a need to reconcile the French self-image with French social realities" ]
0
The author asserts that during the period of Watteau's artistic activity French society was experiencing which one of the following?
Late-nineteenth-century books about the French artist Watteau (1684–1721) betray a curious blind spot: more than any single artist before or since, Watteau provided his age with an influential image of itself, and nineteenth-century writers accepted this image as genuine. This was largely due to the enterprise of Watteau's friends who, soon after his death, organized the printing of engraved reproductions of the great bulk of his work—both his paintings and his drawings—so that Watteau's total artistic output became and continued to be more accessible than that of any other artist until the twentieth-century advent of art monographs illustrated with photographs. These engravings presented aristocratic (and would-be aristocratic) eighteenth-century French society with an image of itself that was highly acceptable and widely imitated by other artists, however little relationship that image bore to reality. By 1884, the bicentenary of Watteau's birth, it was standard practice for biographers to refer to him as "the personification of the witty and amiable eighteenth century." In fact, Watteau saw little enough of that "witty and amiable" century for which so much nostalgia was generally felt between about 1870 and 1920, a period during which enthusiasm for the artist reached its peak. The eighteenth century's first decades, the period of his artistic activity, were fairly calamitous ones. During his short life, France was almost continually at war: his native region was overrun with foreign troops, and Paris was threatened by siege and by a rampaging army rabble. The dreadful winter of 1709, the year of Watteau's first Paris successes, was marked by military defeat and a disastrous famine. Most of Watteau's nineteenth-century admirers simply ignored the grim background of the works they found so lyrical and charming. Those who took the inconvenient historical facts into consideration did so only in order to refute the widely held deterministic view that the content and style of an artist's work were absolutely dictated by heredity and environment. (For Watteau admirers, such determinism was unthinkable: the artist was born in a Flemish town only six years after it first became part of France, yet Watteau was quintessentially French. As one patriotic French biographer put it, "In Dresden, Potsdam, and Berlin I have never come across a Watteau without feeling refreshed by a breath of native air." ) Even such writers, however, persisted in according Watteau's canvases a privileged status as representative "personifications" of the eighteenth century. The discrepancy between historical fact and artistic vision, useful in refuting the extreme deterministic position, merely forced these writers to seek a new formula that allowed them to preserve the desired identity between image and reality, this time a rather suspiciously psychic one: Watteau did not record the society he knew, but rather "foresaw" a society that developed shortly after his death.
199412_3-RC_3_20
[ "An artist's recognition by a society is most directly determined by the degree to which his or her works are perceived as lyrical and charming.", "An artist will have the greatest influence on a society that values art particularly highly.", "The works of an artist who captures the true and essential nature of a given society will probably have a great impact on that society.", "The degree of influence an artist's vision will have on a society is conditional on the visibility of the artist's work.", "An artist who is much imitated by contemporaries will usually fail to have an impact on a society unless the imitators are talented." ]
3
The information given in the passage suggests that which one of the following principles accurately characterizes the relationship between an artist's work and the impact it is likely to have on a society?
Faced with the problems of insufficient evidence, of conflicting evidence, and of evidence relayed through the flawed perceptual, retentive, and narrative abilities of witnesses, a jury is forced to draw inferences in its attempt to ascertain the truth. By applying the same cognitive tools they have developed and used over a lifetime, jurors engage in the inferential exercise that lawyers call fact-finding. In certain decision-making contexts that are relevant to the trial of lawsuits, however, these normally reliable cognitive tools may cause jurors to commit inferential errors that distort rather than reveal the truth. Although juries can make a variety of inferential errors, most of these mistakes in judgment involve the drawing of an unwarranted conclusion from the evidence, that is, deciding that the evidence proves something that, in reality, it does not prove. For example, evidence that the defendant in a criminal prosecution has a prior conviction may encourage jurors to presume the defendant's guilt, because of their preconception that a person previously convicted of a crime must be inclined toward repeated criminal behavior. That commonly held belief is at least a partial distortion of reality; not all former convicts engage in repeated criminal behavior. Also, a jury may give more probative weight than objective analysis would allow to vivid photographic evidence depicting a shooting victim's wounds, or may underestimate the weight of defense testimony that is not delivered in a sufficiently forceful or persuasive manner. Finally, complex or voluminous evidence might be so confusing to a jury that its members would draw totally unwarranted conclusions or even ignore the evidence entirely. Recent empirical research in cognitive psychology suggests that people tend to commit inferential errors like these under certain predictable circumstances. By examining the available information, the situation, and the type of decision being made, cognitive psychologists can describe the kinds of inferential errors a person or group is likely to make. These patterns of human decision-making may provide the courts with a guide to evaluating the effect of evidence on the reliability of the jury's inferential processes in certain situations. The notion that juries can commit inferential errors that jeopardize the accuracy of the fact-finding process is not unknown to the courts. In fact, one of a presiding judge's duties is to minimize jury inferential error through explanation and clarification. Nonetheless, most judges now employ only a limited and primitive concept of jury inferential error: limited because it fails to recognize the potential for error outside certain traditional situations, primitive because it ignores the research and conclusions of psychologists in favor of notions about human cognition held by lawyers.
199412_3-RC_4_21
[ "When making decisions in certain predictable situations, juries may commit inferential errors that obscure rather than reveal the truth.", "The views of human cognition taken by cognitive psychologists on the one hand and by the legal profession on the other are demonstrably dissimilar.", "When confronting powerful preconceptions, particularly shocking evidence, or complex situations, jurors make errors in judgment.", "The problem of inferential error by juries is typical of the difficulties with cognitive processes that people face in their everyday lives.", "Juries would probably make more reliable decisions if cognitive psychologists, rather than judges, instructed them about the problems inherent in drawing unwarranted conclusions." ]
0
Which one of the following best expresses the main idea of the passage?
Faced with the problems of insufficient evidence, of conflicting evidence, and of evidence relayed through the flawed perceptual, retentive, and narrative abilities of witnesses, a jury is forced to draw inferences in its attempt to ascertain the truth. By applying the same cognitive tools they have developed and used over a lifetime, jurors engage in the inferential exercise that lawyers call fact-finding. In certain decision-making contexts that are relevant to the trial of lawsuits, however, these normally reliable cognitive tools may cause jurors to commit inferential errors that distort rather than reveal the truth. Although juries can make a variety of inferential errors, most of these mistakes in judgment involve the drawing of an unwarranted conclusion from the evidence, that is, deciding that the evidence proves something that, in reality, it does not prove. For example, evidence that the defendant in a criminal prosecution has a prior conviction may encourage jurors to presume the defendant's guilt, because of their preconception that a person previously convicted of a crime must be inclined toward repeated criminal behavior. That commonly held belief is at least a partial distortion of reality; not all former convicts engage in repeated criminal behavior. Also, a jury may give more probative weight than objective analysis would allow to vivid photographic evidence depicting a shooting victim's wounds, or may underestimate the weight of defense testimony that is not delivered in a sufficiently forceful or persuasive manner. Finally, complex or voluminous evidence might be so confusing to a jury that its members would draw totally unwarranted conclusions or even ignore the evidence entirely. Recent empirical research in cognitive psychology suggests that people tend to commit inferential errors like these under certain predictable circumstances. By examining the available information, the situation, and the type of decision being made, cognitive psychologists can describe the kinds of inferential errors a person or group is likely to make. These patterns of human decision-making may provide the courts with a guide to evaluating the effect of evidence on the reliability of the jury's inferential processes in certain situations. The notion that juries can commit inferential errors that jeopardize the accuracy of the fact-finding process is not unknown to the courts. In fact, one of a presiding judge's duties is to minimize jury inferential error through explanation and clarification. Nonetheless, most judges now employ only a limited and primitive concept of jury inferential error: limited because it fails to recognize the potential for error outside certain traditional situations, primitive because it ignores the research and conclusions of psychologists in favor of notions about human cognition held by lawyers.
199412_3-RC_4_22
[ "a move away from jury trials", "the institution of minimum formal educational requirements for jurors", "the development of strict guidelines for defense testimony", "specific training for judges in the area of jury instruction", "restrictions on lawyers' use of psychological research" ]
3
Of the following hypothetical reforms in trial procedure, which one would the author be most likely to support as the best way to address the problem of jury inferential error?
Faced with the problems of insufficient evidence, of conflicting evidence, and of evidence relayed through the flawed perceptual, retentive, and narrative abilities of witnesses, a jury is forced to draw inferences in its attempt to ascertain the truth. By applying the same cognitive tools they have developed and used over a lifetime, jurors engage in the inferential exercise that lawyers call fact-finding. In certain decision-making contexts that are relevant to the trial of lawsuits, however, these normally reliable cognitive tools may cause jurors to commit inferential errors that distort rather than reveal the truth. Although juries can make a variety of inferential errors, most of these mistakes in judgment involve the drawing of an unwarranted conclusion from the evidence, that is, deciding that the evidence proves something that, in reality, it does not prove. For example, evidence that the defendant in a criminal prosecution has a prior conviction may encourage jurors to presume the defendant's guilt, because of their preconception that a person previously convicted of a crime must be inclined toward repeated criminal behavior. That commonly held belief is at least a partial distortion of reality; not all former convicts engage in repeated criminal behavior. Also, a jury may give more probative weight than objective analysis would allow to vivid photographic evidence depicting a shooting victim's wounds, or may underestimate the weight of defense testimony that is not delivered in a sufficiently forceful or persuasive manner. Finally, complex or voluminous evidence might be so confusing to a jury that its members would draw totally unwarranted conclusions or even ignore the evidence entirely. Recent empirical research in cognitive psychology suggests that people tend to commit inferential errors like these under certain predictable circumstances. By examining the available information, the situation, and the type of decision being made, cognitive psychologists can describe the kinds of inferential errors a person or group is likely to make. These patterns of human decision-making may provide the courts with a guide to evaluating the effect of evidence on the reliability of the jury's inferential processes in certain situations. The notion that juries can commit inferential errors that jeopardize the accuracy of the fact-finding process is not unknown to the courts. In fact, one of a presiding judge's duties is to minimize jury inferential error through explanation and clarification. Nonetheless, most judges now employ only a limited and primitive concept of jury inferential error: limited because it fails to recognize the potential for error outside certain traditional situations, primitive because it ignores the research and conclusions of psychologists in favor of notions about human cognition held by lawyers.
199412_3-RC_4_23
[ "refute the idea that the fact-finding process is a complicated exercise", "emphasize how carefully evidence must be presented in order to avoid jury inferential error", "explain how commonly held beliefs affect the Jury's ability to ascertain the truth", "provide examples of situations that may precipitate jury errors", "recommend a method for minimizing mistakes by juries" ]
3
In the second paragraph, the author's primary purpose is to
Faced with the problems of insufficient evidence, of conflicting evidence, and of evidence relayed through the flawed perceptual, retentive, and narrative abilities of witnesses, a jury is forced to draw inferences in its attempt to ascertain the truth. By applying the same cognitive tools they have developed and used over a lifetime, jurors engage in the inferential exercise that lawyers call fact-finding. In certain decision-making contexts that are relevant to the trial of lawsuits, however, these normally reliable cognitive tools may cause jurors to commit inferential errors that distort rather than reveal the truth. Although juries can make a variety of inferential errors, most of these mistakes in judgment involve the drawing of an unwarranted conclusion from the evidence, that is, deciding that the evidence proves something that, in reality, it does not prove. For example, evidence that the defendant in a criminal prosecution has a prior conviction may encourage jurors to presume the defendant's guilt, because of their preconception that a person previously convicted of a crime must be inclined toward repeated criminal behavior. That commonly held belief is at least a partial distortion of reality; not all former convicts engage in repeated criminal behavior. Also, a jury may give more probative weight than objective analysis would allow to vivid photographic evidence depicting a shooting victim's wounds, or may underestimate the weight of defense testimony that is not delivered in a sufficiently forceful or persuasive manner. Finally, complex or voluminous evidence might be so confusing to a jury that its members would draw totally unwarranted conclusions or even ignore the evidence entirely. Recent empirical research in cognitive psychology suggests that people tend to commit inferential errors like these under certain predictable circumstances. By examining the available information, the situation, and the type of decision being made, cognitive psychologists can describe the kinds of inferential errors a person or group is likely to make. These patterns of human decision-making may provide the courts with a guide to evaluating the effect of evidence on the reliability of the jury's inferential processes in certain situations. The notion that juries can commit inferential errors that jeopardize the accuracy of the fact-finding process is not unknown to the courts. In fact, one of a presiding judge's duties is to minimize jury inferential error through explanation and clarification. Nonetheless, most judges now employ only a limited and primitive concept of jury inferential error: limited because it fails to recognize the potential for error outside certain traditional situations, primitive because it ignores the research and conclusions of psychologists in favor of notions about human cognition held by lawyers.
199412_3-RC_4_24
[ "apprehensive about whether they are consistent in their instruction of juries", "doubtful of their ability to draw consistently correct conclusions based on the evidence", "critical of their failure to take into account potentially helpful research", "pessimistic about their willingness to make significant changes in trial procedure", "concerned about their allowing the presentation of complex and voluminous evidence in the courtroom" ]
2
Which one of the following best describes the author's attitude toward the majority of judges today?
Faced with the problems of insufficient evidence, of conflicting evidence, and of evidence relayed through the flawed perceptual, retentive, and narrative abilities of witnesses, a jury is forced to draw inferences in its attempt to ascertain the truth. By applying the same cognitive tools they have developed and used over a lifetime, jurors engage in the inferential exercise that lawyers call fact-finding. In certain decision-making contexts that are relevant to the trial of lawsuits, however, these normally reliable cognitive tools may cause jurors to commit inferential errors that distort rather than reveal the truth. Although juries can make a variety of inferential errors, most of these mistakes in judgment involve the drawing of an unwarranted conclusion from the evidence, that is, deciding that the evidence proves something that, in reality, it does not prove. For example, evidence that the defendant in a criminal prosecution has a prior conviction may encourage jurors to presume the defendant's guilt, because of their preconception that a person previously convicted of a crime must be inclined toward repeated criminal behavior. That commonly held belief is at least a partial distortion of reality; not all former convicts engage in repeated criminal behavior. Also, a jury may give more probative weight than objective analysis would allow to vivid photographic evidence depicting a shooting victim's wounds, or may underestimate the weight of defense testimony that is not delivered in a sufficiently forceful or persuasive manner. Finally, complex or voluminous evidence might be so confusing to a jury that its members would draw totally unwarranted conclusions or even ignore the evidence entirely. Recent empirical research in cognitive psychology suggests that people tend to commit inferential errors like these under certain predictable circumstances. By examining the available information, the situation, and the type of decision being made, cognitive psychologists can describe the kinds of inferential errors a person or group is likely to make. These patterns of human decision-making may provide the courts with a guide to evaluating the effect of evidence on the reliability of the jury's inferential processes in certain situations. The notion that juries can commit inferential errors that jeopardize the accuracy of the fact-finding process is not unknown to the courts. In fact, one of a presiding judge's duties is to minimize jury inferential error through explanation and clarification. Nonetheless, most judges now employ only a limited and primitive concept of jury inferential error: limited because it fails to recognize the potential for error outside certain traditional situations, primitive because it ignores the research and conclusions of psychologists in favor of notions about human cognition held by lawyers.
199412_3-RC_4_25
[ "All guidelines about human behavior must take account of variations in the patterns of human decision-making.", "Current models of how humans make decisions apply reliably to individuals but do not hold for decisions made by groups.", "The current conception of jury inferential error employed by judges has been in use for nearly a century.", "Inferential errors can be more easily predicted in controlled situations such as the trial of lawsuits than in other kinds of decision-making processes.", "In certain predictable circumstances, juries are less susceptible to inferential errors than they are in other circumstances." ]
1
Which one of the following statements, if true, would most seriously undermine the author's suggestion about the use of current psychological research in the courtroom?
Faced with the problems of insufficient evidence, of conflicting evidence, and of evidence relayed through the flawed perceptual, retentive, and narrative abilities of witnesses, a jury is forced to draw inferences in its attempt to ascertain the truth. By applying the same cognitive tools they have developed and used over a lifetime, jurors engage in the inferential exercise that lawyers call fact-finding. In certain decision-making contexts that are relevant to the trial of lawsuits, however, these normally reliable cognitive tools may cause jurors to commit inferential errors that distort rather than reveal the truth. Although juries can make a variety of inferential errors, most of these mistakes in judgment involve the drawing of an unwarranted conclusion from the evidence, that is, deciding that the evidence proves something that, in reality, it does not prove. For example, evidence that the defendant in a criminal prosecution has a prior conviction may encourage jurors to presume the defendant's guilt, because of their preconception that a person previously convicted of a crime must be inclined toward repeated criminal behavior. That commonly held belief is at least a partial distortion of reality; not all former convicts engage in repeated criminal behavior. Also, a jury may give more probative weight than objective analysis would allow to vivid photographic evidence depicting a shooting victim's wounds, or may underestimate the weight of defense testimony that is not delivered in a sufficiently forceful or persuasive manner. Finally, complex or voluminous evidence might be so confusing to a jury that its members would draw totally unwarranted conclusions or even ignore the evidence entirely. Recent empirical research in cognitive psychology suggests that people tend to commit inferential errors like these under certain predictable circumstances. By examining the available information, the situation, and the type of decision being made, cognitive psychologists can describe the kinds of inferential errors a person or group is likely to make. These patterns of human decision-making may provide the courts with a guide to evaluating the effect of evidence on the reliability of the jury's inferential processes in certain situations. The notion that juries can commit inferential errors that jeopardize the accuracy of the fact-finding process is not unknown to the courts. In fact, one of a presiding judge's duties is to minimize jury inferential error through explanation and clarification. Nonetheless, most judges now employ only a limited and primitive concept of jury inferential error: limited because it fails to recognize the potential for error outside certain traditional situations, primitive because it ignores the research and conclusions of psychologists in favor of notions about human cognition held by lawyers.
199412_3-RC_4_26
[ "They have a less sophisticated understanding of human cognition than do psychologists.", "They often present complex or voluminous information merely in order to confuse a jury.", "They are no better at making logical inferences from the testimony at a trial than are most judges.", "They have worked to help judges minimize jury inferential error.", "They are unrealistic about the ability of jurors to ascertain the truth." ]
0
It can be inferred from the passage that the author would be most likely to agree with which one of the following generalizations about lawyers?
Faced with the problems of insufficient evidence, of conflicting evidence, and of evidence relayed through the flawed perceptual, retentive, and narrative abilities of witnesses, a jury is forced to draw inferences in its attempt to ascertain the truth. By applying the same cognitive tools they have developed and used over a lifetime, jurors engage in the inferential exercise that lawyers call fact-finding. In certain decision-making contexts that are relevant to the trial of lawsuits, however, these normally reliable cognitive tools may cause jurors to commit inferential errors that distort rather than reveal the truth. Although juries can make a variety of inferential errors, most of these mistakes in judgment involve the drawing of an unwarranted conclusion from the evidence, that is, deciding that the evidence proves something that, in reality, it does not prove. For example, evidence that the defendant in a criminal prosecution has a prior conviction may encourage jurors to presume the defendant's guilt, because of their preconception that a person previously convicted of a crime must be inclined toward repeated criminal behavior. That commonly held belief is at least a partial distortion of reality; not all former convicts engage in repeated criminal behavior. Also, a jury may give more probative weight than objective analysis would allow to vivid photographic evidence depicting a shooting victim's wounds, or may underestimate the weight of defense testimony that is not delivered in a sufficiently forceful or persuasive manner. Finally, complex or voluminous evidence might be so confusing to a jury that its members would draw totally unwarranted conclusions or even ignore the evidence entirely. Recent empirical research in cognitive psychology suggests that people tend to commit inferential errors like these under certain predictable circumstances. By examining the available information, the situation, and the type of decision being made, cognitive psychologists can describe the kinds of inferential errors a person or group is likely to make. These patterns of human decision-making may provide the courts with a guide to evaluating the effect of evidence on the reliability of the jury's inferential processes in certain situations. The notion that juries can commit inferential errors that jeopardize the accuracy of the fact-finding process is not unknown to the courts. In fact, one of a presiding judge's duties is to minimize jury inferential error through explanation and clarification. Nonetheless, most judges now employ only a limited and primitive concept of jury inferential error: limited because it fails to recognize the potential for error outside certain traditional situations, primitive because it ignores the research and conclusions of psychologists in favor of notions about human cognition held by lawyers.
199412_3-RC_4_27
[ "The more evidence that a jury has, the more likely it is that the jury will reach a reliable verdict.", "Juries usually overestimate the value of visual evidence such as photographs.", "Jurors have preconceptions about the behavior of defendants that prevent them from making an objective analysis of the evidence in a criminal trial.", "Most of the jurors who make inferential errors during a trial do so because they are unaccustomed to having to make difficult decisions based on inferences.", "The manner in which evidence is presented to a jury may influence the jury either to overestimate or to underestimate the value of that evidence." ]
4
The author would be most likely to agree with which one of the following generalizations about a jury's decision-making process?
It is a fundamental tenet of geophysics that the Earth's magnetic field can exist in either of two polarity states: a "normal" state, in which north-seeking compass needles point to the geographic north, and a "reverse" state, in which they point to the geographic south. Geological evidence shows that periodically the field's polarity reverses, and that these reversals have been taking place at an increasing rate. Evidence also indicates that the field does not reverse instantaneously from one polarity state to another; rather, the process involves a transition period that typically spans a few thousand years. Though this much is known, the underlying causes of the reversal phenomenon are not well understood. It is generally accepted that the magnetic field itself is generated by the motion of free electrons in the outer core, a slowly churning mass of molten metal sandwiched between the Earth's mantle (the region of the Earth's interior lying below the crust) and its solid inner core. In some way that is not completely understood, gravity and the Earth's rotation, acting on temperature and density differences within the outer core fluid, provide the driving forces behind the generation of the field. The reversal phenomenon may be triggered when something disturbs the heat circulation pattern of the outer core fluid, and with it the magnetic field. Several explanations for this phenomenon have been proposed. One proposal, the "heat-transfer hypothesis," is that the triggering process is intimately related to the way the outer core vents its heat into the mantle. For example, such heat transfer could create hotter (rising) or cooler (descending) blobs of material from the inner and outer boundaries of the fluid core, thereby perturbing the main heat-circulation pattern. A more controversial alternative proposal is the "asteroid-impact hypothesis." In this scenario an extended period of cold and darkness results from the impact of an asteroid large enough to send a great cloud of dust into the atmosphere. Following this climatic change, ocean temperatures drop and the polar ice caps grow, redistributing the Earth's seawater. This redistribution increases the rotational acceleration of the mantle, causing friction and turbulence near the outer core-mantle boundary and initiating a reversal of the magnetic field. How well do these hypotheses account for such observations as the long-term increase in the frequency of reversal? In support of the asteroid-impact model, it has been argued that the gradual cooling of the average ocean temperature would enable progressively smaller asteroid impacts (which are known to occur more frequently than larger impacts) to cool the Earth's climate sufficiently to induce ice-cap growth and reversals. But theories that depend on extraterrestrial intervention seem less convincing than theories like the first, which account for the phenomenon solely by means of the thermodynamic state of the outer core and its effect on the mantle.
199502_3-RC_1_1
[ "Heat circulation in the outer core controls the growth and diminution of the polar ice caps.", "Impact of asteroids on the Earth's surface alters the way in which the outer core vents its heat into the mantle.", "Motion of electrons within the metallic fluid in the outer core produces the Earth's magnetic field.", "Friction and turbulence near the boundary between the outer core and the mantle are typically caused by asteroid impacts.", "Cessation of heat circulation within the outer core brings on multiple reversals in the Earth's magnetic field." ]
2
Which one of the following statements regarding the Earth's outer core is best supported by information presented in the passage?
It is a fundamental tenet of geophysics that the Earth's magnetic field can exist in either of two polarity states: a "normal" state, in which north-seeking compass needles point to the geographic north, and a "reverse" state, in which they point to the geographic south. Geological evidence shows that periodically the field's polarity reverses, and that these reversals have been taking place at an increasing rate. Evidence also indicates that the field does not reverse instantaneously from one polarity state to another; rather, the process involves a transition period that typically spans a few thousand years. Though this much is known, the underlying causes of the reversal phenomenon are not well understood. It is generally accepted that the magnetic field itself is generated by the motion of free electrons in the outer core, a slowly churning mass of molten metal sandwiched between the Earth's mantle (the region of the Earth's interior lying below the crust) and its solid inner core. In some way that is not completely understood, gravity and the Earth's rotation, acting on temperature and density differences within the outer core fluid, provide the driving forces behind the generation of the field. The reversal phenomenon may be triggered when something disturbs the heat circulation pattern of the outer core fluid, and with it the magnetic field. Several explanations for this phenomenon have been proposed. One proposal, the "heat-transfer hypothesis," is that the triggering process is intimately related to the way the outer core vents its heat into the mantle. For example, such heat transfer could create hotter (rising) or cooler (descending) blobs of material from the inner and outer boundaries of the fluid core, thereby perturbing the main heat-circulation pattern. A more controversial alternative proposal is the "asteroid-impact hypothesis." In this scenario an extended period of cold and darkness results from the impact of an asteroid large enough to send a great cloud of dust into the atmosphere. Following this climatic change, ocean temperatures drop and the polar ice caps grow, redistributing the Earth's seawater. This redistribution increases the rotational acceleration of the mantle, causing friction and turbulence near the outer core-mantle boundary and initiating a reversal of the magnetic field. How well do these hypotheses account for such observations as the long-term increase in the frequency of reversal? In support of the asteroid-impact model, it has been argued that the gradual cooling of the average ocean temperature would enable progressively smaller asteroid impacts (which are known to occur more frequently than larger impacts) to cool the Earth's climate sufficiently to induce ice-cap growth and reversals. But theories that depend on extraterrestrial intervention seem less convincing than theories like the first, which account for the phenomenon solely by means of the thermodynamic state of the outer core and its effect on the mantle.
199502_3-RC_1_2
[ "The extinction of the dinosaurs was the result of gradual changes in the composition of the Earth's atmosphere that occurred over millions of years.", "The dinosaurs became extinct when their food supply was disrupted following the emergence of mammals.", "The dinosaurs succumbed to the new, colder environment brought about by a buildup of volcanic ash in the atmosphere.", "After massively overpopulating the planet, dinosaurs disappeared due to widespread starvation and the rapid spread of disease.", "After radical climatic changes resulted from the impact of a comet, dinosaurs disappeared from the Earth." ]
4
The author's objection to the second hypothesis discussed in the passage is most applicable to which one of the following explanations concerning the extinction of the dinosaurs'?
It is a fundamental tenet of geophysics that the Earth's magnetic field can exist in either of two polarity states: a "normal" state, in which north-seeking compass needles point to the geographic north, and a "reverse" state, in which they point to the geographic south. Geological evidence shows that periodically the field's polarity reverses, and that these reversals have been taking place at an increasing rate. Evidence also indicates that the field does not reverse instantaneously from one polarity state to another; rather, the process involves a transition period that typically spans a few thousand years. Though this much is known, the underlying causes of the reversal phenomenon are not well understood. It is generally accepted that the magnetic field itself is generated by the motion of free electrons in the outer core, a slowly churning mass of molten metal sandwiched between the Earth's mantle (the region of the Earth's interior lying below the crust) and its solid inner core. In some way that is not completely understood, gravity and the Earth's rotation, acting on temperature and density differences within the outer core fluid, provide the driving forces behind the generation of the field. The reversal phenomenon may be triggered when something disturbs the heat circulation pattern of the outer core fluid, and with it the magnetic field. Several explanations for this phenomenon have been proposed. One proposal, the "heat-transfer hypothesis," is that the triggering process is intimately related to the way the outer core vents its heat into the mantle. For example, such heat transfer could create hotter (rising) or cooler (descending) blobs of material from the inner and outer boundaries of the fluid core, thereby perturbing the main heat-circulation pattern. A more controversial alternative proposal is the "asteroid-impact hypothesis." In this scenario an extended period of cold and darkness results from the impact of an asteroid large enough to send a great cloud of dust into the atmosphere. Following this climatic change, ocean temperatures drop and the polar ice caps grow, redistributing the Earth's seawater. This redistribution increases the rotational acceleration of the mantle, causing friction and turbulence near the outer core-mantle boundary and initiating a reversal of the magnetic field. How well do these hypotheses account for such observations as the long-term increase in the frequency of reversal? In support of the asteroid-impact model, it has been argued that the gradual cooling of the average ocean temperature would enable progressively smaller asteroid impacts (which are known to occur more frequently than larger impacts) to cool the Earth's climate sufficiently to induce ice-cap growth and reversals. But theories that depend on extraterrestrial intervention seem less convincing than theories like the first, which account for the phenomenon solely by means of the thermodynamic state of the outer core and its effect on the mantle.
199502_3-RC_1_3
[ "present a way in which the venting of heat from the outer core might disturb the heat-circulation pattern within the outer core", "provide proof for the proposal that ventilation of heat from the outer core into the mantle triggers polarity reversal", "give an example of the way in which heat circulates between the Earth's outer core and the Earth's exterior", "describe how the outer core maintains its temperature by venting its excess heat into the Earth's mantle", "argue in favor of the theory that heat circulation in the Earth's interior produces the magnetic field" ]
0
The author mentions the creation of blobs of different temperatures in the Earth's outer core (lines 34–38) primarily in order to
It is a fundamental tenet of geophysics that the Earth's magnetic field can exist in either of two polarity states: a "normal" state, in which north-seeking compass needles point to the geographic north, and a "reverse" state, in which they point to the geographic south. Geological evidence shows that periodically the field's polarity reverses, and that these reversals have been taking place at an increasing rate. Evidence also indicates that the field does not reverse instantaneously from one polarity state to another; rather, the process involves a transition period that typically spans a few thousand years. Though this much is known, the underlying causes of the reversal phenomenon are not well understood. It is generally accepted that the magnetic field itself is generated by the motion of free electrons in the outer core, a slowly churning mass of molten metal sandwiched between the Earth's mantle (the region of the Earth's interior lying below the crust) and its solid inner core. In some way that is not completely understood, gravity and the Earth's rotation, acting on temperature and density differences within the outer core fluid, provide the driving forces behind the generation of the field. The reversal phenomenon may be triggered when something disturbs the heat circulation pattern of the outer core fluid, and with it the magnetic field. Several explanations for this phenomenon have been proposed. One proposal, the "heat-transfer hypothesis," is that the triggering process is intimately related to the way the outer core vents its heat into the mantle. For example, such heat transfer could create hotter (rising) or cooler (descending) blobs of material from the inner and outer boundaries of the fluid core, thereby perturbing the main heat-circulation pattern. A more controversial alternative proposal is the "asteroid-impact hypothesis." In this scenario an extended period of cold and darkness results from the impact of an asteroid large enough to send a great cloud of dust into the atmosphere. Following this climatic change, ocean temperatures drop and the polar ice caps grow, redistributing the Earth's seawater. This redistribution increases the rotational acceleration of the mantle, causing friction and turbulence near the outer core-mantle boundary and initiating a reversal of the magnetic field. How well do these hypotheses account for such observations as the long-term increase in the frequency of reversal? In support of the asteroid-impact model, it has been argued that the gradual cooling of the average ocean temperature would enable progressively smaller asteroid impacts (which are known to occur more frequently than larger impacts) to cool the Earth's climate sufficiently to induce ice-cap growth and reversals. But theories that depend on extraterrestrial intervention seem less convincing than theories like the first, which account for the phenomenon solely by means of the thermodynamic state of the outer core and its effect on the mantle.
199502_3-RC_1_4
[ "Most, but not all, geophysicists agree that the Earth's magnetic field may exist in two distinct polarity states.", "Changes in the polarity of the Earth's magnetic field have occurred more often in the recent past than in the distant past.", "Heat transfer would cause reversals of the polarity of the Earth's magnetic field to occur more quickly than would asteroid impact.", "Geophysicists' understanding of the reversal of the Earth's magnetic field has increased significantly since the introduction of the heat-transfer hypothesis.", "Friction near the boundary of the inner and outer cores brings on reversal of the polarity of the geomagnetic field." ]
1
Which one of the following statements regarding the polarity of the Earth's magnetic field is best supported by information in the passage?
It is a fundamental tenet of geophysics that the Earth's magnetic field can exist in either of two polarity states: a "normal" state, in which north-seeking compass needles point to the geographic north, and a "reverse" state, in which they point to the geographic south. Geological evidence shows that periodically the field's polarity reverses, and that these reversals have been taking place at an increasing rate. Evidence also indicates that the field does not reverse instantaneously from one polarity state to another; rather, the process involves a transition period that typically spans a few thousand years. Though this much is known, the underlying causes of the reversal phenomenon are not well understood. It is generally accepted that the magnetic field itself is generated by the motion of free electrons in the outer core, a slowly churning mass of molten metal sandwiched between the Earth's mantle (the region of the Earth's interior lying below the crust) and its solid inner core. In some way that is not completely understood, gravity and the Earth's rotation, acting on temperature and density differences within the outer core fluid, provide the driving forces behind the generation of the field. The reversal phenomenon may be triggered when something disturbs the heat circulation pattern of the outer core fluid, and with it the magnetic field. Several explanations for this phenomenon have been proposed. One proposal, the "heat-transfer hypothesis," is that the triggering process is intimately related to the way the outer core vents its heat into the mantle. For example, such heat transfer could create hotter (rising) or cooler (descending) blobs of material from the inner and outer boundaries of the fluid core, thereby perturbing the main heat-circulation pattern. A more controversial alternative proposal is the "asteroid-impact hypothesis." In this scenario an extended period of cold and darkness results from the impact of an asteroid large enough to send a great cloud of dust into the atmosphere. Following this climatic change, ocean temperatures drop and the polar ice caps grow, redistributing the Earth's seawater. This redistribution increases the rotational acceleration of the mantle, causing friction and turbulence near the outer core-mantle boundary and initiating a reversal of the magnetic field. How well do these hypotheses account for such observations as the long-term increase in the frequency of reversal? In support of the asteroid-impact model, it has been argued that the gradual cooling of the average ocean temperature would enable progressively smaller asteroid impacts (which are known to occur more frequently than larger impacts) to cool the Earth's climate sufficiently to induce ice-cap growth and reversals. But theories that depend on extraterrestrial intervention seem less convincing than theories like the first, which account for the phenomenon solely by means of the thermodynamic state of the outer core and its effect on the mantle.
199502_3-RC_1_5
[ "Since their introduction they have sharply divided the scientific community.", "Both were formulated in order to explain changes in the frequency of polarity reversal.", "Although no firm conclusions regarding them have yet been reached, both have been extensively investigated.", "They are not the only proposals scientists have put forward to explain the phenomenon of polarity reversal.", "Both were introduced some time ago and have since fallen into disfavor among geophysicists." ]
3
Which one of the following can be inferred regarding the two proposals discussed in the passage?
It is a fundamental tenet of geophysics that the Earth's magnetic field can exist in either of two polarity states: a "normal" state, in which north-seeking compass needles point to the geographic north, and a "reverse" state, in which they point to the geographic south. Geological evidence shows that periodically the field's polarity reverses, and that these reversals have been taking place at an increasing rate. Evidence also indicates that the field does not reverse instantaneously from one polarity state to another; rather, the process involves a transition period that typically spans a few thousand years. Though this much is known, the underlying causes of the reversal phenomenon are not well understood. It is generally accepted that the magnetic field itself is generated by the motion of free electrons in the outer core, a slowly churning mass of molten metal sandwiched between the Earth's mantle (the region of the Earth's interior lying below the crust) and its solid inner core. In some way that is not completely understood, gravity and the Earth's rotation, acting on temperature and density differences within the outer core fluid, provide the driving forces behind the generation of the field. The reversal phenomenon may be triggered when something disturbs the heat circulation pattern of the outer core fluid, and with it the magnetic field. Several explanations for this phenomenon have been proposed. One proposal, the "heat-transfer hypothesis," is that the triggering process is intimately related to the way the outer core vents its heat into the mantle. For example, such heat transfer could create hotter (rising) or cooler (descending) blobs of material from the inner and outer boundaries of the fluid core, thereby perturbing the main heat-circulation pattern. A more controversial alternative proposal is the "asteroid-impact hypothesis." In this scenario an extended period of cold and darkness results from the impact of an asteroid large enough to send a great cloud of dust into the atmosphere. Following this climatic change, ocean temperatures drop and the polar ice caps grow, redistributing the Earth's seawater. This redistribution increases the rotational acceleration of the mantle, causing friction and turbulence near the outer core-mantle boundary and initiating a reversal of the magnetic field. How well do these hypotheses account for such observations as the long-term increase in the frequency of reversal? In support of the asteroid-impact model, it has been argued that the gradual cooling of the average ocean temperature would enable progressively smaller asteroid impacts (which are known to occur more frequently than larger impacts) to cool the Earth's climate sufficiently to induce ice-cap growth and reversals. But theories that depend on extraterrestrial intervention seem less convincing than theories like the first, which account for the phenomenon solely by means of the thermodynamic state of the outer core and its effect on the mantle.
199502_3-RC_1_6
[ "changes in the way heat circulates within the outer core fluid", "extended periods of colder temperatures on the Earth's surface", "the creation of circulating blobs of outer core material of different temperatures", "changes in circulation patterns in the Earth's oceans", "clouding of the Earth's atmosphere by a large amount of dust" ]
3
The author mentions each of the following as possible contributing causes to reversals of the Earth's magnetic field EXCEPT
Innovations in language are never completely new. When the words used for familiar things change, or words for new things enter the language, they are usually borrowed or adapted from stock. Assuming new roles, they drag their old meanings along behind them like flickering shadows. This seems especially true of the language of the contemporary school of literary criticism that now prefers to describe its work simply and rather presumptuously as "theory" but is still popularly referred to as poststructuralism or deconstruction. The first neologisms adopted by this movement were signifier and signified, employed to distinguish words from their referents, and to illustrate the arbitrariness of the terms we choose. The use of these particular terms (rather than, respectively, word and thing) underlined the seriousness of the naming process and its claim on our attention. Since in English "to signify" can also mean "to portend," these terms also suggest that words predict coming events. With the use of the term deconstruction we move into another and more complex realm of meaning. The most common use of the terms construction and deconstruction is in the building trades, and their borrowing by literary theorists for a new type of criticism cannot help but have certain overtones to the outsider. First, the usage suggests that the creation and critical interpretation of literature are not organic but mechanical processes; that the author of any piece of writing is not an inspired, intuitive artist, but merely a laborer who cobbles existing materials (words) into more or less conventional structures. The term deconstruction implies that the text has been put together like a building or a piece of machinery, and that it is in need of being taken apart, not so much in order to repair it as to demonstrate underlying inadequacies, false assumptions, and inherent contradictions. This process can supposedly be repeated many times and by many literary hard hats; it is expected that each deconstruction will reveal additional flaws and expose the illusions or bad faith of the builder. The fact that deconstructionists prefer to describe their activities as deconstruction rather than criticism is also revealing. Criticism and critic derive from the Greek kritikos, "skillful in judging, decisive." Deconstruction, on the other hand, has no overtones of skill or wisdom; it merely suggests demolition of an existing building. In popular usage criticism suggests censure but not change. If we find fault with a building, we may condemn it, but we do not carry out the demolition ourselves. The deconstructionist, by implication, is both judge and executioner who leaves a text totally dismantled, if not reduced to a pile of rubble.
199502_3-RC_2_7
[ "Implicit in the terminology of the school of criticism known as deconstruction are meanings that reveal the true nature of the deconstructionist's endeavor.", "The appearance of the terms signifier and signified in the field of literary theory anticipated the appearance of an even more radical idea known as deconstruction.", "Innovations in language and the relations between old and new meanings of terms are a special concern of the new school of criticism known as deconstruction.", "Deconstructionists maintain that it is insufficient merely to judge a work; the critic must actively dismantle it.", "Progress in the field of literary theory is best achieved by looking for new terms like signifier and deconstruction that might suggest new critical approaches to a work." ]
0
Which one of the following best expresses the main idea of the passage?
Innovations in language are never completely new. When the words used for familiar things change, or words for new things enter the language, they are usually borrowed or adapted from stock. Assuming new roles, they drag their old meanings along behind them like flickering shadows. This seems especially true of the language of the contemporary school of literary criticism that now prefers to describe its work simply and rather presumptuously as "theory" but is still popularly referred to as poststructuralism or deconstruction. The first neologisms adopted by this movement were signifier and signified, employed to distinguish words from their referents, and to illustrate the arbitrariness of the terms we choose. The use of these particular terms (rather than, respectively, word and thing) underlined the seriousness of the naming process and its claim on our attention. Since in English "to signify" can also mean "to portend," these terms also suggest that words predict coming events. With the use of the term deconstruction we move into another and more complex realm of meaning. The most common use of the terms construction and deconstruction is in the building trades, and their borrowing by literary theorists for a new type of criticism cannot help but have certain overtones to the outsider. First, the usage suggests that the creation and critical interpretation of literature are not organic but mechanical processes; that the author of any piece of writing is not an inspired, intuitive artist, but merely a laborer who cobbles existing materials (words) into more or less conventional structures. The term deconstruction implies that the text has been put together like a building or a piece of machinery, and that it is in need of being taken apart, not so much in order to repair it as to demonstrate underlying inadequacies, false assumptions, and inherent contradictions. This process can supposedly be repeated many times and by many literary hard hats; it is expected that each deconstruction will reveal additional flaws and expose the illusions or bad faith of the builder. The fact that deconstructionists prefer to describe their activities as deconstruction rather than criticism is also revealing. Criticism and critic derive from the Greek kritikos, "skillful in judging, decisive." Deconstruction, on the other hand, has no overtones of skill or wisdom; it merely suggests demolition of an existing building. In popular usage criticism suggests censure but not change. If we find fault with a building, we may condemn it, but we do not carry out the demolition ourselves. The deconstructionist, by implication, is both judge and executioner who leaves a text totally dismantled, if not reduced to a pile of rubble.
199502_3-RC_2_8
[ "Deconstructionists would not have been able to formulate their views adequately without the terms signifier and signified.", "Deconstructionists had no particular purpose in mind in choosing to use neologisms.", "Deconstructionists do not recognize that their own theory contains inherent contradictions.", "Deconstructionists find little interest in the relationship between words and their referents.", "Deconstructionists use the terms signifier and signified to stress the importance of the process of naming." ]
4
Which one of the following is a claim that the author passage makes about deconstructionists?
Innovations in language are never completely new. When the words used for familiar things change, or words for new things enter the language, they are usually borrowed or adapted from stock. Assuming new roles, they drag their old meanings along behind them like flickering shadows. This seems especially true of the language of the contemporary school of literary criticism that now prefers to describe its work simply and rather presumptuously as "theory" but is still popularly referred to as poststructuralism or deconstruction. The first neologisms adopted by this movement were signifier and signified, employed to distinguish words from their referents, and to illustrate the arbitrariness of the terms we choose. The use of these particular terms (rather than, respectively, word and thing) underlined the seriousness of the naming process and its claim on our attention. Since in English "to signify" can also mean "to portend," these terms also suggest that words predict coming events. With the use of the term deconstruction we move into another and more complex realm of meaning. The most common use of the terms construction and deconstruction is in the building trades, and their borrowing by literary theorists for a new type of criticism cannot help but have certain overtones to the outsider. First, the usage suggests that the creation and critical interpretation of literature are not organic but mechanical processes; that the author of any piece of writing is not an inspired, intuitive artist, but merely a laborer who cobbles existing materials (words) into more or less conventional structures. The term deconstruction implies that the text has been put together like a building or a piece of machinery, and that it is in need of being taken apart, not so much in order to repair it as to demonstrate underlying inadequacies, false assumptions, and inherent contradictions. This process can supposedly be repeated many times and by many literary hard hats; it is expected that each deconstruction will reveal additional flaws and expose the illusions or bad faith of the builder. The fact that deconstructionists prefer to describe their activities as deconstruction rather than criticism is also revealing. Criticism and critic derive from the Greek kritikos, "skillful in judging, decisive." Deconstruction, on the other hand, has no overtones of skill or wisdom; it merely suggests demolition of an existing building. In popular usage criticism suggests censure but not change. If we find fault with a building, we may condemn it, but we do not carry out the demolition ourselves. The deconstructionist, by implication, is both judge and executioner who leaves a text totally dismantled, if not reduced to a pile of rubble.
199502_3-RC_2_9
[ "A new invention usually consists of components that are specifically manufactured for the new invention.", "A new invention is usually behind the times, never making as much use of all the available modern technology as it could.", "A new invention usually consists of components that are already available but are made to function in new ways.", "A new invention is most useful when it is created with attention to the historical tradition established by implements previously used to do the same job.", "A new invention is rarely used to its full potential because it is surrounded by out-of-date technology that hinders its application." ]
2
Which one of the following generalizations about inventions is most analogous to the author's point about innovation in language?
Innovations in language are never completely new. When the words used for familiar things change, or words for new things enter the language, they are usually borrowed or adapted from stock. Assuming new roles, they drag their old meanings along behind them like flickering shadows. This seems especially true of the language of the contemporary school of literary criticism that now prefers to describe its work simply and rather presumptuously as "theory" but is still popularly referred to as poststructuralism or deconstruction. The first neologisms adopted by this movement were signifier and signified, employed to distinguish words from their referents, and to illustrate the arbitrariness of the terms we choose. The use of these particular terms (rather than, respectively, word and thing) underlined the seriousness of the naming process and its claim on our attention. Since in English "to signify" can also mean "to portend," these terms also suggest that words predict coming events. With the use of the term deconstruction we move into another and more complex realm of meaning. The most common use of the terms construction and deconstruction is in the building trades, and their borrowing by literary theorists for a new type of criticism cannot help but have certain overtones to the outsider. First, the usage suggests that the creation and critical interpretation of literature are not organic but mechanical processes; that the author of any piece of writing is not an inspired, intuitive artist, but merely a laborer who cobbles existing materials (words) into more or less conventional structures. The term deconstruction implies that the text has been put together like a building or a piece of machinery, and that it is in need of being taken apart, not so much in order to repair it as to demonstrate underlying inadequacies, false assumptions, and inherent contradictions. This process can supposedly be repeated many times and by many literary hard hats; it is expected that each deconstruction will reveal additional flaws and expose the illusions or bad faith of the builder. The fact that deconstructionists prefer to describe their activities as deconstruction rather than criticism is also revealing. Criticism and critic derive from the Greek kritikos, "skillful in judging, decisive." Deconstruction, on the other hand, has no overtones of skill or wisdom; it merely suggests demolition of an existing building. In popular usage criticism suggests censure but not change. If we find fault with a building, we may condemn it, but we do not carry out the demolition ourselves. The deconstructionist, by implication, is both judge and executioner who leaves a text totally dismantled, if not reduced to a pile of rubble.
199502_3-RC_2_10
[ "give an example", "introduce a contrast", "undermine an argument", "codify a system", "dismiss an objection" ]
1
The author of the passage uses the word "criticism" in lines 46–56 primarily in order to
Innovations in language are never completely new. When the words used for familiar things change, or words for new things enter the language, they are usually borrowed or adapted from stock. Assuming new roles, they drag their old meanings along behind them like flickering shadows. This seems especially true of the language of the contemporary school of literary criticism that now prefers to describe its work simply and rather presumptuously as "theory" but is still popularly referred to as poststructuralism or deconstruction. The first neologisms adopted by this movement were signifier and signified, employed to distinguish words from their referents, and to illustrate the arbitrariness of the terms we choose. The use of these particular terms (rather than, respectively, word and thing) underlined the seriousness of the naming process and its claim on our attention. Since in English "to signify" can also mean "to portend," these terms also suggest that words predict coming events. With the use of the term deconstruction we move into another and more complex realm of meaning. The most common use of the terms construction and deconstruction is in the building trades, and their borrowing by literary theorists for a new type of criticism cannot help but have certain overtones to the outsider. First, the usage suggests that the creation and critical interpretation of literature are not organic but mechanical processes; that the author of any piece of writing is not an inspired, intuitive artist, but merely a laborer who cobbles existing materials (words) into more or less conventional structures. The term deconstruction implies that the text has been put together like a building or a piece of machinery, and that it is in need of being taken apart, not so much in order to repair it as to demonstrate underlying inadequacies, false assumptions, and inherent contradictions. This process can supposedly be repeated many times and by many literary hard hats; it is expected that each deconstruction will reveal additional flaws and expose the illusions or bad faith of the builder. The fact that deconstructionists prefer to describe their activities as deconstruction rather than criticism is also revealing. Criticism and critic derive from the Greek kritikos, "skillful in judging, decisive." Deconstruction, on the other hand, has no overtones of skill or wisdom; it merely suggests demolition of an existing building. In popular usage criticism suggests censure but not change. If we find fault with a building, we may condemn it, but we do not carry out the demolition ourselves. The deconstructionist, by implication, is both judge and executioner who leaves a text totally dismantled, if not reduced to a pile of rubble.
199502_3-RC_2_11
[ "It introduces a hypothesis that the author later expands upon.", "It qualifies a claim made earlier by the author.", "It develops an initial example of the author's general thesis.", "It predicts a development.", "It presents a contrasting view." ]
2
Which one of the following best describes the function of the second paragraph within the passage as a whole?
Innovations in language are never completely new. When the words used for familiar things change, or words for new things enter the language, they are usually borrowed or adapted from stock. Assuming new roles, they drag their old meanings along behind them like flickering shadows. This seems especially true of the language of the contemporary school of literary criticism that now prefers to describe its work simply and rather presumptuously as "theory" but is still popularly referred to as poststructuralism or deconstruction. The first neologisms adopted by this movement were signifier and signified, employed to distinguish words from their referents, and to illustrate the arbitrariness of the terms we choose. The use of these particular terms (rather than, respectively, word and thing) underlined the seriousness of the naming process and its claim on our attention. Since in English "to signify" can also mean "to portend," these terms also suggest that words predict coming events. With the use of the term deconstruction we move into another and more complex realm of meaning. The most common use of the terms construction and deconstruction is in the building trades, and their borrowing by literary theorists for a new type of criticism cannot help but have certain overtones to the outsider. First, the usage suggests that the creation and critical interpretation of literature are not organic but mechanical processes; that the author of any piece of writing is not an inspired, intuitive artist, but merely a laborer who cobbles existing materials (words) into more or less conventional structures. The term deconstruction implies that the text has been put together like a building or a piece of machinery, and that it is in need of being taken apart, not so much in order to repair it as to demonstrate underlying inadequacies, false assumptions, and inherent contradictions. This process can supposedly be repeated many times and by many literary hard hats; it is expected that each deconstruction will reveal additional flaws and expose the illusions or bad faith of the builder. The fact that deconstructionists prefer to describe their activities as deconstruction rather than criticism is also revealing. Criticism and critic derive from the Greek kritikos, "skillful in judging, decisive." Deconstruction, on the other hand, has no overtones of skill or wisdom; it merely suggests demolition of an existing building. In popular usage criticism suggests censure but not change. If we find fault with a building, we may condemn it, but we do not carry out the demolition ourselves. The deconstructionist, by implication, is both judge and executioner who leaves a text totally dismantled, if not reduced to a pile of rubble.
199502_3-RC_2_12
[ "demonstrate false assumptions and inherent contradictions", "employ skill and insight", "be carried out by one critic rather than many", "reveal how a text is put together like a building", "point out the superiority of conventional text structures" ]
1
The passage suggests that the author most probably holds the view that an important characteristic of literary criticism is that it
Innovations in language are never completely new. When the words used for familiar things change, or words for new things enter the language, they are usually borrowed or adapted from stock. Assuming new roles, they drag their old meanings along behind them like flickering shadows. This seems especially true of the language of the contemporary school of literary criticism that now prefers to describe its work simply and rather presumptuously as "theory" but is still popularly referred to as poststructuralism or deconstruction. The first neologisms adopted by this movement were signifier and signified, employed to distinguish words from their referents, and to illustrate the arbitrariness of the terms we choose. The use of these particular terms (rather than, respectively, word and thing) underlined the seriousness of the naming process and its claim on our attention. Since in English "to signify" can also mean "to portend," these terms also suggest that words predict coming events. With the use of the term deconstruction we move into another and more complex realm of meaning. The most common use of the terms construction and deconstruction is in the building trades, and their borrowing by literary theorists for a new type of criticism cannot help but have certain overtones to the outsider. First, the usage suggests that the creation and critical interpretation of literature are not organic but mechanical processes; that the author of any piece of writing is not an inspired, intuitive artist, but merely a laborer who cobbles existing materials (words) into more or less conventional structures. The term deconstruction implies that the text has been put together like a building or a piece of machinery, and that it is in need of being taken apart, not so much in order to repair it as to demonstrate underlying inadequacies, false assumptions, and inherent contradictions. This process can supposedly be repeated many times and by many literary hard hats; it is expected that each deconstruction will reveal additional flaws and expose the illusions or bad faith of the builder. The fact that deconstructionists prefer to describe their activities as deconstruction rather than criticism is also revealing. Criticism and critic derive from the Greek kritikos, "skillful in judging, decisive." Deconstruction, on the other hand, has no overtones of skill or wisdom; it merely suggests demolition of an existing building. In popular usage criticism suggests censure but not change. If we find fault with a building, we may condemn it, but we do not carry out the demolition ourselves. The deconstructionist, by implication, is both judge and executioner who leaves a text totally dismantled, if not reduced to a pile of rubble.
199502_3-RC_2_13
[ "The author is guardedly optimistic about the ability of deconstruction to reveal the intentions and biases of a writer.", "The author endorses the utility of deconstruction for revealing the role of older meanings of words.", "The author is enthusiastic about the significant neologisms that deconstruction has introduced into literary criticism.", "The author regards deconstruction's tendency to focus only on the problems and faults of literary texts as too mechanical.", "The author condemns deconstruction's attempts to define literary criticism as a creative act." ]
3
The passage suggests that which one of the following most accurately describes the author's view of deconstructionist thought?
The legislature of a country recently considered a bill designed to reduce the uncertainty inherent in the ownership of art by specifying certain conditions that must be met before an allegedly stolen work of art can be reclaimed by a plaintiff. The bill places the burden of proof in reclamation litigation entirely on the plaintiff, who must demonstrate that the holder of an item knew at the time of purchase that it had been stolen. Additionally, the bill creates a uniform national statute of limitations for reclamation of stolen cultural property. Testifying in support of the bill, James D. Burke, a citizen of the country and one of its leading art museum directors, specifically praised the inclusion of a statute of limitations; otherwise, he said, other countries could seek to reclaim valuable art objects, no matter how long they have been held by the current owner or how legitimately they were acquired. Any country could enact a patrimony law stating that anything ever made within the boundaries of that country is its cultural property. Burke expressed the fear that widespread reclamation litigation would lead to ruinous legal defense costs for museums. However, because such reclamation suits have not yet been a problem, there is little basis for Burke's concern. In fact, the proposed legislation would establish too many unjustifiable barriers to the location and recovery of stolen objects. The main barrier is that the bill considers the announcement of an art transaction in a museum publication to be adequate evidence of an attempt to notify a possible owner. There are far too many such publications for the victim of a theft to survey, and with only this form of disclosure, a stolen object could easily remain unlocated even If assiduously searched for. Another stipulation requires that a purchaser show the object to a scholar for verification that it is not stolen, but it is a rare academic who is aware of any but the most publicized art thefts. Moreover, the time limit specified by the statute of limitations is very short, and the requirement that the plaintiff demonstrate that the holder had knowledge of the theft is unrealistic. Typically, stolen art changes hands several times before rising to the level in the marketplace where a curator or collector would see it. At that point, the object bears no trace of the initial transaction between the thief and the first purchaser, perhaps the only one in the chain who knowingly acquired a stolen work of art. Thus, the need for new legislation to protect holders of art is not obvious. Rather, what is necessary is legislation remedying the difficulties that legitimate owners of works of art, and countries from which such works have been stolen, have in locating and reclaiming these stolen works.
199502_3-RC_3_14
[ "Various legal disputes have recently arisen that demonstrate the need for legislation clarifying the legal position of museums in suits involving the repossession of cultural property.", "A bill intended to prevent other governments from recovering cultural property was recently introduced into the legislature of a country at the behest of its museum directors.", "A bill intended to protect good-faith purchasers of works of art from reclamation litigation is unnecessary and fails to address the needs of legitimate owners attempting to recover stolen art works.", "Clashes between museum professionals and members of the academic community regarding governmental legislation of the arts can best be resolved by negotiation and arbitration, not by litigation.", "The desire of some governments to use legislation and litigation to recover cultural property stolen from their counties has led to abuses in international patrimony legislation." ]
2
Which one of the following most accurately summarizes the main point of the passage?
The legislature of a country recently considered a bill designed to reduce the uncertainty inherent in the ownership of art by specifying certain conditions that must be met before an allegedly stolen work of art can be reclaimed by a plaintiff. The bill places the burden of proof in reclamation litigation entirely on the plaintiff, who must demonstrate that the holder of an item knew at the time of purchase that it had been stolen. Additionally, the bill creates a uniform national statute of limitations for reclamation of stolen cultural property. Testifying in support of the bill, James D. Burke, a citizen of the country and one of its leading art museum directors, specifically praised the inclusion of a statute of limitations; otherwise, he said, other countries could seek to reclaim valuable art objects, no matter how long they have been held by the current owner or how legitimately they were acquired. Any country could enact a patrimony law stating that anything ever made within the boundaries of that country is its cultural property. Burke expressed the fear that widespread reclamation litigation would lead to ruinous legal defense costs for museums. However, because such reclamation suits have not yet been a problem, there is little basis for Burke's concern. In fact, the proposed legislation would establish too many unjustifiable barriers to the location and recovery of stolen objects. The main barrier is that the bill considers the announcement of an art transaction in a museum publication to be adequate evidence of an attempt to notify a possible owner. There are far too many such publications for the victim of a theft to survey, and with only this form of disclosure, a stolen object could easily remain unlocated even If assiduously searched for. Another stipulation requires that a purchaser show the object to a scholar for verification that it is not stolen, but it is a rare academic who is aware of any but the most publicized art thefts. Moreover, the time limit specified by the statute of limitations is very short, and the requirement that the plaintiff demonstrate that the holder had knowledge of the theft is unrealistic. Typically, stolen art changes hands several times before rising to the level in the marketplace where a curator or collector would see it. At that point, the object bears no trace of the initial transaction between the thief and the first purchaser, perhaps the only one in the chain who knowingly acquired a stolen work of art. Thus, the need for new legislation to protect holders of art is not obvious. Rather, what is necessary is legislation remedying the difficulties that legitimate owners of works of art, and countries from which such works have been stolen, have in locating and reclaiming these stolen works.
199502_3-RC_3_15
[ "doubt that owners of works of art often harbor over whether individuals have a moral right to possess great art", "concern that owners of works of art often have that their possession of such objects may be legally challenged at any time", "questions that owners of works of art often have concerning the correct identification of the age and origin of their objects", "disputes that often arise between cultural institutions vying for the opportunity to purchase a work of art", "apprehension that owners of works of art often feel concerning the possibility that their objects may be damaged or stolen from them" ]
1
The uncertainty mentioned in line 2 of the passage refers to the
The legislature of a country recently considered a bill designed to reduce the uncertainty inherent in the ownership of art by specifying certain conditions that must be met before an allegedly stolen work of art can be reclaimed by a plaintiff. The bill places the burden of proof in reclamation litigation entirely on the plaintiff, who must demonstrate that the holder of an item knew at the time of purchase that it had been stolen. Additionally, the bill creates a uniform national statute of limitations for reclamation of stolen cultural property. Testifying in support of the bill, James D. Burke, a citizen of the country and one of its leading art museum directors, specifically praised the inclusion of a statute of limitations; otherwise, he said, other countries could seek to reclaim valuable art objects, no matter how long they have been held by the current owner or how legitimately they were acquired. Any country could enact a patrimony law stating that anything ever made within the boundaries of that country is its cultural property. Burke expressed the fear that widespread reclamation litigation would lead to ruinous legal defense costs for museums. However, because such reclamation suits have not yet been a problem, there is little basis for Burke's concern. In fact, the proposed legislation would establish too many unjustifiable barriers to the location and recovery of stolen objects. The main barrier is that the bill considers the announcement of an art transaction in a museum publication to be adequate evidence of an attempt to notify a possible owner. There are far too many such publications for the victim of a theft to survey, and with only this form of disclosure, a stolen object could easily remain unlocated even If assiduously searched for. Another stipulation requires that a purchaser show the object to a scholar for verification that it is not stolen, but it is a rare academic who is aware of any but the most publicized art thefts. Moreover, the time limit specified by the statute of limitations is very short, and the requirement that the plaintiff demonstrate that the holder had knowledge of the theft is unrealistic. Typically, stolen art changes hands several times before rising to the level in the marketplace where a curator or collector would see it. At that point, the object bears no trace of the initial transaction between the thief and the first purchaser, perhaps the only one in the chain who knowingly acquired a stolen work of art. Thus, the need for new legislation to protect holders of art is not obvious. Rather, what is necessary is legislation remedying the difficulties that legitimate owners of works of art, and countries from which such works have been stolen, have in locating and reclaiming these stolen works.
199502_3-RC_3_16
[ "the passage of a law by another country forbidding the future export of any archaeological objects uncovered at sites within its territory", "an international accord establishing strict criteria for determining whether a work of art can be considered stolen and specifying the circumstances under which it must be returned to its country of origin", "the passage of a law by another country declaring that all objects created by its aboriginal people are the sole property of that country", "an increase in the acquisition of culturally significant works of art by private collectors, who are more capable than museums of bearing the cost of litigation but who rarely display their collections to the public", "the recommendation of a United Nations committee studying the problem of art theft that all international sales of cultural property be coordinated by a central regulatory body" ]
2
Which one of the following is an example of the kind of action that Burke feared would pose a serious threat to museums in his country?
The legislature of a country recently considered a bill designed to reduce the uncertainty inherent in the ownership of art by specifying certain conditions that must be met before an allegedly stolen work of art can be reclaimed by a plaintiff. The bill places the burden of proof in reclamation litigation entirely on the plaintiff, who must demonstrate that the holder of an item knew at the time of purchase that it had been stolen. Additionally, the bill creates a uniform national statute of limitations for reclamation of stolen cultural property. Testifying in support of the bill, James D. Burke, a citizen of the country and one of its leading art museum directors, specifically praised the inclusion of a statute of limitations; otherwise, he said, other countries could seek to reclaim valuable art objects, no matter how long they have been held by the current owner or how legitimately they were acquired. Any country could enact a patrimony law stating that anything ever made within the boundaries of that country is its cultural property. Burke expressed the fear that widespread reclamation litigation would lead to ruinous legal defense costs for museums. However, because such reclamation suits have not yet been a problem, there is little basis for Burke's concern. In fact, the proposed legislation would establish too many unjustifiable barriers to the location and recovery of stolen objects. The main barrier is that the bill considers the announcement of an art transaction in a museum publication to be adequate evidence of an attempt to notify a possible owner. There are far too many such publications for the victim of a theft to survey, and with only this form of disclosure, a stolen object could easily remain unlocated even If assiduously searched for. Another stipulation requires that a purchaser show the object to a scholar for verification that it is not stolen, but it is a rare academic who is aware of any but the most publicized art thefts. Moreover, the time limit specified by the statute of limitations is very short, and the requirement that the plaintiff demonstrate that the holder had knowledge of the theft is unrealistic. Typically, stolen art changes hands several times before rising to the level in the marketplace where a curator or collector would see it. At that point, the object bears no trace of the initial transaction between the thief and the first purchaser, perhaps the only one in the chain who knowingly acquired a stolen work of art. Thus, the need for new legislation to protect holders of art is not obvious. Rather, what is necessary is legislation remedying the difficulties that legitimate owners of works of art, and countries from which such works have been stolen, have in locating and reclaiming these stolen works.
199502_3-RC_3_17
[ "commercial dealers in art", "law enforcement officials in his own country", "governments of other countries", "private collectors of art", "museums in other countries" ]
2
According to the passage, Burke envisaged the most formidable potential adversaries of his country's museums in reclamation litigation to be
The legislature of a country recently considered a bill designed to reduce the uncertainty inherent in the ownership of art by specifying certain conditions that must be met before an allegedly stolen work of art can be reclaimed by a plaintiff. The bill places the burden of proof in reclamation litigation entirely on the plaintiff, who must demonstrate that the holder of an item knew at the time of purchase that it had been stolen. Additionally, the bill creates a uniform national statute of limitations for reclamation of stolen cultural property. Testifying in support of the bill, James D. Burke, a citizen of the country and one of its leading art museum directors, specifically praised the inclusion of a statute of limitations; otherwise, he said, other countries could seek to reclaim valuable art objects, no matter how long they have been held by the current owner or how legitimately they were acquired. Any country could enact a patrimony law stating that anything ever made within the boundaries of that country is its cultural property. Burke expressed the fear that widespread reclamation litigation would lead to ruinous legal defense costs for museums. However, because such reclamation suits have not yet been a problem, there is little basis for Burke's concern. In fact, the proposed legislation would establish too many unjustifiable barriers to the location and recovery of stolen objects. The main barrier is that the bill considers the announcement of an art transaction in a museum publication to be adequate evidence of an attempt to notify a possible owner. There are far too many such publications for the victim of a theft to survey, and with only this form of disclosure, a stolen object could easily remain unlocated even If assiduously searched for. Another stipulation requires that a purchaser show the object to a scholar for verification that it is not stolen, but it is a rare academic who is aware of any but the most publicized art thefts. Moreover, the time limit specified by the statute of limitations is very short, and the requirement that the plaintiff demonstrate that the holder had knowledge of the theft is unrealistic. Typically, stolen art changes hands several times before rising to the level in the marketplace where a curator or collector would see it. At that point, the object bears no trace of the initial transaction between the thief and the first purchaser, perhaps the only one in the chain who knowingly acquired a stolen work of art. Thus, the need for new legislation to protect holders of art is not obvious. Rather, what is necessary is legislation remedying the difficulties that legitimate owners of works of art, and countries from which such works have been stolen, have in locating and reclaiming these stolen works.
199502_3-RC_3_18
[ "less common than Burke fears it will become without passage of a national statute of limitations for reclamation of stolen cultural property", "increasing as a result of the passage of legislation that aids legitimate owners of art in their attempts to recover stolen works", "a serious threat to museums and cultural institutions that have unwittingly added stolen artifacts to their collections", "a signal of the legitimate frustrations of victims of art theft", "increasing as a result of an increase in the amount of art theft" ]
0
The author suggests that in the country mentioned in line 1, litigation involving the reclamation of stolen works of art has been